Pinned Repositories
awesome-llm-interpretability
A curated list of Large Language Model (LLM) Interpretability resources.
bonito
A lightweight library for generating synthetic instruction tuning datasets for your data without GPT.
distilling-step-by-step
flores
Facebook Low Resource (FLoRes) MT Benchmark
hallucination-leaderboard
Leaderboard Comparing LLM Performance at Producing Hallucinations when Summarizing Short Documents
Jamba
PyTorch Implementation of Jamba: "Jamba: A Hybrid Transformer-Mamba Language Model"
Ladder-Side-Tuning
PyTorch codes for "LST: Ladder Side-Tuning for Parameter and Memory Efficient Transfer Learning"
llm_distillation_playbook
Best practices for distilling large language models.
mamba.py
A Mamba with parallel scan in PyTorch.
MeZO
[NeurIPS 2023] MeZO: Fine-Tuning Language Models with Just Forward Passes. https://arxiv.org/abs/2305.17333
ColdFusion2001's Repositories
ColdFusion2001/awesome-llm-interpretability
A curated list of Large Language Model (LLM) Interpretability resources.
ColdFusion2001/bonito
A lightweight library for generating synthetic instruction tuning datasets for your data without GPT.
ColdFusion2001/distilling-step-by-step
ColdFusion2001/flores
Facebook Low Resource (FLoRes) MT Benchmark
ColdFusion2001/hallucination-leaderboard
Leaderboard Comparing LLM Performance at Producing Hallucinations when Summarizing Short Documents
ColdFusion2001/Jamba
PyTorch Implementation of Jamba: "Jamba: A Hybrid Transformer-Mamba Language Model"
ColdFusion2001/Ladder-Side-Tuning
PyTorch codes for "LST: Ladder Side-Tuning for Parameter and Memory Efficient Transfer Learning"
ColdFusion2001/llm_distillation_playbook
Best practices for distilling large language models.
ColdFusion2001/mamba.py
A Mamba with parallel scan in PyTorch.
ColdFusion2001/MeZO
[NeurIPS 2023] MeZO: Fine-Tuning Language Models with Just Forward Passes. https://arxiv.org/abs/2305.17333
ColdFusion2001/openchat
OpenChat: Advancing Open-source Language Models with Imperfect Data
ColdFusion2001/OpenMoE
A family of open-sourced Mixture-of-Experts (MoE) Large Language Models
ColdFusion2001/sae
Sparse autoencoders
ColdFusion2001/schedule_free
Schedule-Free Optimization in PyTorch
ColdFusion2001/snowflake-arctic
ColdFusion2001/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
ColdFusion2001/trl
Train transformer language models with reinforcement learning.
ColdFusion2001/xlstm-cuda
Cuda implementation of Extended Long Short Term Memory (xLSTM) with C++ and PyTorch ports