Kaiseem's Stars
OptimalScale/LMFlow
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.
FedML-AI/FedML
FEDML - The unified and scalable ML library for large-scale distributed training, model serving, and federated learning. FEDML Launch, a cross-cloud scheduler, further enables running any AI jobs on any GPU cloud or on-premise cluster. Built on this library, TensorOpera AI (https://TensorOpera.ai) is your generative AI platform at scale.
neuralmagic/deepsparse
Sparsity-aware deep learning inference runtime for CPUs
jiaweizzhao/GaLore
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
alibaba/FederatedScope
An easy-to-use federated learning platform
HuangOwen/Awesome-LLM-Compression
Awesome LLM compression research papers and tools.
AGI-Edgerunners/LLM-Adapters
Code for our EMNLP 2023 Paper: "LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models"
princeton-nlp/MeZO
[NeurIPS 2023] MeZO: Fine-Tuning Language Models with Just Forward Passes. https://arxiv.org/abs/2305.17333
ThuCCSLab/Awesome-LM-SSP
A reading list for large models safety, security, and privacy (including Awesome LLM Security, Safety, etc.).
OpenLMLab/LOMO
LOMO: LOw-Memory Optimization
SafeAILab/EAGLE
Official Implementation of EAGLE-1 (ICML'24) and EAGLE-2 (EMNLP'24)
YuxinWenRick/hard-prompts-made-easy
intel/xFasterTransformer
weavel-ai/Ape
Your first AI prompt engineer
rui-ye/OpenFedLLM
princeton-nlp/AutoCompressors
[EMNLP 2023] Adapting Language Models to Compress Long Contexts
jeffreysijuntan/lloco
The official repo for "LLoCo: Learning Long Contexts Offline"
Hannibal046/xRAG
[Neurips2024] Source code for xRAG: Extreme Context Compression for Retrieval-augmented Generation with One Token
AvivNavon/DWSNets
Official implementation for Equivariant Architectures for Learning in Deep Weight Spaces [ICML 2023]
shizhediao/Black-Box-Prompt-Learning
Source code for the TMLR paper "Black-Box Prompt Learning for Pre-trained Language Models"
declare-lab/della
DELLA-Merging: Reducing Interference in Model Merging through Magnitude-Based Sampling
uiuctml/Localize-and-Stitch
Localize-and-Stitch: Efficient Model Merging via Sparse Task Arithmetic
MrGGLS/BlockPruner
A block pruning framework for LLMs.
arumaekawa/DiLM
Implementaiton of "DiLM: Distilling Dataset into Language Model for Text-level Dataset Distillation" (accepted by NAACL2024 Findings)".
arumaekawa/text-dataset-distillation
Dereck0602/Bolaco
amazon-science/mezo_svrg
Code the ICML 2024 paper: "Variance-reduced Zeroth-Order Methods for Fine-Tuning Language Models"
Eyr3/PrivacyAsst
PrivacyAsst: Safeguarding User Privacy in Tool-Using Large Language Model Agents (TDSC 2024)
Kaiseem/IST
Official implement for EMNLP24 "Layer-wise Importance Matters: Less Memory for Better Performance in Parameter-efficient Fine-tuning of Large Language Models"
kiwi12138/kiwi12138.github.io