xinlong-yang's Stars
QwenLM/Qwen2-VL
Qwen2-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.
QwenLM/Qwen-VL
The official repo of Qwen-VL (通义千问-VL) chat & pretrained large vision language model proposed by Alibaba Cloud.
UbiquitousLearning/Efficient_Foundation_Model_Survey
Survey Paper List - Efficient LLM and Foundation Models
mit-han-lab/smoothquant
[ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
mit-han-lab/duo-attention
DuoAttention: Efficient Long-Context LLM Inference with Retrieval and Streaming Heads
facebookresearch/lingua
Meta Lingua: a lean, efficient, and easy-to-hack codebase to research LLMs.
alibaba/EasyRec
A framework for large scale recommendation algorithms.
Infini-AI-Lab/TriForce
[COLM 2024] TriForce: Lossless Acceleration of Long Sequence Generation with Hierarchical Speculative Decoding
zhentingqi/rStar
THUDM/ReST-MCTS
ReST-MCTS*: LLM Self-Training via Process Reward Guided Tree Search (NeurIPS 2024)
feifeibear/LLMSpeculativeSampling
Fast inference from large lauguage models via speculative decoding
hijkzzz/Awesome-LLM-Strawberry
A collection of LLM papers, blogs, and projects, with a focus on OpenAI o1 and reasoning techniques.
haroldsultan/MCTS
Python Implementations of Monte Carlo Tree Search
YunjiaXi/DARE_code
EurekaLabsAI/micrograd
The Autograd Engine
magpie-align/magpie
Official repository for "Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing". Your efficient and high-quality synthetic data generation pipeline!
JackHCC/PKU-Lessons-Summary
北京大学软件与微电子学院硕士生课程知识点、作业等汇总【Summary of Knowledge Points and Assignments of Peking University Integrated Circuit Major Courses】
AMD-AIG-AIMA/AMD-LLM
QwenLM/Qwen2.5
Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.
GATECH-EIC/Linearized-LLM
[ICML 2024] When Linear Attention Meets Autoregressive Decoding: Towards More Effective and Efficient Linearized Large Language Models
EurekaLabsAI/tensor
The Tensor (or Array)
Equationliu/Kangaroo
Implementation of Kangaroo: Lossless Self-Speculative Decoding via Double Early Exiting
ShiArthur03/ShiArthur03
astramind-ai/Mixture-of-depths
Unofficial implementation for the paper "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"
karpathy/minbpe
Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.
GAIR-NLP/anole
Anole: An Open, Autoregressive and Native Multimodal Models for Interleaved Image-Text Generation
xinlong-yang/Noise_Dense_Retrieval
[ICCV2023] Prototypical Mixing and Retrieval-based Refinement for Label Noise-resistant Image Retrieval
karpathy/LLM101n
LLM101n: Let's build a Storyteller
facebookresearch/chameleon
Repository for Meta Chameleon, a mixed-modal early-fusion foundation model from FAIR.
WillDreamer/LOG