tongyao-zhu's Stars
meta-llama/llama-recipes
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama for WhatsApp & Messenger.
RUCAIBox/LLMSurvey
The official GitHub page for the survey paper "A Survey of Large Language Models".
arcee-ai/mergekit
Tools for merging pretrained large language models.
open-compass/opencompass
OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.
microsoft/LMOps
General technology for enabling AI capabilities w/ LLMs and MLLMs
lucidrains/vector-quantize-pytorch
Vector (and Scalar) Quantization, in Pytorch
RUC-NLPIR/FlashRAG
⚡FlashRAG: A Python Toolkit for Efficient RAG Research
Xnhyacinth/Awesome-LLM-Long-Context-Modeling
📰 Must-read papers and blogs on LLM based Long Context Modeling 🔥
THUDM/LongBench
[ACL 2024] LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding
datamllab/LongLM
[ICML'24 Spotlight] LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning
texttron/tevatron
Tevatron - A flexible toolkit for neural retrieval research and development.
catid/self-discover
Implementation of Google's SELF-DISCOVER
neulab/knn-transformers
PyTorch + HuggingFace code for RetoMaton: "Neuro-Symbolic Language Modeling with Automaton-augmented Retrieval" (ICML 2022), including an implementation of kNN-LM and kNN-MT
booydar/babilong
BABILong is a benchmark for LLM evaluation using the needle-in-a-haystack approach.
chentong0/factoid-wiki
Dense X Retrieval: What Retrieval Granularity Should We Use?
RulinShao/retrieval-scaling
Official repository for "Scaling Retrieval-Based Langauge Models with a Trillion-Token Datastore".
sail-sg/sailor-llm
[EMNLP-2024] ⚓️ Sailor: Open Language Models for South-East Asia
jxmorris12/bm25_pt
minimal pytorch implementation of bm25 (with sparse tensors)
nkandpa2/long_tail_knowledge
Repo for the paper "Large Language Models Struggle to Learn Long-Tail Knowledge"
joeljang/temporalwiki
[EMNLP 2022] TemporalWiki: A Lifelong Benchmark for Training and Evaluating Ever-Evolving Language Models
frankxu2004/knnlm-why
Repo for ICML23 "Why do Nearest Neighbor Language Models Work?"
October2001/ProLong
[ACL 2024 (Oral)] A Prospector of Long-Dependency Data for Large Language Models
bbuing9/ICLR24_SuRe
Official Code for the paper "SuRe: Summarizing Retrievals using Answer Candidates for Open-domain QA of LLMs" (ICLR 2024)
yuzhaouoe/pretraining-data-packing
[ACL'24 Oral] Analysing The Impact of Sequence Composition on Language Model Pre-Training
yasumasaonoe/entity_knowledge_propagation
YisongMiao/DiSQ-Score
The Dataset and Official Implementation for <Discursive Socratic Questioning: Evaluating the Faithfulness of Language Models’ Understanding of Discourse Relations> @ ACL 2024
trestad/mitigating-reversal-curse
Code for paper 'Are We Falling in a Middle-Intelligence Trap? An Analysis and Mitigation of the Reversal Curse'
amy-hyunji/lora-for-retrieval
xiangyue9607/C-MORE
Code for the ACL2022 paper "C-MORE: Pretraining to Answer Open-Domain Questions by Consulting Millions of References"
Fantabulous-J/Self-Training-DPR