intfloat's Stars
xai-org/grok-1
Grok open release
microsoft/autogen
A programming framework for agentic AI 🤖 PyPi: autogen-agentchat Discord: https://aka.ms/autogen-discord Office Hour: https://aka.ms/autogen-officehour
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
unslothai/unsloth
Finetune Llama 3.3, Mistral, Phi, Qwen 2.5 & Gemma LLMs 2-5x faster with 70% less memory
ItzCrazyKns/Perplexica
Perplexica is an AI-powered search engine. It is an Open source alternative to Perplexity AI
openai/swarm
Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team.
microsoft/BitNet
Official inference framework for 1-bit LLMs
LargeWorldModel/LWM
Large World Model -- Modeling Text and Video with Millions Context
hijkzzz/Awesome-LLM-Strawberry
A collection of LLM papers, blogs, and projects, with a focus on OpenAI o1 🍓 and reasoning techniques.
google/gemma_pytorch
The official PyTorch implementation of Google's Gemma models
huggingface/alignment-handbook
Robust recipes to align language models with human and AI preferences
mit-han-lab/llm-awq
[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
facebookresearch/chameleon
Repository for Meta Chameleon, a mixed-modal early-fusion foundation model from FAIR.
gkamradt/LLMTest_NeedleInAHaystack
Doing simple retrieval from LLM models at various context lengths to measure accuracy
RUC-NLPIR/FlashRAG
⚡FlashRAG: A Python Toolkit for Efficient RAG Research
facebookresearch/MetaCLIP
ICLR2024 Spotlight: curation/training code, metadata, distribution and pre-trained models for MetaCLIP; CVPR 2024: MoDE: CLIP Data Experts via Clustering
Xnhyacinth/Awesome-LLM-Long-Context-Modeling
📰 Must-read papers and blogs on LLM based Long Context Modeling 🔥
jxzhangjhu/Awesome-LLM-RAG
Awesome-LLM-RAG: a curated list of advanced retrieval augmented generation (RAG) in Large Language Models
srush/awesome-o1
A bibliography and survey of the papers surrounding o1
mistralai/megablocks-public
NVIDIA/RULER
This repo contains the source code for RULER: What’s the Real Context Size of Your Long-Context Language Models?
jzhang38/EasyContext
Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
haoliuhl/ringattention
Large Context Attention
huggingface/huggingface-llama-recipes
ContextualAI/gritlm
Generative Representational Instruction Tuning
nomic-ai/contrastors
Train Models Contrastively in Pytorch
google-deepmind/loft
LOFT: A 1 Million+ Token Long-Context Benchmark
princeton-nlp/ProLong
Homepage for ProLong (Princeton long-context language models) and paper "How to Train Long-Context Language Models (Effectively)"
microsoft/SmartPlay
SmartPlay is a benchmark for Large Language Models (LLMs). Uses a variety of games to test various important LLM capabilities as agents. SmartPlay is designed to be easy to use, and to support future development of LLMs.
dwzhu-pku/LongEmbed
LongEmbed: Extending Embedding Models for Long Context Retrieval (EMNLP 2024)