RifleZhang's Stars
langchain-ai/langchain
🦜🔗 Build context-aware reasoning applications
meta-llama/llama3
The official Meta Llama 3 GitHub site
stanfordnlp/dspy
DSPy: The framework for programming—not prompting—language models
haotian-liu/LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
OpenBMB/MiniCPM-o
MiniCPM-o 2.6: A GPT-4o Level MLLM for Vision, Speech and Multimodal Live Streaming on Your Phone
LLaVA-VL/LLaVA-NeXT
InternLM/InternLM-XComposer
InternLM-XComposer2.5-OmniLive: A Comprehensive Multimodal System for Long-term Streaming Video and Audio Interactions
openai/simple-evals
EvolvingLMMs-Lab/lmms-eval
Accelerating the development of large multimodal models (LMMs) with one-click evaluation module - lmms-eval.
HazyResearch/ThunderKittens
Tile primitives for speedy kernels
baaivision/Emu3
Next-Token Prediction is All You Need
GAIR-NLP/O1-Journey
O1 Replication Journey
PKU-YuanGroup/LLaVA-CoT
LLaVA-CoT, a visual language model capable of spontaneous, systematic reasoning
laekov/fastmoe
A fast MoE impl for PyTorch
Zefan-Cai/KVCache-Factory
Unified KV Cache Compression Methods for Auto-Regressive Models
BradyFU/Video-MME
✨✨[CVPR 2025] Video-MME: The First-Ever Comprehensive Evaluation Benchmark of Multi-modal LLMs in Video Analysis
xiaoachen98/Open-LLaVA-NeXT
An open-source implementation for training LLaVA-NeXT.
RLHF-V/RLAIF-V
[CVPR'25] RLAIF-V: Open-Source AI Feedback Leads to Super GPT-4V Trustworthiness
RLHF-V/RLHF-V
[CVPR'24] RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback
TideDra/VL-RLHF
A RLHF Infrastructure for Vision-Language Models
NVIDIA/ngpt
Normalized Transformer (nGPT)
ByungKwanLee/Meteor
[NeurIPS 2024] Official PyTorch implementation code for realizing the technical part of Mamba-based traversal of rationale (Meteor) to improve performance of numerous vision language performances for diverse capabilities.
thu-pacman/FasterMoE
luyug/magix
Supercharge huggingface transformers with model parallelism.
hamishivi/EasyLM
Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Flax.
RifleZhang/LLaVA-Reasoner-DPO
HazyResearch/train-tk
train with kittens!
shawntan/stickbreaking-attention
Stick-breaking attention
hyhieu/easy_pybind
princeton-nlp/PTP
Improving Language Understanding from Screenshots. Paper: https://arxiv.org/abs/2402.14073