smilelite's Stars
showlab/Awesome-MLLM-Hallucination
📖 A curated list of resources dedicated to hallucination of multimodal large language models (MLLM).
PKU-Alignment/safe-rlhf
Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
trotsky1997/MathBlackBox
hsiehjackson/RULER
This repo contains the source code for RULER: What’s the Real Context Size of Your Long-Context Language Models?
MozerWang/Loong
[EMNLP 2024 Main]Leave No Document Behind: Benchmarking Long-Context LLMs with Extended Multi-Doc QA
Xnhyacinth/Awesome-LLM-Long-Context-Modeling
📰 Must-read papers and blogs on LLM based Long Context Modeling 🔥
TIGER-AI-Lab/LongICLBench
Code and Data for "Long-context LLMs Struggle with Long In-context Learning"
THUDM/LongAlign
[EMNLP 2024] LongAlign: A Recipe for Long Context Alignment of LLMs
DefTruth/Awesome-LLM-Inference
📖A curated list of Awesome LLM Inference Paper with codes, TensorRT-LLM, vLLM, streaming-llm, AWQ, SmoothQuant, WINT8/4, Continuous Batching, FlashAttention, PagedAttention etc.
2noise/ChatTTS
A generative speech model for daily dialogue.
zhuzilin/ring-flash-attention
Ring attention implementation with flash attention
feifeibear/long-context-attention
USP: Unified (a.k.a. Hybrid, 2D) Sequence Parallel Attention for Long Context Transformers Model Training and Inference
NVIDIA/Megatron-LM
Ongoing research training transformer models at scale
jzhang38/EasyContext
Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
huggingface/alignment-handbook
Robust recipes to align language models with human and AI preferences
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
hpcaitech/ColossalAI
Making large AI models cheaper, faster and more accessible
infinigence/LVEval
Repository of LV-Eval Benchmark
InternLM/InternLM
Official release of InternLM2.5 base and chat models. 1M context support
owenliang/qwen-vllm
通义千问VLLM推理部署DEMO
wuhy68/Parameter-Efficient-MoE
Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks
HIT-SCIR/Chinese-Mixtral-8x7B
中文Mixtral-8x7B(Chinese-Mixtral-8x7B)
hiyouga/LLaMA-Factory
Efficiently Fine-Tune 100+ LLMs in WebUI (ACL 2024)
modelscope/ms-swift
Use PEFT or Full-parameter to finetune 350+ LLMs or 90+ MLLMs. (LLM: Qwen2.5, Llama3.2, GLM4, Internlm2.5, Yi1.5, Mistral, Baichuan2, DeepSeek, Gemma2, ...; MLLM: Qwen2-VL, Qwen2-Audio, Llama3.2-Vision, Llava, InternVL2, MiniCPM-V-2.6, GLM4v, Xcomposer2.5, Yi-VL, DeepSeek-VL, Phi3.5-Vision, ...)
jquesnelle/yarn
YaRN: Efficient Context Window Extension of Large Language Models
HKUNLP/ChunkLlama
[ICML'24] Data and code for our paper "Training-Free Long-Context Scaling of Large Language Models"
HqWu-HITCS/Awesome-Chinese-LLM
整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。
jackaduma/awesome_LLMs_interview_notes
LLMs interview notes and answers:该仓库主要记录大模型(LLMs)算法工程师相关的面试题和参考答案
amusi/CVPR2024-Papers-with-Code
CVPR 2024 论文和开源项目合集
HCIILAB/Scene-Text-Recognition-Recommendations
Papers, Datasets, Algorithms, SOTA for STR. Long-time Maintaining