tzt101's Stars
pprp/Awesome-LLM-Prune
Awesome list for LLM pruning.
microsoft/LMOps
General technology for enabling AI capabilities w/ LLMs and MLLMs
EleutherAI/the-pile
thoppe/The-Pile-PubMed
Download, parse, and filter data PubMed, data-ready for The-Pile
EleutherAI/pile-pubmedcentral
A script for collecting the PubMed Central dataset in a language modelling friendly format.
NVIDIA/RULER
This repo contains the source code for RULER: What’s the Real Context Size of Your Long-Context Language Models?
LAION-AI/Open-Instruction-Generalist
Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks
microsoft/Phi-3CookBook
This is a Phi-3 book for getting started with Phi-3. Phi-3, a family of open sourced AI models developed by Microsoft. Phi-3 models are the most capable and cost-effective small language models (SLMs) available, outperforming models of the same size and next size up across a variety of language, reasoning, coding, and math benchmarks.
microsoft/rho
Repo for Rho-1: Token-level Data Selection & Selective Pretraining of LLMs.
nikhil-ghosh-berkeley/loraplus
tuna/thuthesis
LaTeX Thesis Template for Tsinghua University
eric-mitchell/direct-preference-optimization
Reference implementation for DPO (Direct Preference Optimization)
tianyi-lab/Mosaic-IT
Mosaic IT: Enhancing Instruction Tuning with Data Mosaics
meta-llama/llama-recipes
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama for WhatsApp & Messenger.
meta-llama/PurpleLlama
Set of tools to assess and improve LLM security.
pytorch/torchtune
PyTorch native finetuning library
kongds/MoRA
MoRA: High-Rank Updating for Parameter-Efficient Fine-Tuning
deepseek-ai/DeepSeek-V2
DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model
ZJLab-DataHub-Security/LLaMA-Factory
Unify Efficient Fine-Tuning of 100+ LLMs
ymcui/Chinese-LLaMA-Alpaca-3
中文羊驼大模型三期项目 (Chinese Llama-3 LLMs) developed from Meta Llama 3
bigcode-project/bigcode-evaluation-harness
A framework for the evaluation of autoregressive code generation language models.
EleutherAI/lm-evaluation-harness
A framework for few-shot evaluation of language models.
jzhang38/TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
apple/corenet
CoreNet: A library for training deep neural networks
open-compass/opencompass
OpenCompass is an LLM evaluation platform, supporting a wide range of models (Llama3, Mistral, InternLM2,GPT-4,LLaMa2, Qwen,GLM, Claude, etc) over 100+ datasets.
LianjiaTech/BELLE
BELLE: Be Everyone's Large Language model Engine(开源中文对话大模型)
tatsu-lab/stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
ymcui/Chinese-LLaMA-Alpaca
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
ArtificialZeng/Qwen-Tuning
Qwen-Efficient-Tuning
shikiw/OPERA
[CVPR 2024 Highlight] OPERA: Alleviating Hallucination in Multi-Modal Large Language Models via Over-Trust Penalty and Retrospection-Allocation