jeesoobang's Stars
hkust-nlp/deita
Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]
princeton-nlp/SimPO
SimPO: Simple Preference Optimization with a Reference-Free Reward
NVIDIA/NeMo-Aligner
Scalable toolkit for efficient model alignment
RLHFlow/RLHF-Reward-Modeling
Recipes to train reward model for RLHF.
huggingface/datatrove
Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.
huggingface/data-is-better-together
Let's build better datasets, together!
NVIDIA/TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
prometheus-eval/prometheus-eval
Evaluate your LLM's response with Prometheus and GPT4 💯
declare-lab/instruct-eval
This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.
ytyz1307zzh/PLUG
Code for the ACL 2024 paper "PLUG: Leveraging Pivot Language in Cross-Lingual Instruction Tuning"
meta-llama/llama-recipes
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama for WhatsApp & Messenger.
mosaicml/llm-foundry
LLM training code for Databricks foundation models
tatsu-lab/alpaca_eval
An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.
allenai/open-instruct
kaistAI/FLASK
[ICLR 2024 Spotlight] FLASK: Fine-grained Language Model Evaluation based on Alignment Skill Sets
tcapelle/llm_recipes
A set of scripts and notebooks on LLM finetunning and dataset creation
togethercomputer/Llama-2-7B-32K-Instruct
unslothai/unsloth
Finetune Llama 3.1, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory
xingyaoww/mint-bench
Official Repo for ICLR 2024 paper MINT: Evaluating LLMs in Multi-turn Interaction with Tools and Language Feedback by Xingyao Wang*, Zihan Wang*, Jiateng Liu, Yangyi Chen, Lifan Yuan, Hao Peng and Heng Ji.
HeegyuKim/open-korean-instructions
언어모델을 학습하기 위한 공개 한국어 instruction dataset들을 모아두었습니다.
jianzhnie/awesome-instruction-datasets
A collection of awesome-prompt-datasets, awesome-instruction-dataset, to train ChatLLM such as chatgpt 收录各种各样的指令数据集, 用于训练 ChatLLM 模型。
huggingface/peft
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
meta-llama/llama
Inference code for Llama models
microsoft/unilm
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
THUDM/LongBench
[ACL 2024] LongBench: A Bilingual, Multitask Benchmark for Long Context Understanding
mathllm/MathCoder
Family of LLMs for mathematical reasoning.
IBM/Dromedary
Dromedary: towards helpful, ethical and reliable LLMs.
atfortes/Awesome-LLM-Reasoning
Reasoning in Large Language Models: Papers and Resources, including Chain-of-Thought and OpenAI o1 🍓
aiwaves-cn/agents
An Open-source Framework for Data-centric, Self-evolving Autonomous Language Agents
nlpxucan/WizardLM
LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMath