mihara-bot's Stars
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
QwenLM/Qwen2.5
Qwen2.5 is the large language model series developed by Qwen team, Alibaba Cloud.
THUDM/GLM-4
GLM-4 series: Open Multilingual Multimodal Chat LMs | 开源多语言多模态对话模型
togethercomputer/RedPajama-Data
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
eliahuhorwitz/Academic-project-page-template
A project page template for academic papers. Demo at https://eliahuhorwitz.github.io/Academic-project-page-template/
huggingface/datatrove
Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.
huggingface/nanotron
Minimalistic large language model 3D-parallelism training
mlfoundations/dclm
DataComp for Language Models
huggingface/lighteval
Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backends
CarperAI/OpenELM
Evolution Through Large Models
hkust-nlp/deita
Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]
mlfoundations/open_lm
A repository for research on medium sized language models.
huggingface/cosmopedia
microsoft/rho
Repo for Rho-1: Token-level Data Selection & Selective Pretraining of LLMs.
eth-sri/language-model-arithmetic
Controlled Text Generation via Language Model Arithmetic
tomekkorbak/pretraining-with-human-feedback
Code accompanying the paper Pretraining Language Models with Human Preferences
SALT-NLP/demonstrated-feedback
msclar/formatspread
Code accompanying "How I learned to start worrying about prompt formatting".
yifanzhang-pro/AutoMathText
Official implementation of paper "Autonomous Data Selection with Language Models for Mathematical Texts" (As Huggingface Daily Papers: https://huggingface.co/papers/2402.07625)
locuslab/scaling_laws_data_filtering
skzhang1/IDEAL
IDEAL: Influence-Driven Selective Annotations Empower In-Context Learners in Large Language Models
cxcscmu/MATES
Official repository for MATES: Model-Aware Data Selection for Efficient Pretraining with Data Influence Models [NeurIPS 2024]
adymaharana/d2pruning
UCSB-NLP-Chang/llm_uncertainty
cohere-ai/human-feedback-paper
Code and data from the paper 'Human Feedback is not Gold Standard'
daeveraert/gradient-information-optimization
Implementation of Gradient Information Optimization (GIO) for effective and scalable training data selection
kothasuhas/understanding-forgetting
Understanding Catastrophic Forgetting in Language Models via Implicit Inference
zijian678/TDD
luffy06/ReFusion
[ICLR 2024] ReFusion: Improving Natural Language Understanding with Computation-Efficient Retrieval Representation Fusion
xlhex/acl2024_xicl