Shawn-Guo-CN
Fourth-year CDT-NLP student at University of Edinburgh.
School of Informatics, University of EdinburghEdinburgh, UK
Shawn-Guo-CN's Stars
codecrafters-io/build-your-own-x
Master programming by recreating your favorite technologies from scratch.
xai-org/grok-1
Grok open release
hpcaitech/Open-Sora
Open-Sora: Democratizing Efficient Video Production for All
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
alshedivat/al-folio
A beautiful, simple, clean, and responsive Jekyll theme for academics
NVIDIA/Megatron-LM
Ongoing research training transformer models at scale
huggingface/trl
Train transformer language models with reinforcement learning.
mistralai/mistral-inference
Official inference library for Mistral models
leptonai/search_with_lepton
Building a quick conversation-based search demo with Lepton AI.
huggingface/alignment-handbook
Robust recipes to align language models with human and AI preferences
BAAI-Agents/Cradle
The Cradle framework is a first attempt at General Computer Control (GCC). Cradle supports agents to ace any computer task by enabling strong reasoning abilities, self-improvment, and skill curation, in a standardized general environment with minimal requirements.
gkamradt/LLMTest_NeedleInAHaystack
Doing simple retrieval from LLM models at various context lengths to measure accuracy
ZhiningLiu1998/awesome-imbalanced-learning
😎 Everything about class-imbalanced/long-tail learning: papers, codes, frameworks, and libraries | 有关类别不平衡/长尾学习的一切:论文、代码、框架与库
RLHFlow/RLHF-Reward-Modeling
Recipes to train reward model for RLHF.
ContextualAI/HALOs
A library with extensible implementations of DPO, KTO, PPO, ORPO, and other human-aware loss functions (HALOs).
jzhang38/EasyContext
Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
bojone/bytepiece
更纯粹、更高压缩率的Tokenizer
123penny123/Awesome-LLM-RL
A comprehensive list of PAPERS, CODEBASES, and, DATASETS on Decision Making using Foundation Models including LLMs and VLMs.
LLaMafia/llamafia.github
sanderwood/bgpt
Beyond Language Models: Byte Models are Digital World Simulators
sangmichaelxie/doremi
Pytorch implementation of DoReMi, a method for optimizing the data mixture weights in language modeling datasets
rwitten/HighPerfLLMs2024
allenai/WildBench
Benchmarking LLMs with Challenging Tasks from Real Users
zorazrw/awesome-tool-llm
Edward-Sun/easy-to-hard
Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision
kddubey/cappr
Completion After Prompt Probability. Make your LLM make a choice
facebookresearch/LIGHT
LIGHT is a platform for text-situated dialogue research. We originally hosted LIGHT as a live game with dialogue models in a grounded setting. This repo contains all of the code to get the LIGHT game running, as well as reproducible code for the research projects along the way of getting LIGHT to where it was.
jiahe7ay/infini-mini-transformer
This is a personal reimplementation of Google's Infini-transformer, utilizing a small 2b model. The project includes both model and training code.
allenai/easy-to-hard-generalization
Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"
LLaMafia/SFT_function_learning
Explore what LLMs are really leanring over SFT