Zeng-WH's Stars
Significant-Gravitas/AutoGPT
AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.
hiyouga/LLaMA-Factory
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
THUDM/ChatGLM2-6B
ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型
jzhang38/TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
OpenRLHF/OpenRLHF
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT)
hkust-nlp/simpleRL-reason
Simple RL training for reasoning
allenai/open-instruct
AllenAI's post-training codebase
PhoebusSi/Alpaca-CoT
We unified the interfaces of instruction-tuning data (e.g., CoT data), multiple LLMs and parameter-efficient methods (e.g., lora, p-tuning) together for easy use. We welcome open-source enthusiasts to initiate any meaningful PR on this repo and integrate as many LLM related technologies as possible. 我们打造了方便研究人员上手和使用大模型等微调平台,我们欢迎开源爱好者发起任何有意义的pr!
thunlp/UltraChat
Large-scale, Informative, and Diverse Multi-round Chat Data (and Models)
RayeRen/acad-homepage.github.io
AcadHomepage: A Modern and Responsive Academic Personal Homepage
tatsu-lab/alpaca_eval
An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.
AetherCortex/Llama-X
Open Academic Research on Improving LLaMA to SOTA LLM
srush/awesome-o1
A bibliography and survey of the papers surrounding o1
GAIR-NLP/factool
FacTool: Factuality Detection in Generative AI
hkust-nlp/deita
Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]
salesforce/DialogStudio
DialogStudio: Towards Richest and Most Diverse Unified Dataset Collection and Instruction-Aware Models for Conversational AI
allenai/FineGrainedRLHF
OFA-Sys/InsTag
InsTag: A Tool for Data Analysis in LLM Supervised Fine-tuning
arcee-ai/EvolKit
EvolKit is an innovative framework designed to automatically enhance the complexity of instructions used for fine-tuning Large Language Models (LLMs).
Edward-Sun/easy-to-hard
Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision
CJReinforce/PURE
SOTA RL fine-tuning solution for advanced math reasoning of LLM
hkust-nlp/B-STaR
B-STAR: Monitoring and Balancing Exploration and Exploitation in Self-Taught Reasoners
Edward-Sun/gpt-accelera
Simple and efficient pytorch-native transformer training and inference (batched)
OpenLMLab/MOSS_WebSearchTool
MOSS 003 WebSearchTool: A simple but reliable implementation
csbench/csbench
HypherX/Evolution-Analysis
dongguanting/SKP-for-KBQA
The code of CIKM 2023 short paper : Bridging the KB-Text Gap: Leveraging Structured Knowledge-aware Pre-training for KBQA
alycialee/beyond-scale-language-data-diversity
mjain72/brainwave
The code in this repository allows the visualization of brainwaves detected using TGAM brainwave sensor module. It uses python, d3.js and Bootstrap. A detailed description of the code can be found in this article - https://medium.com/@mohitjain72/detection-and-visualization-of-brainwaves-using-python-d3-and-bootstrap-742129f9ed97
Zeng-WH/FutureTOD
Official codes for ACL 2023 paper "FutureTOD: Teaching Future Knowledge to Pre-trained Language Model for Task-Oriented Dialogue"