pskun's Stars
GanjinZero/RRHF
[NIPS2023] RRHF & Wombat
microsoft/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Instruction-Tuning-with-GPT-4/GPT-4-LLM
Instruction Tuning with GPT-4
AI4Finance-Foundation/FinNLP
Democratizing Internet-scale financial data.
Tencent/TencentPretrain
Tencent Pre-training framework in PyTorch & Pre-trained Model Zoo
CVI-SZU/Linly
Chinese-LLaMA 1&2、Chinese-Falcon 基础模型;ChatFlow中文对话模型;中文OpenLLaMA模型;NLP预训练/指令微调数据集
IllDepence/unarXive
A data set based on all arXiv publications, pre-processed for NLP, including structured full-text and citation network
radi-cho/datasetGPT
A command-line interface to generate textual and conversational datasets with LLMs.
radi-cho/botbots
A dataset featuring diverse dialogues between two ChatGPT (gpt-3.5-turbo) instances with system messages written by GPT-4. Covering various contexts and tasks (task-oriented dialogue systems, abstract reasoning, brainstorming).
yaodongC/awesome-instruction-dataset
A collection of open-source dataset to train instruction-following LLMs (ChatGPT,LLaMA,Alpaca)
thunlp/UltraChat
Large-scale, Informative, and Diverse Multi-round Chat Data (and Models)
lm-sys/FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
AetherCortex/Llama-X
Open Academic Research on Improving LLaMA to SOTA LLM
bigcode-project/bigcode-dataset
sahil280114/codealpaca
databrickslabs/dolly
Databricks’ Dolly, a large language model trained on the Databricks Machine Learning Platform
getcursor/cursor
The AI Code Editor
ymcui/Chinese-LLaMA-Alpaca
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
PhoebusSi/Alpaca-CoT
We unified the interfaces of instruction-tuning data (e.g., CoT data), multiple LLMs and parameter-efficient methods (e.g., lora, p-tuning) together for easy use. We welcome open-source enthusiasts to initiate any meaningful PR on this repo and integrate as many LLM related technologies as possible. 我们打造了方便研究人员上手和使用大模型等微调平台,我们欢迎开源爱好者发起任何有意义的pr!
mymusise/ChatGLM-Tuning
基于ChatGLM-6B + LoRA的Fintune方案
run-llama/llama_index
LlamaIndex is a data framework for your LLM applications
tatsu-lab/stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
THUDM/ChatGLM-6B
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
ad-freiburg/large-qa-datasets
A collection of large question answering datasets
arian-askari/ChatGPT-RetrievalQA
A dataset for training/evaluating Question Answering Retrieval models on ChatGPT responses with the possibility to training/evaluating on real human responses.
togethercomputer/OpenChatKit
c-box/KnowledgeLifecycle
Paper list of "The Life Cycle of Knowledge in Big Language Models: A Survey"
EleutherAI/gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
atfortes/Awesome-LLM-Reasoning
Reasoning in Large Language Models: Papers and Resources, including Chain-of-Thought and OpenAI o1 🍓
NVIDIA/NeMo
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)