llgithubll's Stars
langchain-ai/langchain
🦜🔗 Build context-aware reasoning applications
meta-llama/llama
Inference code for Llama models
LAION-AI/Open-Assistant
OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so.
chenfei-wu/TaskMatrix
tatsu-lab/stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
Hannibal046/Awesome-LLM
Awesome-LLM: a curated list of Large Language Model
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
BlinkDL/RWKV-LM
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
FMInference/FlexGen
Running large language models on a single GPU for throughput-oriented scenarios.
nebuly-ai/nebuly
The user analytics platform for LLMs
lucidrains/PaLM-rlhf-pytorch
Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM
cloneofsimo/lora
Using Low-rank adaptation to quickly fine-tune diffusion models.
modelscope/modelscope
ModelScope: bring the notion of Model-as-a-Service to life.
EleutherAI/gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
open-mmlab/mmcv
OpenMMLab Computer Vision Foundation
CarperAI/trlx
A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)
dalinvip/Awesome-ChatGPT
ChatGPT资料汇总学习,持续更新......
SurviveSJTU/SurviveSJTUManual
更新2008年版本的《上海交通大学生存手册》gitbook发布于https://survivesjtu.gitbook.io/survivesjtumanual/
huggingface/deep-rl-class
This repo contains the syllabus of the Hugging Face Deep Reinforcement Learning Course.
gligen/GLIGEN
Open-Set Grounded Text-to-Image Generation
microsoft/DeBERTa
The implementation of DeBERTa
openai/sparse_attention
Examples of using sparse attention, as in "Generating Long Sequences with Sparse Transformers"
bigscience-workshop/Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
xlang-ai/UnifiedSKG
[EMNLP 2022] Unifying and multi-tasking structured knowledge grounding with language models
voidful/TextRL
Implementation of ChatGPT RLHF (Reinforcement Learning with Human Feedback) on any generation model in huggingface's transformer (blommz-176B/bloom/gpt/bart/T5/MetaICL)
codecaution/Awesome-Mixture-of-Experts-Papers
A curated reading list of research in Mixture-of-Experts(MoE).
THU-KEG/EvaluationPapers4ChatGPT
Resource, Evaluation and Detection Papers for ChatGPT
shizhediao/ChatGPTPapers
Must-read papers, related blogs and API tools on the pre-training and tuning methods for ChatGPT.
KSESEU/LLMPapers
Papers & Works for large languange models (ChatGPT, GPT-3, Codex etc.).
microsoft/DialogLM
Official Implementation of "DialogLM: Pre-trained Model for Long Dialogue Understanding and Summarization."