houyili's Stars
OpenLMLab/GAOKAO-Bench
GAOKAO-Bench is an evaluation framework that utilizes GAOKAO questions as a dataset to evaluate large language models.
OpenLMLab/GAOKAO-Bench-2023
GAOGAO-Bench-2023 is a supplement to the GAOKAO-Bench, a dataset to evaluate large language models.
state-spaces/mamba
Mamba SSM architecture
esbatmop/MNBVC
MNBVC(Massive Never-ending BT Vast Chinese corpus)超大规模中文语料集。对标chatGPT训练的40T数据。MNBVC数据集不但包括主流文化,也包括各个小众文化甚至火星文的数据。MNBVC数据集包括新闻、作文、小说、书籍、杂志、论文、台词、帖子、wiki、古诗、歌词、商品介绍、笑话、糗事、聊天记录等一切形式的纯文本中文数据。
jxxghp/MoviePilot
NAS媒体库自动化管理工具
LLaMafia/llamafia.github
databricks/megablocks
microsoft/Tutel
Tutel MoE: An Optimized Mixture-of-Experts Implementation
NVIDIA/cutlass
CUDA Templates for Linear Algebra Subroutines
triton-lang/triton
Development repository for the Triton language and compiler
XueFuzhao/OpenMoE
A family of open-sourced Mixture-of-Experts (MoE) Large Language Models
kyegomez/FlashAttention20
Get down and dirty with FlashAttention2.0 in pytorch, plug in and play no complex CUDA kernels
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
deepseek-ai/DeepSeek-LLM
DeepSeek LLM: Let there be answers
openai/tiktoken
tiktoken is a fast BPE tokeniser for use with OpenAI's models.
tatsu-lab/gpt_paper_assistant
GPT4 based personalized ArXiv paper assistant bot
Alibaba-NLP/EcomGPT
An Instruction-tuned Large Language Model for E-commerce
GPT-Fathom/GPT-Fathom
GPT-Fathom is an open-source and reproducible LLM evaluation suite, benchmarking 10+ leading open-source and closed-source LLMs as well as OpenAI's earlier models on 20+ curated benchmarks under aligned settings.
google-research/xtreme
XTREME is a benchmark for the evaluation of the cross-lingual generalization ability of pre-trained multilingual models that covers 40 typologically diverse languages and includes nine tasks.
NVIDIA/apex
A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorch
pluto-junzeng/C4-zh
大规模中文语料
openlm-research/open_llama
OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset
bigscience-workshop/data-preparation
Code used for sourcing and cleaning the BigScience ROOTS corpus
alibaba/Megatron-LLaMA
Best practice for training LLaMA models in Megatron-LM
baichuan-inc/Baichuan2
A series of large language models developed by Baichuan Intelligent Technology
THUDM/GLM
GLM (General Language Model)
zhenbench/z-bench
Z-Bench 1.0 by 真格基金:一个麻瓜的大语言模型中文测试集。Z-Bench is a LLM prompt dataset for non-technical users, developed by an enthusiastic AI-focused team in Zhenfund.
baichuan-inc/Baichuan-7B
A large-scale 7B pretraining language model developed by BaiChuan-Inc.
baichuan-inc/Baichuan-13B
A 13B large language model developed by Baichuan Intelligent Technology
meta-llama/llama
Inference code for Llama models