rtao499's Stars
QingruZhang/AdaLoRA
AdaLoRA: Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning (ICLR 2023).
FederatedAI/FATE-LLM
Federated Learning for LLMs.
SMILELab-FL/FedLab
A flexible Federated Learning Framework based on PyTorch, simplifying your Federated Learning research.
IbrahimYang/Hybrid-neural-networks
CGCL-codes/FedGKD
MuiseDestiny/zotero-style
Ethereal Style for Zotero
haozhaowang/DaFKD2023
Code for CVPR2023 DaFKD : Domain-aware Federated Knowledge Distillation
langchain-ai/langchain
🦜🔗 Build context-aware reasoning applications
jzhang38/TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
ridgerchu/SpikeGPT
Implementation of "SpikeGPT: Generative Pre-trained Language Model with Spiking Neural Networks"
RUCAIBox/LLMSurvey
The official GitHub page for the survey paper "A Survey of Large Language Models".
BlinkDL/RWKV-LM
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
Stability-AI/StableLM
StableLM: Stability AI Language Models
binary-husky/gpt_academic
为GPT/GLM等LLM大语言模型提供实用化交互接口,特别优化论文阅读/润色/写作体验,模块化设计,支持自定义快捷按钮&函数插件,支持Python和C++等项目剖析&自译解功能,PDF/LaTex论文翻译&总结功能,支持并行问询多种LLM模型,支持chatglm3等本地模型。接入通义千问, deepseekcoder, 讯飞星火, 文心一言, llama2, rwkv, claude2, moss等。
MLGroupJLU/LLM-eval-survey
The official GitHub page for the survey paper "A Survey on Evaluation of Large Language Models".
hkust-nlp/ceval
Official github repo for C-Eval, a Chinese evaluation suite for foundation models [NeurIPS 2023]
CLUEbenchmark/CLUE
中文语言理解测评基准 Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained models, corpus and leaderboard
Brain-Cog-Lab/Conversion_Burst
This repository contains code from our paper Efficient and Accurate Conversion of Spiking Neural Network with Burst Spikes, published in IJCAI, 2022.
KarhouTam/FL-bench
Benchmark of federated learning. Dedicated to the community. 🤗
hzc1208/ANN2SNN_SRP
liguodongiot/llm-action
本项目旨在分享大模型相关技术原理以及实战经验(大模型工程化、大模型应用落地)
lixin4ever/Conference-Acceptance-Rate
Acceptance rates for the major AI conferences
Zhuzzq/FedLP
IST-DASLab/gptq
Code for the ICLR 2023 paper "GPTQ: Accurate Post-training Quantization of Generative Pretrained Transformers".
microsoft/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
shibing624/MedicalGPT
MedicalGPT: Training Your Own Medical GPT Model with ChatGPT Training Pipeline. 训练医疗大模型,实现了包括增量预训练(PT)、有监督微调(SFT)、RLHF、DPO、ORPO。
EleutherAI/gpt-neo
An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.
microsoft/Swin-Transformer
This is an official implementation for "Swin Transformer: Hierarchical Vision Transformer using Shifted Windows".
microsoft/LoRA
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
ZK-Zhou/spikformer
ICLR 2023, Spikformer: When Spiking Neural Network Meets Transformer