zhangyuansen's Stars
moymix/TaskMatrix
haotian-liu/LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
ymcui/Chinese-LLaMA-Alpaca
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
WooooDyy/LLM-Agent-Paper-List
The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.
microsoft/DeepSpeedExamples
Example models using DeepSpeed
llm-attacks/llm-attacks
Universal and Transferable Attacks on Aligned Language Models
modelscope/swift
ms-swift: Use PEFT or Full-parameter to finetune 300+ LLMs or 50+ MLLMs. (Qwen2, GLM4v, Internlm2.5, Yi, Llama3.1, Llava-Video, Internvl2, MiniCPM-V-2.6, Deepseek, Baichuan2, Gemma2, Phi3-Vision, ...)
eric-mitchell/direct-preference-optimization
Reference implementation for DPO (Direct Preference Optimization)
Timothyxxx/Chain-of-ThoughtsPapers
A trend starts from "Chain of Thought Prompting Elicits Reasoning in Large Language Models".
zjunlp/EasyEdit
[ACL 2024] An Easy-to-use Knowledge Editing Framework for LLMs.
yechens/NL2SQL
Text2SQL 语义解析数据集、解决方案、paper资源整合项目
zixian2021/AI-interview-cards
最完整的AI算法面试题目仓库,1000道,25个类目
yule-BUAA/MergeLM
Codebase for Merging Language Models (ICML 2024)
ucinlp/autoprompt
AutoPrompt: Automatic Prompt Construction for Masked Language Models.
facebookresearch/GradientEpisodicMemory
Continuum Learning with GEM: Gradient Episodic Memory
BeyonderXX/InstructUIE
Universal information extraction with instruction learning
JH-LEE-KR/l2p-pytorch
PyTorch Implementation of Learning to Prompt (L2P) for Continual Learning @ CVPR22
yunqing-me/AttackVLM
[NeurIPS-2023] Annual Conference on Neural Information Processing Systems
thunlp/OpenBackdoor
An open-source toolkit for textual backdoor attack and defense (NeurIPS 2022 D&B, Spotlight)
cmnfriend/O-LoRA
neural-dialogue-metrics/Distinct-N
Compute Distinct-N metric proposed by Jiwei Li et al.
arazd/ProgressivePrompts
Progressive Prompts: Continual Learning for Language Models
reasoning-machines/CoCoGen
Language Models of Code are Few-Shot Commonsense Learners (EMNLP 2022)
hkust-nlp/PEM_composition
[NeurIPS 2023] Github repository for "Composing Parameter-Efficient Modules with Arithmetic Operations"
BeyonderXX/TRACE
TRACE: A Comprehensive Benchmark for Continual Learning in Large Language Models
h3lio5/episodic-lifelong-learning
Implementation of "Episodic Memory in Lifelong Language Learning"(NeurIPS 2019) in Pytorch
SALT-NLP/IDBR
Codes for the paper: "Continual Learning for Text Classification with Information Disentanglement Based Regularization"
MehdiAbbanaBennani/continual-learning-ogdplus
Code for "Generalisation Guarantees for Continual Learning with Orthogonal Gradient Descent" (ICML 2020 - Lifelong Learning Workshop)
cambridgeltl/ClaPS
Survival of the Most Influential Prompts: Efficient Black-Box Prompt Search via Clustering and Pruning (Zhou et al.; EMNLP 2023 Findings)
meryemmhamdi1/x-continuous-learning
This is to analyze and visualize different parameters and components in the architectures of different cross-lingual downstream tasks to detect and treat catastrophic forgetting.