exlaw
Aiwei Liu, a Ph.D. student at Tsinghua University Focus on Natural Language Processing
Tsinghua UniversityBeijing, China
exlaw's Stars
hiyouga/LLaMA-Factory
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
meta-llama/llama-recipes
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama for WhatsApp & Messenger.
huggingface/trl
Train transformer language models with reinforcement learning.
EleutherAI/lm-evaluation-harness
A framework for few-shot evaluation of language models.
predibase/lorax
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
Niek/chatgpt-web
ChatGPT web interface using the OpenAI API
yule-BUAA/MergeLM
Codebase for Merging Language Models (ICML 2024)
sail-sg/lorahub
[COLM 2024] LoraHub: Efficient Cross-Task Generalization via Dynamic LoRA Composition
THUYimingLi/BackdoorBox
The open-sourced Python toolbox for backdoor attacks and defenses.
THU-BPM/MarkLLM
MarkLLM: An Open-Source Toolkit for LLM Watermarking.(EMNLP 2024 Demo)
chrisliu298/awesome-llm-unlearning
A resource repository for machine unlearning in large language models
Ablustrund/LoRAMoE
LoRAMoE: Revolutionizing Mixture of Experts for Maintaining World Knowledge in Language Model Alignment
Libr-AI/do-not-answer
Do-Not-Answer: A Dataset for Evaluating Safeguards in LLMs
ICTMCG/LLM-for-misinformation-research
Paper list of misinformation research using (multi-modal) large language models, i.e., (M)LLMs.
uukuguy/multi_loras
Load multiple LoRA modules simultaneously and automatically switch the appropriate combination of LoRA modules to generate the best answer based on user queries.
andyrdt/refusal_direction
Code and results accompanying the paper "Refusal in Language Models Is Mediated by a Single Direction".
wuhy68/Parameter-Efficient-MoE
Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks
ericwtodd/function_vectors
Function Vectors in Large Language Models (ICLR 2024)
GCYZSL/MoLA
PKU-Alignment/beavertails
BeaverTails is a collection of datasets designed to facilitate research on safety alignment in large language models (LLMs).
TUDB-Labs/MixLoRA
State-of-the-art Parameter-Efficient MoE Fine-tuning Method
LucyDYu/Awesome-Multimodal-Continual-Learning
chrisliu298/awesome-representation-engineering
A resource repository for representation engineering in large language models
vinusankars/Reliability-of-AI-text-detectors
Can AI-Generated Text be Reliably Detected?
rishub-tamirisa/tamper-resistance
Official Repository for "Tamper-Resistant Safeguards for Open-Weight LLMs"
ydyjya/LLM-IHS-Explanation
THU-BPM/WaterSeeker
WaterSeeker: Pioneering Efficient Detection of Watermarked Segments in Large Documents
LeiLiLab/llm_watermark_tutorial
survey-text-watermark/survey-text-watermark.github.io
exlaw/DLMA
Code and data for paper "Direct Large Language Model Alignment Through Self-Rewarding Contrastive Prompt Distillation" accepted by ACL 2024.