Pinned Repositories
AdaLoRA
AdaLoRA: Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning (ICLR 2023).
AdaShift_Release
The original implementation of the experiments in the paper of AdaShift (See https://arxiv.org/abs/1810.00143)
alignment-handbook
Robust recipes to align language models with human and AI preferences
alpaca_eval
An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.
chain-of-thought-hub
Benchmarking large language models' complex reasoning ability with chain-of-thought prompting
Data-Mining-LianJia-HousePrice-Prediction-Project
Data Mining Project: extract raw data from LianJia; data proprocess and feather engineering; build model with MLP and BiLSTM.
PASTA
PASTA: Post-hoc Attention Steering for LLMs
PLATON
This pytorch package implements PLATON: Pruning Large Transformer Models with Upper Confidence Bound of Weight Importance (ICML 2022).
Thanos
A Biased Graph Neural Network Sampler with Near-Optimal Regret.
web-attack-deep-detection
QingruZhang's Repositories
QingruZhang/AdaLoRA
AdaLoRA: Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning (ICLR 2023).
QingruZhang/PASTA
PASTA: Post-hoc Attention Steering for LLMs
QingruZhang/PLATON
This pytorch package implements PLATON: Pruning Large Transformer Models with Upper Confidence Bound of Weight Importance (ICML 2022).
QingruZhang/Data-Mining-LianJia-HousePrice-Prediction-Project
Data Mining Project: extract raw data from LianJia; data proprocess and feather engineering; build model with MLP and BiLSTM.
QingruZhang/chain-of-thought-hub
Benchmarking large language models' complex reasoning ability with chain-of-thought prompting
QingruZhang/alignment-handbook
Robust recipes to align language models with human and AI preferences
QingruZhang/alpaca_eval
An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.
QingruZhang/baseline-lora
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
QingruZhang/cv_hwk_submission
QingruZhang/demonstrated-feedback
QingruZhang/exllama
A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.
QingruZhang/FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
QingruZhang/lm-evaluation-harness
A framework for few-shot evaluation of language models.
QingruZhang/LoftQ
QingruZhang/lost-in-the-middle
Code and data for "Lost in the Middle: How Language Models Use Long Contexts"
QingruZhang/MAmmoTH
Code and data for "MAmmoTH: Building Math Generalist Models through Hybrid Instruction Tuning" (ICLR 2024)
QingruZhang/math
The MATH Dataset (NeurIPS 2021)
QingruZhang/OpenRLHF
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & Mixtral)
QingruZhang/OptiMUS
Optimization Modeling Using mip Solvers and large language models
QingruZhang/peft
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
QingruZhang/ReAlign
Reformatted Alignment
QingruZhang/SoftwareConfiguration
QingruZhang/Sophia
The official implementation of “Sophia: A Scalable Stochastic Second-order Optimizer for Language Model Pre-training”
QingruZhang/summarize-from-feedback
Code for "Learning to summarize from human feedback"
QingruZhang/text-generation-webui
A Gradio web UI for Large Language Models. Supports transformers, GPTQ, AWQ, llama.cpp (GGUF), Llama models.
QingruZhang/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
QingruZhang/trl
Train transformer language models with reinforcement learning.
QingruZhang/TrustLLM
TrustLLM: Trustworthiness in Large Language Models
QingruZhang/unify-parameter-efficient-tuning
Implementation of paper "Towards a Unified View of Parameter-Efficient Transfer Learning" (ICLR 2022)
QingruZhang/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs