JiahangXu's Stars
maitrix-org/llm-reasoners
A library for advanced large language model reasoning
princeton-nlp/tree-of-thought-llm
[NeurIPS 2023] Tree of Thoughts: Deliberate Problem Solving with Large Language Models
Significant-Gravitas/AutoGPT
AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.
microsoft/promptflow
Build high-quality LLM apps - from prototyping, testing to production deployment and monitoring.
jxzhangjhu/Awesome-LLM-Prompt-Optimization
Awesome-LLM-Prompt-Optimization: a curated list of advanced prompt optimization and tuning methods in Large Language Models
microsoft/sammo
A library for prompt engineering and optimization (SAMMO = Structure-aware Multi-Objective Metaprompt Optimization)
stanfordnlp/dspy
DSPy: The framework for programming—not prompting—foundation models
AGI-Edgerunners/LLM-Optimizers-Papers
Must-read Papers on Large Language Model (LLM) as Optimizers and Automatic Optimization for Prompting LLMs.
hwchase17/auto-openai-prompter
horseee/LLM-Pruner
[NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support LLaMA, Llama-2, BLOOM, Vicuna, Baichuan, etc.
justinchiu/openlogprobs
Extract full next-token probabilities via language model APIs
horseee/DeepCache
[CVPR 2024] DeepCache: Accelerating Diffusion Models for Free
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
nlpxucan/WizardLM
LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMath
microsoft/CoML
Interactive coding assistant for data scientists and machine learning developers, empowered by large language models.
princeton-nlp/LLM-Shearing
[ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning
microsoft/Moonlit
This is a collection of our research on efficient AI, covering hardware-aware NAS and model compression.
getao/icae
The repo for In-context Autoencoder
declare-lab/instruct-eval
This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.
THUDM/WebGLM
WebGLM: An Efficient Web-enhanced Question Answering System (KDD 2023)
EleutherAI/lm-evaluation-harness
A framework for few-shot evaluation of language models.
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
microsoft/LoRA
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
HuangOwen/Awesome-LLM-Compression
Awesome LLM compression research papers and tools.
OpenGVLab/LLaMA-Adapter
[ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters
AGI-Edgerunners/LLM-Adapters
Code for our EMNLP 2023 Paper: "LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models"
lm-sys/FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
yizhongw/self-instruct
Aligning pretrained language models with instruction data generated by themselves.
Instruction-Tuning-with-GPT-4/GPT-4-LLM
Instruction Tuning with GPT-4
tatsu-lab/stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.