kugwzk's Stars
apple/ml-stable-diffusion
Stable Diffusion with Core ML on Apple Silicon
state-spaces/mamba
mistralai/mistral-src
Reference implementation of Mistral AI 7B v0.1 model.
sgl-project/sglang
SGLang is a structured generation language designed for large language models (LLMs). It makes your interaction with models faster and more controllable.
openai/weak-to-strong
predibase/lorax
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
ytongbai/LVM
microsoft/Cream
This is a collection of our NAS and Vision Transformer work.
siliconflow/onediff
OneDiff: An out-of-the-box acceleration library for diffusion models.
abertsch72/unlimiformer
Public repo for the NeurIPS 2023 paper "Unlimiformer: Long-Range Transformers with Unlimited Length Input"
open-compass/MixtralKit
A toolkit for inference and evaluation of 'mixtral-8x7b-32kseqlen' from Mistral AI
LTH14/rcg
PyTorch implementation of RCG https://arxiv.org/abs/2312.03701
pjlab-sys4nlp/llama-moe
⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training
VILA-Lab/ATLAS
A principled instruction benchmark on formulating effective queries and prompts for large language models (LLMs). Our paper: https://arxiv.org/abs/2312.16171
abacaj/fine-tune-mistral
Fine-tune mistral-7B on 3090s, a100s, h100s
horseee/DeepCache
[CVPR 2024] DeepCache: Accelerating Diffusion Models for Free
srush/annotated-mamba
Annotated version of the Mamba paper
kyegomez/MultiModalMamba
A novel implementation of fusing ViT with Mamba into a fast, agile, and high performance Multi-Modal Model. Powered by Zeta, the simplest AI framework ever.
hkust-nlp/deita
Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]
huggingface/open-muse
Open reproduction of MUSE for fast text2image generation.
FuxiaoLiu/LRV-Instruction
[ICLR'24] Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning
zchuz/CoT-Reasoning-Survey
A Survey of Chain of Thought Reasoning: Advances, Frontiers and Future
facebookresearch/Shepherd
This is the repo for the paper Shepherd -- A Critic for Language Model Generation
OpenBMB/InfiniteBench
Codes for the paper "∞Bench: Extending Long Context Evaluation Beyond 100K Tokens": https://arxiv.org/abs/2402.13718
Q-Future/Q-Align
③[ICML2024] [IQA, IAA, VQA] All-in-one Foundation Model for visual scoring. Can efficiently fine-tune to downstream datasets.
IBM/SALMON
Self-Alignment with Principle-Following Reward Models
huggingface/amused
eclipse-t2i/eclipse-inference
[CVPR 2024] Official PyTorch implementation of "ECLIPSE: Revisiting the Text-to-Image Prior for Efficient Image Generation"
isXinLiu/MM-SafetyBench
zchuz/TimeBench
The repository for paper "TimeBench: A Comprehensive Evaluation of Temporal Reasoning Abilities in Large Language Models"