yaodongyu's Stars
facebookresearch/searchformer
Official codebase for the paper "Beyond A* Better Planning with Transformers via Search Dynamics Bootstrapping".
togethercomputer/MoA
Together Mixture-Of-Agents (MoA) – 65.1% on AlpacaEval with OSS models
facebookresearch/nevergrad
A Python toolbox for performing gradient-free optimization
facebookresearch/MemoryMosaics
Memory Mosaics are networks of associative memories working in concert to achieve a prediction task.
openai/sparse_autoencoder
THUDM/AgentBench
A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)
SafeAILab/EAGLE
Official Implementation of EAGLE-1 (ICML'24) and EAGLE-2 (EMNLP'24)
test-time-training/ttt-lm-pytorch
Official PyTorch implementation of Learning to (Learn at Test Time): RNNs with Expressive Hidden States
test-time-training/ttt-lm-jax
Official JAX implementation of Learning to (Learn at Test Time): RNNs with Expressive Hidden States
zhaoyue-zephyrus/bsq-vit
[BSQ-ViT] Image and Video Tokenization with Binary Spherical Quantization
TencentARC/LLaMA-Pro
[ACL 2024] Progressive LLaMA with Block Expansion.
FoundationVision/LlamaGen
Autoregressive Model Beats Diffusion: 🦙 Llama for Scalable Image Generation
apple/ml-4m
4M: Massively Multimodal Masked Modeling
haizelabs/llama3-jailbreak
A trivial programmatic Llama 3 jailbreak. Sorry Zuck!
cambrian-mllm/cambrian
Cambrian-1 is a family of multimodal LLMs with a vision-centric design.
PixArt-alpha/PixArt-alpha
PixArt-α: Fast Training of Diffusion Transformer for Photorealistic Text-to-Image Synthesis
locuslab/ect
Consistency Models Made Easy
DruvPai/MCR2
An implementation of the principle of Maximal Coding Rate Reduction (MCR2).
facebookresearch/chameleon
Repository for Meta Chameleon, a mixed-modal early-fusion foundation model from FAIR.
eth-sri/SynthPAI
A Synthetic Dataset for Personal Attribute Inference
haotian-liu/LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
Tsingularity/dift
[NeurIPS'23] Emergent Correspondence from Image Diffusion
KMnP/vpt
❄️🔥 Visual Prompt Tuning [ECCV 2022] https://arxiv.org/abs/2203.12119
kvfrans/jax-diffusion-transformer
Implementation of Diffusion Transformer (DiT) in JAX
OpenBMB/ToolBench
[ICLR'24 spotlight] An open platform for training, serving, and evaluating large language model for tool learning.
princeton-nlp/SimPO
SimPO: Simple Preference Optimization with a Reference-Free Reward
peterljq/Parsimonious-Concept-Engineering
Parsimonious Concept Engineering (PaCE) uses sparse coding on a large-scale concept dictionary to effectively improve the trustworthiness of Large Language Models by precisely controlling and modifying their neural activations.
stanford-cs336/spring2024-lectures
redwoodresearch/Easy-Transformer
lucidrains/alphafold3-pytorch
Implementation of Alphafold 3 in Pytorch