wuwukongkong's Stars
huggingface/trl
Train transformer language models with reinforcement learning.
ccfddl/ccf-deadlines
⏰ Collaboratively track deadlines of conferences recommended by CCF (Website, Python Cli, Wechat Applet) / If you find it useful, please star this project, thanks~
TencentARC/InstantMesh
InstantMesh: Efficient 3D Mesh Generation from a Single Image with Sparse-view Large Reconstruction Models
google-research/parti
THUDM/ImageReward
[NeurIPS 2023] ImageReward: Learning and Evaluating Human Preferences for Text-to-image Generation
nianticlabs/mickey
[CVPR 2024 - Oral] Matching 2D Images in 3D: Metric Relative Pose from Metric Correspondences
yuvalkirstain/PickScore
cvpr-org/author-kit
tgxs002/HPSv2
Human Preference Score v2: A Solid Benchmark for Evaluating Human Preferences of Text-to-Image Synthesis
mindspore-lab/mindone
one for all, Optimal generator with No Exception
yfzhang114/Awesome-Multimodal-Large-Language-Models
Reading notes about Multimodal Large Language Models, Large Language Models, and Diffusion Models
openpsi-project/ReaLHF
Super-Efficient RLHF Training of LLMs with Parameter Reallocation
yk7333/d3po
[CVPR 2024] Code for the paper "Using Human Feedback to Fine-tune Diffusion Models without Any Reward Model"
RockeyCoss/SPO
Aesthetic Post-Training Diffusion Models from Generic Preferences with Step-by-step Preference Optimization
lyndonzheng/Free3D
[CVPR'24] Consistent Novel View Synthesis without 3D Representation
jiawei-ren/insactor
[NeurIPS 2023] InsActor: Instruction-driven Physics-based Characters
ExplainableML/ReNO
[NeurIPS 2024] ReNO: Enhancing One-step Text-to-Image Models through Reward-based Noise Optimization
sjtuplayer/SaRA
SaRA: High-Efficient Diffusion Model Fine-tuning with Progressive Sparse Low-Rank Adaptation
rom1504/embedding-reader
Efficiently read embedding in streaming from any filesystem
abacusai/smaug
liujf69/Classic-Generative-Model
Simple code demos about classic AIGC models/Compilation of blogs and papers on classic AIGC models.
pinterest/atg-research
jacklishufan/diffusion-kto
The official implementation of Diffusion-KTO: Aligning Diffusion Models by Optimizing Human Utility
Shentao-YANG/Dense_Reward_T2I
Source code for "A Dense Reward View on Aligning Text-to-Image Diffusion with Preference" (ICML'24).
cccedric/cpql
This is the official PyTorch implementation of the paper "Boosting Continuous Control with Consistency Policy".
general-preference/general-preference-model
Official implementation of paper "Beyond Bradley-Terry Models: A General Preference Model for Language Model Alignment" (https://arxiv.org/abs/2410.02197)
princeton-nlp/unintentional-unalignment
[ICLR 2025] Unintentional Unalignment: Likelihood Displacement in Direct Preference Optimization
ZiyiZhang27/sdpo
Code for the paper "Aligning Few-Step Diffusion Models with Dense Reward Difference Learning"
catalpaaa/DeMansia
atelion/DPO-Stable-Diffusion