shu-le's Stars
baojudezeze/Generative-Virtual-Try-On
Generative virtual try on (VTON), try-on images of characters can be generated by text prompt.
ali-vilab/FlashFace
kyegomez/Vit-RGTS
Open source implementation of "Vision Transformers Need Registers"
deepinsight/insightface
State-of-the-art 2D and 3D Face Analysis Project
bghira/SimpleTuner
A general fine-tuning kit geared toward diffusion models.
hpcaitech/ColossalAI
Making large AI models cheaper, faster and more accessible
lucidrains/magvit2-pytorch
Implementation of MagViT2 Tokenizer in Pytorch
HL-hanlin/VideoDirectorGPT
official implementation of VideoDirectorGPT: Consistent Multi-scene Video Generation via LLM-Guided Planning (COLM 2024)
XLabs-AI/x-flux
Eris2025/awesome-flux
A curated list of awesome resources for FLUX, the state-of-the-art text-to-image model by Black Forest Labs.
ruanyf/weekly
科技爱好者周刊,每周五发布
minar09/awesome-virtual-try-on
A curated list of awesome research papers, projects, code, dataset, workshops etc. related to virtual try-on.
yipoh/AesExpert
[ACMMM 2024] AesExpert: Towards Multi-modality Foundation Model for Image Aesthetics Perception
THUDM/CogVideo
text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)
dair-ai/ml-visuals
🎨 ML Visuals contains figures and templates which you can reuse and customize to improve your scientific writing.
lencx/ChatGPT
🔮 ChatGPT Desktop Application (Mac, Windows and Linux)
IDEA-Research/GroundingDINO
[ECCV 2024] Official implementation of the paper "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection"
gligen/GLIGEN
Open-Set Grounded Text-to-Image Generation
Picsart-AI-Research/StreamingT2V
StreamingT2V: Consistent, Dynamic, and Extendable Long Video Generation from Text
tencent-ailab/IP-Adapter
The image prompt adapter is designed to enable a pretrained text-to-image diffusion model to generate images with image prompt.
kousw/experimental-consistory
google/prompt-to-prompt
HVision-NKU/StoryDiffusion
Accepted as [NeurIPS 2024] Spotlight Presentation Paper
TencentARC/MasaCtrl
[ICCV 2023] Consistent Image Synthesis and Editing
IDEA-Research/Grounded-SAM-2
Grounded SAM 2: Ground and Track Anything in Videos with Grounding DINO, Florence-2 and SAM 2
lllyasviel/Paints-UNDO
Understand Human Behavior to Align True Needs
black-forest-labs/flux
Official inference repo for FLUX.1 models
facebookresearch/sam2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
Breakthrough/PySceneDetect
:movie_camera: Python and OpenCV-based scene cut/transition detection program & library.
YangLing0818/VideoTetris
[NeurIPS 2024] VideoTetris: Towards Compositional Text-To-Video Generation