weTerence's Stars
XPixelGroup/DepictQA
DepictQA: Depicted Image Quality Assessment with Vision Language Models
geekyutao/Inpaint-Anything
Inpaint anything using Segment Anything and inpainting models.
yisol/IDM-VTON
[ECCV2024] IDM-VTON : Improving Diffusion Models for Authentic Virtual Try-on in the Wild
levihsu/OOTDiffusion
Official implementation of OOTDiffusion: Outfitting Fusion based Latent Diffusion for Controllable Virtual Try-on
uncbiag/Awesome-Foundation-Models
A curated list of foundation models for vision and language tasks
Ji4chenLi/t2v-turbo
Code repository for T2V-Turbo and T2V-Turbo-v2
HL-hanlin/Ctrl-Adapter
Official implementation of Ctrl-Adapter: An Efficient and Versatile Framework for Adapting Diverse Controls to Any Diffusion Model
aigc-apps/CogVideoX-Fun
📹 A more flexible CogVideoX that can generate videos at any resolution and creates videos from images.
facebookresearch/sapiens
High-resolution models for human tasks.
brycedrennan/imaginAIry
Pythonic AI generation of images and videos
hiyouga/LLaMA-Factory
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
stanfordnlp/dspy
DSPy: The framework for programming—not prompting—foundation models
black-forest-labs/flux
Official inference repo for FLUX.1 models
PKU-YuanGroup/Open-Sora-Plan
This project aim to reproduce Sora (Open AI T2V model), we wish the open source community contribute to this project.
aigc-apps/EasyAnimate
📺 An End-to-End Solution for High-Resolution and Long Video Generation Based on Transformer Diffusion
anonymous0769/DreamVideo
pydn/ComfyUI-to-Python-Extension
A powerful tool that translates ComfyUI workflows into executable Python code.
ostris/ai-toolkit
Various AI scripts. Mostly Stable Diffusion stuff.
XLabs-AI/x-flux
dvlab-research/ControlNeXt
Controllable video and image Generation, SVD, Animate Anyone, ControlNet, ControlNeXt, LoRA
Akegarasu/lora-scripts
LoRA & Dreambooth training scripts & GUI use kohya-ss's trainer, for diffusion model.
facebookresearch/sam2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
ICTMCG/Make-Your-Anchor
[CVPR 2024] Make-Your-Anchor: A Diffusion-based 2D Avatar Generation Framework.
Vchitect/VBench
[CVPR2024 Highlight] VBench - We Evaluate Video Generation
open-mmlab/PIA
[CVPR 2024] PIA, your Personalized Image Animator. Animate your images by text prompt, combing with Dreambooth, achieving stunning videos. PIA,你的个性化图像动画生成器,利用文本提示将图像变为奇妙的动画
TIGER-AI-Lab/ConsistI2V
ConsistI2V: Enhancing Visual Consistency for Image-to-Video Generation (TMLR 2024)
HVision-NKU/StoryDiffusion
Accepted as [NeurIPS 2024] Spotlight Presentation Paper
thu-ml/cond-image-leakage
Official implementation for "Identifying and Solving Conditional Image Leakage in Image-to-Video Diffusion Model" (NeurIPS 2024)
wdndev/llm_interview_note
主要记录大语言大模型(LLMs) 算法(应用)工程师相关的知识及面试题
Tencent/MimicMotion
High-Quality Human Motion Video Generation with Confidence-aware Pose Guidance