weigq's Stars
ultralytics/ultralytics
NEW - YOLOv8 🚀 in PyTorch > ONNX > OpenVINO > CoreML > TFLite
guoyww/AnimateDiff
Official implementation of AnimateDiff.
jzhang38/TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
LargeWorldModel/LWM
mikel-brostrom/yolo_tracking
BoxMOT: pluggable SOTA tracking modules for segmentation, object detection and pose estimation models
luosiallen/latent-consistency-model
Latent Consistency Models: Synthesizing High-Resolution Images with Few-Step Inference
tyxsspa/AnyText
Official implementation code of the paper <AnyText: Multilingual Visual Text Generation And Editing>
viliusle/miniPaint
online image editor
sybrenjansen/mpire
A Python package for easy multiprocessing, but faster than multiprocessing
yformer/EfficientSAM
EfficientSAM: Leveraged Masked Image Pretraining for Efficient Segment Anything
ytongbai/LVM
VainF/pytorch-msssim
Fast and differentiable MS-SSIM and SSIM for pytorch.
sczhou/Upscale-A-Video
Upscale-A-Video: Temporal-Consistent Diffusion Model for Real-World Video Super-Resolution
ProjectNUWA/DragNUWA
haofanwang/Lora-for-Diffusers
The most easy-to-understand tutorial for using LoRA (Low-Rank Adaptation) within diffusers framework for AI Generation Researchers🔥
lichao-sun/SoraReview
The official GitHub page for the review paper "Sora: A Review on Background, Technology, Limitations, and Opportunities of Large Vision Models".
Zhen-Dong/Magic-Me
Codes for ID-Specific Video Customized Diffusion
TencentYoutuResearch/CrowdCounting-P2PNet
The official codes for the ICCV2021 Oral presentation "Rethinking Counting and Localization in Crowds: A Purely Point-Based Framework"
pixeli99/SVD_Xtend
Stable Video Diffusion Training Code and Extensions.
MCG-NJU/EMA-VFI
[CVPR 2023] Extracting Motion and Appearance via Inter-Frame Attention for Efficient Video Frame Interpolatio
TencentARC/CustomNet
TiankaiHang/Min-SNR-Diffusion-Training
[ICCV 2023] Efficient Diffusion Training via Min-SNR Weighting Strategy
xk-huang/segment-caption-anything
[CVPR 24] The repository provides code for running inference and training for "Segment and Caption Anything" (SCA) , links for downloading the trained model checkpoints, and example notebooks / gradio demo that show how to use the model.
layer6ai-labs/dgm-eval
Codebase for evaluation of deep generative models as presented in Exposing flaws of generative model evaluation metrics and their unfair treatment of diffusion models
ai-forever/MoVQGAN
MoVQGAN - model for the image encoding and reconstruction
showlab/T2VScore
T2VScore: Towards A Better Metric for Text-to-Video Generation
Hritikbansal/videocon
JunjieYang97/Meta-ControlNet
georgia-tech-db/eva-decord
An efficient video loader for deep learning with smart shuffling that's super easy to digest
makepixelsdance/makepixelsdance.github.io
Homepage for PixelDance. Paper -> https://arxiv.org/abs/2311.10982