hejingwenhejingwen
Currently a research intern in Shenzhen Institutes of Advanced Technology.
Shenzhen Institutes of Advanced Technology
hejingwenhejingwen's Stars
hpcaitech/Open-Sora
Open-Sora: Democratizing Efficient Video Production for All
THUDM/CogVideo
text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)
OpenGVLab/LLaMA-Adapter
[ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters
AILab-CVC/VideoCrafter
VideoCrafter2: Overcoming Data Limitations for High-Quality Video Diffusion Models
XPixelGroup/DiffBIR
Official codes of DiffBIR: Towards Blind Image Restoration with Generative Diffusion Prior
ali-vilab/VGen
Official repo for VGen: a holistic video generation ecosystem for video generation building on diffusion models
williamyang1991/Rerender_A_Video
[SIGGRAPH Asia 2023] Rerender A Video: Zero-Shot Text-Guided Video-to-Video Translation
PixArt-alpha/PixArt-alpha
PixArt-α: Fast Training of Diffusion Transformer for Photorealistic Text-to-Image Synthesis
Alpha-VLLM/LLaMA2-Accessory
An Open-source Toolkit for LLM Development
Alpha-VLLM/Lumina-T2X
Lumina-T2X is a unified framework for Text to Any Modality Generation
Vchitect/Latte
Latte: Latent Diffusion Transformer for Video Generation.
NVlabs/edm
Elucidating the Design Space of Diffusion-Based Generative Models (EDM)
sczhou/Upscale-A-Video
[CVPR 2024] Upscale-A-Video: Temporal-Consistent Diffusion Model for Real-World Video Super-Resolution
ali-vilab/videocomposer
Official repo for VideoComposer: Compositional Video Synthesis with Motion Controllability
OpenGVLab/SAM-Med2D
Official implementation of SAM-Med2D
lucidrains/muse-maskgit-pytorch
Implementation of Muse: Text-to-Image Generation via Masked Generative Transformers, in Pytorch
NUS-HPC-AI-Lab/Neural-Network-Parameter-Diffusion
We introduce a novel approach for parameter generation, named neural network parameter diffusion (p-diff), which employs a standard latent diffusion model to synthesize a new set of parameters
Vchitect/Vchitect-2.0
Vchitect-2.0: Parallel Transformer for Scaling Up Video Diffusion Models
Vchitect/VBench
[CVPR2024 Highlight] VBench - We Evaluate Video Generation
lucidrains/magvit2-pytorch
Implementation of MagViT2 Tokenizer in Pytorch
YingqingHe/ScaleCrafter
[ICLR 2024 Spotlight] Official implementation of ScaleCrafter for higher-resolution visual generation at inference time.
TianxingWu/FreeInit
[ECCV 2024] FreeInit: Bridging Initialization Gap in Video Diffusion Models
Zj-BinXia/DiffIR
This project is the official implementation of 'Diffir: Efficient diffusion model for image restoration', ICCV2023
Vchitect/VEnhancer
Official codes of VEnhancer: Generative Space-Time Enhancement for Video Generation
VINHYU/CoSeR
[CVPR 2024] CoSeR: Bridging Image and Language for Cognitive Super-Resolution
city-super/MatrixCity
bornfly-detachment/asymmetric_magvitv2
In 2024, the strongest open-source implementation of asymmetric magvit_v2 supports inference code but excludes VQVAE. It supports the joint encoding of images and videos, accommodating arbitrary video lengths and resolutions. It surpasses all open-source models in FID and FVD, with 4z and 16z models available on huggingface.
Shuweis/ResMaster
XPixelGroup/SEAL
ICLR 2024 (Spotlight) - SEAL: A Framework for Systematic Evaluation of Real-World Super-Resolution
Vchitect/LiteGen
A light-weight and high-efficient training framework for accelerating diffusion tasks.