Aleafy
Research Intern@Shanghai AI Lab in CV Phd candidate@FDU, CS Bachelor@HIT
Fudan University, Shanghai AI LaboratoryShanghai
Aleafy's Stars
ubc-vision/vivid123
[CVPR 2024 Highlight] ViVid-1-to-3: Novel View Synthesis with Video Diffusion Models
qizekun/ReCon
[ICML 2023] Contrast with Reconstruct: Contrastive 3D Representation Learning Guided by Generative Pretraining
continental/6Img-to-3D
Few-Image to 3D Method
3DTopia/ThemeStation
[SIGGRAPH 2024] ThemeStation: Generating Theme-Aware 3D Assets from Few Exemplars
TQTQliu/MVSGaussian
[ECCV 2024] MVSGaussian: Fast Generalizable Gaussian Splatting Reconstruction from Multi-View Stereo
ali-vilab/VGen
Official repo for VGen: a holistic video generation ecosystem for video generation building on diffusion models
facebookresearch/chameleon
Repository for Meta Chameleon, a mixed-modal early-fusion foundation model from FAIR.
ShiArthur03/ShiArthur03
Qi-Zhangyang/Tailor3D
This is the official code for the paper Tailor3D
Fangkang515/CE3D
Official Implementation of ECCV2024 paper: Chat Edit 3D: Interactive 3D Scene Editing via Text Prompts
VDIGPKU/GALA3D
[ICML 2024] GALA3D: Towards Text-to-3D Complex Scene Generation via Layout-guided Generative Gaussian Splatting
KwaiVGI/LivePortrait
Bring portraits to life!
huggingface/cosmopedia
DepthAnything/Depth-Anything-V2
Depth Anything V2. A More Capable Foundation Model for Monocular Depth Estimation
xchhuang/bndm
Blue noise for diffusion models [SIGGRAPH 2024]
frank-xwang/UnSAM
[NeurIPS 2024] Code release for "Segment Anything without Supervision"
FreedomIntelligence/ALLaVA
Harnessing 1.4M GPT4V-synthesized Data for A Lite Vision-Language Model
RunpeiDong/DreamLLM
[ICLR 2024 Spotlight] DreamLLM: Synergistic Multimodal Comprehension and Creation
OpenGVLab/MM-NIAH
[NeurIPS 2024] Needle In A Multimodal Haystack (MM-NIAH): A comprehensive benchmark designed to systematically evaluate the capability of existing MLLMs to comprehend long multimodal documents.
VITA-Group/Diffusion4D
"Diffusion4D: Fast Spatial-temporal Consistent 4D Generation via Video Diffusion Models", Hanwen Liang*, Yuyang Yin*, Dejia Xu, Hanxue Liang, Zhangyang Wang, Konstantinos N. Plataniotis, Yao Zhao, Yunchao Wei
AiuniAI/Unique3D
Official implementation of Unique3D: High-Quality and Efficient 3D Mesh Generation from a Single Image
xk-huang/segment-caption-anything
[CVPR 24] The repository provides code for running inference and training for "Segment and Caption Anything" (SCA) , links for downloading the trained model checkpoints, and example notebooks / gradio demo that show how to use the model.
modelscope/DiffSynth-Studio
Enjoy the magic of Diffusion models!
fuxiao0719/GeoWizard
[ECCV'24] GeoWizard: Unleashing the Diffusion Priors for 3D Geometry Estimation from a Single Image
buaacyw/MeshAnything
From anything to mesh like human artists. Official impl. of "MeshAnything: Artist-Created Mesh Generation with Autoregressive Transformers"
modelscope/richdreamer
Live Demo:https://modelscope.cn/studios/Damo_XR_Lab/3D_AIGC
FoundationVision/LlamaGen
Autoregressive Model Beats Diffusion: 🦙 Llama for Scalable Image Generation
Jeff-LiangF/streamv2v
Official Pytorch implementation of StreamV2V.
ShareGPT4Omni/ShareGPT4Video
[NeurIPS 2024 D&B Track] An official implementation of ShareGPT4Video: Improving Video Understanding and Generation with Better Captions
SunzeY/Bootstrap3D
Official implementation of Bootstrap3D: Improving 3D Content Creation with Synthetic Data