kevinchiu19's Stars
IDEA-Research/Grounded-Segment-Anything
Grounded SAM: Marrying Grounding DINO with Segment Anything & Stable Diffusion & Recognize Anything - Automatically Detect , Segment and Generate Anything
facebookresearch/sam2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
FoundationVision/VAR
[NeurIPS 2024 Oral][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ultra-simple, user-friendly yet state-of-the-art* codebase for autoregressive image generation!
gaomingqi/Track-Anything
Track-Anything is a flexible and interactive tool for video object tracking and segmentation, based on Segment Anything, XMem, and E2FGVI.
luosiallen/latent-consistency-model
Latent Consistency Models: Synthesizing High-Resolution Images with Few-Step Inference
DepthAnything/Depth-Anything-V2
[NeurIPS 2024] Depth Anything V2. A More Capable Foundation Model for Monocular Depth Estimation
facebookresearch/co-tracker
CoTracker is a model for tracking any point (pixel) on a video.
Doubiiu/DynamiCrafter
[ECCV 2024, Oral] DynamiCrafter: Animating Open-domain Images with Video Diffusion Priors
prs-eth/Marigold
[CVPR 2024 - Oral, Best Paper Award Candidate] Marigold: Repurposing Diffusion-Based Image Generators for Monocular Depth Estimation
3DTopia/LGM
[ECCV 2024 Oral] LGM: Large Multi-View Gaussian Model for High-Resolution 3D Content Creation.
IDEA-Research/Grounded-SAM-2
Grounded SAM 2: Ground and Track Anything in Videos with Grounding DINO, Florence-2 and SAM 2
Drexubery/ViewCrafter
Official implementation of "ViewCrafter: Taming Video Diffusion Models for High-fidelity Novel View Synthesis"
Tencent/DepthCrafter
DepthCrafter: Generating Consistent Long Depth Sequences for Open-world Videos
Junyi42/monst3r
Official Implementation of paper "MonST3R: A Simple Approach for Estimating Geometry in the Presence of Motion"
ziyc/drivestudio
A 3DGS framework for omni urban scene reconstruction and simulation.
NVlabs/EmerNeRF
PyTorch Implementation of EmerNeRF: Emergent Spatial-Temporal Scene Decomposition via Self-Supervision
nv-tlabs/XCube
[CVPR 2024 Highlight] XCube: Large-Scale 3D Generative Modeling using Sparse Voxel Hierarchies
GaussianCube/GaussianCube
[NeurIPS 2024] GaussianCube: A Structured and Explicit Radiance Representation for 3D Generative Modeling
Jyxarthur/flowsam
[ACCV 2024 (Oral)] Official Implementation of "Moving Object Segmentation: All You Need Is SAM (and Flow)" Junyu Xie, Charig Yang, Weidi Xie, Andrew Zisserman
davyneven/SpatialEmbeddings
Instance Segmentation by Jointly Optimizing Spatial Embeddings and Clustering Bandwidth
stelzner/srt
Independent PyTorch implementation of Scene Representation Transformer
QianyiWu/objsdf
:t-rex: [ECCV‘22] Pytorch implementation of 'Object-Compositional Neural Implicit Surfaces'
wzzheng/OccSora
OccSora: 4D Occupancy Generation Models as World Simulators for Autonomous Driving
snap-research/discoscene
CVPR 2023 Highlight: DiscoScene
coltonstearns/dynamic-gaussian-marbles
jimazeyu/GraspSplats
GraspSplats: Efficient Manipulation with 3D Feature Splatting
lifuguan/GGRt_official
[ECCV 2024] GGRt: Towards Pose-free Generalizable 3D Gaussian Splatting in Real-time
YuLiu-LY/BO-QSA
This repository is the official implementation of Improving Object-centric Learning With Query Optimization
singhgautam/steve
Official code for Slot-Transformer for Videos (STEVE)
Shamdan17/CarFormer
The official repository for the ECCV2024 paper "CarFormer: Self-Driving with Learned Object-Centric Representations"