ZhaoYuanQi-1's Stars
graphdeco-inria/gaussian-splatting
Original reference implementation of "3D Gaussian Splatting for Real-Time Radiance Field Rendering"
JonasSchult/Mask3D
Mask3D predicts accurate 3D semantic instances achieving state-of-the-art on ScanNet, ScanNet200, S3DIS and STPLS3D.
TianhaoFu/Awesome-3D-Semantic-Segmentation
Papers, code and datasets about deep learning for 3D Semantic Segmentation.
salesforce/ULIP
XuyangBai/awesome-point-cloud-registration
A curated list of point cloud registration.
danieljf24/awesome-video-text-retrieval
A curated list of deep learning resources for video-text retrieval.
knwng/awesome-vehicle-re-identification
collection of dataset&paper&code on Vehicle Re-Identification
whwu95/Cap4Video
【CVPR'2023 Highlight & TPAMI】Cap4Video: What Can Auxiliary Captions Do for Text-Video Retrieval?
jianzongwu/Awesome-Open-Vocabulary
(TPAMI 2024) A Survey on Open Vocabulary Learning
AnasEmad11/CLAP
Collaborative Learning of Anomalies with Privacy (CLAP) for Unsupervised Video Anomaly Detection: A New Baseline
cbsudux/awesome-human-pose-estimation
A collection of awesome resources in Human Pose estimation.
jxbbb/TOD3Cap
[ECCV 2024] TOD3Cap: Towards 3D Dense Captioning in Outdoor Scenes
ch3cook-fdu/Vote2Cap-DETR
[CVPR 2023] Vote2Cap-DETR and [T-PAMI 2024] Vote2Cap-DETR++; A set-to-set perspective towards 3D Dense Captioning; State-of-the-Art 3D Dense Captioning methods
daveredrum/Scan2Cap
[CVPR 2021] Scan2Cap: Context-aware Dense Captioning in RGB-D Scans
leolyj/3D-VLP
This is the code related to "Context-aware Alignment and Mutual Masking for 3D-Language Pre-training" (CVPR 2023).
3d-vista/3D-VisTA
Official implementation of ICCV 2023 paper "3D-VisTA: Pre-trained Transformer for 3D Vision and Text Alignment"
CurryYuan/ZSVG3D
[CVPR 2024] Visual Programming for Zero-shot Open-Vocabulary 3D Visual Grounding
ZzZZCHS/WS-3DVG
Code for "Distilling Coarse-to-fine Semantic Matching Knowledge for Weakly Supervised 3D Visual Grounding" (ICCV 2023)
liudaizong/Awesome-3D-Visual-Grounding
😎 up-to-date & curated list of awesome 3D Visual Grounding papers, methods & resources.
jianghaojun/Awesome-3D-Vision-and-Language
A collection of 3D vision and language (e.g., 3D Visual Grounding, 3D Question Answering and 3D Dense Caption) papers and datasets.
TheShadow29/awesome-grounding
awesome grounding: A curated list of research papers in visual grounding
daveredrum/ScanRefer
[ECCV 2020] ScanRefer: 3D Object Localization in RGB-D Scans using Natural Language
daveredrum/D3Net
[ECCV2022] D3Net: A Unified Speaker-Listener Architecture for 3D Dense Captioning and Visual Grounding
zlccccc/3DVL_Codebase
[CVPR2022 Oral] 3DJCG: A Unified Framework for Joint Dense Captioning and Visual Grounding on 3D Point Clouds
zyang-ur/SAT
SAT: 2D Semantics Assisted Training for 3D Visual Grounding, ICCV 2021 (Oral)
yanmin-wu/EDA
[CVPR 2023] EDA: Explicit Text-Decoupling and Dense Alignment for 3D Visual Grounding
zlccccc/3DVG-Transformer
[ICCV2021] 3DVG-Transformer: Relation Modeling for Visual Grounding on Point Clouds
Cecile-hi/Multimodal-Learning-with-Alternating-Unimodal-Adaptation
Multimodal Learning Method MLA for CVPR 2024
johnarevalo/gmu-mmimdb
Source code for training Gated Multimodal Units on MM-IMDb dataset
google-research/lanistr