Pinned Repositories
concept-graphs
Official code release for ConceptGraphs
universal_manipulation_interface
Universal Manipulation Interface: In-The-Wild Robot Teaching Without In-The-Wild Robots
CLIP
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
clip-fields
Teaching robots to respond to open-vocab queries with CLIP and NeRF-like neural fields
Clip-Forge
grok-1
Grok open release
Grounded-SAM-2
Grounded SAM 2: Ground and Track Anything in Videos with Grounding DINO, Florence-2 and SAM 2
Implicit3DUnderstanding_front3d
train im3d for front3d dataset
sam2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
YOLO-World
[CVPR 2024] Real-Time Open-Vocabulary Object Detection
vvcatstar's Repositories
vvcatstar/CLIP
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
vvcatstar/clip-fields
Teaching robots to respond to open-vocab queries with CLIP and NeRF-like neural fields
vvcatstar/Clip-Forge
vvcatstar/grok-1
Grok open release
vvcatstar/Grounded-SAM-2
Grounded SAM 2: Ground and Track Anything in Videos with Grounding DINO, Florence-2 and SAM 2
vvcatstar/Implicit3DUnderstanding_front3d
train im3d for front3d dataset
vvcatstar/sam2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
vvcatstar/YOLO-World
[CVPR 2024] Real-Time Open-Vocabulary Object Detection