Reagan1311's Stars
datawhalechina/easy-rl
强化学习中文教程(蘑菇书🍄),在线阅读地址:https://datawhalechina.github.io/easy-rl/
facebookresearch/dinov2
PyTorch code and models for the DINOv2 self-supervised learning method.
SysCV/sam-hq
Segment Anything in High Quality [NeurIPS 2023]
prs-eth/Marigold
[CVPR 2024 - Oral, Best Paper Award Candidate] Marigold: Repurposing Diffusion-Based Image Generators for Monocular Depth Estimation
opendilab/PPOxFamily
PPO x Family DRL Tutorial Course(决策智能入门级公开课:8节课帮你盘清算法理论,理顺代码逻辑,玩转决策AI应用实践 )
jrin771/Everything-LLMs-And-Robotics
The world's largest GitHub Repository for LLMs + Robotics
lxtGH/Awesome-Segmentation-With-Transformer
[T-PAMI-2024] Transformer-Based Visual Segmentation: A Survey
kerrj/lerf
Code for LERF: Language Embedded Radiance Fields
Tsingularity/dift
[NeurIPS'23] Emergent Correspondence from Image Diffusion
facebookresearch/eai-vc
The repository for the largest and most comprehensive empirical study of visual foundation models for Embodied AI (EAI).
real-stanford/scalingup
[CoRL 2023] This repository contains data generation and training code for Scaling Up & Distilling Down
facebookresearch/VLPart
[ICCV2023] VLPart: Going Denser with Open-Vocabulary Part Segmentation
OpenGVLab/Instruct2Act
Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Model
Jiawei-Yang/Denoising-ViT
This is the official code release for our work, Denoising Vision Transformers.
JeffreyYH/robotics-fm-survey
Survey Paper of foundation models for robotics
NVlabs/sim-web-visualizer
Web Based Visualizer for Simulation Environments
zc-alexfan/arctic
[CVPR 2023] Official repository for downloading, processing, visualizing, and training models on the ARCTIC dataset.
NVlabs/RVT
Official Code for RVT-2 and RVT
Junyi42/sd-dino
Official Implementation of paper "A Tale of Two Features: Stable Diffusion Complements DINO for Zero-Shot Semantic Correspondence"
rail-berkeley/bridge_data_v2
afford-motion/afford-motion
Official implementation of CVPR24 highlight paper "Move as You Say, Interact as You Can: Language-guided Human Motion Generation with Scene Affordance"
jiuntian/interactdiffusion
[CVPR 2024] Official repo for "InteractDiffusion: Interaction-Control for Text-to-Image Diffusion Model".
TEA-Lab/Robo-ABC
[ECCV 2024] 🎉 Official repository of "Robo-ABC: Affordance Generalization Beyond Categories via Semantic Correspondence for Robot Manipulation"
soCzech/GenHowTo
Code for the paper "GenHowTo: Learning to Generate Actions and State Transformations from Instructional Videos" published at CVPR 2024
meteorshowers/Sora-Generates-Videos-with-Stunning-Geometrical-Consistency
Sora Generates Videos with Stunning Geometrical Consistency
JasonQSY/3DOI
[ICCV 2023] Understanding 3D Object Interaction from a Single Image
SShowbiz/ZSP3A
tsagkas/click2grasp
Click to Grasp takes calibrated RGB-D images of a tabletop and user-defined part instances in diverse source images as input, and produces gripper poses for interaction, effectively disambiguating between visually similar but semantically different concepts (e.g., left vs right arms).
AntreasAntoniou/infk8s-tutorial
BenjaminJonghyun/RUCGAN
Referenceless User Controllable Semantic Image Synthesis (IJCNN 2023)