yangjiangeyjg
Ph.D Student @MCG-NJU,Reasearch Intern @OpenGVLab and @researchmm
Nanjing UniversityShanghai, China
yangjiangeyjg's Stars
facebookresearch/segment-anything-2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
DepthAnything/Depth-Anything-V2
[NeurIPS 2024] Depth Anything V2. A More Capable Foundation Model for Monocular Depth Estimation
YvanYin/Metric3D
The repo for "Metric3D: Towards Zero-shot Metric 3D Prediction from A Single Image" and "Metric3Dv2: A Versatile Monocular Geometric Foundation Model..."
ARISE-Initiative/robosuite
robosuite: A Modular Simulation Framework and Benchmark for Robot Learning
openvla/openvla
OpenVLA: An open-source vision-language-action model for robotic manipulation.
Farama-Foundation/Metaworld
Collections of robotics environments geared towards benchmarking multi-task and meta reinforcement learning
facebookresearch/PoseDiffusion
[ICCV 2023] PoseDiffusion: Solving Pose Estimation via Diffusion-aided Bundle Adjustment
TencentARC/Open-MAGVIT2
Open-MAGVIT2: Democratizing Autoregressive Visual Generation
lucidrains/mixture-of-experts
A Pytorch implementation of Sparsely-Gated Mixture of Experts, for massively increasing the parameter count of language models
zubair-irshad/Awesome-Robotics-3D
A curated list of 3D Vision papers relating to Robotics domain in the era of large models i.e. LLMs/VLMs, inspired by awesome-computer-vision, including papers, codes, and related websites
huangwl18/ReKep
ReKep: Spatio-Temporal Reasoning of Relational Keypoint Constraints for Robotic Manipulation
OpenRobotLab/GRUtopia
GRUtopia: Dream General Robots in a City at Scale
AssafSinger94/dino-tracker
Official Pytorch Implementation for “DINO-Tracker: Taming DINO for Self-Supervised Point Tracking in a Single Video”
lucidrains/st-moe-pytorch
Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch
lucidrains/soft-moe-pytorch
Implementation of Soft MoE, proposed by Brain's Vision team, in Pytorch
OpenGVLab/video-mamba-suite
The suite of modeling video with Mamba
real-stanford/umi-on-legs
UMI on Legs: Making Manipulation Policies Mobile with Manipulation-Centric Whole-body Controllers
rail-berkeley/crossformer
bdaiinstitute/theia
Theia: Distilling Diverse Vision Foundation Models for Robot Learning
RayYoh/OCRM_survey
A Survey of Embodied Learning for Object-Centric Robotic Manipulation
Open3DVLab/NeuRodin
[NeurIPS'24] NeuRodin: A Two-stage Framework for High-Fidelity Neural Surface Reconstruction
chernyadev/bigym
Demo-Driven Mobile Bi-Manual Manipulation Benchmark.
Open3DVLab/GigaGS
GigaGS: Scaling up Planar-Based 3D Gaussians for Large Scene Surface Reconstruction
UMass-Foundation-Model/Mod-Squad
2toinf/DecisionNCE
[ICML 2024] The offical Implementation of "DecisionNCE: Embodied Multimodal Representations via Implicit Preference Learning"
Genesis-Embodied-AI/DiffTactile
[ICLR 2024] DiffTactile: A Physics-based Differentiable Tactile Simulator for Contact-rich Robotic Manipulation
rvp-group/learning-where-to-look
Learning Where to Look: Self-supervised Viewpoint Selection for Active Localization using Geometrical Information
AirExo/collector
[ICRA 2024] AirExo: Low-Cost Exoskeletons for Learning Whole-Arm Manipulation in the Wild
HaoyiZhu/RealRobot
Open-source implementations on real robots
Yingdong-Hu/PVM-Robotics
The repository for a thorough empirical evaluation of pre-trained vision model performance across different downstream policy learning methods.