skskgrowl's Stars
openvla/openvla
OpenVLA: An open-source vision-language-action model for robotic manipulation.
robodhruv/visualnav-transformer
Official code and checkpoint release for mobile robot foundation models: GNM, ViNT, and NoMaD.
zubair-irshad/Awesome-Robotics-3D
A curated list of 3D Vision papers relating to Robotics domain in the era of large models i.e. LLMs/VLMs, inspired by awesome-computer-vision, including papers, codes, and related websites
LMD0311/Awesome-World-Model
Collect some World Models for Autonomous Driving papers.
wzzheng/OccWorld
[ECCV 2024] 3D World Model for Autonomous Driving
BraveGroup/Drive-WM
[CVPR 2024] A world model for autonomous driving.
OpenDriveLab/ViDAR
[CVPR 2024 Highlight] Visual Point Cloud Forecasting
NVlabs/OmniDrive
HuaiyuanXu/3D-Occupancy-Perception
[Information Fusion 2024] A Survey on Occupancy Perception for Autonomous Driving: The Information Fusion Perspective
haomo-ai/Cam4DOcc
[CVPR 2024] Cam4DOcc: Benchmark for Camera-Only 4D Occupancy Forecasting in Autonomous Driving Applications
GANWANSHUI/GaussianOcc
GaussianOcc: Fully Self-supervised and Efficient 3D Occupancy Estimation with Gaussian Splatting
NVlabs/BEV-Planner
astra-vision/PaSCo
[CVPR 2024 Oral, Best Paper Award Candidate] Official repository of "PaSCo: Urban 3D Panoptic Scene Completion with Uncertainty Awareness"
GengzeZhou/NavGPT
[AAAI 2024] Official implementation of NavGPT: Explicit Reasoning in Vision-and-Language Navigation with Large Language Models
OpenDriveLab/ELM
[ECCV 2024] Embodied Understanding of Driving Scenarios
yuhengliu02/pyramid-discrete-diffusion
Official implementation of paper "Pyramid Diffusion for Fine 3D Large Scene Generation" (ECCV 2024 Oral)
zympsyche/BevWorld
BraveGroup/LAW
Enhancing End-to-End Autonomous Driving with Latent World Model
UT-Austin-RPL/Coopernaut
Coopernaut: End-to-End Driving with Cooperative Perception for Networked Vehicles
hustvl/osp
[ECCV 2024] Occupancy as Set of Points
GengzeZhou/NavGPT-2
[ECCV 2024] Official implementation of NavGPT-2: Unleashing Navigational Reasoning Capability for Large Vision-Language Models
EMZucas/minidrive
rolsheng/MM-VUFM4DS
A systematic survey of multi-modal and multi-task visual understanding foundation models for driving scenarios
AIR-THU/UniV2X
wuyang98/Text2LiDAR
Arlo0o/HTCL
Official PyTorch Implementation of HTCL (ECCV 2024): Hierarchical Temporal Context Learning for Camera-based Semantic Scene Completion
rruisong/CoHFF
[CVPR 2024] Collaborative Semantic Occupancy Prediction with Hybrid Feature Fusion in Connected Automated Vehicles
Hub-Tian/MambaOcc
MambaOcc: Visual State Space Model for BEV-based Occupancy Prediction with Local Adaptive Reordering
refkxh/C-Instructor
[ECCV 2024] Official implementation of C-Instructor: Controllable Navigation Instruction Generation with Chain of Thought Prompting
qiaoguanren/MARL-CCE
ECCV[2024] "Modelling Competitive Behaviors in Autonomous Driving Under Generative World Model" official implement