TikaToka's Stars
HCPLab-SYSU/Embodied_AI_Paper_List
[Embodied-AI-Survey-2024] Paper list and projects for Embodied AI
NVlabs/VILA
VILA is a family of state-of-the-art vision language models (VLMs) for diverse multimodal AI tasks across the edge, data center, and cloud.
Thinklab-SJTU/Bench2Drive
[NeurIPS 2024 Datasets and Benchmarks Track] Closed-Loop E2E-AD Benchmark Enhanced by World Model RL Expert
autowarefoundation/autoware
Autoware - the world's leading open-source software project for autonomous driving
MichalZawalski/embodied-CoT
Embodied Chain of Thought: A robotic policy that reason to solve the task.
NVIDIA/Cosmos-Tokenizer
A suite of image and video neural tokenizers
NVIDIA/Cosmos
Cosmos is a world model development platform that consists of world foundation models, tokenizers and video processing pipeline to accelerate the development of Physical AI at Robotics & AV labs. Cosmos is purpose built for physical AI. The Cosmos repository will enable end users to run the Cosmos models, run inference scripts and generate videos.
leggedrobotics/viplanner
ViPlanner: Visual Semantic Imperative Learning for Local Navigation
Junyi42/monst3r
Official Implementation of paper "MonST3R: A Simple Approach for Estimating Geometry in the Presence of Motion"
naver/mast3r
Grounding Image Matching in 3D with MASt3R
lgsvl/Autoware
Open-Source To Self-Driving.
google-deepmind/mujoco
Multi-Joint dynamics with Contact. A general purpose physics simulator.
sntubix/robocar
RoboCar is a modular, low footprint and easy to deploy autonomous driving software based on ROS2
chen-judge/MapGPT
[ACL 24] The official implementation of MapGPT: Map-Guided Prompting with Adaptive Path Planning for Vision-and-Language Navigation.
allenai/PoliFormer
PoliFormer: Scaling On-Policy RL with Transformers Results in Masterful Navigators
karpathy/LLM101n
LLM101n: Let's build a Storyteller
virattt/ai-hedge-fund
An AI Hedge Fund Team
Genesis-Embodied-AI/Genesis
A generative world for general-purpose robotics & embodied AI learning.
facebookresearch/open-eqa
OpenEQA Embodied Question Answering in the Era of Foundation Models
snumprlab/realfred
Official Implementation of ReALFRED (ECCV'24)
tulerfeng/Awesome-Embodied-Multimodal-LLMs
Latest Advances on Embodied Multimodal LLMs (or Vison-Language-Action Models).
google-research-datasets/RxR
Room-across-Room (RxR) is a large-scale, multilingual dataset for Vision-and-Language Navigation (VLN) in Matterport3D environments. It contains 126k navigation instructions in English, Hindi and Telugu, and 126k navigation following demonstrations. Both annotation types include dense spatiotemporal alignments between the text and the visual perceptions of the annotators
gicheonkang/clip-rt
π + π¦Ύ CLIP-RT: Learning Language-Conditioned Robotic Policies from Natural Language Supervision
CrystalSixone/VLN-GOAT
Repository for Vision-and-Language Navigation via Causal Learning (Accepted by CVPR 2024)
facebookresearch/partnr-planner
A repository accompanying the PARTNR benchmark for using Large Planning Models (LPMs) to solve Human-Robot Collaboration or Robot Instruction Following tasks in the Habitat simulator.
robotpilot/ros-seminar
ROS μμ , μΈλ―Έλ, κ°μ°, κ°μ λ±μ 보쑰 μλ£
GengzeZhou/NavGPT-2
[ECCV 2024] Official implementation of NavGPT-2: Unleashing Navigational Reasoning Capability for Large Vision-Language Models
SakanaAI/AI-Scientist
The AI Scientist: Towards Fully Automated Open-Ended Scientific Discovery π§βπ¬
facebookresearch/sam2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
deeplearningfromscratch2/deep-learning-from-scratch-2
λ°λ°λ₯λΆν° μμνλ λ₯λ¬λ 2! νκ΅μμ μ§νμ€ <3