Li-XD-Pro's Stars
bdaiinstitute/vlfm
The repository provides code associated with the paper VLFM: Vision-Language Frontier Maps for Zero-Shot Semantic Navigation (ICRA 2024)
cassieqiuyd/MJOLNIR
Python implementation of the paper Learning hierarchical relationships for object-goal navigation
devendrachaplot/Object-Goal-Navigation
Pytorch code for NeurIPS-20 Paper "Object Goal Navigation using Goal-Oriented Semantic Exploration"
gunagg/zson
ZSON: Zero-Shot Object-Goal Navigation using Multimodal Goal Embeddings. NeurIPS 2022
jialuli-luka/PanoGen
Code and Data for Paper: PanoGen: Text-Conditioned Panoramic Environment Generation for Vision-and-Language Navigation
facebookresearch/habitat-lab
A modular high-level library to train embodied AI agents across a variety of tasks and environments.
pioneer-innovation/Zero-Shot-Object-Navigation
Gabesarch/TIDEE
code for TIDEE: Novel Room Reorganization using Visuo-Semantic Common Sense Priors
stepjam/RLBench
A large-scale benchmark and learning environment.
TheMTank/cups-rl
Customisable Unified Physical Simulations (CUPS) for Reinforcement Learning. Experiments run on the ai2thor environment (http://ai2thor.allenai.org/) e.g. using A3C, RainbowDQN and A3C_GA (Gated Attention multi-modal fusion) for Task-Oriented Language Grounding (tasks specified by natural language instructions) e.g. "Pick up the Cup or else"
lbaa2022/LLMTaskPlanning
LoTa-Bench: Benchmarking Language-oriented Task Planners for Embodied Agents (ICLR 2024)
HCPLab-SYSU/Embodied_AI_Paper_List
[Embodied-AI-Survey-2024] Paper list and projects for Embodied AI
facebookresearch/habitat-sim
A flexible, high-performance 3D simulator for Embodied AI research.
snumprlab/realfred
Official Implementation of ReALFRED (ECCV'24)
alfworld/alfworld
ALFWorld: Aligning Text and Embodied Environments for Interactive Learning
alexa/teach
TEACh is a dataset of human-human interactive dialogues to complete tasks in a simulated household environment.
StanfordVL/bddl
pucrs-automated-planning/pddl-parser
:snake: Classical Planning in Python
feifeiobama/Awesome-Embodied-Instruction-Following
A leaderboard for Embodied Instruction Following papers and BibTeX entries
hitachi-rd-cv/prompter-alfred
Prompter for Embodied Instruction Following
OpenRobotLab/EmbodiedScan
[CVPR 2024 & NeurIPS 2024] EmbodiedScan: A Holistic Multi-Modal 3D Perception Suite Towards Embodied AI
gistvision/moca
Code and models of MOCA (Modular Object-Centric Approach) proposed in "Factorizing Perception and Policy for Interactive Instruction Following" (ICCV 2021). We address the task of long horizon instruction following with a modular architecture that decouples a task into visual perception and action policy prediction.
askforalfred/alfred
ALFRED - A Benchmark for Interpreting Grounded Instructions for Everyday Tasks
omron-sinicx/ViLaIn
An official implementation of Vision-Language Interpreter (ViLaIn)
NVlabs/progprompt-vh
ProgPrompt for Virtualhome
yding25/GPT-Planner
Paper: Integrating Action Knowledge and LLMs for Task Planning and Situation Handling in Open Worlds
urchade/graph-neural-nets
Graph neural networks tutorial in pytorch (GCN, GAT, Node2vec, GraphSAge, ClusterGCN, ...)
shenweichen/GraphEmbedding
Implementation and experiments of graph embedding algorithms.
pyg-team/pytorch_geometric
Graph Neural Network Library for PyTorch
bertjiazheng/awesome-scene-understanding
😎 A list of awesome scene understanding papers.