embodied-agent

There are 57 repositories under embodied-agent topic.

  • hyp1231/awesome-llm-powered-agent

    Awesome things about LLM-powered agents. Papers / Repos / Blogs / ...

  • zchoi/Awesome-Embodied-Robotics-and-Agent

    This is a curated list of "Embodied AI or robot with Large Language Models" research. Watch this repository for the latest updates! šŸ”„

  • TheShadow29/awesome-grounding

    awesome grounding: A curated list of research papers in visual grounding

  • Autonomous-Agents

    tmgthb/Autonomous-Agents

    Autonomous Agents (LLMs) research papers. Updated Daily.

  • eric-ai-lab/awesome-vision-language-navigation

    A curated list for vision-and-language navigation. ACL 2022 paper "Vision-and-Language Navigation: A Survey of Tasks, Methods, and Future Directions"

  • kyegomez/RT-2

    Democratization of RT-2 "RT-2: New model translates vision and language into action"

    Language:Python52672466
  • haoranD/Awesome-Embodied-AI

    A curated list of awesome papers on Embodied AI and related research/industry-driven resources.

  • RobotecAI/rai

    RAI is a vendor-agnostic agentic framework for robotics, utilizing ROS 2 tools to perform complex actions, defined scenarios, free interface execution, log summaries, voice interaction and more.

    Language:Python407716351
  • allenact

    allenai/allenact

    An open source framework for research in Embodied-AI from AI2.

    Language:Python37099458
  • zju-vipa/Odyssey

    Odyssey: Empowering Minecraft Agents with Open-World Skills

    Language:Python34851319
  • mbodiai/embodied-agents

    Seamlessly integrate state-of-the-art transformer models into robotics stacks

    Language:Python25631529
  • Yuxing-Wang-THU/SurveyBrainBody

    Embodied Co-Design for Rapidly Evolving Agents: Taxonomy, Frontiers, and Challenges

  • Gary3410/TaPA

    [arXiv 2023] Embodied Task Planning with Large Language Models

    Language:Python19251213
  • iris0329/SeeGround

    [CVPR'25] SeeGround: See and Ground for Zero-Shot Open-Vocabulary 3D Visual Grounding

    Language:Python1827214
  • hanxunyu/Inst3D-LMM

    [CVPR 2025 HighlightšŸ”„] Official code repository for "Inst3D-LMM: Instance-Aware 3D Scene Understanding with Multi-modal Instruction Tuning"

    Language:Python1187155
  • AoqunJin/Awesome-VLA-Post-Training

    A collection of vision-language-action model post-training methods.

  • Zhoues/MineDreamer

    [IROS'25 Oral & NeurIPSw'24] Official implementation of "MineDreamer: Learning to Follow Instructions via Chain-of-Imagination for Simulated-World Control "

    Language:Python97676
  • Linketic/TC-Light

    [NeurIPS`25] TC-Light: Temporally Coherent Generative Rendering for Realistic World Transfer

    Language:Python95032
  • bigai-nlco/langsuite

    Official Repo of LangSuitE

    Language:Python84613
  • mazpie/genrl

    [NeurIPS 2024] GenRL: Multimodal-foundation world models enable grounding language and video prompts into embodied domains, by turning them into sequences of latent world model states. Latent state sequences can be decoded using the decoder of the model, allowing visualization of the expected behavior, before training the agent to execute it.

    Language:Python832223
  • wendell0218/GVA-Survey

    Official repository of the paper "Generalist Virtual Agents: A Survey on Autonomous Agents Across Digital Platforms"

  • declare-lab/Emma-X

    Emma-X: An Embodied Multimodal Action Model with Grounded Chain of Thought and Look-ahead Spatial Reasoning

    Language:Python75186
  • Josh00-Lu/DiffusionVeteran

    [ICLR 2025 Spotlight] Official PyTorch Implementation of "What Makes a Good Diffusion Planner for Decision Making?"

    Language:Python713
  • xyz9911/FLAME

    [AAAI-25 Oral] Official Implementation of "FLAME: Learning to Navigate with Multimodal LLM in Urban Environments"

    Language:Python651102
  • ZJLAB-AMMI/LLM4Teach

    Python code to implement LLM4Teach, a policy distillation approach for teaching reinforcement learning agents with Large Language Model

    Language:Python512315
  • Josh00-Lu/BodyGen

    [ICLR 2025 Spotlight] Official PyTorch Implementation of "BodyGen: Advancing Towards Efficient Embodiment Co-Design"

    Language:Python45323
  • BioRAILab/Neural-Brain-for-Embodied-Agents

    Project Page for Paper "Neural Brain: A Neuroscience-inspired Framework for Embodied Agents".

  • automatika-robotics/embodied-agents

    EmbodiedAgents is a fully-loaded ROS2 based framework for creating interactive physical agents that can understand, remember, and act upon contextual information from their environment.

    Language:Python39
  • rese1f/STEVE

    [ECCV 2024] STEVE in Minecraft is for See and Think: Embodied Agent in Virtual Environment

  • OceanGPT/OceanGym

    OceanGym: A Benchmark Environment for Underwater Embodied Agents

    Language:Python37
  • opendilab/OpenPaL

    Building open-ended embodied agent in battle royale FPS game

  • Josh00-Lu/Habi

    [ICML 2025 Poster] Official PyTorch Implementation of "Habitizing Diffusion Planning for Efficient and Effective Decision Making"

    Language:Python35
  • CEC-Agent/CEC

    Official Implementation of NeurIPS'23 Paper "Cross-Episodic Curriculum for Transformer Agents"

    Language:Python31014
  • LoopMind-AI/loopquest

    A Production Tool for Embodied AI

    Language:Python29221
  • airs-cuhk/airsoul

    Next-gen Foundation Model for Embodied AI

    Language:Python22228
  • robosense2025/track2

    Track 2: Social Navigation