vision-and-language-navigation
There are 22 repositories under vision-and-language-navigation topic.
eric-ai-lab/awesome-vision-language-navigation
A curated list for vision-and-language navigation. ACL 2022 paper "Vision-and-Language Navigation: A Survey of Tasks, Methods, and Future Directions"
daqingliu/awesome-vln
A curated list of research papers in Vision-Language Navigation (VLN)
YicongHong/Recurrent-VLN-BERT
Code of the CVPR 2021 Oral paper: A Recurrent Vision-and-Language BERT for Navigation
YicongHong/Discrete-Continuous-VLN
Code and Data of the CVPR 2022 paper: Bridging the Gap Between Learning in Discrete and Continuous Environments for Vision-and-Language Navigation
GT-RIPL/robo-vln
Pytorch code for ICRA'21 paper: "Hierarchical Cross-Modal Agent for Robotics Vision-and-Language Navigation"
YicongHong/Entity-Graph-VLN
Code of the NeurIPS 2021 paper: Language and Visual Entity Relationship Graph for Agent Navigation
YicongHong/Fine-Grained-R2R
Code and data of the Fine-Grained R2R Dataset proposed in the EMNLP 2021 paper Sub-Instruction Aware Vision-and-Language Navigation
CrystalSixone/VLN-GOAT
Repository for Vision-and-Language Navigation via Causal Learning (Accepted by CVPR 2024)
Feliciaxyao/ICML2024-FSTTA
Fast-Slow Test-time Adaptation for Online Vision-and-Language Navigation
joeyy5588/planning-as-inpainting
Planning as In-Painting: A Diffusion-Based Embodied Task Planning Framework for Environments under Uncertainty
PrithivirajDamodaran/vision-language-modelling-series
Companion Repo for the Vision Language Modelling YouTube series - https://bit.ly/3PsbsC2 - by Prithivi Da. Open to PRs and collaborations
raphael-sch/map2seq_vln
Code for ORAR Agent for Vision and Language Navigation on Touchdown and map2seq
eric-ai-lab/FedVLN
[ECCV 2022] Official pytorch implementation of the paper "FedVLN: Privacy-preserving Federated Vision-and-Language Navigation"
xyz9911/FLAME
FLAME: Learning to Navigate with Multimodal LLM in Urban Environments (arXiv:2408.11051)
batra-mlp-lab/vln-chasing-ghosts
Code for 'Chasing Ghosts: Instruction Following as Bayesian State Tracking' published at NeurIPS 2019
joeyy5588/LACMA
LACMA: Language-Aligning Contrastive Learning with Meta-Actions for Embodied Instruction Following
intelligolabs/R2RIE-CE
Official repository of "Mind the Error! Detection and Localization of Instruction Errors in Vision-and-Language Navigation". We present the first dataset - R2R-IE-CE - to benchmark instructions errors in VLN. We then propose a method, IEDL.
marialymperaiou/knowledge-enhanced-multimodal-learning
A list of research papers on knowledge-enhanced multimodal learning
tsujuifu/model_aps
Models of APS
eric-ai-lab/Naivgation-as-wish
Official implementation of the NAACL 2024 paper "Navigation as Attackers Wish? Towards Building Robust Embodied Agents under Federated Learning"
XieZilongAI/HSPR
Hierarchical Spatial Proximity Reasoning for Vision-and-Language Navigation
Gtothemoon/Contrastive-VisionVAE-Follower
Contrastive-VisionVAE-Follower is a model used for multi-modal task called Vision-and-Language Navigation (VLN).