Pinned Repositories
Awesome-LLM-Robotics
A comprehensive list of papers using large language/multi-modal models for Robotics/RL, including papers, codes, and related websites
bottom-up-attention
Bottom-up attention model for image captioning and VQA, based on Faster R-CNN and Visual Genome
Discrete-Continuous-VLN
Code and Data of the CVPR 2022 paper: Bridging the Gap Between Learning in Discrete and Continuous Environments for Vision-and-Language Navigation
Draw-the-line
A graph plotter that works with any pdf containing one (or more) graphs :chart_with_upwards_trend:
Ego2Map-NaViT
Official Implementation of Learning Navigational Visual Representations with Semantic Map Supervision (ICCV2023)
Entity-Graph-VLN
Code of the NeurIPS 2021 paper: Language and Visual Entity Relationship Graph for Agent Navigation
Fine-Grained-R2R
Code and data of the Fine-Grained R2R Dataset proposed in the EMNLP 2021 paper Sub-Instruction Aware Vision-and-Language Navigation
mae
PyTorch implementation of MAE https//arxiv.org/abs/2111.06377
Recurrent-VLN-BERT
Code of the CVPR 2021 Oral paper: A Recurrent Vision-and-Language BERT for Navigation
Thinking-VLN
Ideas and thoughts about the fascinating Vision-and-Language Navigation
YicongHong's Repositories
YicongHong/Thinking-VLN
Ideas and thoughts about the fascinating Vision-and-Language Navigation
YicongHong/Recurrent-VLN-BERT
Code of the CVPR 2021 Oral paper: A Recurrent Vision-and-Language BERT for Navigation
YicongHong/Discrete-Continuous-VLN
Code and Data of the CVPR 2022 paper: Bridging the Gap Between Learning in Discrete and Continuous Environments for Vision-and-Language Navigation
YicongHong/Entity-Graph-VLN
Code of the NeurIPS 2021 paper: Language and Visual Entity Relationship Graph for Agent Navigation
YicongHong/Fine-Grained-R2R
Code and data of the Fine-Grained R2R Dataset proposed in the EMNLP 2021 paper Sub-Instruction Aware Vision-and-Language Navigation
YicongHong/Ego2Map-NaViT
Official Implementation of Learning Navigational Visual Representations with Semantic Map Supervision (ICCV2023)
YicongHong/Awesome-LLM-Robotics
A comprehensive list of papers using large language/multi-modal models for Robotics/RL, including papers, codes, and related websites
YicongHong/bottom-up-attention
Bottom-up attention model for image captioning and VQA, based on Faster R-CNN and Visual Genome
YicongHong/Draw-the-line
A graph plotter that works with any pdf containing one (or more) graphs :chart_with_upwards_trend:
YicongHong/mae
PyTorch implementation of MAE https//arxiv.org/abs/2111.06377
YicongHong/REVERIE
REVERIE: Remote Embodied Visual Referring Expression in Real Indoor Environments
YicongHong/sound-spaces
A first-of-its-kind acoustic simulation platform for audio-visual embodied AI research. It supports training and evaluating multiple tasks and applications.
YicongHong/starter-hugo-academic
🎓 Hugo Academic Theme 创建一个学术网站. Easily create a beautiful academic résumé or educational website using Hugo, GitHub, and Netlify.
YicongHong/website
The DB Group Website
YicongHong/YicongHong.github.io
My academic personal website. https://yiconghong.me