Pinned Repositories
airbert-recurrentvln
LaBERT
A length-controllable and non-autoregressive image captioning model.
language-planner
Official Code for "Language Models as Zero-Shot Planners: Extracting Actionable Knowledge for Embodied Agents"
PREVALENT_R2R-1
Apply PREVALENT pretrained model on R2R task
Recurrent-VLN-BERT
Code of the CVPR 2021 Oral paper: A Recurrent Vision-and-Language BERT for Navigation
ScaleVLN
[ICCV 2023 Oral]: Scaling Data Generation in Vision-and-Language Navigation
video-mamba-suite
VideoMamba
VideoMamba: State Space Model for Efficient Video Understanding
VLN-HAMT
VMamba
VMamba: Visual State Space Models,code is based on mamba
biubiuisacat's Repositories
biubiuisacat/airbert-recurrentvln
biubiuisacat/LaBERT
A length-controllable and non-autoregressive image captioning model.
biubiuisacat/language-planner
Official Code for "Language Models as Zero-Shot Planners: Extracting Actionable Knowledge for Embodied Agents"
biubiuisacat/PREVALENT_R2R-1
Apply PREVALENT pretrained model on R2R task
biubiuisacat/Recurrent-VLN-BERT
Code of the CVPR 2021 Oral paper: A Recurrent Vision-and-Language BERT for Navigation
biubiuisacat/ScaleVLN
[ICCV 2023 Oral]: Scaling Data Generation in Vision-and-Language Navigation
biubiuisacat/video-mamba-suite
biubiuisacat/VideoMamba
VideoMamba: State Space Model for Efficient Video Understanding
biubiuisacat/VLN-HAMT
biubiuisacat/VMamba
VMamba: Visual State Space Models,code is based on mamba