wangjw55
Research Interest: Embodied AI, Visual Language Navigation, Reinforcement Learning. PhD candidate in Tongji University, China
Tongji UniversityShanghai,China
wangjw55's Stars
THUDM/ChatGLM-6B
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
THUDM/ChatGLM3
ChatGLM3 series: Open Bilingual Chat LLMs | 开源双语对话语言模型
datawhalechina/llm-cookbook
面向开发者的 LLM 入门教程,吴恩达大模型系列课程中文版
OpenGVLab/InternVL
[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4o. 接近GPT-4o表现的开源多模态对话模型
huawei-noah/Pretrained-Language-Model
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
DustinWin/clash_singbox-tools
自动同步 mihomo 内核、sing-box 内核、sing-box PuerNya 版内核、Clash dashboard 面板和 AdGuard Home 最新版
OpenRobotLab/GRUtopia
GRUtopia: Dream General Robots in a City at Scale
jacobkrantz/VLN-CE
Vision-and-Language Navigation in Continuous Environments using Habitat
PeterJaq/Awesome-Autonomous-Driving
MarSaKi/ETPNav
[TPAMI 2024] Official repo of "ETPNav: Evolving Topological Planning for Vision-Language Navigation in Continuous Environments"
MarSaKi/VLN-BEVBert
[ICCV 2023} Official repo of "BEVBert: Multimodal Map Pre-training for Language-guided Navigation"
huggingface/jat
General multi-task deep RL Agent
GengzeZhou/NavGPT
[AAAI 2024] Official implementation of NavGPT: Explicit Reasoning in Vision-and-Language Navigation with Large Language Models
YicongHong/Thinking-VLN
Ideas and thoughts about the fascinating Vision-and-Language Navigation
PhoenixZ810/MG-LLaVA
Official repository for paper MG-LLaVA: Towards Multi-Granularity Visual Instruction Tuning(https://arxiv.org/abs/2406.17770).
google-research-datasets/RxR
Room-across-Room (RxR) is a large-scale, multilingual dataset for Vision-and-Language Navigation (VLN) in Matterport3D environments. It contains 126k navigation instructions in English, Hindi and Telugu, and 126k navigation following demonstrations. Both annotation types include dense spatiotemporal alignments between the text and the visual perceptions of the annotators
cshizhe/VLN-DUET
Official implementation of Think Global, Act Local: Dual-scale GraphTransformer for Vision-and-Language Navigation (CVPR'22 Oral).
cshizhe/VLN-HAMT
Official implementation of History Aware Multimodal Transformer for Vision-and-Language Navigation (NeurIPS'21).
GengzeZhou/NavGPT-2
[ECCV 2024] Official implementation of NavGPT-2: Unleashing Navigational Reasoning Capability for Large Vision-Language Models
wzcai99/Pixel-Navigator
Official GitHub Repository for Paper "Bridging Zero-shot Object Navigation and Foundation Models through Pixel-Guided Navigation Skill", ICRA 2024
LYX0501/InstructNav
AirVLN/AirVLN
YicongHong/Fine-Grained-R2R
Code and data of the Fine-Grained R2R Dataset proposed in the EMNLP 2021 paper Sub-Instruction Aware Vision-and-Language Navigation
ggeorgak11/CM2
batra-mlp-lab/vln-sim2real
Code for sim-to-real transfer of a pretrained Vision-and-Language Navigation (VLN) agent to a robot using ROS.
MrZihan/Sim2Real-VLN-3DFF
Official implementation of Sim-to-Real Transfer via 3D Feature Fields for Vision-and-Language Navigation (CoRL'24).
LYX0501/DiscussNav
chengaopro/AZHP
3dlg-hcvc/LAW-VLNCE
Language-Aligned Waypoint (LAW) Supervision for Vision-and-Language Navigation in Continuous Environments
RavenKiller/MLANet