farlit's Stars
AILab-CVC/YOLO-World
[CVPR 2024] Real-Time Open-Vocabulary Object Detection
zubair-irshad/Awesome-Robotics-3D
A curated list of 3D Vision papers relating to Robotics domain in the era of large models i.e. LLMs/VLMs, inspired by awesome-computer-vision, including papers, codes, and related websites
maitrix-org/Pandora
Pandora: Towards General World Model with Natural Language Actions and Video States
mbzuai-oryx/GeoChat
[CVPR 2024 🔥] GeoChat, the first grounded Large Vision Language Model for Remote Sensing
jlin816/dynalang
Code for "Learning to Model the World with Language." ICML 2024 Oral.
SHI-Labs/VCoder
[CVPR 2024] VCoder: Versatile Vision Encoders for Multimodal Large Language Models
chengzhag/PanFusion
🍳 [CVPR'24 Highlight] Pytorch implementation of "Taming Stable Diffusion for Text to 360° Panorama Image Generation"
haruishi43/equilib
🌎→🗾Equirectangular (360/panoramic) image processing library for Python with minimal dependencies only using Numpy and PyTorch
arnold-benchmark/arnold
[ICCV 2023] Official code repository for ARNOLD benchmark
H-Freax/Awesome-Video-Robotic-Papers
This repository compiles a list of papers related to the application of video technology in the field of robotics! Star⭐ the repo and follow me if you like what you see🤩.
UMass-Foundation-Model/Mod-Squad
XiaohanLei/GaussNav
PyTorch implementation of paper: GaussNav: Gaussian Splatting for Visual Navigation
cshizhe/HM3DAutoVLN
Official implementation of Learning from Unlabeled 3D Environments for Vision-and-Language Navigation (ECCV'22).
HaochenZ11/VLA-3D
OpenRobotLab/OVExp
OVExp: Open Vocabulary Exploration for Object-Oriented Navigation
amazon-science/AdaSlot
Official implementation of the CVPR'24 paper [Adaptive Slot Attention: Object Discovery with Dynamic Slot Number]
cshizhe/robot_sugar
Official implementation of "SUGAR: Pre-training 3D Visual Representations for Robotics" (CVPR'24).
FanScy/BEVInstructor
[ECCV24] Navigation Instruction Generation with BEV Perception and Large Language Models
jinga-lala/DAMEX
Code for "DAMEX: Dataset-aware Mixture-of-Experts for visual understanding of mixture-of-datasets", accepted at Neurips 2023 (Main conference).
iminolee/Awesome-Vision-and-Language-Navigation
A curated list of awesome Vision-and-Language Navigation(VLN) resources (continually updated)
lpercc/HA3D_simulator
Official implementation of Human-Aware Vision-and-Language Navigation: Bridging Simulation to Reality with Dynamic Human Interactions (NeurIPS DB Track'24 Spotlight).
zehao-wang/LAD
Official implementation of Layout-aware Dreamer for Embodied Referring Expression Grounding (AAAI'23).
xyz9911/FLAME
[AAAI-25] FLAME: Learning to Navigate with Multimodal LLM in Urban Environments (arXiv:2408.11051)
qizhust/esceme
iSEE-Laboratory/VLN-PRET
NonvolatileMemory/flash_tree_attn
gmuraleekrishna/SAS
roomtour3d/roomtour3d-NaviLLM
RoomTour3D - Geometry-aware, cheap and automatic data from web videos for embodied navigation
HanqingWangAI/GRUtopia
GRUtopia: Dream General Robots in a City at Scale
HLR/Dual-Action-VLN-CE