fanjing8's Stars
PRBonn/kiss-slam
A LiDAR SLAM system that just works
StanfordVL/3DSceneGraph
The data skeleton from "3D Scene Graph: A Structure for Unified Semantics, 3D Space, and Camera" http://3dscenegraph.stanford.edu
microsoft/CogACT
A Foundational Vision-Language-Action Model for Synergizing Cognition and Action in Robotic Manipulation
PKU-HMI-Lab/Hybrid-VLA
HybridVLA: Collaborative Diffusion and Autoregression in a Unified Vision-Language-Action Model
naver/goal-co
naver/roma
RoMa: A lightweight library to deal with 3D rotations in PyTorch.
IDEA-Research/Grounded-SAM-2
Grounded SAM 2: Ground and Track Anything in Videos with Grounding DINO, Florence-2 and SAM 2
gramuah/ros4vsn
Evaluation of Visual Semantic Navigation Models in Real Robots
luhengshiwo/LLMForEverybody
每个人都能看懂的大模型知识分享,LLMs春/秋招大模型面试前必看,让你和面试官侃侃而谈
google-research/big_vision
Official codebase used to develop Vision Transformer, SigLIP, MLP-Mixer, LiT and more.
Yixin-F/better_fastlio2
Postgraduate Thesis: fast_lio_sam + dynamic removal (T-GRS 2024) + multi-session mapping (ICRA 2022 Kim) + object-level update + online relocalization (ICRA 2025) + real-world application (MD-LVIO)
sdwyc/ROLO
This is a ROS package for lidar odometry implementation using rotation optimization method.
NVIDIA/NeMo
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
chengwei0427/Adaptive-LIO
[JIOT 2024] Adaptive-LIO: Enhancing Robustness and Precision through Environmental Adaptation in LiDAR Inertial Odometry
CoMMALab/pRRTC
GPU-Parallel RRT-Connect
xuxw98/ESAM
[ICLR 2025, Oral] EmbodiedSAM: Online Segment Any 3D Thing in Real Time
qimao7213/TGH-Planner
Topology-Guided Hierarchical Planner for Nonholonomic Robots in Unknown and Complex Environments
yanmin-wu/OpenGaussian
[NeurIPS 2024] OpenGaussian: Towards Point-Level 3D Gaussian-based Open Vocabulary Understanding
mannaandpoem/OpenManus
No fortress, purely open ground. OpenManus is Coming.
OpenRobotLab/GRUtopia
GRUtopia: Dream General Robots in a City at Scale
ybgdgh/VLN-Game
A new zero-shot framework to explore and search for the language descriptive targets in unknown environment based on Large Vision Language Model.
OpenGVLab/VisionLLM
VisionLLM Series
ybgdgh/L3MVN
Leveraging Large Language Models for Visual Target Navigation
xz00/fast-lio2-map-based-localization
map-based localization.Modified from fast-lio2.
facebookresearch/fast3r
[CVPR 2025] Fast3R: Towards 3D Reconstruction of 1000+ Images in One Forward Pass
SenseRoboticsLab/CURL-MAP
CURL-MAP: Continuous Mapping and Positioning with CURL Representation
AIT-Assistive-Autonomous-Systems/ros2bag_tools
Tool extensions for ros2bag cli
QwenLM/Qwen2.5-VL
Qwen2.5-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.
Deep-Agent/R1-V
Witness the aha moment of VLM with less than $3.
Liuziyu77/Visual-RFT
Official repository of ’Visual-RFT: Visual Reinforcement Fine-Tuning’