This repository contains the open-source code for the paper "Semantic SLAM for mobile robots in dynamic environments based on visual camera sensors" by Qi Zhang and Changdi Li, published in the Measurement Science and Technology journal in 2023. You can access the original paper here.
(Save your time to read the whole paper, I show the results in a more informatic and intereting way, like a tiktok short video!):
b站: https://www.bilibili.com/video/BV1jP411R7S8/?share_source=copy_web&vd_source=12d45e19826de0471391d3db9d6c9491
微博: https://weibo.com/7348088749/N0fxQaCFQ
抖音: https://v.douyin.com/U89e9po/
小红书:https://www.xiaohongshu.com/user/profile/630844c0000000001200f745
Youtube: https://youtu.be/9g6-N-uYeno
TikTok: https://www.tiktok.com/@zhangqi11/video/7232040785514958107?is_from_webapp=1&sender_device=pc&web_id=7232038248343225883
Visual simultaneous localization and mapping (vSLAM) is inherently constrained by the static world assumption, which renders success in the presence of dynamic objects rather challenging. In our work, we propose a real-time semantic vSLAM system designed for both indoor and outdoor dynamic environments.
- Based on pretain-model of YOLOX
- Real-time semantic vSLAM system suitable for both indoor and outdoor environments.
- Object detection across 80 categories.
- Motion consistency checks for identifying outliers.
- Specialized handling of human and non-human objects.
- Algorithms for assessing human postures (seated or standing).
- Automatic threshold adjustments for non-human objects.
- Tested on TUM, Bonn RGB-D, and KITTI datasets.
Our system is evaluated on the indoor TUM and Bonn RGB-D datasets, with further testing conducted on the outdoor stereo KITTI dataset. The results reveal that our SLAM outperforms most SLAM systems in dynamic environments. We also tested our system in real-world environments with a monocular camera, demonstrating its robustness and universality across diverse settings.
If you use this code in your research, please cite our paper:
@article{Zhang_2023,
doi = {10.1088/1361-6501/acd1a4},
url = {https://dx.doi.org/10.1088/1361-6501/acd1a4},
year = {2023},
month = {may},
publisher = {IOP Publishing},
volume = {34},
number = {8},
pages = {085202},
author = {Qi Zhang and Changdi Li},
title = {Semantic SLAM for mobile robots in dynamic environments based on visual camera sensors},
journal = {Measurement Science and Technology},
}
pip install torch==1.8.2+cu111 torchvision==0.9.2+cu111 torchaudio===0.8.2 -f https://download.pytorch.org/whl/lts/1.8/torch_lts.html
pip install torch==1.10.0+cu113 torchvision==0.11.1+cu113 torchaudio===0.10.0+cu113 -f https://download.pytorch.org/whl/cu113/torch_stable.html -i http://pypi.douban.com/simple/ --trusted-host pypi.douban.com
sudo dpkg -i nv-tensorrt-repo-ubuntu2004-cuda11.6-trt8.4.0.6-ea-20220212_1-1_amd64*
cd /
cd ./var
cd ./nv*
sudo apt-key add 7fa2af80.pub
sudo apt-get update
sudo apt-get install tensorrt
python3 -m pip install numpy
sudo apt-get install python3-libnvinfer-dev
python3 -m pip install protobuf
sudo apt-get install uff-converter-tf
python3 -m pip install numpy onnx
sudo apt-get install onnx-graphsurgeon