Basic version code for "Attention-Enhanced Cross-modal Localization Between Spherical Images and Point Clouds“

Data

The dataset is built based on KITTI360

Stitch dual-fisheye into spherical image by FFmpeg

ffmpeg -y -i $file -vf v360=dfisheye:e:yaw=-90:ih_fov=187.8:iv_fov=185 -c:v libx265 -b:v 40000k -bufsize 5000k -preset ultrafast -c:a copy out.mp4
ffmpeg -i out.mp4 -f image2 ./$(basename $file .png).png

For global Lidar map making and sub-maps division, refer to this

How to use

the ML environment is based on PyTorch 1.7.0

Before running mytrain.py, specific 2D dataset directory path in mytrain.py or when input the command.

--dataset_root_dir refers to the root directory of the 2D dataset

Specify the 3D dataset directory in ./mycode/msls.py

path_to_3d = " "

Train the network

python mytrain.py