/DLAFNet

DLAFNet: Direct LiDAR-Aerial Fusion Network for Semantic Segmentation of 2D Aerial Image and 3D LiDAR Point Cloud

Primary LanguagePythonApache License 2.0Apache-2.0

DLAFNet: Direct LiDAR-Aerial Fusion Network for Semantic Segmentation of 2D Aerial Image and 3D LiDAR Point Cloud

Abstract

Semantic segmentation of high-resolution remote sensing images (RSIs) is developing rapidly. For the semantic segmentation of 2D images, various architectures based on convolutional neural networks have emerged. Nowadays, the accuracy of segmentation has been continuously improved with the introduction of transformer-based models such as SegFormer and Swin Transformer. The methods of semantic segmentation of 3D LiDAR point cloud data, e.g. PointNet and Kernel Point Convolution (KPConv), also develop rapidly due to the emergence of 3D convolution. Very high resolution RGB imagery (RGB) can provide rich spectral information for semantic segmentation, while 3D LiDAR point cloud data can provide depth information. Thus, the semantic segmentation accuracy could be improved by fusing multispectral images and 3D LiDAR point cloud data. The traditional aerial image and LiDAR data fusion uses the DSM or other information obtained from the LiDAR point cloud and fuses the RSIs by adding depth channels. In this paper, we propose a method titled Direct LiDAR-Aerial Fusion Network (DLAFNet) which directly uses the multispectral image and LiDAR point cloud data for semantic Segmentation tasks. In particular, since the sparse features extracted from the KPConv branch are not as essential as features from RGB, we design LiDAR Assisted Attention Module (L-AAM). Our experiments on the modified GRSS18 dataset prove that our method is proper and obtains the best results by comparing with its components and other methods.      

 

Data

The original data is available from 2018 IEEE GRSS Data Fusion Challenge – Fusion of Multispectral LiDAR and Hyperspectral Data | Hyperspectral Image Analysis Lab (uh.edu). The processed 2D GRSS18 data can be found in here(password for BaiduNetdisk: cwpo)  

You can download the transformed data and put it under your path, and change the path of the dataset in the dataset configuration file \DLAFNet\configs\_base_\datasets\GRSS18_5.py to the path you have set yourself.    

You can refer to the KPConv for 3D branching data processing, and get the features of different stages and process them through the pseudo-code of the paper, and save them as corresponding files for your convenience. We also provide the test data, you can get them from here and refer to it to construct your own train data.

Results

Data Model OA(%) mIoU(%)
RGB FCN 75.03 28.23
RGB DANet 78.08 28.66
RGB PSPNet 78.45 30.97
RGB Deeplab v3 78.45 26.37
RGB SegFormer 78.67 30.83
LiDAR PointNet 71.40 15.83
LiDAR PointNet++ 75.32 18.06
LiDAR KPConv 76.97 22.57
RGB-LiDAR EDFT 79.25 31.15
RGB-LiDAR DLAFNet 79.88 31.94

DLAFNet best mIoU results for models able to be found from here

Installation

Please refer to get_started.md for installation

Training

# Single-gpu training 
python tools\train.py configs\dlafnet\dlafnet_mit_b0_512_80k_GRSS18_5.py 

Acknowledgement

We acknowledged the IEEE Geoscience and Remote Sensing Society Image Analysis and Data Fusion Technical Committee and the Hyperspectral Image Analysis Lab at the University of Houston for the GRSS18 dataset. We also acknowledged paper EDFT and forked its code. We are also grateful for the use of KPConv

Citation

@INPROCEEDINGS{10282837,
  author={Liu, Wei and Wang, He and Qiao, Yicheng and Liang, Bin and Yang, Junli and Zhang, Haopeng}, 
  booktitle={IGARSS 2023 - 2023 IEEE International Geoscience and Remote Sensing Symposium},  
  title={Dlafnet: A Direct Fusion Method of 2D Aerial Image and 3D Lidar Point Cloud for Semantic Segmentation}, 
  year={2023},
  volume={},
  number={},
  pages={5922-5925},
  doi={10.1109/IGARSS52108.2023.10282837}}