/xView2_fifth_place

Dual-HRNet

Primary LanguagePythonMIT LicenseMIT

Dual-HRNet for Building Localization and DamageClassification

Abstact

Comparison of pre-disaster and post-disaster on satellites imagery play an important roles for the humanitarian assistance and disaster relief (HADR) efforts while severe global events. Automatic comparison using AI helps identify the affected area. In this study, we propose Dual-HRNets for both localizing buildings and classifying their damage level on the satellite imagery simultaneously. Since the goal of this study is to predict the location of buildings on the pre-disaster image as well as classify the damage level in accordance with the post-disaster image, our Dual-HRNets take a pair of pre- and post-disaster images over each HRNet. Each intermediate stage of Dual-HRNet, the features of each HRNet are fused by a fusion block. Our experiments present a good performance of Dual-HRNets for the task of both localization and classification. Our approach with Dual-HRNet achieved the 5th place in xView2 challenge over 3,500 participants.

Qualitative results of Dual-HRNet

From left to right, the figures shows pre-disaster images, results of Dual-HRnet and post-disaster images. White, yellow, orange and red regions represent nodamage, minor damage, major damage and destroyed respectively.

Quick start

Install

Install PyTorch=1.4 following the official instructions.

Please refer to submission/Dockerfile

Train and test

Please batch size in configuration file(dual-hrnet.yaml) for your environment.

Train

python train_net.py --config_file configs/dual-hrnet.yaml \
                    --ckpt_save_dir=PATH_TO_SAVE_CHECKPOINTS

Test

python test_net.py --config_file configs/dual-hrnet.yaml \
                   --ckpt_path=PATH_TO_LOAD_CHECKPOINTS

License

Dual-HRNet is released under the MIT license.

Acknowledgement

We modified code based on HRNet Semantic Segmentation(https://github.com/HRNet/HRNet-Semantic-Segmentation)