This repository provides the code for the paper Hyperspectral Image Reconstruction Using a Deep Spatial-Spectral Prior. (CVPR 2019).[Link]
Python 3.7.9
CUDA 10.1
TensorFlow 1.14.0
h5py 2.10.0
To train the DNU model for hyperspectral imaging, the datasets should be downloaded to your computer in advance. (e.g., CAVE, KAIST, ICVL, and Harvard.)
For ICVL dataset, you can randomly select 100 spectral images for training and 50 spectral images for testing. For Harvard dataset, you should remove 6 deteriorated spectral images with large-area saturated pixels, and randomly select 35 spectral images for training and 9 spectral images for testing. The training and test images in the ICVL dataset and Harvard dataset are 48 * 48 inclined image blocks.
In the newly increased experiments, the CAVE dataset is used for training and the KAIST dataset is used for test. Then, you should modify the original CAVE and KAIST datasets by spectral interpolation, which have 28 spectral bands ranging from 450nm to 650nm. The patch size for training is 4848, and the patch size for test is 256256.
Then, edit the Training_data_Name
in train.py to indicate the name and path to your dataset. Here is an example:
Training_data_Name = "/PATH/DATSET",
And there should be two directories in your dataset path: [train, test] to indicate which part should be used for training and testing.
The training can be started with the following commands:
# For Harvard and ICVL dataset, using the trian set
python Train.py
# For CAVE and KAIST dataset, using the train set
python Train_cave.py
When the training starts, the trainer will save checkpoints into ./ckpt/
The checkpoint of three DSSP models trained on Harvard and CAVE is provided in ./pretrained_model/. You can directly apply them to subsequent test.
After training, reconstruction image can be generated using the following commands:
# For Harvard and ICVL dataset, using the test set
python Test.py
# For CAVE and KAIST dataset, using the test set
python Test_kaist.py
The results of paper on Harvard Dataset and ICVL Dataset. In this stage, the mask is randomly generated for each batch. And the size of patches is 48 * 48 * 31. In addition, only the central areas with 512* 512 * 31 are compared in testing.
Harvard | ICVL | |
PSNR | 32.84 | 34.13 |
SSIM | 0.979 | 0.992 |
SAM | 0.089 | 0.028 |
For academic reference, we have added some comparisons with the latest methods on CAVE Dataset and KAIST Dataset. Methods for comparison include TSA and DGSM, and our method is completely consistent with the experimental setup of these methods. In addition, we have also increased the comparison of using different masks. In "Real-mask", a given real mask in the range of 0-1 is utilized, which is provided by TSA. In "Binary-mask", the given real mask is rounded to a binary mask. When training the model, a 48 * 48 sub-mask should be randomly derived from the given real mask for each batch. Note that, images with a size of 256 * 256 * 28, matched the given real mask, are used for comparison.
TSA | DGSM | DSSP | ||
Real-mask | Real-mask | Real-mask | Binary-mask | |
PSNR | 31.46 | 32.63 | 32.39 | 32.84 |
SSIM | 0.894 | 0.917 | 0.971 | 0.974 |
SAM | - | - | 0.177 | 0.163 |
If our code is useful in your reseach work, please consider citing our paper.
@inproceedings{DSSP,
title={Hyperspectral Image Reconstruction Using a Deep Spatial-Spectral Prior},
author={Lizhi Wang, Chen Sun, Ying Fu, Min H. Kim and Hua Huang, },
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
pages={8024-8033},
year={2019}
}