Beruflich Dokumente
Kultur Dokumente
Hyperspectral Classification
ABSTRACT Datasets Results
Dataset-specific Dataset-specific
In this paper, we address the dataset scarcity issue with the Hyperspectral
Analysis Layers
Classification
Layers
1
+ +
3x3 Max Pool ReLU ReLU ReLU ReLU ReLU ReLU ReLU ReLU
C
DO DO
+ : element-wise addition
: Shared layers
2.1. Architecture
Fig. 2: Architecture. Red, blue, green blocks indicate the
Figure 2 shows the proposed cross-domain CNN architecture. non-shared portion of the CNN while the black ones are
shared across all the streams for different datasets. De-
Backbone CNN architecture. For each dataset-specific tailed specifications on the architecture are denoted on the
stream in the proposed architecture, we have used the mod- last (green) stream. B and C indicate the number of spec-
ified version of the 9-layered hyperspectral image classi- tral bands and the number of classes, respectively, which are
fication CNN introduced by Lee and Kwon [4, 5]. The different for all the datasets.
modification has been carried out by adding a batch normal-
ization (BN) layer [8] after each convolutional layer while
removing all the local response normalization layers. The adopting the residual modules within the domain of hyper-
BN layer computes the mean and the variance of the samples spectral image classification.
in each mini-batch and performs normalization, that is, it
fits the samples in the mini-batch to a normal distribution. Cross-domain CNN architecture. To perform an RGB-to-
By introducing the BN, we can bypass the process of data RGB domain adaptation process using CNNs [2, 10], we can
normalization for both training and testing. In addition, bias follow a traditional approach of replacing the classification
terms are no longer required. layers to fit the target dataset. However, when source and tar-
The backbone CNN is a fully convolutional network and get domains are hyperspectral (where each dataset carries its
contains the sequentially connected multi-scale filter bank, unique characteristics as shown in Table 1), simply replacing
one convolutional layer, two residual modules and three con- the classification layer does not work physically. Because of
volutional layers. Each multi-scale filter bank [9] consists this, we have devised dataset-specific layers not only at the
of 1×1, 3×3, and 5×5 filters to analyze the spatial-spectral latter portion of the network, but also before the shared con-
characteristics. Each residual module [3] includes two convo- volutional layers to adaptively intake the different datasets.
lutional layers. The residual modules allow ease of learning The multi-scale filter bank in each dataset-specific stream,
of the deep network. To the best of our knowledge, this ar- which is responsible for analyzing the spatial-spectral char-
chitecture is the first attempt to go deeper than three layers by acteristics, is assigned as the “dataset-specific hyperspectral
Table 1: Specifications for different hyperspectral 1
10
Indian Pines 1
10
Salinas 1
10
University of Pavia
Cross−domain CNN Cross−domain CNN Cross−domain CNN
datasets. Reduced bands are acquired by removing the bands 0
10
Individual CNN Individual CNN Individual CNN
−1
acquired using the same type of sensor can have different re- 10
−1 −1
10 10
duced bands. −2
10
−3 −2 −2
10 10 10
0 50K 100K 0 50K 100K 0 50K 100K
Table 3: Classification accuracy. Pattern Analysis and Machine Intelligence, vol. 38, no.
1, pp. 142–158, 2016.
Dataset Individual CNN Cross-Domain CNN Gain
[3] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual
Indian Pines .907 .922 +.015
learning for image recognition,” in CVPR, 2016.
Salinas .893 .908 +.015
University of Pavia .921 .953 +.032 [4] H. Lee and H. Kwon, “Contextual deep CNN based
hyperspectral classification,” in IGARSS, 2016.
[5] H. Lee and H. Kwon, “Going deeper with contex-
4. CONCLUSION tual CNN for hyperspectral image classification,” IEEE
Transactions on Image Processing, vol. 26, no. 10, pp.
In this paper, we have introduced a novel cross-domain CNN 4843–4855, 2017.
which can concurrently learn and perform the hyperspectral
image classification for multiple datasets. As the shared por- [6] J. Deng, W. Dong, R. Socher, L. J Li, K. Li, and F. Li,
tion of our network is being trained using multiple hyperspec- “ImageNet: A large-scale hierarchical image database,”
tral datasets, the proposed approach is more effective in op- in CVPR, 2009.
timizing the high capacity CNN than the cases where only
a single dataset (domain) is being used. Our approach is [7] M. Everingham, L. V. Gool, C. K. I. Williams, J. Winn,
the first attempt to exploit multiple hyperspectral datasets for and A. Zisserman, “The PASCAL visual object classes
training a CNN in an end-to-end fashion. We have experi- (VOC) challenge,” International Journal on Computer
mentally demonstrated that using the shared layers across the Vision, vol. 88, no. 2, pp. 303–338, 2010.
domains brings notable classification accuracy improvements [8] S. Ioffe and C. Szegedy, “Batch normalization: Acceler-
when compared to the individually trained cases. ating deep network training by reducing internal covari-
ate shift,” in ICML, 2015.
5. REFERENCES [9] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed,
D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabi-
[1] S. Gupta, R. Girshick, P. Arbeláez, and J. Malik, “Learn- novich, “Going deeper with convolutions,” in CVPR,
ing rich features from RGB-D images for object detec- 2015.
tion and segmentation,” in ECCV, 2014.
[10] M. Oquab, L. Bottou, I. Laptev, and J. Sivic, “Learning
[2] R. Girshick, J. Donahue, T. Darrell, and J. Malik, and transferring mid-level image representations using
“Region-based convolutional networks for accurate ob- convolutional neural networks,” in CVPR, 2014.
ject detection and segmentation,” IEEE Transactions on