Multiscale information fusion for hyperspectral image classification based on hybrid 2d‐3d cnn

  • Hang Gong
  • , Qiuxia Li
  • , Chunlai Li
  • , Haishan Dai
  • , Zhiping He
  • , Wenjing Wang
  • , Haoyang Li
  • , Feng Han
  • , Abudusalamu Tuniyazi
  • , Tingkui Mu

Research output: Contribution to journalArticlepeer-review

62 Scopus citations

Abstract

Hyperspectral images are widely used for classification due to its rich spectral information along with spatial information. To process the high dimensionality and high nonlinearity of hyper-spectral images, deep learning methods based on convolutional neural network (CNN) are widely used in hyperspectral classification applications. However, most CNN structures are stacked verti-cally in addition to using a onefold size of convolutional kernels or pooling layers, which cannot fully mine the multiscale information on the hyperspectral images. When such networks meet the practical challenge of a limited labeled hyperspectral image dataset—i.e., “small sample prob-lem”—the classification accuracy and generalization ability would be limited. In this paper, to tackle the small sample problem, we apply the semantic segmentation function to the pixel‐level hyper-spectral classification due to their comparability. A lightweight, multiscale squeeze‐and‐excitation pyramid pooling network (MSPN) is proposed. It consists of a multiscale 3D CNN module, a squeezing and excitation module, and a pyramid pooling module with 2D CNN. Such a hybrid 2D‐ 3D‐CNN MSPN framework can learn and fuse deeper hierarchical spatial–spectral features with fewer training samples. The proposed MSPN was tested on three publicly available hyperspectral classification datasets: Indian Pine, Salinas, and Pavia University. Using 5%, 0.5%, and 0.5% training samples of the three datasets, the classification accuracies of the MSPN were 96.09%, 97%, and 96.56%, respectively. In addition, we also selected the latest dataset with higher spatial resolution, named WHU‐Hi‐LongKou, as the challenge object. Using only 0.1% of the training samples, we could achieve a 97.31% classification accuracy, which is far superior to the state‐of‐the‐art hyper-spectral classification methods.

Original languageEnglish
Article number2268
JournalRemote Sensing
Volume13
Issue number12
DOIs
StatePublished - 2 Jun 2021

Keywords

  • Convolutional neural network (CNN)
  • Hyperspectral image classification
  • Multiscale information

Fingerprint

Dive into the research topics of 'Multiscale information fusion for hyperspectral image classification based on hybrid 2d‐3d cnn'. Together they form a unique fingerprint.

Cite this