TY - JOUR
T1 - Group-spectral superposition and position self-attention transformer for hyperspectral image classification
AU - Zhang, Weitong
AU - Hu, Mingwei
AU - Hou, Sihan
AU - Shang, Ronghua
AU - Feng, Jie
AU - Xu, Songhua
N1 - Publisher Copyright:
© 2024
PY - 2025/3/15
Y1 - 2025/3/15
N2 - At present, the existing Vision Transformer is sensitive to the original spectral data in the hyperspectral semantic segmentation process. Moreover, the weight relationship between the center pixel and the surrounding pixels is not properly handled when using the spatial–spectral feature method. To solve these problems, this paper proposes the Group-Spectral Superposition and Position Self-Attention Transformer (GSPST) for hyperspectral image classification. Firstly, while preserving the original spectral band data, GSPST groups the spectra and superimposes them in the spectral dimension. The superimposed data and the original spectral data are spliced in the channel dimension and then input into the network for learning. Secondly, GSPST processes the original spectral data and the superimposed spectral data in the spatial–spectral features separately using group convolution. This can offset the noise of the low-level features of the adversarial samples in the spectral dimension, thereby reducing its impact on the classification accuracy. Finally, GSPST modified the internal attention mechanism formula of Transformer to avoid the problem of pixels at different positions in hyperspectral images having the same contribution to image classification. A position self-attention mechanism is proposed by introducing a matrix, which can assign different weights to the center pixel and the rest of the pixels to affect its importance for image classification. By introducing position weight information into the self-attention formula to disassemble redundant information in hyperspectral images, thereby improving model robustness. This mechanism can effectively emphasize the role of center pixels in semantic segmentation. Compared with six state-of-the-art algorithms, simulation results demonstrate that GSPST has higher accuracy after adversarial attacks.
AB - At present, the existing Vision Transformer is sensitive to the original spectral data in the hyperspectral semantic segmentation process. Moreover, the weight relationship between the center pixel and the surrounding pixels is not properly handled when using the spatial–spectral feature method. To solve these problems, this paper proposes the Group-Spectral Superposition and Position Self-Attention Transformer (GSPST) for hyperspectral image classification. Firstly, while preserving the original spectral band data, GSPST groups the spectra and superimposes them in the spectral dimension. The superimposed data and the original spectral data are spliced in the channel dimension and then input into the network for learning. Secondly, GSPST processes the original spectral data and the superimposed spectral data in the spatial–spectral features separately using group convolution. This can offset the noise of the low-level features of the adversarial samples in the spectral dimension, thereby reducing its impact on the classification accuracy. Finally, GSPST modified the internal attention mechanism formula of Transformer to avoid the problem of pixels at different positions in hyperspectral images having the same contribution to image classification. A position self-attention mechanism is proposed by introducing a matrix, which can assign different weights to the center pixel and the rest of the pixels to affect its importance for image classification. By introducing position weight information into the self-attention formula to disassemble redundant information in hyperspectral images, thereby improving model robustness. This mechanism can effectively emphasize the role of center pixels in semantic segmentation. Compared with six state-of-the-art algorithms, simulation results demonstrate that GSPST has higher accuracy after adversarial attacks.
KW - Adversarial defense
KW - Adversarial examples
KW - Hyperspectral image classification
KW - Transformer
UR - https://www.scopus.com/pages/publications/85211121953
U2 - 10.1016/j.eswa.2024.125846
DO - 10.1016/j.eswa.2024.125846
M3 - 文章
AN - SCOPUS:85211121953
SN - 0957-4174
VL - 265
JO - Expert Systems with Applications
JF - Expert Systems with Applications
M1 - 125846
ER -