TY - JOUR
T1 - Infrared and Visible Cross-Modal Image Retrieval through Shared Features
AU - Liu, Fangcen
AU - Gao, Chenqiang
AU - Sun, Yongqing
AU - Zhao, Yue
AU - Yang, Feng
AU - Qin, Anyong
AU - Meng, Deyu
N1 - Publisher Copyright:
© 1991-2012 IEEE.
PY - 2021/11/1
Y1 - 2021/11/1
N2 - Image retrieval is one of the key techniques of computer vision, and has been studied for a long time. Nevertheless, little attention is paid to infrared and visible cross-modal retrieval which can be widely used in various applications, e.g., infrared and visible surveillance systems. In this paper, we propose a shared features based infrared-visible cross-modal image retrieval method. The similar visual features are extracted from infrared and visible images as the shared features, and the Euclidean distance is used to measure the similarity between these features. The core of the proposed method comes from three aspects: 1) Feature separation network can separate image features into shared features and exclusive features; 2) Maximum Mean Discrepancy (MMD) loss is employed to constrain the distribution of shared features, which can reduce the retrieval error caused by different imaging angles and similarity of infrared images. 3) The cross-layer fusion encoder compensates for the context loss in the convolution of infrared images. Experimental results on the Infrared-Visible dataset demonstrate the proposed method is effective and outperforms the state-of-the-art approaches.
AB - Image retrieval is one of the key techniques of computer vision, and has been studied for a long time. Nevertheless, little attention is paid to infrared and visible cross-modal retrieval which can be widely used in various applications, e.g., infrared and visible surveillance systems. In this paper, we propose a shared features based infrared-visible cross-modal image retrieval method. The similar visual features are extracted from infrared and visible images as the shared features, and the Euclidean distance is used to measure the similarity between these features. The core of the proposed method comes from three aspects: 1) Feature separation network can separate image features into shared features and exclusive features; 2) Maximum Mean Discrepancy (MMD) loss is employed to constrain the distribution of shared features, which can reduce the retrieval error caused by different imaging angles and similarity of infrared images. 3) The cross-layer fusion encoder compensates for the context loss in the convolution of infrared images. Experimental results on the Infrared-Visible dataset demonstrate the proposed method is effective and outperforms the state-of-the-art approaches.
KW - Infrared-visible image retrieval
KW - cross-layer fusion encoder
KW - maximum mean discrepancy
KW - shared feature
UR - https://www.scopus.com/pages/publications/85099222610
U2 - 10.1109/TCSVT.2020.3048945
DO - 10.1109/TCSVT.2020.3048945
M3 - 文章
AN - SCOPUS:85099222610
SN - 1051-8215
VL - 31
SP - 4485
EP - 4496
JO - IEEE Transactions on Circuits and Systems for Video Technology
JF - IEEE Transactions on Circuits and Systems for Video Technology
IS - 11
ER -