Knowledge Synergy Learning for Multi-Modal Tracking

Research output: Contribution to journalArticlepeer-review

12 Scopus citations

Abstract

Benefiting from the rich information provided by different modalities, multi-modal tracking has shown significant improvements compared to single-modal tracking. However, in practical applications, multi-modal tracking still faces two major challenges. Firstly, it is crucial to effectively integrate the complementary information from different modalities in order to improve tracking performance. Secondly, as trackers are often deployed in dynamic environments, it is difficult to ensure complete multi-modal data. Thus, handling modal-missing issues is essential to achieve robust and reliable tracking. To address these challenges, this paper proposes a Knowledge Synergy Network (KSNet) that integrates multi-modal features into a comprehensive representation and incorporates a modal compensation mechanism to handle modal-missing issues. With this framework, a multi-modal tracker (KSTrack) is built and trained using multi-modal data. KSTrack is capable of handling both complete and incomplete multi-modal data during inference. Comprehensive experiments on four large-scale RGB-Thermal (RGB-T) and RGB-Depth (RGB-D) benchmarks show that KSTrack surpasses state-of-the-art multi-modal trackers when using multi-modal data and outperforms single-modal trackers by a large margin when using single-modal data.

Original languageEnglish
Pages (from-to)5519-5532
Number of pages14
JournalIEEE Transactions on Circuits and Systems for Video Technology
Volume34
Issue number7
DOIs
StatePublished - 2024

Keywords

  • Multi-modal tracking
  • knowledge synergy learning
  • modality missing
  • recurrent modal compensation

Fingerprint

Dive into the research topics of 'Knowledge Synergy Learning for Multi-Modal Tracking'. Together they form a unique fingerprint.

Cite this