Distributed Fusion of Multiple Model Estimators Using Minimum Forward Kullback-Leibler Divergence Sum

Research output: Contribution to journalArticlepeer-review

6 Scopus citations

Abstract

The problem of distributed fusion of Gaussian mixture models (GMMs) provided by the local multiple model (MM) estimators is addressed in this article. Taking GMMs instead of combined Gaussian assumed probability density functions (pdfs) as the output of local MM estimators can retain more detailed (or internal) information about local estimations, but the accompanying challenge is to perform the fusion of GMMs. For this problem, a distributed fusion framework of GMMs under the minimum forward Kullback-Leibler (KL) divergence sum criterion is proposed first. Then, because the KL divergence between GMMs is not analytically tractable, two suboptimal distributed fusion algorithms are further developed within this framework. These two fusion algorithms all have closed forms. Numerical examples verify their effectiveness in terms of both computational efficiency and estimation accuracy.

Original languageEnglish
Pages (from-to)2934-2947
Number of pages14
JournalIEEE Transactions on Aerospace and Electronic Systems
Volume60
Issue number3
DOIs
StatePublished - 1 Jun 2024

Keywords

  • Distributed fusion
  • Kullback-Leibler (KL) divergence
  • maneuvering target tracking
  • multiple model (MM) estimation

Fingerprint

Dive into the research topics of 'Distributed Fusion of Multiple Model Estimators Using Minimum Forward Kullback-Leibler Divergence Sum'. Together they form a unique fingerprint.

Cite this