Skip to main navigation Skip to search Skip to main content

Multimodal Information Fusion Approach for Noncontact Heart Rate Estimation Using Facial Videos and Graph Convolutional Network

  • Zijie Yue
  • , Shuai Ding
  • , Shanlin Yang
  • , Linjie Wang
  • , Yinghui Li
  • Hefei University of Technology
  • Chinese Astronaut Research and Training Center

Research output: Contribution to journalArticlepeer-review

24 Scopus citations

Abstract

Heart rate (HR) is a critical signal for reflecting human physical and mental conditions, and it is beneficial for diagnosing neurological and cardiovascular diseases due to its excellent accessibility. However, traditional HR measurement devices have limited usability and convenience. Recent studies have shown that the optical absorption variation of human skin due to blood volume variation in cardiac cycles can be acquired from facial videos and used to estimate HR in a noncontact manner. However, the advanced noncontact HR estimation approaches are based on a single HR information source, resulting in unsatisfactory estimation results due to noise corruption and insufficient information. To address these problems, this article proposes a multimodal information fusion framework for noncontact HR estimation. First, feature representation maps are used to effectively extract periodic signals from facial visible-light and thermal infrared videos. Then, a temporal-information-aware HR feature extraction network (THR-Net) for encoding discriminative spatiotemporal information from the representation maps is presented. Finally, based on a graph convolution network (GCN), an information fusion model is proposed for feature integration and HR estimation. Experimental and evaluation results of five different metrics on two datasets show that the proposed approach outperforms the state-of-the-art approaches. This article demonstrates the advantage of multimodal information fusion for noncontact HR estimation.

Original languageEnglish
JournalIEEE Transactions on Instrumentation and Measurement
Volume71
DOIs
StatePublished - 2022
Externally publishedYes

UN SDGs

This output contributes to the following UN Sustainable Development Goals (SDGs)

  1. SDG 3 - Good Health and Well-being
    SDG 3 Good Health and Well-being

Keywords

  • Attention mechanism
  • Deep learning
  • Graph convolution network (GCN)
  • Multimodal information fusion
  • Noncontact heart rate (HR) estimation

Fingerprint

Dive into the research topics of 'Multimodal Information Fusion Approach for Noncontact Heart Rate Estimation Using Facial Videos and Graph Convolutional Network'. Together they form a unique fingerprint.

Cite this