As the demand for long-term health evaluation grows, researchers show increased interest in remote photoplethysmography studies. However, conventional methods are vulnerable to noise interference caused by non-rigid facial movements (facial expression, talking, etc.). Consequently, avoiding these interferences and improving the remote photoplethysmography (rPPG) signal quality become important tasks during heart rate (HR) estimation. We propose an approach that extracts high-quality rPPG signals from various subregions of the face by fusing static and dynamic weights and then employs the convolutional neural network to estimate HR value by converting the 1D rPPG signal into 2D time-frequency analysis maps. Specifically, chrominance features from various regions of interest are used to generate the raw subregion rPPG signal set that is further utilized to estimate the static weights of different regions through a clustering method. Additionally, a measurement method called enclosed area distance is proposed to perform static weights estimation. The dynamic weights of different regions are calculated using the 3D-gradient descriptor to eliminate motion interference, which evaluates the inactivation degree under regional movement situations. The final rPPG signal is reconstructed by combining the rPPG signals from the different subregions using the static and dynamic weights. The experiments are conducted on two widely used public datasets, i.e., MAHNOB-HCI and PURE. The results demonstrate that the proposed method achieves 3.12 MAE and 3.78 SD on MAHNOB-HCI and the best r on the PURE, which significantly outperforms state-of-the-art methods. |
ACCESS THE FULL ARTICLE
No SPIE Account? Create one
CITATIONS
Cited by 1 scholarly publication.
Heart
Motion estimation
Video
Beam propagation method
RGB color model
Databases
Electronic filtering