{"title":"基于自适应去噪和跨模态融合的加速度计数据虚拟PPG重建","authors":"Illia Fedorin","doi":"10.1016/j.inffus.2025.103781","DOIUrl":null,"url":null,"abstract":"<div><div>Accurate heart rate (HR) monitoring during high-intensity activity is essential for performance optimization and physiological tracking in wearable devices. While photoplethysmography (PPG) remains the standard for HR estimation, it is prone to motion artifacts, power constraints, and temporary signal loss. Accelerometers (ACC), by contrast, offer motion-resilient and energy-efficient sensing, but estimating HR from ACC alone remains a challenging task. In this study, we introduce a cross-modal virtual sensing framework for HR estimation and spectral reconstruction using only ACC signals. The framework includes: (1) a high-fidelity variational autoencoder (VAE) for offline PPG spectrum reconstruction from ACC input, and (2) a lightweight real-time attention-based denoising model for HR prediction. Both models are trained with a fusion-aware loss to enforce alignment between motion-driven and cardiovascular signal features. Experimental results on public and proprietary datasets demonstrate strong performance and generalization under varying sensor configurations and motion conditions. The real-time model achieves 7.0 BPM mean absolute error (MAE) with only 2.6K parameters, making it suitable for embedded deployment. While PPG remains superior under ideal conditions, the proposed system serves as a fallback modality when optical sensing is unreliable or unavailable-enabling gap-filling, post-processing correction, and low-power monitoring. More broadly, this work positions virtual PPG reconstruction as a proof-of-concept for physiological virtual sensing: a paradigm where one modality can be inferred from another, and potentially reversed, supporting robust multimodal inference in real-world mobile health scenarios.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"127 ","pages":"Article 103781"},"PeriodicalIF":15.5000,"publicationDate":"2025-09-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Virtual PPG reconstruction from accelerometer data via adaptive denoising and cross-Modal fusion\",\"authors\":\"Illia Fedorin\",\"doi\":\"10.1016/j.inffus.2025.103781\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Accurate heart rate (HR) monitoring during high-intensity activity is essential for performance optimization and physiological tracking in wearable devices. While photoplethysmography (PPG) remains the standard for HR estimation, it is prone to motion artifacts, power constraints, and temporary signal loss. Accelerometers (ACC), by contrast, offer motion-resilient and energy-efficient sensing, but estimating HR from ACC alone remains a challenging task. In this study, we introduce a cross-modal virtual sensing framework for HR estimation and spectral reconstruction using only ACC signals. The framework includes: (1) a high-fidelity variational autoencoder (VAE) for offline PPG spectrum reconstruction from ACC input, and (2) a lightweight real-time attention-based denoising model for HR prediction. Both models are trained with a fusion-aware loss to enforce alignment between motion-driven and cardiovascular signal features. Experimental results on public and proprietary datasets demonstrate strong performance and generalization under varying sensor configurations and motion conditions. The real-time model achieves 7.0 BPM mean absolute error (MAE) with only 2.6K parameters, making it suitable for embedded deployment. While PPG remains superior under ideal conditions, the proposed system serves as a fallback modality when optical sensing is unreliable or unavailable-enabling gap-filling, post-processing correction, and low-power monitoring. More broadly, this work positions virtual PPG reconstruction as a proof-of-concept for physiological virtual sensing: a paradigm where one modality can be inferred from another, and potentially reversed, supporting robust multimodal inference in real-world mobile health scenarios.</div></div>\",\"PeriodicalId\":50367,\"journal\":{\"name\":\"Information Fusion\",\"volume\":\"127 \",\"pages\":\"Article 103781\"},\"PeriodicalIF\":15.5000,\"publicationDate\":\"2025-09-26\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Information Fusion\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1566253525008437\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525008437","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Virtual PPG reconstruction from accelerometer data via adaptive denoising and cross-Modal fusion
Accurate heart rate (HR) monitoring during high-intensity activity is essential for performance optimization and physiological tracking in wearable devices. While photoplethysmography (PPG) remains the standard for HR estimation, it is prone to motion artifacts, power constraints, and temporary signal loss. Accelerometers (ACC), by contrast, offer motion-resilient and energy-efficient sensing, but estimating HR from ACC alone remains a challenging task. In this study, we introduce a cross-modal virtual sensing framework for HR estimation and spectral reconstruction using only ACC signals. The framework includes: (1) a high-fidelity variational autoencoder (VAE) for offline PPG spectrum reconstruction from ACC input, and (2) a lightweight real-time attention-based denoising model for HR prediction. Both models are trained with a fusion-aware loss to enforce alignment between motion-driven and cardiovascular signal features. Experimental results on public and proprietary datasets demonstrate strong performance and generalization under varying sensor configurations and motion conditions. The real-time model achieves 7.0 BPM mean absolute error (MAE) with only 2.6K parameters, making it suitable for embedded deployment. While PPG remains superior under ideal conditions, the proposed system serves as a fallback modality when optical sensing is unreliable or unavailable-enabling gap-filling, post-processing correction, and low-power monitoring. More broadly, this work positions virtual PPG reconstruction as a proof-of-concept for physiological virtual sensing: a paradigm where one modality can be inferred from another, and potentially reversed, supporting robust multimodal inference in real-world mobile health scenarios.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.