Tianqi Liu , Hanguang Xiao , Yisha Sun , Kun Zuo , Qihang Zhang , Zhipeng Li , Feizhong Zhou
{"title":"Syn-rPPG:利用生成模型改进合成视频的无监督远程光容积脉搏波提取","authors":"Tianqi Liu , Hanguang Xiao , Yisha Sun , Kun Zuo , Qihang Zhang , Zhipeng Li , Feizhong Zhou","doi":"10.1016/j.engappai.2025.110504","DOIUrl":null,"url":null,"abstract":"<div><div>Remote photoplethysmography (rPPG) is a non-contact technology used to capture cardiac activity from the face, providing measurements of physiological parameters. Current unsupervised methods for rPPG tasks often focus on contrastive learning, which highlights relationships between samples but struggles with a lack of diverse training data, particularly in terms of varying skin colors and motion types. This limits model effectiveness in complex real-world scenarios. Generative models offer a potential solution by creating synthetic samples to enrich the training data. In this study, we explore the impact of using synthetic videos generated by style transfer and motion transfer techniques to enhance unsupervised rPPG tasks. We generate two types of synthetic videos: skin color synthetic videos and motion synthetic videos. These address the key challenges in rPPG, namely skin color variations and motion artifacts. Our analysis shows that these synthetic videos provide valuable physiological information, improving the performance and robustness of unsupervised models. Additionally, we propose a novel lightweight rPPG network, Style-Aware rPPG Fusion Net (SAFNet), based on an encoder–decoder structure, which is optimized for joint training with synthetic videos. By incorporating a feature fusion approach, SAFNet captures richer spatiotemporal information, resulting in superior performance and robustness. Extensive experiments on four public benchmark datasets demonstrate that our method achieves excellent results, particularly in challenging conditions, proving the effectiveness of using synthetic data to enhance remote physiological measurements.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"149 ","pages":"Article 110504"},"PeriodicalIF":8.0000,"publicationDate":"2025-03-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Syn-rPPG: Improving unsupervised remote photoplethysmography extraction with synthesized videos using generative models\",\"authors\":\"Tianqi Liu , Hanguang Xiao , Yisha Sun , Kun Zuo , Qihang Zhang , Zhipeng Li , Feizhong Zhou\",\"doi\":\"10.1016/j.engappai.2025.110504\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Remote photoplethysmography (rPPG) is a non-contact technology used to capture cardiac activity from the face, providing measurements of physiological parameters. Current unsupervised methods for rPPG tasks often focus on contrastive learning, which highlights relationships between samples but struggles with a lack of diverse training data, particularly in terms of varying skin colors and motion types. This limits model effectiveness in complex real-world scenarios. Generative models offer a potential solution by creating synthetic samples to enrich the training data. In this study, we explore the impact of using synthetic videos generated by style transfer and motion transfer techniques to enhance unsupervised rPPG tasks. We generate two types of synthetic videos: skin color synthetic videos and motion synthetic videos. These address the key challenges in rPPG, namely skin color variations and motion artifacts. Our analysis shows that these synthetic videos provide valuable physiological information, improving the performance and robustness of unsupervised models. Additionally, we propose a novel lightweight rPPG network, Style-Aware rPPG Fusion Net (SAFNet), based on an encoder–decoder structure, which is optimized for joint training with synthetic videos. By incorporating a feature fusion approach, SAFNet captures richer spatiotemporal information, resulting in superior performance and robustness. Extensive experiments on four public benchmark datasets demonstrate that our method achieves excellent results, particularly in challenging conditions, proving the effectiveness of using synthetic data to enhance remote physiological measurements.</div></div>\",\"PeriodicalId\":50523,\"journal\":{\"name\":\"Engineering Applications of Artificial Intelligence\",\"volume\":\"149 \",\"pages\":\"Article 110504\"},\"PeriodicalIF\":8.0000,\"publicationDate\":\"2025-03-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Engineering Applications of Artificial Intelligence\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0952197625005044\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Engineering Applications of Artificial Intelligence","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0952197625005044","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
Syn-rPPG: Improving unsupervised remote photoplethysmography extraction with synthesized videos using generative models
Remote photoplethysmography (rPPG) is a non-contact technology used to capture cardiac activity from the face, providing measurements of physiological parameters. Current unsupervised methods for rPPG tasks often focus on contrastive learning, which highlights relationships between samples but struggles with a lack of diverse training data, particularly in terms of varying skin colors and motion types. This limits model effectiveness in complex real-world scenarios. Generative models offer a potential solution by creating synthetic samples to enrich the training data. In this study, we explore the impact of using synthetic videos generated by style transfer and motion transfer techniques to enhance unsupervised rPPG tasks. We generate two types of synthetic videos: skin color synthetic videos and motion synthetic videos. These address the key challenges in rPPG, namely skin color variations and motion artifacts. Our analysis shows that these synthetic videos provide valuable physiological information, improving the performance and robustness of unsupervised models. Additionally, we propose a novel lightweight rPPG network, Style-Aware rPPG Fusion Net (SAFNet), based on an encoder–decoder structure, which is optimized for joint training with synthetic videos. By incorporating a feature fusion approach, SAFNet captures richer spatiotemporal information, resulting in superior performance and robustness. Extensive experiments on four public benchmark datasets demonstrate that our method achieves excellent results, particularly in challenging conditions, proving the effectiveness of using synthetic data to enhance remote physiological measurements.
期刊介绍:
Artificial Intelligence (AI) is pivotal in driving the fourth industrial revolution, witnessing remarkable advancements across various machine learning methodologies. AI techniques have become indispensable tools for practicing engineers, enabling them to tackle previously insurmountable challenges. Engineering Applications of Artificial Intelligence serves as a global platform for the swift dissemination of research elucidating the practical application of AI methods across all engineering disciplines. Submitted papers are expected to present novel aspects of AI utilized in real-world engineering applications, validated using publicly available datasets to ensure the replicability of research outcomes. Join us in exploring the transformative potential of AI in engineering.