{"title":"Acute Pain Recognition from Facial Expression Videos using Vision Transformers.","authors":"Ghazal Bargshady, Calvin Joseph, Niraj Hirachan, Roland Goecke, Raul Fernandez Rojas","doi":"10.1109/EMBC53108.2024.10781616","DOIUrl":null,"url":null,"abstract":"<p><p>Pain assessment is significant for patients and clinicians in diagnosis and treatment injuries and disease. It could facilitate a patient's treatment process by monitoring patients' pain levels in an accurate and regular manner. Automated detection of pain from facial expressions is a useful technique to assess pain of patients with communication disabilities. In this study, video vision transformers (ViViT) enhanced for pain recognition tasks are presented to capture spatio-temporal, facial information relevant to estimating the binary classification of pain and, thus, to provide valuable insights for automated estimation. The developed model has been trained and evaluated on two acute pain datasets, including 51 subjects using a newly collected pain intensity dataset designated as the AI4PAIN Challenge dataset, and 87 subjects from the BioVid Pain dataset. As an ablation study we used two baseline models, ResNet50 and a hybrid deep learning model based on the pretrained ResNet50+3DCNN. The results demonstrated that the proposed ViViT outperform the other models in pain detection by achieving accuracy = 66.96% for AI4PAIN dataset and accuracy = 79.95% for BioVid dataset.</p>","PeriodicalId":72237,"journal":{"name":"Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference","volume":"2024 ","pages":"1-4"},"PeriodicalIF":0.0000,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/EMBC53108.2024.10781616","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Pain assessment is significant for patients and clinicians in diagnosis and treatment injuries and disease. It could facilitate a patient's treatment process by monitoring patients' pain levels in an accurate and regular manner. Automated detection of pain from facial expressions is a useful technique to assess pain of patients with communication disabilities. In this study, video vision transformers (ViViT) enhanced for pain recognition tasks are presented to capture spatio-temporal, facial information relevant to estimating the binary classification of pain and, thus, to provide valuable insights for automated estimation. The developed model has been trained and evaluated on two acute pain datasets, including 51 subjects using a newly collected pain intensity dataset designated as the AI4PAIN Challenge dataset, and 87 subjects from the BioVid Pain dataset. As an ablation study we used two baseline models, ResNet50 and a hybrid deep learning model based on the pretrained ResNet50+3DCNN. The results demonstrated that the proposed ViViT outperform the other models in pain detection by achieving accuracy = 66.96% for AI4PAIN dataset and accuracy = 79.95% for BioVid dataset.