Comparing Emotional Valence from Human Quantitative Ratings and Qualitative Narrative Data on Using Artificial Intelligence to Reduce Caregiving Disparity.
Sunmoo Yoon, Robert Crupi, Frederick Sun, Dante Tipiani, Melissa Patterson, Tess Pottinger, Milea Kim, Ncole Davis
{"title":"Comparing Emotional Valence from Human Quantitative Ratings and Qualitative Narrative Data on Using Artificial Intelligence to Reduce Caregiving Disparity.","authors":"Sunmoo Yoon, Robert Crupi, Frederick Sun, Dante Tipiani, Melissa Patterson, Tess Pottinger, Milea Kim, Ncole Davis","doi":"10.3233/SHTI250036","DOIUrl":null,"url":null,"abstract":"<p><p>We compared emotional valence scores as determined via machine vs human ratings from a survey conducted from April to May 2024 on perceived attitudes on the use of artificial intelligence (AI) for African American family caregivers of persons with Alzheimer's disease and related dementias (ADRD) (N=627). The participants answered risks, benefits and possible solutions qualitatively on the open-ended questions on ten AI use cases, followed by a rating of each. Then, we applied three machine learning algorithms to detect emotional valence scores from the text data and compared their mean to the human ratings. The mean emotional valence scores from text data via natural language processing (NLP) were negative regardless of algorithms (AFINN: -1.61 ± 2.76, Bing: -1.40 ± 1.52, and Syuzhet: -0.67 ± 1.14), while the mean score of human ratings was positive (2.30 ± 1.48, p=0.0001). Our findings have implications for the practice of survey design using self-rated instruments and open-ended questions in an NLP era.</p>","PeriodicalId":94357,"journal":{"name":"Studies in health technology and informatics","volume":"323 ","pages":"1-5"},"PeriodicalIF":0.0000,"publicationDate":"2025-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Studies in health technology and informatics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3233/SHTI250036","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
We compared emotional valence scores as determined via machine vs human ratings from a survey conducted from April to May 2024 on perceived attitudes on the use of artificial intelligence (AI) for African American family caregivers of persons with Alzheimer's disease and related dementias (ADRD) (N=627). The participants answered risks, benefits and possible solutions qualitatively on the open-ended questions on ten AI use cases, followed by a rating of each. Then, we applied three machine learning algorithms to detect emotional valence scores from the text data and compared their mean to the human ratings. The mean emotional valence scores from text data via natural language processing (NLP) were negative regardless of algorithms (AFINN: -1.61 ± 2.76, Bing: -1.40 ± 1.52, and Syuzhet: -0.67 ± 1.14), while the mean score of human ratings was positive (2.30 ± 1.48, p=0.0001). Our findings have implications for the practice of survey design using self-rated instruments and open-ended questions in an NLP era.