Evaluating the alignment of AI with human emotions

J. Derek Lomas , Willem van der Maden , Sohhom Bandyopadhyay , Giovanni Lion , Nirmal Patel , Gyanesh Jain , Yanna Litowsky , Haian Xue , Pieter Desmet
{"title":"Evaluating the alignment of AI with human emotions","authors":"J. Derek Lomas ,&nbsp;Willem van der Maden ,&nbsp;Sohhom Bandyopadhyay ,&nbsp;Giovanni Lion ,&nbsp;Nirmal Patel ,&nbsp;Gyanesh Jain ,&nbsp;Yanna Litowsky ,&nbsp;Haian Xue ,&nbsp;Pieter Desmet","doi":"10.1016/j.ijadr.2024.10.002","DOIUrl":null,"url":null,"abstract":"<div><div>Generative AI systems are increasingly capable of expressing emotions through text, imagery, voice, and video. Effective emotional expression is particularly relevant for AI systems designed to provide care, support mental health, or promote wellbeing through emotional interactions. This research aims to enhance understanding of the alignment between AI-expressed emotions and human perception. How can we assess whether an AI system successfully conveys a specific emotion? To address this question, we designed a method to measure the alignment between emotions expressed by generative AI and human perceptions.</div><div>Three generative image models—DALL-E 2, DALL-E 3, and Stable Diffusion v1—were used to generate 240 images expressing five positive and five negative emotions in both humans and robots. Twenty-four participants recruited via Prolific rated the alignment of AI-generated emotional expressions with a string of text (e.g., “A robot expressing the emotion of amusement”).</div><div>Our results suggest that generative AI models can produce emotional expressions that align well with human emotions; however, the degree of alignment varies significantly depending on the AI model and the specific emotion expressed. We analyze these variations to identify areas for future improvement. The paper concludes with a discussion of the implications of our findings on the design of emotionally expressive AI systems.</div></div>","PeriodicalId":100031,"journal":{"name":"Advanced Design Research","volume":"2 2","pages":"Pages 88-97"},"PeriodicalIF":0.0000,"publicationDate":"2024-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advanced Design Research","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2949782524000185","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Generative AI systems are increasingly capable of expressing emotions through text, imagery, voice, and video. Effective emotional expression is particularly relevant for AI systems designed to provide care, support mental health, or promote wellbeing through emotional interactions. This research aims to enhance understanding of the alignment between AI-expressed emotions and human perception. How can we assess whether an AI system successfully conveys a specific emotion? To address this question, we designed a method to measure the alignment between emotions expressed by generative AI and human perceptions.
Three generative image models—DALL-E 2, DALL-E 3, and Stable Diffusion v1—were used to generate 240 images expressing five positive and five negative emotions in both humans and robots. Twenty-four participants recruited via Prolific rated the alignment of AI-generated emotional expressions with a string of text (e.g., “A robot expressing the emotion of amusement”).
Our results suggest that generative AI models can produce emotional expressions that align well with human emotions; however, the degree of alignment varies significantly depending on the AI model and the specific emotion expressed. We analyze these variations to identify areas for future improvement. The paper concludes with a discussion of the implications of our findings on the design of emotionally expressive AI systems.
评估人工智能与人类情感的一致性
生成式人工智能系统越来越有能力通过文本、图像、语音和视频来表达情感。有效的情感表达对于旨在通过情感互动提供护理、支持心理健康或促进福祉的人工智能系统尤为重要。这项研究旨在加强对人工智能表达的情感与人类感知之间一致性的理解。我们如何评估AI系统是否成功地传达了特定的情感?为了解决这个问题,我们设计了一种方法来测量生成式人工智能所表达的情感与人类感知之间的一致性。使用DALL-E 2、DALL-E 3和Stable Diffusion v1三种生成图像模型,生成240幅图像,分别表达人类和机器人的5种积极情绪和5种消极情绪。24名参与者通过多产的方式对人工智能生成的情感表达与一串文本(例如,“一个机器人表达了娱乐的情感”)的一致性进行了评分。我们的研究结果表明,生成式人工智能模型可以产生与人类情绪非常一致的情绪表达;然而,根据人工智能模型和所表达的特定情感,对齐程度会有很大差异。我们分析这些变化,以确定未来需要改进的领域。本文最后讨论了我们的研究结果对情感表达AI系统设计的影响。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信