Zahra Ashktorab, Qian Pan, Werner Geyer, Michael Desmond, Marina Danilevsky, James M. Johnson, Casey Dugan, Michelle Bachman
{"title":"人类-人工智能文本生成中新出现的依赖行为:幻觉、数据质量评估和认知强迫功能","authors":"Zahra Ashktorab, Qian Pan, Werner Geyer, Michael Desmond, Marina Danilevsky, James M. Johnson, Casey Dugan, Michelle Bachman","doi":"arxiv-2409.08937","DOIUrl":null,"url":null,"abstract":"In this paper, we investigate the impact of hallucinations and cognitive\nforcing functions in human-AI collaborative text generation tasks, focusing on\nthe use of Large Language Models (LLMs) to assist in generating high-quality\nconversational data. LLMs require data for fine-tuning, a crucial step in\nenhancing their performance. In the context of conversational customer support,\nthe data takes the form of a conversation between a human customer and an agent\nand can be generated with an AI assistant. In our inquiry, involving 11 users\nwho each completed 8 tasks, resulting in a total of 88 tasks, we found that the\npresence of hallucinations negatively impacts the quality of data. We also find\nthat, although the cognitive forcing function does not always mitigate the\ndetrimental effects of hallucinations on data quality, the presence of\ncognitive forcing functions and hallucinations together impacts data quality\nand influences how users leverage the AI responses presented to them. Our\nanalysis of user behavior reveals distinct patterns of reliance on AI-generated\nresponses, highlighting the importance of managing hallucinations in\nAI-generated content within conversational AI contexts.","PeriodicalId":501541,"journal":{"name":"arXiv - CS - Human-Computer Interaction","volume":"15 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Emerging Reliance Behaviors in Human-AI Text Generation: Hallucinations, Data Quality Assessment, and Cognitive Forcing Functions\",\"authors\":\"Zahra Ashktorab, Qian Pan, Werner Geyer, Michael Desmond, Marina Danilevsky, James M. Johnson, Casey Dugan, Michelle Bachman\",\"doi\":\"arxiv-2409.08937\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, we investigate the impact of hallucinations and cognitive\\nforcing functions in human-AI collaborative text generation tasks, focusing on\\nthe use of Large Language Models (LLMs) to assist in generating high-quality\\nconversational data. LLMs require data for fine-tuning, a crucial step in\\nenhancing their performance. In the context of conversational customer support,\\nthe data takes the form of a conversation between a human customer and an agent\\nand can be generated with an AI assistant. In our inquiry, involving 11 users\\nwho each completed 8 tasks, resulting in a total of 88 tasks, we found that the\\npresence of hallucinations negatively impacts the quality of data. We also find\\nthat, although the cognitive forcing function does not always mitigate the\\ndetrimental effects of hallucinations on data quality, the presence of\\ncognitive forcing functions and hallucinations together impacts data quality\\nand influences how users leverage the AI responses presented to them. Our\\nanalysis of user behavior reveals distinct patterns of reliance on AI-generated\\nresponses, highlighting the importance of managing hallucinations in\\nAI-generated content within conversational AI contexts.\",\"PeriodicalId\":501541,\"journal\":{\"name\":\"arXiv - CS - Human-Computer Interaction\",\"volume\":\"15 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Human-Computer Interaction\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2409.08937\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Human-Computer Interaction","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.08937","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Emerging Reliance Behaviors in Human-AI Text Generation: Hallucinations, Data Quality Assessment, and Cognitive Forcing Functions
In this paper, we investigate the impact of hallucinations and cognitive
forcing functions in human-AI collaborative text generation tasks, focusing on
the use of Large Language Models (LLMs) to assist in generating high-quality
conversational data. LLMs require data for fine-tuning, a crucial step in
enhancing their performance. In the context of conversational customer support,
the data takes the form of a conversation between a human customer and an agent
and can be generated with an AI assistant. In our inquiry, involving 11 users
who each completed 8 tasks, resulting in a total of 88 tasks, we found that the
presence of hallucinations negatively impacts the quality of data. We also find
that, although the cognitive forcing function does not always mitigate the
detrimental effects of hallucinations on data quality, the presence of
cognitive forcing functions and hallucinations together impacts data quality
and influences how users leverage the AI responses presented to them. Our
analysis of user behavior reveals distinct patterns of reliance on AI-generated
responses, highlighting the importance of managing hallucinations in
AI-generated content within conversational AI contexts.