{"title":"You Had Better Check the Facts: Reader Agency in the Identification of Machine-Generated Medical Fake News","authors":"Barbora Daňková","doi":"10.31273/reinvention.v16i1.964","DOIUrl":null,"url":null,"abstract":"During the COVID-19 pandemic, much fake news emerged in the medical field (Naeem et al., 2020: 1). Nowadays, computers can generate text considered to be more trustworthy than text written by a person (Zellers et al., 2019). This means that laypeople are able to produce disinformation; however, they may not understand the implications. This study revealed the most reliable clues as guidance to spot machine writing. While natural-language processing (NLP) research focuses on L1 speakers, studies in second language acquisition demonstrate that L1 and L2 speakers attend to different aspects of English (Scarcella, 1984; Tsang, 2017). In this study, social media users completed a Turing-test style quiz, guessed whether news excerpts were machine generated or human written (Saygin et al., 2000) and identified errors that guided their decision. Quantitative analysis revealed that although both L1 and L2 speakers were equally able to defend themselves against machine-generated fake news, L2 participants were more sceptical, labelling more human-written texts as being machine generated. This is possibly due to concern about the stigma associated with being fooled by a machine due to lower language levels. However, factual errors and internal contradictions were the most reliable indicators of machine writing for both groups. This emphasises the importance of fact-checking when news articles prioritise exaggerated headlines, and NLP tools enable production of popular content in areas like medicine.","PeriodicalId":183531,"journal":{"name":"Reinvention: an International Journal of Undergraduate Research","volume":"3 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Reinvention: an International Journal of Undergraduate Research","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.31273/reinvention.v16i1.964","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
During the COVID-19 pandemic, much fake news emerged in the medical field (Naeem et al., 2020: 1). Nowadays, computers can generate text considered to be more trustworthy than text written by a person (Zellers et al., 2019). This means that laypeople are able to produce disinformation; however, they may not understand the implications. This study revealed the most reliable clues as guidance to spot machine writing. While natural-language processing (NLP) research focuses on L1 speakers, studies in second language acquisition demonstrate that L1 and L2 speakers attend to different aspects of English (Scarcella, 1984; Tsang, 2017). In this study, social media users completed a Turing-test style quiz, guessed whether news excerpts were machine generated or human written (Saygin et al., 2000) and identified errors that guided their decision. Quantitative analysis revealed that although both L1 and L2 speakers were equally able to defend themselves against machine-generated fake news, L2 participants were more sceptical, labelling more human-written texts as being machine generated. This is possibly due to concern about the stigma associated with being fooled by a machine due to lower language levels. However, factual errors and internal contradictions were the most reliable indicators of machine writing for both groups. This emphasises the importance of fact-checking when news articles prioritise exaggerated headlines, and NLP tools enable production of popular content in areas like medicine.
在COVID-19大流行期间,医疗领域出现了许多假新闻(Naeem et al., 2020: 1)。如今,计算机可以生成被认为比人写的文本更可信的文本(Zellers et al., 2019)。这意味着外行可以制造虚假信息;然而,他们可能不明白其中的含义。这项研究揭示了最可靠的线索,作为指导点机器写作。虽然自然语言处理(NLP)研究侧重于母语使用者,但第二语言习得研究表明,母语使用者和第二语言使用者关注英语的不同方面(Scarcella, 1984;曾荫权,2017)。在这项研究中,社交媒体用户完成了一个图灵测试风格的测验,猜测新闻摘录是机器生成的还是人类编写的(Saygin et al., 2000),并找出指导他们决策的错误。定量分析显示,尽管第一语言和第二语言的说话者对机器生成的假新闻都有同样的防御能力,但第二语言的参与者更怀疑,他们将更多的人类书写的文本标记为机器生成的。这可能是由于担心由于语言水平较低而被机器愚弄的耻辱。然而,对于两组人来说,事实错误和内部矛盾是机器写作最可靠的指标。这强调了事实核查的重要性,当新闻文章优先考虑夸张的标题时,NLP工具可以在医学等领域生产流行内容。