Experts fail to reliably detect AI-generated histological data

Jan Hartung, Stefanie Reuter, Vera Anna Kulow, Michael Fahling, Cord Spreckelsen, Ralf Mrowka
{"title":"Experts fail to reliably detect AI-generated histological data","authors":"Jan Hartung, Stefanie Reuter, Vera Anna Kulow, Michael Fahling, Cord Spreckelsen, Ralf Mrowka","doi":"10.1101/2024.01.23.576647","DOIUrl":null,"url":null,"abstract":"AI-based methods to generate images have seen unprecedented advances in recent years challenging both image forensic and human perceptual capabilities. Accordingly, they are expected to play an increasingly important role in the fraudulent fabrication of data. This includes images with complicated intrinsic structures like histological tissue samples, which are harder to forge manually. We use stable diffusion, one of the most recent generative algorithms, to create such a set of artificial histological samples and in a large study with over 800 participants, we study the ability of human subjects to discriminate between such artificial and genuine histological images. Although they perform better than naive participants, we find that even experts fail to reliably identify fabricated data. While participant performance depends on the amount of training data used, even low quantities result in convincing images, necessitating methods to detect fabricated data and technical standards such as C2PA to secure data integrity.","PeriodicalId":501568,"journal":{"name":"bioRxiv - Scientific Communication and Education","volume":"27 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-01-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"bioRxiv - Scientific Communication and Education","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1101/2024.01.23.576647","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

AI-based methods to generate images have seen unprecedented advances in recent years challenging both image forensic and human perceptual capabilities. Accordingly, they are expected to play an increasingly important role in the fraudulent fabrication of data. This includes images with complicated intrinsic structures like histological tissue samples, which are harder to forge manually. We use stable diffusion, one of the most recent generative algorithms, to create such a set of artificial histological samples and in a large study with over 800 participants, we study the ability of human subjects to discriminate between such artificial and genuine histological images. Although they perform better than naive participants, we find that even experts fail to reliably identify fabricated data. While participant performance depends on the amount of training data used, even low quantities result in convincing images, necessitating methods to detect fabricated data and technical standards such as C2PA to secure data integrity.
专家无法可靠地检测人工智能生成的组织学数据
近年来,基于人工智能的图像生成方法取得了前所未有的进步,对图像取证能力和人类感知能力都提出了挑战。因此,它们有望在数据伪造方面发挥越来越重要的作用。这包括像组织学组织样本这样具有复杂内在结构的图像,因为这些图像更难人工伪造。我们使用最新的生成算法之一--稳定扩散来创建这样一组人工组织样本,并在一项有 800 多人参与的大型研究中,研究了人类受试者分辨此类人工组织图像和真实组织图像的能力。我们发现,虽然人类受试者的表现要好于天真的受试者,但即使是专家也无法可靠地识别伪造数据。虽然受试者的表现取决于所使用的训练数据量,但即使数量很少,也能生成令人信服的图像,因此有必要采用检测伪造数据的方法和 C2PA 等技术标准来确保数据的完整性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信