From data extraction to analysis: a comparative study of ELISE capabilities in scientific literature.

IF 3 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Frontiers in Artificial Intelligence Pub Date : 2025-05-12 eCollection Date: 2025-01-01 DOI:10.3389/frai.2025.1587244
Maxime Gobin, Muriel Gosnat, Seindé Toure, Lina Faik, Joel Belafa, Antoine Villedieu de Torcy, Florence Armstrong
{"title":"From data extraction to analysis: a comparative study of ELISE capabilities in scientific literature.","authors":"Maxime Gobin, Muriel Gosnat, Seindé Toure, Lina Faik, Joel Belafa, Antoine Villedieu de Torcy, Florence Armstrong","doi":"10.3389/frai.2025.1587244","DOIUrl":null,"url":null,"abstract":"<p><p>The exponential growth of scientific literature presents challenges for pharmaceutical, biotechnological, and Medtech industries, particularly in regulatory documentation, clinical research, and systematic reviews. Ensuring accurate data extraction, literature synthesis, and compliance with industry standards require AI tools that not only streamline workflows but also uphold scientific rigor. This study evaluates the performance of AI tools designed for bibliographic review, data extraction, and scientific synthesis, assessing their impact on decision-making, regulatory compliance, and research productivity. The AI tools assessed include general-purpose models like ChatGPT and specialized solutions such as ELISE (Elevated LIfe SciencEs), SciSpace/Typeset, Humata, and Epsilon. The evaluation is based on three main criteria: Extraction, Comprehension, and Analysis with Compliance and Traceability (ECACT) as additional dimensions. Human experts established reference benchmarks, while AI Evaluator models ensure objective performance measurement. The study introduces the ECACT score, a structured metric assessing AI reliability in scientific literature analysis, regulatory reporting and clinical documentation. Results demonstrate that ELISE consistently outperforms other AI tools, excelling in precise data extraction, deep contextual comprehension, and advanced content analysis. ELISE's ability to generate traceable, well-reasoned insights makes it particularly well-suited for high-stakes applications such as regulatory affairs, clinical trials, and medical documentation, where accuracy, transparency, and compliance are paramount. Unlike other AI tools, ELISE provides expert-level reasoning and explainability, ensuring AI-generated insights align with industry best practices. ChatGPT is efficient in data retrieval but lacks precision in complex analysis, limiting its use in high-stakes decision-making. Epsilon, Humata, and SciSpace/Typeset exhibit moderate performance, with variability affecting their reliability in critical applications. In conclusion, while AI tools such as ELISE enhance literature review, regulatory writing, and clinical data interpretation, human oversight remains essential to validate AI outputs and ensure compliance with scientific and regulatory standards. For pharmaceutical, biotechnological, and Medtech industries, AI integration must strike a balance between automation and expert supervision to maintain data integrity, transparency, and regulatory adherence.</p>","PeriodicalId":33315,"journal":{"name":"Frontiers in Artificial Intelligence","volume":"8 ","pages":"1587244"},"PeriodicalIF":3.0000,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12104259/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Frontiers in Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3389/frai.2025.1587244","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/1/1 0:00:00","PubModel":"eCollection","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

The exponential growth of scientific literature presents challenges for pharmaceutical, biotechnological, and Medtech industries, particularly in regulatory documentation, clinical research, and systematic reviews. Ensuring accurate data extraction, literature synthesis, and compliance with industry standards require AI tools that not only streamline workflows but also uphold scientific rigor. This study evaluates the performance of AI tools designed for bibliographic review, data extraction, and scientific synthesis, assessing their impact on decision-making, regulatory compliance, and research productivity. The AI tools assessed include general-purpose models like ChatGPT and specialized solutions such as ELISE (Elevated LIfe SciencEs), SciSpace/Typeset, Humata, and Epsilon. The evaluation is based on three main criteria: Extraction, Comprehension, and Analysis with Compliance and Traceability (ECACT) as additional dimensions. Human experts established reference benchmarks, while AI Evaluator models ensure objective performance measurement. The study introduces the ECACT score, a structured metric assessing AI reliability in scientific literature analysis, regulatory reporting and clinical documentation. Results demonstrate that ELISE consistently outperforms other AI tools, excelling in precise data extraction, deep contextual comprehension, and advanced content analysis. ELISE's ability to generate traceable, well-reasoned insights makes it particularly well-suited for high-stakes applications such as regulatory affairs, clinical trials, and medical documentation, where accuracy, transparency, and compliance are paramount. Unlike other AI tools, ELISE provides expert-level reasoning and explainability, ensuring AI-generated insights align with industry best practices. ChatGPT is efficient in data retrieval but lacks precision in complex analysis, limiting its use in high-stakes decision-making. Epsilon, Humata, and SciSpace/Typeset exhibit moderate performance, with variability affecting their reliability in critical applications. In conclusion, while AI tools such as ELISE enhance literature review, regulatory writing, and clinical data interpretation, human oversight remains essential to validate AI outputs and ensure compliance with scientific and regulatory standards. For pharmaceutical, biotechnological, and Medtech industries, AI integration must strike a balance between automation and expert supervision to maintain data integrity, transparency, and regulatory adherence.

从数据提取到分析:科学文献中ELISE能力的比较研究。
科学文献的指数级增长给制药、生物技术和医疗技术行业带来了挑战,特别是在法规文件、临床研究和系统评价方面。确保准确的数据提取、文献合成和符合行业标准,需要人工智能工具不仅要简化工作流程,还要保持科学的严谨性。本研究评估了用于书目审查、数据提取和科学综合的人工智能工具的性能,评估了它们对决策、法规遵从性和研究生产力的影响。评估的人工智能工具包括ChatGPT等通用模型和ELISE(高级生命科学)、SciSpace/Typeset、Humata和Epsilon等专业解决方案。评估基于三个主要标准:提取、理解和分析,并将遵从性和可追溯性(ECACT)作为附加维度。人类专家建立了参考基准,而人工智能评估器模型确保了客观的绩效衡量。该研究引入了ECACT评分,这是一种评估人工智能在科学文献分析、监管报告和临床文件中的可靠性的结构化指标。结果表明,ELISE始终优于其他人工智能工具,在精确的数据提取、深度上下文理解和高级内容分析方面表现出色。ELISE生成可追溯的、合理的见解的能力使其特别适合高风险应用,例如法规事务、临床试验和医疗文档,这些应用的准确性、透明度和合规性至关重要。与其他人工智能工具不同,ELISE提供专家级的推理和可解释性,确保人工智能生成的见解与行业最佳实践保持一致。ChatGPT在数据检索方面效率很高,但在复杂分析方面缺乏精度,限制了其在高风险决策中的应用。Epsilon、Humata和SciSpace/Typeset表现出中等的性能,但在关键应用程序中,可变性会影响它们的可靠性。总之,尽管ELISE等人工智能工具增强了文献综述、法规写作和临床数据解释,但人类监督对于验证人工智能输出并确保符合科学和监管标准仍然至关重要。对于制药、生物技术和医疗技术行业,人工智能集成必须在自动化和专家监督之间取得平衡,以保持数据完整性、透明度和法规遵从性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
6.10
自引率
2.50%
发文量
272
审稿时长
13 weeks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信