BenchCouncil Transactions on Benchmarks, Standards and Evaluations最新文献

筛选
英文 中文
Evaluatology’s perspective on AI evaluation in critical scenarios: From tail quality to landscape 评估学在关键场景下的人工智能评估视角:从尾部质量到景观
BenchCouncil Transactions on Benchmarks, Standards and Evaluations Pub Date : 2025-03-01 Epub Date: 2025-05-27 DOI: 10.1016/j.tbench.2025.100203
Zhengxin Yang
{"title":"Evaluatology’s perspective on AI evaluation in critical scenarios: From tail quality to landscape","authors":"Zhengxin Yang","doi":"10.1016/j.tbench.2025.100203","DOIUrl":"10.1016/j.tbench.2025.100203","url":null,"abstract":"<div><div>Tail Quality, as a metric for evaluating AI inference performance in critical scenarios, reveals the extreme behaviors of AI inference systems in real-world applications, offering significant practical value. However, its adoption has been limited due to the lack of systematic theoretical support. To address this issue, this paper analyzes AI inference system evaluation activities from the perspective of Evaluatology, bridging the gap between theory and practice. Specifically, we begin by constructing a rigorous, consistent, and comprehensive evaluation system for AI inference systems, with a focus on defining the evaluation subject and evaluation conditions. We then refine the Quality@Time-Threshold (Q@T) statistical evaluation framework by formalizing these components, thereby enhancing its theoretical rigor and applicability. By integrating the principles of Evaluatology, we extend Q@T to incorporate stakeholder considerations, ensuring its adaptability to varying time tolerance. Through refining the Q@T evaluation framework and embedding it within Evaluatology, we provide a robust theoretical foundation that enhances the accuracy and reliability of AI system evaluations, making the approach both scientifically rigorous and practically reliable. Experimental results further validate the effectiveness of this refined framework, confirming its scientific rigor and practical applicability. The theoretical analysis presented in this paper provides valuable guidance for researchers aiming to apply Evaluatology in practice.</div></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"5 1","pages":"Article 100203"},"PeriodicalIF":0.0,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144168926","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Tensor databases empower AI for science: A case study on retrosynthetic analysis 张量数据库为科学赋予AI力量:一个关于反合成分析的案例研究
BenchCouncil Transactions on Benchmarks, Standards and Evaluations Pub Date : 2025-03-01 Epub Date: 2025-06-16 DOI: 10.1016/j.tbench.2025.100216
Xueya Zhang , Guoxin Kang , Boyang Xiao , Jianfeng Zhan
{"title":"Tensor databases empower AI for science: A case study on retrosynthetic analysis","authors":"Xueya Zhang ,&nbsp;Guoxin Kang ,&nbsp;Boyang Xiao ,&nbsp;Jianfeng Zhan","doi":"10.1016/j.tbench.2025.100216","DOIUrl":"10.1016/j.tbench.2025.100216","url":null,"abstract":"<div><div>Retrosynthetic analysis is highly significant in chemistry, biology, and materials science, providing essential support for the rational design, synthesis, and optimization of compounds across diverse Artificial Intelligence for Science (AI4S) applications. Retrosynthetic analysis focuses on exploring pathways from products to reactants, and this is typically conducted using deep learning-based generative models. However, existing retrosynthetic analysis often overlooks how reaction conditions significantly impact chemical reactions. This causes existing work to lack unified models that can provide full-cycle services for retrosynthetic analysis, and also greatly limits the overall prediction accuracy of retrosynthetic analysis. These two issues cause users to depend on various independent models and tools, leading to high labor time and cost overhead.</div><div>To solve these issues, we define the boundary conditions of chemical reactions based on the Evaluatology theory and propose BigTensorDB, the first tensor database which integrates storage, prediction generation, search, and analysis functions. BigTensorDB designs the tensor schema for efficiently storing all the key information related to chemical reactions, including reaction conditions. BigTensorDB supports a full-cycle retrosynthetic analysis pipeline. It begins with predicting generation reaction paths, searching for approximate real reactions based on the tensor schema, and concludes with feasibility analysis, which enhances the interpretability of prediction results. BigTensorDB can effectively reduce usage costs and improve efficiency for users during the full-cycle retrosynthetic analysis process. Meanwhile, it provides a potential solution to the low accuracy issue, encouraging researchers to focus on improving full-cycle accuracy.</div></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"5 1","pages":"Article 100216"},"PeriodicalIF":0.0,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144307785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ethical and regulatory challenges in machine learning-based healthcare systems: A review of implementation barriers and future directions 基于机器学习的医疗保健系统中的伦理和监管挑战:对实施障碍和未来方向的回顾
BenchCouncil Transactions on Benchmarks, Standards and Evaluations Pub Date : 2025-03-01 Epub Date: 2025-05-28 DOI: 10.1016/j.tbench.2025.100215
Shehu Mohammed, Neha Malhotra
{"title":"Ethical and regulatory challenges in machine learning-based healthcare systems: A review of implementation barriers and future directions","authors":"Shehu Mohammed,&nbsp;Neha Malhotra","doi":"10.1016/j.tbench.2025.100215","DOIUrl":"10.1016/j.tbench.2025.100215","url":null,"abstract":"<div><div>Machine learning significantly enhances clinical decision-making quality, directly impacting patient care with early diagnosis, personalized treatment,  and predictive analytics. Nonetheless, the increasing proliferation of such ML applications in practice raises potential ethical and regulatory obstacles that may prevent their widespread adoption in healthcare. Key issues concern patient data privacy, algorithmic bias, absence of transparency, and ambiguous legal liability. Fortunately, regulations like the General Data Protection Regulation (GDPR), the Health Insurance Portability and Accountability Act (HIPAA),  and the FDA AI/ML guidance have raised important ways of addressing things like fairness, explainability, legal compliance, etc.; however, the landscape is far from risk-free. AI liability is another one of the gray areas approaching black, worrying about who is liable for an AI medical error — the developers, the physicians, or the institutions. The study reviews ethical risks and potential opportunities, as well as regulatory frameworks and emerging challenges in AI-driven healthcare. It proposes solutions to reduce bias, improve transparency, and enhance legal accountability. This research addresses these challenges to support the safe, fair, and effective deployment of ML-based systems in clinical practice, guaranteeing that patients can trust, regulators can approve, and healthcare can use them.</div></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"5 1","pages":"Article 100215"},"PeriodicalIF":0.0,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144271090","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AICB: A benchmark for evaluating the communication subsystem of LLM training clusters AICB:一个评估LLM训练集群通信子系统的基准
BenchCouncil Transactions on Benchmarks, Standards and Evaluations Pub Date : 2025-03-01 Epub Date: 2025-06-02 DOI: 10.1016/j.tbench.2025.100212
Xinyue Li, Heyang Zhou, Qingxu Li, Sen Zhang, Gang Lu
{"title":"AICB: A benchmark for evaluating the communication subsystem of LLM training clusters","authors":"Xinyue Li,&nbsp;Heyang Zhou,&nbsp;Qingxu Li,&nbsp;Sen Zhang,&nbsp;Gang Lu","doi":"10.1016/j.tbench.2025.100212","DOIUrl":"10.1016/j.tbench.2025.100212","url":null,"abstract":"<div><div>AICB (Artificial Intelligence Communication Benchmark) is a benchmark for evaluating the communication subsystem of GPU clusters, which includes representative workloads in the fields of Large Language Model (LLM) training. Guided by the theories and methodologies of Evaluatology, we simplified the real-workload LLM training systems through AICB that maintain good representativeness and usability. AICB bridges the gap between application benchmarks and microbenchmarks in the scope of LLM training. In addition, we constructed a new GPU-free evaluation system that helps researchers evaluate the communication system of the LLM training systems. To help the urgent demand on this evaluation subject, we open-source AICB and make it available at <span><span>https://github.com/aliyun/aicb</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"5 1","pages":"Article 100212"},"PeriodicalIF":0.0,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144255424","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predicting the number of call center incoming calls using deep learning 使用深度学习预测呼叫中心呼入的数量
BenchCouncil Transactions on Benchmarks, Standards and Evaluations Pub Date : 2025-03-01 Epub Date: 2025-06-04 DOI: 10.1016/j.tbench.2025.100213
Armaghan Nikfar , Javad Mohammadzadeh
{"title":"Predicting the number of call center incoming calls using deep learning","authors":"Armaghan Nikfar ,&nbsp;Javad Mohammadzadeh","doi":"10.1016/j.tbench.2025.100213","DOIUrl":"10.1016/j.tbench.2025.100213","url":null,"abstract":"<div><div>One of the main problems in call centers is the call queue. This can lead to long waiting times for customers, increased frustration and call abandonment. The important role that predictive analytics plays in optimizing call center operations is increasingly recognized. Advanced models can be trained by training datasets such as the number of calls that have occurred throughout history, and by estimating how religious and public holidays have affected the weight of hours and the number of calls, and this study provides an analysis of 4 years. Call center data from Shatel, an Internet service provider. Predictive deep learning models, specifically the Bidirectional Short-Term Memory Model (BLSTM), were used to predict the number of incoming calls, predict the number of calls to centers, and prevent call queues with an accuracy of 90.56.</div></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"5 1","pages":"Article 100213"},"PeriodicalIF":0.0,"publicationDate":"2025-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144336039","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Open Source Evaluatology: An evaluation framework and methodology for open source ecosystems based on evaluatology 开源评估学:基于评估学的开源生态系统评估框架和方法
BenchCouncil Transactions on Benchmarks, Standards and Evaluations Pub Date : 2024-12-01 Epub Date: 2025-02-23 DOI: 10.1016/j.tbench.2025.100190
Fanyu Han , Shengyu Zhao , Wei Wang , Aoying Zhou , Weining Qian , Xuan Zhou , Jiaheng Peng , Lan You , Yang Chen , Xiaoya Xia , Yenan Tang , Liyun Yang , Chunqi Tian
{"title":"Open Source Evaluatology: An evaluation framework and methodology for open source ecosystems based on evaluatology","authors":"Fanyu Han ,&nbsp;Shengyu Zhao ,&nbsp;Wei Wang ,&nbsp;Aoying Zhou ,&nbsp;Weining Qian ,&nbsp;Xuan Zhou ,&nbsp;Jiaheng Peng ,&nbsp;Lan You ,&nbsp;Yang Chen ,&nbsp;Xiaoya Xia ,&nbsp;Yenan Tang ,&nbsp;Liyun Yang ,&nbsp;Chunqi Tian","doi":"10.1016/j.tbench.2025.100190","DOIUrl":"10.1016/j.tbench.2025.100190","url":null,"abstract":"<div><div>The open-source ecosystem, as an important component of the modern software industry, has increasingly attracted attention from both academia and industry regarding its evaluation. However, current open-source evaluation methods face several issues, such as inconsistent evaluation standards, lack of theoretical support in the evaluation process, and poor comparability of evaluation results. Guided by the foundational theories of evaluatology, this paper proposes a new interdisciplinary research field, Open Source Evaluatology, and constructs an evaluation theoretical framework and methodology for open-source ecosystems. The main contributions of this paper include: (1) Based on the five axioms of evaluation theory, a theoretical system for Open Source Evaluatology is developed, and the basic concepts, evaluation dimensions, and evaluation standards for the open-source ecosystem are proposed; (2) An evaluation conditions (EC) framework is designed, encompassing five levels: problem definition, task instances, algorithm mechanisms, implementation examples, and supporting systems. A combined evaluation model (EM) based on statistical metrics and network metrics is also introduced; (3) Experimental validation using the GitHub dataset shows that the proposed evaluation framework effectively assesses various features of open-source projects, developers, and communities, and has been verified in multiple practical application scenarios. The research demonstrates that Open Source Evaluatology provides a standardized theoretical guide and methodological support for open-source ecosystem evaluation, which can be widely applied in various scenarios, such as open-source project selection, developer evaluation, and community management, and plays a significant role in promoting the healthy and sustainable development of open-source ecosystems.</div></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"4 4","pages":"Article 100190"},"PeriodicalIF":0.0,"publicationDate":"2024-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145871577","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
COADBench: A benchmark for revealing the relationship between AI models and clinical outcomes COADBench:揭示人工智能模型与临床结果之间关系的基准
BenchCouncil Transactions on Benchmarks, Standards and Evaluations Pub Date : 2024-12-01 Epub Date: 2025-03-13 DOI: 10.1016/j.tbench.2025.100198
Jiyue Xie , Wenjing Liu , Li Ma , Caiqin Yao , Qi Liang , Suqin Tang , Yunyou Huang
{"title":"COADBench: A benchmark for revealing the relationship between AI models and clinical outcomes","authors":"Jiyue Xie ,&nbsp;Wenjing Liu ,&nbsp;Li Ma ,&nbsp;Caiqin Yao ,&nbsp;Qi Liang ,&nbsp;Suqin Tang ,&nbsp;Yunyou Huang","doi":"10.1016/j.tbench.2025.100198","DOIUrl":"10.1016/j.tbench.2025.100198","url":null,"abstract":"<div><div>Alzheimer’s disease (AD), due to its irreversible nature and the severe social burden it causes, has garnered significant attention from AI researchers. Numerous auxiliary diagnostic models have been developed with the aim of improving AD diagnostic services and thereby reducing the social burden. However, due to a lack of validation regarding the clinical value of these models, no AD diagnostic model has been widely accepted by clinicians or officially approved for use in enhancing AD diagnostic services. The clinical value of traditional medical devices is validated through rigorous randomized controlled trials to prove their impact on clinical outcomes. In contrast, current AD diagnostic models are only validated based on their accuracy, and the relationship between these models and patient outcomes remains unknown. This gap has hindered the acceptance and clinical use of AD diagnostic models by healthcare professionals. To address this issue, we introduce the COADBench, a benchmark centered on clinical outcomes for evaluating the clinical value of AD diagnostic models. COADBench curated subjects from the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database who have at least two cognitive score records (the most commonly used clinical endpoint in AD clinical trials) from different follow-up visits. To the best of our knowledge, for the first time, it links the cognitive scores of subjects with model performance, using patient cognitive scores as clinical outcomes after intervention to evaluate the models. Through the benchmarking of current mainstream AD diagnostic algorithms using COADBench, we find that there was no significant correlation between the subjects’ cognitive improvement and the model’s performance, which means that the current performance evaluation criteria of mainstream AD diagnostic algorithms are not combined with clinical value.</div></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"4 4","pages":"Article 100198"},"PeriodicalIF":0.0,"publicationDate":"2024-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145871523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluating long-term usage patterns of open source datasets: A citation network approach 评估开源数据集的长期使用模式:引文网络方法
BenchCouncil Transactions on Benchmarks, Standards and Evaluations Pub Date : 2024-12-01 Epub Date: 2025-03-28 DOI: 10.1016/j.tbench.2025.100199
Jiaheng Peng, Fanyu Han, Wei Wang
{"title":"Evaluating long-term usage patterns of open source datasets: A citation network approach","authors":"Jiaheng Peng,&nbsp;Fanyu Han,&nbsp;Wei Wang","doi":"10.1016/j.tbench.2025.100199","DOIUrl":"10.1016/j.tbench.2025.100199","url":null,"abstract":"<div><div>The evaluation of datasets serves as a fundamental basis for tasks in evaluatology. Evaluating the usage patterns of datasets has a significant impact on the selection of appropriate datasets. Many renowned Open Source datasets are well-established and have not been updated for many years, yet they continue to be widely used by a large number of researchers. Due to this characteristic, conventional Open Source metrics (e.g., number of stars, issues, and activity) are insufficient for evaluating the long-term usage patterns based on log activity data from their GitHub repositories.</div><div>Researchers often encounter significant challenges in selecting appropriate datasets due to the lack of insight into how these datasets are being utilized. To address this challenge, this paper proposes establishing a connection between Open Source datasets and the citation networks of their corresponding academic papers. By mining the citation network of the corresponding academic paper, we can obtain rich graph-structured information, such as citation times, authors, and more. Utilizing this information, we can evaluate the long-term usage patterns of the associated Open Source dataset.</div><div>Furthermore, this paper conducts extensive experiments based on five major dataset categories (Texts, Images, Videos, Audio, Medical) to demonstrate that the proposed method effectively evaluates the long-term usage patterns of Open Source datasets. Additionally, the insights gained from the experimental results can serve as a valuable reference for future researchers in selecting appropriate datasets for their work.</div></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"4 4","pages":"Article 100199"},"PeriodicalIF":0.0,"publicationDate":"2024-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145871525","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Patrick Star: A comprehensive benchmark for multi-modal image editing Patrick Star:多模态图像编辑的综合基准
BenchCouncil Transactions on Benchmarks, Standards and Evaluations Pub Date : 2024-12-01 Epub Date: 2025-04-30 DOI: 10.1016/j.tbench.2025.100201
Di Cheng , ZhengXin Yang , ChunJie Luo , Chen Zheng , YingJie Shi
{"title":"Patrick Star: A comprehensive benchmark for multi-modal image editing","authors":"Di Cheng ,&nbsp;ZhengXin Yang ,&nbsp;ChunJie Luo ,&nbsp;Chen Zheng ,&nbsp;YingJie Shi","doi":"10.1016/j.tbench.2025.100201","DOIUrl":"10.1016/j.tbench.2025.100201","url":null,"abstract":"<div><div>Generative image editing enhances and automates traditional image designing methods. However, there is a significant imbalance in existing research, where the development of sketch-guided and example-guided image editing has not been sufficiently explored compared to text-guided image editing, despite the former being equally important in real-world applications. The leading cause of this phenomenon is the severe lack of corresponding benchmark datasets. To address this issue, this paper proposes a comprehensive and unified benchmark dataset, Patrick Star, which consists of approximately 500 test images, to promote balanced development in this field across multi-task and multi-modal settings. First, theoretical analysis grounded in Evaluatology highlights the importance of establishing a balanced benchmark dataset to advance research in image editing. Building on this theoretical foundation, the dataset’s construction methodology is explained in detail, ensuring it addresses critical gaps in existing studies. Next, statistical analyses are conducted to verify the dataset’s usability and diversity. Finally, comparative experiments underscore the dataset’s potential as a comprehensive benchmark, demonstrating its capacity to support balanced development in image editing.</div></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"4 4","pages":"Article 100201"},"PeriodicalIF":0.0,"publicationDate":"2024-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145871576","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AI-powered Mathematical Sentiment Model and graph theory for social media trends 社交媒体趋势的人工智能数学情感模型和图论
BenchCouncil Transactions on Benchmarks, Standards and Evaluations Pub Date : 2024-12-01 Epub Date: 2025-05-17 DOI: 10.1016/j.tbench.2025.100202
M. VENKATACHALAM , R․VIKRAMA PRASAD
{"title":"AI-powered Mathematical Sentiment Model and graph theory for social media trends","authors":"M. VENKATACHALAM ,&nbsp;R․VIKRAMA PRASAD","doi":"10.1016/j.tbench.2025.100202","DOIUrl":"10.1016/j.tbench.2025.100202","url":null,"abstract":"<div><div>Significant issues have arisen as a result of the global spread of monkeypox, such as the extensive transmission of false information, public fear, and stigmatization on social media. Increased fear, prejudice, stigmatization of minority groups, and opposition to public health initiatives are frequently the results of these problems. Furthermore, health authorities are unable to provide correct information and prompt actions due to a lack of efficient methods for analyzing the enormous amounts of unstructured social media data. This disparity weakens crisis management initiatives and increases public skepticism of health guidelines. In order to address these issues, this study looks into the attitude around monkeypox on social media in order to pinpoint public worries, counter false information, and enhance communication tactics. The study intends to improve public comprehension, offer practical insights, and help health authorities manage the outbreak by fusing graph theory with AI-driven sentiment analysis. In order to facilitate semantic analysis of tweets through structured information extraction, graph theory is used to organize unstructured or semi-structured data by creating meaningful links between entities. Furthermore, opinions on monkeypox infection in social media are analyzed and user sentiments are detected using a reinforcement Markov decision process. According to experimental results, the suggested model's accuracy on the Monkeypox tweet dataset was 98 %. These results help raise awareness of monkeypox among the general population and promote an educated and robust social response.</div></div>","PeriodicalId":100155,"journal":{"name":"BenchCouncil Transactions on Benchmarks, Standards and Evaluations","volume":"4 4","pages":"Article 100202"},"PeriodicalIF":0.0,"publicationDate":"2024-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145871528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信
小红书