Lida Shi , Fausto Giunchiglia , Ran Luo , Daqian Shi , Rui Song , Xiaolei Diao , Hao Xu
{"title":"An empirical study of LLMs via in-context learning for stance classification","authors":"Lida Shi , Fausto Giunchiglia , Ran Luo , Daqian Shi , Rui Song , Xiaolei Diao , Hao Xu","doi":"10.1016/j.ipm.2025.104322","DOIUrl":null,"url":null,"abstract":"<div><div>The rapid advancement of large language models (LLMs) creates new research opportunities in stance classification. However, existing studies often lack a systematic evaluation and empirical analysis of the performance of mainstream large models. In this paper, we systematically evaluate the performance of 5 SOTA large language models, including LLaMA, DeepSeek, Qwen, GPT, and Gemini, on stance classification using 13 benchmark datasets. We explore the effectiveness of two strategies — random selection and semantic similarity selection — within the framework of in-context learning. By comparing these approaches through cross-domain and in-domain experiments, we reveal how they impact model performance and provide insights for future optimization. Overall, this study clarifies the influence of different models and sampling strategies on stance classification performance and suggests directions for further research. Our code is available at: <span><span>https://github.com/shilida/In-context4Stance</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50365,"journal":{"name":"Information Processing & Management","volume":"63 1","pages":"Article 104322"},"PeriodicalIF":6.9000,"publicationDate":"2025-08-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Processing & Management","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0306457325002638","RegionNum":1,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
The rapid advancement of large language models (LLMs) creates new research opportunities in stance classification. However, existing studies often lack a systematic evaluation and empirical analysis of the performance of mainstream large models. In this paper, we systematically evaluate the performance of 5 SOTA large language models, including LLaMA, DeepSeek, Qwen, GPT, and Gemini, on stance classification using 13 benchmark datasets. We explore the effectiveness of two strategies — random selection and semantic similarity selection — within the framework of in-context learning. By comparing these approaches through cross-domain and in-domain experiments, we reveal how they impact model performance and provide insights for future optimization. Overall, this study clarifies the influence of different models and sampling strategies on stance classification performance and suggests directions for further research. Our code is available at: https://github.com/shilida/In-context4Stance.
期刊介绍:
Information Processing and Management is dedicated to publishing cutting-edge original research at the convergence of computing and information science. Our scope encompasses theory, methods, and applications across various domains, including advertising, business, health, information science, information technology marketing, and social computing.
We aim to cater to the interests of both primary researchers and practitioners by offering an effective platform for the timely dissemination of advanced and topical issues in this interdisciplinary field. The journal places particular emphasis on original research articles, research survey articles, research method articles, and articles addressing critical applications of research. Join us in advancing knowledge and innovation at the intersection of computing and information science.