Meidai Xuanyuan , Tao Yang , Jingwen Fu , Sicheng Zhao , Yuwang Wang
{"title":"Insufficient task description can impair in-context learning: A study from information perspective","authors":"Meidai Xuanyuan , Tao Yang , Jingwen Fu , Sicheng Zhao , Yuwang Wang","doi":"10.1016/j.inffus.2025.103116","DOIUrl":null,"url":null,"abstract":"<div><div>In-context learning, an essential technique in transformer-based models, relies on two main sources of information: in-context examples and task descriptions. While extensive research has focused on the influence of in-context examples, the role of task descriptions remains underexplored, despite its practical significance. This paper investigates how task descriptions impact the in-context learning performance of transformers and how these two sources of information can be effectively fused. We design a synthetic experimental framework to control the information provided in task descriptions and conduct a series of experiments where task description details are systematically varied. Our findings reveal the dual roles of task descriptions: an insufficient task description will cause the model to overlook in-context examples, leading to poor in-context performance; once the amount of information in the task description exceeds a certain threshold, the impact of the task description shifts from negative to positive, and a performance emergence can be observed. We replicate these findings on GPT-4, observing a similar double-sided effect. This study highlights the critical role of task descriptions in in-context learning, offering valuable insights for future applications of transformer models.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"120 ","pages":"Article 103116"},"PeriodicalIF":14.7000,"publicationDate":"2025-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525001897","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
In-context learning, an essential technique in transformer-based models, relies on two main sources of information: in-context examples and task descriptions. While extensive research has focused on the influence of in-context examples, the role of task descriptions remains underexplored, despite its practical significance. This paper investigates how task descriptions impact the in-context learning performance of transformers and how these two sources of information can be effectively fused. We design a synthetic experimental framework to control the information provided in task descriptions and conduct a series of experiments where task description details are systematically varied. Our findings reveal the dual roles of task descriptions: an insufficient task description will cause the model to overlook in-context examples, leading to poor in-context performance; once the amount of information in the task description exceeds a certain threshold, the impact of the task description shifts from negative to positive, and a performance emergence can be observed. We replicate these findings on GPT-4, observing a similar double-sided effect. This study highlights the critical role of task descriptions in in-context learning, offering valuable insights for future applications of transformer models.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.