Antoine Didisheim , Martina Fraschini , Luciano Somoza
{"title":"AI’s predictable memory in financial analysis","authors":"Antoine Didisheim , Martina Fraschini , Luciano Somoza","doi":"10.1016/j.econlet.2025.112602","DOIUrl":null,"url":null,"abstract":"<div><div>Look-ahead bias in Large Language Models (LLMs) arises when information that would not have been available at the time of prediction is included in the training data and inflates prediction performance. This paper proposes a practical methodology to quantify look-ahead bias in financial applications. By prompting LLMs to retrieve historical stock returns without context, we construct a proxy to estimate memorization-driven predictability. We show that the bias varies predictably with data frequency, model size, and aggregation level: smaller models and finer data granularity exhibit negligible bias. Our results help researchers navigate the trade-off between statistical power and bias in LLMs.</div></div>","PeriodicalId":11468,"journal":{"name":"Economics Letters","volume":"256 ","pages":"Article 112602"},"PeriodicalIF":1.8000,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Economics Letters","FirstCategoryId":"96","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0165176525004392","RegionNum":4,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ECONOMICS","Score":null,"Total":0}
引用次数: 0
Abstract
Look-ahead bias in Large Language Models (LLMs) arises when information that would not have been available at the time of prediction is included in the training data and inflates prediction performance. This paper proposes a practical methodology to quantify look-ahead bias in financial applications. By prompting LLMs to retrieve historical stock returns without context, we construct a proxy to estimate memorization-driven predictability. We show that the bias varies predictably with data frequency, model size, and aggregation level: smaller models and finer data granularity exhibit negligible bias. Our results help researchers navigate the trade-off between statistical power and bias in LLMs.
期刊介绍:
Many economists today are concerned by the proliferation of journals and the concomitant labyrinth of research to be conquered in order to reach the specific information they require. To combat this tendency, Economics Letters has been conceived and designed outside the realm of the traditional economics journal. As a Letters Journal, it consists of concise communications (letters) that provide a means of rapid and efficient dissemination of new results, models and methods in all fields of economic research.