The Structure of Financial Equity Research Reports -- Identification of the Most Frequently Asked Questions in Financial Analyst Reports to Automate Equity Research Using Llama 3 and GPT-4

Adria Pop, Jan Spörer, Siegfried Handschuh
{"title":"The Structure of Financial Equity Research Reports -- Identification of the Most Frequently Asked Questions in Financial Analyst Reports to Automate Equity Research Using Llama 3 and GPT-4","authors":"Adria Pop, Jan Spörer, Siegfried Handschuh","doi":"arxiv-2407.18327","DOIUrl":null,"url":null,"abstract":"This research dissects financial equity research reports (ERRs) by mapping\ntheir content into categories. There is insufficient empirical analysis of the questions answered in ERRs.\nIn particular, it is not understood how frequently certain information appears,\nwhat information is considered essential, and what information requires human\njudgment to distill into an ERR. The study analyzes 72 ERRs sentence-by-sentence, classifying their 4940\nsentences into 169 unique question archetypes. We did not predefine the\nquestions but derived them solely from the statements in the ERRs. This\napproach provides an unbiased view of the content of the observed ERRs.\nSubsequently, we used public corporate reports to classify the questions'\npotential for automation. Answers were labeled \"text-extractable\" if the\nanswers to the question were accessible in corporate reports. 78.7% of the questions in ERRs can be automated. Those automatable question\nconsist of 48.2% text-extractable (suited to processing by large language\nmodels, LLMs) and 30.5% database-extractable questions. Only 21.3% of questions\nrequire human judgment to answer. We empirically validate using Llama-3-70B and GPT-4-turbo-2024-04-09 that\nrecent advances in language generation and information extraction enable the\nautomation of approximately 80% of the statements in ERRs. Surprisingly, the\nmodels complement each other's strengths and weaknesses well. The research confirms that the current writing process of ERRs can likely\nbenefit from additional automation, improving quality and efficiency. The\nresearch thus allows us to quantify the potential impacts of introducing large\nlanguage models in the ERR writing process. The full question list, including the archetypes and their frequency, will be\nmade available online after peer review.","PeriodicalId":501294,"journal":{"name":"arXiv - QuantFin - Computational Finance","volume":"50 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-07-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - QuantFin - Computational Finance","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.18327","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

This research dissects financial equity research reports (ERRs) by mapping their content into categories. There is insufficient empirical analysis of the questions answered in ERRs. In particular, it is not understood how frequently certain information appears, what information is considered essential, and what information requires human judgment to distill into an ERR. The study analyzes 72 ERRs sentence-by-sentence, classifying their 4940 sentences into 169 unique question archetypes. We did not predefine the questions but derived them solely from the statements in the ERRs. This approach provides an unbiased view of the content of the observed ERRs. Subsequently, we used public corporate reports to classify the questions' potential for automation. Answers were labeled "text-extractable" if the answers to the question were accessible in corporate reports. 78.7% of the questions in ERRs can be automated. Those automatable question consist of 48.2% text-extractable (suited to processing by large language models, LLMs) and 30.5% database-extractable questions. Only 21.3% of questions require human judgment to answer. We empirically validate using Llama-3-70B and GPT-4-turbo-2024-04-09 that recent advances in language generation and information extraction enable the automation of approximately 80% of the statements in ERRs. Surprisingly, the models complement each other's strengths and weaknesses well. The research confirms that the current writing process of ERRs can likely benefit from additional automation, improving quality and efficiency. The research thus allows us to quantify the potential impacts of introducing large language models in the ERR writing process. The full question list, including the archetypes and their frequency, will be made available online after peer review.
金融股票研究报告的结构 -- 利用 Llama 3 和 GPT-4 识别金融分析师报告中最常见的问题,实现股票研究自动化
本研究将金融股票研究报告(ERRs)的内容分门别类,对其进行剖析。特别是,人们不了解某些信息出现的频率、哪些信息被认为是必要的、哪些信息需要人为判断才能提炼成 ERR。本研究对 72 篇 ERR 进行了逐句分析,将其中的 4940 句分为 169 种独特的问题原型。我们没有预先定义这些问题,而是完全根据 ERR 中的语句得出这些问题。随后,我们利用公开的公司报告对问题的自动化潜力进行了分类。如果问题的答案可以在公司报告中获取,则答案被标记为 "可提取文本"。企业资源报告中 78.7% 的问题可以自动化。这些可自动处理的问题包括 48.2% 的可文本提取问题(适合大型语言模型处理)和 30.5% 的可数据库提取问题。只有 21.3% 的问题需要人工判断才能回答。我们使用 Llama-3-70B 和 GPT-4-turbo-2024-04-09 进行了实证验证,语言生成和信息提取方面的最新进展使得ERR 中约 80% 的语句可以实现自动化。令人惊奇的是,这些模型很好地互补了彼此的优缺点。研究证实,目前的《紧急救济报告》撰写过程有可能从更多的自动化中获益,从而提高质量和效率。因此,这项研究使我们能够量化在 ERR 撰写过程中引入大语言模型的潜在影响。完整的问题清单,包括原型及其频率,将在同行评审后在网上公布。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信