Felix Drinkall, Janet B. Pierrehumbert, Stefan Zohren
{"title":"传统方法在预测信用评级方面优于生成式 LLM","authors":"Felix Drinkall, Janet B. Pierrehumbert, Stefan Zohren","doi":"arxiv-2407.17624","DOIUrl":null,"url":null,"abstract":"Large Language Models (LLMs) have been shown to perform well for many\ndownstream tasks. Transfer learning can enable LLMs to acquire skills that were\nnot targeted during pre-training. In financial contexts, LLMs can sometimes\nbeat well-established benchmarks. This paper investigates how well LLMs perform\nin the task of forecasting corporate credit ratings. We show that while LLMs\nare very good at encoding textual information, traditional methods are still\nvery competitive when it comes to encoding numeric and multimodal data. For our\ntask, current LLMs perform worse than a more traditional XGBoost architecture\nthat combines fundamental and macroeconomic data with high-density text-based\nembedding features.","PeriodicalId":501128,"journal":{"name":"arXiv - QuantFin - Risk Management","volume":"23 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-07-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Traditional Methods Outperform Generative LLMs at Forecasting Credit Ratings\",\"authors\":\"Felix Drinkall, Janet B. Pierrehumbert, Stefan Zohren\",\"doi\":\"arxiv-2407.17624\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Large Language Models (LLMs) have been shown to perform well for many\\ndownstream tasks. Transfer learning can enable LLMs to acquire skills that were\\nnot targeted during pre-training. In financial contexts, LLMs can sometimes\\nbeat well-established benchmarks. This paper investigates how well LLMs perform\\nin the task of forecasting corporate credit ratings. We show that while LLMs\\nare very good at encoding textual information, traditional methods are still\\nvery competitive when it comes to encoding numeric and multimodal data. For our\\ntask, current LLMs perform worse than a more traditional XGBoost architecture\\nthat combines fundamental and macroeconomic data with high-density text-based\\nembedding features.\",\"PeriodicalId\":501128,\"journal\":{\"name\":\"arXiv - QuantFin - Risk Management\",\"volume\":\"23 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-07-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - QuantFin - Risk Management\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2407.17624\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - QuantFin - Risk Management","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2407.17624","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Traditional Methods Outperform Generative LLMs at Forecasting Credit Ratings
Large Language Models (LLMs) have been shown to perform well for many
downstream tasks. Transfer learning can enable LLMs to acquire skills that were
not targeted during pre-training. In financial contexts, LLMs can sometimes
beat well-established benchmarks. This paper investigates how well LLMs perform
in the task of forecasting corporate credit ratings. We show that while LLMs
are very good at encoding textual information, traditional methods are still
very competitive when it comes to encoding numeric and multimodal data. For our
task, current LLMs perform worse than a more traditional XGBoost architecture
that combines fundamental and macroeconomic data with high-density text-based
embedding features.