Sohini Roychowdhury, Marko Krema, Brian Moore, Xingjian Lai, Dike Effedua, Bharat Jethwani
{"title":"FiST--金融风格转移与幻觉和创造力控制框架","authors":"Sohini Roychowdhury, Marko Krema, Brian Moore, Xingjian Lai, Dike Effedua, Bharat Jethwani","doi":"arxiv-2408.05365","DOIUrl":null,"url":null,"abstract":"Financial report generation using general purpose large language models pose\ntwo major challenges, including the lack of compound sentences and\nhallucinations. Advanced prompt engineering and retrieval augmented generation\n(RAG) techniques are incapable of curing the writing style discrepancies. In\nthis work we propose a novel two-stage fine-tuning process wherein public\ndomain financial reports are processed into prompt-completions and augmented\nusing simple LLM prompts to then enable sectional financial report generation\nusing minimal instructions and tabular data inputs. Our proposed fine-tuning\nframework results doubles the number of correct questions answers and reduces\nhallucinations by over 50%. Additionally, the two-stage fine tuned models have\nlower perplexity, improved ROUGE, TER and BLEU scores, higher creativity and\nknowledge density with lower uncertainty and cross entropy.","PeriodicalId":501309,"journal":{"name":"arXiv - CS - Computational Engineering, Finance, and Science","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"FiST-Financial Style Transfer with Hallucination and Creativity Control Framework\",\"authors\":\"Sohini Roychowdhury, Marko Krema, Brian Moore, Xingjian Lai, Dike Effedua, Bharat Jethwani\",\"doi\":\"arxiv-2408.05365\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Financial report generation using general purpose large language models pose\\ntwo major challenges, including the lack of compound sentences and\\nhallucinations. Advanced prompt engineering and retrieval augmented generation\\n(RAG) techniques are incapable of curing the writing style discrepancies. In\\nthis work we propose a novel two-stage fine-tuning process wherein public\\ndomain financial reports are processed into prompt-completions and augmented\\nusing simple LLM prompts to then enable sectional financial report generation\\nusing minimal instructions and tabular data inputs. Our proposed fine-tuning\\nframework results doubles the number of correct questions answers and reduces\\nhallucinations by over 50%. Additionally, the two-stage fine tuned models have\\nlower perplexity, improved ROUGE, TER and BLEU scores, higher creativity and\\nknowledge density with lower uncertainty and cross entropy.\",\"PeriodicalId\":501309,\"journal\":{\"name\":\"arXiv - CS - Computational Engineering, Finance, and Science\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-08-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Computational Engineering, Finance, and Science\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2408.05365\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Computational Engineering, Finance, and Science","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.05365","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
FiST-Financial Style Transfer with Hallucination and Creativity Control Framework
Financial report generation using general purpose large language models pose
two major challenges, including the lack of compound sentences and
hallucinations. Advanced prompt engineering and retrieval augmented generation
(RAG) techniques are incapable of curing the writing style discrepancies. In
this work we propose a novel two-stage fine-tuning process wherein public
domain financial reports are processed into prompt-completions and augmented
using simple LLM prompts to then enable sectional financial report generation
using minimal instructions and tabular data inputs. Our proposed fine-tuning
framework results doubles the number of correct questions answers and reduces
hallucinations by over 50%. Additionally, the two-stage fine tuned models have
lower perplexity, improved ROUGE, TER and BLEU scores, higher creativity and
knowledge density with lower uncertainty and cross entropy.