{"title":"NumLLM: Numeric-Sensitive Large Language Model for Chinese Finance","authors":"Huan-Yi Su, Ke Wu, Yu-Hao Huang, Wu-Jun Li","doi":"arxiv-2405.00566","DOIUrl":null,"url":null,"abstract":"Recently, many works have proposed various financial large language models\n(FinLLMs) by pre-training from scratch or fine-tuning open-sourced LLMs on\nfinancial corpora. However, existing FinLLMs exhibit unsatisfactory performance\nin understanding financial text when numeric variables are involved in\nquestions. In this paper, we propose a novel LLM, called numeric-sensitive\nlarge language model (NumLLM), for Chinese finance. We first construct a\nfinancial corpus from financial textbooks which is essential for improving\nnumeric capability of LLMs during fine-tuning. After that, we train two\nindividual low-rank adaptation (LoRA) modules by fine-tuning on our constructed\nfinancial corpus. One module is for adapting general-purpose LLMs to financial\ndomain, and the other module is for enhancing the ability of NumLLM to\nunderstand financial text with numeric variables. Lastly, we merge the two LoRA\nmodules into the foundation model to obtain NumLLM for inference. Experiments\non financial question-answering benchmark show that NumLLM can boost the\nperformance of the foundation model and can achieve the best overall\nperformance compared to all baselines, on both numeric and non-numeric\nquestions.","PeriodicalId":501372,"journal":{"name":"arXiv - QuantFin - General Finance","volume":"12 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - QuantFin - General Finance","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2405.00566","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Recently, many works have proposed various financial large language models
(FinLLMs) by pre-training from scratch or fine-tuning open-sourced LLMs on
financial corpora. However, existing FinLLMs exhibit unsatisfactory performance
in understanding financial text when numeric variables are involved in
questions. In this paper, we propose a novel LLM, called numeric-sensitive
large language model (NumLLM), for Chinese finance. We first construct a
financial corpus from financial textbooks which is essential for improving
numeric capability of LLMs during fine-tuning. After that, we train two
individual low-rank adaptation (LoRA) modules by fine-tuning on our constructed
financial corpus. One module is for adapting general-purpose LLMs to financial
domain, and the other module is for enhancing the ability of NumLLM to
understand financial text with numeric variables. Lastly, we merge the two LoRA
modules into the foundation model to obtain NumLLM for inference. Experiments
on financial question-answering benchmark show that NumLLM can boost the
performance of the foundation model and can achieve the best overall
performance compared to all baselines, on both numeric and non-numeric
questions.