The paper proposes a novel large language model called NumLLM for Chinese finance. The key contributions are:
Construction of a financial corpus called Fin-Textbooks from financial textbooks, which is essential for improving the numeric capability of language models during fine-tuning.
Development of a novel fine-tuning method with two individual low-rank adaptation (LoRA) modules. One module is for adapting the general-purpose language model to the financial domain, and the other is for enhancing the model's ability to understand financial text with numeric variables.
Experiments on a financial question-answering benchmark show that NumLLM can outperform existing financial language models, including both general-purpose and domain-specific models, on both numeric and non-numeric questions.
The paper first introduces related works on financial corpora and financial language models. It then details the architecture and training process of NumLLM, including the construction of Fin-Textbooks, continual pre-training, and the numeric-sensitive choice tuning (NumCT) method. Finally, the experimental results demonstrate the superior performance of NumLLM compared to various baselines.
To Another Language
from source content
arxiv.org
Key Insights Distilled From
by Huan-Yi Su,K... at arxiv.org 05-02-2024
https://arxiv.org/pdf/2405.00566.pdfDeeper Inquiries