Chao Liu , Yuxia Miao , Qi Zhao , Chao Wang , Xiangyu Zhu
{"title":"Multimodal stock market emotion recognition model trained with a large language model","authors":"Chao Liu , Yuxia Miao , Qi Zhao , Chao Wang , Xiangyu Zhu","doi":"10.1016/j.engappai.2025.111035","DOIUrl":null,"url":null,"abstract":"<div><div>Stock market emotion recognition models are often trained on public textual datasets. These datasets are well-labelled but may not have the same distribution as real-world data and consequently fail to reflect the real world. Furthermore, stock market emotion is reflected not only in textual comments but also in images; emotion recognition on the basis of textual comments alone is one-sided. Considering these issues, in this paper, we propose a multimodal emotion recognition framework that is trained via imitation learning from a large language model. The proposed framework contains two main innovations. First, by leveraging large language model's compositionality capability, the proposed framework generates pseudo labels from textual comments. Through imitating the patterns of the pseudo labels, the framework can be trained directly on unlabelled real-world data, which addresses current models' distribution drift between public datasets and the real-world. Second, multimodal fusion is equipped with the proposed framework, enabling emotion recognition from textual stock market comments and images simultaneously. Compared with existing methods, the proposed method in this paper significantly improves the performance of stock market emotion recognition by leveraging large language models and multimodal stock market data. The experimental results demonstrate that the emotion recognition framework proposed in this paper outperforms the existing single-mode models and multimodal models, with the accuracy, precision, recall, and F1 score being 82.97 %, 83.03 %, 83.05 %, and 82.88 % respectively. This framework provides a new pathway for multimodal and unsupervised/semi-supervised emotion recognition in the stock market.</div></div>","PeriodicalId":50523,"journal":{"name":"Engineering Applications of Artificial Intelligence","volume":"154 ","pages":"Article 111035"},"PeriodicalIF":8.0000,"publicationDate":"2025-05-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Engineering Applications of Artificial Intelligence","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0952197625010358","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Stock market emotion recognition models are often trained on public textual datasets. These datasets are well-labelled but may not have the same distribution as real-world data and consequently fail to reflect the real world. Furthermore, stock market emotion is reflected not only in textual comments but also in images; emotion recognition on the basis of textual comments alone is one-sided. Considering these issues, in this paper, we propose a multimodal emotion recognition framework that is trained via imitation learning from a large language model. The proposed framework contains two main innovations. First, by leveraging large language model's compositionality capability, the proposed framework generates pseudo labels from textual comments. Through imitating the patterns of the pseudo labels, the framework can be trained directly on unlabelled real-world data, which addresses current models' distribution drift between public datasets and the real-world. Second, multimodal fusion is equipped with the proposed framework, enabling emotion recognition from textual stock market comments and images simultaneously. Compared with existing methods, the proposed method in this paper significantly improves the performance of stock market emotion recognition by leveraging large language models and multimodal stock market data. The experimental results demonstrate that the emotion recognition framework proposed in this paper outperforms the existing single-mode models and multimodal models, with the accuracy, precision, recall, and F1 score being 82.97 %, 83.03 %, 83.05 %, and 82.88 % respectively. This framework provides a new pathway for multimodal and unsupervised/semi-supervised emotion recognition in the stock market.
期刊介绍:
Artificial Intelligence (AI) is pivotal in driving the fourth industrial revolution, witnessing remarkable advancements across various machine learning methodologies. AI techniques have become indispensable tools for practicing engineers, enabling them to tackle previously insurmountable challenges. Engineering Applications of Artificial Intelligence serves as a global platform for the swift dissemination of research elucidating the practical application of AI methods across all engineering disciplines. Submitted papers are expected to present novel aspects of AI utilized in real-world engineering applications, validated using publicly available datasets to ensure the replicability of research outcomes. Join us in exploring the transformative potential of AI in engineering.