{"title":"MSMF: Multi-Scale Multi-Modal Fusion for Enhanced Stock Market Prediction","authors":"Jiahao Qin","doi":"arxiv-2409.07855","DOIUrl":null,"url":null,"abstract":"This paper presents MSMF (Multi-Scale Multi-Modal Fusion), a novel approach\nfor enhanced stock market prediction. MSMF addresses key challenges in\nmulti-modal stock analysis by integrating a modality completion encoder,\nmulti-scale feature extraction, and an innovative fusion mechanism. Our model\nleverages blank learning and progressive fusion to balance complementarity and\nredundancy across modalities, while multi-scale alignment facilitates direct\ncorrelations between heterogeneous data types. We introduce Multi-Granularity\nGates and a specialized architecture to optimize the integration of local and\nglobal information for different tasks. Additionally, a Task-targeted\nPrediction layer is employed to preserve both coarse and fine-grained features\nduring fusion. Experimental results demonstrate that MSMF outperforms existing\nmethods, achieving significant improvements in accuracy and reducing prediction\nerrors across various stock market forecasting tasks. This research contributes\nvaluable insights to the field of multi-modal financial analysis and offers a\nrobust framework for enhanced market prediction.","PeriodicalId":501480,"journal":{"name":"arXiv - CS - Multimedia","volume":"34 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Multimedia","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07855","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This paper presents MSMF (Multi-Scale Multi-Modal Fusion), a novel approach
for enhanced stock market prediction. MSMF addresses key challenges in
multi-modal stock analysis by integrating a modality completion encoder,
multi-scale feature extraction, and an innovative fusion mechanism. Our model
leverages blank learning and progressive fusion to balance complementarity and
redundancy across modalities, while multi-scale alignment facilitates direct
correlations between heterogeneous data types. We introduce Multi-Granularity
Gates and a specialized architecture to optimize the integration of local and
global information for different tasks. Additionally, a Task-targeted
Prediction layer is employed to preserve both coarse and fine-grained features
during fusion. Experimental results demonstrate that MSMF outperforms existing
methods, achieving significant improvements in accuracy and reducing prediction
errors across various stock market forecasting tasks. This research contributes
valuable insights to the field of multi-modal financial analysis and offers a
robust framework for enhanced market prediction.