{"title":"A BERT-Based Pre-Training Model for Solving Math Application Problems","authors":"Yuhao Jia, Pingheng Wang, Zhen Zhang, Chi Cheng, Zhifei Li, Xinguo Yu","doi":"10.1109/IEIR56323.2022.10050073","DOIUrl":null,"url":null,"abstract":"Solving the math application problem is hot research in intelligence education. An increasing number of research scholars are using pre-trained models to tackle machine solution problems. Noteworthily, the semantic relationships required in the machine solution task are for describing math problems, while those of the BERT model with pre-training weights are of general significance, which will cause a mismatched word vector representation. To solve this problem, we proposed a self-supervised pre-training method based on loss priority. We use the input data from the downstream task datasets to fine-tune the existing BERT model so that the dynamic word vector it obtained can better match the downstream tasks. And the size of the loss value of each data batch in each round of training will be recorded to decide which data should be trained in the next round, so that the model has a faster convergence speed. Furthermore, considering that in large-scale mathematics application problems, some problems have almost the same forms of solution. We proposed a machine solution model training algorithm based on the analogy of the same problem type. Extensive experiments on two well-known datasets show the superiority of our proposed algorithms compared to other state-of-the-art algorithms.","PeriodicalId":183709,"journal":{"name":"2022 International Conference on Intelligent Education and Intelligent Research (IEIR)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Conference on Intelligent Education and Intelligent Research (IEIR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IEIR56323.2022.10050073","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Solving the math application problem is hot research in intelligence education. An increasing number of research scholars are using pre-trained models to tackle machine solution problems. Noteworthily, the semantic relationships required in the machine solution task are for describing math problems, while those of the BERT model with pre-training weights are of general significance, which will cause a mismatched word vector representation. To solve this problem, we proposed a self-supervised pre-training method based on loss priority. We use the input data from the downstream task datasets to fine-tune the existing BERT model so that the dynamic word vector it obtained can better match the downstream tasks. And the size of the loss value of each data batch in each round of training will be recorded to decide which data should be trained in the next round, so that the model has a faster convergence speed. Furthermore, considering that in large-scale mathematics application problems, some problems have almost the same forms of solution. We proposed a machine solution model training algorithm based on the analogy of the same problem type. Extensive experiments on two well-known datasets show the superiority of our proposed algorithms compared to other state-of-the-art algorithms.