{"title":"基于选择编码和融合语义丢失的数据到文本生成方法","authors":"Yuelin Chen, ZhuCheng Gao, XiaoDong Cai","doi":"10.1109/ISCEIC53685.2021.00038","DOIUrl":null,"url":null,"abstract":"A method with data to text generation based on selecting encoding and fusing semantic loss is proposed in this paper. By highlighting key content and reducing the redundancy of text description information, the quality of the generated text is significantly improved. First, a new selection network is designed, which uses the amount of information related to data records as the coding basis for content importance, and multiple rounds of dynamic iterations of the results to achieve accurate and comprehensive selection of important information. Secondly, in the decoding process using Long Short-Term Memory (LSTM), a hierarchical attention mechanism is designed to assign dynamic selection weights to different entities and their attributes in the hidden layer output to obtain the best generated text Recall rate. Finally, a method of calculating the semantic similarity loss between the generated text and the reference text is introduced. By calculating the cosine distance of the semantic vectors of the two and iteratively feedback to the training process to obtain the optimization of key features, while reducing the redundancy of description information and improving the model BLEU performance. The experimental results show that the test Precision rate, Recall rate and BLEU is up to 94.58%, 53.72% and 17.24, which are better than existing models.","PeriodicalId":342968,"journal":{"name":"2021 2nd International Symposium on Computer Engineering and Intelligent Communications (ISCEIC)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Method with Data to Text Generation Based on Selecting Encoding and Fusing Semantic Loss\",\"authors\":\"Yuelin Chen, ZhuCheng Gao, XiaoDong Cai\",\"doi\":\"10.1109/ISCEIC53685.2021.00038\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"A method with data to text generation based on selecting encoding and fusing semantic loss is proposed in this paper. By highlighting key content and reducing the redundancy of text description information, the quality of the generated text is significantly improved. First, a new selection network is designed, which uses the amount of information related to data records as the coding basis for content importance, and multiple rounds of dynamic iterations of the results to achieve accurate and comprehensive selection of important information. Secondly, in the decoding process using Long Short-Term Memory (LSTM), a hierarchical attention mechanism is designed to assign dynamic selection weights to different entities and their attributes in the hidden layer output to obtain the best generated text Recall rate. Finally, a method of calculating the semantic similarity loss between the generated text and the reference text is introduced. By calculating the cosine distance of the semantic vectors of the two and iteratively feedback to the training process to obtain the optimization of key features, while reducing the redundancy of description information and improving the model BLEU performance. The experimental results show that the test Precision rate, Recall rate and BLEU is up to 94.58%, 53.72% and 17.24, which are better than existing models.\",\"PeriodicalId\":342968,\"journal\":{\"name\":\"2021 2nd International Symposium on Computer Engineering and Intelligent Communications (ISCEIC)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 2nd International Symposium on Computer Engineering and Intelligent Communications (ISCEIC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ISCEIC53685.2021.00038\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 2nd International Symposium on Computer Engineering and Intelligent Communications (ISCEIC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISCEIC53685.2021.00038","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Method with Data to Text Generation Based on Selecting Encoding and Fusing Semantic Loss
A method with data to text generation based on selecting encoding and fusing semantic loss is proposed in this paper. By highlighting key content and reducing the redundancy of text description information, the quality of the generated text is significantly improved. First, a new selection network is designed, which uses the amount of information related to data records as the coding basis for content importance, and multiple rounds of dynamic iterations of the results to achieve accurate and comprehensive selection of important information. Secondly, in the decoding process using Long Short-Term Memory (LSTM), a hierarchical attention mechanism is designed to assign dynamic selection weights to different entities and their attributes in the hidden layer output to obtain the best generated text Recall rate. Finally, a method of calculating the semantic similarity loss between the generated text and the reference text is introduced. By calculating the cosine distance of the semantic vectors of the two and iteratively feedback to the training process to obtain the optimization of key features, while reducing the redundancy of description information and improving the model BLEU performance. The experimental results show that the test Precision rate, Recall rate and BLEU is up to 94.58%, 53.72% and 17.24, which are better than existing models.