Transformer-based Question Text Generation in the Learning System

Jiajun Li, Huazhu Song, Jun Li
{"title":"Transformer-based Question Text Generation in the Learning System","authors":"Jiajun Li, Huazhu Song, Jun Li","doi":"10.1145/3529466.3529484","DOIUrl":null,"url":null,"abstract":"Question text generation from the triple in knowledge graph exists some challenges in learning system. One is the generated question text is difficult to be understood; the other is it considers few contexts. Therefore, this paper focuses on question text generation. Based on the traditional Bi-LSTM+Attention network model, we import Transformer model into question generation to get the simple question with some triples. In addition, this paper proposes a method to get the diverse expressions of questions (a variety of expressions of a question), that is, to take advantage of the semantic similarity algorithm based on Bi-LSTM with the help of a question database constructed in advance. Finally, a corresponding comparison experiment is designed, and the experimental results demonstrated that the accuracy of question generation experiment based on the Transformer model is 8.36% higher than the traditional Bi-LSTM + Attention network model.","PeriodicalId":375562,"journal":{"name":"Proceedings of the 2022 6th International Conference on Innovation in Artificial Intelligence","volume":"43 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2022 6th International Conference on Innovation in Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3529466.3529484","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Question text generation from the triple in knowledge graph exists some challenges in learning system. One is the generated question text is difficult to be understood; the other is it considers few contexts. Therefore, this paper focuses on question text generation. Based on the traditional Bi-LSTM+Attention network model, we import Transformer model into question generation to get the simple question with some triples. In addition, this paper proposes a method to get the diverse expressions of questions (a variety of expressions of a question), that is, to take advantage of the semantic similarity algorithm based on Bi-LSTM with the help of a question database constructed in advance. Finally, a corresponding comparison experiment is designed, and the experimental results demonstrated that the accuracy of question generation experiment based on the Transformer model is 8.36% higher than the traditional Bi-LSTM + Attention network model.
学习系统中基于转换的问题文本生成
知识图谱中的三元组生成问题文本在学习系统中存在一些挑战。一是生成的问题文本难以理解;另一个是它很少考虑上下文。因此,本文主要研究问题文本生成。在传统的Bi-LSTM+注意力网络模型的基础上,将Transformer模型引入到问题生成中,得到具有三元组的简单问题。此外,本文提出了一种获取问题的多样化表达(一个问题的多种表达)的方法,即利用基于Bi-LSTM的语义相似度算法,借助事先构建好的问题数据库。最后设计了相应的对比实验,实验结果表明,基于Transformer模型的问题生成实验的准确率比传统的Bi-LSTM +注意力网络模型提高了8.36%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信