利用注意力的神经机器翻译

Dafni Rose, K. Vijayakumar, D. Kirubakaran, R. Pugalenthi, Gotti Balayaswantasaichowdary
{"title":"利用注意力的神经机器翻译","authors":"Dafni Rose, K. Vijayakumar, D. Kirubakaran, R. Pugalenthi, Gotti Balayaswantasaichowdary","doi":"10.1109/ICECONF57129.2023.10083569","DOIUrl":null,"url":null,"abstract":"Machine Translation pertains to translation of one natural language to other by using automated computing. The most common method for dealing with the machine translation problem is Statistical machine translation. This method is convenient for language pairs with similar grammatical structures even so it taken vast datasets.N evertheless, the conventional models do not perform well for languages without similar grammar and contextual meaning. Lately this problemhas been resolved by the neural machine translation (NMT) that has proved to be an effective curative. Only a little amount of data is required for training in NMT and it can translate only a small number of training words. A fixed-length vector is used to identify the important words that contribute for the translation of text, and assigns weights to each word in our proposed system. The Encoder-Decoder architecture with Long- Term and Short- Term Memory (LSTM) Neural Network and trained modelsare employed by calling the previous sequences and states. The proposed model ameliorates translation performance with attention vector and by returning the sequences of previous states unlike LSTM.English-Hindi sentences corpus data for implementing a Model with attention and without attention is considered here. By evaluating the results, the proposed solution, overcomes complexity of training a Neural Network and increases translation performance.","PeriodicalId":436733,"journal":{"name":"2023 International Conference on Artificial Intelligence and Knowledge Discovery in Concurrent Engineering (ICECONF)","volume":"70 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Neural Machine Translation Using Attention\",\"authors\":\"Dafni Rose, K. Vijayakumar, D. Kirubakaran, R. Pugalenthi, Gotti Balayaswantasaichowdary\",\"doi\":\"10.1109/ICECONF57129.2023.10083569\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Machine Translation pertains to translation of one natural language to other by using automated computing. The most common method for dealing with the machine translation problem is Statistical machine translation. This method is convenient for language pairs with similar grammatical structures even so it taken vast datasets.N evertheless, the conventional models do not perform well for languages without similar grammar and contextual meaning. Lately this problemhas been resolved by the neural machine translation (NMT) that has proved to be an effective curative. Only a little amount of data is required for training in NMT and it can translate only a small number of training words. A fixed-length vector is used to identify the important words that contribute for the translation of text, and assigns weights to each word in our proposed system. The Encoder-Decoder architecture with Long- Term and Short- Term Memory (LSTM) Neural Network and trained modelsare employed by calling the previous sequences and states. The proposed model ameliorates translation performance with attention vector and by returning the sequences of previous states unlike LSTM.English-Hindi sentences corpus data for implementing a Model with attention and without attention is considered here. By evaluating the results, the proposed solution, overcomes complexity of training a Neural Network and increases translation performance.\",\"PeriodicalId\":436733,\"journal\":{\"name\":\"2023 International Conference on Artificial Intelligence and Knowledge Discovery in Concurrent Engineering (ICECONF)\",\"volume\":\"70 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-01-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 International Conference on Artificial Intelligence and Knowledge Discovery in Concurrent Engineering (ICECONF)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICECONF57129.2023.10083569\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 International Conference on Artificial Intelligence and Knowledge Discovery in Concurrent Engineering (ICECONF)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICECONF57129.2023.10083569","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

摘要

机器翻译是指通过自动计算将一种自然语言翻译成另一种自然语言。处理机器翻译问题最常用的方法是统计机器翻译。这种方法对于语法结构相似的语言对来说很方便,尽管它需要大量的数据集。然而,对于没有相似语法和上下文含义的语言,传统模型表现不佳。近年来,神经机器翻译(NMT)已被证明是一种有效的治疗方法。NMT的训练只需要少量的数据,也只能翻译少量的训练词。一个固定长度的向量用于识别对文本翻译有贡献的重要单词,并在我们提出的系统中为每个单词分配权重。通过调用之前的序列和状态,采用长短期记忆(LSTM)神经网络和训练模型的编码器-解码器结构。与LSTM模型不同,该模型利用注意力向量和返回前状态序列来改善翻译性能。这里考虑了用于实现有注意和无注意模型的英语-印地语句子语料库数据。结果表明,该方法克服了神经网络训练的复杂性,提高了翻译性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Neural Machine Translation Using Attention
Machine Translation pertains to translation of one natural language to other by using automated computing. The most common method for dealing with the machine translation problem is Statistical machine translation. This method is convenient for language pairs with similar grammatical structures even so it taken vast datasets.N evertheless, the conventional models do not perform well for languages without similar grammar and contextual meaning. Lately this problemhas been resolved by the neural machine translation (NMT) that has proved to be an effective curative. Only a little amount of data is required for training in NMT and it can translate only a small number of training words. A fixed-length vector is used to identify the important words that contribute for the translation of text, and assigns weights to each word in our proposed system. The Encoder-Decoder architecture with Long- Term and Short- Term Memory (LSTM) Neural Network and trained modelsare employed by calling the previous sequences and states. The proposed model ameliorates translation performance with attention vector and by returning the sequences of previous states unlike LSTM.English-Hindi sentences corpus data for implementing a Model with attention and without attention is considered here. By evaluating the results, the proposed solution, overcomes complexity of training a Neural Network and increases translation performance.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信