Computer Speech and Language最新文献

筛选
英文 中文
Unsupervised question-retrieval approach based on topic keywords filtering and multi-task learning 基于主题关键词过滤和多任务学习的无监督问题检索方法
IF 4.3 3区 计算机科学
Computer Speech and Language Pub Date : 2024-03-21 DOI: 10.1016/j.csl.2024.101644
Aiguo Shang , Xinjuan Zhu , Michael Danner , Matthias Rätsch
{"title":"Unsupervised question-retrieval approach based on topic keywords filtering and multi-task learning","authors":"Aiguo Shang ,&nbsp;Xinjuan Zhu ,&nbsp;Michael Danner ,&nbsp;Matthias Rätsch","doi":"10.1016/j.csl.2024.101644","DOIUrl":"10.1016/j.csl.2024.101644","url":null,"abstract":"<div><p>Currently, the majority of retrieval-based question-answering systems depend on supervised training using question pairs. However, there is still a significant need for further exploration of how to employ unsupervised methods to improve the accuracy of retrieval-based question-answering systems. From the perspective of question topic keywords, this paper presents TFCSG, an unsupervised question-retrieval approach based on topic keyword filtering and multi-task learning. Firstly, we design the topic keyword filtering algorithm, which, unlike the topic model, can sequentially filter out the keywords of the question and can provide a training corpus for subsequent unsupervised learning. Then, three tasks are designed in this paper to complete the training of the question-retrieval model. The first task is a question contrastive learning task based on topic keywords repetition strategy, the second is questions and its corresponding sequential topic keywords similarity distribution task, and the third is a sequential topic keywords generation task using questions. These three tasks are trained in parallel in order to obtain quality question representations and thus improve the accuracy of question-retrieval task. Finally, our experimental results on the four publicly available datasets demonstrate the effectiveness of the TFCSG, with an average improvement of 7.1%, 4.4%, and 3.5% in the P@1, MAP, and MRR metrics when using the BERT model compared to the baseline model. The corresponding metrics improved by 5.7%, 3.5% and 3.0% on average when using the RoBERTa model. The accuracy of unsupervised similar question-retrieval task is effectively improved. In particular, the values of P@1, P@5, and P@10 are close, the retrieved similar questions are ranked more advance.</p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":"87 ","pages":"Article 101644"},"PeriodicalIF":4.3,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140280805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A novel joint extraction model based on cross-attention mechanism and global pointer using context shield window 基于交叉注意机制和全局指针的新型联合提取模型(使用上下文屏蔽窗口
IF 4.3 3区 计算机科学
Computer Speech and Language Pub Date : 2024-03-18 DOI: 10.1016/j.csl.2024.101643
Zhengwei Zhai , Rongli Fan , Jie Huang , Neal Xiong , Lijuan Zhang , Jian Wan , Lei Zhang
{"title":"A novel joint extraction model based on cross-attention mechanism and global pointer using context shield window","authors":"Zhengwei Zhai ,&nbsp;Rongli Fan ,&nbsp;Jie Huang ,&nbsp;Neal Xiong ,&nbsp;Lijuan Zhang ,&nbsp;Jian Wan ,&nbsp;Lei Zhang","doi":"10.1016/j.csl.2024.101643","DOIUrl":"10.1016/j.csl.2024.101643","url":null,"abstract":"<div><p>Relational triple extraction is a critical step in knowledge graph construction. Compared to pipeline-based extraction, joint extraction is gaining more attention because it can better utilize entity and relation information without causing error propagation issues. Yet, the challenge with joint extraction lies in handling overlapping triples. Existing approaches adopt sequential steps or multiple modules, which often accumulate errors and interfere with redundant data. In this study, we propose an innovative joint extraction model with cross-attention mechanism and global pointers with context shield window. Specifically, our methodology begins by inputting text data into a pre-trained RoBERTa model to generate word vector representations. Subsequently, these embeddings are passed through a modified cross-attention layer along with entity type embeddings to address missing entity type information. Next, we employ the global pointer to transform the extraction problem into a quintuple extraction problem, which skillfully solves the issue of overlapping triples. It is worth mentioning that we design a context shield window on the global pointer, which facilitates the identification of correct entities within a limited range during the entity extraction process. Finally, the capability of our model against malicious samples is improved by adding adversarial training during the training process. Demonstrating superiority over mainstream models, our approach achieves impressive results on three publicly available datasets.</p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":"87 ","pages":"Article 101643"},"PeriodicalIF":4.3,"publicationDate":"2024-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S0885230824000263/pdfft?md5=6db6d29053e0503fc07e8e1ded002d0e&pid=1-s2.0-S0885230824000263-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140181814","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving linear orthogonal mapping based cross-lingual representation using ridge regression and graph centrality 利用脊回归和图中心性改进基于线性正交映射的跨语言表示法
IF 4.3 3区 计算机科学
Computer Speech and Language Pub Date : 2024-03-16 DOI: 10.1016/j.csl.2024.101640
Deepen Naorem, Sanasam Ranbir Singh, Priyankoo Sarmah
{"title":"Improving linear orthogonal mapping based cross-lingual representation using ridge regression and graph centrality","authors":"Deepen Naorem,&nbsp;Sanasam Ranbir Singh,&nbsp;Priyankoo Sarmah","doi":"10.1016/j.csl.2024.101640","DOIUrl":"10.1016/j.csl.2024.101640","url":null,"abstract":"<div><p>Orthogonal linear mapping is a commonly used approach for generating cross-lingual embedding between two monolingual corpora that uses a word frequency-based seed dictionary alignment approach. While this approach is found to be effective for isomorphic language pairs, they do not perform well for distant language pairs with different sentence structures and morphological properties. For a distance language pair, the existing frequency-aligned orthogonal mapping methods suffer from two problems - (i)the frequency of source and target word are not comparable, and (ii)different word pairs in the seed dictionary may have different contribution. Motivated by the above two concerns, this paper proposes a novel centrality-aligned ridge regression-based orthogonal mapping. The proposed method uses centrality-based alignment for seed dictionary selection and ridge regression framework for incorporating influential weights of different word pairs in the seed dictionary. From various experimental observations over five language pairs (both isomorphic and distant languages), it is evident that the proposed method outperforms baseline methods in the Bilingual Dictionary Induction(BDI) task, Sentence Retrieval Task(SRT), and Machine Translation. Further, several analyses are also included to support the proposed method.</p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":"87 ","pages":"Article 101640"},"PeriodicalIF":4.3,"publicationDate":"2024-03-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140181780","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A closer look at reinforcement learning-based automatic speech recognition 进一步了解基于强化学习的自动语音识别技术
IF 4.3 3区 计算机科学
Computer Speech and Language Pub Date : 2024-03-16 DOI: 10.1016/j.csl.2024.101641
Fan Yang , Muqiao Yang , Xiang Li , Yuxuan Wu , Zhiyuan Zhao , Bhiksha Raj , Rita Singh
{"title":"A closer look at reinforcement learning-based automatic speech recognition","authors":"Fan Yang ,&nbsp;Muqiao Yang ,&nbsp;Xiang Li ,&nbsp;Yuxuan Wu ,&nbsp;Zhiyuan Zhao ,&nbsp;Bhiksha Raj ,&nbsp;Rita Singh","doi":"10.1016/j.csl.2024.101641","DOIUrl":"10.1016/j.csl.2024.101641","url":null,"abstract":"<div><p>Reinforcement learning (RL) has demonstrated effectiveness in improving model performance and robustness for automatic speech recognition (ASR) tasks. Researchers have employed RL-based training strategies to enhance performance beyond conventional supervised or semi-supervised learning. However, existing approaches treat RL as a supplementary tool, leaving the untapped potential of RL training largely unexplored. In this paper, we formulate a novel pure RL setting where an ASR model is trained exclusively through RL via human feedback metrics, e.g., word error rate (WER) or binary reward. This approach promises to significantly simplify the annotation process if we could replace the conventional onerous annotation with a single numeric value in a human–computer interaction (HCI) way. Our experiments demonstrate the feasibility of this new setting and also identify two main inherent issues in conventional RL-based ASR training that may lead to performance degradation: (1) the mismatch issue between the action and reward has been commonly overlooked in Connectionist Temporal Classification (CTC) based models, which is attributed to the inherent CTC alignment mapping issue; (2) the classic exploration–exploitation trade-off still exists in the sampling stage of RL-based ASR, and finding the balance between them becomes a challenge. To address these issues, we first propose a new RL-based approach named CTC-aligned Policy Gradient (CTC-PG), which provides a unified formulation for different sampling strategies and alleviates the mismatch issue of action and reward in CTC-based models. Moreover, we propose Focal Sampling to balance the trade-off between exploration and exploitation with a flexible temperature parameter. Experiment results on LibriSpeech dataset showcase the effectiveness and robustness of our methods by harnessing the full potential of RL in training ASR models.</p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":"87 ","pages":"Article 101641"},"PeriodicalIF":4.3,"publicationDate":"2024-03-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140181821","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Higher order statistics-driven magnitude and phase spectrum estimation for speech enhancement 用于语音增强的高阶统计驱动的幅度和相位频谱估计
IF 4.3 3区 计算机科学
Computer Speech and Language Pub Date : 2024-03-16 DOI: 10.1016/j.csl.2024.101639
T. Lavanya , P. Vijayalakshmi , K. Mrinalini , T. Nagarajan
{"title":"Higher order statistics-driven magnitude and phase spectrum estimation for speech enhancement","authors":"T. Lavanya ,&nbsp;P. Vijayalakshmi ,&nbsp;K. Mrinalini ,&nbsp;T. Nagarajan","doi":"10.1016/j.csl.2024.101639","DOIUrl":"https://doi.org/10.1016/j.csl.2024.101639","url":null,"abstract":"<div><p>Higher order statistics (HOS), can be effectively employed for noise suppression, provided the noise follows a Gaussian distribution. Since most of the noises are distributed normally, HOS can be effectively used for speech enhancement in noisy environments. In the current work, HOS-based parametric modelling for magnitude spectrum estimation is proposed to improve the SNR under noisy conditions. To establish this, a non-Gaussian reduced ARMA model formulated using third order cumulant sequences (Giannakis, 1990) is used. Here, the AR and MA model orders, <span><math><mi>p</mi></math></span> and <span><math><mi>q</mi></math></span>, are dynamically estimated by the well-established periodicity estimation technique under noisy conditions namely the Ramanujan Filter Bank (RFB) approach. The AR coefficients estimated from the reduced ARMA model are used to obtain the partially enhanced speech output, whose magnitude spectrum is then subjected to second-level enhancement using log MMSE with modified speech presence uncertainty (SPU) estimation technique. The refined magnitude spectrum, is combined with the phase spectrum extracted using proposed bicoherence-based phase compensation (BPC) technique, to estimate the enhanced speech output. The HOS-driven speech enhancement technique proposed in the current work is observed to be efficient for white, pink, babble and buccaneer noises. The objective measures, PESQ and STOI, indicate that the proposed method works well under all the noise conditions considered for evaluation.</p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":"87 ","pages":"Article 101639"},"PeriodicalIF":4.3,"publicationDate":"2024-03-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140162524","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multipath-guided heterogeneous graph neural networks for sequential recommendation 用于顺序推荐的多路径引导异构图神经网络
IF 4.3 3区 计算机科学
Computer Speech and Language Pub Date : 2024-03-16 DOI: 10.1016/j.csl.2024.101642
Fulian Yin , Tongtong Xing , Meiqi Ji , Zebin Yao , Ruiling Fu , Yuewei Wu
{"title":"Multipath-guided heterogeneous graph neural networks for sequential recommendation","authors":"Fulian Yin ,&nbsp;Tongtong Xing ,&nbsp;Meiqi Ji ,&nbsp;Zebin Yao ,&nbsp;Ruiling Fu ,&nbsp;Yuewei Wu","doi":"10.1016/j.csl.2024.101642","DOIUrl":"10.1016/j.csl.2024.101642","url":null,"abstract":"<div><p>With the explosion of information and users’ changing interest, program sequential recommendation becomes increasingly important for TV program platforms to help their users find interesting programs. Existing sequential recommendation methods mainly focus on modeling user preferences from users’ historical interaction behaviors directly, with insufficient learning about the dynamics of programs and users, while ignoring the rich semantic information from the heterogeneous graph. To address these issues, we propose the multipath-guided heterogeneous graph neural networks for TV program sequential recommendation (MHG-PSR), which can enhance the representations of programs and users through multiple paths in heterogeneous graphs. In our method, the auxiliary information is fused to supplement the semantics of program and user to obtain initial representations. Then, we explore the interactive behaviors of programs and users with temporal and auxiliary information to model the collaborative signals in the heterogeneous graph and extract the users’ dynamic preferences of programs. Extensive experiments on real-world datasets verify the proposed method can effectively improve the performance of TV program sequential recommendation.</p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":"87 ","pages":"Article 101642"},"PeriodicalIF":4.3,"publicationDate":"2024-03-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140181846","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Incorporating external knowledge for text matching model 将外部知识纳入文本匹配模型
IF 4.3 3区 计算机科学
Computer Speech and Language Pub Date : 2024-03-12 DOI: 10.1016/j.csl.2024.101638
Kexin Jiang , Guozhe Jin , Zhenguo Zhang , Rongyi Cui , Yahui Zhao
{"title":"Incorporating external knowledge for text matching model","authors":"Kexin Jiang ,&nbsp;Guozhe Jin ,&nbsp;Zhenguo Zhang ,&nbsp;Rongyi Cui ,&nbsp;Yahui Zhao","doi":"10.1016/j.csl.2024.101638","DOIUrl":"10.1016/j.csl.2024.101638","url":null,"abstract":"<div><p>Text matching is a computational task that involves comparing and establishing the semantic relationship between two textual inputs. The prevailing approach in text matching entails the computation of textual representations or employing attention mechanisms to facilitate interaction with the text. These techniques have demonstrated notable efficacy in various text-matching scenarios. However, these methods primarily focus on modeling the sentence pairs themselves and rarely incorporate additional information to enrich the models. In this study, we address the challenge of text matching in natural language processing by proposing a novel approach that leverages external knowledge sources, namely Wiktionary for word definitions and a knowledge graph for text triplet information. Unlike conventional methods that primarily rely on textual representations and attention mechanisms, our approach enhances semantic understanding by integrating relevant external information. We introduce a fusion module to amalgamate the semantic insights derived from the text and the external knowledge. Our methodology’s efficacy is evidenced through comprehensive experiments conducted on diverse datasets, encompassing natural language inference, text classification, and medical natural language inference. The results unequivocally indicate a significant enhancement in model performance, underscoring the effectiveness of incorporating external knowledge into text-matching tasks.</p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":"87 ","pages":"Article 101638"},"PeriodicalIF":4.3,"publicationDate":"2024-03-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140124573","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
KddRES: A Multi-level Knowledge-driven Dialogue Dataset for Restaurant Towards Customized Dialogue System KddRES:面向餐厅的多层次知识驱动型对话数据集--迈向定制化对话系统
IF 4.3 3区 计算机科学
Computer Speech and Language Pub Date : 2024-03-07 DOI: 10.1016/j.csl.2024.101637
Hongru Wang , Wai-Chung Kwan , Min Li , Zimo Zhou , Kam-Fai Wong
{"title":"KddRES: A Multi-level Knowledge-driven Dialogue Dataset for Restaurant Towards Customized Dialogue System","authors":"Hongru Wang ,&nbsp;Wai-Chung Kwan ,&nbsp;Min Li ,&nbsp;Zimo Zhou ,&nbsp;Kam-Fai Wong","doi":"10.1016/j.csl.2024.101637","DOIUrl":"https://doi.org/10.1016/j.csl.2024.101637","url":null,"abstract":"<div><p>To alleviate the shortage of dialogue datasets for Cantonese, one of the low-resource languages, and facilitate the development of customized task-oriented dialogue systems, we propose <strong>KddRES</strong>, the first Cantonese <strong>K</strong>nowledge-driven <strong>d</strong>ialogue <strong>d</strong>ataset for <strong>RES</strong>taurants. It contains 834 multi-turn dialogues, 8000 utterances, and 26 distinct slots. The slots are hierarchical, and beneath the 26 coarse-grained slots are the additional 16 fine-grained slots. Annotations of dialogue states and dialogue actions at both the user and system sides are provided to suit multiple downstream tasks such as natural language understanding and dialogue state tracking. To effectively detect hierarchical slots, we propose a framework HierBERT by modelling label semantics and relationships between different slots. Experimental results demonstrate that KddRES is more challenging compared with existing datasets due to the introduction of hierarchical slots and our framework is particularly effective in detecting secondary slots and achieving a new state-of-the-art. Given the rich annotation and hierarchical slot structure of KddRES, we hope it will promote research on the development of customized dialogue systems in Cantonese and other conversational AI tasks, such as dialogue state tracking and policy learning.</p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":"87 ","pages":"Article 101637"},"PeriodicalIF":4.3,"publicationDate":"2024-03-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140095934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Model discrepancy policy optimization for task-oriented dialogue 面向任务对话的模型差异策略优化
IF 4.3 3区 计算机科学
Computer Speech and Language Pub Date : 2024-03-06 DOI: 10.1016/j.csl.2024.101636
Zhenyou Zhou, Zhibin Liu, Zhaoan Dong, Yuhan Liu
{"title":"Model discrepancy policy optimization for task-oriented dialogue","authors":"Zhenyou Zhou,&nbsp;Zhibin Liu,&nbsp;Zhaoan Dong,&nbsp;Yuhan Liu","doi":"10.1016/j.csl.2024.101636","DOIUrl":"10.1016/j.csl.2024.101636","url":null,"abstract":"<div><p>Task-oriented dialogue systems use deep reinforcement learning (DRL) to learn policies, and agent interaction with user models can help the agent enhance its generalization capacity. But user models frequently lack the language complexity of human interlocutors and contain generative errors, and their design biases can impair the agent’s ability to function well in certain situations. In this paper, we incorporate an evaluator based on inverse reinforcement learning into the model to determine the quality of the dialogue of user models in order to recruit high-quality user models for training. We can successfully regulate the quality of training trajectories while maintaining their diversity by constructing a sampling environment distribution to pick high-quality user models to participate in policy learning. The evaluation on the Multiwoz dataset demonstrates that it is capable of successfully improving the dialogue agents’ performance.</p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":"87 ","pages":"Article 101636"},"PeriodicalIF":4.3,"publicationDate":"2024-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140047069","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Next word prediction for Urdu language using deep learning models 使用深度学习模型预测乌尔都语的下一个单词
IF 4.3 3区 计算机科学
Computer Speech and Language Pub Date : 2024-03-02 DOI: 10.1016/j.csl.2024.101635
Ramish Shahid, Aamir Wali, Maryam Bashir
{"title":"Next word prediction for Urdu language using deep learning models","authors":"Ramish Shahid,&nbsp;Aamir Wali,&nbsp;Maryam Bashir","doi":"10.1016/j.csl.2024.101635","DOIUrl":"https://doi.org/10.1016/j.csl.2024.101635","url":null,"abstract":"<div><p>Deep learning models are being used for natural language processing. Despite their success, these models have been employed for only a few languages. Pretrained models also exist but they are mostly available for the English language. Low resource languages like Urdu are not able to benefit from these pre-trained deep learning models and their effectiveness in Urdu language processing remains a question. This paper investigates the usefulness of deep learning models for the next word prediction and suggestion model for Urdu. For this purpose, this study considers and proposes two word prediction models for Urdu. Firstly, we propose to use LSTM for neural language modeling of Urdu. LSTMs are a popular approach for language modeling due to their ability to process sequential data. Secondly, we employ BERT which was specifically designed for natural language modeling. We train BERT from scratch using an Urdu corpus consisting of 1.1 million sentences thus paving the way for further studies in the Urdu language. We achieved an accuracy of 52.4% with LSTM and 73.7% with BERT. Our proposed BERT model outperformed two other pre-trained BERT models developed for Urdu. Since this is a multi-class problem and the number of classes is equal to the vocabulary size, this accuracy is still promising. Based on the present performance, BERT seems to be effective for the Urdu language, and this paper lays the groundwork for future studies.</p></div>","PeriodicalId":50638,"journal":{"name":"Computer Speech and Language","volume":"87 ","pages":"Article 101635"},"PeriodicalIF":4.3,"publicationDate":"2024-03-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140016379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信