Natural Language Processing Journal最新文献

筛选
英文 中文
BNVGLENET: Hypercomplex Bangla handwriting character recognition with hierarchical class expansion using Convolutional Neural Networks BNVGLENET:利用卷积神经网络进行分层类扩展的超复杂孟加拉语手写字符识别
Natural Language Processing Journal Pub Date : 2024-04-15 DOI: 10.1016/j.nlp.2024.100068
Jabed Omor Bappi , Mohammad Abu Tareq Rony , Mohammad Shariful Islam
{"title":"BNVGLENET: Hypercomplex Bangla handwriting character recognition with hierarchical class expansion using Convolutional Neural Networks","authors":"Jabed Omor Bappi ,&nbsp;Mohammad Abu Tareq Rony ,&nbsp;Mohammad Shariful Islam","doi":"10.1016/j.nlp.2024.100068","DOIUrl":"https://doi.org/10.1016/j.nlp.2024.100068","url":null,"abstract":"<div><p>Object recognition technology has made significant strides where recognizing handwritten Bangla characters including symbols, compounds form, etc. remains a challenging problem due to the prevalence of cursive writing and many ambiguous characters. The complexity and variability of the Bangla script, and individual’s unique handwriting styles make it difficult to achieve satisfactory performance for practical applications, and the best existing recognizers are far less effective than those developed for English alpha-numeric characters. In comparison to other major languages, there are limited options for recognizing handwritten Bangla characters. This study has the potential to improve the accuracy and effectiveness of handwriting recognition systems for the Bengali language, which is spoken by over 200 million people worldwide. This paper aims to investigate the application of Convolutional Neural Networks (CNNs) for recognizing Bangla handwritten characters, with a particular focus on enlarging the recognized character classes. To achieve this, a novel challenging dataset for handwriting recognition is introduced, which is collected from numerous students’ handwriting from two institutions. A novel convolutional neural network-based approach called BNVGLENET is proposed in this paper to recognize Bangla handwritten characters by modifying the LeNet-5 and combining it with the VGG architecture, which has the advantage of significantly identifying the characters from Bengali handwriting. This study systematically evaluated the performance of models not only on custom novel dataset but also on the publicly available Bangla handwritten character dataset called the Grapheme dataset. This research achieved a state-of-the-art recognition accuracy of 98.2% on our custom testing vowel-consonant class and 97.5% on the custom individual class. The improvements achieved in this study bridge a notable disparity between the practical needs and the actual performance of Bangla handwritten character recognition systems.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"7 ","pages":"Article 100068"},"PeriodicalIF":0.0,"publicationDate":"2024-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000165/pdfft?md5=8bf76ee7108a74bfc05ded3f15c3a43e&pid=1-s2.0-S2949719124000165-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140557853","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advancing NLP models with strategic text augmentation: A comprehensive study of augmentation methods and curriculum strategies 通过战略性文本扩增推进 NLP 模型:对扩增方法和课程策略的综合研究
Natural Language Processing Journal Pub Date : 2024-04-13 DOI: 10.1016/j.nlp.2024.100071
Himmet Toprak Kesgin, Mehmet Fatih Amasyali
{"title":"Advancing NLP models with strategic text augmentation: A comprehensive study of augmentation methods and curriculum strategies","authors":"Himmet Toprak Kesgin,&nbsp;Mehmet Fatih Amasyali","doi":"10.1016/j.nlp.2024.100071","DOIUrl":"https://doi.org/10.1016/j.nlp.2024.100071","url":null,"abstract":"<div><p>This study conducts a thorough evaluation of text augmentation techniques across a variety of datasets and natural language processing (NLP) tasks to address the lack of reliable, generalized evidence for these methods. It examines the effectiveness of these techniques in augmenting training sets to improve performance in tasks such as topic classification, sentiment analysis, and offensive language detection. The research emphasizes not only the augmentation methods, but also the strategic order in which real and augmented instances are introduced during training. A major contribution is the development and evaluation of Modified Cyclical Curriculum Learning (MCCL) for augmented datasets, which represents a novel approach in the field. Results show that specific augmentation methods, especially when integrated with MCCL, significantly outperform traditional training approaches in NLP model performance. These results underscore the need for careful selection of augmentation techniques and sequencing strategies to optimize the balance between speed and quality improvement in various NLP tasks. The study concludes that the use of augmentation methods, especially in conjunction with MCCL, leads to improved results in various classification tasks, providing a foundation for future advances in text augmentation strategies in NLP.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"7 ","pages":"Article 100071"},"PeriodicalIF":0.0,"publicationDate":"2024-04-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000190/pdfft?md5=841354620e15317d1fd328df74581e7d&pid=1-s2.0-S2949719124000190-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140551700","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A survey of text summarization: Techniques, evaluation and challenges 文本摘要调查:技术、评估和挑战
Natural Language Processing Journal Pub Date : 2024-04-04 DOI: 10.1016/j.nlp.2024.100070
Supriyono , Aji Prasetya Wibawa , Suyono , Fachrul Kurniawan
{"title":"A survey of text summarization: Techniques, evaluation and challenges","authors":"Supriyono ,&nbsp;Aji Prasetya Wibawa ,&nbsp;Suyono ,&nbsp;Fachrul Kurniawan","doi":"10.1016/j.nlp.2024.100070","DOIUrl":"https://doi.org/10.1016/j.nlp.2024.100070","url":null,"abstract":"<div><p>This paper explores the complex field of text summarization in Natural Language Processing (NLP), with particular attention to the development and importance of semantic understanding. Text summarization is a crucial component of natural language processing (NLP), which helps to translate large amounts of textual data into clear and understandable representations. As the story progresses, it demonstrates the dynamic transition from simple syntactic structures to sophisticated models with semantic comprehension. In order to effectively summarize, syntactic, semantic, and pragmatic concerns become crucial, highlighting the necessity of capturing not only grammar but also the context and underlying meaning. It examines the wide range of summarization models, from conventional extractive techniques to state-of-the-art tools like pre-trained models. Applications are found in many different fields, demonstrating how versatile summarizing techniques are. Semantic drift and domain-specific knowledge remain obstacles, despite progress. In the future, the study predicts developments like artificial intelligence integration and transfer learning, which motivates academics to investigate these prospects for advancement. The approach, which is based on the PRISMA framework, emphasizes a methodical and open literature review. The work attempts to further natural language processing (NLP) and text summarization by combining various research findings and suggesting future research directions in this dynamic subject.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"7 ","pages":"Article 100070"},"PeriodicalIF":0.0,"publicationDate":"2024-04-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000189/pdfft?md5=59f885a43c999d64a8b2382f368be608&pid=1-s2.0-S2949719124000189-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140542600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Utilization of generative AI for the characterization and identification of visual unknowns 利用生成式人工智能表征和识别视觉未知因素
Natural Language Processing Journal Pub Date : 2024-03-25 DOI: 10.1016/j.nlp.2024.100064
Kara Combs , Trevor J. Bihl , Subhashini Ganapathy
{"title":"Utilization of generative AI for the characterization and identification of visual unknowns","authors":"Kara Combs ,&nbsp;Trevor J. Bihl ,&nbsp;Subhashini Ganapathy","doi":"10.1016/j.nlp.2024.100064","DOIUrl":"https://doi.org/10.1016/j.nlp.2024.100064","url":null,"abstract":"<div><p>Current state-of-the-art artificial intelligence (AI) struggles with accurate interpretation of out-of-library objects. One method proposed remedy is analogical reasoning (AR), which utilizes abductive reasoning to draw inferences on an unfamiliar scenario given knowledge about a similar familiar scenario. Currently, applications of visual AR gravitate toward analogy-formatted image problems rather than real-world computer vision data sets. This paper proposes the Image Recognition Through Analogical Reasoning Algorithm (IRTARA) and its “generative AI” version called “GIRTARA” which describes and predicts out-of-library visual objects. IRTARA characterizes the out-of-library object through a list of words called the “term frequency list”. GIRTARA uses the term frequency list to predict what the out-of-library object is. To evaluate the quality of the results of IRTARA, both quantitative and qualitative assessments are used, including a baseline to compare the automated methods with human-generated results. The accuracy of GIRTARA’s predictions is calculated through a cosine similarity analysis. This study observed that IRTARA had consistent results in the term frequency list based on the three evaluation methods for the high-quality results and GIRTARA was able to obtain up to 65% match in terms of cosine similarity when compared to the out-of-library object’s true labels.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"7 ","pages":"Article 100064"},"PeriodicalIF":0.0,"publicationDate":"2024-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000128/pdfft?md5=b907bb3498bdf74554a25eef96b3ee34&pid=1-s2.0-S2949719124000128-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140343887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Claim detection for automated fact-checking: A survey on monolingual, multilingual and cross-lingual research 自动事实核查的索赔检测:单语言、多语言和跨语言研究调查
Natural Language Processing Journal Pub Date : 2024-03-25 DOI: 10.1016/j.nlp.2024.100066
Rrubaa Panchendrarajan, Arkaitz Zubiaga
{"title":"Claim detection for automated fact-checking: A survey on monolingual, multilingual and cross-lingual research","authors":"Rrubaa Panchendrarajan,&nbsp;Arkaitz Zubiaga","doi":"10.1016/j.nlp.2024.100066","DOIUrl":"https://doi.org/10.1016/j.nlp.2024.100066","url":null,"abstract":"<div><p>Automated fact-checking has drawn considerable attention over the past few decades due to the increase in the diffusion of misinformation on online platforms. This is often carried out as a sequence of tasks comprising (i) the detection of sentences circulating in online platforms which constitute claims needing verification, followed by (ii) the verification process of those claims. This survey focuses on the former, by discussing existing efforts towards detecting claims needing fact-checking, with a particular focus on multilingual data and methods. This is a challenging and fertile direction where existing methods are yet far from matching human performance due to the profoundly challenging nature of the issue. Especially, the dissemination of information across multiple social platforms, articulated in multiple languages and modalities demands more generalized solutions for combating misinformation. Focusing on multilingual misinformation, we present a comprehensive survey of existing multilingual claim detection research. We present state-of-the-art multilingual claim detection research categorized into three key factors of the problem, verifiability, priority, and similarity. Further, we present a detailed overview of the existing multilingual datasets along with the challenges and suggest possible future advancements.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"7 ","pages":"Article 100066"},"PeriodicalIF":0.0,"publicationDate":"2024-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000141/pdfft?md5=43cfb5b770cda4c03e5933e454d8f5bd&pid=1-s2.0-S2949719124000141-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140290305","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ensemble learning with soft-prompted pretrained language models for fact checking 利用软提示预训练语言模型进行事实核查的集合学习
Natural Language Processing Journal Pub Date : 2024-03-21 DOI: 10.1016/j.nlp.2024.100067
Shaoqin Huang , Yue Wang , Eugene Y.C. Wong , Lei Yu
{"title":"Ensemble learning with soft-prompted pretrained language models for fact checking","authors":"Shaoqin Huang ,&nbsp;Yue Wang ,&nbsp;Eugene Y.C. Wong ,&nbsp;Lei Yu","doi":"10.1016/j.nlp.2024.100067","DOIUrl":"10.1016/j.nlp.2024.100067","url":null,"abstract":"<div><p>The infectious diseases, such as COVID-19 pandemic, has led to a surge of information on the internet, including misinformation, necessitating fact-checking tools. However, fact-checking infectious diseases related claims pose challenges due to informal claims versus formal evidence and the presence of multiple aspects in a claim. To address these issues, we propose a soft prompt-based ensemble learning framework for COVID-19 fact checking. To understand complex assertions in informal social media texts, we explore various soft prompt structures to take advantage of the T5 language model, and ensemble these prompt structures together. Soft prompts offer flexibility and better generalization compared to hard prompts. The ensemble model captures linguistic cues and contextual information in COVID-19-related data, and thus enhances generalization to new claims. Experimental results demonstrate that prompt-based ensemble learning improves fact-checking accuracy and provides a promising approach to combat misinformation during the pandemic. In addition, the method also shows great zero-shot learning capability and thus can be applied to various fact checking problems.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"7 ","pages":"Article 100067"},"PeriodicalIF":0.0,"publicationDate":"2024-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000153/pdfft?md5=268e2b44eb63a0ef7ca15c1fd64330b7&pid=1-s2.0-S2949719124000153-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140269139","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LeanContext: Cost-efficient domain-specific question answering using LLMs LeanContext:使用 LLM 进行具有成本效益的特定领域问题解答
Natural Language Processing Journal Pub Date : 2024-03-18 DOI: 10.1016/j.nlp.2024.100065
Md Adnan Arefeen , Biplob Debnath , Srimat Chakradhar
{"title":"LeanContext: Cost-efficient domain-specific question answering using LLMs","authors":"Md Adnan Arefeen ,&nbsp;Biplob Debnath ,&nbsp;Srimat Chakradhar","doi":"10.1016/j.nlp.2024.100065","DOIUrl":"https://doi.org/10.1016/j.nlp.2024.100065","url":null,"abstract":"<div><p>Question-answering (QA) is a significant application of Large Language Models (LLMs), shaping chatbot capabilities across healthcare, education, and customer service. However, widespread LLM integration presents a challenge for small businesses due to the high expenses of LLM API usage. Costs rise rapidly when domain-specific data (context) is used alongside queries for accurate domain-specific LLM responses. Extracting context from domain-specific data is implemented by a Retrieval Augmented Generation (RAG) approach. One option is to summarize the RAG context by using LLMs and reduce the context. However, this can also filter out useful information that is necessary to answer some domain-specific queries. In this paper, we shift from human-oriented summarizers to AI model-friendly summaries. Our approach, LeanContext, efficiently extracts <em>k</em> key sentences from the context that are closely aligned with the query. The choice of <em>k</em> is neither static nor random; we introduce a reinforcement learning technique that dynamically determines <em>k</em> based on the query and context. The rest of the less important sentences are either reduced using a free open-source text reduction method or eliminated. We evaluate LeanContext against several recent query-aware and query-unaware context reduction approaches on prominent datasets (arxiv papers and BBC news articles, NarrativeQA). Despite cost reductions of 37.29% to 67.81%, LeanContext’s ROUGE-1 score decreases only by 1.41% to 2.65% compared to a baseline that retains the entire context (no summarization). LeanContext stands out for its ability to provide precise responses, outperforming competitors by leveraging open-source summarization techniques. Human evaluations of the responses further confirm and validate this superiority. Additionally, if open-source pre-trained LLM-based summarizers are used to reduce context (into human consumable summaries), LeanContext can further modify the reduced context to enhance the accuracy (ROUGE-1 score) by 13.22% to 24.61%.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"7 ","pages":"Article 100065"},"PeriodicalIF":0.0,"publicationDate":"2024-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S294971912400013X/pdfft?md5=635c034287e104fec6128cc735fdc367&pid=1-s2.0-S294971912400013X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140180775","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Understanding latent affective bias in large pre-trained neural language models 了解大型预训练神经语言模型中的潜在情感偏差
Natural Language Processing Journal Pub Date : 2024-03-05 DOI: 10.1016/j.nlp.2024.100062
Anoop Kadan , Deepak P. , Sahely Bhadra , Manjary P. Gangan , Lajish V.L.
{"title":"Understanding latent affective bias in large pre-trained neural language models","authors":"Anoop Kadan ,&nbsp;Deepak P. ,&nbsp;Sahely Bhadra ,&nbsp;Manjary P. Gangan ,&nbsp;Lajish V.L.","doi":"10.1016/j.nlp.2024.100062","DOIUrl":"https://doi.org/10.1016/j.nlp.2024.100062","url":null,"abstract":"<div><p>Groundbreaking inventions and highly significant performance improvements in deep learning based Natural Language Processing are witnessed through the development of transformer based large Pre-trained Language Models (PLMs). The wide availability of unlabeled data within human generated data deluge along with self-supervised learning strategy helps to accelerate the success of large PLMs in language generation, language understanding, etc. But at the same time, latent historical bias/unfairness in human minds towards a particular gender, race, etc., encoded unintentionally/intentionally into the corpora harms and questions the utility and efficacy of large PLMs in many real-world applications, particularly for the protected groups. In this paper, we present an extensive investigation towards understanding the existence of <em>“Affective Bias”</em> in large PLMs to unveil any biased association of emotions such as <em>anger</em>, <em>fear</em>, <em>joy</em>, etc., towards a particular gender, race or religion with respect to the downstream task of textual emotion detection. We conduct our exploration of affective bias from the very initial stage of corpus level affective bias analysis by searching for imbalanced distribution of affective words within a domain, in large scale corpora that are used to pre-train and fine-tune PLMs. Later, to quantify affective bias in model predictions, we perform an extensive set of class-based and intensity-based evaluations using various bias evaluation corpora. Our results show the existence of statistically significant affective bias in the PLM based emotion detection systems, indicating biased association of certain emotions towards a particular gender, race, and religion.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"7 ","pages":"Article 100062"},"PeriodicalIF":0.0,"publicationDate":"2024-03-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000104/pdfft?md5=47ed3491ca02f42caa81ecff613ee5f3&pid=1-s2.0-S2949719124000104-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140069620","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Recent advancements and challenges of NLP-based sentiment analysis: A state-of-the-art review 基于 NLP 的情感分析的最新进展和挑战:最新进展综述
Natural Language Processing Journal Pub Date : 2024-03-01 DOI: 10.1016/j.nlp.2024.100059
Jamin Rahman Jim , Md Apon Riaz Talukder , Partha Malakar , Md Mohsin Kabir , Kamruddin Nur , M.F. Mridha
{"title":"Recent advancements and challenges of NLP-based sentiment analysis: A state-of-the-art review","authors":"Jamin Rahman Jim ,&nbsp;Md Apon Riaz Talukder ,&nbsp;Partha Malakar ,&nbsp;Md Mohsin Kabir ,&nbsp;Kamruddin Nur ,&nbsp;M.F. Mridha","doi":"10.1016/j.nlp.2024.100059","DOIUrl":"https://doi.org/10.1016/j.nlp.2024.100059","url":null,"abstract":"<div><p>Sentiment analysis is a method within natural language processing that evaluates and identifies the emotional tone or mood conveyed in textual data. Scrutinizing words and phrases categorizes them into positive, negative, or neutral sentiments. The significance of sentiment analysis lies in its capacity to derive valuable insights from extensive textual data, empowering businesses to grasp customer sentiments, make informed choices, and enhance their offerings. For the further advancement of sentiment analysis, gaining a deep understanding of its algorithms, applications, current performance, and challenges is imperative. Therefore, in this extensive survey, we began exploring the vast array of application domains for sentiment analysis, scrutinizing them within the context of existing research. We then delved into prevalent pre-processing techniques, datasets, and evaluation metrics to enhance comprehension. We also explored Machine Learning, Deep Learning, Large Language Models and Pre-trained models in sentiment analysis, providing insights into their advantages and drawbacks. Subsequently, we precisely reviewed the experimental results and limitations of recent state-of-the-art articles. Finally, we discussed the diverse challenges encountered in sentiment analysis and proposed future research directions to mitigate these concerns. This extensive review provides a complete understanding of sentiment analysis, covering its models, application domains, results analysis, challenges, and research directions.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"6 ","pages":"Article 100059"},"PeriodicalIF":0.0,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000074/pdfft?md5=f2c0dd3a1ae1a2992d955f19909d86a5&pid=1-s2.0-S2949719124000074-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139998912","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Transformer-based text similarity and second language proficiency: A case of written production by learners of Korean 基于转换器的文本相似性与第二语言能力:韩语学习者的书面表达案例
Natural Language Processing Journal Pub Date : 2024-03-01 DOI: 10.1016/j.nlp.2024.100060
Gyu-Ho Shin , Boo Kyung Jung , Seongmin Mun
{"title":"Transformer-based text similarity and second language proficiency: A case of written production by learners of Korean","authors":"Gyu-Ho Shin ,&nbsp;Boo Kyung Jung ,&nbsp;Seongmin Mun","doi":"10.1016/j.nlp.2024.100060","DOIUrl":"https://doi.org/10.1016/j.nlp.2024.100060","url":null,"abstract":"<div><p>The present study applies two transformer models (BERT; GPT-2) to analyse argumentative essays produced by two first-language groups (Czech; English) of second-language learners of Korean and investigates how informative similarity scores of learner writing obtained by these models explain general language proficiency in Korean. Results show three major aspects on model performance. First, the relationships between the similarity scores and the proficiency scores differ from the tendencies between the human rating scores and the proficiency scores. Second, the degree to which the similarity scores obtained by each model explain the proficiency scores is asymmetric and idiosyncratic. Third, the performance of the two models is affected by learners’ native language and essay topic. These findings invite the need for researchers and educators to pay attention to how computational algorithms operate, together with learner language characteristics and language-specific properties of the target language, in utilising Natural Language Processing methods and techniques for their research or instructional purposes.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"6 ","pages":"Article 100060"},"PeriodicalIF":0.0,"publicationDate":"2024-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000086/pdfft?md5=c5357abe0301e49c473990485a85a9a2&pid=1-s2.0-S2949719124000086-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"139998913","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信