Natural Language Processing Journal最新文献

筛选
英文 中文
TransLSTM: A hybrid LSTM-Transformer model for fine-grained suggestion mining TransLSTM:用于细粒度建议挖掘的 LSTM 和 Transformer 混合模型
Natural Language Processing Journal Pub Date : 2024-07-14 DOI: 10.1016/j.nlp.2024.100089
Samad Riaz , Amna Saghir , Muhammad Junaid Khan , Hassan Khan , Hamid Saeed Khan , M. Jaleed Khan
{"title":"TransLSTM: A hybrid LSTM-Transformer model for fine-grained suggestion mining","authors":"Samad Riaz ,&nbsp;Amna Saghir ,&nbsp;Muhammad Junaid Khan ,&nbsp;Hassan Khan ,&nbsp;Hamid Saeed Khan ,&nbsp;M. Jaleed Khan","doi":"10.1016/j.nlp.2024.100089","DOIUrl":"10.1016/j.nlp.2024.100089","url":null,"abstract":"<div><p>Digital platforms on the internet are invaluable for collecting user feedback, suggestions, and opinions about various topics, such as company products and services. This data is instrumental in shaping business strategies, enhancing product development, and refining service delivery. Suggestion mining is a key task in natural language processing, which focuses on extracting and analysing suggestions from these digital sources. Initially, suggestion mining utilized manually crafted features, but recent advancements have highlighted the efficacy of deep learning models, which automatically learn features. Models like Convolutional Neural Networks (CNN), Recurrent Neural Networks (RNN), Long Short-Term Memory (LSTM), and Bidirectional Encoder Representations from Transformers (BERT) have been employed in this field. However, considering the relatively small datasets and the faster training time of LSTM compared to BERT, we introduce TransLSTM, a novel LSTM-Transformer hybrid model for suggestion mining. This model aims to automatically pinpoint and extract suggestions by harnessing both local and global text dependencies. It combines the sequential dependency handling of LSTM with the contextual interaction capabilities of the Transformer, thus effectively identifying and extracting suggestions. We evaluated our method against state-of-the-art approaches using the SemEval Task-9 dataset, a benchmark for suggestion mining. Our model shows promising performance, surpassing existing deep learning methods by 6.76% with an F1 score of 0.834 for SubTask A and 0.881 for SubTask B. Additionally, our paper presents an exhaustive literature review on suggestion mining from digital platforms, covering both traditional and state-of-the-art text classification techniques.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"8 ","pages":"Article 100089"},"PeriodicalIF":0.0,"publicationDate":"2024-07-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000372/pdfft?md5=01d5468c4cb646548ed9ac72a0da2eb9&pid=1-s2.0-S2949719124000372-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141706671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A comprehensive survey on answer generation methods using NLP 使用 NLP 生成答案方法的综合调查
Natural Language Processing Journal Pub Date : 2024-07-02 DOI: 10.1016/j.nlp.2024.100088
Prashant Upadhyay, Rishabh Agarwal, Sumeet Dhiman, Abhinav Sarkar, Saumya Chaturvedi
{"title":"A comprehensive survey on answer generation methods using NLP","authors":"Prashant Upadhyay,&nbsp;Rishabh Agarwal,&nbsp;Sumeet Dhiman,&nbsp;Abhinav Sarkar,&nbsp;Saumya Chaturvedi","doi":"10.1016/j.nlp.2024.100088","DOIUrl":"10.1016/j.nlp.2024.100088","url":null,"abstract":"<div><p>Recent advancements in question-answering systems have significantly enhanced the capability of computers to understand and respond to queries in natural language. This paper presents a comprehensive review of the evolution of question answering systems, with a focus on the developments over the last few years. We examine the foundational aspects of a question answering framework, including question analysis, answer extraction, and passage retrieval. Additionally, we delve into the challenges that question answering systems encounter, such as the intricacies of question processing, the necessity of contextual data sources, and the complexities involved in real-time question answering. Our study categorizes existing question answering systems based on the types of questions they address, the nature of the answers they produce, and the various approaches employed to generate these answers. We also explore the distinctions between opinion-based, extraction-based, retrieval-based, and generative answer generation. The classification provides insight into the strengths and limitations of each method, paving the way for future innovations in the field. This review aims to offer a clear understanding of the current state of question answering systems and to identify the scaling needed to meet the rising expectations and demands of users for coherent and accurate automated responses in natural language.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"8 ","pages":"Article 100088"},"PeriodicalIF":0.0,"publicationDate":"2024-07-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000360/pdfft?md5=57245c441a09df1168241bb40a6f9e06&pid=1-s2.0-S2949719124000360-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141623096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Topic-aware response selection for dialog systems 对话系统的主题感知响应选择
Natural Language Processing Journal Pub Date : 2024-06-24 DOI: 10.1016/j.nlp.2024.100087
Wei Yuan , Zongyang Ma , Aijun An , Jimmy Xiangji Huang
{"title":"Topic-aware response selection for dialog systems","authors":"Wei Yuan ,&nbsp;Zongyang Ma ,&nbsp;Aijun An ,&nbsp;Jimmy Xiangji Huang","doi":"10.1016/j.nlp.2024.100087","DOIUrl":"https://doi.org/10.1016/j.nlp.2024.100087","url":null,"abstract":"<div><p>It is challenging for a persona-based chitchat system to return responses consistent with the dialog context and the persona of the agent. This particularly holds for a retrieval-based chitchat system that selects the most appropriate response from a set of candidates according to the dialog context and the persona of the agent. A persona usually has some dominant topics (e.g., <em>sports</em>, <em>music</em>). Adhering to these topics can enhance the consistency of responses. However, previous studies rarely explore the topical semantics of the agent’s persona in the chitchat system, which often fails to return responses coherent with the persona. In this paper, we propose a Topic-Aware Response Selection (TARS) model, capturing multi-grained matching between the dialog context and a response and also between the persona and a response at both the word and the topic levels, to select the appropriate topic-aware response from the pool of response candidates. Empirical results on the public persona-based empathetic conversation (PEC) data demonstrate the promising performance of the TARS model for response selection.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"8 ","pages":"Article 100087"},"PeriodicalIF":0.0,"publicationDate":"2024-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000359/pdfft?md5=460e17e8ab71eeba6fb71be3795c94c0&pid=1-s2.0-S2949719124000359-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141541701","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A modified Vector Space Model for semantic information retrieval 用于语义信息检索的改进向量空间模型
Natural Language Processing Journal Pub Date : 2024-06-13 DOI: 10.1016/j.nlp.2024.100081
Callistus Ireneous Nakpih
{"title":"A modified Vector Space Model for semantic information retrieval","authors":"Callistus Ireneous Nakpih","doi":"10.1016/j.nlp.2024.100081","DOIUrl":"10.1016/j.nlp.2024.100081","url":null,"abstract":"<div><p>In this research, we present a modified Vector Space Model which focuses on the semantic relevance of words for retrieving documents. The modified VSM resolves the problem of the classical model performing only lexical matching of query terms to document terms for retrievals. This problem also restricts the classical model from retrieving documents that do not have exact match of query terms even if they are semantically relevant to the query. In the modified model, we introduced a Query Relevance Update technique, which pads the original query set with semantically relevant document terms for optimised semantic retrieval results. The modified model also includes a novel <span><math><mrow><mi>t</mi><mi>f</mi><mo>−</mo><mi>p</mi></mrow></math></span> which replaces the <span><math><mrow><mi>t</mi><mi>f</mi><mo>−</mo><mi>i</mi><mi>d</mi><mi>f</mi></mrow></math></span> technique of the classical VSM, which is used to compute the Term Frequency weights. The replacement of the <span><math><mrow><mi>t</mi><mi>f</mi><mo>−</mo><mi>i</mi><mi>d</mi><mi>f</mi></mrow></math></span> resolves the problem of the classical model penalising terms that occur across documents with the assumption that they are stop words, which in practice, there are usually such words which carry relevant semantic information for documents’ retrieval. We also extended the cosine similarity function with a proportionality weight <span><math><msub><mrow><mi>p</mi></mrow><mrow><mi>q</mi><mi>d</mi></mrow></msub></math></span>, which moderates biases for high frequency of terms in longer documents. The <span><math><msub><mrow><mi>p</mi></mrow><mrow><mi>q</mi><mi>d</mi></mrow></msub></math></span> ensures that the frequency of query terms including the updated ones are accounted for in proportionality with documents size for the overall ranking of documents. The simulated results reveal that, the modified VSM does achieve semantic retrieval of documents beyond lexical matching of query and document terms.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"8 ","pages":"Article 100081"},"PeriodicalIF":0.0,"publicationDate":"2024-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000293/pdfft?md5=3a5de846966e83dc34ea6a2e3b7d202f&pid=1-s2.0-S2949719124000293-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141405572","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cutting through the noise to motivate people: A comprehensive analysis of COVID-19 social media posts de/motivating vaccination 穿透噪音,激发人们的积极性:全面分析 COVID-19 社交媒体上关于疫苗接种的帖子
Natural Language Processing Journal Pub Date : 2024-06-13 DOI: 10.1016/j.nlp.2024.100085
Ashiqur Rahman , Ehsan Mohammadi , Hamed Alhoori
{"title":"Cutting through the noise to motivate people: A comprehensive analysis of COVID-19 social media posts de/motivating vaccination","authors":"Ashiqur Rahman ,&nbsp;Ehsan Mohammadi ,&nbsp;Hamed Alhoori","doi":"10.1016/j.nlp.2024.100085","DOIUrl":"10.1016/j.nlp.2024.100085","url":null,"abstract":"<div><p>The COVID-19 pandemic exposed significant weaknesses in the healthcare information system. The overwhelming volume of misinformation on social media and other socioeconomic factors created extraordinary challenges to motivate people to take proper precautions and get vaccinated. In this context, our work explored a novel direction by analyzing an extensive dataset collected over two years, identifying the topics de/motivating the public about COVID-19 vaccination. We analyzed these topics based on time, geographic location, and political orientation. We noticed that while the motivating topics remain the same over time and geographic location, the demotivating topics change rapidly. We also identified that intrinsic motivation, rather than external mandate, is more advantageous to inspire the public. This study addresses scientific communication and public motivation in social media. It can help public health officials, policymakers, and social media platforms develop more effective messaging strategies to cut through the noise of misinformation and educate the public about scientific findings.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"8 ","pages":"Article 100085"},"PeriodicalIF":0.0,"publicationDate":"2024-06-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000335/pdfft?md5=bda60786d7ac110df5894be0ee669f0e&pid=1-s2.0-S2949719124000335-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141392804","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generating dynamic lip-syncing using target audio in a multimedia environment 在多媒体环境中使用目标音频生成动态唇语同步
Natural Language Processing Journal Pub Date : 2024-06-10 DOI: 10.1016/j.nlp.2024.100084
Diksha Pawar, Prashant Borde, Pravin Yannawar
{"title":"Generating dynamic lip-syncing using target audio in a multimedia environment","authors":"Diksha Pawar,&nbsp;Prashant Borde,&nbsp;Pravin Yannawar","doi":"10.1016/j.nlp.2024.100084","DOIUrl":"https://doi.org/10.1016/j.nlp.2024.100084","url":null,"abstract":"<div><p>The presented research focuses on the challenging task of creating lip-sync facial videos that align with a specified target speech segment. A novel deep-learning model has been developed to produce precise synthetic lip movements corresponding to the speech extracted from an audio source. Consequently, there are instances where portions of the visual data may fall out of sync with the updated audio and this challenge is handled through, a novel strategy, leveraging insights from a robust lip-sync discriminator. Additionally, this study introduces fresh criteria and evaluation benchmarks for assessing lip synchronization in unconstrained videos. LipChanger demonstrates improved PSNR values, indicative of enhanced image quality. Furthermore, it exhibits highly accurate lip synthesis, as evidenced by lower LMD values and higher SSIM values. These outcomes suggest that the LipChanger approach holds significant potential for enhancing lip synchronization in talking face videos, resulting in more realistic lip movements. The proposed LipChanger model and its associated evaluation benchmarks show promise and could potentially contribute to advancements in lip-sync technology for unconstrained talking face videos.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"8 ","pages":"Article 100084"},"PeriodicalIF":0.0,"publicationDate":"2024-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000323/pdfft?md5=84516d2e22e4420f113635a3914da66f&pid=1-s2.0-S2949719124000323-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141328741","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluating LLMs on document-based QA: Exact answer selection and numerical extraction using CogTale dataset 评估基于文档的 QA 的 LLM:使用 CogTale 数据集进行精确答案选择和数字提取
Natural Language Processing Journal Pub Date : 2024-06-08 DOI: 10.1016/j.nlp.2024.100083
Zafaryab Rasool , Stefanus Kurniawan , Sherwin Balugo , Scott Barnett , Rajesh Vasa , Courtney Chesser , Benjamin M. Hampstead , Sylvie Belleville , Kon Mouzakis , Alex Bahar-Fuchs
{"title":"Evaluating LLMs on document-based QA: Exact answer selection and numerical extraction using CogTale dataset","authors":"Zafaryab Rasool ,&nbsp;Stefanus Kurniawan ,&nbsp;Sherwin Balugo ,&nbsp;Scott Barnett ,&nbsp;Rajesh Vasa ,&nbsp;Courtney Chesser ,&nbsp;Benjamin M. Hampstead ,&nbsp;Sylvie Belleville ,&nbsp;Kon Mouzakis ,&nbsp;Alex Bahar-Fuchs","doi":"10.1016/j.nlp.2024.100083","DOIUrl":"https://doi.org/10.1016/j.nlp.2024.100083","url":null,"abstract":"<div><p>Document-based Question-Answering (QA) tasks are crucial for precise information retrieval. While some existing work focus on evaluating large language model’s (LLMs) performance on retrieving and answering questions from documents, assessing the LLMs performance on QA types that require exact answer selection from predefined options and numerical extraction is yet to be fully assessed. In this paper, we specifically focus on this underexplored context and conduct empirical analysis of LLMs (GPT-4 and GPT-3.5) on question types, including single-choice, yes–no, multiple-choice, and number extraction questions from documents. We use the CogTale dataset for evaluation, which provide human expert-tagged responses, offering a robust benchmark for precision and factual grounding. We found that LLMs, particularly GPT-4, can precisely answer many single-choice and yes–no questions given relevant context, demonstrating their efficacy in information retrieval tasks. However, their performance diminishes when confronted with multiple-choice and number extraction formats, lowering the overall performance of the models on this task, indicating that these models may not yet be sufficiently reliable for the task. This limits the applications of LLMs on applications demanding precise information extraction and inference from documents, such as meta-analysis tasks. Our work offers a framework for ongoing dataset evaluation, ensuring that LLM applications for information retrieval and document analysis continue to meet evolving standards.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"8 ","pages":"Article 100083"},"PeriodicalIF":0.0,"publicationDate":"2024-06-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000311/pdfft?md5=99895c63882405f8b66929d134da8f31&pid=1-s2.0-S2949719124000311-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141438456","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Topic specificity: A descriptive metric for algorithm selection and finding the right number of topics 主题特异性:用于选择算法和寻找合适主题数量的描述性指标
Natural Language Processing Journal Pub Date : 2024-06-04 DOI: 10.1016/j.nlp.2024.100082
Emil Rijcken , Kalliopi Zervanou , Pablo Mosteiro , Floortje Scheepers , Marco Spruit , Uzay Kaymak
{"title":"Topic specificity: A descriptive metric for algorithm selection and finding the right number of topics","authors":"Emil Rijcken ,&nbsp;Kalliopi Zervanou ,&nbsp;Pablo Mosteiro ,&nbsp;Floortje Scheepers ,&nbsp;Marco Spruit ,&nbsp;Uzay Kaymak","doi":"10.1016/j.nlp.2024.100082","DOIUrl":"10.1016/j.nlp.2024.100082","url":null,"abstract":"<div><p>Topic modeling is a prevalent task for discovering the latent structure of a corpus, identifying a set of topics that represent the underlying themes of the documents. Despite its popularity, issues with its evaluation metric, the coherence score, result in two common challenges: <em>algorithm selection</em> and <em>determining the number of topics</em>. To address these two issues, we propose the <em>topic specificity</em> metric, which captures the relative frequency of topic words in the corpus and is used as a proxy for the specificity of a word. In this work, we formulate the metric firstly. Secondly, we demonstrate that algorithms train topics at different specificity levels. This insight can be used to address algorithm selection as it allows users to distinguish and select algorithms with the desired specificity level. Lastly, we show a strictly positive monotonic correlation between the topic specificity and the number of topics for LDA, FLSA-W, NMF and LSI. This correlation can be used to address the selection of the number of topics, as it allows users to adjust the number of topics to their desired level. Moreover, our descriptive metric provides a new perspective to characterize topic models, allowing them to be understood better.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"8 ","pages":"Article 100082"},"PeriodicalIF":0.0,"publicationDate":"2024-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S294971912400030X/pdfft?md5=af15e6c29d867b39aae58eedf84c6eda&pid=1-s2.0-S294971912400030X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141406979","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Summarizing long scientific documents through hierarchical structure extraction 通过层次结构提取总结长篇科学文献
Natural Language Processing Journal Pub Date : 2024-05-29 DOI: 10.1016/j.nlp.2024.100080
Grishma Sharma , Deepak Sharma , M. Sasikumar
{"title":"Summarizing long scientific documents through hierarchical structure extraction","authors":"Grishma Sharma ,&nbsp;Deepak Sharma ,&nbsp;M. Sasikumar","doi":"10.1016/j.nlp.2024.100080","DOIUrl":"https://doi.org/10.1016/j.nlp.2024.100080","url":null,"abstract":"<div><p>In the realm of academia, staying updated with the latest advancements has become increasingly difficult due to the rapid rise in scientific publications. Text summarization emerges as a solution to this challenge by distilling essential contributions into concise summaries. Despite the structured nature of scientific documents, current summarization techniques often overlook this valuable structural information. Our proposed method addresses this gap through an unsupervised, extractive, user preference-based, and hierarchical iterative graph-based ranking algorithm for summarizing long scientific documents. Unlike existing approaches, our method operates by leveraging the inherent structural information within scientific texts to generate diverse summaries tailored to user preferences. To assess the efficiency of our approach, we conducted evaluations on two distinct long document datasets: ScisummNet and a custom dataset comprising papers from esteemed journals and conferences with human-extracted sentences as gold summaries. The results obtained using automatic evaluation metric Rouge scores as well as human evaluation, demonstrate that our method performs better than other well-known unsupervised algorithms. This emphasizes the need for structural information in text summarization, enabling more effective and customizable solutions.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"8 ","pages":"Article 100080"},"PeriodicalIF":0.0,"publicationDate":"2024-05-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000281/pdfft?md5=7e249fba3a7dd6613770889389366f05&pid=1-s2.0-S2949719124000281-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141291974","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Decoding depression: Analyzing social network insights for depression severity assessment with transformers and explainable AI 解码抑郁症:利用变换器和可解释人工智能分析社交网络洞察,评估抑郁症严重程度
Natural Language Processing Journal Pub Date : 2024-05-13 DOI: 10.1016/j.nlp.2024.100079
Tasnim Ahmed , Shahriar Ivan , Ahnaf Munir , Sabbir Ahmed
{"title":"Decoding depression: Analyzing social network insights for depression severity assessment with transformers and explainable AI","authors":"Tasnim Ahmed ,&nbsp;Shahriar Ivan ,&nbsp;Ahnaf Munir ,&nbsp;Sabbir Ahmed","doi":"10.1016/j.nlp.2024.100079","DOIUrl":"10.1016/j.nlp.2024.100079","url":null,"abstract":"<div><p>Depression is a mental state characterized by recurrent feelings of melancholy, hopelessness, and disinterest in activities, having a significant negative influence on everyday functioning and general well-being. Millions of users express their thoughts and emotions on social media platforms, which can be used as a rich source of data for early detection of depression. In this connection, this work leverages an ensemble of transformer-based architectures for quantifying the severity of depression from social media posts into four categories — non-depressed, mild, moderate, and severe. At first, a diverse range of preprocessing techniques is employed to enhance the quality and relevance of the input. Then, the preprocessed samples are passed through three variants of transformer-based models, namely vanilla BERT, BERTweet, and ALBERT, for generating predictions, which are combined using a weighted soft-voting approach. We conduct a comprehensive explainability analysis to gain deeper insights into the decision-making process, examining both local and global perspectives. Furthermore, to the best of our knowledge, we are the first ones to explore the extent to which a Large Language Model (LLM) like ‘ChatGPT’ can perform this task. Evaluation of the model on the publicly available ‘DEPTWEET’ dataset produces state-of-the-art performance with 13.5% improvement in AUC–ROC score.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"7 ","pages":"Article 100079"},"PeriodicalIF":0.0,"publicationDate":"2024-05-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S294971912400027X/pdfft?md5=5d658d840266d01d808f9f0280aa58df&pid=1-s2.0-S294971912400027X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141047775","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信