{"title":"Revolutionizing Healthcare: NLP, Deep Learning, and WSN Solutions for Managing the COVID-19 Crisis","authors":"Ajay P., Nagaraj B., R. Arun Kumar","doi":"10.1145/3639566","DOIUrl":null,"url":null,"abstract":"<p>The COVID-19 outbreak in 2020 catalyzed a global socio-economic upheaval, compelling nations to embrace digital technologies as a means of countering economic downturns and ensuring efficient communication systems. This paper delves into the role of Natural Language Processing (NLP) in harnessing wireless connectivity during the pandemic. The examination assesses how wireless networks have affected various facets of crisis management, including virus tracking, optimizing healthcare, facilitating remote education, and enabling unified communications. Additionally, the article underscores the importance of digital inclusion in mitigating disease outbreaks and reconnecting marginalized communities. To address these challenges, a Dual CNN-based BERT model is proposed. BERT model is used to extract the text features, the internal layers of BERT excel at capturing intricate contextual details concerning words and phrases, rendering them highly valuable as features for a wide array of text analysis tasks. The significance of dual CNN is capturing the unique capability to seamlessly integrate both character-level and word-level information. This fusion of insights from different levels of textual analysis proves especially valuable in handling text data that is noisy, complex, or presents challenges related to misspellings and domain-specific terminology. The proposed model is evaluated using the simulated WSN-based text data for crisis management.</p>","PeriodicalId":54312,"journal":{"name":"ACM Transactions on Asian and Low-Resource Language Information Processing","volume":"54 1","pages":""},"PeriodicalIF":1.8000,"publicationDate":"2024-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ACM Transactions on Asian and Low-Resource Language Information Processing","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1145/3639566","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
The COVID-19 outbreak in 2020 catalyzed a global socio-economic upheaval, compelling nations to embrace digital technologies as a means of countering economic downturns and ensuring efficient communication systems. This paper delves into the role of Natural Language Processing (NLP) in harnessing wireless connectivity during the pandemic. The examination assesses how wireless networks have affected various facets of crisis management, including virus tracking, optimizing healthcare, facilitating remote education, and enabling unified communications. Additionally, the article underscores the importance of digital inclusion in mitigating disease outbreaks and reconnecting marginalized communities. To address these challenges, a Dual CNN-based BERT model is proposed. BERT model is used to extract the text features, the internal layers of BERT excel at capturing intricate contextual details concerning words and phrases, rendering them highly valuable as features for a wide array of text analysis tasks. The significance of dual CNN is capturing the unique capability to seamlessly integrate both character-level and word-level information. This fusion of insights from different levels of textual analysis proves especially valuable in handling text data that is noisy, complex, or presents challenges related to misspellings and domain-specific terminology. The proposed model is evaluated using the simulated WSN-based text data for crisis management.
期刊介绍:
The ACM Transactions on Asian and Low-Resource Language Information Processing (TALLIP) publishes high quality original archival papers and technical notes in the areas of computation and processing of information in Asian languages, low-resource languages of Africa, Australasia, Oceania and the Americas, as well as related disciplines. The subject areas covered by TALLIP include, but are not limited to:
-Computational Linguistics: including computational phonology, computational morphology, computational syntax (e.g. parsing), computational semantics, computational pragmatics, etc.
-Linguistic Resources: including computational lexicography, terminology, electronic dictionaries, cross-lingual dictionaries, electronic thesauri, etc.
-Hardware and software algorithms and tools for Asian or low-resource language processing, e.g., handwritten character recognition.
-Information Understanding: including text understanding, speech understanding, character recognition, discourse processing, dialogue systems, etc.
-Machine Translation involving Asian or low-resource languages.
-Information Retrieval: including natural language processing (NLP) for concept-based indexing, natural language query interfaces, semantic relevance judgments, etc.
-Information Extraction and Filtering: including automatic abstraction, user profiling, etc.
-Speech processing: including text-to-speech synthesis and automatic speech recognition.
-Multimedia Asian Information Processing: including speech, image, video, image/text translation, etc.
-Cross-lingual information processing involving Asian or low-resource languages.
-Papers that deal in theory, systems design, evaluation and applications in the aforesaid subjects are appropriate for TALLIP. Emphasis will be placed on the originality and the practical significance of the reported research.