News Text Classification Method for Edge Computing Based on BiLSTM-Attention

Zhixun Liang, Peng Chen, Yunfei Yi, Yuanyuan Fan
{"title":"News Text Classification Method for Edge Computing Based on BiLSTM-Attention","authors":"Zhixun Liang, Peng Chen, Yunfei Yi, Yuanyuan Fan","doi":"10.1109/ACAIT56212.2022.10137822","DOIUrl":null,"url":null,"abstract":"In the coming smart city, the explosive growth of data makes the amount of data contained in news texts more and more, which leads to the decrease in the accuracy of traditional machine learning or deep learning models in the news text classification. Therefore, in this paper, we propose a news text classification model based on BiLSTM-Attention. The data set is selected as 30,000 news texts, and the word segmentation is carried out in turn. The stop words are removed, and the word vector is quantified. Then, the data set is cross-validated according to the ratio of training set to validation set of 8:1. Finally, the experiments with the bilstm model, lstm model and bilstm-short text model show that the BiLSTM-Attention model has the highest accuracy and the lowest loss value. In order to further verify the classification performance of BiLSTMAttention model, the experiment is designed again and Bayes and SVM are added to compare. The experimental results show that the accuracy, recall and F1 value of BiLSTM-Attention model are the highest, which proves that BiLSTM-Attention is more suitable for news text classification.","PeriodicalId":398228,"journal":{"name":"2022 6th Asian Conference on Artificial Intelligence Technology (ACAIT)","volume":"64 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 6th Asian Conference on Artificial Intelligence Technology (ACAIT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ACAIT56212.2022.10137822","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In the coming smart city, the explosive growth of data makes the amount of data contained in news texts more and more, which leads to the decrease in the accuracy of traditional machine learning or deep learning models in the news text classification. Therefore, in this paper, we propose a news text classification model based on BiLSTM-Attention. The data set is selected as 30,000 news texts, and the word segmentation is carried out in turn. The stop words are removed, and the word vector is quantified. Then, the data set is cross-validated according to the ratio of training set to validation set of 8:1. Finally, the experiments with the bilstm model, lstm model and bilstm-short text model show that the BiLSTM-Attention model has the highest accuracy and the lowest loss value. In order to further verify the classification performance of BiLSTMAttention model, the experiment is designed again and Bayes and SVM are added to compare. The experimental results show that the accuracy, recall and F1 value of BiLSTM-Attention model are the highest, which proves that BiLSTM-Attention is more suitable for news text classification.
基于bilstm -注意力的边缘计算新闻文本分类方法
在即将到来的智慧城市中,数据的爆炸式增长使得新闻文本中包含的数据量越来越大,这导致传统的机器学习或深度学习模型在新闻文本分类中的准确率下降。因此,本文提出了一种基于BiLSTM-Attention的新闻文本分类模型。选取数据集为3万条新闻文本,依次进行分词。去除停止词,量化词向量。然后,按照训练集与验证集的比例为8:1对数据集进行交叉验证。最后,通过bilstm模型、lstm模型和bilstm-短文本模型的实验表明,bilstm-注意力模型具有最高的准确率和最低的损失值。为了进一步验证bilstattention模型的分类性能,再次设计实验,加入贝叶斯和支持向量机进行比较。实验结果表明,BiLSTM-Attention模型的准确率、查全率和F1值最高,证明了BiLSTM-Attention模型更适合新闻文本分类。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信