An Empirical Study on the Classification of Chinese News Articles by Machine Learning and Deep Learning Techniques

Chuen-Min Huang, Yi-Jun Jiang
{"title":"An Empirical Study on the Classification of Chinese News Articles by Machine Learning and Deep Learning Techniques","authors":"Chuen-Min Huang, Yi-Jun Jiang","doi":"10.1109/ICMLC48188.2019.8949309","DOIUrl":null,"url":null,"abstract":"This study compares Chinese news classification results of machine learning (ML) and deep learning (DL). In processing ML, we chose Support Vector Machine (SVM) and Naive Bayes (NB) to form three models: Word2Vec-SVM, TFIDF-SVM, and TFIDF-NB. Since NB assumes that the words are independent, this is different from the concept of related word distribution in Word2Vec, so the combination with NB is excluded. In processing DL, we adopted Bidirectional Long Short-Term Memory (Bi-LSTM), Long Short-Term Memory (LSTM) and Convolutional Neural Network (CNN), and used Word2Vec for word embedding. Experimental results showed that with proper word preprocessing, the difference of classification accuracy of ML and DL models is actually very small. Although the results show that Bi-LSTM performs the most accurate and has the lowest Loss compared to other DL techniques, its implementation process is the most time consuming. This study affirms the excellent results of CNN, while its Loss is the highest of the DL models. We also found that Word2Vec-SVM was superior to TFIDF-SVM in terms of efficiency, but its accuracy is not as good as expected. To summarize the classification accuracy in Bi-LSTM, LSTM, CNN, Word2vec-SVM, TFIDF-SVM, and NB are 89.3%, 88%, and 87.54%, 85.32%, 87.35%, 86.56%, respectively.","PeriodicalId":221349,"journal":{"name":"2019 International Conference on Machine Learning and Cybernetics (ICMLC)","volume":"74 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 International Conference on Machine Learning and Cybernetics (ICMLC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICMLC48188.2019.8949309","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

This study compares Chinese news classification results of machine learning (ML) and deep learning (DL). In processing ML, we chose Support Vector Machine (SVM) and Naive Bayes (NB) to form three models: Word2Vec-SVM, TFIDF-SVM, and TFIDF-NB. Since NB assumes that the words are independent, this is different from the concept of related word distribution in Word2Vec, so the combination with NB is excluded. In processing DL, we adopted Bidirectional Long Short-Term Memory (Bi-LSTM), Long Short-Term Memory (LSTM) and Convolutional Neural Network (CNN), and used Word2Vec for word embedding. Experimental results showed that with proper word preprocessing, the difference of classification accuracy of ML and DL models is actually very small. Although the results show that Bi-LSTM performs the most accurate and has the lowest Loss compared to other DL techniques, its implementation process is the most time consuming. This study affirms the excellent results of CNN, while its Loss is the highest of the DL models. We also found that Word2Vec-SVM was superior to TFIDF-SVM in terms of efficiency, but its accuracy is not as good as expected. To summarize the classification accuracy in Bi-LSTM, LSTM, CNN, Word2vec-SVM, TFIDF-SVM, and NB are 89.3%, 88%, and 87.54%, 85.32%, 87.35%, 86.56%, respectively.
基于机器学习和深度学习技术的中文新闻分类实证研究
本研究比较了机器学习(ML)和深度学习(DL)的中文新闻分类结果。在ML处理中,我们选择支持向量机(SVM)和朴素贝叶斯(NB),形成Word2Vec-SVM、TFIDF-SVM和TFIDF-NB三个模型。由于NB假设单词是独立的,这与Word2Vec中相关单词分布的概念不同,因此排除了与NB的结合。在深度学习处理中,我们采用了双向长短期记忆(Bi-LSTM)、长短期记忆(LSTM)和卷积神经网络(CNN),并使用Word2Vec进行词嵌入。实验结果表明,通过适当的词预处理,ML和DL模型的分类准确率差异实际上很小。虽然结果表明,与其他深度学习技术相比,Bi-LSTM的准确率最高,损耗最小,但其实现过程耗时最长。本研究肯定了CNN的优异效果,而其Loss是DL模型中最高的。我们还发现,Word2Vec-SVM在效率上优于TFIDF-SVM,但准确率不如预期。综上所示,Bi-LSTM、LSTM、CNN、Word2vec-SVM、TFIDF-SVM、NB的分类准确率分别为89.3%、88%和87.54%、85.32%、87.35%、86.56%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信