DDML: Multi-Student Knowledge Distillation for Hate Speech.

IF 2.1 3区 物理与天体物理 Q2 PHYSICS, MULTIDISCIPLINARY
Entropy Pub Date : 2025-04-11 DOI:10.3390/e27040417
Ze Liu, Zerui Shao, Haizhou Wang, Beibei Li
{"title":"DDML: Multi-Student Knowledge Distillation for Hate Speech.","authors":"Ze Liu, Zerui Shao, Haizhou Wang, Beibei Li","doi":"10.3390/e27040417","DOIUrl":null,"url":null,"abstract":"<p><p>Recent studies have shown that hate speech on social media negatively impacts users' mental health and is a contributing factor to suicide attempts. On a broader scale, online hate speech can undermine social stability. With the continuous growth of the internet, the prevalence of online hate speech is rising, making its detection an urgent issue. Recent advances in natural language processing, particularly with transformer-based models, have shown significant promise in hate speech detection. However, these models come with a large number of parameters, leading to high computational requirements and making them difficult to deploy on personal computers. To address these challenges, knowledge distillation offers a solution by training smaller student networks using larger teacher networks. Recognizing that learning also occurs through peer interactions, we propose a knowledge distillation method called Deep Distill-Mutual Learning (DDML). DDML employs one teacher network and two or more student networks. While the student networks benefit from the teacher's knowledge, they also engage in mutual learning with each other. We trained numerous deep neural networks for hate speech detection based on DDML and demonstrated that these networks perform well across various datasets. We tested our method across ten languages and nine datasets. The results demonstrate that DDML enhances the performance of deep neural networks, achieving an average F1 score increase of 4.87% over the baseline.</p>","PeriodicalId":11694,"journal":{"name":"Entropy","volume":"27 4","pages":""},"PeriodicalIF":2.1000,"publicationDate":"2025-04-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12025758/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Entropy","FirstCategoryId":"101","ListUrlMain":"https://doi.org/10.3390/e27040417","RegionNum":3,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"PHYSICS, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

Abstract

Recent studies have shown that hate speech on social media negatively impacts users' mental health and is a contributing factor to suicide attempts. On a broader scale, online hate speech can undermine social stability. With the continuous growth of the internet, the prevalence of online hate speech is rising, making its detection an urgent issue. Recent advances in natural language processing, particularly with transformer-based models, have shown significant promise in hate speech detection. However, these models come with a large number of parameters, leading to high computational requirements and making them difficult to deploy on personal computers. To address these challenges, knowledge distillation offers a solution by training smaller student networks using larger teacher networks. Recognizing that learning also occurs through peer interactions, we propose a knowledge distillation method called Deep Distill-Mutual Learning (DDML). DDML employs one teacher network and two or more student networks. While the student networks benefit from the teacher's knowledge, they also engage in mutual learning with each other. We trained numerous deep neural networks for hate speech detection based on DDML and demonstrated that these networks perform well across various datasets. We tested our method across ten languages and nine datasets. The results demonstrate that DDML enhances the performance of deep neural networks, achieving an average F1 score increase of 4.87% over the baseline.

仇恨言论的多学生知识蒸馏。
最近的研究表明,社交媒体上的仇恨言论对用户的心理健康产生负面影响,是导致自杀企图的一个因素。在更广泛的范围内,网络仇恨言论会破坏社会稳定。随着互联网的不断发展,网络仇恨言论的流行率不断上升,对其进行检测成为一个紧迫的问题。自然语言处理的最新进展,特别是基于变压器的模型,在仇恨言论检测方面显示出了巨大的希望。然而,这些模型带有大量的参数,导致高计算需求,使它们难以在个人计算机上部署。为了应对这些挑战,知识蒸馏提供了一个解决方案,即使用较大的教师网络来训练较小的学生网络。认识到学习也通过同伴互动发生,我们提出了一种知识蒸馏方法,称为深度蒸馏-相互学习(DDML)。DDML使用一个教师网络和两个或更多的学生网络。当学生网络从老师的知识中受益时,他们也在相互学习。我们训练了许多基于DDML的深度神经网络用于仇恨言论检测,并证明这些网络在各种数据集上表现良好。我们在10种语言和9个数据集上测试了我们的方法。结果表明,DDML增强了深度神经网络的性能,平均F1分数比基线提高了4.87%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Entropy
Entropy PHYSICS, MULTIDISCIPLINARY-
CiteScore
4.90
自引率
11.10%
发文量
1580
审稿时长
21.05 days
期刊介绍: Entropy (ISSN 1099-4300), an international and interdisciplinary journal of entropy and information studies, publishes reviews, regular research papers and short notes. Our aim is to encourage scientists to publish as much as possible their theoretical and experimental details. There is no restriction on the length of the papers. If there are computation and the experiment, the details must be provided so that the results can be reproduced.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信