{"title":"PRCD:全链平行剩余补偿去偏框架","authors":"Chenyang Li , Maoyuan Zhang , Meng Zheng","doi":"10.1016/j.eswa.2025.128544","DOIUrl":null,"url":null,"abstract":"<div><div>Hate speech detection models often suffer from systemic misclassification due to biases in training data, particularly cognitive distortions in semantic associations learned in deep networks, making it especially challenging to identify implicit biases. While existing neuron pruning methods can mitigate explicit biases to some extent, removing parameters weakens the model’s semantic representation capability and struggles to address deeply ingrained cognitively distorted associations in deep networks. To tackle this issue, this paper proposes a full-chain parallel residual compensation debiasing framework (PRCD). This framework introduces a residual compensation method for soft pruning (RCM), which enables soft pruning of bias signals across the entire model, from shallow to deep layers—without compromising the model’s semantic representation ability. Additionally, a toxicity constraint enhancement method based on sensitivity attribution prediction (ToxCon) is incorporated to generate contrastive samples that expose bias, effectively guiding RCM in correcting implicit biases stemming from cognitive distortions in semantic associations. Experimental results on three public datasets demonstrate that the PRCD significantly improves model performance and fairness in detecting hate speech, achieving state-of-the-art performance.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"291 ","pages":"Article 128544"},"PeriodicalIF":7.5000,"publicationDate":"2025-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"PRCD: A full-chain parallel residual compensation debiasing framework\",\"authors\":\"Chenyang Li , Maoyuan Zhang , Meng Zheng\",\"doi\":\"10.1016/j.eswa.2025.128544\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Hate speech detection models often suffer from systemic misclassification due to biases in training data, particularly cognitive distortions in semantic associations learned in deep networks, making it especially challenging to identify implicit biases. While existing neuron pruning methods can mitigate explicit biases to some extent, removing parameters weakens the model’s semantic representation capability and struggles to address deeply ingrained cognitively distorted associations in deep networks. To tackle this issue, this paper proposes a full-chain parallel residual compensation debiasing framework (PRCD). This framework introduces a residual compensation method for soft pruning (RCM), which enables soft pruning of bias signals across the entire model, from shallow to deep layers—without compromising the model’s semantic representation ability. Additionally, a toxicity constraint enhancement method based on sensitivity attribution prediction (ToxCon) is incorporated to generate contrastive samples that expose bias, effectively guiding RCM in correcting implicit biases stemming from cognitive distortions in semantic associations. Experimental results on three public datasets demonstrate that the PRCD significantly improves model performance and fairness in detecting hate speech, achieving state-of-the-art performance.</div></div>\",\"PeriodicalId\":50461,\"journal\":{\"name\":\"Expert Systems with Applications\",\"volume\":\"291 \",\"pages\":\"Article 128544\"},\"PeriodicalIF\":7.5000,\"publicationDate\":\"2025-06-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Expert Systems with Applications\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0957417425021633\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Expert Systems with Applications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0957417425021633","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
PRCD: A full-chain parallel residual compensation debiasing framework
Hate speech detection models often suffer from systemic misclassification due to biases in training data, particularly cognitive distortions in semantic associations learned in deep networks, making it especially challenging to identify implicit biases. While existing neuron pruning methods can mitigate explicit biases to some extent, removing parameters weakens the model’s semantic representation capability and struggles to address deeply ingrained cognitively distorted associations in deep networks. To tackle this issue, this paper proposes a full-chain parallel residual compensation debiasing framework (PRCD). This framework introduces a residual compensation method for soft pruning (RCM), which enables soft pruning of bias signals across the entire model, from shallow to deep layers—without compromising the model’s semantic representation ability. Additionally, a toxicity constraint enhancement method based on sensitivity attribution prediction (ToxCon) is incorporated to generate contrastive samples that expose bias, effectively guiding RCM in correcting implicit biases stemming from cognitive distortions in semantic associations. Experimental results on three public datasets demonstrate that the PRCD significantly improves model performance and fairness in detecting hate speech, achieving state-of-the-art performance.
期刊介绍:
Expert Systems With Applications is an international journal dedicated to the exchange of information on expert and intelligent systems used globally in industry, government, and universities. The journal emphasizes original papers covering the design, development, testing, implementation, and management of these systems, offering practical guidelines. It spans various sectors such as finance, engineering, marketing, law, project management, information management, medicine, and more. The journal also welcomes papers on multi-agent systems, knowledge management, neural networks, knowledge discovery, data mining, and other related areas, excluding applications to military/defense systems.