ConBERT-RL: A policy-driven deep reinforcement learning based approach for detecting homophobia and transphobia in low-resource languages

Vivek Suresh Raj , Chinnaudayar Navaneethakrishnan Subalalitha , Lavanya Sambath , Frank Glavin , Bharathi Raja Chakravarthi
{"title":"ConBERT-RL: A policy-driven deep reinforcement learning based approach for detecting homophobia and transphobia in low-resource languages","authors":"Vivek Suresh Raj ,&nbsp;Chinnaudayar Navaneethakrishnan Subalalitha ,&nbsp;Lavanya Sambath ,&nbsp;Frank Glavin ,&nbsp;Bharathi Raja Chakravarthi","doi":"10.1016/j.nlp.2023.100040","DOIUrl":null,"url":null,"abstract":"<div><p>In this work, we present a novel framework for discriminatory comment classification in targeted low-resource languages thereby enabling identification of discriminatory comments for promoting safer online environment in linguistically diverse contexts. Recently proposed literatures with Bidirectional Encoder Representations from Transformers (BERT) and its variants, have produced promising results, particularly in the case of transliterated Tamil words in English. Such approaches are seen as transfer learning and fine-tuning between a general environment and targeted downstream task. However, for an effective transfer of knowledge, from a source task to a targeted task, their feature space has to be correlated (similarity measure as metric, to measure knowledge transfer), which is unexplored in the previous literature works. In practice, such similarity conditions are often violated. We propose a <strong>Con</strong>catenated representation, powered with <strong>BERT</strong> in <strong>R</strong>einforcement <strong>L</strong>earning (RL) - ConBERT-RL framework, to capture problem-specific features, as well as to understand nuances in case of transliterated Tamil words into English, to make decision in classifying discriminatory comments. ConBERT-RL incorporates a fusion of learned hidden-state representation from our pre-trained Classifier-Model (CM), along with the broader contextualized representation (pooled output) from BERT. The key idea is to utilize this concatenated representation to drive our policy network for hate comment classification. To effectively learn such a policy, we use the REINFORCE algorithm in a reinforcement learning setting, to guide our ConBERT-RL agent, in making informative decisions. To demonstrate the general aspects of ConBERT-RL, we conduct experiments for offensive comment classification on transliterated Tamil words in English dataset. ConBERT-RL obtains results, where it significantly improved the score of micro average accuracy with 90% (<span><math><mrow><mo>≈</mo><mn>1</mn><mo>.</mo><mn>0</mn><mtext>%</mtext></mrow></math></span> absolute improvement over BERT+FC), 93% (<span><math><mrow><mo>≈</mo><mn>3</mn><mo>.</mo><mn>0</mn><mtext>%</mtext></mrow></math></span> absolute improvement over BERT+FC), on transliterated Tamil words in English and an English-only dataset respectively. To further extend support to our previous argument, we present a 2-dimensional t-distributed Stochastic Neighbour Embedding (t-SNE) visualization of ConBERT-RL’s concatenated representation. Additionally, to compare the feature space understanding, specific to the problem of discriminatory comment, we present a graph network comparison our concatenated representation with BERT output embedding. We also design and conduct, a systematic evaluation, to observe the broader capability of ConBERT-RL, in capturing the contextual words in the vicinity of the primary offensive term, that amplifies an offensive term in the given input comment. We show that ConBERT-RL is robust and effective in capturing targeted language specific features for hate comment classification.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"6 ","pages":"Article 100040"},"PeriodicalIF":0.0000,"publicationDate":"2023-11-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719123000377/pdfft?md5=8cc404e7e598ebc58fe4e84a8d0d1037&pid=1-s2.0-S2949719123000377-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Natural Language Processing Journal","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2949719123000377","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

In this work, we present a novel framework for discriminatory comment classification in targeted low-resource languages thereby enabling identification of discriminatory comments for promoting safer online environment in linguistically diverse contexts. Recently proposed literatures with Bidirectional Encoder Representations from Transformers (BERT) and its variants, have produced promising results, particularly in the case of transliterated Tamil words in English. Such approaches are seen as transfer learning and fine-tuning between a general environment and targeted downstream task. However, for an effective transfer of knowledge, from a source task to a targeted task, their feature space has to be correlated (similarity measure as metric, to measure knowledge transfer), which is unexplored in the previous literature works. In practice, such similarity conditions are often violated. We propose a Concatenated representation, powered with BERT in Reinforcement Learning (RL) - ConBERT-RL framework, to capture problem-specific features, as well as to understand nuances in case of transliterated Tamil words into English, to make decision in classifying discriminatory comments. ConBERT-RL incorporates a fusion of learned hidden-state representation from our pre-trained Classifier-Model (CM), along with the broader contextualized representation (pooled output) from BERT. The key idea is to utilize this concatenated representation to drive our policy network for hate comment classification. To effectively learn such a policy, we use the REINFORCE algorithm in a reinforcement learning setting, to guide our ConBERT-RL agent, in making informative decisions. To demonstrate the general aspects of ConBERT-RL, we conduct experiments for offensive comment classification on transliterated Tamil words in English dataset. ConBERT-RL obtains results, where it significantly improved the score of micro average accuracy with 90% (1.0% absolute improvement over BERT+FC), 93% (3.0% absolute improvement over BERT+FC), on transliterated Tamil words in English and an English-only dataset respectively. To further extend support to our previous argument, we present a 2-dimensional t-distributed Stochastic Neighbour Embedding (t-SNE) visualization of ConBERT-RL’s concatenated representation. Additionally, to compare the feature space understanding, specific to the problem of discriminatory comment, we present a graph network comparison our concatenated representation with BERT output embedding. We also design and conduct, a systematic evaluation, to observe the broader capability of ConBERT-RL, in capturing the contextual words in the vicinity of the primary offensive term, that amplifies an offensive term in the given input comment. We show that ConBERT-RL is robust and effective in capturing targeted language specific features for hate comment classification.

ConBERT-RL:基于策略驱动的深度强化学习方法,用于检测低资源语言中的仇视同性恋和仇视变性者现象
在这项工作中,我们提出了一个针对低资源语言的歧视性评论分类新框架,从而能够识别歧视性评论,在语言多样化的背景下促进更安全的网络环境。最近提出的变压器双向编码器表示法(BERT)及其变体取得了可喜的成果,尤其是在泰米尔语单词音译为英语的情况下。这些方法被视为一般环境和目标下游任务之间的迁移学习和微调。然而,为了实现知识从源任务到目标任务的有效转移,它们的特征空间必须具有相关性(以相似性作为衡量知识转移的度量标准),而这一点在以往的文献中还没有得到探讨。在实践中,这种相似性条件经常被违反。我们提出了一种在强化学习(RL)中使用 BERT 的并联表示法--ConBERT-RL 框架,以捕捉特定问题的特征,并理解将泰米尔语单词音译为英语时的细微差别,从而在对歧视性评论进行分类时做出决策。ConBERT-RL 融合了从我们预先训练的分类器模型(CM)中学习到的隐藏状态表示法,以及从 BERT 中学习到的更广泛的上下文表示法(集合输出)。其关键思路是利用这一融合表征来驱动我们的策略网络,以进行仇恨评论分类。为了有效地学习这种策略,我们在强化学习设置中使用了 REINFORCE 算法,以指导我们的 ConBERT-RL 代理做出明智的决定。为了展示 ConBERT-RL 的一般特性,我们对英语中泰米尔语单词的音译数据集进行了攻击性评论分类实验。结果显示,ConBERT-RL 显著提高了微观平均准确率,在英语泰米尔语音译词和纯英语数据集上的准确率分别为 90%(与 BERT+FC 相比绝对值提高了 ≈1.0%)和 93%(与 BERT+FC 相比绝对值提高了 ≈3.0%)。为了进一步支持我们之前的论点,我们展示了 ConBERT-RL 连接表示的二维 t 分布随机邻域嵌入(t-SNE)可视化。此外,为了比较对特征空间的理解,特别是对判别评论问题的理解,我们提出了一个图网络,将我们的串联表示法与 BERT 输出嵌入进行比较。我们还设计并进行了系统评估,以观察 ConBERT-RL 在捕捉主要攻击性词语附近的上下文词语方面的更广泛能力,这些词语在给定的输入评论中放大了攻击性词语。结果表明,ConBERT-RL 在捕捉仇恨评论分类的目标语言特定特征方面既稳健又有效。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信