{"title":"检测儿童性虐待材料(CSAM)中的露骨性内容:端到端分类器和基于区域的网络","authors":"Weronika Gutfeter, Joanna Gajewska, Andrzej Pacut","doi":"arxiv-2406.14131","DOIUrl":null,"url":null,"abstract":"Child sexual abuse materials (CSAM) pose a significant threat to the safety\nand well-being of children worldwide. Detecting and preventing the distribution\nof such materials is a critical task for law enforcement agencies and\ntechnology companies. As content moderation is often manual, developing an\nautomated detection system can help reduce human reviewers' exposure to\npotentially harmful images and accelerate the process of counteracting. This\nstudy presents methods for classifying sexually explicit content, which plays a\ncrucial role in the automated CSAM detection system. Several approaches are\nexplored to solve the task: an end-to-end classifier, a classifier with person\ndetection and a private body parts detector. All proposed methods are tested on\nthe images obtained from the online tool for reporting illicit content. Due to\nlegal constraints, access to the data is limited, and all algorithms are\nexecuted remotely on the isolated server. The end-to-end classifier yields the\nmost promising results, with an accuracy of 90.17%, after augmenting the\ntraining set with the additional neutral samples and adult pornography. While\ndetection-based methods may not achieve higher accuracy rates and cannot serve\nas a final classifier on their own, their inclusion in the system can be\nbeneficial. Human body-oriented approaches generate results that are easier to\ninterpret, and obtaining more interpretable results is essential when analyzing\nmodels that are trained without direct access to data.","PeriodicalId":501168,"journal":{"name":"arXiv - CS - Emerging Technologies","volume":"9 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Detecting sexually explicit content in the context of the child sexual abuse materials (CSAM): end-to-end classifiers and region-based networks\",\"authors\":\"Weronika Gutfeter, Joanna Gajewska, Andrzej Pacut\",\"doi\":\"arxiv-2406.14131\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Child sexual abuse materials (CSAM) pose a significant threat to the safety\\nand well-being of children worldwide. Detecting and preventing the distribution\\nof such materials is a critical task for law enforcement agencies and\\ntechnology companies. As content moderation is often manual, developing an\\nautomated detection system can help reduce human reviewers' exposure to\\npotentially harmful images and accelerate the process of counteracting. This\\nstudy presents methods for classifying sexually explicit content, which plays a\\ncrucial role in the automated CSAM detection system. Several approaches are\\nexplored to solve the task: an end-to-end classifier, a classifier with person\\ndetection and a private body parts detector. All proposed methods are tested on\\nthe images obtained from the online tool for reporting illicit content. Due to\\nlegal constraints, access to the data is limited, and all algorithms are\\nexecuted remotely on the isolated server. The end-to-end classifier yields the\\nmost promising results, with an accuracy of 90.17%, after augmenting the\\ntraining set with the additional neutral samples and adult pornography. While\\ndetection-based methods may not achieve higher accuracy rates and cannot serve\\nas a final classifier on their own, their inclusion in the system can be\\nbeneficial. Human body-oriented approaches generate results that are easier to\\ninterpret, and obtaining more interpretable results is essential when analyzing\\nmodels that are trained without direct access to data.\",\"PeriodicalId\":501168,\"journal\":{\"name\":\"arXiv - CS - Emerging Technologies\",\"volume\":\"9 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-06-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Emerging Technologies\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2406.14131\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Emerging Technologies","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2406.14131","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Detecting sexually explicit content in the context of the child sexual abuse materials (CSAM): end-to-end classifiers and region-based networks
Child sexual abuse materials (CSAM) pose a significant threat to the safety
and well-being of children worldwide. Detecting and preventing the distribution
of such materials is a critical task for law enforcement agencies and
technology companies. As content moderation is often manual, developing an
automated detection system can help reduce human reviewers' exposure to
potentially harmful images and accelerate the process of counteracting. This
study presents methods for classifying sexually explicit content, which plays a
crucial role in the automated CSAM detection system. Several approaches are
explored to solve the task: an end-to-end classifier, a classifier with person
detection and a private body parts detector. All proposed methods are tested on
the images obtained from the online tool for reporting illicit content. Due to
legal constraints, access to the data is limited, and all algorithms are
executed remotely on the isolated server. The end-to-end classifier yields the
most promising results, with an accuracy of 90.17%, after augmenting the
training set with the additional neutral samples and adult pornography. While
detection-based methods may not achieve higher accuracy rates and cannot serve
as a final classifier on their own, their inclusion in the system can be
beneficial. Human body-oriented approaches generate results that are easier to
interpret, and obtaining more interpretable results is essential when analyzing
models that are trained without direct access to data.