{"title":"利用弱监督和 BiGRU 神经网络对无标签新闻标题进行情感分析","authors":"Ahamadali Jamali, Shahin Alipour, Audrey Rah","doi":"10.1109/ICAIC60265.2024.10433844","DOIUrl":null,"url":null,"abstract":"Auto-labeling of text is a useful and necessary technique for creating large and high-quality training data sets for machine learning models. Label-free sentiment classification is a challenging semi-supervised task in the natural language processing domain. This study leveraged the weak supervision framework to generate weak labels in three categories for millions of news headlines from Australian Broadcasting Corporation (ABC). A Bidirectional Gate Recurrent Unit (BiGRU) was then trained with neural network dense layers to achieve a validation accuracy of 96.76% with 99.99% accuracy. The performance of this method was also compared with traditional and deep learning natural language processing techniques.","PeriodicalId":517265,"journal":{"name":"2024 IEEE 3rd International Conference on AI in Cybersecurity (ICAIC)","volume":"3 3","pages":"1-5"},"PeriodicalIF":0.0000,"publicationDate":"2024-02-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Leveraging Weak Supervision and BiGRU Neural Networks for Sentiment Analysis on Label-Free News Headlines\",\"authors\":\"Ahamadali Jamali, Shahin Alipour, Audrey Rah\",\"doi\":\"10.1109/ICAIC60265.2024.10433844\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Auto-labeling of text is a useful and necessary technique for creating large and high-quality training data sets for machine learning models. Label-free sentiment classification is a challenging semi-supervised task in the natural language processing domain. This study leveraged the weak supervision framework to generate weak labels in three categories for millions of news headlines from Australian Broadcasting Corporation (ABC). A Bidirectional Gate Recurrent Unit (BiGRU) was then trained with neural network dense layers to achieve a validation accuracy of 96.76% with 99.99% accuracy. The performance of this method was also compared with traditional and deep learning natural language processing techniques.\",\"PeriodicalId\":517265,\"journal\":{\"name\":\"2024 IEEE 3rd International Conference on AI in Cybersecurity (ICAIC)\",\"volume\":\"3 3\",\"pages\":\"1-5\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-02-07\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2024 IEEE 3rd International Conference on AI in Cybersecurity (ICAIC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICAIC60265.2024.10433844\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2024 IEEE 3rd International Conference on AI in Cybersecurity (ICAIC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICAIC60265.2024.10433844","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Leveraging Weak Supervision and BiGRU Neural Networks for Sentiment Analysis on Label-Free News Headlines
Auto-labeling of text is a useful and necessary technique for creating large and high-quality training data sets for machine learning models. Label-free sentiment classification is a challenging semi-supervised task in the natural language processing domain. This study leveraged the weak supervision framework to generate weak labels in three categories for millions of news headlines from Australian Broadcasting Corporation (ABC). A Bidirectional Gate Recurrent Unit (BiGRU) was then trained with neural network dense layers to achieve a validation accuracy of 96.76% with 99.99% accuracy. The performance of this method was also compared with traditional and deep learning natural language processing techniques.