{"title":"A Comparative Study of Pre-trained Word Embeddings for Arabic Sentiment Analysis","authors":"Mohamed Zouidine, Mohammed Khalil","doi":"10.1109/COMPSAC54236.2022.00196","DOIUrl":null,"url":null,"abstract":"In this paper, we conduct a series of experiments to systematically study both context-independent and context-dependent word embeddings for the purpose of Arabic sentiment analysis. We use pretrained word embeddings as fixed features extractors to provide input features for a CNN model. Experimental results with two different Arabic sentiment analysis datasets indicate that the pre-trained contextualized AraBERT model is the most suitable for such tasks. AraBERT reaches an accuracy score of 91.4% and 95.49% on the large Arabic book reviews dataset (LABR) and the hotel Arabic-reviews dataset (HARD), respectively.","PeriodicalId":330838,"journal":{"name":"2022 IEEE 46th Annual Computers, Software, and Applications Conference (COMPSAC)","volume":"61 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 46th Annual Computers, Software, and Applications Conference (COMPSAC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/COMPSAC54236.2022.00196","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In this paper, we conduct a series of experiments to systematically study both context-independent and context-dependent word embeddings for the purpose of Arabic sentiment analysis. We use pretrained word embeddings as fixed features extractors to provide input features for a CNN model. Experimental results with two different Arabic sentiment analysis datasets indicate that the pre-trained contextualized AraBERT model is the most suitable for such tasks. AraBERT reaches an accuracy score of 91.4% and 95.49% on the large Arabic book reviews dataset (LABR) and the hotel Arabic-reviews dataset (HARD), respectively.