{"title":"Team Wa’ed Al-Shrida at the Mowjaz Multi-Topic Labelling Task","authors":"Wa’ed Al-Shrida","doi":"10.1109/ICICS52457.2021.9464574","DOIUrl":null,"url":null,"abstract":"This paper describes an attempt in the \"Mowjaz Multi-topic Labelling Task\", the competition is about classifying the Arabic articles to its topic by using artificial intelligence and neural networks, the programming language that was used to classify the datasets is \"Python\". The attempt was started by uploading the datasets from the \"Github website\", the datasets that were used in the system include three groups, train, validation, and test datasets. The \"Pyarabic\" and simple-transformers\" libraries were used to allow the system to manipulate Arabic letters and simplify the usage of Transformer models without having to compromise on utility, respectively. The model’s type that I used is \"Bert\" and its name is \"Asafaya/Bert-base-Arabic\". The accuracy of the result that was gotten is as follows: F1 macro: 0.864, F1 micro: 0.869, competition website on Codalab: 0.8430.","PeriodicalId":421803,"journal":{"name":"2021 12th International Conference on Information and Communication Systems (ICICS)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-05-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 12th International Conference on Information and Communication Systems (ICICS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICICS52457.2021.9464574","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This paper describes an attempt in the "Mowjaz Multi-topic Labelling Task", the competition is about classifying the Arabic articles to its topic by using artificial intelligence and neural networks, the programming language that was used to classify the datasets is "Python". The attempt was started by uploading the datasets from the "Github website", the datasets that were used in the system include three groups, train, validation, and test datasets. The "Pyarabic" and simple-transformers" libraries were used to allow the system to manipulate Arabic letters and simplify the usage of Transformer models without having to compromise on utility, respectively. The model’s type that I used is "Bert" and its name is "Asafaya/Bert-base-Arabic". The accuracy of the result that was gotten is as follows: F1 macro: 0.864, F1 micro: 0.869, competition website on Codalab: 0.8430.