{"title":"Cross-Domain Helpfulness Prediction of Online Consumer Reviews by Deep Learning Model","authors":"Shih-Hung Wu, Yi-Kun Chen","doi":"10.1109/IRI49571.2020.00069","DOIUrl":null,"url":null,"abstract":"Customer reviews provide helpful information such as usage experiences or critiques; these are critical information resource for future customers. Since the amount of online review is getting bigger, people need a way to find the most helpful ones automatically. Previous studies addressed on the prediction of the percentage of the helpfulness voting results based on a regression model or classified them into a helpful or unhelpful classes. However, the voting result of an online review is not a constant over time, and we also find that there are many reviews getting zero vote. Therefore, we collect the voting results of the same online customer reviews over time, and observe the change of votes to find a better learning target. We collected a dataset with online reviews in five different product categories (“Apple”, “Video Game”, “Clothing, Shoes & Jewelry”, “Sports & Outdoors”, and “Prime Video”) from Amazon.com with the voting result on the helpfulness of the reviews, and monitor the helpfulness voting for six weeks. Experiments are conducted on the dataset to get a reasonable classification on the zero and non-zero vote reviews. We construct a classification system that can classify the online reviews via the deep learning model BERT. The results show that the classifier can get good result on the helpfulness prediction. We also test the classifier on cross-domain prediction and get promising results.","PeriodicalId":93159,"journal":{"name":"2020 IEEE 21st International Conference on Information Reuse and Integration for Data Science : IRI 2020 : proceedings : virtual conference, 11-13 August 2020. IEEE International Conference on Information Reuse and Integration (21st : 2...","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2020-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE 21st International Conference on Information Reuse and Integration for Data Science : IRI 2020 : proceedings : virtual conference, 11-13 August 2020. IEEE International Conference on Information Reuse and Integration (21st : 2...","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IRI49571.2020.00069","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5
Abstract
Customer reviews provide helpful information such as usage experiences or critiques; these are critical information resource for future customers. Since the amount of online review is getting bigger, people need a way to find the most helpful ones automatically. Previous studies addressed on the prediction of the percentage of the helpfulness voting results based on a regression model or classified them into a helpful or unhelpful classes. However, the voting result of an online review is not a constant over time, and we also find that there are many reviews getting zero vote. Therefore, we collect the voting results of the same online customer reviews over time, and observe the change of votes to find a better learning target. We collected a dataset with online reviews in five different product categories (“Apple”, “Video Game”, “Clothing, Shoes & Jewelry”, “Sports & Outdoors”, and “Prime Video”) from Amazon.com with the voting result on the helpfulness of the reviews, and monitor the helpfulness voting for six weeks. Experiments are conducted on the dataset to get a reasonable classification on the zero and non-zero vote reviews. We construct a classification system that can classify the online reviews via the deep learning model BERT. The results show that the classifier can get good result on the helpfulness prediction. We also test the classifier on cross-domain prediction and get promising results.