{"title":"Bootstrapping both Product Properties and Opinion Words from Chinese Reviews with Cross-Training","authors":"Bo Wang, Houfeng Wang","doi":"10.1109/WI.2007.32","DOIUrl":null,"url":null,"abstract":"We investigate the problem of identifying both product properties and opinion words for sentences in a unified process when only a much small labeled corpus is available. Naive Bayesian method is used in this process. Specifically, considering the fact that product properties and opinion words usually co-occur with high frequency in product review articles, a cross- training method is proposed to bootstrap both of them, in which the two sub-tasks are boosted by each other iteratively. Experiment results show that with a much small labeled corpus cross-training could produce both product properties and opinion words which are very close to what Naive Bayesian Classifiers could do with a large labeled corpus..","PeriodicalId":192501,"journal":{"name":"IEEE/WIC/ACM International Conference on Web Intelligence (WI'07)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2007-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"27","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE/WIC/ACM International Conference on Web Intelligence (WI'07)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WI.2007.32","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 27
Abstract
We investigate the problem of identifying both product properties and opinion words for sentences in a unified process when only a much small labeled corpus is available. Naive Bayesian method is used in this process. Specifically, considering the fact that product properties and opinion words usually co-occur with high frequency in product review articles, a cross- training method is proposed to bootstrap both of them, in which the two sub-tasks are boosted by each other iteratively. Experiment results show that with a much small labeled corpus cross-training could produce both product properties and opinion words which are very close to what Naive Bayesian Classifiers could do with a large labeled corpus..