{"title":"Domain Adaptation in NLP Based on Hybrid Generative and Discriminative Model","authors":"Kang Liu, Jun Zhao","doi":"10.1109/CCPR.2008.11","DOIUrl":null,"url":null,"abstract":"This study investigates the domain adaptation problem for nature language processing tasks in the distributional view. A novel method is proposed for domain adaptation based on the hybrid model which combines the discriminative model with the generative model. The advantage of the discriminative model is to have lower asymptotic error, while the advantage of the generative model can easily incorporate the unlabeled data for better generalization performance. The hybrid model can integrate their advantages. For domain transfer, the proposed method exploits the difference of the distributions in different domains to adjust the weights of the instances in the training set so that the source labeled data is more adaptive to the target domain. Experimental results on several NLP tasks in different domains indicate that our method outperforms both the traditional supervised learning and the semi-supervised method.","PeriodicalId":292956,"journal":{"name":"2008 Chinese Conference on Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2008-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2008 Chinese Conference on Pattern Recognition","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CCPR.2008.11","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This study investigates the domain adaptation problem for nature language processing tasks in the distributional view. A novel method is proposed for domain adaptation based on the hybrid model which combines the discriminative model with the generative model. The advantage of the discriminative model is to have lower asymptotic error, while the advantage of the generative model can easily incorporate the unlabeled data for better generalization performance. The hybrid model can integrate their advantages. For domain transfer, the proposed method exploits the difference of the distributions in different domains to adjust the weights of the instances in the training set so that the source labeled data is more adaptive to the target domain. Experimental results on several NLP tasks in different domains indicate that our method outperforms both the traditional supervised learning and the semi-supervised method.