{"title":"基于生成与判别混合模型的NLP领域自适应","authors":"Kang Liu, Jun Zhao","doi":"10.1109/CCPR.2008.11","DOIUrl":null,"url":null,"abstract":"This study investigates the domain adaptation problem for nature language processing tasks in the distributional view. A novel method is proposed for domain adaptation based on the hybrid model which combines the discriminative model with the generative model. The advantage of the discriminative model is to have lower asymptotic error, while the advantage of the generative model can easily incorporate the unlabeled data for better generalization performance. The hybrid model can integrate their advantages. For domain transfer, the proposed method exploits the difference of the distributions in different domains to adjust the weights of the instances in the training set so that the source labeled data is more adaptive to the target domain. Experimental results on several NLP tasks in different domains indicate that our method outperforms both the traditional supervised learning and the semi-supervised method.","PeriodicalId":292956,"journal":{"name":"2008 Chinese Conference on Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2008-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Domain Adaptation in NLP Based on Hybrid Generative and Discriminative Model\",\"authors\":\"Kang Liu, Jun Zhao\",\"doi\":\"10.1109/CCPR.2008.11\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This study investigates the domain adaptation problem for nature language processing tasks in the distributional view. A novel method is proposed for domain adaptation based on the hybrid model which combines the discriminative model with the generative model. The advantage of the discriminative model is to have lower asymptotic error, while the advantage of the generative model can easily incorporate the unlabeled data for better generalization performance. The hybrid model can integrate their advantages. For domain transfer, the proposed method exploits the difference of the distributions in different domains to adjust the weights of the instances in the training set so that the source labeled data is more adaptive to the target domain. Experimental results on several NLP tasks in different domains indicate that our method outperforms both the traditional supervised learning and the semi-supervised method.\",\"PeriodicalId\":292956,\"journal\":{\"name\":\"2008 Chinese Conference on Pattern Recognition\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2008-10-31\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2008 Chinese Conference on Pattern Recognition\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CCPR.2008.11\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2008 Chinese Conference on Pattern Recognition","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CCPR.2008.11","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Domain Adaptation in NLP Based on Hybrid Generative and Discriminative Model
This study investigates the domain adaptation problem for nature language processing tasks in the distributional view. A novel method is proposed for domain adaptation based on the hybrid model which combines the discriminative model with the generative model. The advantage of the discriminative model is to have lower asymptotic error, while the advantage of the generative model can easily incorporate the unlabeled data for better generalization performance. The hybrid model can integrate their advantages. For domain transfer, the proposed method exploits the difference of the distributions in different domains to adjust the weights of the instances in the training set so that the source labeled data is more adaptive to the target domain. Experimental results on several NLP tasks in different domains indicate that our method outperforms both the traditional supervised learning and the semi-supervised method.