{"title":"On Domain Generalization for Batched Prediction: the Benefit of Contextual Adversarial Training","authors":"Chune Li, Yongyi Mao, Richong Zhang","doi":"10.1109/ICTAI56018.2022.00091","DOIUrl":null,"url":null,"abstract":"This paper considers the domain generalization problem in which the labels of the query batch may be predicted at once and the model is allowed to assign a label to a query example utilizing all information contained in the query batch. In this setting, we identify two problems in the standard adaptive risk minimization (ARM) framework, namely, mismatched context and overfitted context, which are most severely manifested respectively at small and large support batch sizes. The existence of these problems signifies a fundamental tradeoff between within-domain learning and cross-domain generalization. We also propose two context adversarial training approaches to alleviate these problems so as to achieve a better tradeoff. We demonstrate experimentally that the proposed approaches uniformly outperform the standard ARM training for all choices of support batch sizes.","PeriodicalId":354314,"journal":{"name":"2022 IEEE 34th International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"44 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 34th International Conference on Tools with Artificial Intelligence (ICTAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICTAI56018.2022.00091","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
This paper considers the domain generalization problem in which the labels of the query batch may be predicted at once and the model is allowed to assign a label to a query example utilizing all information contained in the query batch. In this setting, we identify two problems in the standard adaptive risk minimization (ARM) framework, namely, mismatched context and overfitted context, which are most severely manifested respectively at small and large support batch sizes. The existence of these problems signifies a fundamental tradeoff between within-domain learning and cross-domain generalization. We also propose two context adversarial training approaches to alleviate these problems so as to achieve a better tradeoff. We demonstrate experimentally that the proposed approaches uniformly outperform the standard ARM training for all choices of support batch sizes.