论文标题
域概括的最大最大跨域可能性约束
Constrained Maximum Cross-Domain Likelihood for Domain Generalization
论文作者
论文摘要
作为最近的显着主题,域的概括旨在学习多个源域的可推广模型,该模型有望在看不见的测试域上表现良好。通过使跨域之间的分布对齐分布来学习域不变特征。但是,现有作品通常是基于一些放松条件的设计,这些条件通常很难满足并且无法实现所需的联合分配对准。在本文中,我们提出了一种新颖的域泛化方法,该方法源于一个直观的想法,即可以通过最大程度地减少来自不同域的后验分布之间的KL差异来学习域不变的分类器。为了增强学习分类器的概括性,我们将优化目标正式化为在地面边缘分布上计算的预期。然而,它还呈现出两个明显的缺陷,其中之一是熵增加的熵增加,另一个是地面边缘分布的不可用。对于前者,我们介绍了一个名为“最大域内可能性”的术语,以维持学习域不变的代表空间的歧视。对于后者,我们在合理的凸面船体假设下与源域近似地面真相边缘分布。最后,通过求解关节分布自然对齐的约束最大跨域可能性(CMCL)优化问题。仔细设计了交替的优化策略,以大致解决此优化问题。在四个标准基准数据集(即Digits-dg,pacs,Office-home和Minidomainnet)上进行了广泛的实验,突出了我们方法的出色性能。
As a recent noticeable topic, domain generalization aims to learn a generalizable model on multiple source domains, which is expected to perform well on unseen test domains. Great efforts have been made to learn domain-invariant features by aligning distributions across domains. However, existing works are often designed based on some relaxed conditions which are generally hard to satisfy and fail to realize the desired joint distribution alignment. In this paper, we propose a novel domain generalization method, which originates from an intuitive idea that a domain-invariant classifier can be learned by minimizing the KL-divergence between posterior distributions from different domains. To enhance the generalizability of the learned classifier, we formalize the optimization objective as an expectation computed on the ground-truth marginal distribution. Nevertheless, it also presents two obvious deficiencies, one of which is the side-effect of entropy increase in KL-divergence and the other is the unavailability of ground-truth marginal distributions. For the former, we introduce a term named maximum in-domain likelihood to maintain the discrimination of the learned domain-invariant representation space. For the latter, we approximate the ground-truth marginal distribution with source domains under a reasonable convex hull assumption. Finally, a Constrained Maximum Cross-domain Likelihood (CMCL) optimization problem is deduced, by solving which the joint distributions are naturally aligned. An alternating optimization strategy is carefully designed to approximately solve this optimization problem. Extensive experiments on four standard benchmark datasets, i.e., Digits-DG, PACS, Office-Home and miniDomainNet, highlight the superior performance of our method.