论文标题

学习半监督领域适应的不变表示和风险

Learning Invariant Representations and Risks for Semi-supervised Domain Adaptation

论文作者

Li, Bo, Wang, Yezhen, Zhang, Shanghang, Li, Dongsheng, Darrell, Trevor, Keutzer, Kurt, Zhao, Han

论文摘要

监督学习的成功取决于假设训练和测试数据来自相同的基础分布,这在实践中通常是由于潜在的分配变化而无效的。鉴于此,大多数现有的无监督域适应性方法都集中在实现域名表示和小源域误差上。但是,最近的作品表明,这不足以保证对目标域的良好概括,实际上,在标签分布变化下证明是有害的。此外,在许多实际应用中,从目标域中获得少量标记数据并使用它们来促进使用源数据的模型培训,通常是可行的。受上述观察的启发,在本文中,我们提出了一种旨在在半监督域适应(Semi-DA)设置下同时学习不变的表示和风险的方法。首先,我们提供了一个有限的样本,以针对半DA下的分类和回归问题绑定。该边界提出了一种获得目标概括的原则方法,即通过在特征空间中跨域之间的边际和条件分布对齐。然后,我们介绍了共同\ textbf {l} ginal \ textbf {i} nvariant \ textbf {r} ePresentations和\ textbf {r} isks的LIRR算法。最后,与仅学习不变表示或不变风险的方法相比,对分类和回归任务进行了广泛的实验,这表明LIRR始终达到最先进的性能和显着改进。

The success of supervised learning hinges on the assumption that the training and test data come from the same underlying distribution, which is often not valid in practice due to potential distribution shift. In light of this, most existing methods for unsupervised domain adaptation focus on achieving domain-invariant representations and small source domain error. However, recent works have shown that this is not sufficient to guarantee good generalization on the target domain, and in fact, is provably detrimental under label distribution shift. Furthermore, in many real-world applications it is often feasible to obtain a small amount of labeled data from the target domain and use them to facilitate model training with source data. Inspired by the above observations, in this paper we propose the first method that aims to simultaneously learn invariant representations and risks under the setting of semi-supervised domain adaptation (Semi-DA). First, we provide a finite sample bound for both classification and regression problems under Semi-DA. The bound suggests a principled way to obtain target generalization, i.e. by aligning both the marginal and conditional distributions across domains in feature space. Motivated by this, we then introduce the LIRR algorithm for jointly \textbf{L}earning \textbf{I}nvariant \textbf{R}epresentations and \textbf{R}isks. Finally, extensive experiments are conducted on both classification and regression tasks, which demonstrates LIRR consistently achieves state-of-the-art performance and significant improvements compared with the methods that only learn invariant representations or invariant risks.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源