论文标题
足够的不变学习以进行分发转移
Sufficient Invariant Learning for Distribution Shift
论文作者
论文摘要
在培训和测试数据集之间进行分配变化的学习强大模型是机器学习的基本挑战。虽然在环境中学习不变特征是一种流行的方法,但通常假设这些功能在训练和测试集中都完全观察到 - 在实践中经常违反。当模型依赖于测试集中缺乏的不变特征时,它们在新环境中的鲁棒性可能会恶化。为了解决这个问题,我们介绍了一种新颖的学习原理,称为足够不变的学习(SIL)框架,该框架着重于学习足够的不变特征子集,而不是依靠单个功能。在证明了现有不变学习方法的局限性之后,我们提出了一种新算法,自适应清晰度 - 感知的群体分布在强大的优化(ASGDRO),以通过在整个环境中寻求常见的平面最小值来学习多样的不变特征。从理论上讲,我们证明,找到一个常见的平面最小值可以基于不同的不变特征实现强大的预测。在包括我们的新基准在内的多个数据集上的经验评估确认了Asgdro对分配变化的鲁棒性,突出了现有方法的局限性。
Learning robust models under distribution shifts between training and test datasets is a fundamental challenge in machine learning. While learning invariant features across environments is a popular approach, it often assumes that these features are fully observed in both training and test sets-a condition frequently violated in practice. When models rely on invariant features absent in the test set, their robustness in new environments can deteriorate. To tackle this problem, we introduce a novel learning principle called the Sufficient Invariant Learning (SIL) framework, which focuses on learning a sufficient subset of invariant features rather than relying on a single feature. After demonstrating the limitation of existing invariant learning methods, we propose a new algorithm, Adaptive Sharpness-aware Group Distributionally Robust Optimization (ASGDRO), to learn diverse invariant features by seeking common flat minima across the environments. We theoretically demonstrate that finding a common flat minima enables robust predictions based on diverse invariant features. Empirical evaluations on multiple datasets, including our new benchmark, confirm ASGDRO's robustness against distribution shifts, highlighting the limitations of existing methods.