论文标题
从互补标签学习减少到概率估计
Reduction from Complementary-Label Learning to Probability Estimates
论文作者
论文摘要
补充标签学习(CLL)是一个弱监督的学习问题,旨在仅从互补标签中学习多级分类器,这表明一个实例不属于的类别。现有方法主要采用简化范式对普通分类的范式,该分类应用了特定的转换和替代损失,以将CLL与普通分类联系起来。但是,这些方法面临着几个局限性,例如过度合适或挂在深层模型上的趋势。在本文中,我们以新颖的视角避开了这些局限性 - 将这些局限性减少到互补类别的概率估计中。我们证明,互补标签的准确概率估计通过一个简单的解码步骤导致良好的分类器。该证明建立了从CLL到概率估计值的还原框架。该框架提供了几种关键CLL方法作为其特殊情况的解释,并使我们能够设计一种在嘈杂环境中更强大的改进算法。该框架还提出了基于概率估计质量的验证过程,从而导致了仅使用互补标签验证模型的另一种方法。灵活的框架在使用深层和非深度模型来解决CLL问题方面开辟了广泛的未开发机会。经验实验进一步验证了该框架在各种环境中的功效和鲁棒性。
Complementary-Label Learning (CLL) is a weakly-supervised learning problem that aims to learn a multi-class classifier from only complementary labels, which indicate a class to which an instance does not belong. Existing approaches mainly adopt the paradigm of reduction to ordinary classification, which applies specific transformations and surrogate losses to connect CLL back to ordinary classification. Those approaches, however, face several limitations, such as the tendency to overfit or be hooked on deep models. In this paper, we sidestep those limitations with a novel perspective--reduction to probability estimates of complementary classes. We prove that accurate probability estimates of complementary labels lead to good classifiers through a simple decoding step. The proof establishes a reduction framework from CLL to probability estimates. The framework offers explanations of several key CLL approaches as its special cases and allows us to design an improved algorithm that is more robust in noisy environments. The framework also suggests a validation procedure based on the quality of probability estimates, leading to an alternative way to validate models with only complementary labels. The flexible framework opens a wide range of unexplored opportunities in using deep and non-deep models for probability estimates to solve the CLL problem. Empirical experiments further verified the framework's efficacy and robustness in various settings.