论文标题

CLS:半监督学习的交叉标签监督

CLS: Cross Labeling Supervision for Semi-Supervised Learning

论文作者

Yao, Yao, Shen, Junyi, Xu, Jin, Zhong, Bin, Xiao, Li

论文摘要

众所周知,深神经网络的成功归因于大规模标记的数据集。但是,在大多数实际应用中收集足够的高质量标签数据可能非常耗时且辛苦。半监督学习(SSL)通过同时利用标签和未标记数据来降低标签成本的有效解决方案。在这项工作中,我们提出了交叉标记监督(CLS),该框架概括了典型的伪标记过程。基于FixMatch,其中伪标签是由弱提名的样本生成的,以教授对同一输入样本的强大增强的预测,CLS允许创建伪和互补标签,以支持正面学习和负面学习。为了减轻自我标记的确认偏差并提高对错误标签的容忍度,同时训练了两个具有相同结构的不同初始化网络。每个网络都利用来自另一个网络的高信心标签作为其他监督信号。在标签生成阶段,自适应样品权重被根据其预测置信度分配给人造标签。样本重量扮演两个角色:量化生成的标签的质量,并减少网络训练中标签不准确的破坏。半监督分类任务上的实验结果表明,我们的框架在CIFAR-10和CIFAR-100数据集上的大幅度优于现有方法。

It is well known that the success of deep neural networks is greatly attributed to large-scale labeled datasets. However, it can be extremely time-consuming and laborious to collect sufficient high-quality labeled data in most practical applications. Semi-supervised learning (SSL) provides an effective solution to reduce the cost of labeling by simultaneously leveraging both labeled and unlabeled data. In this work, we present Cross Labeling Supervision (CLS), a framework that generalizes the typical pseudo-labeling process. Based on FixMatch, where a pseudo label is generated from a weakly-augmented sample to teach the prediction on a strong augmentation of the same input sample, CLS allows the creation of both pseudo and complementary labels to support both positive and negative learning. To mitigate the confirmation bias of self-labeling and boost the tolerance to false labels, two different initialized networks with the same structure are trained simultaneously. Each network utilizes high-confidence labels from the other network as additional supervision signals. During the label generation phase, adaptive sample weights are assigned to artificial labels according to their prediction confidence. The sample weight plays two roles: quantify the generated labels' quality and reduce the disruption of inaccurate labels on network training. Experimental results on the semi-supervised classification task show that our framework outperforms existing approaches by large margins on the CIFAR-10 and CIFAR-100 datasets.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源