论文标题
特征削弱:分类的典型数据增强
Feature Weaken: Vicinal Data Augmentation for Classification
论文作者
论文摘要
深度学习通常依赖于培训大规模数据样本以实现更好的性能。但是,基于培训数据的过度拟合始终是一个问题。学者提出了各种策略,例如功能掉落和特征混合,以连续改善概括。出于相同的目的,我们颠覆性地提出了一种新型的培训方法,特征弱,可以将其视为数据增强方法。特征通过削弱原始样品的特征来削弱具有相同余弦相似性的阴道数据分布。尤其是,特征削弱会改变样品的空间分布,调整样品边界并降低后传播的梯度优化值。这项工作不仅可以改善模型的分类性能和概括,还可以稳定模型训练并加速模型收敛。我们对具有五个常见图像分类数据集的经典深度卷积神经模型进行了广泛的实验,并具有四个常见文本分类数据集的BERT模型。与经典模型或概括改进方法(例如辍学,混合,切割和cutmix)相比,功能弱化显示出良好的兼容性和性能。我们还使用对抗样本执行鲁棒性实验,结果表明特征弱有效地改善了模型的鲁棒性。
Deep learning usually relies on training large-scale data samples to achieve better performance. However, over-fitting based on training data always remains a problem. Scholars have proposed various strategies, such as feature dropping and feature mixing, to improve the generalization continuously. For the same purpose, we subversively propose a novel training method, Feature Weaken, which can be regarded as a data augmentation method. Feature Weaken constructs the vicinal data distribution with the same cosine similarity for model training by weakening features of the original samples. In especially, Feature Weaken changes the spatial distribution of samples, adjusts sample boundaries, and reduces the gradient optimization value of back-propagation. This work can not only improve the classification performance and generalization of the model, but also stabilize the model training and accelerate the model convergence. We conduct extensive experiments on classical deep convolution neural models with five common image classification datasets and the Bert model with four common text classification datasets. Compared with the classical models or the generalization improvement methods, such as Dropout, Mixup, Cutout, and CutMix, Feature Weaken shows good compatibility and performance. We also use adversarial samples to perform the robustness experiments, and the results show that Feature Weaken is effective in improving the robustness of the model.