论文标题
dotfan:用于姿势和照明不变面识别的域转移面部增强网络
DotFAN: A Domain-transferred Face Augmentation Network for Pose and Illumination Invariant Face Recognition
论文作者
论文摘要
基于卷积神经网络(CNN)面部识别模型的性能很大程度上取决于标记的训练数据的丰富性。但是,在不同的姿势和照明变化下,收集具有较大面部身份的训练集非常昂贵,这使得内室内面部图像的多样性在实践中成为关键问题。在本文中,我们提出了一个3D模型辅助域转移的面部增强网络(DOTFAN),该网络可以基于从其他域收集的现有丰富面部数据集中蒸馏出的知识来生成一系列输入面的变体。 Dotfan在结构上是一个有条件的自行车,但还有两个子网,即面对专家网络(FEM)和面部形状回归剂(FSR),用于潜在代码控制。尽管FSR旨在提取面部属性,但FEM旨在捕获面部身份。借助他们的帮助,Dotfan可以学习一个分离的面部表现形式,并有效地产生各种面部属性的面部图像,同时保留增强脸的标识。实验表明,Dotfan有益于增强小面部数据集以改善其内部多样性,因此可以从增强数据集中学习更好的面部识别模型。
The performance of a convolutional neural network (CNN) based face recognition model largely relies on the richness of labelled training data. Collecting a training set with large variations of a face identity under different poses and illumination changes, however, is very expensive, making the diversity of within-class face images a critical issue in practice. In this paper, we propose a 3D model-assisted domain-transferred face augmentation network (DotFAN) that can generate a series of variants of an input face based on the knowledge distilled from existing rich face datasets collected from other domains. DotFAN is structurally a conditional CycleGAN but has two additional subnetworks, namely face expert network (FEM) and face shape regressor (FSR), for latent code control. While FSR aims to extract face attributes, FEM is designed to capture a face identity. With their aid, DotFAN can learn a disentangled face representation and effectively generate face images of various facial attributes while preserving the identity of augmented faces. Experiments show that DotFAN is beneficial for augmenting small face datasets to improve their within-class diversity so that a better face recognition model can be learned from the augmented dataset.