论文标题

正规化还是不规律?正规化AES中的偏差差异权衡

To Regularize or Not To Regularize? The Bias Variance Trade-off in Regularized AEs

论文作者

Mondal, Arnab Kumar, Asnani, Himanshu, Singla, Parag, AP, Prathosh

论文摘要

正则化自动编码器(RAES)形成了丰富的神经生成模型。他们使用编码器组合有效地对数据和潜在空间之间的联合分布进行建模,并根据潜在空间的先验施加正则化。尽管它们具有优势,例如训练中的稳定性,但基于AE的模型的性能并未达到较高的标准,因为其他生成模型(例如生成对抗网络(GAN))的标准。在此激励的情况下,我们研究了潜在先验对本文确定性AE模型的发电质量的影响。具体而言,我们考虑使用确定性编码器对,Wasserstein自动编码器(WAE)的RAE等级,并表明具有固定的先验分布,\ textit {a a先验{a先验{a a priori},涉及“真实”潜在空间的维度,将导致所考虑的优化问题。此外,我们表明,在有限的数据制度中,尽管知道正确的潜在维度,但仍存在任何任意事先强加的偏见方差权衡权衡。作为对上述两个问题的补救措施,我们以灵活的可学习潜在先验的形式引入了一个额外的状态空间,以WAE的优化目标。我们隐含地通过AE培训共同学习潜在先验的分布,这不仅使学习目标可行,而且还促进了在偏见 - 变化曲线不同点上的操作。我们通过多个数据集上的几个实验显示了我们模型的功效,称为Flexae的功效,并证明它是基于AE的生成模型的新最新技术。

Regularized Auto-Encoders (RAEs) form a rich class of neural generative models. They effectively model the joint-distribution between the data and the latent space using an Encoder-Decoder combination, with regularization imposed in terms of a prior over the latent space. Despite their advantages, such as stability in training, the performance of AE based models has not reached the superior standards as that of the other generative models such as Generative Adversarial Networks (GANs). Motivated by this, we examine the effect of the latent prior on the generation quality of deterministic AE models in this paper. Specifically, we consider the class of RAEs with deterministic Encoder-Decoder pairs, Wasserstein Auto-Encoders (WAE), and show that having a fixed prior distribution, \textit{a priori}, oblivious to the dimensionality of the `true' latent space, will lead to the infeasibility of the optimization problem considered. Further, we show that, in the finite data regime, despite knowing the correct latent dimensionality, there exists a bias-variance trade-off with any arbitrary prior imposition. As a remedy to both the issues mentioned above, we introduce an additional state space in the form of flexibly learnable latent priors, in the optimization objective of the WAEs. We implicitly learn the distribution of the latent prior jointly with the AE training, which not only makes the learning objective feasible but also facilitates operation on different points of the bias-variance curve. We show the efficacy of our model, called FlexAE, through several experiments on multiple datasets, and demonstrate that it is the new state-of-the-art for the AE based generative models.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源