论文标题
验证图像数据的流形假设的结合
Verifying the Union of Manifolds Hypothesis for Image Data
论文作者
论文摘要
深度学习在学习高维数据的低维表示方面取得了巨大的成功。如果在感兴趣的数据中没有隐藏的低维结构,那么这一成功将是不可能的。这种存在是由歧管假设提出的,该假设指出数据在于固有维度低的未知流形。在本文中,我们认为该假设不能正确捕获图像数据中通常存在的低维结构。假设数据在于单个流形意味着在整个数据空间中的内在维度相同,并且不允许该空间的子区域具有不同数量的变异因素。为了解决这种缺陷,我们考虑了流形假设的结合,该假设指出数据在于不同固有维度的流形的不相交。我们从经验上验证了在常用图像数据集上的这一假设,发现确实观察到的数据在于断开的集合,并且内在维度不是恒定的。我们还提供了有关多种假设在深度学习中的含义的洞察力,无论是监督还是无监督,表明对这种结构具有诱导性偏见的模型可以改善分类和生成建模任务的性能。我们的代码可在https://github.com/layer6ai-labs/uomh上找到。
Deep learning has had tremendous success at learning low-dimensional representations of high-dimensional data. This success would be impossible if there was no hidden low-dimensional structure in data of interest; this existence is posited by the manifold hypothesis, which states that the data lies on an unknown manifold of low intrinsic dimension. In this paper, we argue that this hypothesis does not properly capture the low-dimensional structure typically present in image data. Assuming that data lies on a single manifold implies intrinsic dimension is identical across the entire data space, and does not allow for subregions of this space to have a different number of factors of variation. To address this deficiency, we consider the union of manifolds hypothesis, which states that data lies on a disjoint union of manifolds of varying intrinsic dimensions. We empirically verify this hypothesis on commonly-used image datasets, finding that indeed, observed data lies on a disconnected set and that intrinsic dimension is not constant. We also provide insights into the implications of the union of manifolds hypothesis in deep learning, both supervised and unsupervised, showing that designing models with an inductive bias for this structure improves performance across classification and generative modelling tasks. Our code is available at https://github.com/layer6ai-labs/UoMH.