论文标题
通过套索人工神经网络在非线性干草堆中寻找针的相过渡
A phase transition for finding needles in nonlinear haystacks with LASSO artificial neural networks
论文作者
论文摘要
为了适应稀疏线性关联,即使样本量小于输入矢量的维度(Haystack),在某些方面诱导较高的特征(Haystack),在某些方面诱导较高的特征(HayStack),诱导单个超参数诱导惩罚的宽带稀少度。最近,被称为人工神经网络(ANN)的学习者在许多机器学习任务(尤其是拟合非线性关联)中取得了巨大的成功。小的学习率,随机梯度下降算法和大型训练套装有助于应对深神经网络中存在的参数数量的爆炸。然而,很少开发和研究ANN学习者,以在非线性干草堆中找到针。在单个超参数的驱动下,我们的ANN学习者(例如稀疏线性关联)表现出相变的概率,我们无法与其他ANN学习者观察到。为了选择我们的惩罚参数,我们概括了Donoho和Johnstone(1994)的普遍门槛,这比保守派(太多的虚假检测)和昂贵的交叉验证更好。本着模拟退火的精神,我们提出了一种诱导算法的温暖启动稀疏性来解决高维,非凸和非差异性优化问题。我们执行精确的蒙特卡洛模拟,以显示我们方法的有效性。
To fit sparse linear associations, a LASSO sparsity inducing penalty with a single hyperparameter provably allows to recover the important features (needles) with high probability in certain regimes even if the sample size is smaller than the dimension of the input vector (haystack). More recently learners known as artificial neural networks (ANN) have shown great successes in many machine learning tasks, in particular fitting nonlinear associations. Small learning rate, stochastic gradient descent algorithm and large training set help to cope with the explosion in the number of parameters present in deep neural networks. Yet few ANN learners have been developed and studied to find needles in nonlinear haystacks. Driven by a single hyperparameter, our ANN learner, like for sparse linear associations, exhibits a phase transition in the probability of retrieving the needles, which we do not observe with other ANN learners. To select our penalty parameter, we generalize the universal threshold of Donoho and Johnstone (1994) which is a better rule than the conservative (too many false detections) and expensive cross-validation. In the spirit of simulated annealing, we propose a warm-start sparsity inducing algorithm to solve the high-dimensional, non-convex and non-differentiable optimization problem. We perform precise Monte Carlo simulations to show the effectiveness of our approach.