论文标题
神经Bregman的分歧用于远程学习
Neural Bregman Divergences for Distance Learning
论文作者
论文摘要
许多公制的学习任务,例如三胞胎学习,最近的邻居检索和可视化,主要是将最终度量的嵌入任务视为欧几里得距离的某种变体(例如余弦或马哈拉诺省),而算法必须学会将点嵌入到预先的空间中。通常不探讨非欧国几何形状的研究,我们认为这是由于缺乏学习非欧国人距离的工具所致。最近的工作表明,可以从数据中学到Bregman的分歧,从而为学习不对称距离提供了有希望的方法。我们通过输入凸神经网络提出了一种以可区分的方式学习任意伯格曼分歧的新方法,并表明它克服了以前的工作的重大局限性。我们还证明,我们的方法更忠实地学习了一组新任务和先前研究的任务,包括不对称回归,排名和聚类。我们的测试进一步扩展到已知的不对称但非BREGMAN任务,尽管规定了错误,但我们的方法仍然具有竞争性的性能,显示了我们的不对称学习方法的一般实用性。
Many metric learning tasks, such as triplet learning, nearest neighbor retrieval, and visualization, are treated primarily as embedding tasks where the ultimate metric is some variant of the Euclidean distance (e.g., cosine or Mahalanobis), and the algorithm must learn to embed points into the pre-chosen space. The study of non-Euclidean geometries is often not explored, which we believe is due to a lack of tools for learning non-Euclidean measures of distance. Recent work has shown that Bregman divergences can be learned from data, opening a promising approach to learning asymmetric distances. We propose a new approach to learning arbitrary Bergman divergences in a differentiable manner via input convex neural networks and show that it overcomes significant limitations of previous works. We also demonstrate that our method more faithfully learns divergences over a set of both new and previously studied tasks, including asymmetric regression, ranking, and clustering. Our tests further extend to known asymmetric, but non-Bregman tasks, where our method still performs competitively despite misspecification, showing the general utility of our approach for asymmetric learning.