论文标题
隐式sindhorn差异化的统一框架
A Unified Framework for Implicit Sinkhorn Differentiation
论文作者
论文摘要
Sinkhorn运营商最近在计算机视觉和相关领域中经历了流行。一个主要原因是它易于将其集成到深度学习框架中。为了允许对各自的神经网络进行有效的培训,我们提出了一种算法,该算法通过隐式分化获得了凹痕层的分析梯度。与先前的工作相比,我们的框架是基于Sinkhorn操作员最通用的公式。它允许任何类型的损失函数,而目标容量和成本矩阵都共同区分。我们进一步构建了所得算法的误差界限,以进行近似输入。最后,我们证明,对于许多应用,只需用我们的算法替换自动分化,可以直接提高所获得梯度的稳定性和准确性。此外,我们表明它在计算上更有效,尤其是当少数GPU内存之类的资源时。
The Sinkhorn operator has recently experienced a surge of popularity in computer vision and related fields. One major reason is its ease of integration into deep learning frameworks. To allow for an efficient training of respective neural networks, we propose an algorithm that obtains analytical gradients of a Sinkhorn layer via implicit differentiation. In comparison to prior work, our framework is based on the most general formulation of the Sinkhorn operator. It allows for any type of loss function, while both the target capacities and cost matrices are differentiated jointly. We further construct error bounds of the resulting algorithm for approximate inputs. Finally, we demonstrate that for a number of applications, simply replacing automatic differentiation with our algorithm directly improves the stability and accuracy of the obtained gradients. Moreover, we show that it is computationally more efficient, particularly when resources like GPU memory are scarce.