论文标题
知识共享和转移的差距最小化
Gap Minimization for Knowledge Sharing and Transfer
论文作者
论文摘要
在过去的二十年中,通过知识共享和转移从多个相关任务中学习变得越来越重要。为了成功地将信息从一个任务转移到另一个任务,了解域之间的相似性和差异至关重要。在本文中,我们介绍了\ emph {performance Gap}的概念,这是对学习任务之间距离的直观而新颖的衡量标准。与现有的措施一起用作绑定任务之间预期风险差异的工具(例如$ \ Mathcal {h} $ - divergence或差异距离),我们从理论上表明,可以将性能差距视为数据和算法依赖性的正规机,从而控制模型复杂性,从而控制了模型的复杂性,并提供了良好的保证。更重要的是,它还提供了新的见解,并激发了设计知识共享和转移策略的新颖原则:差距最小化。我们用两种算法实例化了这一原理:1。Gapboost,一种新颖而有原则的增强算法,可显式地最大程度地减少用于转移学习的源和目标域之间的性能差距;和2。GAPMTNN,一种表示学习算法的表示,该算法将差距最小化为多任务学习的语义条件匹配。我们对转移学习和多任务学习基准数据集的广泛评估表明,我们的方法表现优于现有基准。
Learning from multiple related tasks by knowledge sharing and transfer has become increasingly relevant over the last two decades. In order to successfully transfer information from one task to another, it is critical to understand the similarities and differences between the domains. In this paper, we introduce the notion of \emph{performance gap}, an intuitive and novel measure of the distance between learning tasks. Unlike existing measures which are used as tools to bound the difference of expected risks between tasks (e.g., $\mathcal{H}$-divergence or discrepancy distance), we theoretically show that the performance gap can be viewed as a data- and algorithm-dependent regularizer, which controls the model complexity and leads to finer guarantees. More importantly, it also provides new insights and motivates a novel principle for designing strategies for knowledge sharing and transfer: gap minimization. We instantiate this principle with two algorithms: 1. gapBoost, a novel and principled boosting algorithm that explicitly minimizes the performance gap between source and target domains for transfer learning; and 2. gapMTNN, a representation learning algorithm that reformulates gap minimization as semantic conditional matching for multitask learning. Our extensive evaluation on both transfer learning and multitask learning benchmark data sets shows that our methods outperform existing baselines.