论文标题
基于熵的分层学习模型
An Entropy-Based Model for Hierarchical Learning
论文作者
论文摘要
机器学习是人工智能的主要方法,计算机从数据和经验中学习。在监督学习的框架内,计算机必须通过学习模型为数据分布和目标功能提供有关数据分布和目标函数的辅助信息的必要性。辅助信息的这种概念与统计学习理论中的正则化概念有关。现实世界数据集中的一个共同特征是数据域是多尺度的,目标函数的行为良好且流畅。本文提出了一个基于熵的学习模型,该模型利用了此数据结构并讨论其统计和计算益处。分层学习模型的灵感来自于人类的逻辑和渐进式易于锻炼的学习机制,并且具有可解释的水平。根据数据实例和目标函数的复杂性,模型将计算资源配对。该属性可以具有多种好处,包括在培训许多用户的模型中或培训中断时的推理速度和计算节省。我们使用多尺度熵提供了学习机制的统计分析,并表明它可以比均匀的收敛范围产生明显更强的保证。
Machine learning is the dominant approach to artificial intelligence, through which computers learn from data and experience. In the framework of supervised learning, a necessity for a computer to learn from data accurately and efficiently is to be provided with auxiliary information about the data distribution and target function through the learning model. This notion of auxiliary information relates to the concept of regularization in statistical learning theory. A common feature among real-world datasets is that data domains are multiscale and target functions are well-behaved and smooth. This paper proposes an entropy-based learning model that exploits this data structure and discusses its statistical and computational benefits. The hierarchical learning model is inspired by human beings' logical and progressive easy-to-hard learning mechanism and has interpretable levels. The model apportions computational resources according to the complexity of data instances and target functions. This property can have multiple benefits, including higher inference speed and computational savings in training a model for many users or when training is interrupted. We provide a statistical analysis of the learning mechanism using multiscale entropies and show that it can yield significantly stronger guarantees than uniform convergence bounds.