论文标题
图像和序列上的技巧和插件
Tricks and Plugins to GBM on Images and Sequences
论文作者
论文摘要
近年来,由多个处理层和块组成的卷积神经网络(CNN)和变压器是近年来最成功的机器学习模型。但是,数以百万计的参数和许多块使它们难以接受培训,有时需要几天或几周才能找到理想的架构或调整参数。在本文中,我们提出了一种新的算法,用于增强深度卷积神经网络(BOOSTCNN),以结合动态特征选择和BoostCNN的优点,以及另一个结合了增强和变形金刚的算法系列。为了学习这些新模型,我们介绍了子网格选择和重要性采样策略,并提出了一组算法,以将增强权重纳入基于最小二乘目标功能的深度学习体系结构中。这些算法不仅减少了找到合适的网络体系结构所需的手动努力,而且还会导致较高的性能和较低的运行时间。实验表明,所提出的方法在几个细粒分类任务上的表现优于基准。
Convolutional neural networks (CNNs) and transformers, which are composed of multiple processing layers and blocks to learn the representations of data with multiple abstract levels, are the most successful machine learning models in recent years. However, millions of parameters and many blocks make them difficult to be trained, and sometimes several days or weeks are required to find an ideal architecture or tune the parameters. Within this paper, we propose a new algorithm for boosting Deep Convolutional Neural Networks (BoostCNN) to combine the merits of dynamic feature selection and BoostCNN, and another new family of algorithms combining boosting and transformers. To learn these new models, we introduce subgrid selection and importance sampling strategies and propose a set of algorithms to incorporate boosting weights into a deep learning architecture based on a least squares objective function. These algorithms not only reduce the required manual effort for finding an appropriate network architecture but also result in superior performance and lower running time. Experiments show that the proposed methods outperform benchmarks on several fine-grained classification tasks.