论文标题
逆缩放可能成为U形
Inverse scaling can become U-shaped
论文作者
论文摘要
凭经验证明,扩展语言模型可以提高各种下游任务的性能。但是,如果我们在某些任务上观察到较差的表现(“逆缩放”),这将表明缩放还可以鼓励与人类偏好失误的行为。逆缩放奖(McKenzie等,2022)确定了11个这样的反向缩放任务,对多达280B参数的模型进行了评估,最多可达500个Zettaflops训练计算。本文仔细研究了这些反向缩放任务。我们评估最多540B参数的模型,对计算的训练是倒数奖励中评估的模型的五倍。随着模型尺寸和训练计算的增加范围,这11个任务中只有四个保持倒数。十一项任务中有6个表现出“ U形缩放”,其中性能降低到一定大小,然后再次增加到评估的最大模型(其余任务显示正缩放率)。此外,我们发现1次示例和思想链可以帮助减轻不良缩放模式。 U形缩放表明,McKenzie等人观察到的反向缩放趋势。 (2022)可能无法继续使用较大的模型,我们归因于只有足够大型模型才能避免的干扰器任务。
Scaling up language models has been empirically shown to improve performance on a wide range of downstream tasks. However, if we were to observe worse performance as a function of scale ("inverse scaling") on certain tasks, this would indicate that scaling can also encourage behaviors that are misaligned with human preferences. The Inverse Scaling Prize (McKenzie et al. 2022) identified eleven such inverse scaling tasks, evaluated on models of up to 280B parameters and up to 500 zettaFLOPs of training compute. This paper takes a closer look at these inverse scaling tasks. We evaluate models of up to 540B parameters, trained on five times more compute than those evaluated in the Inverse Scaling Prize. With this increased range of model sizes and training compute, only four out of the eleven tasks remain inverse scaling. Six out of the eleven tasks exhibit "U-shaped scaling", where performance decreases up to a certain size, and then increases again up to the largest model evaluated (the one remaining task displays positive scaling). In addition, we find that 1-shot examples and chain-of-thought can help mitigate undesirable scaling patterns even further. U-shaped scaling suggests that the inverse scaling trend observed in McKenzie et al. (2022) may not continue to hold for larger models, which we attribute to the presence of distractor tasks that only sufficiently large models can avoid.