论文标题
导数信息神经操作员:高维参数学习的有效框架
Derivative-Informed Neural Operator: An Efficient Framework for High-Dimensional Parametric Derivative Learning
论文作者
论文摘要
我们提出了衍生化神经操作员(Dinos),即神经网络的一般家族,将运算符近似于从输入功能空间到输出功能空间或关注数量的无限维映射。离散后,输入和输出都高维。我们的目标不仅近似于具有提高精度的操作员,还要近似其在许多应用程序中的输入功能值值参数方面的衍生物(Jacobians),以增强基于衍生的算法,例如贝叶斯逆问题,在参数不确定性下的优化和最佳实验设计。主要困难包括生成衍生培训数据的计算成本以及问题的高维度,导致大量培训成本。为了应对这些挑战,我们利用了衍生物的内在低维度,并开发出用于压缩衍生信息的算法,并有效地将其强加于神经操作员训练中,从而产生了衍生化的神经操作员。我们证明,这些进步可以大大降低大量问题的数据生成和培训的成本(例如,非线性稳态参数PDE地图),使成本与成本相媲美,而无需使用衍生产品,尤其是独立于输入和输出功能的离散维度。此外,我们表明所提出的恐龙的准确性明显高于未经衍生信息的训练的神经操作员,即功能近似和衍生近似(例如,高斯 - 纽顿·黑森),尤其是当训练数据受到限制时。
We propose derivative-informed neural operators (DINOs), a general family of neural networks to approximate operators as infinite-dimensional mappings from input function spaces to output function spaces or quantities of interest. After discretizations both inputs and outputs are high-dimensional. We aim to approximate not only the operators with improved accuracy but also their derivatives (Jacobians) with respect to the input function-valued parameter to empower derivative-based algorithms in many applications, e.g., Bayesian inverse problems, optimization under parameter uncertainty, and optimal experimental design. The major difficulties include the computational cost of generating derivative training data and the high dimensionality of the problem leading to large training cost. To address these challenges, we exploit the intrinsic low-dimensionality of the derivatives and develop algorithms for compressing derivative information and efficiently imposing it in neural operator training yielding derivative-informed neural operators. We demonstrate that these advances can significantly reduce the costs of both data generation and training for large classes of problems (e.g., nonlinear steady state parametric PDE maps), making the costs marginal or comparable to the costs without using derivatives, and in particular independent of the discretization dimension of the input and output functions. Moreover, we show that the proposed DINO achieves significantly higher accuracy than neural operators trained without derivative information, for both function approximation and derivative approximation (e.g., Gauss-Newton Hessian), especially when the training data are limited.