论文标题

表征和减轻训练物理知识的人工神经网络的困难

Characterizing and Mitigating the Difficulty in Training Physics-informed Artificial Neural Networks under Pointwise Constraints

论文作者

Basir, Shamsulhaq, Senocak, Inanc

论文摘要

神经网络可用于学习任意域上的部分微分方程(PDE)的解决方案,而无需计算网格。通用方法使用结构化损耗函数将差异操作员整合到训练神经网络中。神经网络的最常见训练算法是反向传播,它依赖于网络参数的损耗函数梯度。在这项工作中,我们通过调查差异操作员在破坏背部传播梯度的影响来培训物理网络的困难。特别是,我们表明,在训练的早期阶段,神经网络模型的输出中存在扰动导致结构化损耗函数中的噪声较高,由高阶差异操作员组成。因此,这些扰动会破坏后传播的梯度并阻碍融合。我们通过引入辅助通量参数来获得一阶微分方程系统来缓解此问题。我们使用增强的Lagrangian方法来制定非线性不受约束的优化问题,该方法适当地限制了边界条件,并适应了难以学习的较高梯度区域。我们采用我们的方法来学习各种基准PDE问题的解决方案,并证明对现有方法的数量级有所改善。

Neural networks can be used to learn the solution of partial differential equations (PDEs) on arbitrary domains without requiring a computational mesh. Common approaches integrate differential operators in training neural networks using a structured loss function. The most common training algorithm for neural networks is backpropagation which relies on the gradient of the loss function with respect to the parameters of the network. In this work, we characterize the difficulty of training neural networks on physics by investigating the impact of differential operators in corrupting the back propagated gradients. Particularly, we show that perturbations present in the output of a neural network model during early stages of training lead to higher levels of noise in a structured loss function that is composed of high-order differential operators. These perturbations consequently corrupt the back-propagated gradients and impede convergence. We mitigate this issue by introducing auxiliary flux parameters to obtain a system of first-order differential equations. We formulate a non-linear unconstrained optimization problem using the augmented Lagrangian method that properly constrains the boundary conditions and adaptively focus on regions of higher gradients that are difficult to learn. We apply our approach to learn the solution of various benchmark PDE problems and demonstrate orders of magnitude improvement over existing approaches.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源