论文标题
Lagrangian双重理论引导深度神经网络
A Lagrangian Dual-based Theory-guided Deep Neural Network
论文作者
论文摘要
理论指导的神经网络(TGNN)是一种方法,通过合并科学知识或物理信息来提高神经网络体系结构的有效性和效率。尽管它取得了巨大的成功,但该理论引导(深)神经网络在培训过程中保持训练数据和域知识之间的权衡时仍具有一定的限制。在本文中,提出了基于拉格朗日双基的TGNN(TGNN-LD)来提高TGNN的有效性。我们将原始损失函数转换为受约束的形式,其中较少的项目将部分微分方程(PDE),工程控制(ECS)和专家知识(EK)视为约束,每个约束一个Lagrangian变量。这些Lagrangian变量被合并,以实现观察数据和相应约束之间的公平权衡,以提高预测准确性,并节省按临时过程调整的时间和计算资源。为了研究所提出方法的性能,在地下流量问题上比较了具有一组优化权重值的原始TGNN模型,其L2误差,R Square(R2)和分析计算时间。实验结果证明了Lagrangian双基TGNN的优势。
The theory-guided neural network (TgNN) is a kind of method which improves the effectiveness and efficiency of neural network architectures by incorporating scientific knowledge or physical information. Despite its great success, the theory-guided (deep) neural network possesses certain limits when maintaining a tradeoff between training data and domain knowledge during the training process. In this paper, the Lagrangian dual-based TgNN (TgNN-LD) is proposed to improve the effectiveness of TgNN. We convert the original loss function into a constrained form with fewer items, in which partial differential equations (PDEs), engineering controls (ECs), and expert knowledge (EK) are regarded as constraints, with one Lagrangian variable per constraint. These Lagrangian variables are incorporated to achieve an equitable tradeoff between observation data and corresponding constraints, in order to improve prediction accuracy, and conserve time and computational resources adjusted by an ad-hoc procedure. To investigate the performance of the proposed method, the original TgNN model with a set of optimized weight values adjusted by ad-hoc procedures is compared on a subsurface flow problem, with their L2 error, R square (R2), and computational time being analyzed. Experimental results demonstrate the superiority of the Lagrangian dual-based TgNN.