On the training and generalization of deep operator networks

09/02/2023
by   SangHyun Lee, et al.
0

We present a novel training method for deep operator networks (DeepONets), one of the most popular neural network models for operators. DeepONets are constructed by two sub-networks, namely the branch and trunk networks. Typically, the two sub-networks are trained simultaneously, which amounts to solving a complex optimization problem in a high dimensional space. In addition, the nonconvex and nonlinear nature makes training very challenging. To tackle such a challenge, we propose a two-step training method that trains the trunk network first and then sequentially trains the branch network. The core mechanism is motivated by the divide-and-conquer paradigm and is the decomposition of the entire complex training task into two subtasks with reduced complexity. Therein the Gram-Schmidt orthonormalization process is introduced which significantly improves stability and generalization ability. On the theoretical side, we establish a generalization error estimate in terms of the number of training data, the width of DeepONets, and the number of input and output sensors. Numerical examples are presented to demonstrate the effectiveness of the two-step training method, including Darcy flow in heterogeneous porous media.

READ FULL TEXT
research
08/28/2023

Solving parametric elliptic interface problems via interfaced operator network

Learning operator mapping between infinite-dimensional Banach spaces via...
research
10/08/2019

DeepONet: Learning nonlinear operators for identifying differential equations based on the universal approximation theorem of operators

While it is widely known that neural networks are universal approximator...
research
08/11/2023

Size Lowerbounds for Deep Operator Networks

Deep Operator Networks are an increasingly popular paradigm for solving ...
research
02/12/2023

Generalization Ability of Wide Neural Networks on ℝ

We perform a study on the generalization ability of the wide two-layer R...
research
07/25/2023

Modify Training Directions in Function Space to Reduce Generalization Error

We propose theoretical analyses of a modified natural gradient descent m...
research
05/14/2022

Generalization error bounds for DECONET: a deep unfolding network for analysis Compressive Sensing

In this paper, we propose a new deep unfolding neural network – based on...
research
02/28/2023

DART: Diversify-Aggregate-Repeat Training Improves Generalization of Neural Networks

Generalization of neural networks is crucial for deploying them safely i...

Please sign up or login with your details

Forgot password? Click here to reset