This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Learning the Dynamical Response of Nonlinear Non-Autonomous Dynamical Systems with Deep Operator Learning Neural Networks

Guang Lin 11footnotemark: 122footnotemark: 2 [email protected] Christian Moya 33footnotemark: 3 [email protected] Zecheng Zhang 44footnotemark: 4 [email protected]
Abstract

We propose using operator learning to approximate the dynamical system response with control, such as nonlinear control systems. Unlike classical function learning, operator learning maps between two function spaces, does not require discretization of the output function, and provides flexibility in data preparation and solution prediction. Particularly, we apply and redesign the Deep Operator Neural Network (DeepONet) to recursively learn the solution trajectories of the dynamical systems. Our approach involves constructing and training a DeepONet that approximates the system’s local solution operator. We then develop a numerical scheme that recursively simulates the system’s long/medium-term dynamic response for given inputs and initial conditions, using the trained DeepONet. We accompany the proposed scheme with an estimate for the error bound of the associated cumulative error. Moreover, we propose a data-driven Runge-Kutta (RK) explicit scheme that leverages the DeepONet’s forward pass and automatic differentiation to better approximate the system’s response when the numerical scheme’s step size is small. Numerical experiments on the predator-prey, pendulum, and cart pole systems demonstrate that our proposed DeepONet framework effectively learns to approximate the dynamical response of non-autonomous systems with time-dependent inputs.

keywords:
Neural operator for dynamical systems, operator learning for dynamical systems and control, non-autonomous systems, control systems.
journal: Engineering Applications of Artificial Intelligence
\affiliation

[label1]organization=Department of Mechanical Engineering, Purdue University, addressline=610 Purdue Mall, city=West Lafayette, postcode=47907, state=IN, country=USA \affiliation[label2]organization=Department of Mathematics, Purdue University, addressline=610 Purdue Mall, city=West Lafayette, postcode=47907, state=IN, country=USA

1 Introduction

High-fidelity numerical schemes are the primary computational tools for simulating and predicting complex dynamical systems, such as climate modeling, robotics, or the modern power grid. However, these schemes may be too expensive for control, optimization, and uncertainty quantification tasks, which often require a large number of forward simulations and consume significant computational resources. Therefore, there is a growing interest in developing tools that can accelerate the numerical simulation of complex dynamical systems without compromising accuracy.

By providing faster alternatives to traditional numerical schemes, machine learning-based computational tools hold the promise of accelerating the rate of innovation for complex dynamical systems. Hence, a recent wave of machine and deep learning-based tools has demonstrated the potential of using observational data to construct fast surrogates of complex systems Efendiev et al. (2021); Qin et al. (2021) and providing efficient solutions to complex engineering tasks such as fault diagnosis Choudhary et al. (2023); Mishra et al. (2022b, a) and power engineering Moya and Lin (2023); Moya et al. (2023). These tools aim to (i) learn the governing equations of a dynamical system or (ii) learn to predict the system’s response from data.

On the one hand, several works Brunton et al. (2016a, b); Schaeffer (2017); Sun et al. (2020) use observational data to discover the unknown governing equations of the underlying system. For example, Brunton et al.  Brunton et al. (2016a) proposed identifying nonlinear systems using sparse schemes and a large set of dictionaries. They then extended their work in Brunton et al. (2016b) to identify input/output mappings that describe control systems. In Sun et al. (2020), the authors used sparse approximation schemes to recover the governing partial or ordinary differential equations of unknown systems.

On the other hand, there is growing interest in predicting the future response of a dynamical system using time-series data Qin et al. (2019); Raissi et al. (2018); Qin et al. (2021); Proctor et al. (2018). For instance, the authors of Efendiev et al. (2021); Chung et al. (2021) trained a transformer with data from the early stages of time-dependent partial differential equations to recursively predict the solution of future stages. Qin et al.  Qin et al. (2019) used a recurrent residual neural network (ResNet) to approximate the mapping from the current state to the next state of an unknown autonomous system. Similarly, in Raissi et al. (2018), the authors used feed-forward neural networks (FNN) as a building block of a multi-step scheme that predicts the response of an autonomous system. In Qin et al. (2021), the authors extended their previous work Qin et al. (2019) to non-autonomous systems with time-dependent inputs. To this end, they parametrized the input locally within small time intervals.

Many of the works mentioned above require large amounts of training data to avoid overfitting. However, acquiring this data may be prohibitively expensive for complex dynamical systems. Additionally, if one fails to avoid overfitting when using traditional neural networks for next-state methods, the predicted response may drift and accumulate errors over time. It is therefore crucial to develop deep learning-based frameworks that can efficiently handle the infinite-dimensional nature of predicting the response of non-autonomous systems with time-dependent inputs for long time horizons, as studied in this paper.

Operator Learning Chen and Chen (1995); Lu et al. (2019); Li et al. (2020); Zhang et al. (2022) was first addressed by the seminal work of Chen and Chen (1995). Unlike classical function learning, operator learning involves approximating an operator between two function spaces. Recently, the seminal paper Lu et al. (2021) extended the work of Chen and Chen (1995) and proposed the Deep Operator Neural Network (DeepONet) framework. DeepONet exhibits small generalization errors and learns with limited training data Zhang et al. (2022). This has been demonstrated in many application areas, such as power engineering Moya et al. (2023), multi-physics problems Cai et al. (2021), and turbulent combustion problems Ranade et al. (2021). Extensions to the original DeepONet Lu et al. (2021) have enabled the incorporation of physics-informed models Wang et al. (2021b), handling noisy data Lin et al. (2023), and designing novel optimization methods Lin et al. (2023, 2021).

DeepONet is a prediction-free and domain-free tool Zhang et al. (2022). Specifically, DeepONet can predict the target function value at any point in its domain, and the input and output functions are not necessarily required to share the same domain. This property allows for handling incomplete datasets during the training phase. In Zhang et al. (2022), it was further relaxed the assumptions on input function discretization, improving flexibility in data preparation and prediction accuracy for operator learning.

In this paper, we focus on extending the This paper focuses on extending the original DeepONet framework to learn the solution operator of non-autonomous systems with time-dependent inputs for long-time horizons. By using such a data-driven operator framework, control policies can be designed for continuous nonlinear systems, among other applications.

In particular, one of our motivations behind learning to approximate the solution operator of a non-autonomous system is its application in control, and Model-Based Reinforcement Learning (MBRL) Wang et al. (2019); Sutton (1990). In MBRL, one learns to approximate the system to control (from data) and then uses the learned model to seek an optimal policy without extensive interaction with the actual system. A common approach is to learn a discrete-time forward model that predicts the next state using the current state and selected control action. Such an MBRL framework has delivered successful and efficient results in discrete-time problems such as games Kaiser et al. (2019). However, most control systems in science and engineering (e.g., robotics Nagabandi et al. (2020), unmanned vehicles Hafner et al. (2019), or laminar flows Fan et al. (2020)) are continuous. Of course, one can always discretize the continuous dynamics and apply discrete-time MBRL using traditional neural networks (e.g., see Brockman et al. (2016)). However, if one fails to handle the inherent epistemic uncertainty Deisenroth et al. (2013), such a strategy may lead to error accumulation (due to model bias) and poor asymptotic performance. To alleviate these drawbacks, we build on the original DeepONet Lu et al. (2021) to design an effective and efficient framework that can learn the solution operator of a nonlinear non-autonomous system with time-dependent inputs, which one can then apply within the framework of control or continuous MBRL Du et al. (2020).

Formally, the objectives of this paper are twofold.

  1. 1.

    Approximation of the local solution operator: Our goal is to create a neural operator-based framework that can learn to map (i) the current state of the dynamical system with control and (ii) a local approximation of the input to the next state of the non-autonomous dynamical system.

  2. 2.

    Long/Medium-term simulation: We aim to design an efficient scheme that uses the operator learning framework to simulate the dynamical system’s response to a given input over a long or medium-term horizon.

The contributions that will achieve the above objectives are summarized below.

  1. 1.

    We model the non-autonomous/control dynamical system in the deep operator learning framework and propose a DeepONet-based numerical scheme that effectively and recursively simulates the solution of the system.

  2. 2.

    We also propose a novel data-driven Runge-Kutta (RK) DeepONet-based method that reduces the cumulative error and improves recursive prediction accuracy in long-term simulations.

  3. 3.

    We provide an estimation of the cumulative error for the proposed numerical scheme based on DeepONet. Our estimated error bounds are tighter than those presented in Qin et al. (2021). Additionally, we demonstrate that the novel data-driven RK DeepONet-based scheme achieves better stability bounds than the DeepONet-based scheme.

  4. 4.

    We test the proposed frameworks on state-of-the-art models, such as the predator-prey system, the pendulum, the cart-pole system, and a power engineering task. The effectiveness of the methods is observed in all experiments.

We organize the rest of this paper as follows. In Section 2, we review operator learning and introduce the proposed DeepONet-based algorithms. Next, we present the novel data-driven Runge-Kutta DeepONet-based scheme and estimate its corresponding error bound in Section 3. We present the numerical experiments in Section 4. We provide a discussion of our results and future work in Section 5, and conclude the paper in Section 6.

2 Problem Formulation

We consider the problem of learning from data the solution operator of the following continuous-time non-autonomous system with time-dependent inputs

ddtx(t)=f(x(t),u(t)),t[a,b]x(a)=x0,\displaystyle\begin{aligned} \frac{d}{dt}x(t)&=f(x(t),u(t)),\quad t\in[a,b]\\ x(a)&=x_{0},\end{aligned} (1)

where x(t)𝒳nx(t)\in\mathcal{X}\subseteq\mathbb{R}^{n} is the state vector, u(t)𝒰pu(t)\in\mathcal{U}\subseteq\mathbb{R}^{p} the input vector, and f:𝒳×𝒰𝒳f:\mathcal{X}\times\mathcal{U}\to\mathcal{X} an unknown function. Additional assumptions on the input function u and the function ff will be discussed later. Also, throughout this paper, we assume u(t)u(t) is a scalar input, i.e., p=1p=1. Extending the proposed framework to the vector-valued case is straightforward.

Solution Operator. Let \mathcal{F} denote the solution operator of (1). \mathcal{F} takes as input the initial condition x(a)=x0𝒳x(a)=x_{0}\in\mathcal{X} and the sequence of input function values u[a,t):={u(τ)𝒰:τ[a,t)}u_{[a,t)}:=\{u(\tau)\in\mathcal{U}:\tau\in[a,t)\}, and outputs the state x(t)𝒳x(t)\in\mathcal{X} at time t[a,b]t\in[a,b]. We compute the solution operator via

(x0,u[a,t))(t)x(t)=x0+atf(x(s),u(s))𝑑s.\displaystyle\mathcal{F}\left(x_{0},u_{[a,t)}\right)(t)\equiv x(t)=x_{0}+\int_{a}^{t}f(x(s),u(s))ds. (2)

Approximate System. In practice, to learn the operator \mathcal{F}, we only have access to an approximate/discretized representation of the input function u(t)u(t). Let umu_{m} denote this approximate representation, which yields the following approximate system

ddtx~(t)=f(x~(t),um),t[a,b]x~(a)=x(a),\displaystyle\begin{aligned} \frac{d}{dt}\tilde{x}(t)&=f(\tilde{x}(t),u_{m}),\qquad t\in[a,b]\\ \tilde{x}(a)&=x(a),\end{aligned} (3)

whose solution operator is

(x0,um)(t)x~(t)=x0+atf(x~(s),um)𝑑s.\displaystyle\mathcal{F}\left(x_{0},u_{m}\right)(t)\equiv\tilde{x}(t)=x_{0}+\int_{a}^{t}f(\tilde{x}(s),u_{m})ds. (4)

In the above, with a slight abuse of notation, we denoted umu_{m} as the input discretized using m1m\geq 1 sensors or interpolation points within the interval [a,b][a,b], and the approximate state function as x~(t)\tilde{x}(t).

Remark.

In practice, the approximate system (3) with the solution operator (4) can represent, for example, sampled-data control systems Wittenmark et al. (2002) or semi-Markov Decision Processes Du et al. (2020). Therefore, the methods introduced in this paper can be used to design optimal control policies within the framework of model-based reinforcement learning Wang et al. (2019).

2.1 Learning the Solution Operator

To learn the solution operator \mathcal{F}, we use the Deep Operator Network (DeepONet) framework introduced in Lu et al. (2021). In Lu et al. (2021), the authors used a DeepONet GθG_{\theta} to learn a simplified version of the solution operator \mathcal{F} (2) with x(a)=0x(a)=0, i.e.,

G(u)(t)x(t)=atf(x(s),u(s))𝑑s,t[a,b].G(u)(t)\equiv x(t)=\int_{a}^{t}f(x(s),u(s))ds,\qquad t\in[a,b].

The DeepONet GθG_{\theta} takes as inputs the (i) input u(t)u(t) discretized using mm interpolation points (known as sensors in Lu et al. (2019)) and (ii) time t[a,b]t\in[a,b], and outputs the state x(t)x(t). Clearly, this DeepONet prediction is one-shot; it requires knowledge of the input in the whole interval [a,b][a,b]. For small values of bb, the DeepONet’s GθG_{\theta^{*}} prediction is very accurate. However, as we increase bb, the accuracy deteriorates. To improve accuracy, one can increase the number of interpolation points. This, however, makes DeepONet’s training more challenging.

To alleviate this drawback, we take a different approach in this paper. First, we train a DeepONet θ\mathcal{F}_{\theta}, with the vector of trainable parameters θ\theta, to learn the local solution operator of the system. Then, we design a DeepONet-based numerical scheme that recursively uses the trained DeepONet θ\mathcal{F}_{\theta^{*}} to predict long/medium-term horizons, i.e., for large values of bb.

Learning the Local Solution Operator. We let 𝒫\mathcal{P} denote the possibly irregular and arbitrary time partition

𝒫:a=t0<t1<<tM=b,\mathcal{P}:\leavevmode\nobreak\ a=t_{0}<t_{1}<\ldots<t_{M}=b,

where hn:=tn+1tnh_{n}:=t_{n+1}-t_{n} for all n=0,1,,M1n=0,1,\ldots,M-1 and let h:=maxnhnh:=\max_{n}h_{n}. Then, within the local interval [tn,tn+1][tn,tn+hn][t_{n},t_{n+1}]\equiv[t_{n},t_{n}+h_{n}], the solution operator is given by

(x~(tn),umn)(hn)x~(tn+hn)=x~(tn)+tntn+hnf(x~(s),umn)𝑑s.\displaystyle\begin{aligned} \mathcal{F}(\tilde{x}(t_{n}),u^{n}_{m})(h_{n})&\equiv\tilde{x}(t_{n}+h_{n})\\ &=\tilde{x}(t_{n})+\int_{t_{n}}^{t_{n}+h_{n}}f(\tilde{x}(s),u^{n}_{m})ds.\end{aligned} (5)

In the above, with a slight abuse of notation, we use umnu^{n}_{m} to denote the local discretized representation of the input function umu_{m}, within the interval [tn,tn+1][t_{n},t_{n+1}], using ns1n_{s}\geq 1 interpolation/sensor points or basis.

The DeepONet θ\mathcal{F}_{\theta}. Next, we design a Deep Operator Network (DeepONet) θ\mathcal{F}_{\theta}, with the vector of trainable parameters θ\theta, to approximate the local solution operator (5). Figure 1 illustrates the proposed DeepONet θ\mathcal{F}_{\theta} that has two neural networks: the Branch Net and the Trunk Net.

Refer to caption
Figure 1: The DeepONet architecture for learning the solution operator of non-autonomous system with time-dependent inputs. The Branch Net takes as input a vector resulting from concatenating (i) the current state x~(tn)n\tilde{x}(t_{n})\in\mathbb{R}^{n}, (ii) the discretized local input umnnsu_{m}^{n}\in\mathbb{R}^{n_{s}}, and (iii) the relative/flexible sensor locations (tnd0n,,tndns1n)ns(t_{n}-d_{0}^{n},\ldots,t_{n}-d_{{n_{s}}-1}^{n})^{\top}\in\mathbb{R}^{n_{s}}. Then, the Branch Net outputs the vector of features βnq\beta\in\mathbb{R}^{nq}. The Trunk Net takes as the input hn(0,h]h_{n}\in(0,h] and outputs the vector of features τnq\tau\in\mathbb{R}^{nq}. Finally, we obtain the DeepONet using the dot product between β\beta and τ\tau.

The Branch Net maps the vector that concatenates the (i) current state x~(tn)\tilde{x}(t_{n}) and (ii) input function umnnsu^{n}_{m}\in\mathbb{R}^{n_{s}}, discretized using the mesh of local sensors tn=d0n<d1n<<dns1n=tn+1t_{n}=d_{0}^{n}<d_{1}^{n}<...<d_{n_{s}-1}^{n}=t_{n+1}, to the branch output feature vector βnq\beta\in\mathbb{R}^{nq}. Meanwhile, the Trunk Net maps the scalar step size hn(0,h]h_{n}\in(0,h] to the trunk output feature vector τnq\tau\in\mathbb{R}^{nq}. We compute the DeepONet’s output for the iith component of the state vector x~(tn+1)\tilde{x}(t_{n+1}) using the dot product:

θ(i)(x~(tn),umn)(hn)=k=1qβ(i1)q+kτ(i1)q+k.\mathcal{F}^{(i)}_{\theta}(\tilde{x}(t_{n}),u^{n}_{m})(h_{n})=\sum_{k=1}^{q}\beta_{(i-1)q+k}\tau_{(i-1)q+k}.
Remark.

Sensor locations. One of the problems with the original DeepONet Lu et al. (2019) is that the sensor locations, tn=d0n<d1n<<dns1n=tn+1t_{n}=d_{0}^{n}<d_{1}^{n}<...<d_{n_{s}-1}^{n}=t_{n+1}, are fixed. If we fix these sensor locations, then we cannot predict the response of the system using the arbitrary and irregular partition 𝒫\mathcal{P}. To enable prediction with 𝒫\mathcal{P}, we let the input to the branch umnu^{n}_{m} be the concatenation of the discretized input (u(d0n),,u(dns1n))\left(u(d_{0}^{n}),\ldots,u(d_{{n_{s}}-1}^{n})\right) with the corresponding (flexible) relative sensor locations (tnd0n,,tndns1n)(t_{n}-d_{0}^{n},\ldots,t_{n}-d_{{n_{s}}-1}^{n}).

Remark.

The case of ns=1n_{s}=1 sensors, i.e., the piece-wise constant approximation of uu. If we let ns=1n_{s}=1 sensor, then the discretized input function within the interval [tn,tn+1][t_{n},t_{n+1}] corresponds to the singleton umnu(tn)u^{n}_{m}\equiv u(t_{n}). Such a case is the most challenging to learn because it introduces the largest error that propagates over time. However, it also represents one of the target applications: continuous model-based reinforcement learning with semi-Markov decision processes. We will show (in Section 4) that the proposed DeepONet can effectively handle having only ns=1n_{s}=1 sensor.

Training the DeepONet θ\mathcal{F}_{\theta}. We train the proposed DeepONet θ\mathcal{F}_{\theta} model by minimizing the loss function

(θ;𝒟)=1Ni=1Nxi(tn+hni)θ(xi(tn),umi,n)(hni)22\mathcal{L}(\theta;\mathcal{D})=\frac{1}{N}\sum_{i=1}^{N}||x^{i}(t_{n}+h_{n}^{i})-\mathcal{F}_{\theta}(x^{i}(t_{n}),u^{i,n}_{m})(h^{i}_{n})||_{2}^{2}

over of NN training data triplets

𝒟={(xi(tn),umi,n),hni,xi(tn+hni)}i=1N\mathcal{D}=\{(x^{i}(t_{n}),u^{i,n}_{m}),h_{n}^{i},x^{i}(t_{n}+h^{i}_{n})\}_{i=1}^{N}

generated by the unknown ground truth local solution operator \mathcal{F}.

2.2 Predicting the System’s Response

We predict the response of the system (1) over a long/medium-term horizon (i.e., within the interval [a,b][a,b], with b1b\gg 1) using the DeepONet-based numerical scheme detailed in Algorithm 1 and illustrated in Figure 2. Algorithm 1 takes as inputs the (i) initial condition x(a)=x0x(a)=x_{0}, (ii) partition 𝒫\mathcal{P}, (iii) discretized representation of the input umnu^{n}_{m}, for n=1,,M1n=1,\ldots,M-1, and (iv) trained DeepONet θ\mathcal{F}_{\theta{{}^{*}}}. Then, Algorithm 1 outputs the predicted response of the system over the partition 𝒫\mathcal{P}, i.e., {xˇnxˇ(tn):tn𝒫}\{\check{x}_{n}\equiv\check{x}(t_{n}):t_{n}\in\mathcal{P}\}. Let us conclude this section by estimating a bound for the cumulative error of the proposed DeepONet-based numerical scheme described in Algorithm 1.

1 Require: initial state vector x(a)=x0x(a)=x_{0}, partition 𝒫\mathcal{P}, input umnu^{n}_{m}, for n=1,,M1n=1,\ldots,M-1, and trained DeepONet θ\mathcal{F}_{\theta^{*}}.
2 initialize xˇ(t0)=x0\check{x}(t_{0})=x_{0}
3 for n=0,,M1n=0,\ldots,M-1 do
4       update the independent variable tn+1=tn+hnt_{n+1}=t_{n}+h_{n}
5       update the state vector using the DeepONet’s forward pass
xˇ(tn+1)=θ(xˇ(tn),umn)(hn).\check{x}(t_{n+1})=\mathcal{F}_{\theta^{*}}(\check{x}(t_{n}),u^{n}_{m})(h_{n}).
6 end for
7Return: predicted response {xˇnxˇ(tn):tn𝒫}\{\check{x}_{n}\equiv\check{x}(t_{n}):t_{n}\in\mathcal{P}\}.
Algorithm 1 DeepONet-based Numerical Scheme
Trained recursive DeepONetθ(xˇ(tn),umn)(hn)\mathcal{F}_{\theta^{*}}\left(\check{x}(t_{n}),u_{m}^{n}\right)(h_{n})xˇ(tn+hn)\check{x}(t_{n}+h_{n})xˇ(tn)\check{x}(t_{n})umnu_{m}^{n}hnh_{n}InputsPredicted state
Figure 2: The trained DeepONet θ\mathcal{F}_{\theta^{*}} recursively predicts the next state of a non-autonomous system xˇ(tn+hn)\check{x}(t_{n}+h_{n}) given the current state xˇ(tn)\check{x}(t_{n}), the local approximation of the input umnu_{m}^{n}, and the step-size hnh_{n}.

2.3 Error Bound for the DeepONet-based Numerical Scheme

Assumptions. We let the input function uVC[a,b]u\in V\subset C[a,b] where VV is compact. We also assume the unknown vector field f:𝒳×𝒰𝒳f:\mathcal{X}\times\mathcal{U}\to\mathcal{X} is Lipschitz in xx and uu, i.e.,

f(x1,u)f(x2,u)C1x1x2,\displaystyle\|f(x_{1},u)-f(x_{2},u)\|\leq C_{1}\|x_{1}-x_{2}\|,
f(x,u1)f(x,u2)C1u1u2,\displaystyle\|f(x,u_{1})-f(x,u_{2})\|\leq C_{1}\|u_{1}-u_{2}\|,

where C1>0C_{1}>0 is a constant, and x1,x2,u1,u2x_{1},x_{2},u_{1},u_{2} are in some proper space. Such an assumption is common in engineering, as ff is often differentiable with respect to xx and uu.

The following lemma, presented in Qin et al. (2021), provides us with an alternative form of the local solution operator (4). Such a form will be used later when estimating the error bound.

Lemma 2.1.

Consider the local solution operator (x~(tn),umn)(hn)\mathcal{F}(\tilde{x}(t_{n}),u_{m}^{n})(h_{n}). Then, there exists a function Φ:n×ns×n\Phi:\mathbb{R}^{n}\times\mathbb{R}^{n_{s}}\times\mathbb{R}\to\mathbb{R}^{n}, which depends on ff, such that

x~(tn+1)=(x~(tn),umn)(hn)=Φ(x~(tn),umn,hn),\displaystyle\tilde{x}(t_{n+1})=\mathcal{F}(\tilde{x}(t_{n}),u_{m}^{n})(h_{n})=\Phi(\tilde{x}(t_{n}),u_{m}^{n},h_{n}), (6)

for any tn𝒫t_{n}\in\mathcal{P}. In the above, umnnsu^{n}_{m}\in\mathbb{R}^{n_{s}} locally characterizes umu_{m} on the mesh tn=d0n<d1n<<dns1n=tn+1t_{n}=d_{0}^{n}<d_{1}^{n}<\ldots<d_{n_{s}-1}^{n}=t_{n+1}.

We now provide the error estimation of the proposed DeepONet prediction scheme detailed in Algorithm 1. For the ease of notation, we denote xn=x(tn)x_{n}=x(t_{n}) and xˇn=xˇ(tn)\check{x}_{n}=\check{x}(t_{n}) for all tn𝒫t_{n}\in\mathcal{P}.

Lemma 2.2.

For any tn𝒫t_{n}\in\mathcal{P}, we have

xnx~n1C¯n1C¯e¯(um):=C~e¯m,\displaystyle\|x_{n}-\tilde{x}_{n}\|\leq\frac{1-\bar{C}^{n}}{1-\bar{C}}\bar{e}(u_{m}):=\tilde{C}\bar{e}_{m}, (7)

where C¯=eC1h\bar{C}=e^{C_{1}h}, e¯(um)=maxne¯n(um)\bar{e}(u_{m})=\max_{n}\bar{e}_{n}(u_{m}), and e¯n(um)=C1hnκn(m)eC1hn\bar{e}_{n}(u_{m})=C_{1}h_{n}\kappa_{n}(m)e^{C_{1}h_{n}}.

Proof.

Let a=tna=t_{n} and s[tn,tn+1]s\in[t_{n},t_{n+1}] in the solution operators (2) and (4). Then, subtracting (4) from (2) gives

x(s)x~(s)\displaystyle\|x(s)-\tilde{x}(s)\| xnx~n+tnsf(x(t),u(t))f(x~(t),um)𝑑t\displaystyle\leq\|x_{n}-\tilde{x}_{n}\|+\int_{t_{n}}^{s}\|f(x(t),u(t))-f(\tilde{x}(t),u_{m})\|dt
xnx~n+C1tnsu(t)um𝑑t+C1tnsx(t)x~(t)𝑑t\displaystyle\leq\|x_{n}-\tilde{x}_{n}\|+C_{1}\int_{t_{n}}^{s}\|u(t)-u_{m}\|dt+C_{1}\int_{t_{n}}^{s}\|x(t)-\tilde{x}(t)\|dt
xnx~n+C1hnκn(m)+C1tnsx(t)x~(t)𝑑t,\displaystyle\leq\|x_{n}-\tilde{x}_{n}\|+C_{1}h_{n}\kappa_{n}(m)+C_{1}\int_{t_{n}}^{s}\|x(t)-\tilde{x}(t)\|dt,

where κn(m)\kappa_{n}(m) is the local approximation error of the input within the interval [tn,tn+1][t_{n},t_{n+1}]:

maxs[tn,tn+1]|u(s)um|κn(m),\displaystyle\max_{s\in[t_{n},t_{n+1}]}|u(s)-u_{m}|\leq\kappa_{n}(m),

such that

κn(m)0 when the number of sensorsm.\kappa_{n}(m)\rightarrow 0\text{ when the number of sensors}\leavevmode\nobreak\ m\to\infty.

We refer the interested reader to equation (4) of Lu et al. (2019) for details of such an approximation for the input. Set s=tn+1s=t_{n+1} and apply Gronwall’s inequality, we then have,

xn+1x~n+1\displaystyle\|x_{n+1}-\tilde{x}_{n+1}\| xnx~neC1hn+C1hnκn(m)eC1hne¯n(um).\displaystyle\leq\|x_{n}-\tilde{x}_{n}\|e^{C_{1}h_{n}}+\underbrace{C_{1}h_{n}\kappa_{n}(m)e^{C_{1}h_{n}}}_{\bar{e}_{n}(u_{m})}. (8)

Taking e¯(um)=maxne¯n(um)\bar{e}(u_{m})=\max_{n}\bar{e}_{n}(u_{m}) gives

xn+1x~n+1xnx~neC1hn+e¯(um).\displaystyle\|x_{n+1}-\tilde{x}_{n+1}\|\leq\|x_{n}-\tilde{x}_{n}\|e^{C_{1}h_{n}}+\bar{e}(u_{m}).

The bound then follows immediately due to x(t0)=x~0x(t_{0})=\tilde{x}_{0}. ∎

Before we estimate the cumulative error of the DeepONet-assisted solution xˇn\check{x}_{n}, we review the universal approximation theorem of neural networks for high-dimensional functions Cybenko (1989). To this end, given hnh_{n}, we define the following vector-valued continuous function φ:n×nsn\varphi:\mathbb{R}^{n}\times\mathbb{R}^{n_{s}}\to\mathbb{R}^{n}

φ(yn,umn)=(yn,umn)(hn)=Φ(yn,umn,hn),\displaystyle\varphi(y_{n},u_{m}^{n})=\mathcal{F}(y_{n},u_{m}^{n})(h_{n})=\Phi(y_{n},u_{m}^{n},h_{n}),

where ynny_{n}\in\mathbb{R}^{n}. Then, by the universal approximation theorem, for e¯(um)>0\bar{e}(u_{m})>0, there exist W1K×(n+ns)W_{1}\in\mathbb{R}^{K\times(n+n_{s})}, b1Kb_{1}\in\mathbb{R}^{K}, W2n×KW_{2}\in\mathbb{R}^{n\times K} and b2nb_{2}\in\mathbb{R}^{n} such that

φ(yn,umn)(W2σ(W1[yn,umn]+b1)+b2)<e¯(um).\displaystyle\bigg{\|}\varphi(y_{n},u_{m}^{n})-\left(W_{2}\sigma(W_{1}[y_{n},u_{m}^{n}]^{\top}+b_{1})+b_{2}\right)\bigg{\|}<\bar{e}(u_{m}). (9)

Here, the two-layer network represents the DeepONet for a given hnh_{n}, i.e.,

(W2σ(W1[yn,umn]+b1)+b2)θ(yn,umn).\left(W_{2}\sigma(W_{1}[y_{n},u_{m}^{n}]^{\top}+b_{1})+b_{2}\right)\equiv\mathcal{F}_{\theta^{*}}(y_{n},u_{m}^{n}).

The following lemma estimates the cumulative error between the DeepONet-assisted solution xˇ\check{x} (obtained via Algorithm 1) and the solution x~\tilde{x} of the approximate system that satisfies (6).

Lemma 2.3.

Assume Φ\Phi is Lipschitz in xx with Lipschitz constant C2C_{2}. Suppose the DeepONet is well trained so that the network satisfies (9). Then, we have the following estimate:

xˇnx~nCˇe¯(um),\displaystyle\|\check{x}_{n}-\tilde{x}_{n}\|\leq\check{C}\bar{e}(u_{m}), (10)

where Cˇ=1C2n1C2\check{C}=\frac{1-C_{2}^{n}}{1-C_{2}}.

Proof.

It follows from the universal approximation theorem of neural networks (9) and Φ\Phi being Lipschitz that,

xˇn+1x~n+1)\displaystyle\|\check{x}_{n+1}-\tilde{x}_{n+1})\| =θ(xˇn,umn)Φ(x~n,umn,hn)\displaystyle=\|\mathcal{F}_{\theta^{*}}(\check{x}_{n},u_{m}^{n})-\Phi(\tilde{x}_{n},u_{m}^{n},h_{n})\|
=θ(xˇn,umn)φ(xˇn,umn)\displaystyle=\|\mathcal{F}_{\theta^{*}}(\check{x}_{n},u_{m}^{n})-\varphi(\check{x}_{n},u_{m}^{n})\|
+Φ(xˇn,umn,hn)Φ(x~n,umn,hn)\displaystyle\qquad+\|\Phi(\check{x}_{n},u_{m}^{n},h_{n})-\Phi(\tilde{x}_{n},u_{m}^{n},h_{n})\|
e¯(um)+C2xˇnx~n.\displaystyle\leq\bar{e}(u_{m})+C_{2}\|\check{x}_{n}-\tilde{x}_{n}\|.

The result follows immediately from x~0=xˇ0\tilde{x}_{0}=\check{x}_{0}. ∎

The following theorem summarizes the error of the proposed DeepONet scheme.

Theorem 2.4.

For any tn𝒫t_{n}\in\mathcal{P}, we have

xnxˇnC~e¯(um)+Cˇe¯(um),\displaystyle\|x_{n}-\check{x}_{n}\|\leq\tilde{C}\bar{e}(u_{m})+\check{C}\bar{e}(u_{m}), (11)

where C~\tilde{C}, Cˇ\check{C} and e¯(um)\bar{e}(u_{m}) are, respectively, the constants defined in (7), (10), and (7).

We conclude this section by observing that the error bound found in this section is tighter than the error found in Qin et al. (2021), which behaves like tectte^{ct} where cc is a positive constant.

3 Data-Driven Runge-Kutta DeepONet Prediction Scheme

In this section, we propose a data-driven Runge-Kutta explicit DeepONet-based scheme Iserles (2009) that predicts the new state vector x^(tn+hn)\hat{x}(t_{n}+h_{n}) using the current state value x^(tn)\hat{x}(t_{n}), i.e.,

x^(tn+hn)=x^(tn)+hn2(k1+k2).\hat{x}(t_{n}+h_{n})=\hat{x}(t_{n})+\frac{h_{n}}{2}(k_{1}+k_{2}).

Here k1k_{1} and k2k_{2} are, respectively, the estimates of ff at tnt_{n} and tn+1t_{n+1}. We compute these estimates (see equation (13)) using (i) the forward pass of trained DeepONet θ\mathcal{F}_{\theta^{*}} and (ii) automatic differentiation. Note that in (13b), we use the notation x¯(tn+1)\bar{x}(t_{n+1}) for the estimate of the state at tn+1t_{n+1} obtained using the DeepONet’s θ\mathcal{F}_{\theta^{*}} forward pass.

We detail the proposed data-driven RK explicit DeepONet-based scheme in Algorithm 2 and Figure 3. Two remarks about our algorithm are provided next. (i) For simplicity, we only present our scheme for the improved Euler method or RK-2 Iserles (2009). However, we remark that we can extend our idea to any explicit Runge-Kutta scheme. (ii) If umn+1u_{m}^{n+1} is available at tnt_{n}, then we can compute k2k_{2} as follows:

k2=ddt(θ(xˇ(tn),umn+1)(hn)).\displaystyle k_{2}=\frac{d}{dt}(\mathcal{F}_{\theta^{*}}(\check{x}(t_{n}),u^{n+1}_{m})(h_{n})). (12)

Then, equations (13a) and (12) will work as a predictor-corrector scheme with the updated input information. Other strategies can also be adopted within the proposed RK scheme. However, we let the design of such strategies for our future work.

1 Require: initial state vector x(a)=x0x(a)=x_{0}, partition 𝒫\mathcal{P}, input umnu^{n}_{m}, for n=1,,M1n=1,\ldots,M-1, and trained DeepONet θ\mathcal{F}_{\theta^{*}}.
2 initialize xˇ(t0)=x0\check{x}(t_{0})=x_{0}
3 for n=0,,M1n=0,\ldots,M-1 do
4       update the independent variable tn+1=tn+hnt_{n+1}=t_{n}+h_{n}
5       use the DeepONet’s forward pass to compute
x¯(tn+1)=θ(xˇ(tn),umn)(hn).\bar{x}(t_{n+1})=\mathcal{F}_{\theta^{*}}(\check{x}(t_{n}),u^{n}_{m})(h_{n}).
6 end for
7
8 use automatic differentiation to estimate the vector field ff at t=tnt=t_{n} and t=tn+1t=t_{n+1}
k1\displaystyle k_{1} =ddt(θ(xˇ(tn),umn)(0))=f(x^(tn),umn),\displaystyle=\frac{d}{dt}(\mathcal{F}_{\theta^{*}}(\check{x}(t_{n}),u^{n}_{m})(0))=f(\hat{x}(t_{n}),u_{m}^{n}), (13a)
k2\displaystyle k_{2} =ddtx¯(tn+1)=f(x¯(tn+1),umn).\displaystyle=\frac{d}{dt}\bar{x}(t_{n+1})=f(\bar{x}(t_{n+1}),u_{m}^{n}). (13b)
9 update the state vector with the improved Euler (RK-2) step
xˇ(tn+1)=xˇ(tn)+hn2(k1+k2).\displaystyle\check{x}(t_{n+1})=\check{x}(t_{n})+\frac{h_{n}}{2}(k_{1}+k_{2}). (14)
10 Return: predicted response {xˇ(tn):tn𝒫}\{\check{x}(t_{n}):t_{n}\in\mathcal{P}\}.
Algorithm 2 Data-Driven Runge-Kutta Scheme
Data-driven RK DeepONetx¯(tn+1)=θ(xˇ(tn),umn)(hn)\overline{x}(t_{n+1})=\mathcal{F}_{\theta^{*}}\left(\check{x}(t_{n}),u_{m}^{n}\right)(h_{n})xˇ(tn)\check{x}(t_{n})umnu_{m}^{n}hnh_{n}InputsPredicted statek1=ddt(θ(xˇ(tn),umn)(0))k_{1}=\frac{d}{dt}\left(\mathcal{F}_{\theta^{*}}\left(\check{x}(t_{n}),u_{m}^{n}\right)(0)\right)k2=ddtx¯(tn+1)k_{2}=\frac{d}{dt}\overline{x}(t_{n+1})xˇ(tn+1)=xˇ(tn)+hn2(k1+k2)\check{x}(t_{n+1})=\check{x}(t_{n})+\frac{h_{n}}{2}(k_{1}+k_{2})Automatic differentiation
Figure 3: The data-driven RK DeepONet takes the current state xˇ(tn)\check{x}(t_{n}), the local approximation of the input umnu_{m}^{n}, and the step-size hnh_{n} as inputs. It then recursively predicts the next state of a non-autonomous system using a data-driven RK-2 method, which employs the forward pass of the DeepONet and automatic differentiation to obtain the estimates of the vector field k1k_{1} and k2k_{2}.

3.1 Error Bound for the Data-Driven Runge-Kutta Scheme

Here we derive an improved conditional error bound estimate for x^(tn)\hat{x}(t_{n}). To that end, we start by rephrasing the universal approximation theorem of neural network for high-dimensional functions Cybenko (1989), which we introduced in Section 2.3. For e¯(um)C4hn2>0\bar{e}(u_{m})-C_{4}h_{n}^{2}>0, there exist W1K×(n+ns)W_{1}\in\mathbb{R}^{K\times(n+n_{s})}, b1Kb_{1}\in\mathbb{R}^{K}, W2n×KW_{2}\in\mathbb{R}^{n\times K} and b2nb_{2}\in\mathbb{R}^{n} such that

φ(yn,umn)(W2σ(W1[yn,umn]+b1)+b2)<ϵ,\displaystyle\bigg{\|}\varphi(y_{n},u_{m}^{n})-\left(W_{2}\sigma(W_{1}[y_{n},u_{m}^{n}]^{\top}+b_{1})+b_{2}\right)\bigg{\|}<\epsilon, (15)

where ϵ:=e¯(um)C4h2\epsilon:=\bar{e}(u_{m})-C_{4}h^{2} and C4>0C_{4}>0 is constant. As before, the two-layer network represents the DeepONet θ\mathcal{F}_{\theta^{*}} for a given hnh_{n}.

The following lemma estimates the error between x^\hat{x}, predicted using the RK scheme (Algorithm 2), and x~\tilde{x}, obtained using the solution operator (4) of the approximate system (3).

Lemma 3.1.

Assume Φ\Phi is Lipschitz in xx with Lipschitz constant C2C_{2}. Suppose the DeepONet is well trained so that (15) holds. Then, we have the estimate

x^(tn)x~(tn)C^e¯(um),\displaystyle\|\hat{x}(t_{n})-\tilde{x}(t_{n})\|\leq\hat{C}\bar{e}(u_{m}), (16)

where C^=C1h21C3n1C3\hat{C}=\frac{C_{1}h}{2}\frac{1-C_{3}^{n}}{1-C_{3}}, and C3=(1+(1+C2)C1h2)C_{3}=\left(1+(1+C_{2})\frac{C_{1}h}{2}\right).

Proof.

We denote x~n=x~(tn)\tilde{x}_{n}=\tilde{x}(t_{n}), x¯n=x¯(tn)\bar{x}_{n}=\bar{x}(t_{n}), and x^n=x^(tn)\hat{x}_{n}=\hat{x}(t_{n}). Then, it follows from ff being Lipschitz that,

x~n+1x^n+1\displaystyle\|\tilde{x}_{n+1}-\hat{x}_{n+1}\| =x~n+tntn+1f(x~(s),umn)dsx^n\displaystyle=\|\tilde{x}_{n}+\int_{t_{n}}^{t_{n+1}}f(\tilde{x}(s),u_{m}^{n})ds-\hat{x}_{n}
hn2(f(x¯n+1,umn)+f(x^n,umn))\displaystyle\qquad-\frac{h_{n}}{2}(f(\bar{x}_{n+1},u_{m}^{n})+f(\hat{x}_{n},u_{m}^{n}))\|
x~nx^n+hn2f(x~n,umn)f(x^n,umn)\displaystyle\leq\|\tilde{x}_{n}-\hat{x}_{n}\|+\frac{h_{n}}{2}\|f(\tilde{x}_{n},u_{m}^{n})-f(\hat{x}_{n},u_{m}^{n})
+f(x~n+1,umn)f(x¯n+1,umn)+𝒪(h3)\displaystyle\qquad+f(\tilde{x}_{n+1},u_{m}^{n})-f(\bar{x}_{n+1},u_{m}^{n})\|+{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\mathcal{O}(h^{3})}
x~nx^n+C1hn2x~nx^n\displaystyle\leq\|\tilde{x}_{n}-\hat{x}_{n}\|+\frac{C_{1}h_{n}}{2}\|\tilde{x}_{n}-\hat{x}_{n}\|
+C1hn2x~n+1x¯n+1+𝒪(h3)\displaystyle\qquad+\frac{C_{1}h_{n}}{2}\|\tilde{x}_{n+1}-\bar{x}_{n+1}\|{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}+\mathcal{O}(h^{3})}
(1+C1hn2)x~nx^n\displaystyle\leq\left(1+\frac{C_{1}h_{n}}{2}\right)\|\tilde{x}_{n}-\hat{x}_{n}\|
+C1hn2x~n+1x¯n+1+𝒪(h3).\displaystyle\qquad+\frac{C_{1}h_{n}}{2}\|\tilde{x}_{n+1}-\bar{x}_{n+1}\|+\mathcal{O}(h^{3}). (17)

An estimate of the above term x~n+1x¯n+1\|\tilde{x}_{n+1}-\bar{x}_{n+1}\| is

x~n+1x¯n+1\displaystyle\|\tilde{x}_{n+1}-\bar{x}_{n+1}\| =Φ(x~n,umn,hn)θ(x^n,umn)\displaystyle=\|\Phi(\tilde{x}_{n},u_{m}^{n},h_{n})-\mathcal{F}_{\theta^{*}}(\hat{x}_{n},u_{m}^{n})\|
Φ(x~n,umn,hn)Φ(x^n,umn,hn)\displaystyle\leq\|\Phi(\tilde{x}_{n},u_{m}^{n},h_{n})-\Phi(\hat{x}_{n},u_{m}^{n},h_{n})
+φ(x^n,umn)θ(x^n,umn)\displaystyle\qquad\leavevmode\nobreak\ \quad+\varphi(\hat{x}_{n},u_{m}^{n})-\mathcal{F}_{\theta^{*}}(\hat{x}_{n},u_{m}^{n})\|
C2x~nx^n+e¯(um)C4h2,\displaystyle\leq C_{2}\|\tilde{x}_{n}-\hat{x}_{n}\|+\bar{e}(u_{m}){\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}-C_{4}h^{2}}, (18)

where C4>0C_{4}>0 is a constant. Substituting (18) back into (17) yields

x~n+1x^n+1\displaystyle\|\tilde{x}_{n+1}-\hat{x}_{n+1}\| (1+(1+C2)C1hn2)x~nx^n\displaystyle\leq\left(1+(1+C_{2})\frac{C_{1}h_{n}}{2}\right)\|\tilde{x}_{n}-\hat{x}_{n}\|
+C1hn2e¯(um).\displaystyle\qquad\leavevmode\nobreak\ \qquad+\frac{C_{1}h_{n}}{2}\bar{e}(u_{m}). (19)

Recursive estimation and x^0=x~0\hat{x}_{0}=\tilde{x}_{0} gives the desired error bound (16). ∎

Two remarks about the error bound are as follows. (i) Note that the proposed data-driven RK DeepONet-based scheme provides an improved error bound (16) when compared to the bound obtained in (10). More specifically, the growth factor C3C_{3} behaves like C2C_{2} in (10). However, when hn<1C12C22C2+1h_{n}<\frac{1}{C_{1}}\cdot\frac{2C_{2}-2}{C_{2}+1}, a smaller factor can be derived. (ii) We can extend the proof provided here for the RK-2 scheme to any other RK explicit scheme. We will analyze this in our future work. Let us conclude this section with the following theorem that summarizes the error of the proposed data-driven RK scheme.

Theorem 3.2.

For any tn𝒫t_{n}\in\mathcal{P}, we have

x(tn)x^(tn)C~e¯(um)+C^e¯(um),\displaystyle\|x(t_{n})-\hat{x}(t_{n})\|\leq\tilde{C}\bar{e}(u_{m})+\hat{C}\bar{e}(u_{m}), (20)

where C~\tilde{C}, C^\hat{C} and e¯(um)\bar{e}(u_{m}) are, respectively, the constants defined in (7), (16), and (7).

4 Numerical Experiments

To evaluate our framework, we tested the proposed DeepONets on five tasks: the autonomous Lorentz 63 system (in Section 4.1), the predator-prey dynamics with control (in Section 4.2), the pendulum swing-up (in Section 4.3), the cart-pole system (in Section 4.4), and a power engineering application (in Section 4.5). For the three control tasks, we used only ns=1n_{s}=1 sensor. The reasons for selecting only one sensor are twofold. First, we want to show that DeepONet is effective even when the input signal is encoded with minimal information. For reference, in Qin et al. (2021), the authors encoded the input signals (used in their experiments) with at least ns=3n_{s}=3 interpolation points (sensors). Second, the ns=1n_{s}=1 sensor scenario resembles the scenario of sampled-data control systems Wittenmark et al. (2002) or reinforcement learning tasks Sutton and Barto (2018) with continuous action space. Finally, for the power engineering application task, we used ns=2n_{s}=2 sensors.

Training dataset. For each of the three continuous control tasks (predator-prey, pendulum, and cart-pole systems), we generate the training dataset 𝒟train\mathcal{D}_{\text{train}} as follows. We use Runge-Kutta (RK-4) Iserles (2009) to simulate NtrainN_{\text{train}} trajectories of size two. For each trajectory, the input to RK-4 is the initial condition x(tn)x(t_{n}) uniformly sampled from 𝒳\mathcal{X} and the input u(tn)u(t_{n}) uniformly sampled from the set 𝒰\mathcal{U}. The output from the RK-4 algorithm is the state x(tn+h)x(t_{n}+h), where hh is uniformly sampled from the interval [0,0.25][0,0.25]. Such a procedure gives the dataset:

𝒟train={(xi(tn),ui(tn)),hi,xi(tn+hi)}i=1Ntrain.\mathcal{D}_{\text{train}}=\{(x_{i}(t_{n}),u_{i}(t_{n})),h_{i},x_{i}(t_{n}+h_{i})\}_{i=1}^{N_{\text{train}}}.

Training protocol and neural networks. We implemented our framework using JAX Bradbury et al. (2018)555We will publish the code on GitHub after publication. The neural networks for the Branch and Trunk Nets are the modified fully-connected networks proposed in Wang et al. (2021a) and used in our previous paper Moya et al. (2023). We trained the parameters of the networks using Adam Kingma and Ba (2014). Moreover, we selected (i) the default hyper-parameters for the Adam algorithm and (ii) an initial learning rate of η=0.001\eta=0.001 that exponentially decays every 2000 epochs.

4.1 The Autonomous Lorentz 63 System

To evaluate the proposed framework, we first consider the autonomous and chaotic Lorenz 63 system with the following dynamics:

x˙=σ(yx),y˙=x(ρz)y,z˙=xyβz,\displaystyle\begin{aligned} \dot{x}&=\sigma(y-x),\\ \dot{y}&=x(\rho-z)-y,\\ \dot{z}&=xy-\beta z,\end{aligned} (21)

with parameters σ=10,ρ=28,\sigma=10,\leavevmode\nobreak\ \rho=28, and β=8/3.\beta=8/3. Notice that, compared to non-autonomous systems, autonomous systems only require the previous state to predict the next state, simplifying the learning problem. Nevertheless, we do recognize that the chaotic nature of the Lorentz system can pose a challenge to error accumulation over time. To address this, we have arbitrarily increased the size of the training dataset, which is described next.

The training dataset for this autonomous system is provided as a collection of 20000 scatter one-step responses, or trajectories of size two. Specifically, Dtrain={xi(tn),hi,xi(tn+hi)}i=1NtrainD_{\text{train}}=\{x_{i}(t_{n}),h_{i},x_{i}(t_{n}+h_{i})\}_{i=1}^{N_{\text{train}}}, where xi(tn)x_{i}(t_{n}) is sampled from the state space 𝒳:=[17,20]×[23,28]×[0,50]\mathcal{X}:=[-17,20]\times[-23,28]\times[0,50], and the step sizes hih_{i} are sampled from the interval [0,0.02][0,0.02]. Note that we keep the step size small to track the error accumulation of the chaotic system, but this requires us to increase the size of our training dataset and normalize the state space.

The trained DeepONet is used to predict the response of the autonomous Lorentz 63 system over the time-domain [0,20][0,20] seconds, for the initial condition (x(0),y(0),z(0))=(0,1,1)(x(0),y(0),z(0))=(0,1,1), using a uniform partition 𝒫\mathcal{P} with a fixed step size of h=0.01h=0.01. Figure 4 shows the comparison between the recursive DeepONet prediction obtained from Algorithm 1 and the true response of the Lorentz system’s state variables x(t)x(t) and y(t)y(t). The results demonstrate excellent agreement between the proposed method and the true values, despite the chaotic nature of the autonomous Lorentz system.

Refer to caption
Refer to caption
Figure 4: Comparison of DeepONet prediction with the actual trajectory of the autonomous Lorentz (21) system’s state x=(x(t),y(t))x=(x(t),y(t))^{\top} (left is x(t)x(t) and right is y(t)y(t)) for the initial condition (x(0),y(0),z(0))=(0,1,1)(x(0),y(0),z(0))=(0,1,1) within the partition 𝒫[0,20]\mathcal{P}\subset[0,20] (s) of constant step size h=0.01h=0.01.

4.2 The Predator-Prey Dynamics with Control

To evaluate our framework, we now consider the following Lotka-Volterra Predator-Prey system with input signal u(t)u(t):

x˙1=x1x1x2+u(t)x˙2=x2+x1x2.\displaystyle\begin{aligned} \dot{x}_{1}&=x_{1}-x_{1}x_{2}+u(t)\\ \dot{x}_{2}&=-x_{2}+x_{1}x_{2}.\end{aligned} (22)

The system (22) was also studied in Qin et al. (2021), where the authors encoded u(t)u(t) using three interpolation points. To train our DeepONet, we generated Ntrain=2000N_{\text{train}}=2000 trajectories with the initial condition xi(tn)x_{i}(t_{n}) (resp. input signal ui(tn)u_{i}(t_{n})) sampled from the state space 𝒳:=[0,5]2\mathcal{X}:=[0,5]^{2} (resp. input space 𝒰:=[0,5]\mathcal{U}:=[0,5]).

We use the trained DeepONet to predict the response of the predator-prey system (see equation (22)) to the input signal u(t)=sin(t/3)+cos(t)+2u(t)=\sin(t/3)+\cos(t)+2 within a partition 𝒫[0,100]\mathcal{P}\subset[0,100] (s) with a constant step size h=0.1tn+1tnh=0.1\equiv t_{n+1}-t_{n}, where tn,tn+1𝒫t_{n},t_{n+1}\in\mathcal{P}. Figure 5 compares DeepONet’s long-term prediction with the true trajectory. Note that the predicted trajectory agrees very well with the true trajectory for both states, x=(x1,x2)x=(x_{1},x_{2})^{\top}. The L2L_{2}-relative errors for x1x_{1} and x2x_{2} are summarized in Table 1.

x1x_{1} x2x_{2}
L2L_{2} relative error 2.42 % 0.93 %
Table 1: Relative errors of predator and prey dynamics with control.
Refer to caption
Refer to caption
Figure 5: Comparison of DeepONet prediction with the actual trajectory of the predator-prey (22) system’s state x=(x1(t),x2(t))x=(x_{1}(t),x_{2}(t))^{\top} (left is x1(t)x_{1}(t) and right is x2(t)x_{2}(t)) response to the input signal u(t)=sin(t/3)+cos(t)+2u(t)=\sin(t/3)+\cos(t)+2 within the partition 𝒫[0,100]\mathcal{P}\subset[0,100] (s) of constant step size h=0.1h=0.1.

4.3 Pendulum Swing-Up

Let us now consider the following pendulum swing-up control system:

θ¨(14ml2+I)+12mlgsinθ=u(t)bθ˙,\displaystyle\ddot{\theta}\left(\frac{1}{4}ml^{2}+I\right)+\frac{1}{2}mlg\sin\theta=u(t)-b\dot{\theta}, (23)

where x=(θ,θ˙)𝒳x=(\theta,\dot{\theta})^{\top}\in\mathcal{X} is the state vector, θ\theta the pendulum’s angle, θ˙\dot{\theta} the angular velocity, and u(t)𝒰u(t)\in\mathcal{U} the control torque. We set the parameters to the following values. The pendulum’s mass is m=1m=1 (kg), the length is l=1l=1 (m), the moment of inertia of the pendulum around the midpoint is I=112ml2I=\frac{1}{12}ml^{2}, and the friction coefficient is b=0.01b=0.01 (sNm/rad).

We trained the proposed framework using Ntrain=5000N_{\text{train}}=5000 samples, each consisting of a tuple {(x(tn),u(tn)),hn,x(tn+hn)}\{(x(t_{n}),u(t_{n})),h_{n},x(t_{n}+h_{n})\}. The initial condition, x(tn)x(t_{n}) (the state of the non-autonomous pendulum system), was sampled from the uniform distribution 𝒳:=[π,π]×[8,8]\mathcal{X}:=[-\pi,\pi]\times[-8,8]. The step size hnh_{n} was obtained by uniformly sampling from the interval (0,0.02](0,0.02]. The control torque, u(tn)u(t_{n}), was constant on the interval [tn,tn+1][t_{n},t_{n+1}] and was sampled from the uniform distribution 𝒰:=[2,2]\mathcal{U}:=[-2,2]. Finally, we obtained the terminal state, x(tn+hn)x(t_{n}+h_{n}), by evolving according to the non-autonomous dynamics (see equation (23)). It is worth noting that the the proposed operator learning setting allows us to handle datasets with incomplete data, as hn=tn+1tnh_{n}=t_{n+1}-t_{n} is not identical for all 50005000 training samples. Our methods thus have much more flexibility in terms of data preparation.

Stable response. We begin by using the trained DeepONet to predict the pendulum’s response to the input u1(t)=0.8θ˙(t)u_{1}(t)=-0.8\dot{\theta}(t) within the partition 𝒫[0,10]\mathcal{P}\subset[0,10] (s) with a step size of h=0.1h=0.1 (s). This input yields state trajectories {(θ(tn),θ˙(tn)):tn𝒫}\{(\theta(t_{n}),\dot{\theta}(t_{n})):t_{n}\in\mathcal{P}\} that settle to an asymptotic equilibrium point. Figure 6 shows the excellent agreement between the predicted and actual trajectories.

Refer to caption
Refer to caption
Figure 6: Comparison of DeepONet prediction with the actual trajectories of the pendulum (23) system’s state x=(θ(t),θ˙(t))x=(\theta(t),\dot{\theta}(t))^{\top} (left is angle θ(t)\theta(t) and right is velocity θ˙(t)\dot{\theta}(t)) response to the input signal u(t)=0.8θ˙(t)u(t)=-0.8\dot{\theta}(t) within the partition 𝒫[0,10]\mathcal{P}\subset[0,10] (s) of constant step size h=0.1h=0.1.

To test the predictive power of the proposed framework, we compute the average and standard deviation (st. dev.) of the L2L_{2}-relative error between the predicted and actual response of the pendulum to the following: (i) the control torque u1(t)u_{1}(t), and (ii) 100100 initial conditions sampled from the set 𝒳o:={θ,θ˙:θ[π/2,π/2],θ˙=0}\mathcal{X}_{o}:=\{\theta,\dot{\theta}:\theta\in[-\pi/2,\pi/2],\dot{\theta}=0\}. Table 2 reports the results, demonstrating that DeepONet consistently maintains an average L2L_{2}-relative error of below 1.1%1.1\% and 3.4%3.4\% for the angle θ(t)\theta(t) and the angular velocity θ˙(t)\dot{\theta}(t) trajectories, respectively. Please refer to Table 2 for a summary.

angle θ(t)\theta(t) angular velocity θ˙(t)\dot{\theta}(t)
mean L2L_{2} 1.056 % 3.356 %
st.dev. L2L_{2} 2.509 % 8.234%
Table 2: The average and standard deviation (st.dev.) of the L2L_{2}- relative error between the predicted and actual response trajectories of the pendulum system (23) to (i) the control torque u1(t)=0.8θ(t)˙u_{1}(t)=-0.8\dot{\theta(t)} and (ii) 100100 initial conditions uniformly sampled from the set 𝒳o:={θ,θ˙:θ[π/2,π/2],θ˙=0}\mathcal{X}_{o}:=\{\theta,\dot{\theta}:\theta\in[-\pi/2,\pi/2],\dot{\theta}=0\}.

Oscillatory response. We now test the trained DeepONet for the control torque u2(t)=sin(t/2)u_{2}(t)=\sin(t/2) within the partition 𝒫[0,10]\mathcal{P}\subset[0,10] (s) with a constant step size of h=0.1h=0.1. Figure 7 depicts the excellent agreement between the predicted and actual oscillatory trajectory.

Refer to caption
Refer to caption
Figure 7: Comparison of DeepONet prediction with the actual trajectories of the pendulum (23) system’s state x=(θ(t),θ˙(t))x=(\theta(t),\dot{\theta}(t))^{\top} (left is angle θ(t)\theta(t) and right is velocity θ˙(t)\dot{\theta}(t)) response to the input signal u(t)=sin(t/2)u(t)=\sin(t/2) within the partition 𝒫[0,10]\mathcal{P}\subset[0,10] (s) of constant step size h=0.1h=0.1. The relative errors are presented in Table 4.

Let us now consider a different partition 𝒫\mathcal{P}^{\prime} with a smaller step size h=0.00025h=0.00025. Figure 8 shows that the proposed DeepONet fails to keep up with the oscillatory response. To improve our prediction, we employ the proposed data-driven Runge-Kutta DeepONet method described in Algorithm 2. Figure 8 depicts the agreement between the actual trajectory and the trajectory predicted using the data-driven RK DeepONet method. The corresponding L2L_{2}-relative errors are 7.73%7.73\% and 11.34%11.34\% for the angle θ(t)\theta(t) and the angular velocity θ˙(t)\dot{\theta}(t), respectively.

Refer to caption
Refer to caption
Figure 8: Comparison of DeepONet prediction with the data-driven RK DeepONet prediction, and the actual trajectories of the pendulum (23) system’s state x=(θ(t),θ˙(t))x=(\theta(t),\dot{\theta}(t))^{\top} (left is angle θ(t)\theta(t) and right is velocity θ˙(t)\dot{\theta}(t)) response to the input signal u(t)=sin(t/2)u(t)=\sin(t/2) within the partition 𝒫[0,10]\mathcal{P}^{\prime}\subset[0,10] (s) of constant step size h=0.00025h=0.00025.

Comparison with other benchmarks. We compare the proposed DeepONet numerical scheme (Algorithm 1) with other benchmarks. The authors are only aware of the work developed in Qin et al. (2021), which can approximate the dynamic response of continuous nonlinear non-autonomous systems over an arbitrary time partition 𝒫[0,T]\mathcal{P}\subset[0,T] for arbitrary input functions. Below, we will compare our method with the approach in Qin et al. (2021), which we refer to as FNN.

Additionally, a related field that attempts to learn an environment model of a discrete, non-linear, non-autonomous system is model-based reinforcement learning (MBRL) Wang et al. (2019). However, since they can only approximate the response over a uniform partition, a comparison of traditional MBRL benchmarks with the proposed work is not possible. To address this problem, we transform the ensemble method Wang et al. (2019), a simple but effective MBRL method to learn the environment dynamics in MBRL, into a continuous approach using ideas from Qin et al. (2021). Thus, we also compare the proposed method with this continuous ensemble approach.

We begin by comparing the average and standard deviation of the L2L_{2} relative error between the predicted response from DeepONet, FNN, and Ensemble models. These models were all trained using Ntrain=10000N_{\text{train}}=10000 one-step responses, and compared with the actual response trajectories of the pendulum system (23). The pendulum trajectories are the responses to (i) the control torque u1(t)=0.8θ(t)˙u_{1}(t)=-0.8\dot{\theta(t)} and (ii) 100100 initial conditions uniformly sampled from the set 𝒳o:={θ,θ˙:θ[π/2,π/2],θ˙=0}\mathcal{X}_{o}:=\{\theta,\dot{\theta}:\theta\in[-\pi/2,\pi/2],\dot{\theta}=0\}. Table 3 shows that DeepONet greatly improves generalization when compared with FNN and the continuous ensemble.

angle θ(t)\theta(t) angular velocity θ˙(t)\dot{\theta}(t)
mean L2L_{2} (DeepONet) 0.833 % 1.061 %
st.dev. L2L_{2} (DeepONet) 0.604 % 0.756%
mean L2L_{2} (FNN) 10.076 % 13.318 %
st.dev. L2L_{2} (FNN) 7.626 % 9.817%
mean L2L_{2} (Ensemble) 5.820 % 7.055 %
st.dev. L2L_{2} (Ensemble) 3.817 % 4.654%
Table 3: The average and standard deviation (st.dev.) of the L2L_{2}- relative error between the predicted (DeepONet, FNN, and Ensemble) and actual response trajectories of the pendulum system (23) to (i) the control torque u1(t)=0.8θ(t)˙u_{1}(t)=-0.8\dot{\theta(t)} and (ii) 100100 initial conditions uniformly sampled from the set 𝒳o:={θ,θ˙:θ[π/2,π/2],θ˙=0}\mathcal{X}_{o}:=\{\theta,\dot{\theta}:\theta\in[-\pi/2,\pi/2],\dot{\theta}=0\}.

In addition, we compared the predictions of DeepONet, FNN, and Ensemble with the actual trajectories of the pendulum (23) system’s state x=(θ(t),θ˙(t))x=(\theta(t),\dot{\theta}(t))^{\top}. We analyzed the response to the input signal u(t)=sin(t/2)u(t)=\sin(t/2) within the partition 𝒫[0,10]\mathcal{P}\subset[0,10] (s) with a constant step size of h=0.1h=0.1. Figure 9 (the left plot of Figure 9 shows the angle θ(t)\theta(t), and the right plot shows the velocity θ˙(t)\dot{\theta}(t)) illustrates that only DeepONet can accurately predict oscillatory behavior.

Refer to caption
Refer to caption
Figure 9: Comparison of DeepONet, FNN, and Ensemble prediction with the actual trajectories of the pendulum (23) system’s state x=(θ(t),θ˙(t))x=(\theta(t),\dot{\theta}(t))^{\top} (left is angle θ(t)\theta(t) and right is velocity θ˙(t)\dot{\theta}(t)) response to the input signal u(t)=sin(t/2)u(t)=\sin(t/2) within the partition 𝒫[0,10]\mathcal{P}\subset[0,10] (s) of constant step size h=0.1h=0.1.

4.4 Cart-Pole

In this section, we consider the following cart-pole system Florian (2007) with control:

θ¨=gsinθ+cosθ(u(t)mplθ˙2sinθmc+mp)l(43mpcos2θmc+mp)p¨=u(t)bp˙+mpl(θ˙2sinθθ¨cosθ)mc+mp.\displaystyle\begin{aligned} \ddot{\theta}&=\frac{g\sin\theta+\cos\theta\left(\frac{-u(t)-m_{p}l\dot{\theta}^{2}\sin\theta}{m_{c}+m_{p}}\right)}{l\left(\frac{4}{3}-\frac{m_{p}\cos^{2}\theta}{m_{c}+m_{p}}\right)}\\ \ddot{p}&=\frac{u(t)-b\dot{p}+m_{p}l(\dot{\theta}^{2}\sin\theta-\ddot{\theta}\cos\theta)}{m_{c}+m_{p}}.\end{aligned} (24)

In the above, the state of the cart-pole system is x=(θ,θ˙,p,p˙)𝒳x=(\theta,\dot{\theta},p,\dot{p})^{\top}\in\mathcal{X}, where θ\theta is the angle of the pendulum, θ˙\dot{\theta} the angular velocity of the pendulum, pp and p˙\dot{p} are, respectively, the position and the velocity of the cart. The input u𝒰u\in\mathcal{U} is the horizontal force that makes the cart move to the left or the right. We selected the parameters of the cart-pole system as follows. The pendulum’s length is l=0.5l=0.5 (m), the pendulum’s mass is mp=0.5m_{p}=0.5 (kg), the cart’s mass is mc=0.5m_{c}=0.5 (kg), and the friction coefficient is b=0.01b=0.01 (sNm/rad).

To train the DeepONet, we generated Ntrain=20000N_{\text{train}}=20000 trajectories with the initial condition xi(tn)x_{i}(t_{n}) (resp. input ui(tn)u_{i}(t_{n})) sampled from the state space 𝒳:=[2π,2π]×[π,π]×[2,2]×[1,1]\mathcal{X}:=[-2\pi,2\pi]\times[-\pi,\pi]\times[-2,2]\times[-1,1] (resp. 𝒰:=[5,5]\mathcal{U}:=[-5,5]).

Refer to caption
Refer to caption
Refer to caption
Refer to caption
Figure 10: Comparison of DeepONet prediction with the actual trajectory of the cart-pole (24) system’s state x=(θ(t),θ˙(t),p(t),p˙(t))x=(\theta(t),\dot{\theta}(t),p(t),\dot{p}(t))^{\top} (upper-left is angle θ(t)\theta(t), upper-right is angular velocity θ˙(t)\dot{\theta}(t), lower-left is position p(t)p(t), and lower-right is velocity p˙(t)\dot{p}(t)) response to the input signal u(t)=t/100u(t)=t/100 within the partition 𝒫[0,10]\mathcal{P}\subset[0,10] (s) of constant step size h=0.1h=0.1.

We used a trained DeepONet to predict the response of the cart-pole system to a time-dependent input signal u(t)=t/100u(t)=t/100 within the partition 𝒫[0,10]\mathcal{P}\subset[0,10] (s) with a constant step size of h=0.1h=0.1 (i.e., tn+1tn=ht_{n+1}-t_{n}=h) for all tn,tn+1𝒫t_{n},t_{n+1}\in\mathcal{P}. In Figure 10, we compare the predicted and true state trajectories. We observe a high degree of agreement between the predicted and true state trajectories, with L2L_{2}-relative errors of 0.008%0.008\%, 1.028%1.028\%, 0.478%0.478\%, and 0.296%0.296\% for θ\theta, θ˙\dot{\theta}, pp, and p˙\dot{p}, respectively. Based on these results, we conclude that the proposed DeepONet framework is effective in predicting the response of a nonlinear control system, such as the cart-pole, to given inputs and different initial conditions.

θ\theta θ˙\dot{\theta} pp p˙\dot{p}
L2L_{2} relative error 0.0080.008 % 1.0281.028% 0.478%0.478\% 0.296%0.296\%
Table 4: Relative errors of the cart-pole example. The solutions are plotted in Figure 10.

4.5 A Power Engineering Application

The dynamics of the future power grid will fundamentally differ from today’s grid due to widespread installation of energy storage systems, offshore wind power plants, and electric vehicle fast-charging sites. These devices will connect to the power grid via power electronic devices. This scenario differs greatly from the current paradigm where robust models of traditional power systems exist. Thus, simulating future power grids for planning, optimization, and control will require the interplay of legacy simulators and models of power electronics-based devices (e.g., wind turbines) that can be learned from data.

In this final experiment, we aim to demonstrate that the proposed method can potentially interact with a traditional power engineering simulator, such as the Power System Toolbox. To achieve this, we begin with a simple application of recursive DeepONet that uses Algorithm 1 to approximate the dynamic response of a simple second order model of a generator. We use data collected with the Power System Toolbox (PST) Chow et al. (2000) of generator 1 within the classic two area system. Note that the response of the generator can be modeled as a non-autonomous system (1) by assuming the input function corresponds to the interface variables (stator currents) of the bus where the target generator connects.

Training data. We generated training data 𝒟PST\mathcal{D}_{\text{PST}} by simulating Nexp=600N_{\text{exp}}=600 disturbance experiments on PST. Each experiment consisted of simulating the two area system system using a uniform partition 𝒫[0.0,5.0]\mathcal{P}\subset[0.0,5.0] (s), with a constant step size of h=0.005h=0.005. The disturbances occur at tf=0.01t_{f}=0.01 (s) with a duration sampled from the interval [0.005,0.02][0.005,0.02] (s). Trajectory data was collected after each experiment, including the interacting input trajectories {u(tn):tn𝒫}\{u(t_{n}):t_{n}\in\mathcal{P}\} and state trajectory data {x(tn):tn𝒫}\{x(t_{n}):t_{n}\in\mathcal{P}\}.

We constructed the training dataset by interpolating and sampling this trajectory data (i.e., a time-series). In particular, we discretized the inputs using interpolation and m=2m=2 sensors, i.e., u~mn:={u(tn+d0),u(tn+d1))}\tilde{u}^{n}_{m}:=\{u(t_{n}+d_{0}),u(t_{n}+d_{1}))\}, where d0=0.0d_{0}=0.0 and d1d_{1} was uniformly sampled from the open interval (0,h)(0,h). The final training dataset of size Ntrain=30000N_{\text{train}}=30000 is: 𝒟PST={(xi(tn),u~m,in),{0,dm,i},hi,xi(tn+hi)}i=1Ntrain.\mathcal{D}_{\text{PST}}=\{(x_{i}(t_{n}),\tilde{u}^{n}_{m,i}),\{0,d_{m,i}\},h_{i},x_{i}(t_{n}+h_{i})\}_{i=1}^{N_{\text{train}}}.

We tested the proposed DeepONet using a PST test trajectory that was not included in the training dataset and which experienced a disturbance of duration 0.01 (s). For these test trajectory, we used an uniform partition 𝒫\mathcal{P} of size 100100. As shown in Figure 11, DeepONet-based Algorithm 1 accurately predicted the dynamic response of generator 1 in the classic two area system.

Refer to caption
Refer to caption
Figure 11: Comparison of DeepONet prediction with the actual trajectory of a second-order electromagnetic generator model state x=(δ(t),ω(t))x=(\delta(t),\omega(t))^{\top} (left is δ(t)\delta(t) angle and right is ω(t)\omega(t) speed) for a disturbance of duration 0.010.01 (s) within a uniform partition 𝒫[0,5]\mathcal{P}\subset[0,5] (s) size 100100.

In our future work, we plan to (i) extend this approach to allow for the full interaction of DeepONets and traditional numerical schemes, (ii) enable transfer learning for approximating more than one generator, and (iii) design learning protocols that can withstand error accumulation when predicting more complex and severe discontinuous disturbances.

5 Discussion

This section discusses our main results and outlines our plans for future work.

We have demonstrated through all five tasks outlined in Section 4 that the proposed recursive and data-driven RK DeepONets can effectively approximate the dynamic response of nonlinear, non-autonomous systems for varying inputs and initial conditions. Furthermore, in Section 4.3, we compared the proposed recursive method with two other benchmarks. The first is a neural network approach for approximating non-autonomous systems developed in Qin et al. (2021). The second is an ensemble method Wang et al. (2019) that we extended to the continuous scenario for comparison purposes. Table 3 shows that the proposed recursive DeepONet outperforms the other two benchmarks when the training dataset size is small. In particular, the recursive method keeps the mean L2L_{2}-relative error for 100 test trajectories below 1.11.1 %. Additionally, Figure 9 demonstrates that among all benchmarks, the proposed recursive DeepONet is the only one that can effectively approximate challenging oscillatory solution trajectories. In conclusion, when the dataset is small, which often occurs in engineering systems, the proposed recursive DeepONet provides the best method for approximating the dynamic response of non-autonomous systems.

In our future work, we aim to extend the proposed framework to include (i) reduced-order, (ii) stochastic, and (iii) networked non-autonomous and control systems. We also plan to apply the DeepONet framework to model-based reinforcement learning and control design. Specifically, we aim to use it for learning semi-Markov decision processes, which can be applied to learn suboptimal offline policies. Moreover, we aim to apply the recursive and data-driven RK DeepONets to complex engineering applications such as fluid dynamics, materials engineering, and future power systems.

6 Conclusion

We introduced a Deep Operator Network (DeepONet) framework that can learn (from data) the dynamic response of nonlinear, non-autonomous systems with time-dependent inputs for long-term horizons. The proposed framework approximates the system’s solution operator locally using the DeepONet and then recursively predicts the system’s response for long/medium-term horizons using the trained network. We also estimated the error bound for this DeepONet-based numerical scheme. To improve the predictive accuracy of the scheme when the step size is small, we designed and theoretically validated a data-driven Runge-Kutta DeepONet scheme. This scheme uses estimates of the vector field computed with the DeepONet forward pass and automatic differentiation. Finally, we validated the proposed framework using an autonomous and chaotic system, three continuous control tasks, and a power engineering application.

Acknowledgments

We gratefully acknowledge the support of the National Science Foundation (DMS-1555072, DMS-2053746, and DMS-2134209), Brookhaven National Laboratory Subcontract 382247, and U.S. Department of Energy (DOE) Office of Science Advanced Scientific Computing Research program DE-SC0021142 and DE-SC0023161.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  • Bradbury et al. (2018) Bradbury, J., Frostig, R., Hawkins, P., Johnson, M.J., Leary, C., Maclaurin, D., Necula, G., Paszke, A., VanderPlas, J., Wanderman-Milne, S., Zhang, Q., 2018. JAX: composable transformations of Python+NumPy programs. URL: http://github.com/google/jax.
  • Brockman et al. (2016) Brockman, G., Cheung, V., Pettersson, L., Schneider, J., Schulman, J., Tang, J., Zaremba, W., 2016. Openai gym. arXiv preprint arXiv:1606.01540 .
  • Brunton et al. (2016a) Brunton, S.L., Proctor, J.L., Kutz, J.N., 2016a. Discovering governing equations from data by sparse identification of nonlinear dynamical systems. Proceedings of the national academy of sciences 113, 3932–3937.
  • Brunton et al. (2016b) Brunton, S.L., Proctor, J.L., Kutz, J.N., 2016b. Sparse identification of nonlinear dynamics with control (sindyc). IFAC-PapersOnLine 49, 710–715.
  • Cai et al. (2021) Cai, S., Wang, Z., Lu, L., Zaki, T.A., Karniadakis, G.E., 2021. Deepm&mnet: Inferring the electroconvection multiphysics fields based on operator approximation by neural networks. Journal of Computational Physics 436, 110296.
  • Chen and Chen (1995) Chen, T., Chen, H., 1995. Universal approximation to nonlinear operators by neural networks with arbitrary activation functions and its application to dynamical systems. IEEE Transactions on Neural Networks 6, 911–917.
  • Choudhary et al. (2023) Choudhary, A., Mishra, R.K., Fatima, S., Panigrahi, B., 2023. Multi-input cnn based vibro-acoustic fusion for accurate fault diagnosis of induction motor. Engineering Applications of Artificial Intelligence 120, 105872.
  • Chow et al. (2000) Chow, J., Rogers, G., Cheung, K., 2000. Power system toolbox. Cherry Tree Scientific Software 48, 53.
  • Chung et al. (2021) Chung, E., Leung, W.T., Pun, S.M., Zhang, Z., 2021. A multi-stage deep learning based algorithm for multiscale model reduction. Journal of Computational and Applied Mathematics 394, 113506.
  • Cybenko (1989) Cybenko, G., 1989. Approximation by superpositions of a sigmoidal function. Mathematics of control, signals and systems 2, 303–314.
  • Deisenroth et al. (2013) Deisenroth, M.P., Fox, D., Rasmussen, C.E., 2013. Gaussian processes for data-efficient learning in robotics and control. IEEE transactions on pattern analysis and machine intelligence 37, 408–423.
  • Du et al. (2020) Du, J., Futoma, J., Doshi-Velez, F., 2020. Model-based reinforcement learning for semi-markov decision processes with neural odes. Advances in Neural Information Processing Systems 33, 19805–19816.
  • Efendiev et al. (2021) Efendiev, Y., Leung, W.T., Lin, G., Zhang, Z., 2021. Hei: hybrid explicit-implicit learning for multiscale problems. arXiv preprint arXiv:2109.02147 .
  • Fan et al. (2020) Fan, D., Yang, L., Wang, Z., Triantafyllou, M.S., Karniadakis, G.E., 2020. Reinforcement learning for bluff body active flow control in experiments and simulations. Proceedings of the National Academy of Sciences 117, 26091–26098.
  • Florian (2007) Florian, R.V., 2007. Correct equations for the dynamics of the cart-pole system. Center for Cognitive and Neural Studies (Coneural), Romania .
  • Hafner et al. (2019) Hafner, D., Lillicrap, T., Ba, J., Norouzi, M., 2019. Dream to control: Learning behaviors by latent imagination. arXiv preprint arXiv:1912.01603 .
  • Iserles (2009) Iserles, A., 2009. A first course in the numerical analysis of differential equations. 44, Cambridge university press.
  • Kaiser et al. (2019) Kaiser, L., Babaeizadeh, M., Milos, P., Osinski, B., Campbell, R.H., Czechowski, K., Erhan, D., Finn, C., Kozakowski, P., Levine, S., et al., 2019. Model-based reinforcement learning for atari. arXiv preprint arXiv:1903.00374 .
  • Kingma and Ba (2014) Kingma, D.P., Ba, J., 2014. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980 .
  • Li et al. (2020) Li, Z., Kovachki, N., Azizzadenesheli, K., Liu, B., Bhattacharya, K., Stuart, A., Anandkumar, A., 2020. Fourier neural operator for parametric partial differential equations. arXiv preprint arXiv:2010.08895 .
  • Lin et al. (2023) Lin, G., Moya, C., Zhang, Z., 2023. B-deeponet: An enhanced bayesian deeponet for solving noisy parametric pdes using accelerated replica exchange sgld. Journal of Computational Physics 473, 111713.
  • Lin et al. (2021) Lin, G., Wang, Y., Zhang, Z., 2021. Multi-variance replica exchange stochastic gradient mcmc for inverse and forward bayesian physics-informed neural network. arXiv preprint arXiv:2107.06330 .
  • Lu et al. (2019) Lu, L., Jin, P., Karniadakis, G.E., 2019. Deeponet: Learning nonlinear operators for identifying differential equations based on the universal approximation theorem of operators. arXiv preprint arXiv:1910.03193 .
  • Lu et al. (2021) Lu, L., Jin, P., Pang, G., Zhang, Z., Karniadakis, G.E., 2021. Learning nonlinear operators via deeponet based on the universal approximation theorem of operators. Nature Machine Intelligence 3, 218–229.
  • Mishra et al. (2022a) Mishra, R., Choudhary, A., Fatima, S., Mohanty, A., Panigrahi, B., 2022a. A fault diagnosis approach based on 2d-vibration imaging for bearing faults. Journal of Vibration Engineering & Technologies , 1–14.
  • Mishra et al. (2022b) Mishra, R.K., Choudhary, A., Fatima, S., Mohanty, A.R., Panigrahi, B.K., 2022b. A self-adaptive multiple-fault diagnosis system for rolling element bearings. Measurement Science and Technology 33, 125018.
  • Moya and Lin (2023) Moya, C., Lin, G., 2023. Dae-pinn: a physics-informed neural network model for simulating differential algebraic equations with application to power networks. Neural Computing and Applications 35, 3789–3804.
  • Moya et al. (2023) Moya, C., Zhang, S., Lin, G., Yue, M., 2023. Deeponet-grid-uq: A trustworthy deep operator framework for predicting the power grid’s post-fault trajectories. Neurocomputing 535, 166–182.
  • Nagabandi et al. (2020) Nagabandi, A., Konolige, K., Levine, S., Kumar, V., 2020. Deep dynamics models for learning dexterous manipulation, in: Conference on Robot Learning, PMLR. pp. 1101–1112.
  • Proctor et al. (2018) Proctor, J.L., Brunton, S.L., Kutz, J.N., 2018. Generalizing koopman theory to allow for inputs and control. SIAM Journal on Applied Dynamical Systems 17, 909–930.
  • Qin et al. (2021) Qin, T., Chen, Z., Jakeman, J.D., Xiu, D., 2021. Data-driven learning of nonautonomous systems. SIAM Journal on Scientific Computing 43, A1607–A1624.
  • Qin et al. (2019) Qin, T., Wu, K., Xiu, D., 2019. Data driven governing equations approximation using deep neural networks. Journal of Computational Physics 395, 620–635.
  • Raissi et al. (2018) Raissi, M., Perdikaris, P., Karniadakis, G.E., 2018. Multistep neural networks for data-driven discovery of nonlinear dynamical systems. arXiv preprint arXiv:1801.01236 .
  • Ranade et al. (2021) Ranade, R., Gitushi, K., Echekki, T., 2021. Generalized joint probability density function formulation inturbulent combustion using deeponet. arXiv preprint arXiv:2104.01996 .
  • Schaeffer (2017) Schaeffer, H., 2017. Learning partial differential equations via data discovery and sparse optimization. Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences 473, 20160446.
  • Sun et al. (2020) Sun, Y., Zhang, L., Schaeffer, H., 2020. Neupde: Neural network based ordinary and partial differential equations for modeling time-dependent data, in: Mathematical and Scientific Machine Learning, PMLR. pp. 352–372.
  • Sutton (1990) Sutton, R.S., 1990. Integrated architectures for learning, planning, and reacting based on approximating dynamic programming, in: Machine learning proceedings 1990. Elsevier, pp. 216–224.
  • Sutton and Barto (2018) Sutton, R.S., Barto, A.G., 2018. Reinforcement learning: An introduction. MIT press.
  • Wang et al. (2021a) Wang, S., Teng, Y., Perdikaris, P., 2021a. Understanding and mitigating gradient flow pathologies in physics-informed neural networks. SIAM Journal on Scientific Computing 43, A3055–A3081.
  • Wang et al. (2021b) Wang, S., Wang, H., Perdikaris, P., 2021b. Learning the solution operator of parametric partial differential equations with physics-informed deeponets. Science advances 7, eabi8605.
  • Wang et al. (2019) Wang, T., Bao, X., Clavera, I., Hoang, J., Wen, Y., Langlois, E., Zhang, S., Zhang, G., Abbeel, P., Ba, J., 2019. Benchmarking model-based reinforcement learning. arXiv preprint arXiv:1907.02057 .
  • Wittenmark et al. (2002) Wittenmark, B., Åström, K.J., Årzén, K.E., 2002. Computer control: An overview. IFAC Professional Brief 1, 2.
  • Zhang et al. (2022) Zhang, Z., Leung, W.T., Schaeffer, H., 2022. Belnet: Basis enhanced learning, a mesh-free neural operator. arXiv preprint arXiv:2212.07336 .