This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Prediction of fitness in bacteria with causal jump dynamic mode decomposition

Shara Balakrishnan, Aqib Hasnain, Nibodh Boddupalli, Dennis M. Joshy, Robert G. Egbert, and Enoch Yeung Shara Balakrishnan ([email protected]) is with the Department of Electrical and Computer Engineering, University of California, Santa Barbara
Aqib Hasnain, Nibodh Boddupalli and Dennis Joshy are with the Department of Mechanical Engineering, University of California, Santa Barbara, Robert G. Egbert is with the Environmental and Biological Sciences Directorate at the Pacific Northwest National Laboratory, Richland, WA.
Enoch Yeung ([email protected]) is with the Department of Mechanical Engineering, Center for Control, Dynamical Systems, and Computation, and Biomolecular Science and Engineering, University of California, Santa Barbara
Abstract

In this paper, we consider the problem of learning a predictive model for population cell growth dynamics as a function of the media conditions. We first introduce a generic data-driven framework for training operator-theoretic models to predict cell growth rate. We then introduce the experimental design and data generated in this study, namely growth curves of Pseudomonas putida as a function of casein and glucose concentrations. We use a data driven approach for model identification, specifically the nonlinear autoregressive (NAR) model to represent the dynamics. We show theoretically that Hankel DMD can be used to obtain a solution of the NAR model. We show that it identifies a constrained NAR model and to obtain a more general solution, we define a causal state space system using 1-step, 2-step,…, τ\tau-step predictors of the NAR model and identify a Koopman operator for this model using extended dynamic mode decomposition. The hybrid scheme we call causal-jump dynamic mode decomposition, which we illustrate on a growth profile or fitness prediction challenge as a function of different input growth conditions. We show that our model is able to recapitulate training growth curve data with 96.6% accuracy and predict test growth curve data with 91% accuracy.

I Introduction

One of the most fundamental processes in life is the ability to replicate and pass on hereditary material [1]. From viral particles to bacteria to mammalian cells, cell division is fundamental to growth, maintenance of physiological health, and intrinsically tied to the notion of senescence [2].

The mechanisms for controlling growth in organisms are determined by metabolic networks [3, 4], namely their topological structure and parametric realization. Known metabolic networks in well studied model organisms such as E. coli [5] and S. cerevisiae [6, 7] have given rise to predictive models that translate environmental activity to metabolic network state, and ultimately to predictions of growth rate. For canonical biological model systems, these models have been highly accurate in predicting growth rate and found utility in industrial microbiology applications, e.g. in the design of bioreactors or informing best practices in food safety.

For many biological life forms, relatively little is known about their metabolic network or structure. This is especially the case when developing bioengineering tools in novel host microbes [8, 9]. For new organisms, canonical metabolic networks are lacking and often obtained through a process of sequence alignment and comparative analysis with existing metabolic network models in relative species. However, many novel strains do not exhibit significant similarity, and even in the case of sequence similarity, small mutations can lead to dramatically different growth phenotypes, e.g. growth of non-pathogenic soil strains[10, 11] versus pathogenic counterparts [12]. The absence of predictive cross-species models, as well as the inability to predict growth phenotype wholly from sequence data, motivates the need for data-driven methods to accelerate the discovery of metabolic models and growth rate prediction models.

Due to advances in high-throughput experimental techniques, it is relatively easy to characterize growth rates as a function of exposure to environment. Liquid and acoustic-liquid handling robotics enables interrogation of thousands of growth conditions in a single microtiter plate, which in turn opens the door for using data-driven approaches [13] to predict growth rate as a function of environmental state. Is it possible to accurately predict the growth rate of a microbe, entirely from the chemical composition and environmental parameters of its growth condition? In this paper we explore a data-driven operator theoretic approach that utilizes microtiter plate reader data, and more generally multi-variate time-series data, to develop predictive models of growth rate in Pseudomonas putida, a broadly used strain for commercial bioreactors and a target workhorse for tractable genetic engineering [14].

A broadly successful class of data-driven modeling approaches stem from the study of Koopman operators, a mathematical construct for representing the time-evolution of nonlinear dynamical systems. In Koopman operator theory, the time-evolution of a nonlinear system is defined on a function space, acting on the original state of the system. In this function space, known the observables space, the Koopman operator is a linear operator, enabling spectral analysis, the decomposition of eigenspaces, and study of nonlinear structure [15]. The Koopman operator framework has been developed for continuous [16] and discrete time systems [17, 18], for open-loop [17] and input-controlled [19, 20] dynamic systems. Thus, Koopman operators present a powerful framework for analyzing the behavior of nonlinear systems, including predicting how experimental conditions regulate growth dynamics.

Many numerical methods for identifying Koopman operators from data have been developed in the last two decades [21, 22, 23, 24, 25, 26, 27, 28]. The most common approach is to use dynamic mode decomposition (DMD), which models nonlinear dynamics via an approximate local linear expansion [25]. In [17] an extended dictionary of basis functions with universal function approximation properties is used to discover an approximation of the lifting map or observables. These techniques suffer from combinatorial explosion, which generally has prohibited analysis of high-dimensional nonlinear systems [29]. The most recent developments in the field of DMD integrate established advances in deep learning with DMD [27, 30, 31, 32] where the deep neural networks have the capacity to approximate exponentially many distinct observable functions. Recent work has shown deep Koopman learning algorithms can be extended to synthesize controllers for systems subject to uncertainty [33], suggesting that deep Koopman learning can be used broadly for robust controller synthesis.

The existing algorithms assume a full state measurement and construct observables from that. With partial state observables like in biological systems, we construct a state space model for an output nonlinear difference equation model and identify a Koopman operator for that model. In Section II we briefly introduce the Koopman operator and DMD and the existing literature. In Section III we describe the experimental setup for obtaining the growth curve data of Pseudomonas putida by adding different concentrations of casein and glucose substrates to the media. In Section IV, we justify nonlinear autoregressive difference equation model is an appropriate choice for this system. In Section V, we formulate the Hankel DMD as a solution of the NAR model and bring out its issues and in Section VI, we formulate a state space model for the NAR model and use extended dynamic mode decomposition to identify a Koopman operator for the NAR model. In Section VII, we show that the algorithm is able to train a predictive Koopman operator, that predicts with 3.4% on the training data and 9% on the test data on extended forecasting tasks approximately 500 time steps ahead.

II Mathematical Preliminaries

Consider a discrete-time autonomous nonlinear dynamical system

xk+1=f(xk)\begin{split}x_{k+1}&=f(x_{k})\ \end{split} (1)

with f:nnf:\mathbb{R}^{n}\rightarrow\mathbb{R}^{n} is analytic. There exists a Koopman operator [34] of (1), which acts on a function space \mathcal{F} as 𝒦\mathcal{K} : \mathcal{F} \rightarrow \mathcal{F}. This action can be given by

𝒦ψ(xk)=ψf(xk).\mathcal{K}\psi(x_{k})=\psi\circ f(x_{k}). (2)

where the function ψ:n\psi:\mathbb{R}^{n}\rightarrow\mathbb{R} is called an observable of the system and the set of all observables ψ{ψi}i=1p,p\psi\triangleq\{\psi_{i}\}_{i=1}^{p},p\leq\infty on the system. Here \mathcal{F} is invariant under the action of 𝒦\mathcal{K}.

The most important property of the Koopman operator that we utilize is the linearity of the operator, in other words,

𝒦(αψ1+βψ2)=αψ1f+βψ2f=α𝒦ψ1+β𝒦ψ2\mathcal{K}(\alpha\psi_{1}+\beta\psi_{2})=\alpha\psi_{1}\circ f+\beta\psi_{2}\circ f=\alpha\mathcal{K}\psi_{1}+\beta\mathcal{K}\psi_{2}

which follows from (2) since the composition operator is linear. Thus, we have that the Koopman operator of (1) is a linear operator that acts on observable functions ψ(xk)\psi(x_{k}) and propagates them forward in time.

II-A DMD and relevant variants

The practical identification of Koopman operator for a nonlinear system from input-output data is commonly done using DMD [25] or extended DMD [17] which constructs an approximate Koopman operator KK. Rowley et. al showed that the finite-dimensional approximation to the Koopman operator obtained from DMD is closely related to a spectral analysis of the linear but infinite-dimensional Koopman operator [18]. The approach taken to compute an approximation to the Koopman operator in both DMD and extended DMD is as follows

K=minKΨ(Xf)KΨ(Xp)=Ψ(Xf)Ψ(Xp)K=\min_{K}||\Psi(X_{f})-K\Psi(X_{p})||=\Psi(X_{f})\Psi(X_{p})^{{\dagger}} (3)

where Xf[x1xN1],X_{f}\equiv\begin{bmatrix}x_{1}&\ldots&x_{N-1}\end{bmatrix}, Xp[x2xN]X_{p}\equiv\begin{bmatrix}x_{2}&\ldots&x_{N}\end{bmatrix} are snapshot matrices formed from the discrete-time dynamical system (1), Ψ(X)[ψ1(x)ψR(x)]\Psi(X)\equiv\begin{bmatrix}\psi_{1}(x)&\ldots&\psi_{R}(x)\end{bmatrix} is the mapping from physical space into the space of observables and {\dagger} denotes the Moore-Penrose pseudoinverse. Here NN is the number of snapshots i.e. timepoints. We note that DMD is a special case of extended DMD where ψ(x)=x\psi(x)=x. Throughout the rest of the paper, when we refer to the Koopman operator we mean the finite dimensional approximation to the infinite-dimensional Koopman operator.

III Experimental Setup

We describe the procedure adopted to obtain P. putida’s growth curve for varying concentrations of glucose and casein substrates in the media.

Incubation: We revived P. putida cryopreserved at 80oC-80^{o}C in 30% (vol/vol) glycerol stock by suspending a small portion into a polypropylene test tube containing 4 mL of Lysogeny Broth (LB). We cultured it at 30oC30^{o}C spinning with a speed of 200 revolutions per minute (rpm) for 12 hours overnight. A visual inspection of the culture tube resulted in a cloudy culture medium, suggesting subsequent growth with the seed P. putida culture in a plate reader was feasible.

Solution Preparation: We prepared 300 g/L solution of glucose solution and 225 g/L casein acid hydrolysate solution. Once the bacteria reached a certain optical density(OD), we shifted the culture from LB media to R2A 2x media obtained from Teknova Inc to 2x the required initial OD.

Serial dilution setup for P. putida culture: We use a 630 μL\mu L 96 well plate to create media with different substrate concentrations. Each well of this plate contained 500 μL\mu L of modified media - 250 μL\mu L of culture in 2x R2A at 0.4 OD and 120 μL\mu L containing a mixture of casein and glucose solutions. To vary casein and glucose across the 96 well plate, we perform 2D serial dilution such that the concentration of glucose was halved across columns and concentration of casein is halved across rows as shown in Figure 1. Then, the culture was mixed into each well to get a starting OD of 0.2 in 1x R2A media since equal volumes of culture media and substrate solutions were added.

Refer to caption
Figure 1: Different initial conditions of substrates obtained by two dimensional serial dilution of casein and glucose and the corresponding growth curves are obtained for a period of 27 hours.

Data Collection: The microplate reader was set to 30o30^{o}C and the shaker to 807 cycles per minute, with continuous double orbital mixing. The absorbance at 600 nanometers(nm), which is termed as the Optical Density at 600 nm (OD600OD_{600}), was measured as a function of time for 27 hours. We assume in this work, as is widely accepted, that OD600OD_{600} measurements were collected in a linear regime, where cell population is proportional OD600 measurements. The obtained data along with the varying substrate concentrations are shown in Figure. 1.

IV Growth Curve Dynamics Model

The dynamics of the bacterial cell growth can be represented by

[Nk+1(b)Ck+1Gk+1]=f(Nk(b),Ck,Gk)\displaystyle\begin{bmatrix}N^{(b)}_{k+1}\\ C_{k+1}\\ G_{k+1}\ \end{bmatrix}=f(N^{(b)}_{k},C_{k},G_{k}) (4)

where the bacterial cell count (N(b)N^{(b)}), casein substrate concentration (CC) and glucose substrate concentration (GG) are the states of the system, ff is the nonlinear dynamics and kk is the discrete time index. We measure the OD600OD_{600} data as mentioned in section III and the output equation is given by

yk=h(Nk(b))\displaystyle y_{k}=h(N_{k}^{(b)}) (5)

as OD600OD_{600} is a function of only the number of cells. Some of the existing empirical nonlinear models for growth curve dynamics include the Monod’s model [35] which uses a single substrate to form the foundation of the growth curve dynamics and in [36] and [37] multiple substrates are incorporated. Monod’s model is a two-state nonlinear dynamical system comprising of the substrate (S) and the number of bacteria (N(b)N^{(b)}):

N˙(b)(t)\displaystyle\dot{N}^{(b)}(t) =rmaxS(t)N(b)(t)Ks+S(t)\displaystyle=r_{max}\frac{S(t)N^{(b)}(t)}{K_{s}+S(t)}
S˙\displaystyle\dot{S} =γN˙(b)\displaystyle=-\gamma\dot{N}^{(b)}\ (6)

where rmaxr_{max} is the maximum growth rate and KsK_{s} is the half velocity constant. As N(b)N^{(b)} is the only variable of measurement in (5), we convert the model to a single differential equation containing only N(b)N^{(b)}

N¨(b)(t)=1rmaxKsN(b)(KsrmaxN˙(b)2γN˙(b)3+2γrmaxN(b)N˙(b)2γrmax2N(b)2N˙(b))\begin{split}\ddot{N}^{(b)}(t)&=\frac{1}{r_{max}K_{s}N^{(b)}}(K_{s}r_{max}\dot{N}^{(b)^{2}}-\gamma\dot{N}^{(b)^{3}}\\ &+2\gamma r_{max}N^{(b)}\dot{N}^{(b)^{2}}-\gamma r_{max}^{2}N^{(b)^{2}}\dot{N}^{(b)})\end{split} (7)

The existing models though heuristic, suggest that N(b)N^{(b)} at any point in time is a function of the past

Nk+1(b)=f(Nk(b),Nk1(b),)\displaystyle N^{(b)}_{k+1}=f(N^{(b)}_{k},N^{(b)}_{k-1},\cdots)

Nk(b)N^{(b)}_{k} to be a function of its finite past. This is the general structure of the discrete nonlinear autoregressive (NAR) model given by

yk=f(yk1,yk2,,ykτ)yipi>0f:p×p××pτ timesp\begin{split}y_{k}&=f(y_{k-1},y_{k-2},\cdots,y_{k-\tau})\\ y_{i}&\in\mathbb{R}^{p}\quad\forall i\in\mathbb{Z}_{>0}\\ f&:\underbrace{\mathbb{R}^{p}\times\mathbb{R}^{p}\times\cdots\times\mathbb{R}^{p}}_{\tau\text{ times}}\rightarrow\mathbb{R}^{p}\end{split} (8)

where the current output is a function of the past τ\tau outputs.

V Hankel dynamic mode decomposition

Given the nonlinear system (4) with the state measurement given by (5) and modeled by the discrete time difference equation (8), Hankel DMD [38] is a suitable algorithm to solve the model identification problem with the NAR structure. The promising feature of using a DMD algorithm is that it identifies a linear state space representation which has a theoretical foundation in Koopman operator theory.

Given the autonomous state space system

x~k+1=f~(x~k)yk=h(x~k)\begin{split}\tilde{x}_{k+1}&=\tilde{f}(\tilde{x}_{k})\\ y_{k}&=h(\tilde{x}_{k})\end{split} (9)

where xknx_{k}\in\mathbb{R}^{n} is the state, f~:nn\tilde{f}:\mathbb{R}^{n}\rightarrow\mathbb{R}^{n} is the dynamics, ykpy_{k}\in\mathbb{R}^{p} is the output and h:nph:\mathbb{R}^{n}\rightarrow\mathbb{R}^{p} is a nonlinear function that maps the state directly to itself, i.e. xx is identical to the output yy, Hankel DMD constructs a Koopman model of the form

[ψ(yk+1)ψ(yk+2)ψ(yk+τ)]=K[ψ(yk)ψ(yk+1)ψ(yk+τ1)]\begin{bmatrix}\psi(y_{k+1})\\ \psi(y_{k+2})\\ \vdots\\ \psi(y_{k+\tau})\end{bmatrix}=K\begin{bmatrix}\psi(y_{k})\\ \psi(y_{k+1})\\ \vdots\\ \psi(y_{k+\tau-1})\end{bmatrix} (10)

such that ψ:pNp\psi:\mathbb{R}^{p}\rightarrow\mathbb{R}^{N_{p}} is the dictionary of state inclusive observables of the state x~k\tilde{x}_{k} constructed by a nonlinear transformation of the corresponding output yky_{k} and KK is the Koopman operator. Regardless of full-state measurements, we nonetheless cast Hankel DMD in this form to compare it with our subsequent causal jump DMD algorithm.

Given the output measurements {y1,y2,..,yN}\{y_{1},y_{2},..,y_{N}\}, to identify an approximate Koopman operator KK using Hankel DMD, the time shifted Hankel matrices are constructed as

Ψ(Yp)=[ψ(y1)ψ(y2)ψ(yNτ)ψ(y2)ψ(y3)ψ(yNτ+1)ψ(yτ)ψ(yτ+1)ψ(yN1)]Ψ(Yf)=[ψ(y2)ψ(y3)ψ(yNτ+1)ψ(y3)ψ(y4)ψ(yNτ+2)ψ(yτ+1)ψ(yτ+2)ψ(yN)]\begin{split}\Psi(Y_{p})&=\begin{bmatrix}\psi(y_{1})&\psi(y_{2})&\ldots&\psi(y_{N-\tau})\\ \psi(y_{2})&\psi(y_{3})&\ldots&\psi(y_{N-\tau+1})\\ \vdots&\vdots&\ddots&\vdots\\ \psi(y_{\tau})&\psi(y_{\tau+1})&\ldots&\psi(y_{N-1})\end{bmatrix}\\ \Psi(Y_{f})&=\begin{bmatrix}\psi(y_{2})&\psi(y_{3})&\ldots&\psi(y_{N-\tau+1})\\ \psi(y_{3})&\psi(y_{4})&\ldots&\psi(y_{N-\tau+2})\\ \vdots&\vdots&\ddots&\vdots\\ \psi(y_{\tau+1})&\psi(y_{\tau+2})&\ldots&\psi(y_{N})\end{bmatrix}\end{split} (11)

and the optimization problem

minKΨ(Zf)KΨ(Zp)\min_{K}||\Psi(Z_{f})-K\Psi(Z_{p})|| (12)

is solved using the Moore-Penrose pseudoinverse method mentioned in section II. This yields a solution of the form

[ψ(yk+1)ψ(yk+2)ψ(yk+τ)]=[0𝕀Np000000000𝕀Npk1k2kτ1kτ][ψ(yk)ψ(yk+1)ψ(yk+τ1)]\begin{bmatrix}\psi(y_{k+1})\\ \psi(y_{k+2})\\ \vdots\\ \psi(y_{k+\tau})\end{bmatrix}=\begin{bmatrix}0&\mathbb{I}_{N_{p}}&\cdots&0&0\\ 0&0&\cdots&0&0\\ \vdots&\vdots&\ddots&\vdots&\vdots\\ 0&0&\cdots&0&\mathbb{I}_{N_{p}}\\ k_{1}&k_{2}&\cdots&k_{\tau-1}&k_{\tau}\\ \end{bmatrix}\begin{bmatrix}\psi(y_{k})\\ \psi(y_{k+1})\\ \vdots\\ \psi(y_{k+\tau-1})\end{bmatrix}

Other than the last NpN_{p} equations, the others are trivial. To construct an output predictor, we take the component yk+τy_{k+\tau} of ψ(yk+τ)\psi(y_{k+\tau}) to get

yk+τ=k~1ψ(yk)+k~2ψ(yk+1)++k~τψ(yk+τ+1)y_{k+\tau}=\tilde{k}_{1}\psi(y_{k})+\tilde{k}_{2}\psi(y_{k+1})+\cdots+\tilde{k}_{\tau}\psi(y_{k+\tau+1}) (13)

where k~i\tilde{k}_{i} are the components of kik_{i} that map ψ(yk+τ1)\psi(y_{k+\tau-1}) to yk+τy_{k+\tau}. More generally, this yields a nonlinear equation of the form

yk=f~1(yk1)+f~2(yk2)++f~τ(ykτ)y_{k}=\tilde{f}_{1}(y_{k-1})+\tilde{f}_{2}(y_{k-2})+\cdots+\tilde{f}_{\tau}(y_{k-\tau}) (14)

where the functions f~1,f~2,,f~τ\tilde{f}_{1},\tilde{f}_{2},\cdots,\tilde{f}_{\tau} have the same basis functions with different coefficients. This identifies a constrained NAR model as it imposes an additive structure on the basis of nonlinear models across time.

VI Dynamic mode decomposition of nonlinear autoregressive models

To identify a Koopman operator for the unconstrained NAR model (8), we formulate a state space representation for the NAR model with full state observation and identify an approximate Koopman operator for that model using the general class of DMD algorithms like extended DMD and deep DMD.

In this methodology, the problem is broken into two pieces —— the system identification aspect where we select the model structure and the dynamic mode decomposition aspect where we have to construct the dictionary of observables. We define a window parameter τ>0\tau\in\mathbb{Z}_{>0} indicating how many past output snapshots are used to define a new extended dictionary of monomial observable functions, up to order no>0n_{o}\in\mathbb{Z}_{>0}. The new τ\tau-dictionary defines a general extended dynamic mode decomposition problem, which we then solve using classical methods.

We proceed as follows: given the NAR model (8) with the system identification parameter τ\tau, we construct a state defined by

zk:=[yk+1yk+2yk+τ]Tz_{k}:=\begin{bmatrix}y_{k+1}&y_{k+2}&\cdots&y_{k+\tau}\end{bmatrix}^{T} (15)

with zkpτz_{k}\in\mathbb{R}^{p\tau}.This yields the state space representation

zk+1\displaystyle z_{k+1} =[yk+2yk+3yk+τyk+τ+1]:=[f1(yk+1,yk+2,,yk+τ)f2(yk+1,yk+2,,yk+τ)fτ1(yk+1,yk+2,,yk+τ)fτ(yk+1,yk+2,,yk+τ)]\displaystyle=\begin{bmatrix}y_{k+2}\\ y_{k+3}\\ \vdots\\ y_{k+\tau}\\ y_{k+\tau+1}\end{bmatrix}:=\begin{bmatrix}f_{1}(y_{k+1},y_{k+2},\cdots,y_{k+\tau})\\ f_{2}(y_{k+1},y_{k+2},\cdots,y_{k+\tau})\\ \vdots\\ f_{\tau-1}(y_{k+1},y_{k+2},\cdots,y_{k+\tau})\\ f_{\tau}(y_{k+1},y_{k+2},\cdots,y_{k+\tau})\\ \end{bmatrix}
:=[yk+2yk+3yk+τf(yk+1,yk+2,,yk+τ)]=F(zk)\displaystyle:=\begin{bmatrix}y_{k+2}\\ y_{k+3}\\ \vdots\\ y_{k+\tau}\\ f_{(}y_{k+1},y_{k+2},\cdots,y_{k+\tau})\end{bmatrix}=F(z_{k})
zk+1\displaystyle\Rightarrow z_{k+1} =F~(zk)\displaystyle=\tilde{F}(z_{k}) (16)

where F~:pτpτ\tilde{F}:\mathbb{R}^{p\tau}\rightarrow\mathbb{R}^{p\tau} represents the dynamics of the lifted ”state” model. The approximate EDMD model for the full output observable model is given by

ψ(zk+1)=Kψ(zk)\psi(z_{k+1})=K\psi(z_{k}) (17)

where ψ(zk)\psi(z_{k}) is the state inclusive dictionary of observables defined as

ψ(zk)=[zkφ(zk)]\psi(z_{k})=\begin{bmatrix}z_{k}\\ \varphi(z_{k})\end{bmatrix} (18)

with φ:pτNp\varphi:\mathbb{R}^{p\tau}\rightarrow\mathbb{R}^{N_{p}} being a nonlinear transformation that constructs the nonlinear observables. Since the only additional information in the state zk+1z_{k+1} when compared to the state zkz_{k} is yk+τ+1y_{k+\tau+1}, the output predictor form for the Koopman model can be identified considering the complete Koopman model and extracting the equation that corresponds to yk+τ+1y_{k+\tau+1} given by

ψ(zk+1)\displaystyle\psi(z_{k+1}) =Kψ(zk)\displaystyle=K\psi(z_{k})
[yk+2yk+τyk+τ+1φ(zk+1)]\displaystyle\Rightarrow\begin{bmatrix}y_{k+2}\\ \vdots\\ y_{k+\tau}\\ y_{k+\tau+1}\\ \varphi(z_{k+1})\end{bmatrix} =[k1kτ1kτk11][yk+1yk+τ1yk+τφ(zk)]\displaystyle=\begin{bmatrix}\bullet&\cdots&\bullet&\bullet&\bullet\\ \vdots&\ddots&\vdots&\vdots&\vdots\\ \bullet&\cdots&\bullet&\bullet&\bullet\\ k_{1}&\cdots&k_{\tau-1}&k_{\tau}&k_{11}\\ \bullet&\cdots&\bullet&\bullet&\bullet\\ \end{bmatrix}\begin{bmatrix}y_{k+1}\\ \vdots\\ y_{k+\tau-1}\\ y_{k+\tau}\\ \varphi(z_{k})\end{bmatrix}
yk+τ+1\displaystyle\Rightarrow y_{k+\tau+1} =k1yk+1++kτyk+τ\displaystyle=k_{1}y_{k+1}+\cdots+k_{\tau}y_{k+\tau}
+k11Tφ(yk+1,,yk+τ)\displaystyle+k^{T}_{11}\varphi(y_{k+1},\cdots,y_{k+\tau}) (19)

The output predictor form keeps the general structure of the NAR model intact as opposed to the predictor identified by Hankel DMD which identified a constrained model. But, the issue with this model is the causality. It can be seen from (VI) that the Koopman model is non causal due to the overlap of outputs yky_{k} between the states zk+1z_{k+1} and zkz_{k}. This identifies models that use future outputs to predict past outputs which are inadmissible as our system is causal. To identify a causal model, the property of (VI) proved in proposition 1 is very important.

Proposition 1

Given the state space model (VI) for the nonlinear autoregressive (NAR) model (8), if the state is propagated ii time steps where i{1,2,,τ}i\in\{1,2,...,\tau\}

zk+i=F~i(zk)=F~F~F~i times(zk),\begin{split}z_{k+i}&=\tilde{F}^{i}(z_{k})=\underbrace{\tilde{F}\circ\tilde{F}\circ\cdots\circ\tilde{F}}_{\text{i times}}(z_{k}),\end{split}

then the last ii functions of F~Hi(zk)\tilde{F}_{H}^{i}(z_{k}) are such that

(F~i)(τi+j)(zk)=f(j)(zk)j{1,2,,i}yy+τ+j=f(j)(zk)=f(j)(yk+1,yk+2,,yk+τ)\begin{split}(\tilde{F}^{i})^{(\tau-i+j)}(z_{k})&=f^{(j)}(z_{k})\quad j\in\{1,2,\ldots,i\}\\ y_{y+\tau+j}=f^{(j)}(z_{k})&=f^{(j)}(y_{k+1},y_{k+2},\ldots,y_{k+\tau})\end{split} (20)

where (F~i)(b)(zk)(\tilde{F}^{i})^{(b)}(z_{k}) corresponds to the bthb^{th} function of F~i(zk)\tilde{F}^{i}(z_{k}) and f(j)(zk)f^{(j)}(z_{k}) is the jj-step predictor of the NAR model (8).

Proof:

Given the state zkz_{k} defined in (15), the state propagated ii time steps i0\forall i\in\mathbb{Z}_{\geq 0} is given by

zk+i=[yk+i+1yk+i+2yk+i+τ]Tz_{k+i}=\begin{bmatrix}y_{k+i+1}&y_{k+i+2}&\cdots&y_{k+i+\tau}\end{bmatrix}^{T}

and the mthm^{th} component of zk+iz_{k+i} is given by zk+i(m)=yk+i+mz_{k+i}^{(m)}=y_{k+i+m} where m{1,2,,τ}m\in\{1,2,...,\tau\}.

A function f(j):p×p××pτ timespf^{(j)}:\underbrace{\mathbb{R}^{p}\times\mathbb{R}^{p}\times\cdots\times\mathbb{R}^{p}}_{\tau\text{ times}}\rightarrow\mathbb{R}^{p} is a j-step predictor of the NAR model (8) if it has the following form

yk+τ+j=f(j)(xk)=f(j)(yk+1,yk+2,,yk+τ).\begin{split}y_{k+\tau+j}&=f^{(j)}(x_{k})=f^{(j)}(y_{k+1},y_{k+2},\cdots,y_{k+\tau}).\end{split}

Now that we have the state definitions and the predictor function definitions in place, we prove (20) by induction. For i=1i=1,

zk+1=F~(1)(zk)(F~1)(τi+j)(zk)=(F~1)(τ)(zk)=f(1)(xk)j{1}zk+1(τ)=yk+τ+1=f(1)(zk)\begin{split}z_{k+1}&=\tilde{F}^{(1)}(z_{k})\\ (\tilde{F}^{1})^{(\tau-i+j)}(z_{k})&=(\tilde{F}^{1})^{(\tau)}(z_{k})=f^{(1)}(x_{k})\quad j\in\{1\}\\ \Rightarrow z_{k+1}^{(\tau)}&=y_{k+\tau+1}=f^{(1)}(z_{k})\end{split}

Hence (20) is satisfied for i=1i=1. We assume the result is true for i=pi=p. This yields

(F~p(zk))(τp+j)(zk)=f(j)(zk)j{1,2,,p}zk+p=[yk+p+1yk+τyk+τ+1yk+τ+p]=F~p(zk)=[yk+p+1yk+τf(1)(zk)f(p)(zk)].\begin{split}(\tilde{F}^{p}(z_{k}))^{(\tau-p+j)}(z_{k})&=f^{(j)}(z_{k})\quad j\in\{1,2,...,p\}\\ \Rightarrow z_{k+p}=\begin{bmatrix}\vspace{-2pt}y_{k+p+1}\\ \vspace{-2pt}\vdots\\ y_{k+\tau}\\ \vspace{-2pt}y_{k+\tau+1}\\ \vspace{-2pt}\vdots\\ y_{k+\tau+p}\end{bmatrix}&=\tilde{F}^{p}(z_{k})=\begin{bmatrix}\vspace{-2pt}y_{k+p+1}\\ \vspace{-2pt}\vdots\\ y_{k+\tau}\\ \vspace{-2pt}f^{(1)}(z_{k})\\ \vspace{-2pt}\vdots\\ f^{(p)}(z_{k})\end{bmatrix}.\end{split}

For i=p+1i=p+1, the state zk+p+1z_{k+p+1} becomes

zk+p+1=F~p+1(zk)=F~F~p(zk)[yk+p+2yk+τyk+τ+1yk+τ+pyk+τ+p+1]=F~([yk+p+1yk+τf(1)(zk)f(p1)(zk)f(p)(zk)])=[yk+p+2yk+τf(1)(zk)f(p)(zk)g]\begin{split}z_{k+p+1}&=\tilde{F}^{p+1}(z_{k})=\tilde{F}\circ\tilde{F}^{p}(z_{k})\\ \Rightarrow\begin{bmatrix}\vspace{-2pt}y_{k+p+2}\\ \vspace{-2pt}\vdots\\ y_{k+\tau}\\ \vspace{-2pt}y_{k+\tau+1}\\ \vspace{-2pt}\vdots\\ y_{k+\tau+p}\\ y_{k+\tau+p+1}\end{bmatrix}&=\tilde{F}\left(\begin{bmatrix}\vspace{-2pt}y_{k+p+1}\\ \vspace{-2pt}\vdots\\ y_{k+\tau}\\ \vspace{-2pt}f^{(1)}(z_{k})\\ \vspace{-2pt}\vdots\\ f^{(p-1)}(z_{k})\\ f^{(p)}(z_{k})\\ \end{bmatrix}\right)=\begin{bmatrix}\vspace{-2pt}y_{k+p+2}\\ \vspace{-2pt}\vdots\\ y_{k+\tau}\\ \vspace{-2pt}f^{(1)}(z_{k})\\ \vspace{-2pt}\vdots\\ f^{(p)}(z_{k})\\ g\end{bmatrix}\\ \end{split}

where

g=f(yk+p+1,,yk+τ,f(1)(zk),,f(p)(zk))=f(zk(p+1),,zk(τ),f(1)(zk),,f(p)(zk)):=g(zk).\begin{split}g&=f(y_{k+p+1},...,y_{k+\tau},f^{(1)}(z_{k}),...,f^{(p)}(z_{k}))\\ &=f(z_{k}^{(p+1)},...,z_{k}^{(\tau)},f^{(1)}(z_{k}),...,f^{(p)}(z_{k}))\\ &:=g(z_{k}).\end{split}

Since gg is a function of only zkz_{k} and since yk+τ+p+1=gy_{k+\tau+p+1}=g, g(zk)g(z_{k}) satisfies the definition of a predictor function and hence is a (p+1)(p+1)-step predictor of (8)

yk+τ+p+1=zk+p+1(τ)=(F~p+1(zk))(τ)=f(p+1)(zk).\begin{split}y_{k+\tau+p+1}&=z_{k+p+1}^{(\tau)}=(\tilde{F}^{p+1}(z_{k}))^{(\tau)}=f^{(p+1)}(z_{k}).\end{split}

Therefore, for i=p+1i=p+1,

(F~i(zk))(τp1+j)=f(j)(zk)j{1,2,,(p+1)}(\tilde{F}^{i}(z_{k}))^{(\tau-p-1+j)}=f^{(j)}(z_{k})\quad j\in\{1,2,...,(p+1)\}

stating that the last (p+1)(p+1) entries of zk+p+1z_{k+p+1} are f(1)(x)f^{(1)}(x), f(2)(x)f^{(2)}(x), …, f(p+1)(x)f^{(p+1)}(x). Hence the proof. ∎

To identify a causal Koopman model for the NAR system (8), we propagate the model (VI) by τ\tau time steps to ensure no intersection of outputs between the states zk+1z_{k+1} and zkz_{k}. We define a new state xk=zkτx_{k}=z_{k\tau} which yields

xk\displaystyle x_{k} =zkτ\displaystyle=z_{k\tau}
xk+1\displaystyle\Rightarrow x_{k+1} =zkτ+τ=F~τ(zkτ)=F(xk)\displaystyle=z_{k\tau+\tau}=\tilde{F}^{\tau}(z_{k\tau})=F(x_{k}) (21)
where F\displaystyle\text{where }F =F~τ=F~F~F~τ times\displaystyle=\tilde{F}^{\tau}=\underbrace{\tilde{F}\circ\tilde{F}\circ\cdots\tilde{F}}_{\tau\text{ times}}

Using proposition 1, we can say that the nonlinear state space model contains functions that are 1-step, 2-step, …, τ\tau-step predictors in the following form

xk+1\displaystyle x_{k+1} =[ykτ+τ+1ykτ+τ+2ykτ+2τ1ykτ+2τ]:=[f1(ykτ+1,ykτ+2,,ykτ+τ)f2(ykτ+1,ykτ+2,,ykτ+τ)fτ1(ykτ+1,ykτ+2,,ykτ+τ)fτ(ykτ+1,ykτ+2,,ykτ+τ)]\displaystyle=\begin{bmatrix}y_{k\tau+\tau+1}\\ y_{k\tau+\tau+2}\\ \vdots\\ y_{k\tau+2\tau-1}\\ y_{k\tau+2\tau}\end{bmatrix}:=\begin{bmatrix}f_{1}(y_{k\tau+1},y_{k\tau+2},\cdots,y_{k\tau+\tau})\\ f_{2}(y_{k\tau+1},y_{k\tau+2},\cdots,y_{k\tau+\tau})\\ \vdots\\ f_{\tau-1}(y_{k\tau+1},y_{k\tau+2},\cdots,y_{k\tau+\tau})\\ f_{\tau}(y_{k\tau+1},y_{k\tau+2},\cdots,y_{k\tau+\tau})\\ \end{bmatrix}
:=[f(1)(ykτ+1,ykτ+2,,ykτ+τ)f(2)(ykτ+1,ykτ+2,,ykτ+τ)f(τ1)(ykτ+1,ykτ+2,,ykτ+τ)f(τ)(ykτ+1,ykτ+2,,ykτ+τ)]=F(xk)\displaystyle:=\begin{bmatrix}f^{(1)}(y_{k\tau+1},y_{k\tau+2},\cdots,y_{k\tau+\tau})\\ f^{(2)}(y_{k\tau+1},y_{k\tau+2},\cdots,y_{k\tau+\tau})\\ \vdots\\ f^{(\tau-1)}(y_{k\tau+1},y_{k\tau+2},\cdots,y_{k\tau+\tau})\\ f^{(\tau)}(y_{k\tau+1},y_{k\tau+2},\cdots,y_{k\tau+\tau})\end{bmatrix}=F(x_{k})

where f(i)f^{(i)} is the ii-step predictor of the NAR model. We prove the existence of a Koopman operator for this model in Proposition 2.

Proposition 2

If the function f(x)f(x) in the NAR model (8) is analytic, then a Koopman operator exists for (VI) and (VI).

Proof:

Since ff in (8) is analytic, F~\tilde{F} in (VI) is analytic since all the entries of F~\tilde{F} are either linear functions or are equal to ff. Since FF is obtained by the composition of F~\tilde{F} τ\tau times, FF is also analytic.

F(x)F(x) admits a countable-dimension Koopman operator KxK_{x}, with an invariant subspace isomorphic to either a finite or an infinite Taylor polynomial basis [34]. Moreover, isomorphism with a Taylor polynomial basis ensures that the Koopman observable space contains the full state observable, i.e. it is state inclusive.

There are two easy arguments to conclude the proof. First, note that since ff is analytic, fτf^{\tau} is analytic and thus by the same reasoning as in [34], fτf^{\tau} thus must admit a Koopman operator. The second argument is a constructive one, noting that equation

ψ(x[(k)τ])=Kτψ(x[(k1)(τ))\psi(x[(k)\tau])=K^{\tau}\psi(x[(k-1)(\tau)) (23)

must hold due to τ\tau applications of the 1-step Koopman equation. This means therefore that the following matrix equation must hold

ψ([x[(k)τ]x[kτ+1]x[(k+1)τ1]])=𝐊Jψ([(x[(k1)(τ))](x[(k1)]τ+1])(x[(k)τ1)]])\psi\left(\begin{bmatrix}x[(k)\tau]\\ x[k\tau+1]\\ \vdots\\ x[(k+1)\tau-1]\end{bmatrix}\right)={\bf K}_{J}\psi\left(\begin{bmatrix}(x[(k-1)(\tau))]\\ (x[(k-1)]\tau+1])\\ \vdots\\ (x[(k)\tau-1)]\end{bmatrix}\right) (24)

where 𝐊J=diag(Kτ,Kτ,Kτ).{\bf K}_{J}=\text{diag}\left(K^{\tau},K^{\tau},\ldots K^{\tau}\right). This concludes the proof. ∎

Since the existence of a Koopman operator has been proved for the model (VI) in Proposition 2, we construct a state inclusive dictionary of observables

ψ(xk)=[xkφ(xk)]\psi(x_{k})=\begin{bmatrix}x_{k}\\ \varphi(x_{k})\end{bmatrix} (25)

with φ:pτNp\varphi:\mathbb{R}^{p\tau}\rightarrow\mathbb{R}^{N_{p}} to define a Koopman model

ψ(xk+1)=Kψ(xk)\psi(x_{k+1})=K\psi(x_{k}) (26)

This Koopman model is causal since there is no intersection of outputs between xk+1x_{k+1} and xkx_{k}. The added feature of this model is that the DMD algorithm while identifying a Koopman operator, also simultaneously minimizes the 1-step, 2-step, …, τ\tau-step prediction error of the NAR model.

Now that we have a theoretical state space representation of a NAR model and established the conditions under which a Koopman operator exists, we turn our attention to the algorithm for identification of the Koopman operator. Given the data with M data sets and N data points in each data set {y1(i),y2(i),,yN(i)}\{y^{(i)}_{1},y^{(i)}_{2},...,y^{(i)}_{N}\} where i{1,2,M}i\in\{1,2,...M\} is the index of the data set, we construct the Hankel states zkz_{k} and the dictionary of observables allowing the intermixing of states. We compile the observables into snapshot matrices Ψ~f(z)\tilde{\Psi}_{f}(z) and Ψ~p(z)\tilde{\Psi}_{p}(z) with a τ\tau time step jump and solve the Koopman learning problem

Ψ~f(z)KΨ~p(z)F||\tilde{\Psi}_{f}(z)-K\tilde{\Psi}_{p}(z)||_{F}

using the methodology in Algorithm 1.

Algorithm 1 Extended DMD for NAR models
1:Get NAR model parameter τ\tau
2:Get extended DMD parameter non_{o} for monomial observables
3:for dataset i=1,2,,Mi=1,2,\ldots,M do
4:     for time index j=1,2,,Nτj=1,2,\ldots,N-\tau do
5:         Construct the Hankel state
zj(i)=[yj+1(i)yj+2(i)yj+τ(i)]z^{(i)}_{j}=\begin{bmatrix}y^{(i)}_{j+1}&y^{(i)}_{j+2}&\cdots y^{(i)}_{j+\tau}\end{bmatrix}
6:         Construct the dictionary of observables ψ(zj(i))\psi(z^{(i)}_{j})
7:     end for
8:     Construct the snapshot matrices for each data set with the τ\tau-jump
Ψp(i)(x)=[ψ(z1(i))ψ(z2(i))ψ(zN2τ(i))]Ψf(i)(x)=[ψ(z1+τ(i))ψ(z2+τ(i))ψ(zNτ(i))]\begin{split}\Psi^{(i)}_{p}(x)&=\begin{bmatrix}\psi(z^{(i)}_{1})&\psi(z^{(i)}_{2})&...&\psi(z^{(i)}_{N-2\tau})\end{bmatrix}\\ \Psi^{(i)}_{f}(x)&=\begin{bmatrix}\psi(z^{(i)}_{1+\tau})&\psi(z^{(i)}_{2+\tau})&...&\psi(z^{(i)}_{N-\tau})\end{bmatrix}\\ \end{split}
9:end for
10:Compile the snapshot matrices across data sets
Ψ~p(x)=[Ψp(1)(x)Ψp(2)(x)Ψp(M)(x)]Ψ~f(x)=[Ψf(1)(x)Ψf(2)(x)Ψf(M)(x)]\begin{split}\tilde{\Psi}_{p}(x)=\begin{bmatrix}\Psi^{(1)}_{p}(x)&\Psi_{p}^{(2)}(x)&...&\Psi_{p}^{(M)}(x)\end{bmatrix}\\ \tilde{\Psi}_{f}(x)=\begin{bmatrix}\Psi_{f}^{(1)}(x)&\Psi_{f}^{(2)}(x)&...&\Psi_{f}^{(M)}(x)\end{bmatrix}\ \end{split}
11:Compute the SVD of Ψ~p(x)=USV\tilde{\Psi}_{p}(x)=USV^{*}
12:Truncate to required number of singular values and identify the Koopman operator
K^=Ψ~f(x)V~S~1U~\hat{K}=\tilde{\Psi}_{f}(x)\tilde{V}\tilde{S}^{-1}\tilde{U}^{*}

VII Results

From the data-sets obtained in the plate reader experiments shown in Fig. 1, we used Algorithm 1 to implement extended DMD using monomials as the dictionary of observables

ψ(zk)=[yk+1,,yk+τ,yk+12,yk+1yk+2,,yk+τ2yk+13,yk+12yk+2,]T.\psi(z_{k})=[y_{k+1},...,y_{k+\tau},y^{2}_{k+1},y_{k+1}y_{k+2},...,y^{2}_{k+\tau}\\ y^{3}_{k+1},y^{2}_{k+1}y_{k+2},...]^{T}.\ (27)

to identify an approximate Koopman operator for the state space model (VI) as a solution to the identification of the NAR model (8).

We use all the datasets in Fig. 1 to find a Koopman operator invariant to the substrate concentrations. They are broken equally into training, validation and test set. Given the two parameters τ\tau (NAR model parameter) and non_{o} (extended DMD parameter), we can find the optimal approximate Koopman operator by cumulatively iterating through the principal components and evaluating the summation of the mean squared error(MSE) of training and validation data. The number of principal components corresponding to the minimum MSE yields the optimal Koopman operator for a given τ\tau and non_{o}. We then iterate through the two parameters to find the optimal model that minimizes the

By choosing τ=9\tau=9 and keeping the maximum order of monomials to 3, the Koopman operator has been identified and the prediction on the training data is shown in Figure 2 and it has an MSE of 3.4%. The identified Koopman operator has an MSE of 9% and the fit is shown in Figure 3.

Refer to caption
Figure 2: The identified Koopman operator is tested on the training sets with 9 point initial condition and up to 3rd3^{rd} order monomials to get a MSE of 3.4%
Refer to caption
Figure 3: The identified Koopman operator is tested on the test sets by using the initial observablesψ(x0)\psi(x_{0}) and the mean squared error remains the same as that of the training set.

The results on the experimental data suggest that Causal Jump DMD is a suitable candidate algorithm for identifying the Koopman operator of the population growth dynamics of bacteria and can also be extended in general to identify Koopman operators for NAR models.

VIII Conclusion

In this paper, we introduced the microbial growth curve dynamics to motivate the usage of DMD algorithms to identify Koopman operators for NAR models. We formulated Hankel DMD as a state space representations of the NAR model and showed that it is restrictive in its structure. We construct a causal state space model for the NAR model and identify a Koopman operator for it using extended dynamic mode decomposition with a monomial dictionary of observables. We showed that it does a good job in predicting the population growth dynamics of Pseudomonas putida invariant to substrate concentrations. The future goals of this work is to use this model to identify the optimal media conditions for maximal and minimal growth of the microbe thereby enabling us to develop a general methodology to develop an external growth harness for microbes for dynamic growth control. To achieve this, we need to extend the mathematical models to allow for inputs and extend the identification to NARX and NARMAX models. Further, if we integrate this framework with deepDMD which aids in finding the observable functions in a parsimonious fashion, it renders a useful tool for identifying high dimensional linear models for nonlinear systems.

IX Acknowledgments

The authors gratefully acknowledge the funding of DARPA grants FA8750-17-C-0229, HR001117C0092, HR001117C0094, DEAC0576RL01830. The authors would also like to thank Professors Arun K. Tangirala, Igor Mezic, Milan Korda, Alexandre Mauroy, Nathan Kutz, Steve Haase, John Harer, Devin Strickland, and Eric Klavins for insightful discussions. Any opinions, findings, conclusions, or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the Defense Advanced Research Project Agency, the Department of Defense, or the United States government. This material is based on work supported by DARPA and AFRL under contract numbers FA8750-17-C-0229, HR001117C0092, HR001117C0094, DEAC0576RL01830.

References

  • [1] D. Kornberg and D. TA, “Replication,” San Francisco: W H. Freeman, 1980.
  • [2] N. F. Mathon and A. C. Lloyd, “Cell senescence and cancer,” Nature Reviews Cancer, vol. 1, no. 3, p. 203, 2001.
  • [3] G. Wu, Q. Yan, J. A. Jones, Y. J. Tang, S. S. Fong, and M. A. Koffas, “Metabolic burden: cornerstones in synthetic biology and metabolic engineering applications,” Trends in biotechnology, vol. 34, no. 8, pp. 652–664, 2016.
  • [4] D. S. Glazier, “Is metabolic rate a universal ‘pacemaker’for biological processes?” Biological Reviews, vol. 90, no. 2, pp. 377–407, 2015.
  • [5] D. De Martino, F. Capuani, and A. De Martino, “Growth against entropy in bacterial metabolism: the phenotypic trade-off behind empirical growth rate distributions in e. coli,” Physical biology, vol. 13, no. 3, p. 036005, 2016.
  • [6] B. J. Sanchez, C. Zhang, A. Nilsson, P.-J. Lahtvee, E. J. Kerkhoven, and J. Nielsen, “Improving the phenotype predictions of a yeast genome-scale metabolic model by incorporating enzymatic constraints,” Molecular systems biology, vol. 13, no. 8, 2017.
  • [7] M. Zwietering, I. Jongenburger, F. Rombouts, and K. Van’t Riet, “Modeling of the bacterial growth curve,” Appl. Environ. Microbiol., vol. 56, no. 6, pp. 1875–1881, 1990.
  • [8] T. Tschirhart, V. Shukla, E. E. Kelly, Z. Schultzhaus, E. NewRingeisen, J. S. Erickson, Z. Wang, W. garcia, E. Curl, R. G. Egbert et al., “Synthetic biology tools for the fast-growing marine bacterium vibrio natriegens,” ACS synthetic biology, 2019.
  • [9] N. Khan, E. Yeung, Y. Farris, S. J. Fansler, and H. C. Bernstein, “A broad-host-range event detector: expanding and quantifying performance across bacterial species,” bioRxiv, p. 369967, 2018.
  • [10] C. Gill and K. Tan, “Effect of carbon dioxide on growth of pseudomonas fluorescens.” Appl. Environ. Microbiol., vol. 38, no. 2, pp. 237–240, 1979.
  • [11] D. M. Gulliver, G. V. Lowry, and K. B. Gregory, “Comparative study of effects of co2 concentration and ph on microbial communities from a saline aquifer, a depleted oil reservoir, and a freshwater aquifer,” Environmental Engineering Science, vol. 33, no. 10, pp. 806–816, 2016.
  • [12] A. E. LaBauve and M. J. Wargo, “Growth and laboratory maintenance of pseudomonas aeruginosa,” Current protocols in microbiology, vol. 25, no. 1, pp. 6E–1, 2012.
  • [13] A. P. Palacios, J. M. Marín, E. J. Quinto, M. P. Wiper et al., “Bayesian modeling of bacterial growth for multiple populations,” The Annals of Applied Statistics, vol. 8, no. 3, pp. 1516–1537, 2014.
  • [14] H. H. Lee, N. Ostrov, B. G. Wong, M. A. Gold, A. Khalil, and G. M. Church, “Vibrio natriegens, a new genomic powerhouse,” bioRxiv, p. 058487, 2016.
  • [15] I. Mezic, “Spectral properties of dynamical systems, model reduction and decompositions,” Nonlinear Dynamics, vol. 41, no. 1-3, pp. 309–325, 2005.
  • [16] M. Budišić, R. Mohr, and I. Mezić, “Applied koopmanism,” Chaos: An Interdisciplinary Journal of Nonlinear Science, vol. 22, no. 4, p. 047510, 2012.
  • [17] M. O. Williams, I. G. Kevrekidis, and C. W. Rowley, “A data–driven approximation of the koopman operator: Extending dynamic mode decomposition,” Journal of Nonlinear Science, vol. 25, no. 6, pp. 1307–1346, 2015.
  • [18] C. W. Rowley, I. Mezić, S. Bagheri, P. Schlatter, and D. S. Henningson, “Spectral analysis of nonlinear flows,” Journal of fluid mechanics, vol. 641, pp. 115–127, 2009.
  • [19] J. L. Proctor, S. L. Brunton, and J. N. Kutz, “Dynamic mode decomposition with control,” SIAM Journal on Applied Dynamical Systems, vol. 15, no. 1, pp. 142–161, 2016.
  • [20] M. O. Williams, M. S. Hemati, S. T. Dawson, I. G. Kevrekidis, and C. W. Rowley, “Extending data-driven koopman analysis to actuated systems,” IFAC-PapersOnLine, vol. 49, no. 18, pp. 704–709, 2016.
  • [21] T. Askham and J. N. Kutz, “Variable projection methods for an optimized dynamic mode decomposition,” SIAM Journal on Applied Dynamical Systems, vol. 17, no. 1, pp. 380–416, 2018.
  • [22] Y. Kaneko, S. Muramatsu, H. Yasuda, K. Hayasaka, Y. Otake, S. Ono, and M. Yukawa, “Convolutional-sparse-coded dynamic mode decomposition and its application to river state estimation,” in ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).   IEEE, 2019, pp. 1872–1876.
  • [23] O. Azencot, W. Yin, and A. Bertozzi, “Consistent dynamic mode decomposition,” arXiv preprint arXiv:1905.09736, 2019.
  • [24] K. Manohar, E. Kaiser, S. L. Brunton, and J. N. Kutz, “Optimized sampling for multiscale dynamics,” Multiscale Modeling & Simulation, vol. 17, no. 1, pp. 117–136, 2019.
  • [25] P. J. Schmid, “Dynamic mode decomposition of numerical and experimental data,” Journal of fluid mechanics, vol. 656, pp. 5–28, 2010.
  • [26] S. Sinha and E. Yeung, “On computation of koopman operator from sparse data,” arXiv:1901.03024, 2019.
  • [27] E. Yeung, S. Kundu, and N. Hodas, “Learning deep neural network representations for koopman operators of nonlinear dynamical systems,” in 2019 American Control Conference (ACC).   IEEE, 2019, pp. 4832–4839.
  • [28] A. Hasnain, S. Sinha, Y. Dorfan, A. E. Borujeni, Y. Park, P. Maschhoff, U. Saxena, J. Urrutia, N. Gaffney, D. Becker et al., “A data-driven method for quantifying the impact of a genetic circuit on its host,” arXiv preprint arXiv:1909.06455, 2019.
  • [29] C. A. Johnson and E. Yeung, “A class of logistic functions for approximating state-inclusive koopman operators,” in 2018 Annual American Control Conference (ACC).   IEEE, 2018, pp. 4803–4810.
  • [30] S. E. Otto and C. W. Rowley, “Linearly recurrent autoencoder networks for learning dynamics,” SIAM Journal on Applied Dynamical Systems, vol. 18, no. 1, pp. 558–593, 2019.
  • [31] N. Takeishi, Y. Kawahara, and T. Yairi, “Learning koopman invariant subspaces for dynamic mode decomposition,” in Advances in Neural Information Processing Systems, 2017, pp. 1130–1140.
  • [32] Q. Li, F. Dietrich, E. M. Bollt, and I. G. Kevrekidis, “Extended dynamic mode decomposition with dictionary learning: A data-driven adaptive spectral decomposition of the koopman operator,” Chaos: An Interdisciplinary Journal of Nonlinear Science, vol. 27, no. 10, p. 103111, 2017.
  • [33] P. You, J. Pang, and E. Yeung, “Deep koopman controller synthesis for cyber-resilient market-based frequency regulation,” IFAC-PapersOnLine, vol. 51, no. 28, pp. 720–725, 2018.
  • [34] E. Yeung, Z. Liu, and N. O. Hodas, “A koopman operator approach for computing and balancing gramians for discrete time nonlinear systems,” in 2018 Annual American Control Conference (ACC).   IEEE, 2018, pp. 337–344.
  • [35] J. Monod, “The growth of bacterial cultures,” Annual review of microbiology, vol. 3, no. 1, pp. 371–394, 1949.
  • [36] B. W. Brandt, I. M. van Leeuwen, and S. A. Kooijman, “A general model for multiple substrate biodegradation. application to co-metabolism of structurally non-analogous compounds,” Water research, vol. 37, no. 20, pp. 4843–4854, 2003.
  • [37] D. S. Kompala, D. Ramkrishna, N. B. Jansen, and G. T. Tsao, “Investigation of bacterial growth on mixed substrates: experimental evaluation of cybernetic models,” Biotechnology and Bioengineering, vol. 28, no. 7, pp. 1044–1055, 1986.
  • [38] H. Arbabi and I. Mezic, “Ergodic theory, dynamic mode decomposition, and computation of spectral properties of the koopman operator,” SIAM Journal on Applied Dynamical Systems, vol. 16, no. 4, pp. 2096–2126, 2017.