This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Interpolatory tensorial reduced order models for parametric dynamical systems

Alexander V. Mamonov Department of Mathematics, University of Houston, Houston, Texas 77204 ([email protected]).    Maxim A. Olshanskii Department of Mathematics, University of Houston, Houston, Texas 77204 ([email protected]).
Abstract

The paper introduces a reduced order model (ROM) for numerical integration of a dynamical system which depends on multiple parameters. The ROM is a projection of the dynamical system on a low dimensional space that is both problem-dependent and parameter-specific. The ROM exploits compressed tensor formats to find a low rank representation for a sample of high-fidelity snapshots of the system state. This tensorial representation provides ROM with an orthogonal basis in a universal space of all snapshots and encodes information about the state variation in parameter domain. During the online phase and for any incoming parameter, this information is used to find a reduced basis that spans a parameter-specific subspace in the universal space. The computational cost of the online phase then depends only on tensor compression ranks, but not on space or time resolution of high-fidelity computations. Moreover, certain compressed tensor formats enable to avoid the adverse effect of parameter space dimension on the online costs (known as the curse of dimension). The analysis of the approach includes an estimate for the representation power of the acquired ROM basis. We illustrate the performance and prediction properties of the ROM with several numerical experiments, where tensorial ROM’s complexity and accuracy is compared to those of conventional POD-ROM.

keywords:
Model order reduction, parametric PDEs, low-rank tensors, dynamical systems, proper orthogonal decomposition

1 Introduction

In numerical optimal control, inverse modeling or uncertainty quantification, one commonly needs to integrate a parameter-dependent dynamical system for various values of the parameter vector. For example, inverse modeling may require repeated solutions of the forward problem represented by a dynamical system, along the search path in a high-dimensional parameter space. This may lead to extreme-scale computations that, if implemented straightforwardly, often result in overwhelming computational costs. Reduced order models (ROMs) offer a possibility to alleviate these costs by replacing a high-fidelity model with a low-dimensional surrogate model [3, 33]. Thanks to this practical value and apparent success, ROMs for parametric dynamical systems have already attracted considerable attention; see, e.g., [10, 39, 14, 5, 9, 13].

In this paper, we are interested in projection based ROMs that build the surrogate model by projecting a high-fidelity model onto a low-dimensional problem-dependent vector space [10]. Projection-based ROMs for dynamical systems include such well-known model order reduction techniques as proper orthogonal decomposition (POD) ROMs [54, 64] (and its variants such as POD-DEIM [17] and balanced POD [61]) and PGD-ROMs [18, 19]. In these approaches the basis for the projection space is computed by building on the information about the dynamical system provided through high-fidelity solutions sampled for certain time instances and/or parameter values, the so-called solution snapshots. However, building a general low-dimensional space for all times and parameters of interest might be challenging, if possible at all, for wide parameter ranges and long times. Several studies aimed to address this challenge: In [26, 25, 2] the authors considered partitioning strategies which introduce a subdivision of the parameter domain and assign an individual local reduced-order basis to each subdomain offline. Another idea [1, 65] is to adapt precomputed equal-dimension reduced order spaces by interpolating them for out-of-sample parameters along geodesics on the Grassmann manifold. The present paper introduces a different approach that builds on recent developments in tensor decompositions and low-rank approximations to quickly compute parameter-specific reduced bases for projection based ROMs.

For parametrized systems of time-dependent differential equations, the generated data (the input of ROM) naturally takes a form of a multi-dimensional tensor of solution snapshots, with dimensionality D+2D+2, where DD is a the dimension of the parameter space and 2 accounts for the spatial and time-wise distributions. Modern ROMs often proceed by unfolding such tensors into a matrix to perform standard POD based on truncated SVD. This leads to the loss of information about the dependency of solutions on parameters. We propose to overcome these issues by working directly with tensor data, and by exploiting low-rank tensor approximations based on the canonical polyadic (CP), high order SVD (HOSVD), and Tensor Train (TT) decompositions. The approach consists of two stages. First, at the offline stage, the compressed snapshot tensor is computed using one of the three tensor decompositions, thus preserving the essential information about variation of the solution with respect to parameters. Up to the compression accuracy, each of these decompositions provides a (global) basis for the universal space spanned by all observed snapshots. The so-called core of the compressed representation is then transmitted to the second stage, referred to as the online stage. At the online stage the transmitted part of compressed tensor allows for a fast computation of a parameter-specific reduced basis for any incoming out-of-sample parameter vector through an interpolation and fast linear algebra routines. The reduced order basis is then given in terms of its coordinates in the global basis that can be stored offline. For CP and TT formats, the cost of these computations is free of exponential growth with respect to the parameter space dimension. On analysis side of this work, we prove an estimate for prediction power of the parameter-specific reduced order basis. The estimate explicitly depends on the approximation accuracy of the original tensor by the compressed one, parameter interpolation error, and singular values of a small-size parameter-specific matrix.

Despite an outstanding recent progress in numerical multi-linear algebra and, in particular, in understanding tensor decompositions (see, e.g., review articles  [46, 29, 63]), the application of tensor methods in reduced order modeling of dynamical systems is still rather scarce. We mention two reports by Nouy [55, 56], who reviewed tensor compressed formats and discussed their possible use for sparse function representation and reduced order modeling, as well as a series of publications on the treatment in compressed tensor formats of algebraic systems resulting from the stochastic and parametric Galerkin finite element method, see e.g.  [11, 7, 8, 49, 48]. A POD-ROM was combined with a low-rank tensor representation of a mapping from a parameter space onto an output domain in [42]. The authors of survey [10] observe that “The combination of tensor calculus \ldots and parametric model reduction techniques for time dependent problems is still in its infancy, but offers a promising research direction”. We believe the statement holds true, and the present study contributes to this largely open research field.

The remainder of the paper is organized as follows. In Section 2 we set up a parameter-dependent Cauchy problem and recall the basics of POD-ROM approach that is needed for reference purpose later in the text. Section 3 introduces a general idea of the interpolatory tensorial ROM and considers its realization using three popular tensor compression formats. Details are worked out for a Cartesian grid-based sampling of the parameter domain, and then the approach is extended to a more general parameter sampling scheme. A separate subsection discusses online–offline complexity and storage requirements of the method. An estimate on the prediction power of the reduced order basis is proved in Section 4. Numerical examples in Section 5 illustrate the analysis and performance of the method. In particular, we compare the delivered accuracy with standard POD-ROM that employs a global low-dimensional basis.

2 Parameterized Cauchy problem and the conventional POD-ROM

To fix ideas, consider the following multi-parameter initial value problem. For a vector of parameters 𝜶=(α1,,αD)\mbox{\boldmath$\alpha$\unboldmath}=(\alpha_{1},\dots,\alpha_{D}) from the parameter domain 𝒜D\mathcal{A}\subset\mathbb{R}^{D} find the trajectory 𝐮=𝐮(t,𝜶):[0,T)M\mathbf{u}=\mathbf{u}(t,\mbox{\boldmath$\alpha$\unboldmath}):[0,T)\to\mathbb{R}^{M} solving

𝐮t=F(t,𝐮,𝜶),t(0,T),and𝐮|t=0=𝐮0,\mathbf{u}_{t}=F(t,\mathbf{u},\mbox{\boldmath$\alpha$\unboldmath}),\quad t\in(0,T),\quad\text{and}~{}\mathbf{u}|_{t=0}=\mathbf{u}_{0}, (1)

with a given continuous flow field F:(0,T)×M×𝒜MF:(0,T)\times\mathbb{R}^{M}\times\mathcal{A}\to\mathbb{R}^{M}. Hereafter we denote all vector quantities by bold lowercase letters. We assume that the unique solution exists on (0,T)(0,T) for all 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A}. Examples considered in this paper include parameter-dependent parabolic equations, in which case one can think of (1) as a system of ODEs for nodal values of the finite volume or finite element solution to the PDE problem, where material coefficients, boundary conditions, or the computational domain (via a mapping into a reference domain) are parameterized by 𝜶\alpha.

We are interested in projection based ROMs, where for an arbitrary but fixed 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A} an approximation to 𝐮\mathbf{u} is sought as a solution to equations projected onto a reduced space. Projection based approaches aim at retaining the structure of the model and thus at preserving the physics present in the high-fidelity model [10]. Among the projection based approaches to model reduction for time-dependent differential equations, Proper Orthogonal Decomposition (POD) and its variants are likely the most widely used ROM technique, which provides tools to represent trajectories of a dynamical system in a low-dimensional, problem-dependent basis [43, 60, 50, 52]. We summarize the POD-ROM below for further reference and for the purpose of comparison to our approach in Section 5.

Assume for a moment that 𝜶\alpha is fixed. The POD-ROM computes a representative collection of states ϕk(𝜶)=𝐮(tk,𝜶)M\boldsymbol{\phi}_{k}(\mbox{\boldmath$\alpha$\unboldmath})=\mathbf{u}(t_{k},\mbox{\boldmath$\alpha$\unboldmath})\in\mathbb{R}^{M} at times 0t1,,tN<T0\leq t_{1},\dots,t_{N}<T, referred to as snapshots, through high-fidelity numerical simulations. Next, one finds a parameter-specific low-dimensional basis {𝐳ipod(𝜶)}i=1nM\{\mathbf{z}_{i}^{\rm pod}(\mbox{\boldmath$\alpha$\unboldmath})\}_{i=1}^{n}\subset\mathbb{R}^{M}, nNn\ll N, referred to hereafter as the reduced basis, such that the projection subspace span{𝐳1pod(𝜶),,𝐳npod(𝜶)}\mbox{span}\big{\{}\mathbf{z}_{1}^{\rm pod}(\mbox{\boldmath$\alpha$\unboldmath}),\dots,\mathbf{z}_{n}^{\rm pod}(\mbox{\boldmath$\alpha$\unboldmath})\big{\}} approximates the snapshot space span{ϕ1(𝜶),,ϕN(𝜶)}\mbox{span}\{\boldsymbol{\phi}_{1}(\mbox{\boldmath$\alpha$\unboldmath}),\dots,\boldsymbol{\phi}_{N}(\mbox{\boldmath$\alpha$\unboldmath})\} in the best possible way.

To determine the reduced basis, form a matrix of snapshots

Φpod(𝜶)=[ϕ1(𝜶),,ϕN(𝜶)]M×N,\Phi_{\text{pod}}(\mbox{\boldmath$\alpha$\unboldmath})=[\boldsymbol{\phi}_{1}(\mbox{\boldmath$\alpha$\unboldmath}),\ldots,\boldsymbol{\phi}_{N}(\mbox{\boldmath$\alpha$\unboldmath})]\in\mathbb{R}^{M\times N}, (2)

compute its SVD

Φpod(𝜶)=UΣVT,\Phi_{\text{pod}}(\mbox{\boldmath$\alpha$\unboldmath})=\mathrm{U}\Sigma\mathrm{V}^{T}, (3)

and define 𝐳ipod(𝜶)\mathbf{z}_{i}^{\rm pod}(\mbox{\boldmath$\alpha$\unboldmath}), i=1,,ni=1,\dots,n, to be the first nn left singular vectors of Φpod(𝜶)\Phi_{\text{pod}}(\mbox{\boldmath$\alpha$\unboldmath}), i.e., the first nn columns of U\mathrm{U}. Hereafter we denote all matrices with upright capital letters. The singular values in Σ\Sigma provide information about the approximation power of span{𝐳1pod(𝜶),,𝐳npod(𝜶)}\mbox{span}\big{\{}\mathbf{z}_{1}^{\rm pod}(\mbox{\boldmath$\alpha$\unboldmath}),\dots,\mathbf{z}_{n}^{\rm pod}(\mbox{\boldmath$\alpha$\unboldmath})\big{\}}. We refer to [51, 43] and references therein for a discussion about algebraically different ways to define POD and their equivalence.

For parameters 𝜶\alpha varying in 𝒜\mathcal{A}, a parametric POD-ROM builds a global reduced basis by sampling the parameter domain, generating snapshots for each sampled parameter value and proceeding with SVD  (3) for a cumulative matrix of all snapshots. Possible sampling strategies include using a Cartesian grid in 𝒜\mathcal{A}, Monte–Carlo methods, and greedy algorithms based on a posteriori error estimates; see, e.g., [10, 39]. Regardless of the sampling procedure, the resulting basis can accurately reproduce only the data from which it originated. Without parameter-specificity, the basis may lack robustness for out-of-sample parameters, i.e., away from the reference simulations. This is a serious limitation for using POD based ROMs in inverse modeling. We plan to address this limitation by introducing tensorial techniques for finding reduced bases that are both problem- and parameter-specific.

3 Tensorial ROMs

We first consider in Section 3.1 a Cartesian grid-based sampling of the parameter domain 𝒜\mathcal{A} in the case when 𝒜\mathcal{A} is the DD-dimensional box

𝒜=i=1D[αimin,αimax],\mathcal{A}=\bigotimes_{i=1}^{D}[\alpha_{i}^{\min},\alpha_{i}^{\max}], (4)

and the sampling points are placed at the nodes of a Cartesian grid. Next, we describe three tensorial ROMs (TROMs) based on three different tensor decompositions, canonical polyadic (CP, Section 3.5), high order SVD (HOSVD, Section 3.6) and tensor train (TT, Section 3.7).

3.1 Cartesian grid-based parameter sampling

To generate the sampling set 𝒜^{\widehat{\mathcal{A}}}, we distribute nin_{i} nodes {α^ij}j=1,,ni\{\widehat{\alpha}_{i}^{j}\}_{j=1,\dots,n_{i}} within each of the intervals [αimin,αimax][\alpha_{i}^{\min},\alpha_{i}^{\max}] in (4) for i=1,,Di=1,\dots,D, and define

𝒜^={𝜶^=(α^1,,α^D)T:α^i{α^ij}j=1,,ni,i=1,,D}.{\widehat{\mathcal{A}}}=\left\{\widehat{\boldsymbol{\alpha}}=(\widehat{\alpha}_{1},\dots,\widehat{\alpha}_{D})^{T}\,:\,\widehat{\alpha}_{i}\in\{\widehat{\alpha}_{i}^{j}\}_{j=1,\dots,n_{i}},~{}i=1,\dots,D\right\}. (5)

Hereafter we use hats to denote parameters from the sampling set 𝒜^{\widehat{\mathcal{A}}}, and the cardinality of 𝒜^{\widehat{\mathcal{A}}} is denoted by

K=i=1Dni.K=\prod_{i=1}^{D}n_{i}. (6)

The corresponding snapshots ϕk(𝜶^)=𝐮(tk,𝜶^)\boldsymbol{\phi}_{k}(\widehat{\boldsymbol{\alpha}})=\mathbf{u}(t_{k},\widehat{\boldsymbol{\alpha}}), 𝜶^𝒜^\widehat{\boldsymbol{\alpha}}\in{\widehat{\mathcal{A}}}, are organized in a multi-dimensional array

(𝚽):,i1,,iD,k=ϕk(α^1i1,,α^DiD),(\boldsymbol{\Phi})_{:,i_{1},\dots,i_{D},k}=\boldsymbol{\phi}_{k}(\widehat{\alpha}_{1}^{i_{1}},\dots,\widehat{\alpha}_{D}^{i_{D}}), (7)

which is a tensor of order D+2D+2 and size M×n1××nD×NM\times n_{1}\times\dots\times n_{D}\times N. We reserve the first and the last indices of 𝚽\boldsymbol{\Phi} for the spatial and temporal distributions, respectively. All tensors hereafter are denoted with bold uppercase letters.

Unfolding 𝚽\boldsymbol{\Phi} along the first index in an M×(n1nD)NM\times(n_{1}\cdots n_{D})N matrix and applying (truncated) SVD to determine the first nn left singular vectors is equivalent to the POD with grid-based parameter sampling. The disadvantage of this approach for ROM construction is that it neglects any information about the dependence of snapshots on parameter variation reflected in the tensor structure of 𝚽\boldsymbol{\Phi}. To preserve this information, we proceed with a compressed approximation 𝚽~\widetilde{\boldsymbol{\Phi}} of 𝚽\boldsymbol{\Phi} rather than with the low rank approximation of the unfolded matrix.

3.2 Tensor compression and universal space

The notion of a tensor rank and low-rank tensor approximation is ambiguous and later in this section we consider three popular compressed tensor formats. For now we only assume that 𝚽~\widetilde{\boldsymbol{\Phi}} satisfies

𝚽~𝚽Fε~𝚽F\big{\|}\widetilde{\boldsymbol{\Phi}}-\boldsymbol{\Phi}\big{\|}_{F}\leq\widetilde{\varepsilon}\big{\|}\boldsymbol{\Phi}\big{\|}_{F} (8)

for some small ε~>0\widetilde{\varepsilon}>0, where tensor Frobenius norm is simply

𝚽F:=(j=1Mi1=1n1iD=1nDk=1N𝚽j,i1,,iD,k2)1/2\|\boldsymbol{\Phi}\|_{F}:=\Big{(}\sum_{j=1}^{M}\sum_{i_{1}=1}^{n_{1}}\dots\sum_{i_{D}=1}^{n_{D}}\sum_{k=1}^{N}\boldsymbol{\Phi}_{j,i_{1},\dots,i_{D},k}^{2}\Big{)}^{1/2} (9)

The ”low-rank” (compressed) tensor 𝚽~\widetilde{\boldsymbol{\Phi}} is computed during the first, offline stage of TROM construction and a part of 𝚽~\widetilde{\boldsymbol{\Phi}} is passed on to the second, online stage which uses this information about variation of snapshots with respect to changes in parameters to compute a parameter-specific TROM.

We call universal space the space V~\widetilde{V} spanned by the first-mode fibers of 𝚽~\widetilde{\boldsymbol{\Phi}}, i.e., V~\widetilde{V} is the column space of the mode-1 unfolding matrix. For the exact decomposition (i.e., for ε~=0\widetilde{\varepsilon}=0), V~\widetilde{V} is the space of all observed system states. In general, V~\widetilde{V} depends on a compression format, dimension of V~\widetilde{V} does not exceed MM and depends on ε~\widetilde{\varepsilon} and snapshot variation. We shall see that V~\widetilde{V} does approximate the full space of high-fidelity snapshots, while the CP, HOSVD and TT formats all deliver an orthogonal basis for V~\widetilde{V}. In the online stage of TROM we find a local (parameter-specific) ROM basis by specifying its coordinates in V~\widetilde{V}.

3.3 In-sample prediction

During the online stage we wish to be able to approximately solve (1) for an arbitrary parameter 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A} in an 𝛂\alpha-specific reduced basis. For this step we need to introduce the notion of kk-mode tensor-vector product 𝚿×k𝐚\boldsymbol{\Psi}\times_{k}\mathbf{a} of a tensor 𝚿N1××Nm\boldsymbol{\Psi}\in\mathbb{R}^{N_{1}\times\dots\times N_{m}} of order mm and a vector 𝐚Nk\mathbf{a}\in\mathbb{R}^{N_{k}}: the resulting tensor 𝚿×k𝐚\boldsymbol{\Psi}\times_{k}\mathbf{a} has order m1m-1 and size N1××Nk1×Nk+1××NmN_{1}\times\dots\times N_{k-1}\times N_{k+1}\times\dots\times N_{m}. Specifically, elementwise

(𝚿×k𝐚)j1,,jk1,jk+1,,jm=jk=1Nk𝚿j1,,jmajk.(\boldsymbol{\Psi}\times_{k}\mathbf{a})_{j_{1},\dots,j_{k-1},j_{k+1},\dots,j_{m}}=\sum_{j_{k}=1}^{N_{k}}\boldsymbol{\Psi}_{j_{1},\dots,j_{m}}a_{j_{k}}. (10)

For a moment, consider some 𝜶^=(α^1,,α^D)T\widehat{\boldsymbol{\alpha}}=\left(\widehat{\alpha}_{1},\dots,\widehat{\alpha}_{D}\right)^{T} from the sampling set 𝒜^{\widehat{\mathcal{A}}} and define DD vectors, 𝐞i(𝜶^)=(e1i(𝜶^),,enii(𝜶^))Tni\mathbf{e}^{i}(\widehat{\boldsymbol{\alpha}})=\left(e_{1}^{i}(\widehat{\boldsymbol{\alpha}}),\dots,e_{n_{i}}^{i}(\widehat{\boldsymbol{\alpha}})\right)^{T}\in\mathbb{R}^{n_{i}}, i=1,,Di=1,\dots,D, as

eji(𝜶^)={1ifα^i=α^ij0otherwise,j=1,,ni.e^{i}_{j}(\widehat{\boldsymbol{\alpha}})=\begin{cases}1&\text{if}~{}\widehat{\alpha}_{i}=\widehat{\alpha}_{i}^{j}\\ 0&\text{otherwise}\end{cases},\qquad j=1,\dots,n_{i}. (11)

In other words, 𝐞i(𝜶^)\mathbf{e}^{i}(\widehat{\boldsymbol{\alpha}}) encodes the position of α^i\widehat{\alpha}_{i} among the grid nodes on [αimin,αimax][\alpha_{i}^{\min},\alpha_{i}^{\max}], i=1,,Di=1,\ldots,D.

Vectors 𝐞i(𝜶^)\mathbf{e}^{i}(\widehat{\boldsymbol{\alpha}}) defined above allow us to extract the snapshots corresponding to a particular 𝜶^𝒜^\widehat{\boldsymbol{\alpha}}\in{\widehat{\mathcal{A}}}. Specifically, we introduce the following extraction operation

Φe(𝜶^)=𝚽×2𝐞1(𝜶^)×3𝐞2(𝜶^)×D+1𝐞D(𝜶^)M×N,\Phi_{e}(\widehat{\boldsymbol{\alpha}})=\boldsymbol{\Phi}\times_{2}\mathbf{e}^{1}(\widehat{\boldsymbol{\alpha}})\times_{3}\mathbf{e}^{2}(\widehat{\boldsymbol{\alpha}})\dots\times_{D+1}\mathbf{e}^{D}(\widehat{\boldsymbol{\alpha}})\in\mathbb{R}^{M\times N}, (12)

which extracts from tensor of all snapshots 𝚽\boldsymbol{\Phi} the matrix of snapshots (2) for the particular 𝜶^𝒜^\widehat{\boldsymbol{\alpha}}\in{\widehat{\mathcal{A}}}, i.e., Φe(𝜶^)=Φpod(𝜶^)\Phi_{e}(\widehat{\boldsymbol{\alpha}})=\Phi_{\text{pod}}(\widehat{\boldsymbol{\alpha}}).

Combining (12) with compressed approximation (8), we conclude that is should be possible to extract from 𝚽~\widetilde{\boldsymbol{\Phi}} the information about the space spanned by the snapshots {𝐮(ti,𝜶^)}i=1N\{\mathbf{u}(t_{i},\widehat{\boldsymbol{\alpha}})\}_{i=1}^{N}, for a particular 𝜶^𝒜^\widehat{\boldsymbol{\alpha}}\in{\widehat{\mathcal{A}}} up to the accuracy of approximation in (8). Indeed, let ϕi(𝜶^)=𝐮(ti,𝜶^)\boldsymbol{\phi}_{i}(\widehat{\boldsymbol{\alpha}})=\mathbf{u}(t_{i},\widehat{\boldsymbol{\alpha}}), i=1,,Ni=1,\dots,N, and denote by {𝐳j(𝜶^)}j=1N~\{\mathbf{z}_{j}(\widehat{\boldsymbol{\alpha}})\}_{j=1}^{\widetilde{N}}, N~N\widetilde{N}\leq N, an orthonormal basis for the column space of

Φ~e(𝜶^)=𝚽~×2𝐞1(𝜶^)×3𝐞2(𝜶^)×D+1𝐞D(𝜶^)M×N,\widetilde{\Phi}_{e}(\widehat{\boldsymbol{\alpha}})=\widetilde{\boldsymbol{\Phi}}\times_{2}\mathbf{e}^{1}(\widehat{\boldsymbol{\alpha}})\times_{3}\mathbf{e}^{2}(\widehat{\boldsymbol{\alpha}})\dots\times_{D+1}\mathbf{e}^{D}(\widehat{\boldsymbol{\alpha}})\in\mathbb{R}^{M\times N}, (13)

where N~=rank(Φ~e(𝜶^))\widetilde{N}=\mbox{rank}\big{(}\widetilde{\Phi}_{e}(\widehat{\boldsymbol{\alpha}})\big{)}. Then, it holds

i=1Nϕij=1N~ϕi,𝐳j𝐳j22ε~2𝚽F2,\sum_{i=1}^{N}\left\|\boldsymbol{\phi}_{i}-\sum_{j=1}^{\widetilde{N}}\langle\boldsymbol{\phi}_{i},\mathbf{z}_{j}\rangle\mathbf{z}_{j}\right\|^{2}_{\ell^{2}}\leq\widetilde{\varepsilon}^{2}\big{\|}\boldsymbol{\Phi}\big{\|}_{F}^{2}, (14)

where we use the shortcut ϕi=ϕi(𝜶^)\boldsymbol{\phi}_{i}=\boldsymbol{\phi}_{i}(\widehat{\boldsymbol{\alpha}}), 𝐳i=𝐳i(𝜶^)\mathbf{z}_{i}=\mathbf{z}_{i}(\widehat{\boldsymbol{\alpha}}). To establish (14), consider (thin) SVD Φ~e(𝜶^)=U~Σ~V~T\widetilde{\Phi}_{e}(\widehat{\boldsymbol{\alpha}})=\widetilde{\mathrm{U}}\widetilde{\Sigma}\widetilde{\mathrm{V}}^{T} and compute

i=1Nϕij=1N~ϕi,𝐳j𝐳j22=(IU~U~T)Φe(𝜶^)F2=(IU~U~T)(Φe(𝜶^)Φ~e(𝜶^))F2IU~U~T2Φe(𝜶^)Φ~e(𝜶^)F2Φe(𝜶^)Φ~e(𝜶^)F2=(𝚽𝚽~)×2𝐞1(𝜶^)×3𝐞2(𝜶^)×D+1𝐞D(𝜶^)F2𝚽𝚽~F2ε~2𝚽F2,\begin{split}\sum_{i=1}^{N}\left\|\boldsymbol{\phi}_{i}-\sum_{j=1}^{\widetilde{N}}\langle\boldsymbol{\phi}_{i},\mathbf{z}_{j}\rangle\mathbf{z}_{j}\right\|^{2}_{\ell^{2}}&=\left\|(\mathrm{I}-\widetilde{\mathrm{U}}\widetilde{\mathrm{U}}^{T}){\Phi}_{e}(\widehat{\boldsymbol{\alpha}})\right\|^{2}_{F}\\ &=\left\|(\mathrm{I}-\widetilde{\mathrm{U}}\widetilde{\mathrm{U}}^{T})\left({\Phi}_{e}(\widehat{\boldsymbol{\alpha}})-\widetilde{\Phi}_{e}(\widehat{\boldsymbol{\alpha}})\right)\right\|^{2}_{F}\\ &\leq\left\|\mathrm{I}-\widetilde{\mathrm{U}}\widetilde{\mathrm{U}}^{T}\right\|^{2}\left\|{\Phi}_{e}(\widehat{\boldsymbol{\alpha}})-\widetilde{\Phi}_{e}(\widehat{\boldsymbol{\alpha}})\right\|^{2}_{F}\leq\left\|\Phi_{e}(\widehat{\boldsymbol{\alpha}})-\widetilde{\Phi}_{e}(\widehat{\boldsymbol{\alpha}})\right\|^{2}_{F}\\ &=\left\|(\boldsymbol{\Phi}-\widetilde{\boldsymbol{\Phi}})\times_{2}\mathbf{e}^{1}(\widehat{\boldsymbol{\alpha}})\times_{3}\mathbf{e}^{2}(\widehat{\boldsymbol{\alpha}})\dots\times_{D+1}\mathbf{e}^{D}(\widehat{\boldsymbol{\alpha}})\right\|^{2}_{F}\\ &\leq\left\|\boldsymbol{\Phi}-\widetilde{\boldsymbol{\Phi}}\right\|^{2}_{F}\leq\widetilde{\varepsilon}^{2}\big{\|}\boldsymbol{\Phi}\big{\|}_{F}^{2},\end{split}

where we used linearity of (10) and the inequality ABFABF\|\mathrm{A}\mathrm{B}\|_{F}\leq\|\mathrm{A}\|\|\mathrm{B}\|_{F} for matrices A\mathrm{A}, B\mathrm{B}, and spectral matrix norm \|\cdot\|. We also used P1\|\mathrm{P}\|\leq 1 for an orthogonal projection matrix P=IU~U~T\mathrm{P}=\mathrm{I}-\widetilde{\mathrm{U}}\widetilde{\mathrm{U}}^{T}.

Since 𝐳i(𝜶^)V~\mathbf{z}_{i}(\widehat{\boldsymbol{\alpha}})\in\widetilde{V} for all in-sample 𝜶^\widehat{\boldsymbol{\alpha}}, the bound in (14) provides an estimate on how accurate the true snapshots can be approximated in the universal space. Furthermore, from (14) we conclude that given 𝚽~\widetilde{\boldsymbol{\Phi}} and 𝜶^𝒜^\widehat{\boldsymbol{\alpha}}\in{\widehat{\mathcal{A}}} we can obtain a parameter-specific (quasi)-optimal reduced basis by taking the first nn left singular vectors of Φ~e(𝜶^)\widetilde{\Phi}_{e}(\widehat{\boldsymbol{\alpha}}). Representation power of this basis is determined by ε~\widetilde{\varepsilon} from (8) and σi\sigma_{i}, i>ni>n, the singular values of Φ~e(𝜶^)\widetilde{\Phi}_{e}(\widehat{\boldsymbol{\alpha}}). If ε~\widetilde{\varepsilon} is sufficiently small, i.e., the snapshot tensor admits an efficient low-rank representation, then the computed reduced basis better represents the snapshot space for a given 𝛂^\widehat{\boldsymbol{\alpha}} than the first nn left singular vectors of the unfolded snapshot matrix (i.e., better than the POD basis); see numerical results in Section 5 which show up to several orders of accuracy gain for some of the examples.

The arguments above apply only to parameter values 𝜶^\widehat{\boldsymbol{\alpha}} from the sampling set 𝒜^𝒜{\widehat{\mathcal{A}}}\subset\mathcal{A}. Next, we consider ROM basis computation for an arbitrary 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A} that may not necessarily come from the training set 𝒜^{\widehat{\mathcal{A}}}, the so-called out-of-sample 𝜶\alpha. Below we explore the option of building ROM basis for an out-of-sample 𝜶\alpha using interpolation in the parameter space. This approach is based on an assumption of smooth dependence of the solution 𝐮(t,𝜶)\mathbf{u}(t,\mbox{\boldmath$\alpha$\unboldmath}) of (1) on 𝜶\alpha. We refer to the corresponding tensorial ROMs as interpolatory TROMs.

3.4 Interpolatory TROM

To construct parameter-specific ROM basis for an arbitrary 𝜶=(α1,,αD)T𝒜\mbox{\boldmath$\alpha$\unboldmath}=(\alpha_{1},\dots,\alpha_{D})^{T}\in\mathcal{A} we introduce the interpolation procedure defined by

𝐞i:𝜶ni,i=1,,D.\mathbf{e}^{i}\,:\,\mbox{\boldmath$\alpha$\unboldmath}\to\mathbb{R}^{n_{i}},\quad i=1,\dots,D. (15)

Entrywise, we write 𝐞i(𝜶)=(e1i(𝜶),,enii(𝜶))Tni\mathbf{e}^{i}(\mbox{\boldmath$\alpha$\unboldmath})=\big{(}e_{1}^{i}(\mbox{\boldmath$\alpha$\unboldmath}),\ldots,e_{n_{i}}^{i}(\mbox{\boldmath$\alpha$\unboldmath})\big{)}^{T}\in\mathbb{R}^{n_{i}}. The interpolation procedure should satisfy the following property. For a smooth function g:[αimin,αimax]{g}:[\alpha_{i}^{\min},\alpha_{i}^{\max}]\to\mathbb{R}, it holds

g(αi)j=1nieji(𝜶)g(α^ij),i=1,,D,{g}(\alpha_{i})\approx\sum_{j=1}^{n_{i}}e_{j}^{i}(\mbox{\boldmath$\alpha$\unboldmath}){g}(\widehat{\alpha}_{i}^{j}),\quad i=1,\ldots,D, (16)

where α^ij\widehat{\alpha}_{i}^{j}, j=1,,nij=1,\ldots,n_{i}, are the grid nodes on [αimin,αimax][\alpha_{i}^{\min},\alpha_{i}^{\max}].

We further consider Lagrangian interpolation of order pp: for a given 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A} let α^ii1,,α^iip\widehat{\alpha}_{i}^{i_{1}},\ldots,\widehat{\alpha}_{i}^{i_{p}} be the pp closest grid nodes to αi\alpha_{i} on [αimin,αimax][\alpha_{i}^{\min},\alpha_{i}^{\max}], for i=1,,Di=1,\ldots,D. Then

eji(𝜶)={m=1,mkp(α^iimαi)/m=1,mkp(α^iimα^ij),if j=ik{i1,,ip},0,otherwise,e_{j}^{i}(\mbox{\boldmath$\alpha$\unboldmath})=\begin{cases}\prod\limits_{\begin{subarray}{c}m=1,\\ m\neq k\end{subarray}}^{p}(\widehat{\alpha}_{i}^{i_{m}}-\alpha_{i})\Big{/}\prod\limits_{\begin{subarray}{c}m=1,\\ m\neq k\end{subarray}}^{p}(\widehat{\alpha}_{i}^{i_{m}}-\widehat{\alpha}_{i}^{j}),&\text{if }j=i_{k}\in\{i_{1},\ldots,i_{p}\},\\ \qquad\qquad\qquad\qquad\qquad\qquad\qquad 0,&\text{otherwise},\end{cases} (17)

are the entries of 𝐞i(𝜶)\mathbf{e}^{i}(\mbox{\boldmath$\alpha$\unboldmath}) for j=1,,nij=1,\dots,n_{i}. For the numerical experiments in Section 5 we use p=2p=2 or 33, i.e., linear or quadratic interpolation. The Lagrangian interpolation is not the only option and depending on parameter sampling and solution smoothness other fitting procedures can be more suitable.

Vectors 𝐞i\mathbf{e}^{i} extend the notion of position vectors 𝐞i\mathbf{e}^{i} defined in (11) for out-of-sample vectors. Indeed, from (17) it is easy to see that both vectors coincide if 𝜶^𝒜^\widehat{\boldsymbol{\alpha}}\in{\widehat{\mathcal{A}}} and so we use the same notation hereafter. Therefore, we can define a snapshot matrix Φ~e(𝜶)\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}) through the extraction–interpolation procedure:

Φ~e(𝜶)=𝚽~×2𝐞1(𝜶)×3𝐞2(𝜶)×D+1𝐞D(𝜶)M×N,\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})=\widetilde{\boldsymbol{\Phi}}\times_{2}\mathbf{e}^{1}(\mbox{\boldmath$\alpha$\unboldmath})\times_{3}\mathbf{e}^{2}(\mbox{\boldmath$\alpha$\unboldmath})\dots\times_{D+1}\mathbf{e}^{D}(\mbox{\boldmath$\alpha$\unboldmath})\in\mathbb{R}^{M\times N}, (18)

to generalize (12) for any 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A}. Note that (18) and (12) are the same for 𝜶=𝜶^𝒜^𝒜\mbox{\boldmath$\alpha$\unboldmath}=\widehat{\boldsymbol{\alpha}}\in{\widehat{\mathcal{A}}}\subset\mathcal{A}, while (18) defines Φ~e(𝜶)\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}) also for out-of-sample parameter vectors. If the low-rank representation of the snapshot tensor is exact, i.e., 𝚽~=𝚽\widetilde{\boldsymbol{\Phi}}=\boldsymbol{\Phi}, then Φ~e(𝜶)\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}) is the interpolation of the snapshot matrices Φpod(^𝜶){\Phi}_{\rm pod}(\widehat{}\mbox{\boldmath$\alpha$\unboldmath}).

Once 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A} is fixed and Φ~e(𝜶)\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}) is given by (18), our parameter-specific reduced basis {𝐳i(𝜶)}i=1n\{\mathbf{z}_{i}(\mbox{\boldmath$\alpha$\unboldmath})\}_{i=1}^{n} is defined as the first nn left singular vectors of Φ~e(𝜶)\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}). Later we demonstrate that the coordinates of this basis in the universal space can be calculated quickly (i.e., using only low-dimensional calculations) online without actually computing Φ~e(𝜶)\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}).

In a non-interpolatory TROM, a parameter-specific reduced basis can be constructed as follows. Choose p2p\geq 2 and fix 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A}, then let α^ii1,,α^iip\widehat{\alpha}_{i}^{i_{1}},\ldots,\widehat{\alpha}_{i}^{i_{p}} be the pp closest grid nodes to αi\alpha_{i} on [αimin,αimax][\alpha_{i}^{\min},\alpha_{i}^{\max}], for i=1,,Di=1,\ldots,D, similarly to the interpolatory construction above. Define the set

𝒜^p:={𝜶^=(α^1,,α^D)T:α^i{α^ij}j{i1,,ip},i=1,,D}𝒜^.{\widehat{\mathcal{A}}}_{p}:=\left\{\widehat{\boldsymbol{\alpha}}=(\widehat{\alpha}_{1},\dots,\widehat{\alpha}_{D})^{T}\,:\,\widehat{\alpha}_{i}\in\{\widehat{\alpha}_{i}^{j}\}_{j\in\{i_{1},\ldots,i_{p}\}},~{}i=1,\dots,D\right\}\subset{\widehat{\mathcal{A}}}. (19)

Then, assemble a large matrix by concatenating Φ~e(𝜶^)\widetilde{\Phi}_{e}(\widehat{\boldsymbol{\alpha}}) for all 𝜶^𝒜^p\widehat{\boldsymbol{\alpha}}\in{\widehat{\mathcal{A}}}_{p} and take {𝐳i(𝜶)}i=1n\{\mathbf{z}_{i}(\mbox{\boldmath$\alpha$\unboldmath})\}_{i=1}^{n} to be its first nn left singular vectors. Of course, hybrid strategies (e.g., interpolation only in some parameter directions) are also possible. For non-interpolatory or hybrid TROMs it is also possible to compute local basis online with only low-dimensional calculations following same steps as considered below.

In the rest of paper we focus on the interpolatory TROM and consider three well-known compressed formats for low rank tensor approximation 𝚽~𝚽\widetilde{\boldsymbol{\Phi}}\approx\boldsymbol{\Phi}: canonical polyadic (CP), Tucker, a.k.a. higher order SVD (HOSVD), and tensor train (TT) decomposition formats. Note that the notion of tensor rank(s) differs among these formats. When applied to TROM computation, these formats lead to different offline computational costs to build 𝚽~\widetilde{\boldsymbol{\Phi}}, different amounts of information transmitted from the offline stage to the online stage (measured by the compression rate, as explained in Section 3.9), and slightly varying amounts of online computations for finding the reduced basis given an incoming 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A}.

3.5 CP-TROM

The first tensor decomposition that we consider is the canonical polyadic decomposition of a tensor into the sum of rank one tensors [40, 16, 45, 46]. In CP-TROM we approximate 𝚽\boldsymbol{\Phi} by the sum of RR (where RR is the so-called canonical tensor rank) direct products of D+2D+2 vectors 𝐮rM\mathbf{u}^{r}\in\mathbb{R}^{M}, 𝝈irni\mbox{\boldmath$\sigma$\unboldmath}^{r}_{i}\in\mathbb{R}^{n_{i}}, i=1,,Di=1,\dots,D, and 𝐯rN\mathbf{v}^{r}\in\mathbb{R}^{N},

𝚽𝚽~=r=1R𝐮r𝝈1r𝝈Dr𝐯r,\boldsymbol{\Phi}\approx\widetilde{\boldsymbol{\Phi}}=\sum_{r=1}^{R}\mathbf{u}^{r}\circ\mbox{\boldmath$\sigma$\unboldmath}^{r}_{1}\circ\dots\circ\mbox{\boldmath$\sigma$\unboldmath}^{r}_{D}\circ\mathbf{v}^{r}, (20)

or entry-wise

(𝚽~)j,i1,,iD,k=r=1Rujrσ1,i1rσD,iDrvkr.(\widetilde{\boldsymbol{\Phi}})_{j,i_{1},\dots,i_{D},k}=\sum_{r=1}^{R}u_{j}^{r}\sigma_{1,i_{1}}^{r}\dots\sigma_{D,i_{D}}^{r}v_{k}^{r}.

CP decomposition often delivers excellent compression. However, there are well-known difficulties in determining the accurate canonical rank RR and working with the CP format, see, e.g., [38, 22]. Since we are interested in the approximation 𝚽~\widetilde{\boldsymbol{\Phi}} to 𝚽\boldsymbol{\Phi}, the alternating least squares (ALS) algorithm [36, 46] can be used to minimize 𝚽𝚽~F\|\boldsymbol{\Phi}-\widetilde{\boldsymbol{\Phi}}\|_{F} for a specified target canonical rank RR to find the approximate factors 𝐮rM\mathbf{u}^{r}\in\mathbb{R}^{M}, 𝝈irni\mbox{\boldmath$\sigma$\unboldmath}^{r}_{i}\in\mathbb{R}^{n_{i}}, and 𝐯rN\mathbf{v}^{r}\in\mathbb{R}^{N}, r=1,,Rr=1,\ldots,R. We further assume RMR\leq M, where MM is the dimension of high-fidelity snapshots.

Note that the second-mode product of a D+2D+2-dimensional rank-one tensor 𝐮r𝝈1r𝝈Dr𝐯r\mathbf{u}^{r}\circ\mbox{\boldmath$\sigma$\unboldmath}^{r}_{1}\circ\dots\circ\mbox{\boldmath$\sigma$\unboldmath}^{r}_{D}\circ\mathbf{v}^{r} and a vector 𝐞1(𝜶)n1\mathbf{e}^{1}(\mbox{\boldmath$\alpha$\unboldmath})\in\mathbb{R}^{n_{1}} is the D+1D+1-dimensional rank-one tensor 𝝈1r,𝐞(𝐮r𝝈2r𝝈Dr𝐯r)\left\langle\mbox{\boldmath$\sigma$\unboldmath}^{r}_{1},\mathbf{e}\right\rangle(\mathbf{u}^{r}\circ\mbox{\boldmath$\sigma$\unboldmath}^{r}_{2}\circ\dots\circ\mbox{\boldmath$\sigma$\unboldmath}^{r}_{D}\circ\mathbf{v}^{r}). Proceeding with this computation for other modes, in the decomposition (20) we find that the definition (18) yields representation of Φ~e(𝜶)\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}) as the sum of rank one matrices for any 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A}:

Φ~e(𝜶)=r=1Rsr𝐮r𝐯rM×N,withsr=i=1D𝝈ir,𝐞i(𝜶).\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})=\sum_{r=1}^{R}s_{r}\mathbf{u}^{r}\circ\mathbf{v}^{r}\in\mathbb{R}^{M\times N},~{}~{}\text{with}~{}~{}s_{r}=\prod_{i=1}^{D}\left\langle\mbox{\boldmath$\sigma$\unboldmath}^{r}_{i},\mathbf{e}^{i}(\mbox{\boldmath$\alpha$\unboldmath})\right\rangle\in\mathbb{R}. (21)

However, (21) is not the SVD of Φ~e(𝜶)\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}), since vectors 𝐮r\mathbf{u}^{r} (and 𝐯r\mathbf{v}^{r}) are not necessarily orthogonal. To avoid computing Φ~e(𝜶)\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}) and its SVD online, the following preparatory offline step is required. Organize the vectors 𝐮r\mathbf{u}^{r} and 𝐯r\mathbf{v}^{r} from (20) into matrices

U^=[𝐮1,,𝐮R]M×R,V^=[𝐯1,,𝐯R]N×R,\widehat{\mathrm{U}}=[\mathbf{u}^{1},\dots,\mathbf{u}^{R}]\in\mathbb{R}^{M\times R},\quad\widehat{\mathrm{V}}=[\mathbf{v}^{1},\dots,\mathbf{v}^{R}]\in\mathbb{R}^{N\times R}, (22)

and compute the thin QR factorizations

U^=URU,V^=VRV,\widehat{\mathrm{U}}={\mathrm{U}}\mathrm{R}_{U},\quad\widehat{\mathrm{V}}={\mathrm{V}}\mathrm{R}_{V}, (23)

if R>NR>N let further RV=V^\mathrm{R}_{V}=\widehat{\mathrm{V}} and V=I{\mathrm{V}}=\mathrm{I}. The columns of U{\mathrm{U}} form an orthogonal basis in the universal space V~\widetilde{V}. Matrix U{\mathrm{U}} is stored offline (V{\mathrm{V}} is not used and can be dropped), while low-dimensional matrices RU\mathrm{R}_{U} and RV\mathrm{R}_{V} together with vectors 𝝈ir\mbox{\boldmath$\sigma$\unboldmath}^{r}_{i} form the online part of 𝚽~\widetilde{\boldsymbol{\Phi}},

online(𝚽~)={RU,RVR×R,𝝈irni,i=1,,D},\mbox{online}(\widetilde{\boldsymbol{\Phi}})=\left\{\mathrm{R}_{U},\mathrm{R}_{V}\in\mathbb{R}^{R\times R},~{}\mbox{\boldmath$\sigma$\unboldmath}^{r}_{i}\in\mathbb{R}^{n_{i}},~{}{\small i=1,\dots,D}\right\}, (24)

which is transmitted to the online stage.

At the online stage and for any incoming 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A}, we compute the SVD of the R×RR\times R core matrix

C(𝜶)=RUS(𝜶)RVT,\mathrm{C}(\mbox{\boldmath$\alpha$\unboldmath})=\mathrm{R}_{U}\mathrm{S}(\mbox{\boldmath$\alpha$\unboldmath})\mathrm{R}_{V}^{T}, (25)

where S(𝜶)=diag(s1,,sR)\mathrm{S}(\mbox{\boldmath$\alpha$\unboldmath})=\mbox{diag}(s_{1},\dots,s_{R}), with srs_{r} from (21): C(𝜶)=UcΣcVcT\mathrm{C}(\mbox{\boldmath$\alpha$\unboldmath})=\mathrm{U}_{c}\Sigma_{c}\mathrm{V}_{c}^{T}. Since

Φ~e(𝜶)=UC(𝜶)VT=(UUc)Σc(VVc)T\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})={\mathrm{U}}\mathrm{C}(\mbox{\boldmath$\alpha$\unboldmath}){\mathrm{V}}^{T}=\left({\mathrm{U}}\mathrm{U}_{c}\right)\Sigma_{c}\left({\mathrm{V}}\mathrm{V}_{c}\right)^{T} (26)

is the SVD of Φ~e(𝜶)\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}), the first nn columns of Uc\mathrm{U}_{c}, denoted by {𝛃1(𝛂),,𝛃n(𝛂)}\left\{\mbox{\boldmath$\beta$\unboldmath}_{1}(\mbox{\boldmath$\alpha$\unboldmath}),\dots,\mbox{\boldmath$\beta$\unboldmath}_{n}(\mbox{\boldmath$\alpha$\unboldmath})\right\}, are the coordinates of the local reduced basis in the universal space V~\widetilde{V}. The parameter-specific basis in the physical space is then {𝐳i(𝜶)}i=1n\{\mathbf{z}_{i}(\mbox{\boldmath$\alpha$\unboldmath})\}_{i=1}^{n}, with 𝐳i(𝜶)=U𝜷i(𝜶)\mathbf{z}_{i}(\mbox{\boldmath$\alpha$\unboldmath})={\mathrm{U}}\mbox{\boldmath$\beta$\unboldmath}_{i}(\mbox{\boldmath$\alpha$\unboldmath}). Note that 𝐳i(𝜶)\mathbf{z}_{i}(\mbox{\boldmath$\alpha$\unboldmath}) are not actually computed.

Under certain assumptions on FF, the dynamical system (1) is projected offline onto V~\widetilde{V} and passed to the online stage, where for any 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A} it is further projected onto the local basis {𝜷1(𝜶),,𝜷n(𝜶)}\left\{\mbox{\boldmath$\beta$\unboldmath}_{1}(\mbox{\boldmath$\alpha$\unboldmath}),\dots,\mbox{\boldmath$\beta$\unboldmath}_{n}(\mbox{\boldmath$\alpha$\unboldmath})\right\}. This avoids any online computations with high-dimensional objects used in high-fidelity simulations; see further discussion in Section 3.10.

We summarize the above in the following algorithm.

Algorithm 1 (CP-TROM).
  • Offline stage.
    Input: snapshot tensor 𝚽M×n1××nD×N\boldsymbol{\Phi}\in\mathbb{R}^{M\times n_{1}\times\ldots\times n_{D}\times N}, target canonical rank RR;
    Output: CP decomposition factors, universal basis matrix UM×R{\mathrm{U}}\in\mathbb{R}^{M\times R}, and upper triangular matrices RU\mathrm{R}_{U}, RVR×R\mathrm{R}_{V}\in\mathbb{R}^{R\times R};
    Compute:

    1. 1.

      Use ALS algorithm to minimize 𝚽𝚽~F\|\boldsymbol{\Phi}-\widetilde{\boldsymbol{\Phi}}\|_{F} to find CP decomposition factors 𝐮r\mathbf{u}^{r}, 𝝈ir\mbox{\boldmath$\sigma$\unboldmath}^{r}_{i}, and 𝐯r\mathbf{v}^{r} of 𝚽~\widetilde{\boldsymbol{\Phi}} satisfying (20);

    2. 2.

      Assemble matrices V^,U^\widehat{\mathrm{V}},\widehat{\mathrm{U}} as in (22) and compute their thin QR factorizations (23) to obtain RU\mathrm{R}_{U}, RV\mathrm{R}_{V} and U{\mathrm{U}}.

  • Online stage.
    Input: online(Φ~)\mbox{\rm online}(\widetilde{\Phi}) as defined in (24), reduced space dimension nRn\leq R, and parameter vector 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A};
    Output: Coordinates of the reduced basis in V~\widetilde{V}: {𝜷i(𝜶)}i=1nR\{\mbox{\boldmath$\beta$\unboldmath}_{i}(\mbox{\boldmath$\alpha$\unboldmath})\}_{i=1}^{n}\subset\mathbb{R}^{R};
    Compute:

    1. 1.

      Use (25) to assemble the core matrix C(𝜶)\mathrm{C}(\mbox{\boldmath$\alpha$\unboldmath});

    2. 2.

      Compute the SVD of the core matrix C(𝜶)=UcΣcVcT\mathrm{C}(\mbox{\boldmath$\alpha$\unboldmath})=\mathrm{U}_{c}\Sigma_{c}\mathrm{V}_{c}^{T}, with Uc=[𝐮~1,𝐮~2,,𝐮~R]\mathrm{U}_{c}=[\widetilde{\mathbf{u}}_{1},\widetilde{\mathbf{u}}_{2},\ldots,\widetilde{\mathbf{u}}_{R}];

    3. 3.

      Set 𝜷i(𝜶)=𝐮~i\mbox{\boldmath$\beta$\unboldmath}_{i}(\mbox{\boldmath$\alpha$\unboldmath})=\widetilde{\mathbf{u}}_{i}, i=1,,ni=1,\dots,n.

Note that we do not have direct control over ALS algorithm to enforce a priori CP decomposition accuracy 𝚽𝚽~F<ε~𝚽F\|\boldsymbol{\Phi}-\widetilde{\boldsymbol{\Phi}}\|_{F}<\widetilde{\varepsilon}\|\boldsymbol{\Phi}\|_{F}. One option is to rerun the offline stage for different trial values of RR. Given that the offline stage is computationally expensive, this may become prohibitive in cases where a desired accuracy ε~\widetilde{\varepsilon} must be strictly enforced. The other two variants of TROM presented below are free from this limitation.

3.6 HOSVD-TROM

As we already mentioned, truncated variant of CP decomposition is not known to satisfy any simple minimization property (unlike the SVD decomposition for matrices). A classical tensor decomposition, known to deliver a (quasi)-minimization property, is the so-called higher order SVD (HOSVD) [21]. In HOSVD-TROM variant we approximate the snapshot tensor with a Tucker tensor [68, 46] Φ~\widetilde{\Phi} of the form

𝚽𝚽~=j=1M~q1=1n~1qD=1n~Dk=1N~(𝐂)j,q1,,qD,k𝐮j𝝈1q1𝝈DqD𝐯k,\boldsymbol{\Phi}\approx\widetilde{\boldsymbol{\Phi}}=\sum_{j=1}^{\widetilde{M}}\sum_{q_{1}=1}^{\widetilde{n}_{1}}\dots\sum_{q_{D}=1}^{\widetilde{n}_{D}}\sum_{k=1}^{\widetilde{N}}(\mathbf{C})_{j,q_{1},\dots,q_{D},k}\mathbf{u}^{j}\circ\mbox{\boldmath$\sigma$\unboldmath}^{q_{1}}_{1}\circ\dots\circ\mbox{\boldmath$\sigma$\unboldmath}^{q_{D}}_{D}\circ\mathbf{v}^{k}, (27)

with 𝐮jM\mathbf{u}^{j}\in\mathbb{R}^{M}, 𝝈iqini\mbox{\boldmath$\sigma$\unboldmath}_{i}^{q_{i}}\in\mathbb{R}^{n_{i}}, and 𝐯kN\mathbf{v}^{k}\in\mathbb{R}^{N}. The numbers M~\widetilde{M}, n~1\widetilde{n}_{1}, \ldots, n~D\widetilde{n}_{D} and N~\widetilde{N} are referred to as Tucker ranks of 𝚽~\widetilde{\boldsymbol{\Phi}}. The HOSVD delivers an efficient compression of the snapshot tensor, provided the size of the core tensor 𝐂M~×n~1××n~D×N~\mathbf{C}\in\mathbb{R}^{\widetilde{M}\times\widetilde{n}_{1}\times\dots\times\widetilde{n}_{D}\times\widetilde{N}} is (much) smaller than the size of 𝚽\boldsymbol{\Phi}.

In what follows, it is helpful to organize the column vectors from (27) into matrices

U=[𝐮1,,𝐮M~]M×M~,V=[𝐯1,,𝐯N~]N×N~,Si=[𝝈i1,,𝝈in~i]Tn~i×ni,i=1,,D.\begin{split}\mathrm{U}&=[\mathbf{u}^{1},\dots,\mathbf{u}^{\widetilde{M}}]\in\mathbb{R}^{M\times\widetilde{M}},\quad\mathrm{V}=[\mathbf{v}^{1},\dots,\mathbf{v}^{\widetilde{N}}]\in\mathbb{R}^{N\times\widetilde{N}},\\ \mathrm{S}_{i}&=[\mbox{\boldmath$\sigma$\unboldmath}^{1}_{i},\dots,\mbox{\boldmath$\sigma$\unboldmath}^{\widetilde{n}_{i}}_{i}]^{T}\in\mathbb{R}^{\widetilde{n}_{i}\times{n}_{i}},\quad i=1,\ldots,D.\end{split} (28)

In contrast with CP decomposition, HOSVD computes vectors 𝐮j\mathbf{u}^{j}, j=1,,M~j=1,\ldots,\widetilde{M}, and 𝐯k\mathbf{v}^{k}, k=1,,N~k=1,\ldots,\widetilde{N}, that are orthonormal. Therefore, the columns of U\mathrm{U} form an orthogonal basis in the universal reduced space V~\widetilde{V}. The dimension of this space is defined by the first Tucker rank, dim(V~)=M~\mbox{\rm dim}(\widetilde{V})=\widetilde{M}. The information about 𝚽~\widetilde{\boldsymbol{\Phi}} to be transmitted to the online stage includes matrices Si\mathrm{S}_{i} and the core tensor 𝐂\mathbf{C}. Explicitly,

online(𝚽~)={𝐂M~×n~1××n~D×N~,Sini×n~i,i=1,,D}.\mbox{online}(\widetilde{\boldsymbol{\Phi}})=\left\{\mathbf{C}\in\mathbb{R}^{\widetilde{M}\times\widetilde{n}_{1}\times\dots\times\widetilde{n}_{D}\times\widetilde{N}},~{}\mathrm{S}_{i}\in\mathbb{R}^{n_{i}\times\widetilde{n}_{i}},~{}{\small i=1,\dots,D}\right\}. (29)

To find coordinates of the local basis for 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A}, define the 𝜶\alpha-specific core matrix Ce(𝜶)\mathrm{C}_{e}(\mbox{\boldmath$\alpha$\unboldmath}) as

Ce(𝜶)=𝐂×2(S1𝐞1(𝜶))×3(S2𝐞2(𝜶))×D+1(SD𝐞D(𝜶))M~×N~.\mathrm{C}_{e}(\mbox{\boldmath$\alpha$\unboldmath})=\mathbf{C}\times_{2}\left(\mathrm{S}_{1}\mathbf{e}^{1}(\mbox{\boldmath$\alpha$\unboldmath})\right)\times_{3}\left(\mathrm{S}_{2}\mathbf{e}^{2}(\mbox{\boldmath$\alpha$\unboldmath})\right)\dots\times_{D+1}\left(\mathrm{S}_{D}\mathbf{e}^{D}(\mbox{\boldmath$\alpha$\unboldmath})\right)\in\mathbb{R}^{\widetilde{M}\times\widetilde{N}}. (30)

Using the definition of kk-mode product, (18) and (27), one computes

Φ~e(𝜶)=j=1M~q1=1n~1qD=1n~Dk=1N~(𝐂)j,q1,,qD,k𝝈1q1,𝐞1(α)𝝈DqD,𝐞D(α)𝐮j𝐯k=j=1M~q1=1n~1qD=1n~Dk=1N~(𝐂)j,q1,,qD,k(S1𝐞1(𝜶))q1(SD𝐞D(𝜶))qD𝐮j𝐯k=j=1M~k=1N~(𝐂×2(S1𝐞1(𝜶))×3(S2𝐞2(𝜶))×D+1(SD𝐞D(𝜶)))jk𝐮j𝐯k=UCe(𝜶)VT\begin{split}\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})&=\sum_{j=1}^{\widetilde{M}}\sum_{q_{1}=1}^{\widetilde{n}_{1}}\dots\sum_{q_{D}=1}^{\widetilde{n}_{D}}\sum_{k=1}^{\widetilde{N}}(\mathbf{C})_{j,q_{1},\dots,q_{D},k}\langle\mbox{\boldmath$\sigma$\unboldmath}^{q_{1}}_{1},\mathbf{e}^{1}(\alpha)\rangle\cdot\dotsc\cdot\langle\mbox{\boldmath$\sigma$\unboldmath}^{q_{D}}_{D},\mathbf{e}^{D}(\alpha)\rangle\,\mathbf{u}^{j}\circ\mathbf{v}^{k}\\ &=\sum_{j=1}^{\widetilde{M}}\sum_{q_{1}=1}^{\widetilde{n}_{1}}\dots\sum_{q_{D}=1}^{\widetilde{n}_{D}}\sum_{k=1}^{\widetilde{N}}(\mathbf{C})_{j,q_{1},\dots,q_{D},k}\left(\mathrm{S}_{1}\mathbf{e}^{1}(\mbox{\boldmath$\alpha$\unboldmath})\right)_{q_{1}}\cdot\dotsc\cdot\left(\mathrm{S}_{D}\mathbf{e}^{D}(\mbox{\boldmath$\alpha$\unboldmath})\right)_{q_{D}}\mathbf{u}^{j}\circ\mathbf{v}^{k}\\ &=\sum_{j=1}^{\widetilde{M}}\sum_{k=1}^{\widetilde{N}}\left(\mathbf{C}\times_{2}\left(\mathrm{S}_{1}\mathbf{e}^{1}(\mbox{\boldmath$\alpha$\unboldmath})\right)\times_{3}\left(\mathrm{S}_{2}\mathbf{e}^{2}(\mbox{\boldmath$\alpha$\unboldmath})\right)\dots\times_{D+1}\left(\mathrm{S}_{D}\mathbf{e}^{D}(\mbox{\boldmath$\alpha$\unboldmath})\right)\right)_{jk}\mathbf{u}^{j}\circ\mathbf{v}^{k}=\mathrm{U}\mathrm{C}_{e}(\mbox{\boldmath$\alpha$\unboldmath})\mathrm{V}^{T}\end{split}

Consider the thin SVD of the core matrix

Ce(𝜶)=UcΣcVcT.\mathrm{C}_{e}(\mbox{\boldmath$\alpha$\unboldmath})=\mathrm{U}_{c}\Sigma_{c}\mathrm{V}_{c}^{T}. (31)

Combining this with the representation above we get

Φ~e(𝜶)=(UUc)Σc(VVc)T,\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})=\left({\mathrm{U}}\mathrm{U}_{c}\right)\Sigma_{c}\left({\mathrm{V}}\mathrm{V}_{c}\right)^{T}, (32)

which is the thin SVD of Φ~e(𝜶)\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}) since both matrices U\mathrm{U} and V\mathrm{V} are orthogonal. We conclude that the coordinates {𝛃1(𝛂),,𝛃n(𝛂)}\left\{\mbox{\boldmath$\beta$\unboldmath}_{1}(\mbox{\boldmath$\alpha$\unboldmath}),\dots,\mbox{\boldmath$\beta$\unboldmath}_{n}(\mbox{\boldmath$\alpha$\unboldmath})\right\} of the local reduced basis in the universal space V~\widetilde{V} are the first nn columns of Uc\mathrm{U}_{c} from (31). The parameter-specific basis is then {𝐳i(𝜶)}i=1n\{\mathbf{z}_{i}(\mbox{\boldmath$\alpha$\unboldmath})\}_{i=1}^{n}, with 𝐳i(𝜶)=U𝜷i(𝜶)\mathbf{z}_{i}(\mbox{\boldmath$\alpha$\unboldmath})={\mathrm{U}}\mbox{\boldmath$\beta$\unboldmath}_{i}(\mbox{\boldmath$\alpha$\unboldmath}) (not actually computed at the online stage).

To compute the low-rank HOSVD approximation (27) we employ the standard algorithm [21] based on repeated computations of truncated SVD for unfolded matrices. In particular, one may compute 𝚽~\widetilde{\boldsymbol{\Phi}} with either prescribed Tucker ranks or prescribed accuracy ε~\widetilde{\varepsilon}. Moreover, for fixed Tucker ranks one can show that the recovered 𝚽~\widetilde{\boldsymbol{\Phi}} satisfies a quasi-minimization property [21] of the form

𝚽𝚽~D+2𝚽𝚽optand𝚽𝚽~(i=1D+1Δi2)12,\|\boldsymbol{\Phi}-\widetilde{\boldsymbol{\Phi}}\|\leq\sqrt{D+2}\|\boldsymbol{\Phi}-\boldsymbol{\Phi}^{\rm opt}\|\quad\text{and}\quad\|\boldsymbol{\Phi}-\widetilde{\boldsymbol{\Phi}}\|\leq\left(\sum_{i=1}^{D+1}{\Delta_{i}^{2}}\right)^{\frac{1}{2}}, (33)

where 𝚽opt\boldsymbol{\Phi}^{\rm opt} is the best approximation to 𝚽\boldsymbol{\Phi} among all Tucker tensors of the given rank (such approximation always exists), and Δi{\Delta_{i}} measures truncated SVD error on the ii-th step of the HOSVD. We summarize the above in the following algorithm.

Algorithm 2 (HOSVD-TROM).
  • Offline stage.
    Input: snapshot tensor 𝚽M×n1××nD×N\boldsymbol{\Phi}\in\mathbb{R}^{M\times n_{1}\times\ldots\times n_{D}\times N} and target accuracy ε~\widetilde{\varepsilon};
    Output: Compressed tensor ranks, HOSVD decomposition matrices as in (28), and core tensor 𝐂\mathbf{C};
    Compute: Use algorithm [21] with prescribed accuracy ε~\widetilde{\varepsilon} to compute HOSVD decomposition matrices and the core tensor.

  • Online stage.
    Input: online(𝚽~)\mbox{\rm online}(\widetilde{\boldsymbol{\Phi}}) as defined in (29), reduced space dimension nmin{M~,N~}n\leq\min\{\widetilde{M},\widetilde{N}\}, and parameter vector 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A};
    Output: Coordinates of the reduced basis in V~\widetilde{V}: {𝜷i(𝜶)}i=1nM~\{\mbox{\boldmath$\beta$\unboldmath}_{i}(\mbox{\boldmath$\alpha$\unboldmath})\}_{i=1}^{n}\subset\mathbb{R}^{\widetilde{M}};
    Compute:

    1. 1.

      Use the core tensor 𝐂\mathbf{C} and matrices Si\mathrm{S}_{i}, i=1,,Di=1,\ldots,D, to assemble the core matrix Ce(𝜶)M~×N~\mathrm{C}_{e}(\mbox{\boldmath$\alpha$\unboldmath})\in\mathbb{R}^{\widetilde{M}\times\widetilde{N}} as in (30);

    2. 2.

      Compute the SVD of the core matrix Ce(𝜶)=UcΣcVcT\mathrm{C}_{e}(\mbox{\boldmath$\alpha$\unboldmath})=\mathrm{U}_{c}\Sigma_{c}\mathrm{V}_{c}^{T} with Uc=[𝐮~1,,𝐮~M~]\mathrm{U}_{c}=[\widetilde{\mathbf{u}}_{1},\ldots,\widetilde{\mathbf{u}}_{\widetilde{M}}];

    3. 3.

      Set 𝜷i(𝜶)=𝐮~i\mbox{\boldmath$\beta$\unboldmath}_{i}(\mbox{\boldmath$\alpha$\unboldmath})=\widetilde{\mathbf{u}}_{i}, i=1,,ni=1,\ldots,n.

3.7 TT-TROM

A third low-rank tensor decomposition of interest is the Tensor Train (TT) decomposition [58]. In TT-TROM we seek to approximate the snapshot tensor with 𝚽~\widetilde{\boldsymbol{\Phi}} in the TT-format, namely

𝚽𝚽~=j1=1r~1jD+1=1r~D+1𝐮j1𝝈1j1,j2𝝈DjD,jD+1𝐯jD+1,\boldsymbol{\Phi}\approx\widetilde{\boldsymbol{\Phi}}=\sum_{j_{1}=1}^{\widetilde{r}_{1}}\dots\sum_{j_{D+1}=1}^{\widetilde{r}_{D+1}}\mathbf{u}^{j_{1}}\circ\mbox{\boldmath$\sigma$\unboldmath}^{j_{1},j_{2}}_{1}\circ\dots\circ\mbox{\boldmath$\sigma$\unboldmath}^{j_{D},j_{D+1}}_{D}\circ\mathbf{v}^{j_{D+1}}, (34)

with 𝐮j1M\mathbf{u}^{j_{1}}\in\mathbb{R}^{M}, 𝝈iji,ji+1ni\mbox{\boldmath$\sigma$\unboldmath}^{j_{i},j_{i+1}}_{i}\in\mathbb{R}^{n_{i}}, and 𝐯jD+1N\mathbf{v}^{j_{D+1}}\in\mathbb{R}^{N}, where the positive integers r~i\widetilde{r}_{i} are referred to as the compression ranks (or TT-ranks) of the decomposition. For higher order tensors the TT format is in general more efficient compared to HOSVD. This may be beneficial for large DD, the dimension of parameter space. In [58, 57] a stable algorithm for finding 𝚽~\widetilde{\boldsymbol{\Phi}} based on truncated SVD for a sequence of unfolding matrices was introduced and the optimality property similar to (33) was proved.

Once an optimal TT approximation (34) is computed, we organize the vectors and matrices from (34) into matrices

U=[𝐮1,,𝐮r~1]M×r~1,V=[𝐯1,,𝐯r~D+1]N×r~D+1,\mathrm{U}=[\mathbf{u}^{1},\dots,\mathbf{u}^{\widetilde{r}_{1}}]\in\mathbb{R}^{M\times\widetilde{r}_{1}},\quad\mathrm{V}=[\mathbf{v}^{1},\dots,\mathbf{v}^{\widetilde{r}_{D+1}}]\in\mathbb{R}^{N\times\widetilde{r}_{D+1}}, (35)

and third order tensors 𝐒ir~i×ni×r~i+1\mathbf{S}_{i}\in\mathbb{R}^{\widetilde{r}_{i}\times n_{i}\times\widetilde{r}_{i+1}}, defined entry-wise as

(𝐒i)jkq=(𝝈ijq)k,j=1,,r~i,k=1,,ni,q=1,,r~i+1,(\mathbf{S}_{i})_{jkq}=(\mbox{\boldmath$\sigma$\unboldmath}^{jq}_{i})_{k},\quad j=1,\ldots,\widetilde{r}_{i},\quad k=1,\ldots,n_{i},\quad q=1,\ldots,\widetilde{r}_{i+1}, (36)

for all i=1,,Di=1,\ldots,D. Note that matrix U\mathrm{U} is orthogonal and so its columns provide an orthogonal basis in the universal space V~\widetilde{V}. The dimension of V~\widetilde{V} is defined by the first TT-rank, dim(V~)=r~1\mbox{dim}(\widetilde{V})=\widetilde{r}_{1}.

While U\mathrm{U} is an orthogonal matrix, the columns of V\mathrm{V} are orthogonal, but not necessarily orthonormal. Thus, we introduce a diagonal scaling matrix

Wc=diag(𝐯1,,𝐯r~D+1)r~D+1×r~D+1.\mathrm{W}_{c}=\mbox{diag}\left(\|\mathbf{v}^{1}\|,\ldots,\|\mathbf{v}^{\widetilde{r}_{D+1}}\|\right)\in\mathbb{R}^{\widetilde{r}_{D+1}\times\widetilde{r}_{D+1}}. (37)

The essential information about 𝚽~\widetilde{\boldsymbol{\Phi}} to be transmitted to the online phase includes 𝐒i\mathbf{S}_{i} tensors and the scaling factors:

online(𝚽~)={𝐒ir~i×ni×r~i+1,i=1,,D,Wcr~D+1×r~D+1}.\mbox{online}(\widetilde{\boldsymbol{\Phi}})=\left\{\mathbf{S}_{i}\in\mathbb{R}^{\widetilde{r}_{i}\times n_{i}\times\widetilde{r}_{i+1}},~{}{\small i=1,\dots,D},~{}\mathrm{W}_{c}\in\mathbb{R}^{\widetilde{r}_{D+1}\times\widetilde{r}_{D+1}}\right\}. (38)

To find the coordinates of the local basis, we define the parameter-specific core matrix Ce(𝜶)r~1×r~D+1\mathrm{C}_{e}(\mbox{\boldmath$\alpha$\unboldmath})\in\mathbb{R}^{\widetilde{r}_{1}\times\widetilde{r}_{D+1}} as the product

Ce(𝜶)=i=1D(𝐒i×2𝐞i(𝜶)).\mathrm{C}_{e}(\mbox{\boldmath$\alpha$\unboldmath})=\prod_{i=1}^{D}\left(\mathbf{S}_{i}\times_{2}\mathbf{e}^{i}(\mbox{\boldmath$\alpha$\unboldmath})\right). (39)

Using the definition of kk-mode product, (18) and (34), one computes

Φ~e(𝜶)=j1=1r~1jD+1=1r~D+1𝝈1j1,j2,𝐞1(𝜶)𝝈DjD,jD+1,𝐞D(𝜶)𝐮j1𝐯jD+1=j1=1r~1jD+1=1r~D+1(𝐒1×2𝐞1(𝜶))j1,j2(𝐒D×D+1𝐞D(𝜶))jD,jD+1𝐮j1𝐯jD+1=j1=1r~1jD+1=1r~D+1(i=1D(𝐒i×2𝐞i(𝜶)))j1jD+1𝐮j1𝐯jD+1=UCe(𝜶)VT.\begin{split}\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})&=\sum_{j_{1}=1}^{\widetilde{r}_{1}}\dots\sum_{j_{D+1}=1}^{\widetilde{r}_{D+1}}\langle\mbox{\boldmath$\sigma$\unboldmath}^{j_{1},j_{2}}_{1},\mathbf{e}_{1}(\mbox{\boldmath$\alpha$\unboldmath})\rangle\cdot\dotsc\cdot\langle\mbox{\boldmath$\sigma$\unboldmath}^{j_{D},j_{D+1}}_{D},\mathbf{e}_{D}(\mbox{\boldmath$\alpha$\unboldmath})\rangle\,\mathbf{u}^{j_{1}}\circ\mathbf{v}^{j_{D+1}}\\ &=\sum_{j_{1}=1}^{\widetilde{r}_{1}}\dots\sum_{j_{D+1}=1}^{\widetilde{r}_{D+1}}\left(\mathbf{S}_{1}\times_{2}\mathbf{e}^{1}(\mbox{\boldmath$\alpha$\unboldmath})\right)_{j_{1},j_{2}}\cdot\dotsc\cdot\left(\mathbf{S}_{D}\times_{D+1}\mathbf{e}^{D}(\mbox{\boldmath$\alpha$\unboldmath})\right)_{j_{D},j_{D+1}}\,\mathbf{u}^{j_{1}}\circ\mathbf{v}^{j_{D+1}}\\ &=\sum_{j_{1}=1}^{\widetilde{r}_{1}}\sum_{j_{D+1}=1}^{\widetilde{r}_{D+1}}\Big{(}\prod_{i=1}^{D}\left(\mathbf{S}_{i}\times_{2}\mathbf{e}^{i}(\mbox{\boldmath$\alpha$\unboldmath})\right)\Big{)}_{j_{1}j_{D+1}}\,\mathbf{u}^{j_{1}}\circ\mathbf{v}^{j_{D+1}}=\mathrm{U}\mathrm{C}_{e}(\mbox{\boldmath$\alpha$\unboldmath})\mathrm{V}^{T}.\end{split}

Consider the SVD of the rescaled core matrix:

Ce(𝜶)Wc=UcΣcVcT.\mathrm{C}_{e}(\mbox{\boldmath$\alpha$\unboldmath})\mathrm{W}_{c}=\mathrm{U}_{c}\Sigma_{c}\mathrm{V}_{c}^{T}. (40)

Using this and the above representation of Φ~e(𝜶)\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}) we compute

Φ~e(𝜶)=UCe(𝜶)WcWc1VT=(UUc)Σc(VWc1Vc)T.\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})=\mathrm{U}\mathrm{C}_{e}(\mbox{\boldmath$\alpha$\unboldmath})\mathrm{W}_{c}\mathrm{W}_{c}^{-1}\mathrm{V}^{T}=\left({\mathrm{U}}\mathrm{U}_{c}\right)\Sigma_{c}\left({\mathrm{V}}\mathrm{W}_{c}^{-1}\mathrm{V}_{c}\right)^{T}. (41)

The right-hand side of (41) is the thin SVD of Φ~e(𝜶)\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}), since matrices U{\mathrm{U}}, Uc\mathrm{U}_{c}, VWc1{\mathrm{V}}\mathrm{W}_{c}^{-1}, and Vc\mathrm{V}_{c} are all orthogonal. We conclude that the coordinates {𝛃1(𝛂),,𝛃n(𝛂)}\left\{\mbox{\boldmath$\beta$\unboldmath}_{1}(\mbox{\boldmath$\alpha$\unboldmath}),\dots,\mbox{\boldmath$\beta$\unboldmath}_{n}(\mbox{\boldmath$\alpha$\unboldmath})\right\} of the local reduced basis in the universal space V~\widetilde{V} are the first nn columns of Uc\mathrm{U}_{c}. The parameter-specific basis is then {𝐳i(𝜶)}i=1n\{\mathbf{z}_{i}(\mbox{\boldmath$\alpha$\unboldmath})\}_{i=1}^{n}, with 𝐳i(𝜶)=U𝜷i(𝜶)\mathbf{z}_{i}(\mbox{\boldmath$\alpha$\unboldmath})={\mathrm{U}}\mbox{\boldmath$\beta$\unboldmath}_{i}(\mbox{\boldmath$\alpha$\unboldmath}) (not actually computed at the online stage).

We summarize the above in the following algorithm.

Algorithm 3 (TT-TROM).
  • Offline stage.
    Input: snapshot tensor 𝚽M×n1××nD×N\boldsymbol{\Phi}\in\mathbb{R}^{M\times n_{1}\times\ldots\times n_{D}\times N} and target accuracy ε~\widetilde{\varepsilon};
    Output: Compression ranks, TT decomposition matrices and third order tensors as in (35)–(36);
    Compute: Use algorithm from [58] with prescribed accuracy ε~\widetilde{\varepsilon} to compute TT decomposition (34).

  • Online stage.
    Input: online(𝚽~)\mbox{\rm online}(\widetilde{\boldsymbol{\Phi}}) as defined in (38), reduced space dimension nmin{r~1,r~D+1}n\leq\min\{\widetilde{r}_{1},\widetilde{r}_{D+1}\}, and parameter vector 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A};
    Output: Coordinates of the reduced basis in V~\widetilde{V}: {𝜷i(𝜶)}i=1nr~1\{\mbox{\boldmath$\beta$\unboldmath}_{i}(\mbox{\boldmath$\alpha$\unboldmath})\}_{i=1}^{n}\subset\mathbb{R}^{\widetilde{r}_{1}};
    Compute:

    1. 1.

      Use tensors 𝐒i\mathbf{S}_{i} to assemble the core matrix Ce(𝜶)r~1×r~D+1\mathrm{C}_{e}(\mbox{\boldmath$\alpha$\unboldmath})\in\mathbb{R}^{\widetilde{r}_{1}\times\widetilde{r}_{D+1}} as in (39);

    2. 2.

      Compute the SVD of the scaled core matrix Ce(𝜶)Wc=UcΣcVcT\mathrm{C}_{e}(\mbox{\boldmath$\alpha$\unboldmath})\mathrm{W}_{c}=\mathrm{U}_{c}\Sigma_{c}\mathrm{V}_{c}^{T} with Uc=[𝐮~1,,𝐮~r~1]\mathrm{U}_{c}=[\widetilde{\mathbf{u}}_{1},\ldots,\widetilde{\mathbf{u}}_{\widetilde{r}_{1}}];

    3. 3.

      Set 𝜷i(𝜶)=𝐮~i\mbox{\boldmath$\beta$\unboldmath}_{i}(\mbox{\boldmath$\alpha$\unboldmath})=\widetilde{\mathbf{u}}_{i}, i=1,,ni=1,\ldots,n.

3.8 General parameter sampling

Grid-based sampling of parameter space can be computationally expensive or not applicable if the set of admissible parameters 𝒜\mathcal{A} is not a box (or an image of a box) in Euclidean space. However, interpolatory TROMs introduced above can be extended to accommodate a more general sampling set 𝒜^{\widehat{\mathcal{A}}}. If 𝒜\mathcal{A} does have the Cartesian structure (a box or an image of a box), then one way to reduce offline computational costs is to compute the snapshots for only a few parameter values from a Cartesian grid 𝒜^𝒜{\widehat{\mathcal{A}}}\subset\mathcal{A}. To recover the missing entries of the full snapshot tensor 𝚽\boldsymbol{\Phi}, one may use a low-rank tensor completion method, e.g., one of those studied in [53, 28, 41, 69, 6]. The low-rank completion can be performed for any of the three compressed tensor formats considered above. We shall investigate this option elsewhere. In this paper, we consider another (more general) approach.

With a slight abuse of notation, let 𝒜^={𝜶^1,,𝜶^K}𝒜{{\widehat{\mathcal{A}}}}=\{\widehat{\boldsymbol{\alpha}}_{1},\dots,\widehat{\boldsymbol{\alpha}}_{K}\}\subset\mathcal{A} be a set of sampled parameter values. We assume that 𝒜^{\widehat{\mathcal{A}}} is a frame in D\mathbb{R}^{D} and so KDK\geq D. Note that KK does not obey (6) for a general sampling. Given an out-of-sample vector of parameters 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A}, let

𝐞:𝜶K\mathbf{e}\,:\,\mbox{\boldmath$\alpha$\unboldmath}\to\mathbb{R}^{K} (42)

be the representation of 𝜶\alpha in 𝒜^{\widehat{\mathcal{A}}}, i.e.,

𝜶=j=1Kaj𝜶^j,𝐞(𝜶)=(a1,,aK)T,\mbox{\boldmath$\alpha$\unboldmath}=\sum_{j=1}^{K}a_{j}\widehat{\boldsymbol{\alpha}}_{j},\quad\mathbf{e}(\mbox{\boldmath$\alpha$\unboldmath})=(a_{1},\dots,a_{K})^{T}, (43)

with an additional constraint enforcing uniqueness of the representation.

Similarly to (16) for the Cartesian grid case, we assume that for a smooth function g:𝒜{g}:\mathcal{A}\to\mathbb{R} it holds

g(𝜶)j=1Kajg(𝜶^j).{g}(\mbox{\boldmath$\alpha$\unboldmath})\approx\sum_{j=1}^{K}a_{j}{g}(\widehat{\boldsymbol{\alpha}}_{j}). (44)

In Section 5.1 we describe one particular choice of 𝐞(𝜶)\mathbf{e}(\mbox{\boldmath$\alpha$\unboldmath}) that is used in all numerical experiments reported in Section 5.

To assemble the snapshot tensor, for each 𝜶^j𝒜^\widehat{\boldsymbol{\alpha}}_{j}\in{\widehat{\mathcal{A}}}, j=1,,Kj=1,\ldots,K, collect the snapshot vectors 𝐮(tk,𝜶^j)=(u1(tk,𝜶^j),,uM(tk,𝜶^j))T\mathbf{u}(t_{k},\widehat{\boldsymbol{\alpha}}_{j})=\left(u_{1}(t_{k},\widehat{\boldsymbol{\alpha}}_{j}),\ldots,u_{M}(t_{k},\widehat{\boldsymbol{\alpha}}_{j})\right)^{T}, k=1,,Nk=1,\dots,N, and arrange them in a third order tensor 𝚽M×K×N\boldsymbol{\Phi}\in\mathbb{R}^{M\times K\times N} with entries

(𝚽)ijk=ui(tk,𝜶^j),i=1,,M,j=1,,K,k=1,,N.(\boldsymbol{\Phi})_{ijk}=u_{i}(t_{k},\widehat{\boldsymbol{\alpha}}_{j}),\quad i=1,\ldots,M,\quad j=1,\ldots,K,\quad k=1,\ldots,N. (45)

Then, for any 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A}, the parameter-specific reduced basis is defined as the first nn left singular vectors of

Φ~e(𝜶)=𝚽~×2𝐞(𝜶),\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})=\widetilde{\boldsymbol{\Phi}}\times_{2}\mathbf{e}(\mbox{\boldmath$\alpha$\unboldmath}), (46)

where 𝚽~\widetilde{\boldsymbol{\Phi}} is a low rank approximation of the snapshot tensor 𝚽\boldsymbol{\Phi} with entries (45). The same three compressed tensor formats considered above (CP, HOSVD and TT) can be used for 𝚽~\widetilde{\boldsymbol{\Phi}}, with TT format being inferior to HOSVD (for 3D tensors both HOSVD- and TT-decomposition are Tucker tensors). An orthogonal basis in the universal space of 𝚽~\widetilde{\boldsymbol{\Phi}} and coordinates of a local parameter-specific basis in it are computed similarly to the Cartesian grid sampling cases considered previously in Sections 3.53.7. The only difference is that all the calculations therein are performed setting D=1D=1 and replacing 𝐞1(𝜶)\mathbf{e}^{1}(\mbox{\boldmath$\alpha$\unboldmath}) with 𝐞(𝜶)\mathbf{e}(\mbox{\boldmath$\alpha$\unboldmath}). This includes the optimality result (33), where the factor D+2\sqrt{D+2} becomes 3\sqrt{3}.

3.9 Complexity and compression analysis

Projection-based parametric ROM framework consists in general of the following steps.

  • (i)

    High-fidelity simulations of (1) to generate the snapshot tensor 𝚽\boldsymbol{\Phi};

  • (ii)

    Offline stage: computing the compressed approximation 𝚽~\widetilde{\boldsymbol{\Phi}} to 𝚽\boldsymbol{\Phi} in one of low-rank tensor formats;

  • (iii)

    Passing the online(𝚽~)\mbox{online}(\widetilde{\boldsymbol{\Phi}}) part of the compressed tensor to the online stage;

  • (iv)

    Online stage: using online(𝚽~)\mbox{online}(\widetilde{\boldsymbol{\Phi}}) to compute the coordinates of the parameter-specific reduced basis for an input 𝜶\alpha;

  • (v)

    Solving (1) projected onto the reduced space.

Since steps (i) and (v) are common for all projection-based ROM approaches, we focus below on the computational and storage/transmission costs invoked in steps (ii)–(iv). The necessary details on step (v) are included in Section 3.10.

First, we discuss briefly the computational costs at the more expensive offline stage. For CP-TROM, the standard algorithm for finding 𝚽~\widetilde{\boldsymbol{\Phi}} in CP format (20) is the ALS method [36] which for a given CP rank RR iteratively fits a rank RR tensor Φ~\widetilde{\Phi} by solving on each iteration D+2D+2 least squares problems for the factors 𝐮r\mathbf{u}^{r}, 𝝈ir\mbox{\boldmath$\sigma$\unboldmath}^{r}_{i}, i=1,,Di=1,\dots,D, and 𝐯r\mathbf{v}^{r}, r=1,,Rr=1,\ldots,R. While straightforward to implement, the method is sensitive to the choice of initial guess and may converge slowly. We refer the reader to [46] for a guidance on the literature on improving the efficiency of ALS and possible alternatives. On the other hand, computing 𝚽~\widetilde{\boldsymbol{\Phi}} in either HOSVD or TT formats relies on finding truncated SVDs for matrix unfoldings of 𝚽\boldsymbol{\Phi} [21, 58]. Therefore, the computational complexity and cost of step (ii) for HOSVD- and TT-TROM is essentially the same as that of standard POD-ROM.

Second, to measure the amount of information transmitted to the online stage at step (iii), we introduce the compression factor CF, defined as

CF=#(𝚽)#(online(𝚽~)),\text{CF}=\frac{\#(\boldsymbol{\Phi})}{\#(\mbox{online}(\widetilde{\boldsymbol{\Phi}}))}\,, (47)

where we denote by #(𝚿)\#(\boldsymbol{\Psi}) the number of floating point numbers needed to store a tensor 𝚿\boldsymbol{\Psi}. Specifically, #(𝚽)=MKN\#(\boldsymbol{\Phi})=MKN is simply the total number of entries in 𝚽\boldsymbol{\Phi}, while #(online(𝚽~))\#(\mbox{online}(\widetilde{\boldsymbol{\Phi}})) is the number of entries needed to store all the factors passed to the online stage, as defined in (24), (29) and (38) for CP-, HOSVD- and TT- TROMs, which we summarize in Table 1.

Table 1: Number of entries needed to store online(𝚽~)\mbox{online}(\widetilde{\boldsymbol{\Phi}}) for CP, HOSVD and TT formats.
#(online(𝚽~))\#(\mbox{online}(\widetilde{\boldsymbol{\Phi}}))
Format Cartesian grid-based General
CP R(i=1Dni+R+1)R\big{(}\sum\limits_{i=1}^{D}n_{i}+R+1\big{)} R(K+R+1)R\big{(}K+R+1\big{)}
HOSVD N~M~i=1Dn~i+i=1Dn~ini\widetilde{N}\widetilde{M}\prod\limits_{i=1}^{D}\widetilde{n}_{i}+\sum\limits_{i=1}^{D}\widetilde{n}_{i}n_{i} N~M~n~1+n~1K\widetilde{N}\widetilde{M}\widetilde{n}_{1}+\widetilde{n}_{1}K
TT r~D+1+i=1Dr~inir~i+1\widetilde{r}_{D+1}+\sum\limits_{i=1}^{D}\widetilde{r}_{i}n_{i}\widetilde{r}_{i+1} r~2+r~1Kr~2\widetilde{r}_{2}+\widetilde{r}_{1}K\widetilde{r}_{2}

Table 1 shows that the compression factor is largely determined by the compression ranks. In turn, the ranks depend on ε~\widetilde{\varepsilon} and variability of observed states.

Third, the computational complexity of finding 𝜶\alpha-specific reduced basis in step (iv) is determined by the interpolation procedure and the computation of first nn left singular vectors of the core matrix. Since vectors 𝐞i(𝜶)\mathbf{e}^{i}(\mbox{\boldmath$\alpha$\unboldmath}) contain very few non-zero entries, e.g., p=2p=2 or 33 of them for the Cartesian sampling, the number of operations for computing core matrices C(𝜶)\mathrm{C}(\mbox{\boldmath$\alpha$\unboldmath}) for CP-, HOSVD- and TT-TROM is

O(R2),O(M~N~i=1Dn~i),andO(i=2Dr~i1r~ir~i+1),O\left(R^{2}\right),~{}~{}O\Big{(}\widetilde{M}\widetilde{N}\prod\limits_{i=1}^{D}\widetilde{n}_{i}\Big{)},~{}~{}\text{and}~{}~{}O\Big{(}\sum\limits_{i=2}^{D}\widetilde{r}_{i-1}\widetilde{r}_{i}\widetilde{r}_{i+1}\Big{)}, (48)

respectively. CP-, HOSVD- and TT-TROM algorithms proceed to compute the SVD of small core matrices of sizes R×RR\times R, M~×N~\widetilde{M}\times\widetilde{N} and r~1×r~D+1\widetilde{r}_{1}\times\widetilde{r}_{D+1}, respectively. If a reduced basis in the physical space is desired, then one finds its vectors as linear combinations of columns of U\mathrm{U}, which requires O(MRn)O(MRn), O(MM~n)O(M\widetilde{M}n) or O(Mr~1n)O(M\widetilde{r}_{1}n) operations for CP-, HOSVD- or TT-TROM, respectively. Section 3.10 below discusses how these costs can be avoided at the online phase. We note that for a fixed compression accuracy ϵ~\widetilde{\epsilon}, it is often observed in practice that the corresponding ranks of HOSVD and TT formats satisfy M~r~1\widetilde{M}\simeq\widetilde{r}_{1}, N~r~D+1\widetilde{N}\simeq\widetilde{r}_{D+1}.

In summary, the computational costs of the offline stage for TROMs are comparable to those of POD-ROM for a multi-parameter problem. At the online stage complexity of all preparatory steps depends only on compressed tensor ranks rather than the size of the snapshot tensor 𝚽\boldsymbol{\Phi}. The amount of information transmitted from offline to online stages is determined by the compressed tensor ranks, as should be clear from Table 1.

3.10 TROM evaluation

Besides finding a suitable reduced basis, a fast evaluation of the reduced model for any incoming 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A} is required for a reduced modeling scheme to be effective. Efficient implementation of a projected parametric model is a well-known challenge that have been addressed in the literature with various approaches; see, e.g.,  [12, 30, 17, 24, 15, 10, 39, 47]. The tensorial approach presented here does not directly contribute to resolving this issue, but it does not make it harder either and so techniques known from the literature can be adapted in the TROM framework.

For example, assume that F(t,𝐮,𝜶)F(t,\mathbf{u},\mbox{\boldmath$\alpha$\unboldmath}) from (1) has an affine dependence on parameters and linear dependence on 𝐮\mathbf{u}:

F(t,𝐮,𝜶)=i=1Pfi(𝜶)Ai𝐮,F(t,\mathbf{u},\mbox{\boldmath$\alpha$\unboldmath})=\sum_{i=1}^{P}f_{i}(\mbox{\boldmath$\alpha$\unboldmath})\mathrm{A}_{i}\mathbf{u},

with some fi:𝒜f_{i}:\mathcal{A}\to\mathbb{R} and parameter-independent AiM×M\mathrm{A}_{i}\in\mathbb{R}^{M\times M}. We assume that PP is not too large, at least independent of other dimensions. Then the offline stage of model reduction consists of projecting matrices onto the universal space by computing A^i=UTAiU\widehat{\mathrm{A}}_{i}=\mathrm{U}^{T}\mathrm{A}_{i}\mathrm{U}, where U\mathrm{U} is an orthogonal basis matrix for V~\widetilde{V} provided by the tensor decompositions. The new matrices A^i\widehat{\mathrm{A}}_{i} have the reduced size Tr×TrT_{r}\times T_{r}, with Tr{R,M~,r~1}T_{r}\in\{R,\widetilde{M},\widetilde{r}_{1}\} for CP-, HOSVD- and TT-TROMs, respectively.

For each of TROMs, denote by Uc(n)\mathrm{U}_{c}(n) the matrix of the first nn columns of Uc\mathrm{U}_{c}, left singular vectors of α\alpha-specific core matrices. During the online stage one solves the system projected further on the parameter-specific local basis:

𝐯t=i=1Pfi(𝜶)UcT(n)A^iUc(n)𝐯,\mathbf{v}_{t}=\sum_{i=1}^{P}f_{i}(\mbox{\boldmath$\alpha$\unboldmath}){\mathrm{U}^{T}_{c}(n)}\widehat{\mathrm{A}}_{i}{\mathrm{U}_{c}(n)}\mathbf{v},

where 𝐯(t)\mathbf{v}(t) is the trajectory in a space spanned by the columns of Uc(n)Tr×n{\mathrm{U}_{c}(n)}\in\mathbb{R}^{T_{r}\times n}, i.e., the corresponding physical states are given by 𝐮(t)=UUc(n)𝐯(t)\mathbf{u}(t)=\mathrm{U}{\mathrm{U}_{c}(n)}\mathbf{v}(t). We see that online computations depend only on reduced dimensions (tensor ranks) and the small dimension nn of parameter-specific basis. This observation can be extended to the case when FF has a low order polynomial non-linearity with respect to 𝐮\mathbf{u}. For example, quadratic nonlinear terms, as in Burgers or Navier-Stokes equations, can be evaluated in O(Tr2)O(T_{r}^{2}) operations on each time step given a vector 𝐯\mathbf{v} in the local reduced basis.

To evaluate more general nonlinear terms, one can use a hyper-reduction technique such as the discrete empirical interpolation method (DEIM) [17]. In this approach, the nonlinear term is approximated in a basis of its snapshots. As an example, consider F(t,𝐮,𝜶)=A𝐮+f(t,𝐮(t),𝜶)F(t,\mathbf{u},\mbox{\boldmath$\alpha$\unboldmath})=\mathrm{A}\mathbf{u}+f(t,\mathbf{u}(t),\mbox{\boldmath$\alpha$\unboldmath}), with A𝐮\mathrm{A}\mathbf{u} representing linear and f(t,𝐮(t),𝜶)f(t,\mathbf{u}(t),\mbox{\boldmath$\alpha$\unboldmath}) representing the non-linear part of FF. Define the snapshots 𝐟i=f(ti,𝐮(ti),𝜶i)\mathbf{f}_{i}=f(t_{i},\mathbf{u}(t_{i}),\mbox{\boldmath$\alpha$\unboldmath}_{i}), i=1,,NDEIMi=1,\dots,N_{\rm DEIM} for some “greedy” choice of parameters and time instances and high-fidelity solution 𝐮\mathbf{u}. Denote by Q\mathrm{Q} an orthogonal basis matrix for span{𝐟1,,𝐟NDEIM}\mbox{span}\{\mathbf{f}_{1},\dots,\mathbf{f}_{N_{\rm DEIM}}\}, then DEIM approximates

f(t,𝐮,𝜶)Q(PQ)1Pf(t,𝐮,𝜶),f(t,\mathbf{u},\mbox{\boldmath$\alpha$\unboldmath})\approx\mathrm{Q}(\mathrm{P}\mathrm{Q})^{-1}\mathrm{P}f(t,\mathbf{u},\mbox{\boldmath$\alpha$\unboldmath}),

where PT\mathrm{P}^{T} is an NDEIM×MN_{\rm DEIM}\times M “selection” matrix such that for any fMf\in\mathbb{R}^{M}, Pf\mathrm{P}f contains NDEIMN_{\rm DEIM} selected entries of ff. A particular P\mathrm{P} corresponds to the choice of spatial interpolation nodes, cf. [17]. In TROM one may pre-compute Q^=UTQ\widehat{\mathrm{Q}}=\mathrm{U}^{T}\mathrm{Q} and A^=UTAU\widehat{\mathrm{A}}=\mathrm{U}^{T}\mathrm{A}\mathrm{U} during the offline stage, then solve at the online stage

𝐯t=UcT(n)A^Uc(n)𝐯+UcT(n)Q^(PQ)1Pf(t,UUc(n)𝐯,𝜶),\mathbf{v}_{t}={\mathrm{U}_{c}^{T}(n)}\widehat{\mathrm{A}}\mathrm{U}_{c}(n)\mathbf{v}+{\mathrm{U}^{T}_{c}(n)}\widehat{\mathrm{Q}}(\mathrm{P}\mathrm{Q})^{-1}\mathrm{P}f(t,\mathrm{U}{\mathrm{U}_{c}(n)}\mathbf{v},\mbox{\boldmath$\alpha$\unboldmath}),

with costs depending on compressed tensor ranks, nn, and the dimension of DEIM space, but not on the dimensions of high-fidelity simulations. It is an interesting question, whether the tensor technique can be applied to make the DEIM space parameter-specific for more efficient reduce online computations. We plan to address this question elsewhere.

4 Prediction analysis

In this section we assess the prediction power of the reduced basis 𝒵n(𝜶)={𝐳1,,𝐳n}\mathcal{Z}_{n}(\mbox{\boldmath$\alpha$\unboldmath})=\{\mathbf{z}_{1},\ldots,\mathbf{z}_{n}\} consisting of the first nn left singular vectors of Φ~e(𝜶)\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}) from (18), for a parameter 𝜶=(α1,,αD)T𝒜\mbox{\boldmath$\alpha$\unboldmath}=(\alpha_{1},\ldots,\alpha_{D})^{T}\in\mathcal{A}, not necessarily from a sampling set; i.e., [𝐳1,,𝐳n]=UUc(n)[\mathbf{z}_{1},\ldots,\mathbf{z}_{n}]=\mathrm{U}{\mathrm{U}_{c}(n)}.

For the discussion below we also need the following notation. Given an 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A}, we denote by 𝝍i=𝐮(ti,𝜶)M\boldsymbol{\psi}_{i}=\mathbf{u}(t_{i},\mbox{\boldmath$\alpha$\unboldmath})\in\mathbb{R}^{M}, i=1,,Ni=1,\dots,N, the snapshots of a high-fidelity solution to (1) and let Ψ(𝜶)=[𝝍1,,𝝍N]M×N\Psi(\mbox{\boldmath$\alpha$\unboldmath})=[\boldsymbol{\psi}_{1},\dots,\boldsymbol{\psi}_{N}]\in\mathbb{R}^{M\times N} be the corresponding snapshot matrix. Note that in practice the snapshots for out-of-sample parameters are not available, so the matrix Ψ(𝜶)\Psi(\mbox{\boldmath$\alpha$\unboldmath}) should be treated as unknown.

We estimate the prediction power of 𝒵n(𝜶)\mathcal{Z}_{n}(\mbox{\boldmath$\alpha$\unboldmath}) in terms of the quantity

En(𝜶)=1NMi=1N𝝍ij=1n𝝍i,𝐳j𝐳j22,E_{n}(\mbox{\boldmath$\alpha$\unboldmath})=\frac{1}{NM}\sum_{i=1}^{N}\left\|\boldsymbol{\psi}_{i}-\sum_{j=1}^{n}\langle\boldsymbol{\psi}_{i},\mathbf{z}_{j}\rangle\mathbf{z}_{j}\right\|^{2}_{\ell^{2}}, (49)

which measures how accurate the solution 𝐮(t,𝜶)\mathbf{u}(t,\mbox{\boldmath$\alpha$\unboldmath}) at time instances tit_{i} can be represented in the reduced basis for the arbitrary but fixed 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A}. The scaling 1/(NM)1/(NM) accounts for the variation of dimensions NN and MM, which may correspond to the number of temporal and spatial degrees of freedom, respectively, if (1) comes from a discretization of a parabolic PDE defined in a spatial domain Ω\Omega. In this case and for uniform grids, the quantity in (49) is consistent with the L2(0,T,L2(Ω))L^{2}(0,T,L^{2}(\Omega)) norm.

Below we prove an estimate for En(𝜶)E_{n}(\mbox{\boldmath$\alpha$\unboldmath}) in terms of ε~\widetilde{\varepsilon} from (8), the singular values of Φ~e(𝜶)\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}) and interpolation properties of 𝐞i(𝜶)\mathbf{e}^{i}(\mbox{\boldmath$\alpha$\unboldmath}), i=1,,Di=1,\ldots,D. To make use of the latter, we introduce the following quantities related to the interpolation procedure. For Cartesian grid-based sampling we define the maximum grid step

δi=max1jni1|α^jiα^j+1i|,i=1,,D.\delta_{i}=\max\limits_{1\leq j\leq n_{i}-1}\left|\widehat{\alpha}_{j}^{i}-\widehat{\alpha}_{j+1}^{i}\right|,\quad i=1,\ldots,D. (50)

Relation (17) implies that the interpolation procedure (15)–(17) is of order pp, i.e., for any sufficiently smooth f:[αimin,αimax]f:[\alpha_{i}^{\min},\alpha_{i}^{\max}]\to\mathbb{R} it holds

supa[αimin,αimax]|f(a)j=1nieji(a𝐞i)f(α^ij)|Caf(p)C([αimin,αimax])δip,\sup_{a\in[\alpha_{i}^{\min},\alpha_{i}^{\max}]}\Big{|}f(a)-\sum_{j=1}^{n_{i}}e^{i}_{j}\left(a\mathbf{e}_{i}\right)f(\widehat{\alpha}_{i}^{j})\Big{|}\leq C_{a}\|f^{(p)}\|_{C([\alpha_{i}^{\min},\alpha_{i}^{\max}])}\delta_{i}^{p}, (51)

for i=1,,Di=1,\dots,D, where 𝐞ini\mathbf{e}_{i}\in\mathbb{R}^{n_{i}} is the ithi^{\mbox{\scriptsize th}} column of an ni×nin_{i}\times n_{i} identity matrix. The constant CaC_{a} does not depend on ff. We let δp=i=1Dδip\delta^{p}=\sum_{i=1}^{D}\delta_{i}^{p} and also assume that the interpolation procedure is stable in the sense that

(j=1ni|eji(a𝐞i)|2)12Ce,\left(\sum_{j=1}^{n_{i}}\left|e^{i}_{j}(a\mathbf{e}_{i})\right|^{2}\right)^{\frac{1}{2}}\leq C_{e}, (52)

with some CeC_{e} independent of a[αimin,αimax]a\in[\alpha_{i}^{\min},\alpha_{i}^{\max}] and i=1,,Di=1,\ldots,D. For the example of linear interpolation with p=2p=2 and αimin\alpha_{i}^{\min}, αimax\alpha_{i}^{\max} included among the grid nodes, bounds (51) and (52) hold with Ca=18C_{a}=\frac{1}{8} and Ce=1C_{e}=1.

To estimate En(𝜶)E_{n}(\mbox{\boldmath$\alpha$\unboldmath}), consider the SVD of Φ~e(𝜶)M×N\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})\in\mathbb{R}^{M\times N} given by

Φ~e(𝜶)=U~Σ~V~T,withΣ~=diag(σ~1,,σ~N).\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})=\widetilde{\mathrm{U}}\widetilde{\Sigma}\widetilde{\mathrm{V}}^{T},~{}~{}\text{with}~{}~{}\widetilde{\Sigma}=\text{diag}(\widetilde{\sigma}_{1},\dots,\widetilde{\sigma}_{N}). (53)

Then Z=[𝐳1,,𝐳n]M×n\mathrm{Z}=[\mathbf{z}_{1},\ldots,\mathbf{z}_{n}]\in\mathbb{R}^{M\times n} is build as the first nn columns of U~\widetilde{\mathrm{U}}, the reduced basis vectors, i.e., Z=UUc\mathrm{Z}=\mathrm{U}\mathrm{U}_{c}. Then,

En(𝜶)\displaystyle E_{n}(\mbox{\boldmath$\alpha$\unboldmath}) =1NM(IZZT)Ψ(𝜶)F2\displaystyle=\frac{1}{NM}\left\|(\mathrm{I}-\mathrm{Z}\mathrm{Z}^{T})\Psi(\mbox{\boldmath$\alpha$\unboldmath})\right\|^{2}_{F}
1NM((IZZT)(Ψ(𝜶)Φ~e(𝜶))F+(IZZT)Φ~e(𝜶)F)2\displaystyle\leq\frac{1}{NM}\left(\left\|(\mathrm{I}-\mathrm{Z}\mathrm{Z}^{T})(\Psi(\mbox{\boldmath$\alpha$\unboldmath})-\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}))\right\|_{F}+\left\|(\mathrm{I}-\mathrm{Z}\mathrm{Z}^{T})\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})\right\|_{F}\right)^{2}
1NM(Ψ(𝜶)Φ~e(𝜶)F+(IZZT)Φ~e(𝜶)F)2,\displaystyle\leq\frac{1}{NM}\left(\left\|{\Psi}(\mbox{\boldmath$\alpha$\unboldmath})-\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})\right\|_{F}+\left\|(\mathrm{I}-\mathrm{Z}\mathrm{Z}^{T})\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})\right\|_{F}\right)^{2}, (54)

where we used triangle inequality and IZZT1\|\mathrm{I}-\mathrm{Z}\mathrm{Z}^{T}\|\leq 1 for the spectral norm of the projector. For the last term in (54), we observe

(IZZT)Φ~e(𝜶)F=U~diag(0,,0,σ~n+1,,σ~N)V~TF=(j=n+1Nσ~j2)12.\left\|(\mathrm{I}-\mathrm{Z}\mathrm{Z}^{T})\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})\right\|_{F}=\left\|\widetilde{\mathrm{U}}\;\text{diag}(0,\dots,0,\widetilde{\sigma}_{n+1},\dots,\widetilde{\sigma}_{N})\;\widetilde{\mathrm{V}}^{T}\right\|_{F}=\left(\sum_{j=n+1}^{N}\widetilde{\sigma}_{j}^{2}\right)^{\frac{1}{2}}. (55)

To handle the first term of (54), consider the extraction

Φe(𝜶)=𝚽×2𝐞1(𝜶)×3𝐞2(𝜶)×D+1𝐞D(𝜶)\Phi_{e}(\mbox{\boldmath$\alpha$\unboldmath})=\boldsymbol{\Phi}\times_{2}\mathbf{e}^{1}(\mbox{\boldmath$\alpha$\unboldmath})\times_{3}\mathbf{e}^{2}(\mbox{\boldmath$\alpha$\unboldmath})\dots\times_{D+1}\mathbf{e}^{D}(\mbox{\boldmath$\alpha$\unboldmath}) (56)

and proceed using the triangle inequality

Ψ(𝜶)Φ~e(𝜶)FΨ(𝜶)Φe(𝜶)F+Φe(𝜶)Φ~e(𝜶)F.\big{\|}\Psi(\mbox{\boldmath$\alpha$\unboldmath})-\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})\big{\|}_{F}\leq\big{\|}\Psi(\mbox{\boldmath$\alpha$\unboldmath})-\Phi_{e}(\mbox{\boldmath$\alpha$\unboldmath})\big{\|}_{F}+\big{\|}\Phi_{e}(\mbox{\boldmath$\alpha$\unboldmath})-\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})\big{\|}_{F}. (57)

We use the stability of interpolation (52) and (8) to bound the second term of (57). Specifically,

Φe(𝜶)Φ~e(𝜶)F=(𝚽𝚽~)×2𝐞1(𝜶)×3𝐞2(𝜶)×D+1𝐞D(𝜶)F𝚽𝚽~F𝐞1(𝜶)2𝐞2(𝜶)2𝐞D(𝜶)2(Ce)D𝚽𝚽~F(Ce)Dε~𝚽F.\begin{split}\left\|\Phi_{e}(\mbox{\boldmath$\alpha$\unboldmath})-\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath})\right\|_{F}&=\left\|(\boldsymbol{\Phi}-\widetilde{\boldsymbol{\Phi}})\times_{2}\mathbf{e}^{1}(\mbox{\boldmath$\alpha$\unboldmath})\times_{3}\mathbf{e}^{2}(\mbox{\boldmath$\alpha$\unboldmath})\dots\times_{D+1}\mathbf{e}^{D}(\mbox{\boldmath$\alpha$\unboldmath})\right\|_{F}\\ &\leq\left\|\boldsymbol{\Phi}-\widetilde{\boldsymbol{\Phi}}\right\|_{F}\|\mathbf{e}^{1}(\mbox{\boldmath$\alpha$\unboldmath})\|_{\ell^{2}}\|\mathbf{e}^{2}(\mbox{\boldmath$\alpha$\unboldmath})\|_{\ell^{2}}\dots\|\mathbf{e}^{D}(\mbox{\boldmath$\alpha$\unboldmath})\|_{\ell^{2}}\\ &\leq(C_{e})^{D}\left\|\boldsymbol{\Phi}-\widetilde{\boldsymbol{\Phi}}\right\|_{F}\leq(C_{e})^{D}\widetilde{\varepsilon}\left\|\boldsymbol{\Phi}\right\|_{F}.\end{split} (58)

It remains to handle the first term in (57). At this point we need more precise assumptions on the smoothness of 𝐮(t,𝜶)\mathbf{u}(t,\mbox{\boldmath$\alpha$\unboldmath}), the solution of (1). In particular,

𝐮C([0,T]×𝒜¯)M,𝐣𝐮α1j1αDjDC([0,T]×𝒜¯)M,|𝐣|p.\mathbf{u}\in C([0,T]\times\overline{\mathcal{A}})^{M},\quad\frac{\partial^{\mathbf{j}}\mathbf{u}}{\partial\alpha^{j_{1}}_{1}\dots\alpha^{j_{D}}_{D}}\in C([0,T]\times\overline{\mathcal{A}})^{M},\quad|\mathbf{j}|\leq p. (59)

We note that (59) is guaranteed to hold if the unique solution to (1) exists on (0,T1)(0,T_{1}) for all 𝜶𝒜1\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A}_{1}, with T<T1T<T_{1} and 𝒜¯𝒜1\overline{\mathcal{A}}\subset\mathcal{A}_{1} and FF is continuous with continuous partial derivatives in components of 𝐮\mathbf{u} and 𝜶\alpha of order up to pp [37].

Using interpolation property (51), we compute

(𝚽×2𝐞1(𝜶)):,i2,,iD,k=j=1n1ej1(𝜶)𝐮(tk,α^1j,α^2i2,,α^DiD)=𝐮(tk,α1,α^2i2,α^DiD)+Δ:,i2,,iD,k1,\begin{split}\left(\boldsymbol{\Phi}\times_{2}\mathbf{e}^{1}(\mbox{\boldmath$\alpha$\unboldmath})\right)_{:,i_{2},\dots,i_{D},k}&=\sum_{j=1}^{n_{1}}e^{1}_{j}(\mbox{\boldmath$\alpha$\unboldmath})\mathbf{u}(t_{k},\widehat{\alpha}_{1}^{j},\widehat{\alpha}_{2}^{i_{2}},\dots,\widehat{\alpha}_{D}^{i_{D}})\\ &=\mathbf{u}(t_{k},\alpha_{1},\widehat{\alpha}_{2}^{i_{2}}\dots,\widehat{\alpha}_{D}^{i_{D}})+\Delta^{1}_{:,i_{2},\dots,i_{D},k},\end{split} (60)

with the remainder term obeying a component-wise bound

|Δ:,i2,,iD,k1|Casupa[α1min,α1max]|p𝐮α1p(tk,a,α^2i2,,α^DiD)|δ1p,|\Delta^{1}_{:,i_{2},\dots,i_{D},k}|\leq C_{a}\sup_{a\in[\alpha_{1}^{\min},\alpha_{1}^{\max}]}\left|\frac{\partial^{p}\mathbf{u}}{\partial\alpha^{p}_{1}}(t_{k},a,\widehat{\alpha}_{2}^{i_{2}},\dots,\widehat{\alpha}_{D}^{i_{D}})\right|\delta_{1}^{p}, (61)

where the absolute value of vectors is understood entry-wise. Analogously, we compute

(𝚽×2𝐞1(𝜶)×3𝐞2(𝜶)):,i3,,iD,k=((𝚽×2𝐞1(𝜶))×2𝐞2(𝜶)):,i3,,iD,k=j=1n2ej2(𝜶)(𝐮(tk,α1,α^2j,α^3i3,,α^DiD)+Δ:,j,i3,,iD,k1)=𝐮(tk,α1,α2,α^3i3,,α^DiD)+Δ:,i3,,iD,k2+j=1n2ej2(𝜶)Δ:,j,i3,,iD,k1,\begin{split}\big{(}\boldsymbol{\Phi}\times_{2}\mathbf{e}^{1}(\mbox{\boldmath$\alpha$\unboldmath})&\times_{3}\mathbf{e}^{2}(\mbox{\boldmath$\alpha$\unboldmath})\big{)}_{:,i_{3},\dots,i_{D},k}=\left((\boldsymbol{\Phi}\times_{2}\mathbf{e}^{1}(\mbox{\boldmath$\alpha$\unboldmath}))\times_{2}\mathbf{e}^{2}(\mbox{\boldmath$\alpha$\unboldmath})\right)_{:,i_{3},\dots,i_{D},k}\\ &=\sum_{j=1}^{n_{2}}e^{2}_{j}(\mbox{\boldmath$\alpha$\unboldmath})\left(\mathbf{u}(t_{k},\alpha_{1},\widehat{\alpha}_{2}^{j},\widehat{\alpha}_{3}^{i_{3}},\dots,\widehat{\alpha}_{D}^{i_{D}})+\Delta^{1}_{:,j,i_{3},\dots,i_{D},k}\right)\\ &=\mathbf{u}(t_{k},\alpha_{1},\alpha_{2},\widehat{\alpha}_{3}^{i_{3}},\dots,\widehat{\alpha}_{D}^{i_{D}})+\Delta^{2}_{:,i_{3},\dots,i_{D},k}+\sum_{j=1}^{n_{2}}e^{2}_{j}(\mbox{\boldmath$\alpha$\unboldmath})\Delta^{1}_{:,j,i_{3},\dots,i_{D},k},\end{split}

with a component-wise bound for the remainder

|Δ:,i3,,iD,k2+j=1n2ej2(𝜶)Δ:,j,i3,,iD,k1|Casupa[α2min,α2max]|p𝐮α2p(tk,α1,a,α^3i3,,α^DiD)|δ2p+CeCasupa[α1min,α1max]|p𝐮α1p(tk,a,α^2i2,,α^DiD)|δ1p.\begin{split}\Big{|}\Delta^{2}_{:,i_{3},\dots,i_{D},k}+\sum_{j=1}^{n_{2}}e^{2}_{j}(\mbox{\boldmath$\alpha$\unboldmath})&\Delta^{1}_{:,j,i_{3},\dots,i_{D},k}\Big{|}\\ \leq C_{a}&\sup_{a\in[\alpha_{2}^{\min},\alpha_{2}^{\max}]}\left|\frac{\partial^{p}\mathbf{u}}{\partial\alpha^{p}_{2}}(t_{k},\alpha_{1},a,\widehat{\alpha}_{3}^{i_{3}},\dots,\widehat{\alpha}_{D}^{i_{D}})\right|\delta_{2}^{p}\\ +\;C_{e}\;C_{a}&\sup_{a\in[\alpha_{1}^{\min},\alpha_{1}^{\max}]}\left|\frac{\partial^{p}\mathbf{u}}{\partial\alpha^{p}_{1}}(t_{k},a,\widehat{\alpha}_{2}^{i_{2}},\dots,\widehat{\alpha}_{D}^{i_{D}})\right|\delta_{1}^{p}.\end{split} (62)

Applying the same argument repeatedly, we obtain

(Φe(𝜶)):,k=(𝚽×2𝐞1(𝜶)×3𝐞2(𝜶)×D+1𝐞D(𝜶)):,k=𝐮(tk,α1,α2,,αD)+Δ:,k=(Ψ(𝜶)):,k+Δ:,k,\begin{split}\left(\Phi_{e}(\mbox{\boldmath$\alpha$\unboldmath})\right)_{:,k}&=\left(\boldsymbol{\Phi}\times_{2}\mathbf{e}^{1}(\mbox{\boldmath$\alpha$\unboldmath})\times_{3}\mathbf{e}^{2}(\mbox{\boldmath$\alpha$\unboldmath})\dots\times_{D+1}\mathbf{e}^{D}(\mbox{\boldmath$\alpha$\unboldmath})\right)_{:,k}\\ &=\mathbf{u}(t_{k},\alpha_{1},\alpha_{2},\dots,\alpha_{D})+\Delta_{:,k}=\left(\Psi(\mbox{\boldmath$\alpha$\unboldmath})\right)_{:,k}+\Delta_{:,k},\end{split} (63)

with a component-wise bound for the remainder

|Δ:,k|Ca(\displaystyle\left|\Delta_{:,k}\right|\leq C_{a}\Big{(} supa[αDmin,αDmax]|p𝐮αDp(tk,α1,,αD1,a)|δDp+\displaystyle\sup_{a\in[\alpha_{D}^{\min},\alpha_{D}^{\max}]}\left|\frac{\partial^{p}\mathbf{u}}{\partial\alpha^{p}_{D}}(t_{k},\alpha_{1},\dots,\alpha_{D-1},a)\right|\delta_{D}^{p}+\dots
+(Ce)D2\displaystyle+\,(C_{e})^{D-2} supa[α2min,α2max]|p𝐮α2p(tk,α1,a,α^3i3,,α^DiD)|δ2p\displaystyle\sup_{a\in[\alpha_{2}^{\min},\alpha_{2}^{\max}]}\left|\frac{\partial^{p}\mathbf{u}}{\partial\alpha^{p}_{2}}(t_{k},\alpha_{1},a,\widehat{\alpha}_{3}^{i_{3}},\dots,\widehat{\alpha}_{D}^{i_{D}})\right|\delta_{2}^{p}
+(Ce)(D1)\displaystyle+\,(C_{e})^{(D-1)} supa[α1min,α1max]|p𝐮α1p(tk,a,α^2i2,,α^DiD)|δ1p).\displaystyle\sup_{a\in[\alpha_{1}^{\min},\alpha_{1}^{\max}]}\left|\frac{\partial^{p}\mathbf{u}}{\partial\alpha^{p}_{1}}(t_{k},a,\widehat{\alpha}_{2}^{i_{2}},\dots,\widehat{\alpha}_{D}^{i_{D}})\right|\delta_{1}^{p}\Big{)}.

Using the definition of the Frobenius norm, we arrive at

Ψ(𝜶)Φe(𝜶)FNMCaC𝐮max{(Ce)(D1),1}δp,\left\|\Psi(\mbox{\boldmath$\alpha$\unboldmath})-\Phi_{e}(\mbox{\boldmath$\alpha$\unboldmath})\right\|_{F}\leq\sqrt{NM}\;C_{a}C_{\mathbf{u}}\max\left\{(C_{e})^{(D-1)},1\right\}\;\delta^{p}, (64)

where C𝐮C_{\mathbf{u}} depends only on the smoothness of 𝐮\mathbf{u} with respect to the variations of parameters 𝜶\alpha. More precisely, we can take C𝐮=𝐮C(0,T;Cp(𝒜))C_{\mathbf{u}}=\|\mathbf{u}\|_{C(0,T;C^{p}(\mathcal{A}))}, which is bounded due to assumption (59).

Summarizing (54)–(64), we proved the following result.

Theorem 1.

Assume the solution 𝐮\mathbf{u} to (1) satisfies (59), 𝒜^{\widehat{\mathcal{A}}} is a Cartesian grid in parameter domain 𝒜\mathcal{A}. Then for any 𝛂𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A} the interpolatory TROM reduced basis 𝒵(𝛂)={𝐳1,,𝐳n}\mathcal{Z}(\mbox{\boldmath$\alpha$\unboldmath})=\{\mathbf{z}_{1},\ldots,\mathbf{z}_{n}\} delivers the following representation estimate

13NMi=1N𝐮(ti,𝜶)j=1n𝐮(ti,𝜶),𝐳j𝐳j221NM((Ce)2Dε~2𝚽F+i=n+1Nσ~i2)+CaC𝐮max{(Ce)2(D1),1}δ2p,\frac{1}{3NM}\sum_{i=1}^{N}\left\|\mathbf{u}(t_{i},\mbox{\boldmath$\alpha$\unboldmath})-\sum_{j=1}^{n}\left\langle\mathbf{u}(t_{i},\mbox{\boldmath$\alpha$\unboldmath}),\mathbf{z}_{j}\right\rangle\mathbf{z}_{j}\right\|^{2}_{\ell^{2}}\\ \leq\frac{1}{NM}\left((C_{e})^{2D}\widetilde{\varepsilon}^{2}\left\|\boldsymbol{\Phi}\right\|_{F}+\sum_{i=n+1}^{N}\widetilde{\sigma}_{i}^{2}\right)+C_{a}C_{\mathbf{u}}\max\left\{(C_{e})^{2(D-1)},1\right\}\delta^{2p}, (65)

with C𝐮=𝐮C(0,T;Cp(𝒜))C_{\mathbf{u}}=\|\mathbf{u}\|_{C(0,T;C^{p}(\mathcal{A}))} independent of 𝛂\alpha, sampling grid and nn.

We summarize here the definitions of quantities that appear in Theorem 1: NN is a number of time steps for snapshot collection, while MM is the spatial dimension of snapshots, i.e., 𝐮(ti,𝜶)M\mathbf{u}(t_{i},\mbox{\boldmath$\alpha$\unboldmath})\in\mathbb{R}^{M}, i=1,,Ni=1,\ldots,N; ε~\widetilde{\varepsilon} is the relative accuracy of the snapshot tensor compression from (8); σ~i\widetilde{\sigma}_{i} are the singular values of Φ~e(𝜶)\widetilde{\Phi}_{e}(\mbox{\boldmath$\alpha$\unboldmath}) from (18) (note that in TROMs we have an access to σ~i\widetilde{\sigma}_{i} as the singular values of core matrices (25), (31) and (40) for CP-, HOSVD-, and TT-TROM, respectively); δ=(i=1Dδip)1p\delta=\big{(}\sum_{i=1}^{D}\delta_{i}^{p}\big{)}^{\frac{1}{p}} is the grid step parameter of the Cartesian grid in 𝒜\mathcal{A}; pp is both the number of nearest grid points and the order of interpolation of the interpolation procedure (15)–(17); CeC_{e} is the interpolation stability constant from (52); and DD is the dimension of parameter space.

For the general parameter sampling, prediction power analysis follows the same lines as above, simply setting D=1D=1. However, the order of interpolation pp is slightly more difficult to formalize, so instead of (51)–(52) we rather assume

sup𝜶𝒜|f(𝜶)j=1K(𝐞(𝜶))jf(𝜶^j)|fCp(𝒜)δ,(j=1K|(𝐞(𝜶))j|2)12Ce\displaystyle\sup_{\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A}}\left|f(\mbox{\boldmath$\alpha$\unboldmath})-\sum_{j=1}^{K}(\mathbf{e}(\mbox{\boldmath$\alpha$\unboldmath}))_{j}f(\widehat{\boldsymbol{\alpha}}_{j})\right|\leq\|f\|_{C^{p}(\mathcal{A})}\delta,\quad\Big{(}\sum_{j=1}^{K}|(\mathbf{e}(\mbox{\boldmath$\alpha$\unboldmath}))_{j}|^{2}\Big{)}^{\frac{1}{2}}\leq C_{e} (66)

with some δ\delta depending on 𝒜^{\widehat{\mathcal{A}}}. The prediction estimate then becomes

13NMi=1N𝐮(ti,𝜶)j=1n𝐮(ti,𝜶),𝐳j𝐳j221NM((Ce)2ε~2𝚽F+i=n+1Nσ~i2)+C𝐮max{(Ce)2,1}δ2,\frac{1}{3NM}\sum_{i=1}^{N}\left\|\mathbf{u}(t_{i},\mbox{\boldmath$\alpha$\unboldmath})-\sum_{j=1}^{n}\left\langle\mathbf{u}(t_{i},\mbox{\boldmath$\alpha$\unboldmath}),\mathbf{z}_{j}\right\rangle\mathbf{z}_{j}\right\|^{2}_{\ell^{2}}\\ \leq\frac{1}{NM}\left((C_{e})^{2}\widetilde{\varepsilon}^{2}\left\|\boldsymbol{\Phi}\right\|_{F}+\sum_{i=n+1}^{N}\widetilde{\sigma}_{i}^{2}\right)+C_{\mathbf{u}}\max\{(C_{e})^{2},1\}\delta^{2}, (67)

with C𝐮=𝐮C(0,T;Cp(𝒜))C_{\mathbf{u}}=\|\mathbf{u}\|_{C(0,T;C^{p}(\mathcal{A}))}.

We finally, note that the feasibility of a sufficiently accurate lower rank representation of 𝚽\boldsymbol{\Phi} depends on the smoothness of 𝐮\mathbf{u} as a function of 𝐱,\mathbf{x}, tt and 𝜶\alpha. This question can be addressed by considering tensor decompositions of multivariate functions, e.g. [34, 55]. For these functional CP, HOSVD and hierarchical Tucker (including TT) formats, the dependence of compression ranks on ε~\widetilde{\varepsilon} from (8) and the regularity (smoothness) of 𝐮\mathbf{u} was studied in [32, 67, 62, 35, 66]. This compression property for multivariate functions was exploited to effectively represent solutions of parametric elliptic PDEs using tensor formats in [44, 20, 4, 27, 23] among other publications.

5 Numerical experiments

We perform several numerical experiments to assess the performance of the three TROM approches and compare them to the conventional POD-ROM. The testing in Section 5.2 is performed for a dynamical system originating from a discretization of linear parameter-dependent heat equation. In Section 5.3 a similar set of tests is carried out for a time-dependent parameterized advection-diffusion system.

5.1 General parameter sampling interpolation scheme

For the numerical examples in the general parameter sampling setting we employ the following interpolation scheme. Fix an integer qD+1q\geq D+1 and let 𝜶=(α1,,αD)𝒜\mbox{\boldmath$\alpha$\unboldmath}=(\alpha_{1},\ldots,\alpha_{D})\in\mathcal{A} be an out-of-sample parameter vector. The interpolation scheme is based on the weighted minimum norm fit over qq nearest neighbors of 𝜶\alpha in the sampling set. Thus, we denote by 𝜶^i1,,𝜶^iq\widehat{\boldsymbol{\alpha}}_{i_{1}},\ldots,\widehat{\boldsymbol{\alpha}}_{i_{q}} the qq closest parameter samples in 𝒜^{\widehat{\mathcal{A}}} to 𝜶\alpha and set dk=𝜶^ik𝜶>0d_{k}=\|\widehat{\boldsymbol{\alpha}}_{i_{k}}-\mbox{\boldmath$\alpha$\unboldmath}\|>0, k=1,,qk=1,\ldots,q. Next, define the weighting matrix

D=diag(d11,,dq1)q×q.{\mathrm{D}}=\mbox{diag}(d_{1}^{-1},\ldots,d_{q}^{-1})\in\mathbb{R}^{q\times q}.

Also, assemble the matrix

X=[(𝜶^i1)1(𝜶^i2)1(𝜶^iq)1(𝜶^i1)2(𝜶^i2)2(𝜶^iq)2(𝜶^i1)D(𝜶^i2)D(𝜶^iq)D111]D+1×q.{\mathrm{X}}=\begin{bmatrix}(\widehat{\boldsymbol{\alpha}}_{i_{1}})_{1}&(\widehat{\boldsymbol{\alpha}}_{i_{2}})_{1}&\cdots&(\widehat{\boldsymbol{\alpha}}_{i_{q}})_{1}\\ (\widehat{\boldsymbol{\alpha}}_{i_{1}})_{2}&(\widehat{\boldsymbol{\alpha}}_{i_{2}})_{2}&\cdots&(\widehat{\boldsymbol{\alpha}}_{i_{q}})_{2}\\ \vdots&\vdots&\vdots&\vdots\\ (\widehat{\boldsymbol{\alpha}}_{i_{1}})_{D}&(\widehat{\boldsymbol{\alpha}}_{i_{2}})_{D}&\cdots&(\widehat{\boldsymbol{\alpha}}_{i_{q}})_{D}\\ 1&1&\cdots&1\end{bmatrix}\in\mathbb{R}^{D+1\times q}.

Solve the weighted minimum norm fitting problem to obtain

𝐚^=D(XD)[𝜶1]q.\widehat{\mathbf{a}}={\mathrm{D}}({\mathrm{X}}{\mathrm{D}})^{\dagger}\begin{bmatrix}\mbox{\boldmath$\alpha$\unboldmath}\\ 1\end{bmatrix}\in\mathbb{R}^{q}. (68)

Note that the last row of X{\mathrm{X}} and the last entry of (𝜶,1)T(\mbox{\boldmath$\alpha$\unboldmath},1)^{T} enforces the condition that the entries of 𝐚^\widehat{\mathbf{a}} sum to one. Meanwhile, the presence of the weighting matrix puts more emphasis on the neighbors of 𝜶\alpha that are closest to it.

Once 𝐚^\widehat{\mathbf{a}} is obtained from (68), we define aja_{j}, j=1,,Kj=1,\ldots,K, the entries of 𝐞(𝜶)\mathbf{e}(\mbox{\boldmath$\alpha$\unboldmath}) as

aj={a^k,if j=ik{i1,,iq}0,otherwisea_{j}=\begin{cases}\widehat{a}_{k},&\text{if }j=i_{k}\in\{i_{1},\ldots,i_{q}\}\\ 0,&\text{otherwise}\end{cases}

Clearly, such construction enforces representation (43).

5.2 Parameterized heat equation

We first assess performance of the three TROM approaches on a dynamical system resulting from the discretization of a heat equation

wt=Δw,w_{t}=\Delta w, (69)

in a rectangular domain with three holes Ω=Ωr(Ω1Ω2Ω3)2\Omega=\Omega_{r}\setminus(\Omega_{1}\cup\Omega_{2}\cup\Omega_{3})\subset\mathbb{R}^{2}, where Ωr=[0,10]×[0,4]\Omega_{r}=[0,10]\times[0,4], and the holes are Ω1=[1,3]×[1,3]\Omega_{1}=[1,3]\times[1,3], Ω2=[4,6]×[1,3]\Omega_{2}=[4,6]\times[1,3], Ω3=[7,9]×[1,3]\Omega_{3}=[7,9]\times[1,3]. The PDE and geometry of Ω\Omega follow that of [31], while the boundary conditions are modified from those used in [31], as described below.

We parametrize the system with D=4D=4 parameters that enter the boundary conditions. Convection boundary conditions are enforced on the left side of the rectangle Γo=0×[0,4]\Gamma_{o}=0\times[0,4] and on the boundaries of each hole Ωj\partial\Omega_{j}, j=1,2,3j=1,2,3. Explicitly,

(𝐧w+α1(w1))|Γo=0,\left.(\mathbf{n}\cdot\nabla w+\alpha_{1}(w-1)\,)\right|_{\Gamma_{o}}=0, (70)

and

(𝐧w+12w)|Ωj=12αj+1,j=1,2,3,\left.\left(\mathbf{n}\cdot\nabla w+\frac{1}{2}w\right)\right|_{\partial\Omega_{j}}=\frac{1}{2}\alpha_{j+1},\quad j=1,2,3, (71)

i.e., the first parameter in 𝜶4\mbox{\boldmath$\alpha$\unboldmath}\in\mathbb{R}^{4} is Biot number at Γo\Gamma_{o} with a fixed outside temperature to=1t_{o}=1, while the other three parameters are the temperatures at Ωj\partial\Omega_{j}, j=1,2,3j=1,2,3, respectively, with Biot numbers equal to 12\frac{1}{2} on all three hole boundaries. The rest of the boundary of Ω\Omega is assumed to be insulated

(𝐧w)|ΩrΓo=0.\left.(\mathbf{n}\cdot\nabla w)\right|_{\partial\Omega_{r}\setminus\Gamma_{o}}=0. (72)

In (70)–(72), 𝐧\mathbf{n} is the outer unit normal. Observe that the boundary conditions (70)–(72) can be combined into

(𝐧w+q(𝐱,𝜶)w)|Ω=g(𝐱,𝜶),\left.(\mathbf{n}\cdot\nabla w+q(\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})w)\right|_{\partial\Omega}=g(\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath}),

for the appropriate choices of q(𝐱,𝜶)q(\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath}) and g(𝐱,𝜶)g(\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath}) defined on Ω\partial\Omega with 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A}, a parameter domain that we take to be the 4D box 𝒜=[0.01,0.5]×[0,0.9]3\mathcal{A}=[0.01,0.5]\times[0,0.9]^{3}. The initial temperature is taken to be zero throughout Ω\Omega.

The system (69)–(72) is discretized with P2P_{2} finite elements on a quasi-uniform triangulation of Ω\Omega resulting in M=3,562M=3,562 spatial degrees of freedom. The choice of standard nodal basis functions {θj(𝐱)}j=1M\{\theta_{j}(\mathbf{x})\}_{j=1}^{M} defines the mass MM×M\mathrm{M}\in\mathbb{R}^{M\times M} and stiffness KM×M\mathrm{K}\in\mathbb{R}^{M\times M} matrices, as well as boundary terms Q(𝜶)M×M\mathrm{Q}(\mbox{\boldmath$\alpha$\unboldmath})\in\mathbb{R}^{M\times M}, 𝐠(𝜶)M\mathbf{g}(\mbox{\boldmath$\alpha$\unboldmath})\in\mathbb{R}^{M} with entries given by

(Q)ij(𝜶)=Ωq(𝐱,𝜶)θj(𝐱)θi(𝐱)𝑑s𝐱,(𝐠)j(𝜶)=Ωg(𝐱,𝜶)θj(𝐱)𝑑s𝐱,(\mathrm{Q})_{ij}(\mbox{\boldmath$\alpha$\unboldmath})=\int_{\partial\Omega}q(\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})\theta_{j}(\mathbf{x})\theta_{i}(\mathbf{x})ds_{\mathbf{x}},\quad(\mathbf{g})_{j}(\mbox{\boldmath$\alpha$\unboldmath})=\int_{\partial\Omega}g(\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})\theta_{j}(\mathbf{x})ds_{\mathbf{x}},

for i,j=1,,Mi,j=1,\ldots,M.

The vector-valued function of nodal values 𝐮(t,𝜶):[0,T)×𝒜M\mathbf{u}(t,\mbox{\boldmath$\alpha$\unboldmath}):[0,T)\times\mathcal{A}\to\mathbb{R}^{M} solves

M𝐮t+(K+Q(𝜶))𝐮=𝐠(𝜶),\mathrm{M}\mathbf{u}_{t}+\left(\mathrm{K}+\mathrm{Q}(\mbox{\boldmath$\alpha$\unboldmath})\right)\mathbf{u}=\mathbf{g}(\mbox{\boldmath$\alpha$\unboldmath}), (73)

i.e., it satisfies the dynamical system of the form (1) with

F(t,𝐮,𝜶)=M1(K+Q(𝜶))𝐮+M1𝐠(𝜶)F(t,\mathbf{u},\mbox{\boldmath$\alpha$\unboldmath})=-\mathrm{M}^{-1}\left(\mathrm{K}+\mathrm{Q}(\mbox{\boldmath$\alpha$\unboldmath})\right)\mathbf{u}+\mathrm{M}^{-1}\mathbf{g}(\mbox{\boldmath$\alpha$\unboldmath})

and the initial condition 𝐮(0,𝜶)=𝐮0=𝟎M\mathbf{u}(0,\mbox{\boldmath$\alpha$\unboldmath})=\mathbf{u}_{0}=\boldsymbol{0}\in\mathbb{R}^{M} corresponding to zero initial temperature condition for ww.

We compute the snapshots ϕk=𝐮(tk,𝜶)\boldsymbol{\phi}_{k}=\mathbf{u}(t_{k},\mbox{\boldmath$\alpha$\unboldmath}) by time-stepping (73) at tk=0.2kt_{k}=0.2k, k=1,2,,Nk=1,2,\ldots,N, with N=100N=100 time steps and T=20T=20 using Crank-Nicolson scheme. Setting Θ(𝐱)=[θ1(𝐱),,θM(𝐱)]\Theta(\mathbf{x})=[\theta_{1}(\mathbf{x}),\ldots,\theta_{M}(\mathbf{x})] allows to express the solution w(t,𝐱,𝜶)w(t,\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath}) of (69)–(72) as w(t,𝐱,𝜶)=Θ(𝐱)𝐮(t,𝜶)w(t,\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})=\Theta(\mathbf{x})\mathbf{u}(t,\mbox{\boldmath$\alpha$\unboldmath}), hence the solution snapshots are

w(tk,𝐱,𝜶)=Θ(𝐱)𝐮(tk,𝜶).w(t_{k},\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})=\Theta(\mathbf{x})\mathbf{u}(t_{k},\mbox{\boldmath$\alpha$\unboldmath}). (74)

The setting is illustrated in Figure 1, where we display the domain Ω\Omega along with solution w(T,𝐱,𝜶)w(T,\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath}) corresponding to parameter values 𝜶=(0.5,0,0,0.9)T\mbox{\boldmath$\alpha$\unboldmath}=(0.5,0,0,0.9)^{T}.

Refer to caption
Fig. 1: Domain Ω\Omega and the solution w(T,𝐱,𝜶)w(T,\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath}) of the heat equation (69)–(72) corresponding to 𝜶=(0.5,0,0,0.9)T\mbox{\boldmath$\alpha$\unboldmath}=(0.5,0,0,0.9)^{T}.

For an arbitrary but fixed 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A} let Z=[𝐳1,,𝐳n]M×n\mathrm{Z}=[\mathbf{z}_{1},\ldots,\mathbf{z}_{n}]\in\mathbb{R}^{M\times n} be a matrix with columns being vectors constituting the reduced basis, i.e. Z=UUc\mathrm{Z}=\mathrm{U}\mathrm{U}_{c} for TROM. Then, the projection ROM of (73) is

M~𝐮~t+(K~+Q~(𝜶))𝐮~=𝐠~(𝜶),\widetilde{\mathrm{M}}\widetilde{\mathbf{u}}_{t}+\left(\widetilde{\mathrm{K}}+\widetilde{\mathrm{Q}}(\mbox{\boldmath$\alpha$\unboldmath})\right)\widetilde{\mathbf{u}}=\widetilde{\mathbf{g}}(\mbox{\boldmath$\alpha$\unboldmath}), (75)

where

M~\displaystyle\widetilde{\mathrm{M}} =ZTMZn×n,K~=ZTKZn×n,\displaystyle=\mathrm{Z}^{T}\mathrm{M}\mathrm{Z}\in\mathbb{R}^{n\times n},\quad\widetilde{\mathrm{K}}=\mathrm{Z}^{T}\mathrm{K}\mathrm{Z}\in\mathbb{R}^{n\times n},
Q~(𝜶)\displaystyle\widetilde{\mathrm{Q}}(\mbox{\boldmath$\alpha$\unboldmath}) =ZTQ(𝜶)Zn×n,𝐠~(𝜶)=ZT𝐠(𝜶)n,\displaystyle=\mathrm{Z}^{T}\mathrm{Q}(\mbox{\boldmath$\alpha$\unboldmath})\mathrm{Z}\in\mathbb{R}^{n\times n},\quad\widetilde{\mathbf{g}}(\mbox{\boldmath$\alpha$\unboldmath})=\mathrm{Z}^{T}\mathbf{g}(\mbox{\boldmath$\alpha$\unboldmath})\in\mathbb{R}^{n},

and the initial condition is 𝐮~(0,𝜶)=ZT𝐮0=𝟎n\widetilde{\mathbf{u}}(0,\mbox{\boldmath$\alpha$\unboldmath})=\mathrm{Z}^{T}\mathbf{u}_{0}=\boldsymbol{0}\in\mathbb{R}^{n}. As discussed in Section 3.10, the evaluation of (75) can be effectively split between the offline and online stages. Solving (75) for 𝐮~(t,𝜶)\widetilde{\mathbf{u}}(t,\mbox{\boldmath$\alpha$\unboldmath}) allows to recover the approximate solution at times tkt_{k} as

w~(tk,𝐱,𝜶)=Θ(𝐱)Z𝐮~(tk,𝜶)w(tk,𝐱,𝜶).\widetilde{w}(t_{k},\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})=\Theta(\mathbf{x})\;\mathrm{Z}\;\widetilde{\mathbf{u}}(t_{k},\mbox{\boldmath$\alpha$\unboldmath})\approx w(t_{k},\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath}). (76)

5.2.1 In-sample prediction and compression study

We begin TROM assessment with in-sample prediction and compression study for the linear parabolic system described Section 5.2. To measure TROM predictive power and to compare it to that of POD-ROM, we sample 𝒜\mathcal{A} uniformly in each direction with n1×n2×n3×n4=9×5×5×5n_{1}\times n_{2}\times n_{3}\times n_{4}=9\times 5\times 5\times 5 samples, for a total of K=1,125K=1,125 samples in the set 𝒜^={𝜶^1,,𝜶^K}{\widehat{\mathcal{A}}}=\{\widehat{\mbox{\boldmath$\alpha$\unboldmath}}_{1},\dots,\widehat{\mbox{\boldmath$\alpha$\unboldmath}}_{K}\}. For each of the three TROMs and for POD-ROM we compute the following in-sample prediction error

EL2(𝒜^)=(1MNKj=1K(IZZT)Φe(𝜶^j)F2)1/2,E_{L^{2}({\widehat{\mathcal{A}}})}=\left(\frac{1}{MNK}\sum_{j=1}^{K}\left\|(\mathrm{I}-\mathrm{Z}\mathrm{Z}^{T})\Phi_{e}(\widehat{\mbox{\boldmath$\alpha$\unboldmath}}_{j})\right\|_{F}^{2}\right)^{1/2}, (77)

to quantify the ability of the CP-TROM, HOSVD-TROM, TT-TROM local bases and POD-ROM basis to represent original snapshots for in-sample parameter values. Note that EL2(𝒜^)2E_{L^{2}({\widehat{\mathcal{A}}})}^{2} is the quantity from (14) averaged over 𝒜^{\widehat{\mathcal{A}}} and scaled by (MN)1(MN)^{-1}.

Table 2: Cartesian grid-based sampling in-sample prediction and compression study results reporting prediction errors EL2(𝒜^)E_{L^{2}({\widehat{\mathcal{A}}})} for HOSVD,TT,POD{\rm HOSVD},~{}{\rm TT},~{}{\rm POD}, the number of compressed tensors elements transmitted to the online stage, and the corresponding compression factors CF. 𝚽~\widetilde{\boldsymbol{\Phi}} ranks comprise: Tucker ranks for HOSVD-TROM in format [M~,n~1,n~2,n~3,n~4,N~][\widetilde{M},\widetilde{n}_{1},\widetilde{n}_{2},\widetilde{n}_{3},\widetilde{n}_{4},\widetilde{N}], and compression ranks for TT-TROM in format [r~1,r~2,r~3,r~4,r~5][\widetilde{r}_{1},\widetilde{r}_{2},\widetilde{r}_{3},\widetilde{r}_{4},\widetilde{r}_{5}].
ε~\widetilde{\varepsilon} 1e-4 1e-5 1e-6 1e-7 1e-9
nn 12 16 19 23 30
HOSVD 3.45e-05 3.85e-06 2.78e-07 5.61e-08
EL2(𝒜^)E_{L^{2}({\widehat{\mathcal{A}}})} TT 6.15e-05 5.58e-06 5.21e-07 5.03e-08 5.41e-10
POD 1.67e-03 9.06e-04 5.84e-04 3.05e-04 7.76e-05
HOSVD [34,4,2,[34,4,2, [46,5,2,[46,5,2, [57,6,2,[57,6,2, [66,7,2,[66,7,2,
𝚽~\widetilde{\boldsymbol{\Phi}} ranks 2,2,12]2,2,12] 2,2,16]2,2,16] 2,2,20]2,2,20] 2,2,23]2,2,23]
TT [34,35,30,[34,35,30, [46,48,41,[46,48,41, [57,61,51,[57,61,51, [69,74,60,[69,74,60, [99,97,80,[99,97,80,
21,12]21,12] 29,16]29,16] 36,19]36,19] 43,23]43,23] 57,30]57,30]
HOSVD 13122 29515 54804 85101
#online(𝚽~)\#\mbox{online}(\widetilde{\boldsymbol{\Phi}}) TT 20382 37993 53877 86022 156607
HOSVD 3.05e+4 1.36e+4 7.31e+3 4.71e+3
CF TT 1.97e+4 1.05e+4 7.42e+3 4.66e+3 2.56e+3
Table 3: General parameter sampling in-sample prediction and compression study results reporting prediction errors EL2(𝒜^)E_{L^{2}({\widehat{\mathcal{A}}})} for HOSVD,TT,POD{\rm HOSVD},~{}{\rm TT},~{}{\rm POD}, the number of compressed tensors elements transmitted to the online stage, and the corresponding compression factors CF. 𝚽~\widetilde{\boldsymbol{\Phi}} ranks comprise: Tucker ranks for HOSVD-TROM in format [M~,N~][\widetilde{M},\widetilde{N}], and compression ranks for TT-TROM in format [r~1,r~2][\widetilde{r}_{1},\widetilde{r}_{2}].
ε~\widetilde{\varepsilon} 1e-4 1e-5 1e-6 1e-7 1e-9
nn 12 16 19 23 30
HOSVD 5.69e-05 5.29e-06 4.66e-07 7.93e-08
EL2(𝒜^)E_{L^{2}({\widehat{\mathcal{A}}})} TT 5.63e-05 5.95e-06 5.72e-07 5.56e-08 5.43e-09
POD 2.05e-03 1.12e-03 5.84e-04 3.05e-04 8.46e-05
𝚽~\widetilde{\boldsymbol{\Phi}} ranks HOSVD [33,11][33,11] [45,16][45,16] [56,19][56,19] [65,23][65,23]
TT [32,11][32,11] [43,15][43,15] [55,19][55,19] [66,23][66,23] [95,29][95,29]
HOSVD 11904 18450 26268 36680
#online(𝚽~)\#\mbox{online}(\widetilde{\boldsymbol{\Phi}}) TT 396011 725640 1175644 1633522 3099404
HOSVD 3.37e+4 2.17e+4 1.53e+4 1.09e+4
CF TT 1.01e+3 5.52e+2 3.41e+2 2.45e+2 1.29e+2

We report in Tables 2 and 3 the in-sample prediction errors and compression factors CF defined in (47) for Cartesian grid-based and general samplings, respectively. For general sampling we organize the sampling parameters from the Cartesian grid in 1D array leading to snapshot tensors or order 3. Experimenting with the same number of randomly sampled parameters showed very similar compression rates and in-sample prediction errors and so those are not reported here. The results are reported in Tables 2 and 3 for a number of decreasing values of ε~\widetilde{\varepsilon} and correspondingly increasing nn, such that nmin(N~,r~D+1)n\leq\min(\widetilde{N},\widetilde{r}_{D+1}), where N~\widetilde{N} and r~D+1\widetilde{r}_{D+1} are the last Tucker and compression ranks, respectively, for HOSVD- and TT-TROM. Available compression algorithm failed to deliver the accuracy of 1e-9 for HOSVD, so we report only TT statistics for this extreme value of ε~\widetilde{\varepsilon}. Note that we leave EL2(𝒜^)E_{L^{2}({\widehat{\mathcal{A}}})} for CP-ROM out since there is no direct way to control its relative error ε~\widetilde{\varepsilon}, as discussed at the end of Section 3.5. Instead, compression factors and canonical ranks for CP-TROM are illustrated in Figure 2.

We observe in Tables 2 and 3 that both HOSVD- and TT-TROM outperform POD-ROM in terms of prediction error up to four orders of magnitude for larger nn and correspondingly small ε~\widetilde{\varepsilon}. This result is consistent across both Cartesian grid-based and general parameter samplings. The #online(𝚽~)\#\mbox{online}(\widetilde{\boldsymbol{\Phi}}) values tell us that for Cartesian based sampling, HOSVD and TT are comparable in terms of memory and data transmission requirements with HOSVD doing somewhat better for lower representation accuracy for the snapshot tensor of order 6. Compression achieved varies in ε~\widetilde{\varepsilon} (as should expected) and gives more than 3 orders of saving even for the finest available representation accuracy. If the Cartesian structure of 𝒜^\widehat{\mathcal{A}} is abandoned and snapshots are organized in tensors of order 3, then HOSVD format has a clear advantage over TT in terms of compression achieved; see #online(𝚽~)\#\mbox{online}(\widetilde{\boldsymbol{\Phi}}) and CF statistics in Table 3.

Refer to caption
Refer to caption
Fig. 2: CP-TROM compression factor CF and canonical rank RR as functions of ε~\widetilde{\varepsilon}. Left: Cartesian grid-based sampling; Right: general sampling.

For CP-TROM we display in Figure 2 compression factors CF and canonical ranks RR for a number of values of ε~\widetilde{\varepsilon}. For comparison we also show on the same plots CF for HOSVD, which shows that HOSVD-TROM is on par with CP-TROM if Cartesian sampling allows to organize snapshots in a higher order tensor. However, we were able to compute 𝚽~\widetilde{\boldsymbol{\Phi}} in CP compressed format only up to moderate values of ε~\widetilde{\varepsilon}, since the corresponding CP rank was growing fast as ε~\widetilde{\varepsilon} decreases (see the left plot in Figure 2). Smaller ε~\widetilde{\varepsilon} become feasible with CP if the snapshots are organized in tensors of order 3, but in this case HOSVD-TROM achieves much better compression than CP-TROM (see the right plot in Figure 2). We conclude that for this example with a relatively small number of parameters (D=4D=4) HOSVD-TROM appears to be the best performing TROM. We finally note that for the same compression accuracy ε~\widetilde{\varepsilon} (if it was achieved) CP-TROM demonstrated very similar in-sample prediction error as HOSVD- and TT-ROMs. This observation largely carries over to out-of-sample representation studied next.

5.2.2 Out-of-sample prediction study

To quantify the ability of the CP-, HOSVD- and TT-TROM local bases to represent the solution of (69) for arbitrary out-of-sample parameter values, we use EL2(𝒜)E_{L^{2}(\mathcal{A})} which is defined as in (77) but with 𝜶\alpha (in place of 𝜶^\widehat{\alpha}) running through a large number of random points from 𝒜\mathcal{A}. We also use

EL(𝒜)=sup𝜶𝒜(1MN(IZZT)Φe(𝜶)F2)1/2,E_{L^{\infty}(\mathcal{A})}=\sup_{\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A}}\left(\frac{1}{MN}\left\|(\mathrm{I}-\mathrm{Z}\mathrm{Z}^{T})\Phi_{e}({\mbox{\boldmath$\alpha$\unboldmath}})\right\|_{F}^{2}\right)^{1/2},

for the maximum of representation error over the parameter domain. An estimate of EL(𝒜)E_{L^{\infty}(\mathcal{A})} is given by Theorem 1. Regarding constants appearing in (65) we note that for our choice of the uniform grid in 𝒜\mathcal{A} and p=2,3p=2,3 one computes Ce=1C_{e}=1, Ca=18C_{a}=\frac{1}{8} (p=2p=2) and Ca=148C_{a}=\frac{1}{48} (p=2p=2), while constant C𝐮C_{\mathbf{u}} is hard to evaluate. Experimentally we found that the grid in the forth parameter should be finer than for the first three parameters to balance the observed error suggesting that the solution is less smooth as a function of α4\alpha_{4}. To study the error dependence on the parameter mesh size δ\delta, we reduce the number of parameters to two, letting α1=α2=α3\alpha_{1}=\alpha_{2}=\alpha_{3} in (71).

TROM and POD prediction error vs nn TROM prediction error vs δ\delta TROM prediction error vs ε~\widetilde{\varepsilon}
Refer to caption Refer to caption Refer to caption
Fig. 3: Out-of-sample prediction error of TROM versus ROM basis dimension nn (left plot), parameter mesh size δ\delta (middle plot) and tensor compression accuracy ε~\widetilde{\varepsilon} (right plot).

In Figure 3 we plot EL(𝒜)E_{L^{\infty}(\mathcal{A})} and EL2(𝒜)E_{L^{2}(\mathcal{A})} versus ROM basis dimension nn, parameter mesh size δ\delta and tensor compression accuracy ε~\widetilde{\varepsilon}. The results were computed using 100 randomly distributed parameters from 𝒜\mathcal{A} to evaluate the error quantities for HOSVD-TROM. For TT-TROM and CP-TROM the error dependence on nn, δ\delta and ε~\widetilde{\varepsilon} was virtually the same and are omitted. Variants of TROM of course may differ by complexity. While the cost of offline phase of computing 𝚽~\widetilde{\boldsymbol{\Phi}} varies significantly depending on the format, the online distribution of costs was persistent for all three formats showing between 30% and 40% of online time spent on finding UcU_{c} (coordinates of local basis), less then 5% on projection to 𝜶\alpha-specific coordinates, and about 60% of time on the integration of the projected system (this last step is common with standard POD approach).

The left plot in Figure 3 shows the EL(𝒜)E_{L^{\infty}(\mathcal{A})} error for different values of nn and compares it to the error of POD-ROM. We use ε~\widetilde{\varepsilon}=1e-7 and 65×3365\times 33 parameter grid. Such fine compression accuracy and grid allows us to isolate the effect of the second term on the right hand side of (65). Indeed, we see that the error curve follows closely the graph of “SVD remainder” (the maximum over all 𝜶\alpha of the second term on the right-hand side of (65)) until the interpolation error starts dominating for larger nn. As can be expected, the interpolation error for p=3p=3 starts affecting EL(𝒜)E_{L^{\infty}(\mathcal{A})} for larger nn than the interpolation error for p=2p=2.

The middle plot in Figure 3 demonstrates that both EL(𝒜)E_{L^{\infty}(\mathcal{A})} and EL2(𝒜)E_{L^{2}(\mathcal{A})} for TROM decrease as O(δ2)O(\delta^{2}) for p=2p=2 (computed with ε~\widetilde{\varepsilon}=1e-7 and n=N~n=\widetilde{N}) just as predicted by (65). Likewise, the right plot in Figure 3 gives evidence for O(ε~)O(\widetilde{\varepsilon}) decrease of EL(𝒜)E_{L^{\infty}(\mathcal{A})} and EL2(𝒜)E_{L^{2}(\mathcal{A})} (computed with 65×3365\times 33 parameter grid and n=N~n=\widetilde{N}) in accordance to (65) as long the first term on the right hand side of (65) dominates.

5.2.3 Out-of-sample TROMs vs POD-ROM performance: heat equation

Performance of TROM and POD-ROM may vary for different out-of-sample parameter values. Therefore, assessment of TROM vs POD-ROM performance in this section is conducted in a statistical setting. The quantities of interest are computed for Nr1N_{r}\gg 1 out-of -sample realizations of 𝜶(r)𝒜\mbox{\boldmath$\alpha$\unboldmath}^{(r)}\in\mathcal{A}, r=1,2,,Nrr=1,2,\ldots,N_{r}, where we use Nr=200N_{r}=200 for the numerical studies below. Realizations 𝜶(r)=(α1(r),α2(r),α3(r),α4(r))T\mbox{\boldmath$\alpha$\unboldmath}^{(r)}=\left(\alpha_{1}^{(r)},\alpha_{2}^{(r)},\alpha_{3}^{(r)},\alpha_{4}^{(r)}\right)^{T} are drawn at random from 𝒜=[0.01,0.5]×[0,0.9]3\mathcal{A}=[0.01,0.5]\times[0,0.9]^{3} with each αi\alpha_{i} distributed uniformly on [αimin,αimax][\alpha_{i}^{\min},\alpha_{i}^{\max}], i=1,2,3,4i=1,2,3,4.

For statistical tests we use the following quantities to measure performance of TROM. First, we introduce the relative L(0,T,L2(Ω))L^{\infty}(0,T,L^{2}(\Omega)) ROM solution error

RX(𝜶)=maxk=1,,Nw~(tk,𝐱,𝜶)w(tk,𝐱,𝜶)L2(Ω)maxk=1,,Nw(tk,𝐱,𝜶)L2(Ω)supt[0,T]w~(t,𝐱,𝜶)w(t,𝐱,𝜶)L2(Ω)supt[0,T]w(t,𝐱,𝜶)L2(Ω),\begin{split}R_{X}(\mbox{\boldmath$\alpha$\unboldmath})&=\frac{\max\limits_{k=1,\ldots,N}\left\|\widetilde{w}(t_{k},\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})-w(t_{k},\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})\right\|_{L^{2}(\Omega)}}{\max\limits_{k=1,\ldots,N}\left\|w(t_{k},\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})\right\|_{L^{2}(\Omega)}}\\ &\approx\frac{\sup\limits_{t\in[0,T]}\left\|\widetilde{w}(t,\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})-w(t,\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})\right\|_{L^{2}(\Omega)}}{\sup\limits_{t\in[0,T]}\left\|w(t,\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})\right\|_{L^{2}(\Omega)}},\end{split} (78)

which we compute for each realization 𝜶(r)\mbox{\boldmath$\alpha$\unboldmath}^{(r)}, r=1,2,,Nrr=1,2,\ldots,N_{r}, for both POD-ROM and each of the three TROMs with X{\in\{POD, CP, HOSVD, TT}\}. The true and reduced order snapshots for (78) are computed as in (74) and (76), respectively. We report the mean, minimum, and standard deviation of the three relative gain distributions

GX(r)=RPOD(𝜶(r))RX(𝜶(r)),r=1,2,,Nr,G_{\text{X}}^{(r)}=\frac{R_{\text{POD}}\big{(}\mbox{\boldmath$\alpha$\unboldmath}^{(r)}\big{)}}{R_{\text{X}}\big{(}\mbox{\boldmath$\alpha$\unboldmath}^{(r)}\big{)}},\quad r=1,2,\ldots,N_{r}, (79)

for X{\in\{CP, HOSVD, TT}\}, which quantify the error decrease of CP-TROM, HOSVD-TROM and TT-TROM, respectively, relative to POD-ROM. We study the dependency of (79) with respect to KK, the number of sampled parameter values in 𝒜\mathcal{A}, and nn, the dimension of the reduced space. The results are reported for both Cartesian grid-based sampling and general parameter sampling.

Table 4: Statistics of relative gain (79) for various values of KK, the number of sampled parameter values in 𝒜\mathcal{A}. The study is performed with n=10n=10.
General parameter sampling Cartesian grid-based sampling
KK GXG_{X} CP HOSVD TT CP HOSVD TT
mean 24.17 24.17 24.17 24.76 25.08 25.08
135=135= min 0.49 0.49 0.49 0.56 0.56 0.56
5×335\times 3^{3} std 17.33 17.33 17.33 16.88 17.31 17.32
mean 34.60 34.60 34.61 35.21 35.52 35.51
1000=1000= min 2.80 2.80 2.80 1.72 1.72 1.72
8×538\times 5^{3} std 15.36 15.36 15.37 15.03 15.11 15.11
mean 38.14 38.15 38.15 37.80 38.80 38.80
3430=3430= min 3.81 3.81 3.81 4.20 4.45 4.43
10×7310\times 7^{3} std 14.20 14.21 14.22 12.96 13.61 13.62
Table 5: Statistics of relative gain (79) for n=10n=10 and 2020. The study is performed for K=10×7×7×7=3430K=10\times 7\times 7\times 7=3430.
General parameter sampling Cartesian grid-based sampling
nn GXG_{X} CP HOSVD TT CP HOSVD TT
mean 38.14 38.15 38.15 37.80 38.80 38.80
1010 min 3.81 3.81 3.81 4.20 4.45 4.43
std 14.20 14.21 14.22 12.96 13.61 13.62
mean 155.00 158.97 161.75 49.80 155.65 154.03
2020 min 4.59 4.54 4.55 1.51 5.26 5.23
std 513.33 530.50 557.86 39.48 551.92 541.88

We present in Table 4 the dependence of relative gain statistics on the values of KK (statistics in the table were computed setting ε~=105\widetilde{\varepsilon}=10^{-5} as the targeted accuracy of HOSVD-TROM, TT-TROM, and R=250R=250 as the targeted rank for CP-TROM). We observe that as KK increases, TROMs become both more accurate on average and more robust. The robustness is observed in both the increase of the minimum relative gain and the decrease of its standard deviation. On average, for K=3,430K=3,430, all three TROMs are almost 4040 times more accurate compared to POD-ROM. The performance difference between the TROMs themselves is basically negligible for this particular study.

The performance of TROMs in the example above is limited by the relatively small value of n=10n=10. The effect of increasing nn to 2020 while keeping K=3,430K=3,430 is shown in Table 5 (statistics in the table were computed setting ε~=107\widetilde{\varepsilon}=10^{-7} as the targeted accuracy of HOSVD-TROM, TT-TROM, and R=250R=250 as the targeted rank for CP-TROM). While the worst case scenario stays relatively unchanged, the average accuracy gain by TROMs is over two orders of magnitude. An outlier here is CP-TROM that underperforms HOSVD- and TT-TROM in case of Cartesian grid-based parameter sampling. Aside from that, performance difference of TROMs for general and Cartesian samplings is negligible. It is also interesting to see if the interpolation of approximate snapshots from the lower-rank tensor 𝚽~\widetilde{\boldsymbol{\Phi}} alone, i.e., without finding a reduced local basis and solving the projected problem, gives reasonable approximation to high-fidelity solutions for out-of-sample parameters. Such interpolation-only predicted solution for incoming 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A} is given by columns of Φ~(𝜶)\widetilde{\Phi}(\mbox{\boldmath$\alpha$\unboldmath}). Repeating the experiment with Cartesian grid-based sampling and other parameters the same as used for results in Tables 4 and 79, we find that the mean relative gain of HOSVD-tROM compared to interpolation-only approach is {2.10, 1.51, 1.13}\{2.10,\,1.51,\,1.13\} for n=10n=10, K={135,1000,3430}K=\{135,1000,3430\} and {7.64, 7.60, 7.54}\{7.64,\,7.60,\,7.54\} for n=20n=20 and same values of KK. The numbers were very close for other two TROMs. We see that TROM based on solving projected problem in general gives more accurate results then pure interpolation, especially if more vectors are included in the reduced basis. If nn is fixed, then for sufficiently fine sampling the interpolation-only approach delivers the same (or even better) accuracy.

Table 6: Statistics of relative gain (79) for POD-ROM with POD-Greedy reduced basis for n=10n=10 and 2020. The study is performed for K=10×7×7×7=3430K=10\times 7\times 7\times 7=3430.
General parameter sampling Cartesian grid-based sampling
nn GXG_{X} CP HOSVD TT CP HOSVD TT
mean 53.14 53.14 53.14 54.02 54.12 54.12
1010 min 7.01 7.01 7.01 7.43 7.42 7.42
std 18.63 18.63 18.63 18.03 18.03 18.03
mean 188.57 195.57 198.77 62.18 193.66 191.86
2020 min 7.39 7.35 7.36 1.87 8.51 8.47
std 571.50 607.44 634.94 49.07 649.81 639.61

We conclude the numerical study for the parameterized heat equation with a comparison between the three TROMs and another variant of POD-ROM often used in practice, the so-called greedy POD or POD-Greedy approach to computing the reduced basis [59]. Replacing the conventional POD-ROM computation with POD-Greedy algorithm, we perform the same out-of-sample performance study as presented in Table 5. The resulting relative gain statistics are reported in Table 6. Qualitatively, the results are very similar to those in Table 5. However, quantitatively we observe 20%20\% to 40%40\% increased relative gain for all TROMs. This is consistent with the fact that POD-Greedy reduced basis is sub-optimal compared to the conventional POD-ROM reduced basis computed from the snapshots corresponding to all parameter values in 𝒜^{\widehat{\mathcal{A}}}.

In terms of computational performance for this specific setting, POD-Greedy algorithm was found to be significantly slower in the offline stage than all three TROM approaches and the conventional POD-ROM even if one includes into the offline cost of TROM and POD-ROM the computation of KNKN snapshots ϕk(𝜶^j)\boldsymbol{\phi}_{k}(\widehat{\mbox{\boldmath$\alpha$\unboldmath}}_{j}), k=1,,Nk=1,\ldots,N, j=1,,Kj=1,\ldots,K. Indeed, the bulk of computational cost of POD-Greedy approach is in the evaluation of error estimator that has to be performed at each of its nn iterations for all 𝜶^j\widehat{\mbox{\boldmath$\alpha$\unboldmath}}_{j}, j=1,,K,j=1,\ldots,K, to determine the sample with the largest error. In turn, each evaluation of the estimator requires the computation of the residual of the chosen time-stepping scheme (e.g., Crank-Nicolson) that needs O(N)O(N) matrix-vector products of a dense M×nM\times n matrix and a vector in n\mathbb{R}^{n}. Thus, error estimator evaluations alone account for O(KNMn2)O(KNMn^{2}) operations of POD-Greedy offline stage. We note that in other situations, when computing the snapshopts is much more expensive compared to the evaluation of the error estimator (e.g., when the matrices of systems to be solved on each time step are dense), both POD-ROM and TROM may benefit from a greedy approach to parameter sampling.

5.3 Advection-diffusion PDE

In the second numerical example, for assessing performance of the three TROM approaches we are interested in a case with a higher order of parameter space compared to D=4D=4 in Section 5.2. To that end we set up a dynamical system resulting from the discretization of a linear advection-diffusion equation

wt=νΔw𝜼(𝐱,𝜶)w+f(𝐱),w_{t}=\nu\Delta w-\boldsymbol{\eta}(\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})\cdot\nabla w+f(\mathbf{x}), (80)

in a unit square domain Ω=[0,1]×[0,1]2\Omega=[0,1]\times[0,1]\subset\mathbb{R}^{2}, 𝐱=(x1,x2)TΩ\mathbf{x}=(x_{1},x_{2})^{T}\in\Omega. Here ν\nu is a constant diffusion coefficient, 𝜼:Ω×𝒜2\boldsymbol{\eta}:\Omega\times\mathcal{A}\to\mathbb{R}^{2} is the advection field and f(𝐱)f(\mathbf{x}) is a Gaussian source

f(𝐱)=12πσs2exp((x1x1s)2+(x2x2s)22σs2),f(\mathbf{x})=\frac{1}{2\pi\sigma_{s}^{2}}\exp\left(-\cfrac{(x_{1}-x_{1}^{s})^{2}+(x_{2}-x_{2}^{s})^{2}}{2\sigma_{s}^{2}}\right), (81)

where we take σs=0.05\sigma_{s}=0.05, x1s=x2s=0.25x_{1}^{s}={x_{2}^{s}}=0.25. We enforce homogeneous Neumann boundary conditions and zero initial condition

(𝐧w)|Ω=0,w(0,𝐱,𝜶)=0.\left.\left(\mathbf{n}\cdot\nabla w\right)\right|_{\partial\Omega}=0,\quad w(0,\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})=0. (82)

The model is parametrized with D=9D=9 parameters with only the advection field 𝜼\boldsymbol{\eta} depending on 𝜶9\mbox{\boldmath$\alpha$\unboldmath}\in\mathbb{R}^{9}. The advection field is given as follows

𝜼(𝐱,𝜶)=(η1(𝐱,𝜶)η2(𝐱,𝜶))=(cosα9sinα9)+1π(x2h(𝐱,𝜶)x1h(𝐱,𝜶)),\boldsymbol{\eta}(\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})=\begin{pmatrix}\eta_{1}(\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})\\ \eta_{2}(\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})\end{pmatrix}=\begin{pmatrix}\cos\alpha_{9}\\ \sin\alpha_{9}\end{pmatrix}+\frac{1}{\pi}\begin{pmatrix}\partial_{x_{2}}h(\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})\\ -\partial_{x_{1}}h(\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})\end{pmatrix}, (83)

where h(𝐱)h(\mathbf{x}) is the cosine trigonometric polynomial

h(𝐱,𝜶)=α1cos(πx1)+α2cos(πx2)+α3cos(πx1)cos(πx2)+α4cos(2πx1)+α5cos(2πx2)+α6cos(2πx1)cos(πx2)+α7cos(πx1)cos(2πx2)+α8cos(2πx1)cos(2πx2).\begin{split}h(\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})=&\;\;\;\;\alpha_{1}\cos(\pi x_{1})+\alpha_{2}\cos(\pi x_{2})+\alpha_{3}\cos(\pi x_{1})\cos(\pi x_{2})\\ &+\alpha_{4}\cos(2\pi x_{1})+\alpha_{5}\cos(2\pi x_{2})+\alpha_{6}\cos(2\pi x_{1})\cos(\pi x_{2})\\ &+\alpha_{7}\cos(\pi x_{1})\cos(2\pi x_{2})+\alpha_{8}\cos(2\pi x_{1})\cos(2\pi x_{2}).\end{split} (84)

Here α9\alpha_{9} determines the angle of the dominant advection direction, while parameters αi\alpha_{i}, i=1,,8i=1,\ldots,8, introduce perturbations into the advection field. The parameter domain is a 9D box, see Section 5.3.1 for details.

The system (80)–(82) is discretized using P2P_{2} finite elements on a grid with either M=1,893M=1,893 or M=4,797M=4,797 nodes (depending on a particular experiment) using the standard nodal basis functions {θj(𝐱)}j=1M\{\theta_{j}(\mathbf{x})\}_{j=1}^{M} that define the mass MM×M\mathrm{M}\in\mathbb{R}^{M\times M}, stiffness KM×M\mathrm{K}\in\mathbb{R}^{M\times M} and advection H(𝜶)M×M\mathrm{H}(\mbox{\boldmath$\alpha$\unboldmath})\in\mathbb{R}^{M\times M} matrices, and the source vector 𝐟M\mathbf{f}\in\mathbb{R}^{M}. The vector-valued function of nodal values 𝐮(t,𝜶):[0,T)×𝒜M\mathbf{u}(t,\mbox{\boldmath$\alpha$\unboldmath}):[0,T)\times\mathcal{A}\to\mathbb{R}^{M} solves

M𝐮t+(K+H(𝜶))𝐮=𝐟,\mathrm{M}\mathbf{u}_{t}+\left(\mathrm{K}+\mathrm{H}(\mbox{\boldmath$\alpha$\unboldmath})\right)\mathbf{u}=\mathbf{f}, (85)

i.e., it satisfies the dynamical system of the form (1) with

F(t,𝐮,𝜶)=M1(K+H(𝜶))𝐮+M1𝐟,F(t,\mathbf{u},\mbox{\boldmath$\alpha$\unboldmath})=-\mathrm{M}^{-1}\left(\mathrm{K}+\mathrm{H}(\mbox{\boldmath$\alpha$\unboldmath})\right)\mathbf{u}+\mathrm{M}^{-1}\mathbf{f}, (86)

and the initial condition 𝐮(0,𝜶)=𝟎M\mathbf{u}(0,\mbox{\boldmath$\alpha$\unboldmath})=\boldsymbol{0}\in\mathbb{R}^{M}.

Similarly to the experiments in Section 5.2, we compute the snapshots ϕk=𝐮(tk,𝜶)\boldsymbol{\phi}_{k}=\mathbf{u}(t_{k},\mbox{\boldmath$\alpha$\unboldmath}) by time-stepping (85) at tk=(1/30)kt_{k}=(1/30)k, k=1,2,,Nk=1,2,\ldots,N, with N=30N=30 time steps and T=1T=1 using Crank-Nicolson scheme. Then, the physical solution snapshots have the form (74). The setting is illustrated in Figure 4 where we display the advection field for a random realization of 𝜶𝒜\mbox{\boldmath$\alpha$\unboldmath}\in\mathcal{A} and the corresponding solution w(T,𝐱,𝜶)w(T,\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath}).

Advection field 𝜼(𝐱,𝜶)\boldsymbol{\eta}(\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath}) Solution w(T,𝐱,𝜶)w(T,\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath})
Refer to caption   Refer to caption
Fig. 4: Advection field (83) (left) and the solution w(T,𝐱,𝜶)w(T,\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath}) of (80)–(82) (right) corresponding to a random realization of 𝜶\alpha from \mathcal{B} and ν=0.01\nu=0.01.

Projection ROM of (85) is obtained similarly to that of (73) using the matrix of reduced basis vectors Z=[𝐳1,,𝐳n]M×n\mathrm{Z}=[\mathbf{z}_{1},\ldots,\mathbf{z}_{n}]\in\mathbb{R}^{M\times n}. Specifically,

M~𝐮~t+(K~+H~(𝜶))𝐮~=𝐟~,\widetilde{\mathrm{M}}\widetilde{\mathbf{u}}_{t}+\left(\widetilde{\mathrm{K}}+\widetilde{\mathrm{H}}(\mbox{\boldmath$\alpha$\unboldmath})\right)\widetilde{\mathbf{u}}=\widetilde{\mathbf{f}}, (87)

where M~\widetilde{\mathrm{M}} and K~\widetilde{\mathrm{K}} are defined as in section 5.2, whereas H~(𝜶)=ZTH(𝜶)Zn×n\widetilde{\mathrm{H}}(\mbox{\boldmath$\alpha$\unboldmath})=\mathrm{Z}^{T}\mathrm{H}(\mbox{\boldmath$\alpha$\unboldmath})\mathrm{Z}\in\mathbb{R}^{n\times n}, 𝐟~=ZT𝐟n\widetilde{\mathbf{f}}=\mathrm{Z}^{T}\mathbf{f}\in\mathbb{R}^{n}, and the initial condition is 𝐮~(0,𝜶)=𝟎n\widetilde{\mathbf{u}}(0,\mbox{\boldmath$\alpha$\unboldmath})=\boldsymbol{0}\in\mathbb{R}^{n}. Efficient evaluation of (87) was discussed in Section 3.10. Solving (87) for 𝐮~(t,𝜶)\widetilde{\mathbf{u}}(t,\mbox{\boldmath$\alpha$\unboldmath}) allows to compute the approximate solution snapshots w~(tk,𝐱,𝜶)\widetilde{w}(t_{k},\mathbf{x},\mbox{\boldmath$\alpha$\unboldmath}) exactly as in (76).

5.3.1 Out-of-sample TROM performance: advection-diffusion equation

The testing of TROMs for out-of-sample parameters for the advection-diffusion system (80)–(82) is performed similarly to that for the heat equation in Section 5.2.3. In particular, we compute the average over 𝒜\mathcal{A} of the relative L(0,T;L2(Ω))L^{\infty}(0,T;L^{2}(\Omega)) and L2(0,T;H1(Ω))L^{2}(0,T;H^{1}(\Omega)) errors for w~\widetilde{w} and use the statistical behavior of the TROM vs POD gain (79) to compare the three TROM variants to conventional POD-ROM.

First, we test TROM in the following setting. The flow behavior is balanced between advection and diffusion, with a diffusion coefficient ν=0.1\nu=0.1 and M=1893M=1893. The parameter domain is 𝒜=[0.05,0.05]8×[0.1π,0.3π]\mathcal{A}=[-0.05,0.05]^{8}\times[0.1\pi,0.3\pi] sampled at a Cartesian grid with K=38×9=59,049K=3^{8}\times 9=59,049 points to obtain the sampling set 𝒜^{\widehat{\mathcal{A}}}. We draw Nr=200N_{r}=200 out-of-sample realizations 𝜶(r)\mbox{\boldmath$\alpha$\unboldmath}^{(r)} from 𝒜\mathcal{A} with each αi(r)\alpha_{i}^{(r)}, i=1,,9i=1,\ldots,9, distributed uniformly in its corresponding interval. The averaged relative L(0,T;L2(Ω))L^{\infty}(0,T;L^{2}(\Omega)) finite element error of HOSVD-TROM is evaluated as Nr1r=1NrRHOSVD(𝜶(r))N_{r}^{-1}\sum\limits_{r=1}^{N_{r}}R_{\rm HOSVD}(\mbox{\boldmath$\alpha$\unboldmath}^{(r)}) with RHOSVD(𝜶)R_{\rm HOSVD}(\mbox{\boldmath$\alpha$\unboldmath}) defined in (78). The average relative L2(0,T;H1(Ω))L^{2}(0,T;H^{1}(\Omega)) finite element error is computed in the same way after modifying RHOSVD(𝜶)R_{\rm HOSVD}(\mbox{\boldmath$\alpha$\unboldmath}) accordingly. The TROM vs POD gain statistic was defined in (79).

For such large values of KK and therefore large 𝚽\boldsymbol{\Phi}, #(𝚽)=\#(\boldsymbol{\Phi})= 3.3534e+09, the algorithm we use for finding CP decomposition turns out to be the most memory-intensive and runs out of memory. Thus, in what follows we only report the results for HOSVD- and TT-TROM approaches.

Table 7: Tensor compression ranks, averaged relative error of the HOSVD-TROM finite element solutions, statistics of the gain (79) for ν=0.1\nu=0.1, K=38×9=59,049K=3^{8}\times 9=59,049.
n=10n=10, ε~=105\widetilde{\varepsilon}=10^{-5} n=12n=12, ε~=106\widetilde{\varepsilon}=10^{-6} n=13n=13, ε~=107\widetilde{\varepsilon}=10^{-7}
HOSVD ranks [78,3,3,3,3,3, [116,3,3,3,3,3, [153,3,3,3,3,3,
3,3,3,6,11] 3,3,3,7,12] 3,3,3,9,13]
TT ranks [77,99,117,121,121, [116,162,204,219,218, [179,251,330,364,362,
107,85,62,37,11] 186,139,91,50,12] 298,209,126,63,14]
TROM FE error
L(0,T;L2(Ω))L^{\infty}(0,T;L^{2}(\Omega)) 1.15e-03 8.36e-04 7.95e-04
L2(0,T;H1(Ω))L^{2}(0,T;H^{1}(\Omega)) 9.86e-04 9.46e-04 9.14e-04
GXG_{X} mean std min mean std min mean std min
HOSVD  9.17  5.87 3.11  11.32  10.42 1.16  10.69  9.06 1.28
TT 9.13 5.85 3.11 11.40 10.61 1.17 10.65 9.06 1.22

We present in Table 7 the averaged relative error of the HOSVD-TROM finite element solutions (for TT-TROM the errors were very close and so are skipped) and the behavior of the gain statistics when tensor compression error ε~\widetilde{\varepsilon} decreases, while simultaneously increasing nn to be slightly less or equal to Tucker rank N~\widetilde{N} for HOSVD or compression rank r~D+1\widetilde{r}_{D+1} for TT, respectively. We observe in Table 7 a relatively weak dependence of the errors and TROM vs POD gain mean on the choice of ε~\widetilde{\varepsilon} and nn, hence, we conclude that higher tensor compression error and smaller nn are more beneficial, since they correspond to higher compression factors and possible faster run times for the offline stage of TROM algorithms.

For the second example we choose an advection-dominated flow with a smaller diffusion coefficient ν=0.01\nu=0.01 and M=4,797M=4,797. The parameter domain is 𝒜=[0.01,0.01]8×[0.1π,0.5π]\mathcal{A}=[-0.01,0.01]^{8}\times[0.1\pi,0.5\pi] sampled on a Cartesian grid with K=20×28=5,120K=20\times 2^{8}=5,120 points to obtain 𝒜^{\widehat{\mathcal{A}}}. This gives the snapshot tensor with #(𝚽)=\#({\boldsymbol{\Phi}})= 1.2280e+09 entries. We draw Nr=100N_{r}=100 out-of-sample realizations 𝜶(r)\mbox{\boldmath$\alpha$\unboldmath}^{(r)} from 𝒜\mathcal{A} with each αi(r)\alpha_{i}^{(r)}, i=1,,9i=1,\ldots,9, distributed uniformly in its corresponding interval.

Table 8: Tensor compression ranks, number of elements passed to the online stage, averaged relative error of the HOSVD-TROM finite element solutions, and statistics of relative gain (79) for ν=0.01\nu=0.01, K=20×28=5,120K=20\times 2^{8}=5,120.

ε~=103\widetilde{\varepsilon}=10^{-3}

HOSVD ranks [76,2,2,2,2,2,2,2,2,11,12] #(online(𝚽~))\#(\mbox{online}(\widetilde{\boldsymbol{\Phi}}))=2568444
TT ranks [75,77,79,76,77,75,71,67,61,11] #(online(𝚽~))\#(\mbox{online}(\widetilde{\boldsymbol{\Phi}}))=100747
nn 5 8 10
TROM FE error
L(0,T;L2(Ω))L^{\infty}(0,T;L^{2}(\Omega)) 3.45e-2 7.07e-3 5.58e-3
L2(0,T;H1(Ω))L^{2}(0,T;H^{1}(\Omega)) 5.59e-2 1.10e-2 5.86e-3
GXG_{X} mean std min mean std min mean std min
HOSVD 6.95 1.10 5.41 22.56 6.97 12.02  32.66  24.70 10.00
TT 6.95 1.09 5.41 22.54 6.94 11.97  31.83  23.58 9.99

ε~=105\widetilde{\varepsilon}=10^{-5}

HOSVD ranks [184,2,2,2,2,2,2,2,2,15,18] #(online(𝚽~))\#(\mbox{online}(\widetilde{\boldsymbol{\Phi}}))=12718412
TT ranks [183,235,288,305,319,295,246,187,128,18] #(online(𝚽~))\#(\mbox{online}(\widetilde{\boldsymbol{\Phi}}))=1110964
nn 5 10 15
TROM FE error
L(0,T;L2(Ω))L^{\infty}(0,T;L^{2}(\Omega)) 3.45e-2 5.56e-3 5.51e-3
L2(0,T;H1(Ω))L^{2}(0,T;H^{1}(\Omega)) 5.59e-2 5.80e-3 5.08e-3
GXG_{X} mean std min mean std min mean std min
HOSVD 6.95 1.10 5.41 33.34 25.90 10.01  19.45  16.33 5.60
TT 6.95 1.10 5.41 33.34 25.90 10.01  19.45  16.33 5.60

ε~=107\widetilde{\varepsilon}=10^{-7}

HOSVD ranks [476,2,2,2,2,2,2,2,2,18,25] #(online(𝚽~))\#(\mbox{online}(\widetilde{\boldsymbol{\Phi}}))=54835592
TT ranks [528,618,753,858,866,725,520,341,211,25] #(online(𝚽~))\#(\mbox{online}(\widetilde{\boldsymbol{\Phi}}))=6975287
nn 5 10 15
TROM FE error
L(0,T;L2(Ω))L^{\infty}(0,T;L^{2}(\Omega)) 3.44e-2 5.56e-3 5.51e-3
L2(0,T;H1(Ω))L^{2}(0,T;H^{1}(\Omega)) 5.59e-2 5.80e-3 5.08e-3
GXG_{X} mean std min mean std min mean std min
HOSVD 6.95 1.10 5.41 33.34 25.90 10.01  19.45  16.33 5.60
TT 6.95 1.10 5.41 33.34 25.90 10.01  19.45  16.33 5.60

Table 8 shows tensor compression ranks, number of elements passed to the online stage, the averaged relative error of the HOSVD-TROM finite element solutions, the relative gain statistics for three different levels of tensor compression error ε~=103,105,107\widetilde{\varepsilon}=10^{-3},10^{-5},10^{-7} with three different reduced space dimensions nn for each case. We observe that it is possible to achieve performance that is very close to the best one with ε~\widetilde{\varepsilon} as large as 10310^{-3}, provided nn is large enough. Similarly to the results for ν=0.1\nu=0.1, we suggest that for the given problem, discretization and parameter sampling, the finite element error is dominated by the interpolation error of the TROM and using a tighter compression threshold or larger nn does not lead to more accurate ROM solutions. It seems beneficial to use low-accuracy tensor decompositions to save on both the computation and storage, while not losing much in terms of relative gain compared to more expensive options. As expected, the TT format becomes more cost-efficient for the higher parameter space dimension DD.

Overall, while the accuracy increase of HOSVD- and TT-TROM over POD-ROM is still substantial in the advection-diffusion setting with D=9D=9 parameters, it is smaller than the one for the heat equation considered in Section 5.2. This is most probably caused by larger variability of the snapshots with respect to parameter variation making the problem a good candidate for non-interpolatory TROM (not studied here).

6 Conclusions

Summarizing the findings of the paper, the tensorial projection ROM for parametric dynamical systems builds on several new ideas:
(i) To approximately represent the set of observed snapshots, it uses low-rank tensor formats, rather than a truncated SVD of the snapshot matrix. The corresponding tensor decompositions provide POD-type universal basis while preserving information about solution variation with respect to parameters.
(ii) This additional information is used to find a local (parameter-specific) ROM basis for any incoming parameter that is not necessarily from the training/sampling set.
(iii) The local basis can be represented by its coordinates in the universal low-dimensional basis allowing an effective split of the ROM evaluation between the online and offline phases.

An interpolation procedure was suggested to extract the information about parameter dependence of the solutions, and thus of the ROM spaces, from the low-rank tensor decompositions. Online stage uses fast linear algebra with complexity depending only on the compression ranks. Non-interpolatory or hybrid approaches are also possible and in fact can produce even more accurate and robust TROMs. We will study these options elsewhere. For interpolatory TROMs, Theorem 1 proves an estimate on the representation power of the local ROM bases. Numerical experiment with parameterized heat equation supported the estimate and illustrated the role of each of its terms.

Three popular compressed tensor formats were considered to represent the low-rank tensor in the TROM. Of course, other low-rank tensor decompositions can be used within the general framework of TROM. Out of the three tested, we found HOSVD to be most user-friendly and cost-efficient provided either the dimension of the parameter space is not too large or no Cartesian structure is exploited in organizing the snapshots. Otherwise, TT-TROM provides necessary tools to handle higher-dimensional parameter spaces. We also observed that the accuracy of TROMs crucially depend on nn, ε~\widetilde{\varepsilon} and parameter domain sampling, but not as much on the particular low-rank tensor format employed.

Finally, for higher-dimensional parameter spaces a grid-based sampling of the parameter domain becomes prohibitively expensive in terms of offline computation costs. Significant offline costs also incur for problems with less smooth dependence of solution on parameters, which would require a denser sampling, and for problems where each high-fidelity solve is expensive because of fine spatial or temporal resolution. We see several ways to develop TROMs addressing these challenges: (i) use a sophisticated sampling, e.g., based on a greedy strategy, and organize the snapshots in 3D tensors, (ii) to benefit from Cartesian structure and higher order tensor decompositions, apply a tensor completion method to find a low-rank representation of the snapshot tensor sampled at a few nodes of the parameter grid, and (iii) combine TROMs with compressed formats to represent high-fidelity snapshots. We leave these options for a future research.

Acknowledgments

M.O. was supported in part by the U.S. National Science Foundation under awards DMS-2011444 and DMS-1953535. This material is based upon research supported in part by the U.S. Office of Naval Research under award number N00014-21-1-2370 to A.M. The authors thank Vladimir Druskin, Traian Iliescu, and Vladimir Kazeev for their comments on the first draft of this paper.

References

  • [1] D. Amsallem and C. Farhat, Interpolation method for adapting reduced-order models and application to aeroelasticity, AIAA journal, 46 (2008), pp. 1803–1813.
  • [2] D. Amsallem, M. J. Zahr, and C. Farhat, Nonlinear model order reduction based on local reduced-order bases, International Journal for Numerical Methods in Engineering, 92 (2012), pp. 891–916.
  • [3] A. C. Antoulas, D. C. Sorensen, and S. Gugercin, A survey of model reduction methods for large-scale systems, tech. report, 2000.
  • [4] J. Ballani, D. Kressner, and M. D. Peters, Multilevel tensor approximation of pdes with random data, Stochastics and Partial Differential Equations: Analysis and Computations, 5 (2017), pp. 400–427.
  • [5] U. Baur, C. Beattie, P. Benner, and S. Gugercin, Interpolatory projection methods for parameterized model reduction, SIAM Journal on Scientific Computing, 33 (2011), pp. 2489–2518.
  • [6] J. A. Bengua, H. N. Phien, H. D. Tuan, and M. N. Do, Efficient tensor completion for color image and video recovery: Low-rank tensor train, IEEE Transactions on Image Processing, 26 (2017), pp. 2466–2479.
  • [7] P. Benner, S. Dolgov, A. Onwunta, and M. Stoll, Low-rank solvers for unsteady stokes–brinkman optimal control problem with random data, Computer Methods in Applied Mechanics and Engineering, 304 (2016), pp. 26–54.
  • [8]  , Solving optimal control problems governed by random navier-stokes equations using low-rank methods, arXiv preprint arXiv:1703.06097, (2017).
  • [9] P. Benner and L. Feng, A robust algorithm for parametric model order reduction based on implicit moment matching, in Reduced order methods for modeling and computational reduction, Springer, 2014, pp. 159–185.
  • [10] P. Benner, S. Gugercin, and K. Willcox, A survey of projection-based model reduction methods for parametric dynamical systems, SIAM review, 57 (2015), pp. 483–531.
  • [11] P. Benner, A. Onwunta, and M. Stoll, Low-rank solution of unsteady diffusion equations with stochastic coefficients, SIAM/ASA Journal on Uncertainty Quantification, 3 (2015), pp. 622–649.
  • [12] S. Brenner and L. Scott, The Mathematical Theory of Finite Element Methods, Springer, New York, second ed., 2002.
  • [13] S. L. Brunton, J. L. Proctor, and J. N. Kutz, Discovering governing equations from data by sparse identification of nonlinear dynamical systems, Proceedings of the national academy of sciences, 113 (2016), pp. 3932–3937.
  • [14] T. Bui-Thanh, K. Willcox, and O. Ghattas, Model reduction for large-scale systems with high-dimensional parametric input space, SIAM Journal on Scientific Computing, 30 (2008), pp. 3270–3288.
  • [15] K. Carlberg, C. Farhat, J. Cortial, and D. Amsallem, The gnat method for nonlinear model reduction: effective implementation and application to computational fluid dynamics and turbulent flows, Journal of Computational Physics, 242 (2013), pp. 623–647.
  • [16] J. D. Carroll and J.-J. Chang, Analysis of individual differences in multidimensional scaling via an n-way generalization of “eckart-young” decomposition, Psychometrika, 35 (1970), pp. 283–319.
  • [17] S. Chaturantabut and D. C. Sorensen, Nonlinear model reduction via discrete empirical interpolation, SIAM Journal on Scientific Computing, 32 (2010), pp. 2737–2764.
  • [18] F. Chinesta, A. Ammar, and E. Cueto, Recent advances and new challenges in the use of the proper generalized decomposition for solving multidimensional models, Archives of Computational methods in Engineering, 17 (2010), pp. 327–350.
  • [19] F. Chinesta, R. Keunings, and A. Leygue, The proper generalized decomposition for advanced numerical simulations: a primer, Springer Science & Business Media, 2013.
  • [20] A. Cohen, R. Devore, and C. Schwab, Analytic regularity and polynomial approximation of parametric and stochastic elliptic pde’s, Analysis and Applications, 9 (2011), pp. 11–47.
  • [21] L. De Lathauwer, B. De Moor, and J. Vandewalle, A multilinear singular value decomposition, SIAM journal on Matrix Analysis and Applications, 21 (2000), pp. 1253–1278.
  • [22] V. De Silva and L.-H. Lim, Tensor rank and the ill-posedness of the best low-rank approximation problem, SIAM Journal on Matrix Analysis and Applications, 30 (2008), pp. 1084–1127.
  • [23] S. V. Dolgov, V. A. Kazeev, and B. N. Khoromskij, Direct tensor-product solution of one-dimensional elliptic equations with parameter-dependent coefficients, Mathematics and computers in simulation, 145 (2018), pp. 136–155.
  • [24] M. Drohmann, B. Haasdonk, and M. Ohlberger, Reduced basis approximation for nonlinear parametrized evolution equations based on empirical operator interpolation, SIAM Journal on Scientific Computing, 34 (2012), pp. A937–A969.
  • [25] J. L. Eftang, D. J. Knezevic, and A. T. Patera, An hp certified reduced basis method for parametrized parabolic partial differential equations, Mathematical and Computer Modelling of Dynamical Systems, 17 (2011), pp. 395–422.
  • [26] J. L. Eftang, A. T. Patera, and E. M. Rønquist, An” hp” certified reduced basis method for parametrized elliptic partial differential equations, SIAM Journal on Scientific Computing, 32 (2010), pp. 3170–3200.
  • [27] M. Eigel, M. Pfeffer, and R. Schneider, Adaptive stochastic galerkin fem with hierarchical tensor representations, Numerische Mathematik, 136 (2017), pp. 765–803.
  • [28] S. Gandy, B. Recht, and I. Yamada, Tensor completion and low-n-rank tensor recovery via convex optimization, Inverse problems, 27 (2011), p. 025010.
  • [29] L. Grasedyck, D. Kressner, and C. Tobler, A literature survey of low-rank tensor approximation techniques, GAMM-Mitteilungen, 36 (2013), pp. 53–78.
  • [30] M. A. Grepl, Y. Maday, N. C. Nguyen, and A. T. Patera, Efficient reduced-basis treatment of nonaffine and nonlinear partial differential equations, ESAIM: Mathematical Modelling and Numerical Analysis, 41 (2007), pp. 575–605.
  • [31] M. A. Grepl and A. T. Patera, A posteriori error bounds for reduced-basis approximations of parametrized parabolic partial differential equations, ESAIM: Mathematical Modelling and Numerical Analysis, 39 (2005), pp. 157–181.
  • [32] M. Griebel and H. Harbrecht, Analysis of tensor approximation schemes for continuous functions, Foundations of Computational Mathematics, (2021), pp. 1–22.
  • [33] S. Gugercin and A. C. Antoulas, A survey of model reduction by balanced truncation and some new results, International Journal of Control, 77 (2004), pp. 748–766.
  • [34] W. Hackbusch, Tensor spaces and numerical tensor calculus, vol. 42, Springer, 2012.
  • [35] W. Hackbusch and B. N. Khoromskij, Tensor-product approximation to operators and functions in high dimensions, Journal of Complexity, 23 (2007), pp. 697–714.
  • [36] R. A. Harshman, Foundations of the parafac procedure: Models and conditions for an” explanatory” multimodal factor analysis, (1970).
  • [37] P. Hartman, Ordinary Differential Equations, vol. 590, John Wiley and Sons, 1964.
  • [38] J. Håstad, Tensor rank is np-complete, Journal of Algorithms, 11 (1990), pp. 644–654.
  • [39] J. S. Hesthaven, G. Rozza, and B. Stamm, Certified reduced basis methods for parametrized partial differential equations, vol. 590, Springer, 2016.
  • [40] F. L. Hitchcock, The expression of a tensor or a polyadic as a sum of products, Journal of Mathematics and Physics, 6 (1927), pp. 164–189.
  • [41] B. Huang, C. Mu, D. Goldfarb, and J. Wright, Provable low-rank tensor recovery, Optimization-Online, 4252 (2014), pp. 455–500.
  • [42] S. Kastian, D. Moser, L. Grasedyck, and S. Reese, A two-stage surrogate model for neo-hookean problems based on adaptive proper orthogonal decomposition and hierarchical tensor approximation, Computer Methods in Applied Mechanics and Engineering, 372 (2020), p. 113368.
  • [43] G. Kerschen, J.-c. Golinval, A. F. Vakakis, and L. A. Bergman, The method of proper orthogonal decomposition for dynamical characterization and order reduction of mechanical systems: an overview, Nonlinear dynamics, 41 (2005), pp. 147–169.
  • [44] B. N. Khoromskij and C. Schwab, Tensor-structured galerkin approximation of parametric and stochastic elliptic pdes, SIAM Journal on Scientific Computing, 33 (2011), pp. 364–385.
  • [45] H. A. Kiers, Towards a standardized notation and terminology in multiway analysis, Journal of Chemometrics: A Journal of the Chemometrics Society, 14 (2000), pp. 105–122.
  • [46] T. G. Kolda and B. W. Bader, Tensor decompositions and applications, SIAM review, 51 (2009), pp. 455–500.
  • [47] B. Kramer and K. E. Willcox, Nonlinear model order reduction via lifting transformations and proper orthogonal decomposition, AIAA Journal, 57 (2019), pp. 2297–2307.
  • [48] D. Kressner and C. Tobler, Low-rank tensor krylov subspace methods for parametrized linear systems, SIAM Journal on Matrix Analysis and Applications, 32 (2011), pp. 1288–1316.
  • [49] K. Lee, H. C. Elman, and B. Sousedik, A low-rank solver for the navier–stokes equations with uncertain viscosity, SIAM/ASA Journal on Uncertainty Quantification, 7 (2019), pp. 1275–1300.
  • [50] Y. Liang, H. Lee, S. Lim, W. Lin, K. Lee, and C. Wu, Proper orthogonal decomposition and its applications—part i: Theory, Journal of Sound and vibration, 252 (2002), pp. 527–544.
  • [51]  , Proper orthogonal decomposition and its applications—part i: Theory, Journal of Sound and vibration, 252 (2002), pp. 527–544.
  • [52] Y. Liang, W. Lin, H. Lee, S. Lim, K. Lee, and H. Sun, Proper orthogonal decomposition and its applications–part ii: Model reduction for mems dynamical analysis, Journal of Sound and Vibration, 256 (2002), pp. 515–532.
  • [53] J. Liu, P. Musialski, P. Wonka, and J. Ye, Tensor completion for estimating missing values in visual data, IEEE transactions on pattern analysis and machine intelligence, 35 (2012), pp. 208–220.
  • [54] J. L. Lumley, The structure of inhomogeneous turbulent flows, Atmospheric turbulence and radio wave propagation, (1967).
  • [55] A. Nouy, Low-rank tensor methods for model order reduction, arXiv preprint arXiv:1511.01555, (2015).
  • [56]  , Low-rank methods for high-dimensional approximation and model order reduction, Model reduction and approximation, P. Benner, A. Cohen, M. Ohlberger, and K. Willcox, eds., SIAM, Philadelphia, PA, (2017), pp. 171–226.
  • [57] I. Oseledets and E. Tyrtyshnikov, Tt-cross approximation for multidimensional arrays, Linear Algebra and its Applications, 432 (2010), pp. 70–88.
  • [58] I. V. Oseledets, Tensor-train decomposition, SIAM Journal on Scientific Computing, 33 (2011), pp. 2295–2317.
  • [59] A. T. Patera, G. Rozza, et al., Reduced basis approximation and a posteriori error estimation for parametrized partial differential equations, 2007.
  • [60] M. Rathinam and L. R. Petzold, A new look at proper orthogonal decomposition, SIAM Journal on Numerical Analysis, 41 (2003), pp. 1893–1925.
  • [61] C. W. Rowley, Model reduction for fluids, using balanced proper orthogonal decomposition, International Journal of Bifurcation and Chaos, 15 (2005), pp. 997–1013.
  • [62] R. Schneider and A. Uschmajew, Approximation rates for the hierarchical tensor format in periodic sobolev spaces, Journal of Complexity, 30 (2014), pp. 56–71.
  • [63] N. D. Sidiropoulos, L. De Lathauwer, X. Fu, K. Huang, E. E. Papalexakis, and C. Faloutsos, Tensor decomposition for signal processing and machine learning, IEEE Transactions on Signal Processing, 65 (2017), pp. 3551–3582.
  • [64] L. Sirovich, Turbulence and the dynamics of coherent structures. i. coherent structures, Quarterly of applied mathematics, 45 (1987), pp. 561–571.
  • [65] N. T. Son, A real time procedure for affinely dependent parametric model order reduction using interpolation on grassmann manifolds, International Journal for Numerical Methods in Engineering, 93 (2013), pp. 818–833.
  • [66] V. N. Temlyakov, Estimates for the best bilinear approximations of periodic functions, Trudy Matematicheskogo Instituta imeni VA Steklova, 181 (1988), pp. 250–267.
  • [67] L. Trefethen, Multivariate polynomial approximation in the hypercube, Proceedings of the American Mathematical Society, 145 (2017), pp. 4837–4844.
  • [68] L. R. Tucker, Some mathematical notes on three-mode factor analysis, Psychometrika, 31 (1966), pp. 279–311.
  • [69] M. Yuan and C.-H. Zhang, On tensor completion via nuclear norm minimization, Foundations of Computational Mathematics, 16 (2016), pp. 1031–1068.