This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Structure-Preserving Model Reduction for Nonlinear Power Grid Network
thanks: This work was supported in parts by National Science Foundation under Grant No. DMS-1923221

Bita Safaee and Serkan Gugercin B. Safaee is with Department of Mechanical Engineering at Virginia Tech, Blacksburg, VA 24061, (e-mail: [email protected]) S. Gugercin is with Department of Mathematics at Virginia Tech, Blacksburg, VA 24061 (e-mail: [email protected])
Abstract

We develop a structure-preserving system-theoretic model reduction framework for nonlinear power grid networks. First, via a lifting transformation, we convert the original nonlinear system with trigonometric nonlinearities to an equivalent quadratic nonlinear model. This equivalent representation allows us to employ the 2\mathcal{H}_{2}-based model reduction approach, Quadratic Iterative Rational Krylov Algorithm (Q-IRKA), as an intermediate model reduction step. Exploiting the structure of the underlying power network model, we show that the model reduction bases resulting from Q-IRKA have a special subspace structure, which allows us to effectively construct the final model reduction basis. This final basis is applied on the original nonlinear structure to yield a reduced model that preserves the physically meaningful (second-order) structure of the original model. The effectiveness of our proposed framework is illustrated via two numerical examples.

Index Terms:
Power network grids, lifting, nonlinear model reduction, structured model reduction, 2\mathcal{H}_{2}-norm.

I Introduction

Power networks are complex and large-scale systems in operation. Simulation of these high dimensional power system models are computationally expensive and demand unmanageable levels of storage in dynamic simulation or trajectory sensitivity analysis. Hence, reducing the order of these models is of a great importance especially for real time applications, stability analysis, and control design [12].

Numerous model order reduction (MOR) techniques have been applied to the linearized power networks models including balanced truncation [20, 11, 32, 36, 18], Krylov subspace (interpolatory) methods [8, 38, 39, 35], proper orthogonal decomposition (POD) [41], singular perturbation theory [27, 13], sparse representation of the system matrices [19], and clustering analysis [10, 37, 16, 23]. For a comparative study of MOR techniques in power systems, see, e.g., [7, 1, 12]. Recently, development of MOR techniques for nonlinear systems including power networks has received great attention. In [28], POD is employed to reduce the hybrid, nonlinear model of a power network for the study of cascading failures. Trajectory piece-wise linearization (TPWL) method to approximate the nonlinear term in swing models together with POD to find a reduced order model for the swing dynamics was employed in [21]. Balanced truncation based on empirical controllability and observability covariances of nonlinear power systems is proposed in [30]. The empirical Gramians and balanced realization were also used in [42, 43] to build a nonlinear power system model suitable for controller design. In [29], real-time phasor measurements are used to cluster generators based on similar behaviors using recursive spectral bipartitioning and spectral clustering. The reduced model is formed by representative generators chosen from each cluster. A model reduction approach that adaptively switches between a hybrid system with selective linearization and a linear system based on the variations of the system state or size of a disturbance was developed in [26]. A structure-preserving, aggregation-based model order reduction framework based on synchronization properties of power systems is introduced in [24]. In [40] an SVD-based approach for real-time dynamic model reduction with the ability of preserving some nonlinearity in the reduced model is presented. The recent work [33] presents the idea of transforming the swing equations into a quadratic system, lifts nonlinear swing equations to a quadratic one, and employs quadratic balanced truncation model order reduction technique. A non-intrusive data-driven modeling framework for power network dynamics using Lift and Learn [31] has been studied in [34].

In this paper, we focus on designing a structure-preserving and input-independent MOR framework for a special class of power grid networks. We obtain the reduction bases using the idea of lifting the nonlinear dynamics to quadratic models and employing 2\mathcal{H}_{2}-quasi-optimal model reduction for quadratic systems (Q-IRKA) [6]. The reduction basis for the second-order nonlinear power network grids model is obtained with the help of the reduction bases obtained by Q-IRKA and their particular subspace structure. This reduction basis is then applied on the second-order nonlinear power network grids model to form a structure-preserving reduced model.

The rest of the paper is organized as follows: Section II briefly introduces the nonlinear dynamical model of the power grid network followed by a quick recap of projection-based structure-preserving model reduction in this setting in Section III. Section IV illustrates how to lift the second-order dynamics of the power grid network to quadratic dynamics and presents an equilibrium analysis. In Section V, we describe the model reduction for quadratic systems via the 2\mathcal{H}_{2}-based model reduction, Q-IRKA, followed by the required modification for the quadratized power grid model before employing 2\mathcal{H}_{2}-based model reduction. In Section VI, we present our structured-preserving model reduction approach for power grid networks by proving and exploiting the specific structure arising in Q-IRKA for these dynamics. Section VII illustrates the feasibility of our approach via numerical examples and by comparing our approach with two other methods, followed by conclusions in Section VIII.

II power grid network model

There are three most commonly used models for describing a dynamical model of a power grid network: synchronous motor (SMSM), effective network (ENEN) and structure-preserving (SPSP). Each model is described as a network of nn coupled oscillators (generators or loads) whose dynamics at the iith node is expressed by the nonlinear second-order equation [25]

2JiωRδi¨+DiωRδi˙+j=1,jinKijsin(δiδjγij)=Bi,\displaystyle\frac{2J_{i}}{\omega_{R}}\ddot{\delta_{i}}+\frac{D_{i}}{\omega_{R}}\dot{\delta_{i}}+\sum_{\begin{subarray}{c}j=1,j\neq i\end{subarray}}^{n}K_{ij}\sin(\delta_{i}-\delta_{j}-\gamma_{ij})=B_{i}, (1)

where δi\delta_{i} is the angle of rotation for the iith oscillator; JiJ_{i} and DiD_{i} are inertia and damping constants, respectively; ωR\omega_{R} is the angular frequency for the system; Kij0K_{ij}\geq 0 is dynamical coupling between oscillator ii and jj; and γij\gamma_{ij} is the phase shift in this coupling. Constants BiB_{i}, KijK_{ij} and γij\gamma_{ij} are computed by solving the power flow equations and applying Kron reduction [25]. The three models mainly differ by the way they model loads. The EN model assumes loads as constant impedances instead of oscillators. The SP model represents loads as first-order oscillators (Ji=0J_{i}=0) while the SM model represent loads as synchronous motors expressed as second-order oscillators.

Define the state-vector δ=[δ1δ2δn]Tn\delta=[\delta_{1}~{}\delta_{2}~{}\dots~{}\delta_{n}]^{T}\in\mathbb{R}^{n}. Then, the dynamics of a network of nn coupled oscillators, given in (1) for the iith node, can be described in the system form

δ¨(t)+𝒟δ˙(t)+f(δ)\displaystyle\mathcal{M}\ddot{\delta}(t)+\mathcal{D}\dot{\delta}(t)+f(\delta) =u(t),\displaystyle=\mathcal{B}u(t), (2)
y(t)\displaystyle y(t) =𝒞δ(t),\displaystyle=\mathcal{C}\delta(t), (3)

where the matrices \mathcal{M}, 𝒟n×n\mathcal{D}{\in\mathbb{R}}^{n\times n}, and n\mathcal{B}\in\mathbb{R}^{n} are defined as

=diag(m1,,mn),𝒟=diag(d1,,dn),=[B1,B2,,Bn]T,\displaystyle\begin{array}[]{rcl}\mathcal{M}&=&\mathrm{diag}(m_{1},\dots,m_{n}),~{}~{}\mathcal{D}=\mathrm{diag}(d_{1},\dots,d_{n}),{}\\ \mathcal{B}&=&\begin{bmatrix}B_{1},&B_{2},&\ldots,&B_{n}\end{bmatrix}^{T},\end{array} (6)

where mi=2JiωRm_{i}=\frac{2J_{i}}{\omega_{R}} and di=DiωRd_{i}=\frac{D_{i}}{\omega_{R}}, u(t)=1u(t)=1 is input to the dynamics; and the nonlinear map f:nnf:\mathbb{R}^{n}\to\mathbb{R}^{n} is defined such that its iith component for i=1,2,,ni=1,2,\ldots,n is

(f(δ))i=j=1,jinKijsin(δiδjγij).\displaystyle\left(f(\delta)\right)_{i}=\sum_{\begin{subarray}{c}j=1,j\neq i\end{subarray}}^{n}K_{ij}\sin(\delta_{i}-\delta_{j}-\gamma_{ij}). (7)

The variable y(t)py(t)\in\mathbb{R}^{p} in (3) corresponds to the output (quantity of interest) of the network dynamics, described by the state-to-output matrix 𝒞p×n\mathcal{C}\in\mathbb{R}^{p\times n}.

III Structure-preserving MOR approach

Large scale power network dynamics, i.e., those with large nn, are computationally expensive to simulate and control. With the aid of the model reduction, we are able to replace these high dimensional models by lower dimensional models that can closely mimic the input-output behavior of the original high dimensional systems. To be more specific, we seek to develop a nonlinear structure-preserving model reduction approach such that the reduced model preserves the physically-meaningful second-order structure. In other words, we directly reduce the second-order dynamics (2)-(3) to obtain a reduced second-order system with dimension rnr\ll n and of the form

rδr¨(t)+𝒟rδr˙(t)+fr(δr)\displaystyle\mathcal{M}_{r}\ddot{\delta_{r}}(t)+\mathcal{D}_{r}\dot{\delta_{r}}(t)+f_{r}(\delta_{r}) =ru(t),\displaystyle=\mathcal{B}_{r}u(t), (8)
yr(t)\displaystyle y_{r}(t) =𝒞rδr(t),\displaystyle=\mathcal{C}_{r}\delta_{r}(t), (9)

where δrr\delta_{r}\in\mathbb{R}^{r} is the reduced state; r,𝒟rr×r\mathcal{M}_{r},\mathcal{D}_{r}\in\mathbb{R}^{r\times r}; rr×1\mathcal{B}_{r}\in\mathbb{R}^{r\times 1}; fr:rrf_{r}:\mathbb{R}^{r}\to\mathbb{R}^{r}; and 𝒞rp×r\mathcal{C}_{r}\in\mathbb{R}^{p\times r}. The goal of this structure-preserving model reduction process is that yr(t)y_{r}(t) approximates the original output y(t)y(t) with high fidelity.

Given the full-order dynamics (2)-(3), the most common approach for constructing the reduced model (8)-(9) is the Petrov-Galerkin projection framework: Construct two model reduction bases 𝒲,𝒱n×r\mathcal{W},\mathcal{V}\in\mathbb{R}^{n\times r} such that δ(t)𝒱δr(t)\delta(t)\approx\mathcal{V}\delta_{r}(t). Insert 𝒱δr(t)\mathcal{V}\delta_{r}(t) into (2) and enforce a Petrov-Galerkin condition on the residual to obtain the reduced matrices in (8) and (9) as

r=𝒲T𝒱,𝒟r=𝒲T𝒟𝒱,fr(δr)=𝒲Tf(𝒱δr),r=𝒲T,and𝒞r=𝒞𝒱.\displaystyle\begin{array}[]{rcl}\mathcal{M}_{r}&=&\mathcal{W}^{T}\mathcal{M}\mathcal{V},\ \mathcal{D}_{r}=\mathcal{W}^{T}\mathcal{D}\mathcal{V},\\ f_{r}(\delta_{r})&=&\mathcal{W}^{T}f(\mathcal{V}\delta_{r}),\ \mathcal{B}_{r}=\mathcal{W}^{T}\mathcal{B},\ \mbox{and}\ \mathcal{C}_{r}=\mathcal{C}\mathcal{V}.\end{array} (12)

To preserve the symmetry and positive definiteness of \mathcal{M} and 𝒟\mathcal{D} in (8), we employ a Galerkin projection by setting 𝒲=𝒱\mathcal{W}=\mathcal{V}. However, unlike the usual MOR via Galerkin projection where the subspace 𝒱\mathcal{V} (and thus 𝒲\mathcal{W}) only contains information from the dynamics equation (2) and ignores the output equation (3), our one-sided Galerkin projection will contain information from both (2) and (3) by taking advantage of the special structure arising in the network dynamics. These issues are discussed in detail in Section VI. The term fr(δr)f_{r}(\delta_{r}) in (12) is usually further approximated to resolve the lifting bottleneck [9]. Due to the page limit, we skip this detail here.

The accuracy of the structure-preserving reduced model (8) depends on the proper choice of 𝒱\mathcal{V}. For general nonlinearities, POD is usually the method of choice. However, for special classes of nonlinear systems such as bilinear and quadratic-bilinear systems, one can indeed use well-established system theoretical methods; see, e.g., [6, 5, 3]. A large class of nonlinear systems, such as those with smooth nonlinearities, e.g., exponential, trigonometric, can indeed be represented as quadratic-bilinear systems by introducing new variables (in a lifting map) [22], [14], [4], [33], [31]. Converting a nonlinear model to its equivalent quadratic form provides an opportunity to perform input-independent model reduction techniques where a system-theoretic norm is defined [5, 6]. Although this transformation is not unique, it is exact in many cases (as in here), i.e., there is no approximation error [5].

Second-order model (1) inherently contains a quadratic nonlinearity in f(δ)f(\delta) and the nonlinear system (2) can be lifted to a quadratic dynamic as we will see next in Section IV. However, as stated in section (III), we seek to obtain a reduced second-order model, not a reduced quadratic model. Therefore, by exploiting the inherent structure of the model reduction bases obtained for the quadratic model, we form the final reduction base 𝒱\mathcal{V} to construct the final structure preserving reduced model (12). This is achieved in Sections VVI.

IV Quadratic representation

As we mentioned in Section III for systems with quadratic nonlinearities, effective and in some cases optimal, systems-theoretic MOR methods already exist. Therefore, in obtaining high-fidelity reduced models, we can significantly benefit from a quadratic representation of the nonlinear swing dynamics.

Towards this goal, recall the original second-order dynamics (1). Expand the nonlinearity sin(δiδjγij)\sin(\delta_{i}-\delta_{j}-\gamma_{ij}) using the trigonometric identity and introducing s:=sins:=\sin and c:=cosc:=\cos

s(δiδjγij)\displaystyle s(\delta_{i}-\delta_{j}-\gamma_{ij}) =(s(δi)c(δj)c(δi)s(δj))c(γij)\displaystyle=(s(\delta_{i})c(\delta_{j})-c(\delta_{i})s(\delta_{j}))c(\gamma_{ij}) (13)
(c(δi)c(δj)+s(δi)s(δj))s(γij).\displaystyle-(c(\delta_{i})c(\delta_{j})+s(\delta_{i})s(\delta_{j}))s(\gamma_{ij}).

Note that equation (13) is quadratic in sin(δi)\sin(\delta_{i}) and cos(δi)\cos(\delta_{i}). This directly hints at choosing sin(δ)\sin(\delta) and cos(δ)\cos(\delta) as auxiliary variables in a new state vector to convert the nonlinear dynamics to a quadratic one. This is precisely what has been done in [33] to find a quadratic representation for (2). However, since the transformation, the so-called lifting, is not unique and we have a slightly different form in the resulting state-space form, we include its derivation here for completeness.

Lemma 1.

Given the second-order dynamics (2), define the new lifted state vector q(t)4nq(t)\in\mathbb{R}^{4n} as

q(t)\displaystyle q(t) =[δ(t)Tδ˙(t)Tsin(δ(t))Tcos(δ(t))T]T\displaystyle=\begin{bmatrix}\delta(t)^{T}&\dot{\delta}(t)^{T}&\sin(\delta(t))^{T}&\cos(\delta(t))^{T}\end{bmatrix}^{T}
=[q1Tq2Tq3Tq4T]T.\displaystyle=\begin{bmatrix}q_{1}^{T}&q_{2}^{T}&q_{3}^{T}&q_{4}^{T}\end{bmatrix}^{T}. (14)

For the new state q(t)q(t) in (1), the original dynamics (2) can be written exactly as a quadratic nonlinear system of the form

Eq˙(t)=Aq(t)+H(q(t)q(t))+Bu(t)y(t)=Cq(t),\displaystyle\begin{array}[]{rcl}E\dot{q}(t)&=&Aq(t)+H(q(t)\otimes q(t))+Bu(t)\\ y(t)&=&Cq(t),\end{array} (17)

where E,AE,A 4n×4n\in\mathbb{R}^{4n\times 4n}, B4n×1B\in\mathbb{R}^{{4n\times 1}} , Cp×4nC\in\mathbb{R}^{p\times 4n}, H4n×(4n)2H\in\mathbb{R}^{4n\times(4n)^{2}}, and \otimes denotes the Kronecker product.

Proof.

The proof is constructive, i.e., it will show how the matrices E,A,H,BE,A,H,B and CC in (17) will be constructed. Using the new state vector qq in (1), we can directly rewrite the original output y(t)=𝒞δ(t)y(t)=\mathcal{C}\delta(t) as y(t)=Cq(t)y(t)=Cq(t) as in (17) with

C=[𝒞000]p×4n.\displaystyle C=\begin{bmatrix}\mathcal{C}&0&0&0\end{bmatrix}\in\mathbb{R}^{p\times 4n}. (18)

It follows from (2) and (1) that

q˙1\displaystyle\dot{q}_{1} =q2,\displaystyle=q_{2}, (19)
q˙2\displaystyle\mathcal{M}\dot{q}_{2} =𝒟q2f(q1)+u,\displaystyle=-\mathcal{D}q_{2}-f(q_{1})+\mathcal{B}u, (20)
q˙3\displaystyle\dot{q}_{3} =q2q4,\displaystyle=q_{2}\circ q_{4}, (21)
q˙4\displaystyle\dot{q}_{4} =q2q3,\displaystyle=-q_{2}\circ q_{3}, (22)

where \circ denotes the Hadamard product.

We start by writing q3˙\dot{q_{3}} in (21) and q˙4\dot{q}_{4} in (22) as

q˙3=Φ(q2q4)andq˙4=Φ(q2q3),\displaystyle\dot{q}_{3}=\Phi(q_{2}\otimes q_{4})\quad\mbox{and}\quad\dot{q}_{4}=-\Phi(q_{2}\otimes q_{3}), (23)

where Φ=[Φ1Φ2Φn]n×n2\Phi=\begin{bmatrix}\Phi_{1}&\Phi_{2}&\dots&\Phi_{n}\end{bmatrix}\in\mathbb{R}^{n\times n^{2}} and Φkn×n,fork=1,,n,\Phi_{k}\in\mathbb{R}^{n\times n},\text{for}~{}k=1,\ldots,n, has zero entries except for the kkth diagonal element, which is 11. Next, we write the nonlinearity f(q1)f(q_{1}) in (20) as a quadratic term. Towards this goal, we use (13) in (7) to obtain

(f(q1))i=\displaystyle(f(q_{1}))_{i}= j=1,jinKij((q3)i(q4)j(q4)i(q3)j)cos(γij)\displaystyle\sum_{\begin{subarray}{c}j=1,j\neq i\end{subarray}}^{n}K_{ij}\left((q_{3})_{i}(q_{4})_{j}-(q_{4})_{i}(q_{3})_{j}\right)\cos(\gamma_{ij})
j=1,jinKij((q4)i(q4)j)sin(γij)\displaystyle-\sum_{\begin{subarray}{c}j=1,j\neq i\end{subarray}}^{n}K_{ij}((q_{4})_{i}(q_{4})_{j})\sin(\gamma_{ij}) (24)
j=1,jinKij((q3)i(q3)j)sin(γij).\displaystyle-\sum_{\begin{subarray}{c}j=1,j\neq i\end{subarray}}^{n}K_{ij}\left((q_{3})_{i}(q_{3})_{j}\right)\sin(\gamma_{ij}).

Using (24), we can write the vector f(q1)f(q_{1}) compactly as

f(q1)=(Z(q3q4)+Ψ(q4q4)+Ψ(q3q3)),\displaystyle f(q_{1})=-\left(Z(q_{3}\otimes q_{4})+\Psi(q_{4}\otimes q_{4})+\Psi(q_{3}\otimes q_{3})\right), (25)

where Z=[Z1Z2Zn]n×n2Z=\begin{bmatrix}Z_{1}&Z_{2}&\dots&Z_{n}\end{bmatrix}\in\mathbb{R}^{n\times n^{2}} consists of nn block matrices Zkn×nZ_{k}\in\mathbb{R}^{n\times n} defined as

Zk(i,j)={Kkjcos(γkj),i=j=1,,n,jkKkjcos(γkj),i=k,j=1,,n,jk0otherwise.\displaystyle Z_{k}(i,j)=\Bigg{\{}\begin{matrix}K_{kj}\cos{(\gamma_{kj})},&i=j=1,...,n,\ j\neq k\\ -K_{kj}\cos{(\gamma_{kj})},&i=k,j=1,...,n,\ j\neq k\\ 0&\text{otherwise}\end{matrix}. (26)

Similarly, we have Ψ=[Ψ1Ψ2Ψn]n×n2\Psi=\begin{bmatrix}\Psi_{1}&\Psi_{2}&\dots&\Psi_{n}\end{bmatrix}\in\mathbb{R}^{n\times n^{2}} and for the kthk^{th} block of Ψ\Psi, denoted by Ψkn×n\Psi_{k}\in\mathbb{R}^{n\times n}, we have

Ψk(i,j)={12Kkjsin(γkj),i=j=1,,n,jk12Kkjsin(γkj),i=k,j=1,,n,jk0otherwise.\displaystyle\Psi_{k}(i,j)=\Bigg{\{}\begin{matrix}-\frac{1}{2}{K_{kj}}\sin{(\gamma_{kj})},&i=j=1,...,n,\ j\neq k\\ -\frac{1}{2}{K_{kj}}\sin{(\gamma_{kj})},&i=k,j=1,...,n,\ j\neq k\\ 0&\text{otherwise}\end{matrix}. (27)

Using (23) and (25), we now rewrite (19)-(22) as

q˙1\displaystyle\dot{q}_{1} =q2,\displaystyle=q_{2}, (28)
q˙2\displaystyle\mathcal{M}\dot{q}_{2} =𝒟q2+Z(q3q4),\displaystyle=-\mathcal{D}q_{2}+Z(q_{3}\otimes q_{4}),
+Ψ(q4q4)+Ψ(q3q3)+u,\displaystyle~{}\qquad\qquad+\Psi(q_{4}\otimes q_{4})+\Psi(q_{3}\otimes q_{3})+\mathcal{B}u, (29)
q˙3\displaystyle\dot{q}_{3} =Φ(q2q4),\displaystyle=\Phi(q_{2}\otimes q_{4}), (30)
q˙4\displaystyle\dot{q}_{4} =Φ(q2q3).\displaystyle=-\Phi(q_{2}\otimes q_{3}). (31)

Note that the new formulation of the dynamics in (28)-(31) only contains linear and quadratic terms in qq, and the input mapping is linear. The output mapping CC is already established in (18). Therefore, there exist matrices E,A,H,BE,A,H,B and CC such that the original power network dynamics can be written as a quadratic nonlinear system as in (17), thus proving the desired result. However, to give a constructive proof, we continue to show how the matrices in (17) can be explicitly constructed.

It immediately follows from (28)-(31) that E4n×4nE\in\mathbb{R}^{4n\times 4n}, A4n×4nA\in\mathbb{R}^{4n\times 4n}, and B4nB\in\mathbb{R}^{4n} in (17) (corresponding to the linear parts of the dynamics) are given by

E=\displaystyle E= blkdiag(I,,I,I),\displaystyle~{}\text{blkdiag}(I,\mathcal{M},I,I), (32)
B\displaystyle\quad B =[000],andA=[0I000𝒟0000000000],\displaystyle=\begin{bmatrix}0\\ \mathcal{B}\\ 0\\ 0\end{bmatrix},\mbox{and}\quad A=\begin{bmatrix}0&I&0&0\\ 0&-\mathcal{D}&0&0\\ 0&0&0&0\\ 0&0&0&0\end{bmatrix}, (33)

where II denotes the identity matrix (of appropriate size). To show how HH is constructed, we define a revised Kronecker product as

q~q=[q1~q;q2~q;q3~q;q4~q],\displaystyle q\,\widetilde{\otimes}\,q=[q_{1}\,\widetilde{\otimes}\,q;\ q_{2}\,\widetilde{\otimes}\,q;\ q_{3}\,\widetilde{\otimes}\,q;\ q_{4}\,\widetilde{\otimes}\,q], (34)

where qi~q=[qiq1;qiq2;qiq3;qiq4]q_{i}\,\widetilde{\otimes}\,q=\begin{bmatrix}q_{i}\otimes q_{1};q_{i}\otimes q_{2};q_{i}\otimes q_{3};q_{i}\otimes q_{4}\end{bmatrix}, for i=1,,4i=1,\dots,4. Clearly, (34) is a permuted form of the regular Kronecker product. It is introduced here to make the derivation of HH easier. The quadratic terms (29)-(31) can be written as

H~(q~q)whereH~4n×(4n)2.\displaystyle\tilde{H}(q\,\widetilde{\otimes}\,q)\quad\mbox{where}\quad\tilde{H}\in\mathbb{R}^{4n\times(4n)^{2}}. (35)

Decompose H~\tilde{H} into four sub-matrices:

H~=[H~1H~2H~3H~4],\displaystyle\tilde{H}=\begin{bmatrix}\tilde{H}_{1}&\tilde{H}_{2}&\tilde{H}_{3}&\tilde{H}_{4}\end{bmatrix}, (36)

where H~k4n×4n2\tilde{H}_{k}\in\mathbb{R}^{4n\times 4n^{2}} for k=1,2,3,4k=1,2,3,4. The first submatrix H~1\tilde{H}_{1} corresponds to the first block of (34), i.e, q1~q=[q1q1;q1q2;q1q3;q1q4]q_{1}\tilde{\otimes}\,q=\begin{bmatrix}q_{1}\otimes q_{1};q_{1}\otimes q_{2};q_{1}\otimes q_{3};q_{1}\otimes q_{4}\end{bmatrix}. There is no q1qiq_{1}\otimes q_{i} term in (28)-(31). Thus, we set H~1=0\tilde{H}_{1}=0.

Similarly, the matrix H~2\tilde{H}_{2} corresponds to the second block of (34), i.e, q2~qq_{2}\tilde{\otimes}q. We can see from (28)-(31) that two terms, namely q2q3q_{2}\otimes q_{3} and q2q4q_{2}\otimes q_{4}, exist in (30) and (31), respectively. One can rewrite (30) as q˙3=Φ2(q2q4)+Φ2(q4q2)\dot{q}_{3}=\frac{\Phi}{2}(q_{2}\otimes q_{4})+\frac{\Phi}{2}(q_{4}\otimes q_{2}) and allocate the first term, i.e., Φ2(q2q4)\frac{\Phi}{2}(q_{2}\otimes q_{4}) in H~(q~q)\tilde{H}(q\,\widetilde{\otimes}\,q), to H~2\tilde{H}_{2}. More specifically and by using the MATLAB notation for row and column indices, we set the H~2(2n+1:3n,3n2+1:4n2)=Φ2\tilde{H}_{2}(2n+1:3n,3n^{2}+1:4n^{2})=\frac{\Phi}{2}. (The second term in q˙3\dot{q}_{3}, i.e.,Φ2(q4q2)\frac{\Phi}{2}(q_{4}\otimes q_{2}), will be matched by H~4\tilde{H}_{4} below). Similarly, we can write (31) as q˙4=Φ2(q2q3)Φ2(q3q2)\dot{q}_{4}=-\frac{\Phi}{2}(q_{2}\otimes q_{3})-\frac{\Phi}{2}(q_{3}\otimes q_{2}) and allocate the first term to H~2\tilde{H}_{2} by setting H~2(3n+1:4n,2n2+1:3n2)=Φ2\tilde{H}_{2}(3n+1:4n,2n^{2}+1:3n^{2})=-\frac{\Phi}{2} (the second term will be matched via H~3\tilde{H}_{3}). Thus, we set

H~2=[00000000000Φ200Φ20].\displaystyle\tilde{H}_{2}=\begin{bmatrix}0&0&0&0\\ 0&0&0&0\\ 0&0&0&\frac{\Phi}{2}\\ 0&0&\frac{-\Phi}{2}&0\end{bmatrix}. (37)

Following similar arguments, we obtain

H~3=[000000ΨZ200000Φ200],H~4=[000000Z2Ψ0Φ2000000].\displaystyle\tilde{H}_{3}=\begin{bmatrix}0&0&0&0\\ 0&0&{\Psi}&\frac{Z}{2}\\ 0&0&0&0\\ 0&\frac{-\Phi}{2}&0&0\end{bmatrix},~{}\tilde{H}_{4}=\begin{bmatrix}0&0&0&0\\ 0&0&\frac{-Z}{2}&{\Psi}\\ 0&\frac{\Phi}{2}&0&0\\ 0&0&0&0\end{bmatrix}. (38)

Since (34) is a permuted form of (qq)(q\otimes q), there exists a permutation matrix Pqn2×n2P_{q}\in\mathbb{R}^{n^{2}\times n^{2}} such that

(q~q)=PT(qq),\displaystyle(q\,\widetilde{\otimes}\,q)=P^{T}(q\otimes q), (39)

where P=blkdiag(Pq,Pq,Pq,Pq)P=\text{blkdiag}\left(P_{q},P_{q},P_{q},P_{q}\right). Then, HH in (17) is

H=H~PT.\displaystyle H=\tilde{H}P^{T}. (40)

IV-A Equilibrium analysis

So far we have represented the power network dynamics in two formats: as second-order dynamics in (2)-(3) and as quadratic dynamics in (17). In this section, we investigate how the lifting transformation (1) affects the equilibrium points.

To find the equilibrium points of (2), first we transform it to a first-order form by defining a state vector 𝒳(t)=[𝒳1(t)T𝒳2(t)T]T=[δ(t)Tδ˙(t)T]T2n\mathcal{X}(t)=[\mathcal{X}_{1}(t)^{T}~{}\mathcal{X}_{2}(t)^{T}]^{T}=[\delta(t)^{T}~{}\dot{\delta}(t)^{T}]^{T}\in\mathbb{R}^{2n} to obtain

𝒳˙(t)=[0I01𝒟]𝒳(t)+[01f(𝒳1)]+[01],\dot{\mathcal{X}}(t)=\begin{bmatrix}0&I\\ 0&-\mathcal{M}^{-1}\mathcal{D}\end{bmatrix}\mathcal{X}(t)+\begin{bmatrix}0\\ -\mathcal{M}^{-1}f(\mathcal{X}_{1})\end{bmatrix}+\begin{bmatrix}0\\ \mathcal{M}^{-1}\mathcal{B}\end{bmatrix}, (41)

using u(t)=1u(t)=1. Let 𝒳=[𝒳1T𝒳2T]T\mathcal{X}^{*}=\begin{bmatrix}{\mathcal{X}_{1}^{*}}^{T}&{\mathcal{X}_{2}^{*}}^{T}\end{bmatrix}^{T} be an equilibrium point of (41). By setting 𝒳˙(t)=0\dot{\mathcal{X}}(t)=0, we obtain that 𝒳\mathcal{X}^{*} satisfies

𝒳=[𝒳1𝒳2]=[δ0]wheref(δ)=.\displaystyle\mathcal{X}^{*}=\begin{bmatrix}\mathcal{X}_{1}^{*}\\ \mathcal{X}_{2}^{*}\end{bmatrix}=\begin{bmatrix}\delta^{*}\\ 0\end{bmatrix}~{}\mbox{where}~{}~{}f(\delta^{*})=\mathcal{B}. (42)
Lemma 2.

Let 𝒳=[δT0T]T\mathcal{X}^{*}=\begin{bmatrix}{\delta^{*}}^{T}&0^{T}\end{bmatrix}^{T} be an equilibrium point of the original dynamics (41). Then,

q=[δT0T(sin(δ))T(cos(δ))T]T\displaystyle q^{*}=\begin{bmatrix}{\delta^{*}}^{T}&{0}^{T}&{(\sin{(\delta^{*})})}^{T}&{(\cos{(\delta^{*})})}^{T}\end{bmatrix}^{T} (43)

is an equilibrium point of (17). Similarly, let

q\displaystyle q^{*} =[q1Tq2Tq3Tq4T]T\displaystyle=\begin{bmatrix}{q_{1}^{*}}^{T}&{q_{2}^{*}}^{T}&{q_{3}^{*}}^{T}&{q_{4}^{*}}^{T}\end{bmatrix}^{T} (44)

be an equilibrium point of (17). Then, q2=0q_{2}^{*}=0 and 𝒳=[q1T0T]T\mathcal{X}^{*}=\begin{bmatrix}{q_{1}^{*}}^{T}&0^{T}\end{bmatrix}^{T} is an equilibrium point of (41).

Proof.

Rewrite (41) as

𝒳˙=(𝒳),\displaystyle\dot{\mathcal{X}}=\mathcal{F}(\mathcal{X}), (45)

where :2n2n\mathcal{F}:\mathbb{R}^{2n}\to\mathbb{R}^{2n}. If 𝒳\mathcal{X}^{*} in (42) is an equilibrium of (45), then we have (𝒳)=0\mathcal{F}(\mathcal{X}^{*})=0. Now, let us write (1) as

q=𝒯(𝒳),\displaystyle q=\mathcal{T}(\mathcal{X}), (46)

where 𝒯\mathcal{T} is the quadratic lifting in (1), given by

𝒯:𝒳=(𝒳1𝒳2)q=(𝒳1𝒳2sin(𝒳1)cos(𝒳1)).\displaystyle\mathcal{T}:\mathcal{X}=\left(\begin{matrix}\mathcal{X}_{1}\\ \mathcal{X}_{2}\end{matrix}\right)\rightarrow q=\left(\begin{matrix}\mathcal{X}_{1}\\ \mathcal{X}_{2}\\ \sin{(\mathcal{X}_{1})}\\ \cos{(\mathcal{X}_{1})}\end{matrix}\right). (47)

Differentiate (46) to obtain

q˙=J𝒯𝒳˙=J𝒯(𝒳):=𝒢(q),\displaystyle\dot{q}=J_{\mathcal{T}}\dot{\mathcal{X}}=J_{\mathcal{T}}\mathcal{F}(\mathcal{X}):=\mathcal{G}(q), (48)

where the Jacobian JTJ_{{T}} is given by

J𝒯=[I00Idiag(cos(𝒳1))0diag(sin(𝒳1))0]4n×2n.\displaystyle J_{\mathcal{T}}=\begin{bmatrix}I&0\\ 0&I\\ \text{diag}(\cos{(\mathcal{X}_{1})})&0\\ -\text{diag}(\sin{(\mathcal{X}_{1})})&0\end{bmatrix}\in\mathbb{R}^{4n\times 2n}. (49)

Insert (𝒳)=0\mathcal{F}(\mathcal{X}^{*})=0 into (48) to obtain q˙=𝒢(q)=J𝒯(𝒳)=0,\dot{q}=\mathcal{G}(q^{*})=J_{\mathcal{T}}\mathcal{F}(\mathcal{X}^{*})=0, proving that (43) is an equilibrium of (17).

Now let qq^{*} in (44) be an equilibrium of (17). Then from (48), we obtain 𝒢(q)=J𝒯(𝒳)=0.\mathcal{G}(q^{*})=J_{\mathcal{T}}\mathcal{F}(\mathcal{X}^{*})=0. Since the Jacobian (49) has full column rank, we have (𝒳)=0,\mathcal{F}(\mathcal{X}^{*})=0, which by (45) yields that (42) is an equilibrium of (41). The fact that q2=0q_{2}^{*}=0 follows from the second row block of (𝒳)=0\mathcal{F}(\mathcal{X}^{*})=0. ∎

V Model reduction for quadratic systems

Now that we have represented the original nonlinear swing dynamics as an equivalent quadratic dynamics, we will investigate the MOR problem for those systems in more details.

Consider the quadratic dynamical system

Σ:={Ex˙(t)=Ax(t)+H(x(t)x(t))+Bu(t)y(t)=Cx(t),x(0)=0,\Sigma:=\begin{cases}$${E}\dot{x}(t)={A}x(t)+H(x(t)\otimes x(t))+{B}{u}(t)$$\\ $$~{}~{}y(t)=Cx(t),~{}~{}~{}x(0)=0$$\end{cases}, (50)

where E,AN×NE,A\in\mathbb{R}^{N\times N}, HN×N2H\in\mathbb{R}^{N\times N^{2}}, BN×lB\in\mathbb{R}^{N\times l}, and Cp×NC\in\mathbb{R}^{p\times N}. One can view HH as the mode-1 matricization of a third-order tensor N×N×N\mathcal{H}\in\mathbb{R}^{N\times N\times N}. In other words, let iN×N\mathcal{H}_{i}\in\mathbb{R}^{N\times N} for i=1,,Ni=1,\dots,N be the frontal slices of \mathcal{H}. Then,

H=(1)\displaystyle H={\mathcal{H}^{(1)}} =[12N],\displaystyle=\begin{bmatrix}\mathcal{H}_{1}&\mathcal{H}_{2}&\dots&\mathcal{H}_{N}\end{bmatrix}, (51)

where the notation (1){\mathcal{H}^{(1)}} denotes the mode-1 matricization.

For the cases where the state-space dimension NN is large, the goal is to find a reduced quadratic system

Σr:={Erxr˙(t)=Arxr(t)+Hr(xr(t)xr(t))+Bru(t)yr(t)=Crxr(t),xr(0)=0,\Sigma_{r}:=\begin{cases}$${E}_{r}\dot{{x}_{r}}(t)={{A}_{r}}{x}_{r}(t)+{H}_{r}({x}_{r}(t)\otimes{x}_{r}(t))+{{B}}_{r}{u}(t)$$\\ $$~{}~{}~{}{y}_{r}(t)={C}_{r}{x}_{r}(t),~{}~{}~{}{x}_{r}(0)=0$$\end{cases}, (52)

where Er,Arrq×rq{E}_{r},{{A}}_{r}\in\mathbb{R}^{r_{q}\times r_{q}} Hrrq×rq2{H}_{r}\in\mathbb{R}^{r_{q}\times r_{q}^{2}}, Crp×rq{C}_{r}\in\mathbb{R}^{p\times r_{q}} and Brrq×l{{B}}_{r}\in\mathbb{R}^{r_{q}\times l} with rqNr_{q}\ll N such that y(t)yr(t)y(t)\approx y_{r}(t) for a wide range of inputs. We construct the reduce system (52) via projection: Construct two model reduction bases VV, WN×rqW\in\mathbb{R}^{N\times r_{q}} such that the reduced matrices in (52) are given by

Er=WTEV,Ar=WTAV,Hr=WTH(VV),\displaystyle{E}_{r}=W^{T}EV,~{}{{A}}_{r}=W^{T}{{A}}V,~{}{H}_{r}=W^{T}H(V\otimes V),
Br=WTB,Cr=CV.\displaystyle{{B}}_{r}=W^{T}{B},~{}C_{r}=CV. (53)

The quality of the reduced system (52) depends on the choice of the model reduction bases VV and WW. There are various model reduction techniques to determine theses bases including trajectory-based method such as POD (see, e.g., [15]); and input-independent systems theoretic methods for quadratic systems such as balanced truncation [5], [33], interpolatory projections [4], [14] and 2\mathcal{H}_{2}-quasi-optimal model reduction [6]. In this paper, we employ the 2\mathcal{H}_{2}-based approach in [6].

V-A 2\mathcal{H}_{2}-based model reduction of quadratic nonlinearities

Consider the original quadratic dynamics (50) with an asymptotically stable matrix E1AE^{-1}{A}. Define a truncated 2\mathcal{H}_{2} norm for (50) denoted by Σ2(τ)\|\Sigma\|_{\mathcal{H}_{2}}(\tau) using the three leading Volterra kernels as [6]:

Σ2(𝒯)2:=\displaystyle\|\Sigma\|^{2}_{{\mathcal{H}}_{2}^{(\mathcal{T})}}:= (54)
tr(i=1300hi(t1,,ti)hiT(t1,,ti)𝑑t1𝑑ti),\displaystyle\text{tr}\left({\sum_{i=1}^{3}\int_{0}^{\infty}\cdots\int_{0}^{\infty}h_{i}(t_{1},\ldots,t_{i})h_{i}^{T}(t_{1},\ldots,t_{i})dt_{1}\cdots dt_{i}}\right),

where h1(t1)=CeE1At1E1Bh_{1}(t_{1})=Ce^{E^{-1}At_{1}}E^{-1}B, h2(t1,t2)=0h_{2}(t_{1},t_{2})=0,

h3(t1,t2,t3)=CeE1At3E1H(eE1At2E1BeE1At1E1B)h_{3}(t_{1},t_{2},t_{3})=Ce^{E^{-1}At_{3}}E^{-1}H(e^{E^{-1}At_{2}}E^{-1}B\otimes e^{E^{-1}At_{1}}E^{-1}B)

and tr()\text{tr}(\cdot) denotes the trace of the argument (for details see, e.g., [6, 2]). Note that when H=0H=0, only the first kernel h1(t1)h_{1}(t_{1}) remains and accordingly (54) boils down to the classical 2\mathcal{H}_{2}-norm of an asymptotically stable linear system.

The goal is now to construct VV and WW such that the truncated 2\mathcal{H}_{2} error norm ΣΣr2(τ)\|\Sigma-\Sigma_{r}\|_{\mathcal{H}_{2}}(\tau) is minimized. This can be achieved by the model reduction algorithm Q-IRKA [6]. A brief sketch of (the two-sided) Q-IRKA is shown in Algorithm 1. In some settings, to retain the positive definiteness and/or symmetry of the original realization in the reduced order matrices (V), a one sided version of Q-IRKA is applied where WW is set to VV as given in Algorithm 2. The proposed approach to structure preserving model reduction of power networks as outlined in the next section will employ Q-IRKA as an intermediate step. Note that the reduced models from Q-IRKA is quadratic, do not preserve the original second-order structure (2), and thus cannot be directly used in a structure preserving setting. In the next section, we will prove that the model reduction basis VV and WW resulting from applying Q-IRKA to the quadratic form of power network have special subspace structure, which will then allow us to develop our structure-preserving model reduction algorithm.

Algorithm 1 Two-sided Q-IRKA
1:Input: EEA{A}HHB{B}CC
2:Output: ErE_{r}Ar{A}_{{r}}Hr{H}_{r}Br{{B}}_{r}CrC_{r}, and VV, WW
3:Symmetrize HH, then transform it to a N×N×NN\times N\times N tensor and determine its mode-2 matricization (2)\mathcal{H}^{(2)} (for details see, e.g., [17], [6])
4:Make an initial guess for ErE_{r}Ar{A}_{{r}}Hr{H}_{r}Br{{B}}_{r}CrC_{r}
5:while not converged do
6:     Perform the spectral decomposition of the pair ErE_{r} and Ar{A}_{{r}}, i.e., ArR=ErRΛ{A}_{{r}}R=E_{r}R\Lambda and define:
7:H^=(ErR)1Hr(RR)\hat{H}=(E_{r}R)^{-1}{H}_{r}(R\otimes R) , B^=(ErR)1Br\hat{B}=(E_{r}R)^{-1}{B}_{r},C^=CrR~{}\hat{C}=C_{r}R
8:     Compute mode-2 matricization ^(2)\hat{\mathcal{H}}^{(2)}
9:     Solve for V1V_{1} and V2V_{2}:
10:EV1ΛAV1=BB^T-EV_{1}\Lambda-{A}V_{1}={B}\hat{B}^{T}
11:EV2ΛAV2=H(V1V1)H^T-EV_{2}\Lambda-{A}V_{2}=H(V_{1}\otimes V_{1})\hat{H}^{T}
12:     Solve for W1W_{1} and W2W_{2}:
13:ETW1ΛATW1=CTC^-{E}^{T}W_{1}\Lambda-{A}^{T}W_{1}=C^{T}\hat{C}
14:ETW2ΛATW2=(2)(V1W1)(^(2))T-{E}^{T}W_{2}\Lambda-{A}^{T}W_{2}=\mathcal{H}^{(2)}(V_{1}\otimes W_{1})(\hat{\mathcal{H}}^{(2)})^{T}
15:     Compute VV1+V2V\coloneqq V_{1}+V_{2} and WW1+W2W\coloneqq W_{1}+W_{2}.
16:     Create an orthogonal basis for each VV and WW
17:     Determine the reduced matrices as (V).
18:end while
Algorithm 2 One-sided Q-IRKA
1:Input: EEA{A}HHB{B}CC
2:Output: ErE_{r}Ar{A}_{{r}}Hr{H}_{r}Br{{B}}_{r}CrC_{r}, and VV, WW
3:Apply Algorithm 1 by replacing Step 7 with
W1=V1andW2=V2W_{1}=V_{1}~{}~{}\text{and}~{}~{}W_{2}=V_{2}

V-B Modifications to perform Q-IRKA for power network dynamics: zero initial conditions and asymptotic stability

Recall the state vector q(t)q(t) in (1). Since sin(δ)\sin(\delta) and cos(δ)\cos(\delta) can not be simultaneously zero, the reformulation (17) as a quadratic model will always have a nonzero initial condition. Since the most system theoretic methods to model reduction assumes a zero initial condition, a natural approach is to consider a shifted state vector that allows the use of the reduction methods developed for zero initial condition. Following [33], we define a shifted state vector x=qq0x=q-q_{0}, such that in the new state xx, we have x(0)=0x(0)=0 and write

qq=xx+((Iq0)+(q0I))x+q0q0.\displaystyle q\otimes q=x\otimes x+\left((I\otimes q_{0})+(q_{0}\otimes I)\right)x+q_{0}\otimes q_{0}. (55)

By substituting (55) in (17) we obtain

Ex˙(t)=A~x(t)+H(x(t)x(t))+B~u~(t)y(t)=Cx(t),x(0)=0,\displaystyle\begin{array}[]{rcl}{E}\dot{x}(t)&=&\tilde{A}x(t)+H(x(t)\otimes x(t))+\tilde{B}\tilde{u}(t)\\ y(t)&=&Cx(t),~{}~{}~{}x(0)=0,\end{array} (58)
withA~\displaystyle\mbox{with}~{}~{}~{}\tilde{A} =A+H((Iq0)+(q0I)),u~=[u1]T,\displaystyle=A+H\left((I\otimes q_{0})+(q_{0}\otimes I)\right),~{}\tilde{u}=\begin{bmatrix}u&1\end{bmatrix}^{T},
B~\displaystyle\tilde{B} =[BAq0+H(q0q0)].\displaystyle=[B~{}~{}Aq_{0}+H\ (q_{0}\otimes q_{0})]. (59)

Given the dynamical system (V-B) with zero initial conditions, one can apply a wide range of available methods, as mentioned at the beginning of Section V, to construct a quadratic reduced model. However, as in the linear case, optimal-2\mathcal{H}_{2} model reduction for quadratic system requires asymptotic stability of E1AE^{-1}{A} [6]. Recall the quadratic dynamics (58) and A~\tilde{A} in (V-B). Consider zero initial conditions for δ\delta and δ˙\dot{\delta}, and thus the initial condition q0=[01×n01×n01×n11×n]T4nq_{0}=[0_{1\times n}~{}0_{1\times n}~{}0_{1\times n}~{}1_{1\times n}]^{T}\in\mathbb{R}^{4n}. Thus, A~\tilde{A} have zero eigenvalues and therefore E1A~E^{-1}\tilde{A} is not asymptotically stable [33]. To overcome this stability issue in the power network setting, we replace A~\tilde{A} in (58) by

Aμ~=A~μE,\displaystyle\tilde{A_{\mu}}=\tilde{A}-\mu E, (60)

where μ>0\mu>0 is real and small, so that all the eigenvalues of the pair Aμ~\tilde{A_{\mu}} and EE have negative real part. Therefore, as the final modification, we replace (58) by

Ex˙(t)=Aμ~x(t)+H(x(t)x(t))+B~u~(t)y(t)=Cx(t),x(0)=0.\displaystyle\begin{array}[]{rcl}{E}\dot{x}(t)&=&\tilde{A_{\mu}}x(t)+H(x(t)\otimes x(t))+\tilde{B}\tilde{u}(t)\\ y(t)&=&Cx(t),~{}~{}~{}x(0)=0.\end{array} (63)

The quadratic dynamical system in (63) has now asymptotically stable linear part and have zero initial conditions. Therefore, it has the structure to perform 2\mathcal{H}_{2} based model reduction using Q-IRKA. However, we re-emphasize that the reduced-model obtained from applying Q-IRKA to (63) is not structure-preserving; it does not inherit the second-order structure. Therefore, we employ Q-IRKA only to obtain potential reduction subspace information. In the next section where we present the proposed framework, we describe how we benefit from and how we use the subspaces resulting from Q-IRKA to obtain a reduced second-order model in (8). We also show that Q-IRKA subspaces applied to (63) has special properties due to the underlying power network structure.

VI The proposed approach for MOR of Power Networks

Our aim is to perform structure-preserving model reduction of the second-order dynamics (2)-(3) using the reduction bases 𝒱,𝒲n×r\mathcal{V},\mathcal{W}\in\mathbb{R}^{n\times r} to obtain the reduced second-order model (8)-(12). To keep the symmetry of the underlying dynamics, we perform a Galerkin projection by choosing 𝒲=𝒱\mathcal{W}=\mathcal{V}. Converting the nonlinear second-order dynamics to the quadratic form has allowed us to use (optimal) systems-theoretic techniques to apply, such as Q-IRKA. Although we cannot simply use the output of Q-IRKA as the reduced model (since it does not preserve the second-order structure), we still would like to employ the high-fidelity model reduction bases VV and WW resulting from Q-IRKA in constructing 𝒱\mathcal{V}. In this section, by analyzing and exploiting the structure of the underlying power network dynamics, we show how to efficiently construct 𝒱\mathcal{V} using VV and WW from Q-IRKA.

VI-A Structures arising in Q-IRKA in the Power Network Setting

Let V4n×rqV\in\mathbb{R}^{4n\times r_{q}} and W4n×rqW\in\mathbb{R}^{4n\times r_{q}} be the model reduction bases obtained from applying Q-IRKA to (63). Let VTn×rqV_{T}\in\mathbb{R}^{n\times r_{q}} and WTn×rqW_{T}\in\mathbb{R}^{n\times r_{q}} denote leading nn rows of VV and WW. Recall that due to (1), the leading nn rows of q(t)q(t), the state of quadratic dynamic (63), corresponds to the original (second-order) state δ(t)\delta(t). Therefore, one can consider choosing 𝒱=VT\mathcal{V}=V_{T} and 𝒲=WT\mathcal{W}=W_{T} in (12). However, in addition to this being a Petrov-Galerkin projection and thus not preserving the symmetry, we will show in the next result WTW_{T} has a particular subspace structure that one can further exploit in constructing the structured reduced model.

Theorem 1.

Let VV and WW be obtained via Q-IRKA as in Algorithm 1 to the quadratized dynamics (63) resulting with a reduced quadratic system with asymptotically stable linear part. Let WTn×rqW_{T}\in\mathbb{R}^{n\times r_{q}} denote the first nn rows of WW. Then

Range(WT)Range(𝒞T).\text{Range}(W_{T})\subseteq\text{Range}(\mathcal{C}^{T}). (64)

In addition, when prqp\leq r_{q}, we have Range(WT)=Range(𝒞T)\text{Range}(W_{T})=\text{Range}(\mathcal{C}^{T}).

Before we prove Theorem 1, we establish a symmetry property of HH that will be used in its proof.

Lemma 3.

Let HH in (63) be the mode-1 matricization of the third order tensor 4n×4n×4n\mathcal{H}\in\mathbb{R}^{4n\times 4n\times 4n}, Then, the tensor \mathcal{H} is symmetric, i.e., for every ν,ρ4n×1\nu,\rho\in\mathbb{R}^{4n\times 1}, it holds

H(νρ)=H(ρν).\displaystyle H(\nu\otimes\rho)=H(\rho\otimes\nu). (65)
Proof.

Recall (39) and (40), and rewrite H(ρν)H(\rho\otimes\nu) as

H(ρν)\displaystyle H(\rho\otimes\nu) =H~PT(ρν)=H~(ρ~ν),\displaystyle=\tilde{H}P^{T}(\rho\otimes\nu)=\tilde{H}(\rho\,\widetilde{\otimes}\,\nu),

where ρ=[ρ1Tρ2Tρ3Tρ4T]T\rho=[\rho_{1}^{T}\,\rho_{2}^{T}\,\rho_{3}^{T}\,\rho_{4}^{T}]^{T} and ν=[ν1Tν2Tν3Tν4T]T\nu=[\nu_{1}^{T}\,\nu_{2}^{T}\,\nu_{3}^{T}\,\nu_{4}^{T}]^{T} with ρi,νin×1.\rho_{i},\nu_{i}\in\mathbb{R}^{n\times 1}. Using (34) and (36) we obtain

H(ρν)=[H~2(ρ2~ν)+H~3(ρ3~ν)+H~4(ρ4~ν)].\displaystyle H(\rho\otimes\nu)=\begin{bmatrix}\tilde{H}_{2}(\rho_{2}\,\widetilde{\otimes}\,\nu)+\tilde{H}_{3}(\rho_{3}\,\widetilde{\otimes}\,\nu)+\tilde{H}_{4}(\rho_{4}\,\widetilde{\otimes}\,\nu)\end{bmatrix}. (66)

Let 1=[I000]n×4n\mathcal{I}_{1}=\begin{bmatrix}I&0&0&0\end{bmatrix}\in\mathbb{R}^{n\times 4n}. Multiply (66) from the right with 1\mathcal{I}_{1} to pick the first nn rows of H(ρν)H(\rho\otimes\nu). Since the first nn rows of H~2,H~3\tilde{H}_{2},\tilde{H}_{3} and H~4\tilde{H}_{4} are zero as can be seen in (37) and (38), we have 1H(ρν)=0n×1.\mathcal{I}_{1}H(\rho\otimes\nu)=0\in\mathbb{R}^{n\times 1}. Similarly, 1H(νρ)=0n×1.\mathcal{I}_{1}H(\nu\otimes\rho)=0\in\mathbb{R}^{n\times 1}. Therefore, we have

1H(ρν)\displaystyle\mathcal{I}_{1}H(\rho\otimes\nu) =1H(νρ).\displaystyle=\mathcal{I}_{1}H(\nu\otimes\rho). (67)

Now let 2=[0I00]n×4n\mathcal{I}_{2}=\begin{bmatrix}0&I&0&0\end{bmatrix}\in\mathbb{R}^{n\times 4n} and multiply (66) from the right with 2\mathcal{I}_{2}:

2H(ρν)\displaystyle\mathcal{I}_{2}H(\rho\otimes\nu) =Ψ(ρ3ν3)+Z2(ρ3ν4)\displaystyle=\Psi(\rho_{3}\otimes\nu_{3})+\frac{Z}{2}(\rho_{3}\otimes\nu_{4})
Z2(ρ4ν3)+Ψ(ρ4ν4).\displaystyle-\frac{Z}{2}(\rho_{4}\otimes\nu_{3})+\Psi(\rho_{4}\otimes\nu_{4}). (68)

Let Sn2×n2S\in\mathbb{R}^{n^{2}\times n^{2}} be the permutation matrix so that

ρν=S(νρ),\displaystyle\rho\otimes\nu=S(\nu\otimes\rho), (69)

where SS is partitioned as S=[S1S2Sn]S=\begin{bmatrix}S_{1}&S_{2}&\dots S_{n}\end{bmatrix} and each Sjn2×nS_{j}\in\mathbb{R}^{n^{2}\times n}, for j=1,,nj=1,\dots,n, is defined as

Sj=[ejen+je2n+jen(n1)+j],\displaystyle S_{j}=\begin{bmatrix}e_{j}&e_{n+j}&e_{2n+j}&\dots&e_{n(n-1)+j}\end{bmatrix}, (70)

where ejn2×1e_{j}\in\mathbb{R}^{n^{2}\times 1} denotes the jjth column of a n2×n2{n^{2}\times n^{2}} identity matrix. Insert (69) into (VI-A) to obtain

2H(ρν)\displaystyle\mathcal{I}_{2}H(\rho\otimes\nu) =ΨS(ν3ρ3)+Z2S(ν3ρ4)\displaystyle=\Psi S(\nu_{3}\otimes\rho_{3})+\frac{Z}{2}S(\nu_{3}\otimes\rho_{4}) (71)
Z2S(ν4ρ3)+ΨS(ν4ρ4).\displaystyle~{}\quad-\frac{Z}{2}S(\nu_{4}\otimes\rho_{3})+\Psi S(\nu_{4}\otimes\rho_{4}).

Then, form ΨS\Psi S,

ΨS=\displaystyle\Psi S= [ΨS1ΨS2ΨSn],\displaystyle\begin{bmatrix}\Psi S_{1}&\Psi S_{2}&\dots\Psi S_{n}\end{bmatrix}, (72)

where, using (70) we have, for j=1,,nj=1,\dots,n,

ΨSj=[Ψ1(:,j)Ψ2(:,j)Ψn(:,j)],\displaystyle\Psi S_{j}=\begin{bmatrix}\Psi_{1}(:,j)&\Psi_{2}(:,j)&\dots&\Psi_{n}(:,j)\end{bmatrix}, (73)

and Ψ1(:,j)\Psi_{1}(:,j) is the MATLAB notation referring to the jjth column of Ψ1\Psi_{1}. Since for the coupling between the oscillators ii and jj, we have Kij=KjiK_{ij}=K_{ji} and γij=γji\gamma_{ij}=\gamma_{ji}, we can write Ψk(:,j)=Ψj(:,k)\Psi_{k}(:,j)=\Psi_{j}(:,k) and accordingly rewrite (73) as

ΨSj=Ψj,j=1,,n.\displaystyle\Psi S_{j}=\Psi_{j},\quad j=1,\dots,n. (74)

Then, inserting (74) into (72) yields

ΨS=Ψ.\displaystyle\Psi S=\Psi. (75)

Similarly, one can show that

ZS=Z.\displaystyle ZS=-Z. (76)

Plug (75) and (76) into (71) to obtain

2H(ρν)=2H(νρ).\displaystyle\mathcal{I}_{2}H(\rho\otimes\nu)=\mathcal{I}_{2}H(\nu\otimes\rho). (77)

Following similar steps, one obtains

kH(ρν)\displaystyle\mathcal{I}_{k}H(\rho\otimes\nu) =kH(νρ),fork=3,4,\displaystyle=\mathcal{I}_{k}H(\nu\otimes\rho),~{}~{}\mbox{for}~{}k=3,4, (78)

where 3=[00I0]n×4n\mathcal{I}_{3}=\begin{bmatrix}0&0&I&0\end{bmatrix}\in\mathbb{R}^{n\times 4n} and 4=[000I]n×4n\mathcal{I}_{4}=\begin{bmatrix}0&0&0&I\end{bmatrix}\in\mathbb{R}^{n\times 4n}. Combining (67), (77), and (78), we conclude (65), i.e., \mathcal{H} is symmetric. ∎

Now we are ready to prove Theorem 1.

Proof.

(of Theorem 1) We start by analyzing the first Sylvester equation involving W1W_{1} in Step 7 of Algorithm 1:

ETW1ΛA~μTW1=CTC^,\displaystyle-{E}^{T}W_{1}\Lambda-\tilde{A}_{\mu}^{T}W_{1}=C^{T}\hat{C}, (79)

where Λ=diag(λ1,,λr)rq×rq\Lambda=\text{diag}(\lambda_{1},\dots,\lambda_{r})\in\mathbb{R}^{r_{q}\times r_{q}} contains the eigenvalues of Er1ArE_{r}^{-1}A_{r}, C^p×rq\hat{C}\in\mathbb{R}^{p\times r_{q}}, and

A~μ=A~μE=A+H((Iq0)+(q0I))μE.\displaystyle\tilde{A}_{\mu}=\tilde{A}-\mu E=A+H\left((I\otimes q_{0})+(q_{0}\otimes I)\right)-\mu E. (80)

Recall H4n×(4n)2H\in\mathbb{R}^{4n\times(4n)^{2}} in (40):

H=[04n×4n2H~2PqTH~3PqTH~4PqT].\displaystyle H=\begin{bmatrix}0_{4n\times 4n^{2}}&\tilde{H}_{2}P_{q}^{T}&\tilde{H}_{3}P_{q}^{T}&\tilde{H}_{4}P_{q}^{T}\end{bmatrix}. (81)

Multiply H((Iq0)+(q0I))H\left((I\otimes q_{0})+(q_{0}\otimes I)\right) by the first unit vector e14n×1e_{1}\in\mathbb{R}^{4n\times 1} from the right to obtain

h1=H((e1q0)+(q0e1)),\displaystyle h_{1}=H\left((e_{1}\otimes q_{0})+(q_{0}\otimes e_{1})\right), (82)

where h14n×1h_{1}\in\mathbb{R}^{4n\times 1} denotes the first column of H((Iq0)+(q0I))H\left((I\otimes q_{0})+(q_{0}\otimes I)\right). Now, we apply Lemma 3 to (82) and use (81) to obtain

h1=2H(e1q0)\displaystyle h_{1}=2H(e_{1}\otimes q_{0}) =2H[q004n(4n1)]=0.\displaystyle=2H\begin{bmatrix}q_{0}\\ 0_{4n(4n-1)}\end{bmatrix}=0. (83)

Similarly, for the kkth column of (81), we obtain

hk=2H(ekq0)=0;k=2,3,,n.\displaystyle h_{k}=2H(e_{k}\otimes q_{0})=0~{}~{}~{};k=2,3,\ldots,n. (84)

Therefore, (83) and (84) reveal that the first nn columns of H((Iq0)+(q0I))H\left((I\otimes q_{0})+(q_{0}\otimes I)\right) are zero. Using this fact, and (32) and (33), we obtain the first nn columns of (80) as

A~μ(:,1:n)=[μI000]T,\displaystyle\tilde{A}_{\mu}(:,1:n)=\begin{bmatrix}-\mu I&0&0&0\end{bmatrix}^{T}, (85)

where Y(:,1:n)Y(:,1:n) refers to the first nn columns of the matrix YY. Decompose W1=[W11TW12TW13TW14T]T4n×rqW_{1}=\begin{bmatrix}W_{11}^{T}&W_{12}^{T}&W_{13}^{T}&W_{14}^{T}\end{bmatrix}^{T}\in\mathbb{R}^{4n\times r_{q}} where W1iW_{1i} n×rq;i=1,2,3,4\in\mathbb{R}^{n\times r_{q}}~{};i=1,2,3,4. Recall (18) and (32), and use (85) in (79) to obtain W11(μIΛ)=𝒞TC^W_{11}(\mu I-\Lambda)=\mathcal{C}^{T}\hat{C}.

Recall that μ\mu is a positive real number and the reduced order poles are assumed in the left-half plane. Therefore, the matrix μIΛ\mu I-\Lambda is invertible and W11=𝒞TC^(μIΛ)1.W_{11}=\mathcal{C}^{T}\hat{C}(\mu I-\Lambda)^{-1}. Similarly, let W2=[W21TW22TW23TW24T]TW_{2}=[W_{21}^{T}~{}W_{22}^{T}~{}W_{23}^{T}~{}W_{24}^{T}]^{T} and solve for the first n×rqn\times r_{q} block of W2W_{2} 4n×rq\in\mathbb{R}^{4n\times r_{q}} denoted by W21n×rqW_{21}\in\mathbb{R}^{n\times r_{q}} in the second Sylvester equation in Step 7 of Algorithm 1:

ETW2ΛA~μTW2=(2)(V1W1)(^(2))T.\displaystyle-{E}^{T}W_{2}\Lambda-\tilde{A}_{\mu}^{T}W_{2}=\mathcal{H}^{(2)}(V_{1}\otimes W_{1})(\hat{\mathcal{H}}^{(2)})^{T}. (86)

Next, we will prove that the first nn rows of (2)\mathcal{H}^{(2)} are zero. Towards this goal, recall (51) and let HH be partitioned as

H=(1)\displaystyle H={\mathcal{H}^{(1)}} =[124n],\displaystyle=\begin{bmatrix}\mathcal{H}_{1}&\mathcal{H}_{2}&\dots&\mathcal{H}_{4n}\end{bmatrix}, (87)

where i4n×4n\mathcal{H}_{i}\in\mathbb{R}^{4n\times 4n} for i=1,2,,4ni=1,2,\ldots,4n. From (81), we know that the leading 4n24n^{2} columns are zero. Therefore, we write

H=(1)=[04n×4n2n+14n],\displaystyle H={\mathcal{H}^{(1)}}=\begin{bmatrix}0_{4n\times 4n^{2}}&\mathcal{H}_{n+1}&\dots&\mathcal{H}_{4n}\end{bmatrix}, (88)

i.e., i=0\mathcal{H}_{i}=0 for i=1,2,,ni=1,2,\ldots,n. Recall that \mathcal{H} is symmetric due to Lemma 3. This means that

(2)=(3)4n×(4n)2,\displaystyle\mathcal{H}^{(2)}=\mathcal{H}^{(3)}~{}\in\mathbb{R}^{4n\times(4n)^{2}}, (89)

where (3)=[vec(1)vec(2)vec(4n)]T.\mathcal{H}^{(3)}=\begin{bmatrix}\text{vec}(\mathcal{H}_{1})&\text{vec}(\mathcal{H}_{2})&\dots&\text{vec}(\mathcal{H}_{4n})\end{bmatrix}^{T}. Since i=0\mathcal{H}_{i}=0 for i=1,2,,ni=1,2,\ldots,n, we obtain

(3)=[016n2×nvec(n+1)vec(4n)]T.\displaystyle\mathcal{H}^{(3)}=\begin{bmatrix}0_{16n^{2}\times n}&\text{vec}(\mathcal{H}_{n+1})\dots&\text{vec}(\mathcal{H}_{4n})\end{bmatrix}^{T}. (90)

Due to (89), the first nn rows of (2)\mathcal{H}^{(2)} are the first nn columns of ((3))T{(\mathcal{H}^{(3)})}^{T}. Then, (90) implies that (2)(1:n,:)=0\mathcal{H}^{(2)}(1:n,:)=0. Using this fact in (86) yields W21(μIΛ)=0W_{21}(\mu I-\Lambda)=0. Since μIΛ\mu I-\Lambda is invertible, we obtain W21=0n×rqW_{21}=0_{n\times r_{q}}. Recall WW in Step 88 of Algorithm (1) defined as W=W1+W2W=W_{1}+W_{2}. Then, for the first nn rows of WW, i.e., for WTW_{T}, we obtain

WT=W11+W21=W11=𝒞TC^(μIΛ)1.\displaystyle W_{T}=W_{11}+W_{21}=W_{11}=\mathcal{C}^{T}\hat{C}(\mu I-\Lambda)^{-1}. (91)

Hence, Range(WT)Range(𝒞T)\text{Range}(W_{T})\subseteq\text{Range}(\mathcal{C}^{T}), which proves (64). In addition, if prqp\leq r_{q}, the matrix C^\hat{C} has a right-inverse and thus the reverse direction holds as well, i.e., Range(𝒞T)Range(WT)\text{Range}(\mathcal{C}^{T})\subseteq\text{Range}(W_{T}), completing the proof of the theorem. ∎

We emphasize that the structure of the range of WTW_{T} due to Q-IRKA as shown in Lemma 1 is specific to the underlying power network dynamics and its quadratic form. The authors are not aware of any other models that yield such a form. This result will be crucial in forming the proposed algorithm next.

VI-B Proposed structure-preserving model reduction algorithm

Lemma 1 states that whenever prqp\leq r_{q} (which will be the common situation), WTW_{T} is rank-deficient. Therefore, even though this means that WTW_{T} is not suitable to be used as a model reduction space as is, we can use it to our advantage.

As noted earlier, we will perform Galerkin projection. In other words, in (12) we will use 𝒲=𝒱\mathcal{W}=\mathcal{V}. One option is to simply ignore WTW_{T} and set 𝒱=VT\mathcal{V}=V_{T}. However, this is not preferred in an input-output based systems-theoretic model reduction setting since the information in WTW_{T} related to the output is fully ignored. Fortunately, due to the subspace result (64) in Theorem 1, we can include both the input-to-state subspace information (VTV_{T}) and the state-to-output subspace information (WTW_{T}) in one subspace 𝒱\mathcal{V} by choosing in (12) as

𝒱=orth([VT𝒞T])n×r,\displaystyle\mathcal{V}=\text{orth}\left(\begin{bmatrix}{V}_{T}&\mathcal{C}^{T}\end{bmatrix}\right)\in\mathbb{R}^{n\times r}, (92)

where “orth” refers to forming an orthonormal basis so that 𝒱T𝒱=I\mathcal{V}^{T}\mathcal{V}=I. Then, we choose 𝒲=𝒱\mathcal{W}=\mathcal{V} and perform a Galerkin projection to compute the reduced second-order structure-preserving model (8)-(12). Thus, despite performing a one-sided Galerkin, by exploiting Theorem 1 and using the reduction basis 𝒱\mathcal{V} in (92), we are able to also incorporate the output information in the reduction framework, as is the standard in systems theoretic model reduction. We note that for the column dimension rr of 𝒱\mathcal{V}, we have r=min(rq+p,2rq)r=\text{min}(r_{q}+p,2r_{q}).

Now, we finally summarize our proposed approach, an 2\mathcal{H}_{2}-based Structure-preserving MOR method for Power Network Dynamics (StrH2), in Algorithm 3. We start by converting the original dynamics to the equivalent quadratic form in Step 1. Then, in Step 2, we either apply two-sided Q-IRKA (Algorithm 1) or one-sided Q-IRKA (Algorithm 2) to the resulting quadratic-system to construct VV. The resulting versions of the proposed method will be labeled as StrH2-A and StrH2-B, respectively. Step 3 constructs the final model reduction basis using the analysis of Theorem 1 and Step 4 constructs the final structured reduced-model.

We note that in (92), the WW-subspace information is not needed since it is already contained in 𝒞T\mathcal{C}^{T}. However, StrH2-A version of Algorithm 3, in Step 2, still uses two-sided Q-IRKA, which computes WW during the iteration unlike in one-sided Q-IRKA where the WW-subspace is never computed. One might think that since the WW-subspace information is replaced by 𝒞T\mathcal{C}^{T} in the end, the two options StrH2-A and StrH2-B should produce equivalent results. This is indeed not the case since the VV-subspace resulting from one-sided and two-sided Q-IRKA implementations are completely different; the former never used the output information. We will see in next section that this distinction does indeed impact the final reduced model.

Algorithm 3 2\mathcal{H}_{2}-based Structure-preserving MOR for Power Network Dynamics (StrH2)
1:Input: Second-order model (2)-(3)
δ¨(t)+𝒟δ˙(t)+f(δ)=u(t),y(t)=𝒞δ(t)\displaystyle\mathcal{M}\ddot{\delta}(t)+\mathcal{D}\dot{\delta}(t)+f(\delta)=\mathcal{B}u(t)~{},~{}y(t)=\mathcal{C}\delta(t)
2:Output: Reduced second-order model (8)-(9)
rδr¨(t)+𝒟rδr˙(t)+fr(δr)=ru(t),yr(t)=𝒞rδr(t)\displaystyle\mathcal{M}_{r}\ddot{\delta_{r}}(t)+\mathcal{D}_{r}\dot{\delta_{r}}(t)+f_{r}(\delta_{r})=\mathcal{B}_{r}u(t)~{},~{}y_{r}(t)=\mathcal{C}_{r}\delta_{r}(t)
3:Transform the second-order dynamic (2)-(3) into quadratic dynamics (63) with asymptotically stable linear part:
Ex˙(t)=Aμ~x(t)+H(x(t)x(t))+B~u~(t)y(t)=Cx(t),x(0)=0.\displaystyle\begin{array}[]{rcl}{E}\dot{x}(t)&=&\tilde{A_{\mu}}x(t)+H(x(t)\otimes x(t))+\tilde{B}\tilde{u}(t)\\ y(t)&=&Cx(t),~{}~{}~{}x(0)=0.\end{array} (95)
4:For the quadratic dynamics (63), apply
5:     Option A: Algorithm 1
6:                 or                     to construct VV.
7:     Option B: Algorithm 2
8:Form the final model reduction basis 𝒱\mathcal{V} using the leading nn rows of VV and the output matrix 𝒞\mathcal{C} as in (92):
𝒱=orth([VT𝒞T])\displaystyle\mathcal{V}=\text{orth}\left(\begin{bmatrix}{V}_{T}&\mathcal{C}^{T}\end{bmatrix}\right)
9:Compute the reduced matrices as in (12) with 𝒲=𝒱\mathcal{W}=\mathcal{V}.

VII Numerical Experiments

We test the proposed approach on two models: (i) the SM model of the 39-bus New England test system and (ii) IEEE 118 bus system, included in the MATPOWER software toolbox [44],[45]. We focus on a single-output system (p=1)(p=1) by choosing the arithmetic mean of all phase angles as the output as in [34, 33]. Models are generated by MATPOWER and MATLAB toolbox pg-sync-models [25] that provide all the necessary parameters to form (2) including JiJ_{i}, DiD_{i}, ωR\omega_{R}, KijK_{ij}, γij\gamma_{ij} and BiB_{i}. To illustrate the effectiveness of our approach, we compare it with two other approaches (i) a structure-preserving formulation of balanced truncation for power systems denoted by Str-QBT and (ii) POD.

For Str-QBT, we follow the approach of [33] in finding VV and WW. However, as opposed to [33], which performs the reduction on the quadratic dynamics (63) and constructs a quadratic reduced model, we directly find the reduced second order dynamic (8) by extracting 𝒱\mathcal{V} and 𝒲\mathcal{W} from VV and WW. To be more precise, [33] forms the reduction bases V{V} as

V\displaystyle V =blkdiag{Vb,Vb,Vb,Vb}\displaystyle=\text{blkdiag}\{{V_{b}},{V_{b}},{V_{b}},{V_{b}}\} (96)

and similarly for WW using WbW_{b} where Vb{V_{b}} and Wbn×rq{W_{b}}\in\mathbb{R}^{n\times r_{q}} are

Vb=RbTU^(:,1:rq)(Σ^(:,1:rq))12Wb=SbTV^(:,1:rq)(Σ^(:,1:rq))12,\displaystyle\begin{array}[]{rcl}{V_{b}}&=&R_{b}^{T}\hat{U}(:,1:r_{q})(\hat{\Sigma}(:,1:r_{q}))^{-\frac{1}{2}}\\ {W_{b}}&=&S_{b}^{T}\hat{V}(:,1:r_{q})(\hat{\Sigma}(:,1:r_{q}))^{-\frac{1}{2}},\end{array} (99)

RbSbT=U^Σ^V^TR_{b}S_{b}^{T}=\hat{U}\hat{\Sigma}\hat{V}^{T} is the SVD of RbSbTR_{b}S_{b}^{T}, and RbR_{b} and SbS_{b} are, respectively, the Cholesky factors of the second n×nn\times n block of the truncated reachability and observability gramians of the quadratic system [5], [33]. Then, [33] performs the reduction on (63) using (96) to obtain a reduced quadratic model for (63). Instead, since the second n×nn\times n blocks of the truncated reachability and observability gramians of the quadratic system correspond to the second order dynamic (2), we pick 𝒱=Vb\mathcal{V}=V_{b} and 𝒲=Wb\mathcal{W}=W_{b} in (12) and directly perform reduction on the second order dynamic (2)-(3) to obtain (8), (9).

To apply POD, we simulate (2) for a given time interval to form state snapshot matrix Δn×L\Delta\in\mathbb{R}^{n\times L} and compute its SVD:

Δ=[δ(t0)δ(t1)δ(tL1)]=UΔΣΔVΔT.\displaystyle\Delta=\begin{bmatrix}\delta(t_{0})&\delta(t_{1})&\dots&\delta(t_{L-1})\end{bmatrix}=U_{\Delta}\Sigma_{\Delta}V_{\Delta}^{T}.

Then, the rr leading left singular vectors of Δ\Delta, i.e., the leading rr columns of UΔU_{\Delta}, form the reduction basis 𝒱=𝒲\mathcal{V}=\mathcal{W} in (12).

To test the accuracy of the reduced models, we will measure the time-domain error between the true output y(t)y(t) and the reduced-model outputs yr(t)y_{r}(t) over a time interval t[0,T]t\in[0,T]. Towards this goal, define the (T)\mathcal{L}_{\infty}(T) norm of y(t)y(t) as

y(T)=maxt[0,T]y(t).\|y\|_{\mathcal{L}_{\infty}(T)}=\max_{t\in[0,T]}\mid y(t)\mid.~{}~{} (100)

The corresponding relative \mathcal{L}_{\infty} output error is defined as

e(T)=yyr(T)y(T).\|e\|_{\mathcal{L}_{\infty}(T)}=\frac{\|y-y_{r}\|_{\mathcal{L}_{\infty}(T)}}{\|y\|_{\mathcal{L}_{\infty}(T)}}.~{}~{} (101)

VII-A New England test system

The model has order n=39n=39 in the original second-order coordinates (2). We choose μ=103\mu=10^{-3} in (60). We apply both formulations of StrH2 in Algorithm 3 (StrH2-A and StrH2-B), Str-QBT, and POD; and reduce the order to r=2, 25r=2,\dots\,25. In Figure 1, the relative \mathcal{L}_{\infty} error e(T)\|e\|_{\mathcal{L}_{\infty}(T)} with T=10T=10 vs. reduction order rr is depicted for each method. The figure shows that for the majority of rr values, StrH2-A yields the lowest relative \mathcal{L}_{\infty} error. We note that this is the best case scenario for POD since the reduced model is tested with the same input that is used to train POD.

Refer to caption
Figure 1: Relative \mathcal{L}_{\infty} error vs. reduction order

For a more detailed comparison, in Figure 2, the output of the original and all the reduced order models of order r=23r=23 (at which the relative error due to StrH2-A drops below 1%1\%), and the corresponding absolute error, y(t)yr(t)\mid y(t)-y_{r}(t)\mid, are depicted. This figure supports the earlier observation that StrH2-A approximates the original model with a better accuracy compared to POD and Str-QBT.

Refer to caption
Figure 2: Original and the reduced order models output and the corresponding absolute error vs. time for r=23r=23

To investigate the robustness of the algorithms to variations in the input, we slightly perturb the input and repeat the same model reduction procedures as before. One might view this perturbation to the input as if the operation conditions are slightly varied. We perturb the input only by 0.1%0.1\%. Note that the choice of input has never entered the proposed algorithm or Str-QBT; however it is used to train the POD model. Figures 3 shows the relative error as rr varies for this slightly varied input. As in the previous case, StrH2-A outperforms the other methods. We also observe that the relative errors for the POD reduced models stagnate after r=13r=13, i.e., not much further improvements, while StrH2-A and StrH2-B, and Str-QBT approximate the original model with almost the same accuracy as in the earlier case. This is expected since the input never entered the model reduction process of the StrH2-A, StrH2-B or Str-QBT. Thus, for this small input variation case, those reduced models provide accurate approximations independent of the input selection as in the linear case.

Refer to caption
Figure 3: Relative \mathcal{L}_{\infty} error vs. reduction order

VII-B IEEE 118 bus system

The second-order model (2) in this case has n=118n=118. We choose μ=102\mu=10^{-2} and apply StrH2-A, StrH2-B and POD for r=2, 10r=2,\dots\,10. We use T=[03]T=[0~{}3]. Str-QBT has been skipped due to its poor performance in this example. The relative \mathcal{L}_{\infty} output error (101) vs. rr is depicted in Figure 4(a). Although POD yields a lower \mathcal{L}_{\infty} error (the testing input is the same as the training input for POD), StrH2-A has an acceptable accuracy as POD.

Refer to caption
(a) uu is not perturbed
Refer to caption
(b) uu is perturbed
Figure 4: Relative \mathcal{L}_{\infty} error vs. reduction order

As in the previous example, to test the robustness of the algorithms in the presence of a variation in the input, we slightly perturb the input uu by 0.5%0.5\%. As one can observe in Figure 4(b), both StrH2-A and StrH2-B outperform POD. Also as in the previous example, the relative errors for the POD reduced models do not show further improvements after r=3r=3, while both StrH2 formulations approximate the original model with similar accuracy as in the former case.

VIII Conclusion

We have developed a structure-preserving system-theoretic MOR technique for nonlinear power grid networks. We lifted the original nonlinear model to its equivalent quadratic form in order to employ Q-IRKA as an intermediate stage of our MOR framework. We have shown that the model reduction bases obtained by Q-IRKA have a specific subspace structure that can be exploited to construct the desired reduction basis for reducing the original nonlinear model. This reduction basis has led to a reduced model that preserved the physically meaningful second-order structure of the original model. We have illustrated the effectiveness of our proposed approach via two numerical examples.

References

  • [1] U. D. Annakkage, N. K. C. Nair, Y. Liang, A. M. Gole, V. Dinavahi, B. Gustavsen, T. Noda, H. Ghasemi, A. Monti, M. Matar, R. Iravani, and J. A. Martinez. Dynamic system equivalents: A survey of available techniques. IEEE Transactions on Power Delivery, 27(1):411–420, 2012.
  • [2] A. C. Antoulas, C. Beattie, and S. Güğercin. Interpolatory methods for model reduction. Computational Science and Engineering 21. SIAM, Philadelphia, 2020.
  • [3] P. Benner and T. Breiten. Interpolation-based 2\mathcal{H}_{2}-model reduction of bilinear control systems. SIAM Journal on Matrix Analysis and Applications, 33(3):859–885, 2012.
  • [4] P. Benner and T. Breiten. Two-sided projection methods for nonlinear model order reduction. SIAM Journal on Scientific Computing, 37(2):B239–B260, 2015.
  • [5] P. Benner and P. Goyal. Balanced truncation model order reduction for quadratic-bilinear control systems. e-print 1705.00160, arXiv, 2017.
  • [6] P. Benner, P. Goyal, and S. Gugercin. 2\mathcal{H}_{2}-quasi-optimal model order reduction for quadratic-bilinear control systems. SIAM Journal on Matrix Analysis and Applications, 39(2):983–1032, 2018.
  • [7] B. Besselink, U. Tabak, A. Lutowska, N. van de Wouw, H. Nijmeijer, D.J. Rixen, M.E. Hochstenbach, and W.H.A. Schilders. A comparison of model reduction techniques from structural dynamics, numerical mathematics and systems and control. Journal of Sound and Vibration, 332(19):4403–4422, 2013.
  • [8] D. Chaniotis and M. A. Pai. Model reduction in power systems using Krylov subspace methods. IEEE Transactions on Power Systems, 20(2):888–894, 2005.
  • [9] S. Chaturantabut and D. C. Sorensen. Nonlinear model reduction via discrete empirical interpolation. SIAM Journal on Scientific Computing, 32(5):2737–2764, 2010.
  • [10] X. Cheng and J. M. A. Scherpen. Clustering approach to model order reduction of power networks with distributed controllers. Advances in Computational Mathematics, 44(6):1917–1939, Dec 2018.
  • [11] A. Cherid and M. Bettayeb. Reduced-order models for the dynamics of a single-machine power system via balancing. Electric Power Systems Research, 22(1):7 – 12, 1991.
  • [12] J. H. Chow. Power system coherency and model reduction, volume 84. Springer, 2013.
  • [13] J. H. Chow, J. R. Winkelman, M. A. Pai, and P. W. Sauer. Singular perturbation analysis of large-scale power systems. International Journal of Electrical Power & Energy Systems, 12(2):117 – 126, 1990.
  • [14] C. Gu. Qlmor: A projection-based nonlinear model order reduction approach using quadratic-linear representation of nonlinear systems. IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, 30(9):1307–1320, 2011.
  • [15] M. Hinze and S. Volkwein. Proper orthogonal decomposition surrogate models for nonlinear dynamical systems: Error estimates and suboptimal control. In P. Benner, D. C. Sorensen, and V. Mehrmann, editors, Dimension Reduction of Large-Scale Systems, pages 261–306. Springer Berlin Heidelberg, 2005.
  • [16] T. Ishizaki, K. Kashima, A. Girard, J. Imura, L. Chen, and K. Aihara. Clustered model reduction of positive directed networks. Automatica, 59:238 – 247, 2015.
  • [17] T. G. Kolda and B. W. Bader. Tensor decompositions and applications. SIAM Review, 51(3):455–500, 2009.
  • [18] J. Leung, M. Kinnaert, J. Maun, and F. Villella. Model reduction of coherent LPV models in power systems. In 2019 IEEE Power Energy Society General Meeting (PESGM), pages 1–5, 2019.
  • [19] Y. Levron and J. Belikov. Reduction of power system dynamic models using sparse representations. IEEE Transactions on Power Systems, 32(5):3893–3900, 2017.
  • [20] S. Liu. Dynamic-data driven real-time identification for electric power systems. PhD thesis, University of Illinois at Urbana-Champaign, 2009.
  • [21] M. H. Malik, D. Borzacchiello, F. Chinesta, and P. Diez. Reduced order modeling for transient simulation of power systems using trajectory piece-wise linear approximation. Advanced Modeling and Simulation in Engineering Sciences, 3(1):31, Dec 2016.
  • [22] G. P. McCormick. Computability of global solutions to factorable nonconvex programs: Part i—convex underestimating problems. Mathematical programming, 10(1):147–175, 1976.
  • [23] P. Mlinarić, S. Grundel, and P. Benner. Efficient model order reduction for multi-agent systems using QR decomposition-based clustering. In Proceedings of 54th IEEE Conference on Decision and Control, pages 4794–4799, 2015.
  • [24] P. Mlinarić, T. Ishizaki, A. Chakrabortty, S. Grundel, P. Benner, and J. Imura. Synchronization and aggregation of nonlinear power systems with consideration of bus network structures. In 2018 European Control Conference (ECC), pages 2266–2271, 2018.
  • [25] T. Nishikawa and A. E Motter. Comparative analysis of existing models for power-grid synchronization. New Journal of Physics, 17(1):015012, 2015.
  • [26] D. Osipov and K. Sun. Adaptive nonlinear model reduction for fast power system simulation. IEEE Transactions on Power Systems, 33(6):6746–6754, 2018.
  • [27] M. A. Pai and R. P. Adgaonkar. Singular perturbation analysis of nonlinear transients in power systems. In 1981 20th IEEE Conference on Decision and Control including the Symposium on Adaptive Processes, pages 221–222, 1981.
  • [28] P. A. Parrilo, S. Lall, F. Paganini, G. C. Verghese, B. C. Lesieutre, and J. E. Marsden. Model reduction for analysis of cascading failures in power systems. In Proceedings of the 1999 American Control Conference (Cat. No. 99CH36251), volume 6, pages 4208–4212, 1999.
  • [29] E. Purvine, E. Cotilla-Sanchez, M. Halappanavar, Z. Huang, G. Lin, S. Lu, and S. Wang. Comparative study of clustering techniques for real-time dynamic model reduction. Statistical Analysis and Data Mining: The ASA Data Science Journal, 10(5):263–276, 2017.
  • [30] J. Qi, J. Wang, H. Liu, and A. D. Dimitrovski. Nonlinear model reduction in power systems by balancing of empirical controllability and observability covariances. IEEE Transactions on Power Systems, 32(1):114–126, 2017.
  • [31] E. Qian, B. Kramer, B. Peherstorfer, and K. Willcox. Lift & learn: Physics-informed machine learning for large-scale nonlinear dynamical systems. Physica D: Nonlinear Phenomena, 406:132401, 2020.
  • [32] A. Ramirez, A. Mehrizi-Sani, D. Hussein, M. Matar, M. Abdel-Rahman, J. Jesus Chavez, A. Davoudi, and S. Kamalasadan. Application of balanced realizations for model-order reduction of dynamic power system equivalents. IEEE Transactions on Power Delivery, 31(5):2304–2312, 2016.
  • [33] T. K.S. Ritschel, F. Weiß, M. Baumann, and S. Grundel. Nonlinear model reduction of dynamical power grid models using quadratization and balanced truncation. at - Automatisierungstechnik, 68(12):1022–1034, 2020.
  • [34] B. Safaee and S. Gugercin. Data-driven modeling of power networks. In 2021 60th IEEE Conference on Decision and Control (CDC), pages 4236–4241, 2021.
  • [35] B. Safaee and S. Gugercin. Structure-preserving model reduction of parametric power networks. In 2021 American Control Conference (ACC), pages 1824–1829, 2021.
  • [36] C. Sturk, L. Vanfretti, Y. Chompoobutrgool, and H. Sandberg. Coherency-independent structured model reduction of power systems. IEEE Transactions on Power Systems, 29(5):2418–2426, 2014.
  • [37] A. van der Schaft. On model reduction of physical network systems. In Proceedings of 21st International Symposium on Mathematical Theory of Networks and Systems (MTNS), pages 1419–1425, 2014.
  • [38] C. Wang, H. Yu, P. Li, C. Ding, C. Sun, X. Guo, F. Zhang, Y. Zhou, and Z. Yu. Krylov subspace based model reduction method for transient simulation of active distribution grid. In 2013 IEEE Power Energy Society General Meeting, pages 1–5, 2013.
  • [39] C. Wang, H. Yu, P. Li, J. Wu, and C. Ding. Model order reduction for transient simulation of active distribution networks. IET Generation, Transmission & Distribution, 9:457–467(10), April 2015.
  • [40] S. Wang, S. Lu, G. Lin, and N. Zhou. Measurement-based coherency identification and aggregation for power systems. In 2012 IEEE Power and Energy Society General Meeting, pages 1–7, 2012.
  • [41] S. Wang, S. Lu, N. Zhou, G. Lin, M. Elizondo, and M. A. Pai. Dynamic-feature extraction, attribution, and reconstruction (DEAR) method for power system model reduction. IEEE Transactions on Power Systems, 29(5):2049–2059, 2014.
  • [42] H. Zhao, X. Lan, and H. Ren. Nonlinear power system model reduction based on empirical gramians. Journal of Electrical Engineering, 68(6):425 – 434, 2017.
  • [43] H. S. Zhao, N. Xue, and N. Shi. Nonlinear Dynamic Power System Model Reduction Analysis Using Balanced Empirical Gramian. Applied Mechanics and Materials, 448-453:2368–2374, October 2013.
  • [44] R. D. Zimmerman and C. E. Murillo-Sánchez. Matpower 6.0 user’s manual. Power Systems Engineering Research Center, 9, 2016.
  • [45] R. D. Zimmerman, C. E. Murillo-Sánchez, and R. J. Thomas. Matpower: Steady-state operations, planning, and analysis tools for power systems research and education. IEEE Transactions on Power Systems, 26(1):12–19, 2010.