This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

\useunder

\ul

Signal Processing over Multilayer Graphs: Theoretical Foundations and Practical Applications

Songyang Zhang, Qinwen Deng, and Zhi Ding This work was supported in part by the National Science Foundation under Grants 2029848 and 2002937.S. Zhang, Q. Deng and Z. Ding are with Department of Electrical and Computer Engineering, University of California, Davis, CA, 95616. (E-mail: [email protected], [email protected], and [email protected]).
Abstract

Signal processing over single-layer graphs has become a mainstream tool owing to its power in revealing obscure underlying structures within data signals. However, many real-life datasets and systems, including those in Internet of Things (IoT), are characterized by more complex interactions among distinct entities, which may represent multi-level interactions that are harder to be captured with a single-layer graph, and can be better characterized by multilayers graph connections. Such multilayer or multi-level data structure can be more naturally modeled by high-dimensional multilayer graphs (MLG). To generalize traditional graph signal processing (GSP) over multilayer graphs for analyzing multi-level signal features and their interactions, this work proposes a tensor-based framework of multilayer graph signal processing (M-GSP). Specifically, we introduce core concepts of M-GSP and study properties of MLG spectrum space, followed by fundamentals of MLG-based filter design. To illustrate novel aspects of M-GSP, we further explore its link with traditional signal processing and GSP. We provide example applications to demonstrate the efficacy and benefits of applying multilayer graphs and M-GSP in practical scenarios.

Index Terms:
Multilayer graph, graph signal processing, tensor, data analysis

I Introduction

Geometric signal processing tools have found broad applications in data analysis to uncover obscure or hidden structures from complex datasets [1]. Various data sources, such as social networks, Internet of Things (IoT) intelligence, traffic flows and biological images, often feature complex structures that pose challenges to traditional signal processing tools. Recently, graph signal processing (GSP) emerges as an effective tool over the graph signal representation [2]. For a signal with NN samples, a graph of NN nodes can be formed to model their underlying interactions [3]. In GSP, a graph Fourier space is also defined from the spectrum space of the representing matrix (adjacency/Laplacian) for signal processing tasks [4], such as denoising [5], resampling [6], and classification [7]. Generalization of the more traditional GSP includes signal processing over hypergraphs [8] and simplicial complexes [9], which are suitable to model high-degree multi-lateral node relationships.

Traditional graph signal processing tools generally describe signals as graph nodes connected by one type of edges. However, real-life systems and datasets may feature multi-facet interactions [10]. For example, in a video dataset modeled by spatial-temporal graph shown in Fig. 1, the nodes may exhibit different types of spatial connections at different temporal steps. It is harder for single-layer graphs to model such multi-facet connections. To model multiple layers of signal connectivity, we explore a high-dimensional graph representation known as multilayer graphs (MLG) [11].

Multilayer graph, also named as multilayer network, is a geometric model containing correlated layers with different structures and physical meanings, unlike traditional single-layer graphs [10]. A typical example is smart grid consisting of two layers shown as Fig. 1: the power grid and the computation network. These two layers have different physical connectivity and rules [12]. Still, signal interactions across the multiple layers in MLG can be strongly correlated. Thus, separate representations by multiple single layer graphs may fail to capture such characteristics. Consider a network consisting of a physical power layer and a cyber layer, the failure of one layer could trigger the failure of the other [13]. One example was the power line damage caused by a storm on September 28th of 2003. Not only did it lead to the failure of several power stations, but also disrupted communications as a result of power station breakdowns that eventually affected 56 million people in Europe [14].

Refer to caption
Refer to caption
Figure 1: Multilayer Graphs and Applications: (a) Video: each layer represents one frame of the video and the edges capture the spatial-temporal relationships; (b) Cyber-Physical System (CPS): each layer represents one component in CPS and the edges capture the physical connections.

The complexity and multi-level interactions of MLG make the data reside on the irregular and high-dimensional structures, which do not directly lend themselves to standard GSP tools. For example, even though one can represent MLG by a supra-graph unfolding all the layers [15], traditional GSP would treat interlayer and intralayer interactions equivalently in one spectrum space without differentiating the spectra of interlayer and intralayer signal correlations. Recently, there has been growing interest in developing advanced GSP tools to process such multi-level structures. In [16], a joint time-vertex Fourier transform (JFT) is defined to process spatial-temporal graphs by applying graph Fourier transform (GFT) and discrete Fourier transform (DFT). Although JFT can process time-varying datasets, it does not provide more general temporal (interlayer) connectivity in a generic multilayer graph. For flexible interlayer structure, a two-way GFT proposed in [17], defines different graph Fourier spaces for interlayer and intralayer connections, respectively. However, its spatial interactions all lie within a single graph structure, thereby limiting the generalization of MLG. Expanding the works of [17], the tensor-based multi-way graph signal processing framework (MWGSP) of [18] relies on product graph. In MWGSP, separate factor graphs are constructed for each mode of a tensor-represented signal, and a joint spectrum combines factor graph spectra. However, both JFT and MWGSP use the same intralayer connections for all layers. They do not admit different layers with heterogeneous graph structures, thereby limiting the ability to represent a general MLG.

Another challenge in MLG signal processing lies in the need for a suitable mathematical representation. Traditional methods start with connectivity matrices. For example, in [19], a supra-adjacency matrix is defined to represent all layers equivalently while ignoring the natures of different layers. One can also represent each layer with an individual adjacency matrix [20]. However, such matrix-based representations mainly focus on the intralayer connections and lack representation for interlayer interactions. A more natural and general way may start with tensor representation [11], which is particularly attractive in handling complex MLG graph analysis.

Our goal is to generalize graph signal processing for multilayer graphs to model, analyze, and process signals based on the intralayer and interlayer signal interactions. To address the aforementioned challenges and to advance MLG processing, we present a novel tensor framework for multilayer graph signal processing (M-GSP). We summarize the main contributions of this work as follows:

  • Leveraging tensor representation of MLG, we introduce M-GSP from a spatial perspective, in which MLG signals and shifting of signals are defined;

  • Taking a spectrum perspective, we introduce new concepts of spectrum space and spectrum transform for MLG. For interpretability of spectrum space, we analyze the resulting MLG spectral properties and their distinction from existing GSP tools.

  • We also present fundamentals of filter design in M-GSP, and suggest several practical applications based on the proposed framework, including those in IoT systems.

We organize the technical presentation as follows. Section II first summarizes preliminaries of traditional GSP and tensor analysis, before presenting representations of multilayer graphs within M-GSP in Section. III. We then introduce the fundamentals of M-GSP spatial and spectrum analysis in Section IV and Section V, respectively. We next discuss MLG filter design in Section VI. We develop the physical insights and spectrum interpretation of M-GSP concepts in Section VII. With the newly proposed M-GSP framework, we provide several example applications to demonstrate its potential in Section VIII, before concluding this paper in Section IX.

II Preliminaries

II-A Overview of Graph Signal Processing

Signal processing on graphs [1, 3, 2] studies signals that are discrete in some dimensions by representing the irregular signal structure using a graph 𝒢={𝒱,𝐅}\mathcal{G}=\{\mathcal{V},\mathbf{F}\}, where 𝒱={v1,v2,,vN}\mathcal{V}=\{v_{1},v_{2},\cdots,v_{N}\} is a set of NN nodes, and 𝐅N×N\mathbf{F}\in\mathbb{R}^{N\times N} is the representing matrix (e.g., adjacency/Laplacian) describing the geometric structure of the graph 𝒢\mathcal{G}. Graph signals are the attributes of nodes that underlie the graph structure. A graph signal can be written as vector 𝐬=[s1,s2,,sN]TN\mathbf{s}=[s_{1},s_{2},\cdots,s_{N}]^{\mathrm{T}}\in\mathbb{R}^{N} where the superscript ()T(\cdot)^{\mathrm{T}} denotes matrix/vector transpose.

With a graph representation 𝐅\mathbf{F} and a signal vector 𝐬\mathbf{s}, the basic graph filtering (shifting) is defined via

𝐬=𝐅𝐬.\mathbf{s}^{\prime}=\mathbf{Fs}. (1)

The graph spectrum space, also known as the graph Fourier space is defined based on the eigenspace of the representing matrix. Let the eigen-decomposition of 𝐅\mathbf{F} be given by 𝐅=𝐕𝚲𝐕1\mathbf{F}=\mathbf{V}\mathbf{\Lambda}\mathbf{V}^{-1}, where 𝐕\mathbf{V} is the matrix with eigenvectors of 𝐅\mathbf{F} as columns, and diagonal matrix 𝚲\mathbf{\Lambda} consists of the corresponding eigenvalues. The graph Fourier transform (GFT) is defined as

𝐬^=𝐕1𝐬,\hat{\mathbf{s}}=\mathbf{V}^{-1}\mathbf{s}, (2)

whereas the inverse GFT is given by 𝐬=𝐕𝐬^\mathbf{s}=\mathbf{V}\hat{\mathbf{s}}.

From definitions of GFT, other concepts, such as sampling theory [21], filter design [22], and frequency analysis [4] can be developed for signal processing and data analysis tasks.

II-B Introduction of Tensor Basics

Before introducing the fundamentals of M-GSP, we first review some basics on tensors that are useful for multilayer graph analysis. Tensors can be viewed as multi-dimensional arrays. The order of a tensor is the number of indices needed to label a component of that array [23]. For example, a third-order tensor has three indices. More specially, a scalar is a zeroth-order tensor; a vector is a first-order tensor; a matrix is a second-order tensor; and an MM-dimensional array is an MMth-order tensor. For convenience, we use bold letters to represent the tensors excluding scalars, i.e., 𝐀I1×I2×IN\mathbf{A}\in\mathbb{R}^{I_{1}\times I_{2}\cdots\times I_{N}} represents an NNth-order tensor with IkI_{k} being the dimension of the kkth order, and use Ai1iNA_{{i_{1}}\cdots i_{N}} to represent the entry of 𝐀\mathbf{A} at position (i1,i2,,iN)(i_{1},i_{2},\cdots,i_{N}) with 1ikIk1\leq i_{k}\leq I_{k} in this work. If 𝐀f\mathbf{A}_{f} has a subscript ff, we use [Af]i1iN[A_{f}]_{{i_{1}}\cdots i_{N}} to denote its entries.

We now start with some useful definitions and tensor operations for the M-GSP framework [23].

II-B1 Super-diagonal Tensor

An NNth-order tensor 𝐀I1×I2×IN\mathbf{A}\in\mathbb{R}^{I_{1}\times I_{2}\cdots\times I_{N}} is super-diagonal if its entries Ai1i2iN0A_{i_{1}i_{2}\cdots i_{N}}\neq 0 only for i1=i2==iNi_{1}=i_{2}=\cdots=i_{N}.

II-B2 Symmetric Tensor

A tensor is super-symmetric if its elements remain constant under index permutation. For example, a third-order 𝐀I×I×I\mathbf{A}\in\mathbb{R}^{I\times I\times I} is super-symmetric if Aijk=Ajik=Akij=Akji=Ajik=AjkiA_{ijk}=A_{jik}=A_{kij}=A_{kji}=A_{jik}=A_{jki}. In addition, tensors can be partially symmetric in two or more modes as well. For example, a third-order tensor 𝐀I×I×J\mathbf{A}\in\mathbb{R}^{I\times I\times J} is symmetric in the order one and two if Aijk=AjikA_{ijk}=A_{jik}, for 1i,jI1\leq i,j\leq I and 1kJ1\leq k\leq J.

II-B3 Tensor Outer Product

The tensor outer product between a PPth-order tensor 𝐔I1××IP\mathbf{U}\in\mathbb{R}^{I_{1}\times\cdots\times I_{P}} with entries Ui1iPU_{i_{1}...i_{P}} and a QQth-order tensor 𝐕J1××JQ\mathbf{V}\in\mathbb{R}^{J_{1}\times\cdots\times J_{Q}} with entries Vj1jQV_{j_{1}...j_{Q}} is denoted by

𝐖=𝐔𝐕.\mathbf{W}=\mathbf{U}\circ\mathbf{V}. (3)

The result 𝐖I1××IP×J1××JQ\mathbf{W}\in\mathbb{R}^{I_{1}\times\cdots\times I_{P}\times J_{1}\times\cdots\times J_{Q}} is a (P+Q)(P+Q)th-order tensor, whose entries are calculated by

Wi1iPj1jQ=Ui1iPVj1jQ.W_{i_{1}...i_{P}j_{1}...j_{Q}}=U_{i_{1}...i_{P}}\cdot V_{j_{1}...j_{Q}}. (4)

The tensor outer product is useful to construct a higher order tensor from several lower order tensors.

II-B4 n-mode Product

The n-mode product between a tensor 𝐔I1××IP\mathbf{U}\in\mathbb{R}^{I_{1}\times\cdots\times I_{P}} and a matrix 𝐕J×In\mathbf{V}\in\mathbb{R}^{J\times I_{n}} is denoted by

𝐖=𝐔×n𝐕I1××In1×J×In+1××IP.\mathbf{W}=\mathbf{U}\times_{n}\mathbf{V}\in\mathbb{R}^{I_{1}\times\cdots\times I_{n-1}\times J\times I_{n+1}\times\cdots\times I_{P}}. (5)

Each element in 𝐖\mathbf{W} is given by

Wi1i2in1jin+1iP=in=1InUi1iPVjin.W_{i_{1}i_{2}\cdots i_{n-1}ji_{n+1}\cdots i_{P}}=\sum_{i_{n}=1}^{I_{n}}U_{i_{1}\cdots i_{P}}V_{ji_{n}}. (6)

Note that the nn-mode product is a different operation from matrix product.

Refer to caption
(a) CP Decomposition.
Refer to caption
(b) Tucker Decomposition.
Figure 2: Diagram of Tensor Decompositions.

II-B5 Tensor Contraction

In M-GSP, the contraction (inner product) between a forth order tensor 𝐀M×N×M×N\mathbf{A}\in\mathbb{R}^{M\times N\times M\times N} and a matrix 𝐱M×N\mathbf{x}\in\mathbb{R}^{M\times N} in the third and forth order is defined as

𝐲=𝐀𝐱M×N,\mathbf{y}=\mathbf{A}\diamond\mathbf{x}\in\mathbb{R}^{M\times N}, (7)

where yαi=β=1Mj=1NAαiβjxβjy_{\alpha i}=\sum_{\beta=1}^{M}\sum_{j=1}^{N}A_{\alpha i\beta j}x_{\beta j}.

In addition, the contraction between two fourth-order tensor 𝐔,𝐕M×N×M×N\mathbf{U},\mathbf{V}\in\mathbb{R}^{M\times N\times M\times N} is defined as

𝐖=𝐔𝐕M×N×M×N,\mathbf{W}=\mathbf{U}\odot\mathbf{V}\in\mathbb{R}^{M\times N\times M\times N}, (8)

whose entries are Wαiϵp=βjUαiβjVβjϵpW_{\alpha i\epsilon p}=\sum_{\beta j}U_{\alpha i\beta j}V_{\beta j\epsilon p}.

II-B6 Tensor Decomposition

Tensor decompositions are useful tools to extract the underlying information of tensors. Particularly, CANDECOMP/PARAFAC (CP) decomposition decomposes a tensor as a sum of the tensor outer product of rank-one tensors [23, 24]. For example, a third order tensor 𝐓I×J×K\mathbf{T}\in\mathbb{R}^{I\times J\times K} is decomposed by CP decomposition into

𝐓r=1R𝐚r𝐛r𝐜r,𝐚rI,𝐛rJ,𝐜rK\mathbf{T}\approx\sum_{r=1}^{R}\mathbf{a}_{r}\circ\mathbf{b}_{r}\circ\mathbf{c}_{r},\quad\mathbf{a}_{r}\in\mathbb{R}^{I},\mathbf{b}_{r}\in\mathbb{R}^{J},\mathbf{c}_{r}\in\mathbb{R}^{K} (9)

where integer RR denotes the rank, i.e., the lowest number of rank-one tensors in the decomposition. We illustrate CP decomposition for a third-order tensor in Fig. 2(a), which could be viewed as factorization of the tensor.

Another important decomposition is the Tucker decomposition, which is in the form of higher-order PCA. More specifically, Tucker decomposition decomposes a tensor into a core tensor multiplied by a matrix along each mode [23]. Defining a core tensor 𝐒=[Spqr]P×Q×R\mathbf{S}=[S_{pqr}]\in\mathbb{R}^{P\times Q\times R}. Defining 𝐀I×P,𝐁J×Q,𝐂K×R\mathbf{A}\in\mathbb{R}^{I\times P},\mathbf{B}\in\mathbb{R}^{J\times Q},\mathbf{C}\in\mathbb{R}^{K\times R}, the Tucker decomposition of a third-order tensor 𝐓I×J×K\mathbf{T}\in\mathbb{R}^{I\times J\times K} is

𝐓\displaystyle\mathbf{T} 𝐒×1𝐀×2𝐁×3𝐂,\displaystyle\approx\mathbf{S}\times_{1}\mathbf{A}\times_{2}\mathbf{B}\times_{3}\mathbf{C},\quad
=p=1Pq=1Qr=1RSpqr𝐚p𝐛q𝐜r,\displaystyle=\sum_{p=1}^{P}\sum_{q=1}^{Q}\sum_{r=1}^{R}S_{pqr}\mathbf{a}_{p}\circ\mathbf{b}_{q}\circ\mathbf{c}_{r}, (10)

where 𝐀=[𝐚1𝐚p]\mathbf{A}=[\mathbf{a}_{1}\;\cdots\;\mathbf{a}_{p}], 𝐁=[𝐛1𝐛Q]\mathbf{B}=[\mathbf{b}_{1}\;\cdots\;\mathbf{b}_{Q}], 𝐂=[𝐜1𝐜R]\mathbf{C}=[\mathbf{c}_{1}\;\cdots\;\mathbf{c}_{R}]. The Tucker decomposition for a third-order tensor is illustrated in Fig. 2(b). Tucker decomposition reduces to CP decomposition if the core tensor is limited to be super-diagonal.

Other typical decompositions include Higher-Order SVD (HOSVD) [25], orthogonal CP-decomposition [26], and Tensor-Train decomposition [27]. Interested readers are referred to the tutorial [23] for more details.

III Representations of MLG in M-GSP

In this section, we introduce the fundamental representations of MLG within the M-GSP framework.

Refer to caption
Refer to caption
Figure 3: Example of multilayer graphs: (a) A three-layer interconnected graph; (b) A three-layer multiplex graph.

III-A Multilayer Graphs

Multilayer graph, also referred to as multilayer network, is an important geometric model in complex systems [10]. In this work, we refrain from using the “network” terminology because of its diverse meanings in various field ranging from communication networking to deep learning. From here onward, unless otherwise specified, we shall use the less ambiguous term of multilayer graph.

We first provide definitions of multilayer graphs (MLG).

Definition 1 (Multilayer Graph).

A multilayer graph with KK nodes and MM layers is defined as ={𝒱,,𝐅}\mathcal{M}=\{\mathcal{V},\mathcal{L},\mathbf{F}\}, where 𝒱={v1,v2,,vK}\mathcal{V}=\{v_{1},v_{2},\cdots,v_{K}\} is the set of nodes, ={l1,l2,,lM}\mathcal{L}=\{l_{1},l_{2},\cdots,l_{M}\} denotes the set of layers with each layer li={vi1,,vin}l_{i}=\{v_{i_{1}},\cdots,v_{i_{n}}\} being the subsets of 𝒱\mathcal{V}, whereas 𝐅\mathbf{F} is the algebraic representation describing node interactions.

Note that, we mainly focus on the layer-disjoint multilayer graph [10] where each node exists exactly in one layer, since layers denote different phenomena. For example, in a smart grid, a station with functions in both power grid and communication network, is usually modeled as two nodes in a two-layer graph for the network analysis [28].

In multilayer graphs, edges connect nodes in the same layer (intralayer edges) or nodes of different layers (interlayer edges) [29]. There are two main types of multilayer graphs: multiplex graph and interconnected graph [30]. In a multiplex graph, each layer has the same number of nodes, and each node only connects with their 1-to-1 matching counterparts in other layers to form interlayer connections. Typically, multiplex graphs characterize different types of interactions among the same (or a similar) set of physical entities. For example, the spatial-temporal connections among a set of nodes can be intuitively modeled as a multiplex graph [30]. In the interconnected graphs, each layer may have different numbers of nodes without a 1-to-1 counterpart. Their interlayer connections could be more flexible. Examples of a three-layer multiplex graph and a three-layer interconnected graph are shown in Fig. 3, where different colors represent different layers, solid lines represent intralayer connections, and dash lines indicate interlayer connections.

III-B Algebraic Representation

To capture the high-dimensional ‘multilayer’ interactions between different nodes, we use tensor as algebraic representation of MLG for the proposed M-GSP framework [11].

III-B1 MLG with same number of nodes in each layer

To better interpret the tensor representation of a multilayer graph, we start from a simpler type of MLG, in which each layer contains the same number of nodes. For a multilayer graph ={𝒱,}\mathcal{M}=\{\mathcal{V},\mathcal{L}\} with ||=M|\mathcal{L}|=M layers and NN nodes in each layer, i.e., |li|=N|l_{i}|=N for 1iM1\leq i\leq M, it could be interpreted as embedding the interactions between a set of NN ‘entities’ (not nodes) into a set of MM layers. The nodes in different layers can be viewed as the projections of the entities. For example, the video datasets could be modeled by the spatial connections between objects (entities) into different temporal frames (layers).

Mathematically, the process of embedding (projecting) entities can be viewed as a tensor product, and graph connections can be represented by tensors [11]. For convenience, we use Greek letters α,β,\alpha,\beta,\cdots to indicate each layer and Latin letters i,j,i,j,\cdots to indicate each interpretable ‘entity’ with corresponding node in each layer. Given a set of entities 𝒳={x1,x2,,xN}\mathcal{X}=\{x_{1},x_{2},\cdots,x_{N}\}, one can construct a vector 𝐞i=[0,,0,1,0,,0]T\mathbf{e}_{i}=[0,\cdots,0,1,0,\cdots,0]^{\mathrm{T}} whose sole nonzero element is its ii-th element (equal to 1) to characterize each entity ii. Thus, interactions of two entities can be represented by a second-order tensor 𝐀X=i,j=1Naij𝐞i𝐞jN×N\mathbf{A}_{X}=\sum_{i,j=1}^{N}a_{ij}\mathbf{e}_{i}\circ\mathbf{e}_{j}\in\mathbb{R}^{N\times N}, where aija_{ij} is the intensity of the relationship between entity ii and jj. Similarly, given a set of layers ={l1,l2,,lM}\mathcal{L}=\{l_{1},l_{2},\cdots,l_{M}\}, a vector 𝐞αM\mathbf{e}_{\alpha}\in\mathbb{R}^{M} can capture the properties of the layer α\alpha, and the connectivity between two layers could be represented by 𝐀L=α,β=1Mbαβ𝐞α𝐞βM×M\mathbf{A}_{L}=\sum_{\alpha,\beta=1}^{M}b_{\alpha\beta}\mathbf{e}_{\alpha}\circ\mathbf{e}_{\beta}\in\mathbb{R}^{M\times M}. Following this approach, connectivity between the projected nodes of the entities in the layers can be represented by a fourth-order tensor

𝐀=α,β=1Mi,j=1Nwαiβj𝐞α𝐞i𝐞β𝐞jM×N×M×N,\mathbf{A}=\sum_{\alpha,\beta=1}^{M}\sum_{i,j=1}^{N}w_{\alpha i\beta j}\mathbf{e}_{\alpha}\circ\mathbf{e}_{i}\circ\mathbf{e}_{\beta}\circ\mathbf{e}_{j}\in\mathbb{R}^{M\times N\times M\times N}, (11)

where wαiβjw_{\alpha i\beta j} is the weight of connection between the entity ii’s projected node on layer α\alpha and the entity jj’s projected node on layer β\beta. More specially, the fourth-order tensor becomes the adjacency tensor of the multilayer graph, where each entry Aαiβj=wαiβjA_{\alpha i\beta j}=w_{\alpha i\beta j} characterizes the edge between the entity ii’s projected node on layer α\alpha and the entity jj’s projected node on layer β\beta. Thus, similar to the adjacency matrix whose 2-D entries indicate whether and how two nodes are pairwise connected by a simple edge in the normal graphs, we adopt an adjacency tensor 𝐀M×N×M×N\mathbf{A}\in\mathbb{R}^{M\times N\times M\times N} to represent the multilayer graph with the same number of nodes in each layer as follows.

Definition 2 (Adjacency Tensor).

A multilayer graph \mathcal{M}, with ||=M|\mathcal{L}|=M layers and |li|=N|l_{i}|=N nodes in each layer ii, can be represented by a fourth-order adjacency tensor 𝐀M×N×M×N\mathbf{A}\in\mathbb{R}^{M\times N\times M\times N} defined as

𝐀=(Aαiβj),1α,βM,1i,jN.\mathbf{A}=(A_{\alpha i\beta j}),\quad 1\leq\alpha,\beta\leq M,1\leq i,j\leq N. (12)

Here, each entry AαiβjA_{\alpha i\beta j} of the adjacency tensor 𝐀\mathbf{A} indicates the intensity of the edge between the entity jj’s projected node on layer β\beta and entity ii’s projected node on layer α\alpha.

Clearly, for a single layer graph, 𝐞α\mathbf{e}_{\alpha} is a scalar 11 and the fourth-order tensor degenerates to the adjacency matrix of a normal graph. Similar to AijA_{ij} in an adjacency matrix which indicates the direction from the node vjv_{j} to viv_{i}, AαiβjA_{\alpha i\beta j} also indicates the direction from the node vβjv_{\beta j} to the node vαiv_{\alpha i} in a MLG. Note that, vectors 𝐞i\mathbf{e}_{i} and 𝐞α\mathbf{e_{\alpha}} are not eigenvectors of the adjacency tensor. They are merely the vectors characterizing features of the entities and layers, respectively. We shall discuss the MLG-based spectrum space in Section V.

Given an adjacency tensor, we can define the Laplacian tensor of the multilayer graph similar to that in a single-layer graph. Denoting the degree (or multi-strength) of the entity’s ii’s projected node vαiv_{\alpha i} on layer α\alpha as d(vαi)d(v_{\alpha i}) which is a summation over weights of different natures (inter- and intra- layer edges), the degree tensor 𝐃M×N×M×N\mathbf{D}\in\mathbb{R}^{M\times N\times M\times N} is defined as a diagonal tensor with entries Dαiαi=d(vαi)D_{\alpha i\alpha i}=d(v_{\alpha i}) for 1iN,1αM1\leq i\leq N,1\leq\alpha\leq M, whereas its other entries are zero. The Laplacian tensor can be defined as follows.

Definition 3 (Laplacian Tensor).

A multilayer graph \mathcal{M}, with ||=M|\mathcal{L}|=M layers and |li|=N|l_{i}|=N nodes in each layer ii, can be represented by a fourth-order Laplacian tensor 𝐋M×N×M×N\mathbf{L}\in\mathbb{R}^{M\times N\times M\times N} defined as 𝐋=𝐃𝐀\mathbf{L=D-A}, where 𝐀\mathbf{A} is the adjacency tensor and 𝐃\mathbf{D} is the degree tensor.

The Laplacian tensor can be useful to analyze propagation processes such as diffusion or random walk [11]. Both adjacency and Laplacian tensors are important algebraic representations of the MLG depending on datasets and user objectives. For convenience, we use a tensor 𝐅M×N×M×N\mathbf{F}\in\mathbb{R}^{M\times N\times M\times N} as a general representation of a given MLG \mathcal{M}. As the adjacency tensor is more general, the representing tensor F refers to the adjacency tensor in this paper unless specified otherwise.

III-B2 Representation of General MLG

Representing a general multilayer graph with different number of nodes in each layer always remains a challenge if one aims to distinguish the interlayer and intralayer connection features. In JFT [16] and MWGSP [18], all layers must reside on the same underlying graph structure which restrict the number of nodes to be the same in each layer. Similarly, a reconstruction is also needed to represent a general MLG by the forth-order tensor in M-GSP. Note that, although M-GSP also needs a reconstruction to represent a general MLG, we allow different layers with heterogeneous graph structures, which provides additional flexibility than JFT and MWGSP.

There are mainly two ways to reconstruct: 1) Add isolated nodes to layers with fewer nodes to reach NN nodes [12] and set the augmented signals as zero; and 2) Aggregate several nodes into super-nodes for layers with |li|>N|l_{i}|>N [31] and merge the corresponding signals. Since isolated nodes do not interact with any other nodes, it does not change the topological structure of the original multilayer architecture in the sense of signal shifting while the corresponding spectrum space could still be changed. The aggregation method depends on how efficiently we can aggregate redundant or similar nodes. Different methods can be applied depending on specific tasks. For example, if one wants to explore the cascading failure in a physical system, the method based on isolated nodes is more suitable. For the applications, such as video analysis where pixels can be intuitively merged as superpixels, the aggregation method can be also practical.

In addition, although the fourth-order representing tensor can be viewed as the projection of several entities into different layers in Eq. (11), the entities and layers can be virtual and not necessarily physical to capture the underlying structures of the datasets. The information within the multilayer graphs, together with definitions of the underlying virtual entities and layers, should only depend on the structure of the multilayer graphs. We will illustrate this further in Section VII-B.

III-C Flattening and Analysis

In this part, we introduce the flattening of the multilayer graph, which could simplify some operations in the tensor-based M-GSP. For a multilayer graph ={𝒱,,𝐅}\mathcal{M}=\{\mathcal{V},\mathcal{L},\mathbf{F}\} with MM layers and NN nodes in each layer, its fourth-order representing tensor 𝐅M×N×M×N\mathbf{F}\in\mathbb{R}^{M\times N\times M\times N} can be flattened into a second-order matrix to capture the overall edge weights. There are two main flattening schemes in the sense of entities and layers, respectively:

  • Layer-wise Flattening: The representing tensor 𝐅\mathbf{F} can be flattened into 𝐅FLMN×MN\mathbf{F}_{FL}\in\mathbb{R}^{MN\times MN} with each element

    [FFL]N(α1)+i,N(β1)+j=Fαiβj.{[F_{FL}]}_{{N(\alpha-1)+i,N(\beta-1)+j}}=F_{\alpha i\beta j}. (13)
  • Entity-wise Flattening: The representing tensor 𝐅\mathbf{F} can be flattened into 𝐅FNNM×NM\mathbf{F}_{FN}\in\mathbb{R}^{NM\times NM} with each element

    [FFN]M(i1)+α,M(j1)+β=Fαiβj.{[F_{FN}]}_{{M(i-1)+\alpha,M(j-1)+\beta}}=F_{\alpha i\beta j}. (14)

These two flattening methods provide two ways to interpret the graph structure. In the first method, the flattened multilayer graph has MM clusters with NN nodes in each cluster. The nodes in the same cluster have the same function (belong to the same layer). In the second method, the flattened graph has NN clusters with MM nodes in each cluster. Here, the nodes in the same cluster are from the same entity. Examples of the tensor flattening of a two-layer graph with 33 nodes in each layer are shown in Fig. 4. From the examples, we can see that the diagonal blocks in N×N\mathbb{R}^{N\times N} are the intralayer connections for each layer and other blocks describe the interlayer connections through layer-wise flattening; and the diagonal block in M×M\mathbb{R}^{M\times M} describe the ‘intra-entity’ connections and other elements represent the ‘inter-entity’ connections in entity-wise flattening. Although these two flattening schemes define the same MLG with a different indexing of vertices, they are still helpful to analyze the MLG spectrum space.

Refer to caption
Figure 4: Example of Multilayer Graph Flattening.

IV Spatial Definitions in M-GSP

Based on the tensor representation, we now define signals and signal shifting over the multilayer graphs. In GSP, each signal sample is the attribute of one node. Typically, a graph signal can be represented by an NN-length vector for a graph with NN nodes. Recall that in traditional GSP [3], basic signal shifting is defined with the representing matrix as the shifting filter. Thus, in M-GSP, we can also define the signals and signal shifting based on the filter implementation.

In M-GSP, each signal sample is also related to one node in the multilayer graph. Intuitively, if there are K=MNK=MN nodes, there are MNMN signal samples in total. Similar to GSP, we use the representing (adjacency/Laplacian) tensor 𝐅M×N×M×N\mathbf{F}\in\mathbb{R}^{M\times N\times M\times N} as the basic MLG-filter. Since the input signal and the output signal of the MLG-filter should be consistent in the tensor size, we define a special form of M-GSP signals to work with the representing tensor as follows.

Definition 4 (Signals over Multilayer Graphs).

For a multilayer graph ={𝒱,,𝐅}\mathcal{M}=\{\mathcal{V},\mathcal{L},\mathbf{F}\}, with ||=M|\mathcal{L}|=M layers and |li|=N|l_{i}|=N nodes in each layer ii, the definition of multilayer graph signals is a second-order tensor

𝐬=(sαi)M×N,1αM,1iN,\mathbf{s}=(s_{\alpha i})\in\mathbb{R}^{M\times N},\quad 1\leq\alpha\leq M,1\leq i\leq N, (15)

where the entry sαis_{\alpha i} is the signal sample in the projected node of entity ii on layer α\alpha.

Note that, if the multilayer graph degenerates to a single-layer graph with M=1M=1, the multilayer graph signal becomes an NN-length vector, which is similar to that in the traditional GSP. Similar to the representing tensor, the tensor signal 𝐬M×N\mathbf{s}\in\mathbb{R}^{M\times N} can also be flattened as a vector in MN\mathbb{R}^{MN}:

  • Layer-wise flattening: 𝐬LMN\mathbf{s}_{L}\in\mathbb{R}^{MN} whose entries are calculated as [sL]N(α1)+i=sαi[s_{L}]_{N(\alpha-1)+i}=s_{\alpha i}.

  • Entity-wise flattening: 𝐬NNM\mathbf{s}_{N}\in\mathbb{R}^{NM} whose entries are calculated as [sN]M(i1)+α=sαi[s_{N}]_{M(i-1)+\alpha}=s_{\alpha i}.

Given the definitions of multilayer graph signals and filters, we now introduce the definitions of signal shifting in M-GSP. In traditional GSP, the signal shifting is defined as product between signal vectors and representing matrix. Similarly, we define the shifting in the multilayer graph based on the contraction (inner product) between the representing tensor and tensor signals.

Definition 5 (Signal Shifting over Multilayer Graphs).

Given the representing tensor 𝐅M×N×M×N\mathbf{F}\in\mathbb{R}^{M\times N\times M\times N} and the tensor signal 𝐬M×N\mathbf{s}\in\mathbb{R}^{M\times N} defined over a multilayer graph \mathcal{M}, the signal shifting is defined as the contraction (inner product) between 𝐅\mathbf{F} and 𝐬\mathbf{s} in one entity-related order and one layer-related order, i.e.,

𝐬=𝐅𝐬M×N,\mathbf{s}^{\prime}=\mathbf{F}\diamond\mathbf{s}\in\mathbb{R}^{M\times N}, (16)

where \diamond is the contraction between 𝐅\mathbf{F} and 𝐬\mathbf{s} defined in Eq. (7).

The elements in the shifted signal 𝐬\mathbf{s}^{\prime} are calculated as

sαi=β=1Mj=1NFαiβjsβj.s^{\prime}_{\alpha i}=\sum_{\beta=1}^{M}\sum_{j=1}^{N}F_{\alpha i\beta j}s_{\beta j}. (17)
Refer to caption
Figure 5: Example of Multilayer Graph Shifting.

From Eq. (17), there are two important factors to construct the shifted signal: 1) The signal in the neighbors (both intra- and inter- layer interactions) of the node vαiv_{\alpha i}; and 2) The intensity of interactions between the node vαiv_{\alpha i} and its neighbors. Then, the signal shifting is related to the diffusion process over the multilayer graphs. More specifically, if 𝐅\mathbf{F} is the adjacency tensor, signals shift in directions of edges. To better illustrate the signal shifting based on the adjacency tensor, we use a two-layer directed network shown in Fig. 5 as an example. In this multilayer graph, the original signal 𝐬\mathbf{s} is defined as

𝐬=[123654],\mathbf{s}=\begin{bmatrix}1&2&3\\ 6&5&4\end{bmatrix}, (18)

and the adjacency tensor 𝐀\mathbf{A} is defined as

𝐀(1,:,1,:)=[000100010],𝐀(2,:,2,:)=[010001000],\mathbf{A}_{(1,:,1,:)}=\begin{bmatrix}0&0&0\\ 1&0&0\\ 0&1&0\end{bmatrix},\mathbf{A}_{(2,:,2,:)}=\begin{bmatrix}0&1&0\\ 0&0&1\\ 0&0&0\end{bmatrix},
𝐀(2,:,1,:)=[000000001],𝐀(1,:,2,:)=[100000000],\mathbf{A}_{(2,:,1,:)}=\begin{bmatrix}0&0&0\\ 0&0&0\\ 0&0&1\end{bmatrix},\mathbf{A}_{(1,:,2,:)}=\begin{bmatrix}1&0&0\\ 0&0&0\\ 0&0&0\end{bmatrix},

for each fiber. Then, the shifted signal is calculated as

𝐬=[612543].\mathbf{s}^{\prime}=\begin{bmatrix}6&1&2\\ 5&4&3\end{bmatrix}. (19)

From Eq. (19), we can see that the signal shift one step following the direction of the links.

Meanwhile, if 𝐅\mathbf{F} is the Laplacian tensor, Eq. (17) can be written as

sαi=β=1Mj=1NAαiβj(sαisβj),s^{\prime}_{\alpha i}=\sum_{\beta=1}^{M}\sum_{j=1}^{N}A_{\alpha i\beta j}(s_{\alpha i}-s_{\beta j}), (20)

which is the weighted average of difference with neighbors.

V Multilayer Graph spectrum space

In traditional GSP, graph spectrum space is defined according to the eigenspace of the representing matrix [3]. Similarly, we define the MLG spectrum space based on the decomposition of the representing tensor. Since tensor decomposition is less stable when exploring the factorization of a specific order or when extracting the separate features in the asymmetric tensors, we will mainly focus on spectral properties of undirected multilayer graphs in this section for simplicity and clarity of presentation. For directed MLG, we provide alternative spectrum definitions in the Appendix and will explore detailed analysis in future works.

V-A Joint Spectral Analysis in M-GSP

For a multilayer graph ={𝒱,,𝐅}\mathcal{M}=\{\mathcal{V},\mathcal{L},\mathbf{F}\} with MM layers and NN nodes, the eigen-tensor 𝐕M×N\mathbf{V}\in\mathbb{R}^{M\times N} of the representing tensor 𝐅\mathbf{F} is defined in the tensor-based multilayer graph theory [11] as

𝐅𝐕=λ𝐕.\mathbf{F}\diamond\mathbf{V}=\lambda\mathbf{V}. (21)

More specifically, 𝐅M×N×M×N\mathbf{F}\in\mathbb{R}^{M\times N\times M\times N} can be decomposed as

𝐅\displaystyle\mathbf{F} =k=1MNλk𝐕k𝐕k\displaystyle=\sum_{k=1}^{MN}\lambda_{k}\mathbf{V}_{k}\circ\mathbf{V}_{k} (22)
=α=1Mi=1Nλαi𝐕αi𝐕αi,\displaystyle=\sum_{\alpha=1}^{M}\sum_{i=1}^{N}\lambda_{\alpha i}\mathbf{V}_{\alpha i}\circ\mathbf{V}_{\alpha i}, (23)

where λk\lambda_{k} is the eigenvalues and 𝐕kM×N\mathbf{V}_{k}\in\mathbb{R}^{M\times N} is the corresponding eigen-tensor. Note that 𝐕αi\mathbf{V}_{\alpha i} just relabels the index of 𝐕k\mathbf{V}_{k}, and there is no specific order for 𝐕αi\mathbf{V}_{\alpha i} here.

Similar to the traditional GSP where the graph Fourier space is defined by the eigenvectors of the representing matrix, we define the joint MLG Fourier space as follows.

Definition 6 (Joint Multilayer Graph Fourier Space).

For a multilayer graph ={𝒱,,𝐅}\mathcal{M}=\{\mathcal{V},\mathcal{L},\mathbf{F}\} with MM layers and NN nodes, the MLG Fourier space is defined as the space consisting of all spectral tensors {𝐕1,,𝐕MN}\{\mathbf{V}_{1},\cdots,\mathbf{V}_{MN}\}, which characterizes the joint features of entities and layers.

Recall that in GSP, the GFT is defined based on the inner product of 𝐕1\mathbf{V}^{-1} and the signals 𝐬\mathbf{s} defined in Eq. (2). Similarly, we can define the M-GFT based on the spectral tensors of the representing tensor 𝐅\mathbf{F} to capture joint features of inter- and intra- layer interactions as follows.

Definition 7 (Joint M-GFT).

Let 𝐔=(𝐕αi)M×N×M×N\mathbf{U}_{\mathcal{F}}=(\mathbf{V}_{\alpha i})\in\mathbb{R}^{M\times N\times M\times N} consist of spectral tensors of the representing tensor 𝐅\mathbf{F}, where [U]αiβj=[Vαi]βj[U_{\mathcal{F}}]_{\alpha i\beta j}=[V_{\alpha i}]_{\beta j}.

The joint M-GFT (M-JGFT) can be defined as the contraction between 𝐔\mathbf{U}_{\mathcal{F}} and the tensor signal 𝐬M×N\mathbf{s}\in\mathbb{R}^{M\times N}, i.e.,

𝐬^=𝐔𝐬.\hat{\mathbf{s}}=\mathbf{U}_{\mathcal{F}}\diamond\mathbf{s}. (24)

Now, we show how to obtain the eigen-tensors. Implementing the flattening analysis, we have the following properties.

Property 1.

The two types of flattened tensor in Eq. (13) and Eq. (14) lead to the same eigenvalues.

Proof.

Suppose (λ,𝐱)(\lambda,\mathbf{x}) is an eigenpair of 𝐀FL\mathbf{A}_{FL}, i.e.,

𝐀FL𝐱=λ𝐱.\mathbf{A}_{FL}\cdot\mathbf{x}=\lambda\mathbf{x}. (25)

Let xN(α1)+i=yM(i1)+αx_{N(\alpha-1)+i}=y_{M(i-1)+\alpha}. Since

Aαiβj\displaystyle A_{\alpha i\beta j} =[AFL]N(α1)+i,N(β1)+j\displaystyle=[A_{FL}]_{N(\alpha-1)+i,N(\beta-1)+j}
=[AFN]M(i1)+α,M(j1)+β,\displaystyle=[A_{FN}]_{M(i-1)+\alpha,M(j-1)+\beta}, (26)

we have

𝐀FN𝐲=λ𝐲.\mathbf{A}_{FN}\cdot\mathbf{y}=\lambda\mathbf{y}. (27)

Thus, λ\lambda is also an eigenvalue of 𝐀FN\mathbf{A}_{FN}. ∎

This property shows that the flattened tensors are the reshaped original representing tensor, and could capture some of the spectral properties as follows.

Property 2.

Given the eigenpair (λFL,𝐱)(\lambda_{FL},\mathbf{x}) of the layer-wise flattened tensors, the eigenpair (λ,𝐕)(\lambda,\mathbf{V}) of the original representing tensor can be calculated as λ=λFL\lambda=\lambda_{FL}, and Vαi=xN(α1)+iV_{\alpha i}=x_{N(\alpha-1)+i}. Similarly, given the eigenpair (λFN,𝐲)(\lambda_{FN},\mathbf{y}) of the entity-wise flattened tensor, the eigenpair (λ,𝐕)(\lambda,\mathbf{V}) of the original representing tensor can be calculated as λ=λFN\lambda=\lambda_{FN}, and Vαi=yM(i1)+αV_{\alpha i}=y_{M(i-1)+\alpha}.

The Property 2 shows that we can calculate the eigen-tensor from the flattened tensor to simplify the decomposition operations. Moreover, the M-JGFT is the bijection of GFT in the flattened MLG, with vertices indexed by both the layers and the entities. However, such M-JGFT analyzes the inter- and intra- layer connections jointly while ignoring the individual features of entities and layers. Next, we will show how to implement the order-wise frequency analysis in M-GSP based on tensor decomposition.

V-B Order-wise Spectral Analysis in M-GSP

In an undirected multilayer graph, the representing tensor (adjacency/Laplacian) 𝐅\mathbf{F} is partially symmetric between orders one and three, and between orders two and four, respectively. Then, the representing tensor can be written with the consideration of the multilayer graph structure under orthogonal CP-decomposition [26] as follows:

𝐅\displaystyle\mathbf{F} α=1Mi=1Nλαi𝐟α𝐞i𝐟α𝐞i\displaystyle\approx\sum_{\alpha=1}^{M}\sum_{i=1}^{N}\lambda_{\alpha i}\cdot\mathbf{f}_{\alpha}\circ\mathbf{e}_{i}\circ\mathbf{f}_{\alpha}\circ\mathbf{e}_{i} (28)
=α=1Mi=1Nλαi𝐕~αi𝐕~αi,\displaystyle=\sum_{\alpha=1}^{M}\sum_{i=1}^{N}\lambda_{\alpha i}\tilde{\mathbf{V}}_{\alpha i}\circ\tilde{\mathbf{V}}_{\alpha i}, (29)

where 𝐟αM\mathbf{f}_{\alpha}\in\mathbb{R}^{M} are orthonormal, 𝐞iN\mathbf{e}_{i}\in\mathbb{R}^{N} are orthonormal and 𝐕~αi=𝐟α𝐞iM×N\tilde{\mathbf{V}}_{\alpha i}=\mathbf{f}_{\alpha}\circ\mathbf{e}_{i}\in\mathbb{R}^{M\times N}.

The CP decomposition factorizes a tensor into a sum of component rank-one tensors, which describe the underlying features of each order. Although approximated algorithms are implemented to obtain the optimal decomposition, CP decomposition still achieves great success in real scenarios, such as feature extraction [32] and tensor-based PCA analysis [33]. A detailed discussion of tensor decomposition and its implementation in M-GSP are provided in Section VII-D. In Eq. (29), 𝐟α\mathbf{f}_{\alpha} and 𝐞i\mathbf{e}_{i} capture the features of layers and entities, respectively, which can be interpreted as the subspaces of the MLG. More discussions about the frequency interpretation of order-wise M-GSP spectrum and connections to MWGSP spectrum are presented in Section VII-C.

Note that, if there is only one layer in the multilayer graph, Eq. (29) reduces to the eigen-decomposition of a normal single-layer graph, i.e., 𝐅=i=1Nλi𝐞i𝐞i\mathbf{F}=\sum_{i=1}^{N}\lambda_{i}\mathbf{e}_{i}\circ\mathbf{e}_{i}

With the decomposed representing tensor in Eq. (29), the order-wise MLG spectrum is defined as follows.

Definition 8 (Order-wise MLG Spectral Pair).

For a multilayer graph ={𝒱,,𝐅}\mathcal{M}=\{\mathcal{V},\mathcal{L},\mathbf{F}\} with MM layers and NN nodes, the order-wise MLG spectral pairs are defined by {λαi,𝐟α,𝐞i}\{\lambda_{\alpha i},\mathbf{f}_{\alpha},\mathbf{e}_{i}\}, where {𝐟1,,𝐟M}\{\mathbf{f}_{1},\cdots,\mathbf{f}_{M}\} and {𝐞1,,𝐞N}\{\mathbf{e}_{1},\cdots,\mathbf{e}_{N}\} characterize features of layers and entities, respectively.

With the definition of order-wise MLG spectral pair, we now explore their properties. Considering 𝐕~αi=𝐟α𝐞i\tilde{\mathbf{V}}_{\alpha i}=\mathbf{f}_{\alpha}\circ\mathbf{e}_{i}, we have the following property, which indicates the availability of a joint MLG analysis based on order-wise spectrum.

Property 3.

The factor tensor 𝐕~αi\tilde{\mathbf{V}}_{\alpha i} of the representing tensor 𝐅\mathbf{F} is the approximated eigen-tensor of 𝐅\mathbf{F}.

Proof.

Suppose that 𝐕~αi\tilde{\mathbf{V}}_{\alpha i} is one factor tensor of 𝐅\mathbf{F} obtained from Eq. (29).

Let δ[k]\delta[k] denote the Kronecker delta. Since 𝐟α\mathbf{f}_{\alpha} forms an orthonormal basis, then the inner product would satisfy

<𝐟α,𝐟β>=k[fα]k[fβ]k=δ[αβ].<\mathbf{f}_{\alpha},\mathbf{f}_{\beta}>=\sum_{k}[f_{\alpha}]_{k}\cdot[f_{\beta}]_{k}=\delta[\alpha-\beta].

Similarly,

<𝐞i,𝐞j>=δ[ij].<\mathbf{e}_{i},\mathbf{e}_{j}>=\delta[i-j].
[𝐅𝐕~αi]βj=\displaystyle[\mathbf{F}\diamond\tilde{\mathbf{V}}_{\alpha i}]_{\beta j}= σ=1Mk=1NFβjσk[V~αi]σk\displaystyle\sum_{\sigma=1}^{M}\sum_{k=1}^{N}F_{\beta j\sigma k}\;[\tilde{V}_{\alpha i}]_{\sigma k} (30)
\displaystyle\approx σ=1Mk=1Nγ=1Mt=1Nλγt\displaystyle\sum_{\sigma=1}^{M}\sum_{k=1}^{N}\sum_{\gamma=1}^{M}\sum_{t=1}^{N}\lambda_{\gamma t}\cdot (31)
[fγ]β[et]j[fγ]σ[et]k[V~αi]σk.\displaystyle[f_{\gamma}]_{\beta}\;[e_{t}]_{j}\;[f_{\gamma}]_{\sigma}\;[e_{t}]_{k}\;[\tilde{V}_{\alpha i}]_{\sigma k}. (32)

Then, we have

σ=1Mk=1N[fγ]σ[et]k[V~αi]σk\displaystyle\sum_{\sigma=1}^{M}\sum_{k=1}^{N}[f_{\gamma}]_{\sigma}\;[e_{t}]_{k}\;[\tilde{V}_{\alpha i}]_{\sigma k} =σ=1Mk=1N[fγ]σ[et]k[fα]σ[ei]k\displaystyle=\sum_{\sigma=1}^{M}\sum_{k=1}^{N}[f_{\gamma}]_{\sigma}\;[e_{t}]_{k}\;[f_{\alpha}]_{\sigma}\;[e_{i}]_{k}
=σ=1M[fγ]σ[fα]σk=1N[et]k[ei]k\displaystyle=\sum_{\sigma=1}^{M}[f_{\gamma}]_{\sigma}\;[f_{\alpha}]_{\sigma}\;\sum_{k=1}^{N}[e_{t}]_{k}\;[e_{i}]_{k}
=<𝐟γ,𝐟α><𝐞t,𝐞i>\displaystyle=<\mathbf{f}_{\gamma},\mathbf{f}_{\alpha}>\cdot<\mathbf{e}_{t},\mathbf{e}_{i}>
=δ[γα]δ[ti]\displaystyle=\delta[\gamma-\alpha]\delta[t-i] (33)

Thus,

[𝐅𝐕~αi]βjλαi[fα]β[ei]j,[\mathbf{F}\diamond\tilde{\mathbf{V}}_{\alpha i}]_{\beta j}\approx\lambda_{\alpha i}[f_{\alpha}]_{\beta}[e_{i}]_{j}, (34)

which indicates

𝐅𝐕~αiλαi𝐕~αi.\mathbf{F}\diamond\tilde{\mathbf{V}}_{\alpha i}\approx\lambda_{\alpha i}\tilde{\mathbf{V}}_{\alpha i}. (35)

Then, 𝐕~αi\tilde{\mathbf{V}}_{\alpha i} is the approximated eigen-tensor. ∎

This property indicates the relationship between the order-wise MLG spectral pair and the joint eigen-tensors.

By constructing a fourth-order tensor 𝐔~M×N×M×N\tilde{\mathbf{U}}_{\mathcal{F}}\in\mathbb{R}^{M\times N\times M\times N} with 𝐕~αi\tilde{\mathbf{V}}_{\alpha i} as its elements, i.e., [U~]αiβj=[V~αi]βj[\tilde{U}_{\mathcal{F}}]_{\alpha i\beta j}=[\tilde{V}_{\alpha i}]_{\beta j}, we can have the following property.

Property 4.

Let 𝐖=𝐔~𝐔~\mathbf{W}=\tilde{\mathbf{U}}_{\mathcal{F}}\otimes\tilde{\mathbf{U}}_{\mathcal{F}}, where \otimes is the contraction in the third and forth order with Wαiβj=p,θ[U~]βjθp×[U~]αiθpW_{\alpha i\beta j}=\sum_{p,\theta}[\tilde{U}_{\mathcal{F}}]_{\beta j\theta p}\times[\tilde{U}_{\mathcal{F}}]_{\alpha i\theta p}. Then, 𝐖\mathbf{W} is super-diagonal with super-diagonal elements all equal to one.

Proof.

Let 𝐕~k=𝐕~αi=𝐟α𝐞i\tilde{\mathbf{V}}_{k}=\tilde{\mathbf{V}}_{\alpha i}=\mathbf{f}_{\alpha}\circ\mathbf{e}_{i} and 𝐕~t=𝐕~βj=𝐟β𝐞j\tilde{\mathbf{V}}_{t}=\tilde{\mathbf{V}}_{\beta j}=\mathbf{f}_{\beta}\circ\mathbf{e}_{j}. Then, we have

Wαiβj\displaystyle W_{\alpha i\beta j} =θ,p[V~k]θp[V~t]θp\displaystyle=\sum_{\theta,p}[\tilde{V}_{k}]_{\theta p}[\tilde{V}_{t}]_{\theta p}
=p[ei]p[ej]pθ[fα]θ[fβ]θ\displaystyle=\sum_{p}[e_{i}]_{p}[e_{j}]_{p}\sum_{\theta}[f_{\alpha}]_{\theta}[f_{\beta}]_{\theta}
=<𝐟α,𝐟β><𝐞i,𝐞j>,\displaystyle=<\mathbf{f}_{\alpha},\mathbf{f}_{\beta}>\cdot<\mathbf{e}_{i},\mathbf{e}_{j}>,
=δ[αβ]δ[ij].\displaystyle=\delta[\alpha-\beta]\delta[i-j]. (36)

This property generalizes the orthogonality of the spectral tensor into a similar definition of matrix.

We now introduce the order-wise MLG spectral transform. Similar to Eq. (24), the joint transform can be defined as

𝐬^=𝐔~𝐬.\hat{\mathbf{s}}=\tilde{\mathbf{U}}_{\mathcal{F}}\diamond\mathbf{s}. (37)

Note that each element of 𝐬^\hat{\mathbf{s}} in Eq. (37) can be calculated as

s^αi\displaystyle\hat{s}_{\alpha i} =β,j[U~]αiβjsβj\displaystyle=\sum_{\beta,j}[\tilde{U}_{\mathcal{F}}]_{\alpha i\beta j}s_{\beta j} (38)
=β,j[V~αi]βjsβj\displaystyle=\sum_{\beta,j}[\tilde{V}_{\alpha i}]_{\beta j}s_{\beta j} (39)
=β,j[fα]β[ei]jsβj.\displaystyle=\sum_{\beta,j}[f_{\alpha}]_{\beta}\cdot[e_{i}]_{j}\cdot s_{\beta j}. (40)

Let 𝐄f=[𝐟1𝐟M]M×M\mathbf{E}_{f}=[\mathbf{f}_{1}\cdots\mathbf{f}_{M}]\in\mathbb{R}^{M\times M} and 𝐄e=[𝐞1𝐞N]N×N\mathbf{E}_{e}=[\mathbf{e}_{1}\cdots\mathbf{e}_{N}]\in\mathbb{R}^{N\times N}. We then have

𝐬^=𝐄fT𝐬𝐄e,\hat{\mathbf{s}}^{\prime}=\mathbf{E}_{f}^{\mathrm{T}}\mathbf{s}\mathbf{E}_{e}, (41)

with each element

s^α=j,β[fα]β[ei]jsβj\hat{s}_{\alpha}^{\prime}=\sum_{j,\beta}[f_{\alpha}]_{\beta}\cdot[e_{i}]_{j}\cdot s_{\beta j} (42)

Clearly, the M-GFT can be obtained as

𝐬^=𝐬^=𝐄fT𝐬𝐄e.\hat{\mathbf{s}}=\hat{\mathbf{s}}^{\prime}=\mathbf{E}_{f}^{\mathrm{T}}\mathbf{s}\mathbf{E}_{e}. (43)

Then, we have the following definition of M-GFT based on order-wise spectrum.

Definition 9 (Order-wise M-GFT).

Given the spectral vectors 𝐄f=[𝐟1𝐟M]M×M\mathbf{E}_{f}=[\mathbf{f}_{1}\cdots\mathbf{f}_{M}]\in\mathbb{R}^{M\times M} and 𝐄e=[𝐞1𝐞N]N×N\mathbf{E}_{e}=[\mathbf{e}_{1}\cdots\mathbf{e}_{N}]\in\mathbb{R}^{N\times N}, the layer-wise M-GFT can be defined as

𝐬^L=𝐄fT𝐬M×N,\hat{\mathbf{s}}_{L}=\mathbf{E}_{f}^{\mathrm{T}}\mathbf{s}\in\mathbb{R}^{M\times N}, (44)

and the entity-wise M-GFT can be defined as

𝐬^N=𝐬𝐄eM×N.\hat{\mathbf{s}}_{N}=\mathbf{s}\mathbf{E}_{e}\in\mathbb{R}^{M\times N}. (45)

The general M-GFT based on order-wise spectrum is defined as

𝐬^=𝐄fT𝐬𝐄eM×N.\hat{\mathbf{s}}=\mathbf{E}_{f}^{\mathrm{T}}\mathbf{s}\mathbf{E}_{e}\in\mathbb{R}^{M\times N}. (46)

If there is only one layer in the multilayer graph, the M-GFT calculated with 𝐬TN\mathbf{s}^{T}\in\mathbb{R}^{N} as (𝐬^N)T=(𝐬𝐄e)TN(\hat{\mathbf{s}}_{N})^{\mathrm{T}}=(\mathbf{s}\mathbf{E}_{e})^{\mathrm{T}}\in\mathbb{R}^{N}, which has the same form as the traditional GFT in Eq. (2).

In addition, since 𝐟α\mathbf{f}_{\alpha} and 𝐞i\mathbf{e}_{i} are orthonormal basis of undirected MLG, the inverse M-GFT can be calculated as

𝐬=𝐄f𝐬^𝐄eT.\mathbf{s}^{\prime}=\mathbf{E}_{f}\hat{\mathbf{s}}\mathbf{E}_{e}^{\mathrm{T}}. (47)

Different from joint MLG Fourier space in Section V-A, the order-wise MLG spectrum provides an individual analysis on layers and entities separately, and a reliable approximated analysis on the underlying MLG structures jointly. To minimize confusion, we abbreviate joint M-GFT in Eq. (24) as M-JGFT. M-GFT refers to the order-wise transform in Eq. (46) in the remaining parts if there is no specification.

V-C MLG Singular Tensor Analysis

In addition to the eigen-decomposition, the singular value decomposition (SVD) is another important decomposition to factorize a matrix. In this part, we provide the higher-order SVD (HOSVD) [25] of the representing tensor as an alternative definition of spectrum for the multilayer graphs.

Given the multilayer graph ={𝒱,,𝐅}\mathcal{M}=\{\mathcal{V},\mathcal{L},\mathbf{F}\} with MM layers and NN nodes in each layer, its representing tensor 𝐅M×N×M×N\mathbf{F}\in\mathbb{R}^{M\times N\times M\times N} can be decomposed via HOSVD as

𝐅=𝐒×1𝐔(1)×2𝐔(2)×3𝐔(3)×4𝐔(4),\mathbf{F}=\mathbf{S}\times_{1}\mathbf{U}^{(1)}\times_{2}\mathbf{U}^{(2)}\times_{3}\mathbf{U}^{(3)}\times_{4}\mathbf{U}^{(4)}, (48)

where 𝐔(n)=[𝐔1(n)𝐔2(n)𝐔In(n)]\mathbf{U}^{(n)}=[\mathbf{U}^{(n)}_{1}\quad\mathbf{U}^{(n)}_{2}\quad\cdots\quad\mathbf{U}^{(n)}_{I_{n}}] is a unitary (In×In)(I_{n}\times I_{n}) matrix, with I1=I3=MI_{1}=I_{3}=M and I2=I4=NI_{2}=I_{4}=N. 𝐒\mathbf{S} is a complex (I1×I2×I3×I4)(I_{1}\times I_{2}\times I_{3}\times I_{4})-tensor of which the subtensor 𝐒in\mathbf{S}_{i_{n}} obtained by fixing nnth index to α\alpha have

  • <𝐒in=α,𝐒in=β>=0<\mathbf{S}_{i_{n}=\alpha},\mathbf{S}_{i_{n}=\beta}>=0 where αβ\alpha\neq\beta.

  • 𝐒in=1𝐒in=2𝐒in=In0||\mathbf{S}_{i_{n}=1}||\geq||\mathbf{S}_{i_{n}=2}||\geq\cdots\geq||\mathbf{S}_{i_{n}=I_{n}}||\geq 0.

The Frobenius-norms σi(n)=𝐒in=i\sigma_{i}^{(n)}=||\mathbf{S}_{i_{n}=i}|| is the nn-mode singular value, and 𝐔(i)\mathbf{U}^{(i)} are the corresponding nn-mode singular vectors. For an undirected multilayer graph, the representing tensor is symmetric for every 2-D combination. Thus, there are two modes of singular spectrum, i.e., (γα,𝐟α)(\gamma_{\alpha},\mathbf{f}_{\alpha}) for mode 1,31,3, and (σi,𝐞i)(\sigma_{i},\mathbf{e}_{i}) for mode 2,42,4. More specifically, 𝐔(1)=𝐔(3)=(𝐟α)\mathbf{U}^{(1)}=\mathbf{U}^{(3)}=(\mathbf{f}_{\alpha}) and 𝐔(2)=𝐔(4)=(𝐞i)\mathbf{U}^{(2)}=\mathbf{U}^{(4)}=(\mathbf{e}_{i}). Since the joint singular tensor captures the consistent information of entities and layers, it can be calculated as

(λαi,𝐕^αi)=(γασi,𝐟α𝐞i).(\lambda_{\alpha i},\hat{\mathbf{V}}_{\alpha i})=(\gamma_{\alpha}\cdot\sigma_{i},\mathbf{f}_{\alpha}\circ\mathbf{e}_{i}). (49)

Note that the diagonal entries of 𝐒\mathbf{S} are not the eigenvalues or frequency coefficients of the representing tensor in general. The multilayer graph singular space is defined as follows.

Definition 10 (Multilayer Graph Singular Space).

For a multilayer graph ={𝒱,,𝐅}\mathcal{M}=\{\mathcal{V},\mathcal{L},\mathbf{F}\} with MM layers and NN nodes, the MLG singular space is defined as the space consisting of all singular tensors {𝐕^1𝐕^MN}\{\hat{\mathbf{V}}_{1}\cdots\hat{\mathbf{V}}_{MN}\} obtained from Eq. (49). The singular vectors {𝐟1,,𝐟M}\{\mathbf{f}_{1},\cdots,\mathbf{f}_{M}\} and {𝐞1,,𝐞N}\{\mathbf{e}_{1},\cdots,\mathbf{e}_{N}\} in Eq. (48) characterize layers and entities, respectively.

Similar to order-wise spectral analysis in Section V-B, we can define the MLG singular tensor transform (M-GST) based on the singular tensors as follows.

Definition 11 (M-GST).

Suppose that 𝐔s=(𝐟α𝐞i)M×N×M×N\mathbf{U}_{s}=(\mathbf{f}_{\alpha}\circ\mathbf{e}_{i})\in\mathbb{R}^{M\times N\times M\times N} consists of the singular vectors of the representing tensor 𝐅\mathbf{F} in Eq. (48), where [Us]αiβj=[fα]β[ei]j[U_{s}]_{\alpha i\beta j}=[f_{\alpha}]_{\beta}\cdot[e_{i}]_{j}. The M-GST can be defined as the contraction between 𝐔s\mathbf{U}_{s} and the tensor signal 𝐬M×N\mathbf{s}\in\mathbb{R}^{M\times N}, i.e.,

𝐬ˇ=𝐔s𝐬.\check{\mathbf{s}}=\mathbf{U}_{s}\diamond\mathbf{s}. (50)

If the singular vectors are included in 𝐖f=[𝐟1𝐟M]M×M\mathbf{W}_{f}=[\mathbf{f}_{1}\cdots\mathbf{f}_{M}]\in\mathbb{R}^{M\times M} and 𝐖e=[𝐞1𝐞N]N×N\mathbf{W}_{e}=[\mathbf{e}_{1}\cdots\mathbf{e}_{N}]\in\mathbb{R}^{N\times N}, the layer-wise M-GST can be defined as

𝐬ˇL=𝐖fT𝐬M×N,\check{\mathbf{s}}_{L}=\mathbf{W}_{f}^{\mathrm{T}}\mathbf{s}\in\mathbb{R}^{M\times N}, (51)

and the entity-wise M-GST can be defined as

𝐬ˇN=𝐬𝐖eM×N.\check{\mathbf{s}}_{N}=\mathbf{s}\mathbf{W}_{e}\in\mathbb{R}^{M\times N}. (52)

Inverse M-GST can be defined similarly as in Eq. (47) with unitary 𝐖e\mathbf{W}_{e} and 𝐖f\mathbf{W}_{f}.

Compared to the eigen-tensors in Eq. (22), the singular tensors come from the combinations of the singular vectors, thus are capable of capturing information of layers and entities more efficiently. Eigen-decomposition, however, focuses more on the joint information and approximate the separate information of layers and entities. We shall provide further discussion on the performance of different decomposition methods in Section VII-D. The intuition of applying HOSVD in MLG analysis and its correlations to GSP are also provided in Section VII-A3.

V-D Spectrum Ranking in the Multilayer Graph

In traditional GSP, the frequencies are defined by the eigenvalues of the shift, whereas the total variation is an alternative measurement of the order of the graph frequencies [3]. Similarly, we use the total variation of λαi\lambda_{\alpha i} based on the spectral tensors to rank the MLG frequencies. Let |λ|max|\lambda|_{max} be the joint eigenvalue of the adjacency tensor 𝐀\mathbf{A} with the largest magnitude. The M-GSP total variation is defined as follows:

TV(𝐕αi)\displaystyle TV(\mathbf{V}_{\alpha i}) =𝐕αi1|λ|max𝐀𝐕αi1\displaystyle=||\mathbf{V}_{\alpha i}-\frac{1}{|\lambda|_{max}}\mathbf{A}\diamond\mathbf{V}_{\alpha i}||_{1} (53)
=|1λ|λ|max|𝐕αi1,\displaystyle=|1-\frac{\lambda}{|\lambda|_{max}}\mathbf{|}\cdot||\mathbf{V}_{\alpha i}||_{1}, (54)

where ||||1||\cdot||_{1} is the element-wise l1l_{1} norm. Other norms could also be used to define the total variation. For example, the l2l_{2} norm could be efficient in signal denoising [3]. The graph frequency related to λαi\lambda_{\alpha i} is said to be a higher frequency if its total variation TV(𝐕αi)TV(\mathbf{V}_{\alpha i}) is larger, and its corresponding spectral tensor 𝐕αi\mathbf{V}_{\alpha i} is a higher frequency spectrum. If the representation tensor refers to Laplacian tensor, i.e., 𝐋=𝐃𝐀\mathbf{L=D-A}, the frequency order is in contract to the adjacency tensor 𝐀\mathbf{A} as GSP [3]. We shall provide more details on interpretation of MLG frequency in Section VII-A.

VI Filter Design

In this section, we introduce an MLG filter design together with its properties based on signal shifting.

VI-A Polynomial Filter Design

Polynomial filters are basic filters in GSP [7, 34]. In M-GSP, first-order filtering consists of basic signal filtering, i.e.,

𝐬=f1(𝐬)=𝐅𝐬.\mathbf{s}^{\prime}=f_{1}(\mathbf{s})=\mathbf{F}\diamond\mathbf{s}. (55)

Similarly, a second order filter can be defined as additional filtering on first-order filtered signal, i.e.,

𝐬′′\displaystyle\mathbf{s}^{\prime\prime} =f2(𝐬)\displaystyle=f_{2}(\mathbf{s}) (56)
=𝐅(𝐅𝐬),\displaystyle=\mathbf{F}\diamond(\mathbf{F}\diamond\mathbf{s}), (57)

whose entries sαi′′s_{\alpha i}^{\prime\prime} are calculated as

sαi′′\displaystyle s_{\alpha i}^{\prime\prime} =β=1Mj=1NFαiβjsβj\displaystyle=\sum_{\beta=1}^{M}\sum_{j=1}^{N}F_{\alpha i\beta j}s_{\beta j}^{\prime} (58)
=β,jFαiβjϵ,pFβjϵpsϵp\displaystyle=\sum_{\beta,j}F_{\alpha i\beta j}\sum_{\epsilon,p}F_{\beta j\epsilon p}s_{\epsilon p} (59)
=ϵ,psϵpβ,jFαiβjFβjϵp\displaystyle=\sum_{\epsilon,p}s_{\epsilon p}\sum_{\beta,j}F_{\alpha i\beta j}F_{\beta j\epsilon p} (60)
=(𝐅𝐅)𝐬,\displaystyle=(\mathbf{F}\odot\mathbf{F})\diamond\mathbf{s}, (61)

where \odot is the contraction defined in Eq. (8).

Let 𝐅[2]=𝐅𝐅\mathbf{F}^{[2]}=\mathbf{F}\odot\mathbf{F}. From Eq. (22), we have:

Fαiβj[2]\displaystyle F^{[2]}_{\alpha i\beta j} =θ,pFαiθpFθpβj\displaystyle=\sum_{\theta,p}F_{\alpha i\theta p}F_{\theta p\beta j} (62)
=θ,p(kλk[Vk]αi[Vk]θp)(tλt[Vt]βj[Vt]θp)\displaystyle=\sum_{\theta,p}(\sum_{k}\lambda_{k}[{V}_{k}]_{\alpha i}[V_{k}]_{\theta p})(\sum_{t}\lambda_{t}[V_{t}]_{\beta j}[V_{t}]_{\theta p})
=k,tλkλt[Vk]αi[Vt]βj(θ,p[Vt]θp[Vk]θp)\displaystyle=\sum_{k,t}\lambda_{k}\lambda_{t}[{V}_{k}]_{\alpha i}[{V}_{t}]_{\beta j}(\sum_{\theta,p}[{V}_{t}]_{\theta p}[{V}_{k}]_{\theta p})
=kλk2[Vk]αi[Vk]βj.\displaystyle=\sum_{k}\lambda_{k}^{2}[{V}_{k}]_{\alpha i}[{V}_{k}]_{\beta j}. (63)

Similarly, for τ\tauth-order term 𝐅[τ]\mathbf{F}^{[\tau]}, its entry Fαiβj[τ]F^{[\tau]}_{\alpha i\beta j} can be calculated as Fαiβj[τ]=kλkτ[𝐕k]αi[𝐕k]βjF^{[\tau]}_{\alpha i\beta j}=\sum_{k}\lambda_{k}^{\tau}[\mathbf{V}_{k}]_{\alpha i}[\mathbf{V}_{k}]_{\beta j}.

Now we have the following property.

Property 5.

The τ\tau-th order basic shifting filter fτ(𝐬)f_{\tau}(\mathbf{s}) can be calculated as

fτ(𝐬)\displaystyle f_{\tau}(\mathbf{s}) =𝐅[τ]𝐬\displaystyle=\mathbf{F}^{[\tau]}\diamond\mathbf{s} (64)
=(k=1MNλkτ𝐕k𝐕k)𝐬.\displaystyle=(\sum_{k=1}^{MN}\lambda_{k}^{\tau}\mathbf{V}_{k}\circ\mathbf{V}_{k})\diamond\mathbf{s}. (65)

This property is the M-GSP counterpart to the traditional linear system interpretation that complex exponential signals are eigenfunctions of linear systems [3], and provide a quicker implementation of higher-order shifting. With the kk-order polynomial term, the adaptive polynomial filter is defined as

h(𝐬)=kαk𝐅[k]𝐬,h(\mathbf{s})=\sum_{k}\alpha_{k}\mathbf{F}^{[k]}\diamond\mathbf{s}, (66)

where {αk}\{\alpha_{k}\} are parameters to be estimated from data.

Adaptive polynomial filter is useful in semi-supervised classification [35] and exploits underlying geometric topologies. We will illustrate further and provide application examples based on MLG polynomial filtering in Section VIII.

VI-B Spectral Filter Design

Filtering in the graph spectrum space is useful in GSP frequency analysis. For example, ordering the Laplacian graph spectrum 𝐕𝒢=[𝐞1,,𝐞N]N×N\mathbf{V}_{\mathcal{G}}=[\mathbf{e}_{1},\cdots,\mathbf{e}_{N}]\in\mathbb{R}^{N\times N} in a descent order by the graph total variation [3], i.e., high frequency to low frequency, the GFT of 𝐬N\mathbf{s}\in\mathbb{R}^{N} is calculated as 𝐬^=𝐕𝒢T𝐬\hat{\mathbf{s}}=\mathbf{V}_{\mathcal{G}}^{\mathrm{T}}\mathbf{s}. By removing kk elements in the low frequency part, i.e., s^=[s^1,,s^Nk,0,,0]\hat{s}^{\prime}=[\hat{s}_{1},\cdots,\hat{s}_{N-k},0,\cdots,0], a high-pass filter can be designed as

𝐬\displaystyle\mathbf{s}^{\prime} =𝐕𝒢s^\displaystyle=\mathbf{V}_{\mathcal{G}}\hat{s}^{\prime} (67)
=𝐕𝒢Σk𝐕𝒢T𝐬\displaystyle=\mathbf{V}_{\mathcal{G}}\Sigma_{k}\mathbf{V}_{\mathcal{G}}^{\mathrm{T}}\mathbf{s} (68)

where Σk=diag([σ1,,σN])\Sigma_{k}=diag([\sigma_{1},\cdots,\sigma_{N}]) is a diagonal matrix with σi=0\sigma_{i}=0 for i=1,,Nki=1,\cdots,N-k; otherwise, σi=0\sigma_{i}=0.

Similarly, in M-GSP, a spectral filter is designed by filtering in the spectrum space together with the inverse M-GFT. With Eq. (46) and Eq. (50), spectral filtering of 𝐬\mathbf{s} is defined as

𝐬=\displaystyle\mathbf{s}^{\prime}=
𝐄f[g(γ1)00g(γM)]𝐄fT𝐬𝐄e[f(σ1)00f(σN)]𝐄eT\displaystyle\mathbf{E}_{f}\begin{bmatrix}g(\gamma_{1})&\cdots&0\\ \vdots&\ddots&\vdots\\ 0&\cdots&g(\gamma_{M})\end{bmatrix}\mathbf{E}_{f}^{\mathrm{T}}\mathbf{s}\mathbf{E}_{e}\begin{bmatrix}f(\sigma_{1})&\cdots&0\\ \vdots&\ddots&\vdots\\ 0&\cdots&f(\sigma_{N})\end{bmatrix}\mathbf{E}_{e}^{\mathrm{T}} (69)

where functions g()g(\cdot) and f()f(\cdot) are designed by the specific tasks. For example, if one wants to design a layer-wise filter capturing the smoothness of signals in the MLG singular space, the function g()g(\cdot) can be designed as 𝐠=[1,,1,0,,0]\mathbf{g}=[1,\cdots,1,0,\cdots,0] by ordering the layer-wise singular vectors in the descent order of singular values.

VI-C Discussion

We briefly discuss the interpretation of polynomial filters and spectral filters. From the spatial perspective, MLG polynomial filter is a weighted sum of messaging passing on the multilayer graph in different orders, shown as Eq. (66). Each node collects information from both inter- and intra- layer neighbors, before combining them with its own information. From the spectrum perspective, M-GSP polynomial filters are eigenfunctions of linear systems, which are special cases of M-GSP spectral filters shown in Eq. (69) The M-GSP spectral filters assign different weights to each M-GSP spectrum via functions f(){f(\cdot)} and g(){g(\cdot)} depending on specific tasks. Both M-GSP polynomial filters and spectral filters can be useful for high-dimensional IoT signal processing. More discussions and examples of M-GSP filters are presented in Section VIII.

VII Discussion and Interpretative Insights

VII-A Interpretation of M-GSP Frequency

VII-A1 Interpretation of Graph Frequency

To better understand its physical meaning, we start with the total variation in digital signal processing (DSP). The total variation in DSP is defined as differences among signals over time [36]. Moreover, the total variations of frequency components have a 1-to-1 correspondence to frequencies in the order of their values. If the total variation of a frequency component is larger, the corresponding frequency with the same index is higher. This means that, a higher frequency component changes faster over time and exhibits a larger total variation. Interested readers could refer to [3, 8] for a detailed interpretation of total variation in DSP.

Now, let us elaborate the graph frequency motivated by the cyclic graph. Rewrite the finite signals in DSP as vectors, i.e., 𝐬=[s1,,sN]N\mathbf{s}=[s_{1},\cdots,s_{N}]\in\mathbb{R}^{N}, the signal shifting can be interpreted as the shift filtering corresponding to a cyclic graph shown in Fig. 6. Suppose that its adjacency matrix is written as

𝐂N=[0001100000000010]\displaystyle{\mathbf{C}_{N}=\begin{bmatrix}0&0&\cdots&0&1\\ 1&0&\cdots&0&0\\ \vdots&\ddots&\ddots&\ddots&\vdots\\ 0&0&\ddots&0&0\\ 0&0&\cdots&1&0\end{bmatrix}} (70)

Then, the shifted signal over the cyclic graph is calculated as 𝐬=𝐂N𝐬=[sNs1sN1]T\mathbf{s}^{\prime}=\mathbf{C}_{N}\mathbf{s}=[s_{N}\quad s_{1}\quad\cdots\quad s_{N-1}]^{\mathrm{T}}, which shifts the signal at each node to its next node. More specifically, 𝐂N\mathbf{C}_{N} can be decomposed as 𝐂N=𝐕𝚲𝐕1\mathbf{C}_{N}=\mathbf{V\Lambda}\mathbf{V}^{-1} where the eigenvalues λn=ej2πnN\lambda_{n}=e^{-j\frac{2\pi n}{N}} and 𝐕1=1N[λNkn]\mathbf{V}^{-1}=\frac{1}{\sqrt{N}}[\lambda^{kn}_{N}] is the discrete Fourier matrix. Inspired by the DSP, the eigenvectors in 𝐕\mathbf{V} are the spectral components (spectrum) of the cyclic graph and the eigenvalues are related to the graph frequencies [3].

Generalizing the adjacency matrix of the cyclic graph to the representing matrix 𝐅M\mathbf{F}_{M} of an arbitrary graph, the graph Fourier space consists of the eigenvectors of 𝐅M\mathbf{F}_{M} and the graph frequencies are related to the eigenvalues. More specifically, the graph Fourier space can be interpreted as the manifold or spectrum space of the representing matrix. As aforementioned, the total variations of frequency components reflect the order of frequencies, we can also use the total variation, i.e., TV(𝐞i)=𝐞i1|λ|max𝐞i1TV(\mathbf{e}_{i})=||\mathbf{e}_{i}-\frac{1}{|\lambda|_{max}}\mathbf{e}_{i}||_{1}, to rank the graph frequencies, where 𝐞i\mathbf{e}_{i} is the spectral component related to the eigenvalue λi\lambda_{i} in 𝐅M\mathbf{F}_{M}. Similar to DSP, the graph frequency indicates the oscillations over the vertex set, i.e., how fast the signals change over the graph shifting.

Refer to caption
Figure 6: Example of Cyclic Graph.

VII-A2 Interpretation of MLG Frequency

Now, return to M-GSP. Given spectral tensors 𝐕kM×N\mathbf{V}_{k}\in\mathbb{R}^{M\times N} of a multilayer graph, a signal 𝐬M×N\mathbf{s}\in\mathbb{R}^{M\times N} can be written in a weighted sum of the spectrum, i.e., 𝐬=kak𝐕k\mathbf{s}=\sum_{k}a_{k}\mathbf{V}_{k}. Viewing the spectral tensor as a signal component, the total variation is in the form of differences between the original signal and its shifted version in Eq. (53). If the signal component changes faster over the multilayer graph, the corresponding total variation is larger. Since we relate higher frequency component with a larger total variation, the MLG frequency indicates how fast the signal propagates over the multilayer graph under the representing tensor. If a signal 𝐬\mathbf{s} contains more high frequency components, it changes faster under the representing tensor.

VII-A3 Interpretation of MLG Singular Tensors

As discussed in Section VII-A1, the name of graph Fourier space arises from the adjacency matrix of the cyclic graph. However, when the algebra representation is generalized to an arbitrary graph, especially the Laplacian matrix, the definition of graph spectrum is less related to the Fourier space in DSP but can be interpreted as the manifold or subspace of the representing matrix instead. In literature, SVD is an efficient method to obtain the spectrum for signal analysis, such as spectral clustering [37] and PCA analysis [38]. It is straightforward to generalize graph spectral analysis to graph singular space, especially for the Laplacian matrix. In M-GSP, the order-wise singular vectors can be interpreted as subspaces characterizing features of layers and entities, respectively. Since HOSVD is robust and efficient, transforming signals to the MLG singular space (M-GST) for the analysis of underlying structures can be a useful alternative for M-GFT.

VII-B Interpretation of Entities and Layers

To gain better physical insight of entities and layers, we discuss two categories of datasets:

  • In most of the physical systems and datasets, signals can be modeled with a specific physical meaning in terms of layers and entities. In smart grid, for example, each station can be an entity, connected in two layers of computation and power transmission, respectively. Another example is video in which each geometric pixel point is an entity and each video frame form a layer. Each layer node denotes the pixel value in that video frame. M-GSP can be intuitive tool for these datasets and systems.

  • In some scenarios, however, the datasets usually only has a definition of layers without meaningful entities. In particular, for multilayer graphs with different numbers of nodes, we may insert some isolated artificial nodes to augment the multilayer graph. Often in such applications, it may be harder to identify the physical meaning of entities. Here, the entities may be virtual and are embedded in the underlying structure of the multilayer graph. Although definition of a virtual entity may vary with the chosen adjacency tensor, it relates to the topological structure in terms of global spectral information. For example, in Fig. 7, we can use two different definitions of virtual entities. Although the representing tensors for these two definitions differ, their eigenvalues remain the same. Considering also layer-wise flattening, the two supra-matrices are related by reshaping, by exchanging the fourth and fifth columns and rows. They still have the same eigenvalues, whereas the eigentensors can also be the same by implementing the reshaping operations. Note that, to capture distinct information from entities, their spectra would change with different definitions of virtual entities.

Refer to caption
Figure 7: Example of Different Entities.

VII-C Distinctions from Existing GSP Works

VII-C1 Graph Signal Processing

Generally, M-GSP extends traditional GSP into multilayer graphs. Although one can stack all MLG layers to represent them with a supra-matrix, such matrix representation makes GSP inefficient in extracting features of layers and entities separately. Given a supra-matrix of the MLG, the layers of nodes can not be identified directly from its index since all the nodes are treated equivalently. However, the tensor representation provides clear identifications on layers in its index. Moreover, in GSP, we can only implement a joint transform to process inter- and intra- layer connections together, while the M-GSP provide a more flexible choice on joint and order-wise analysis. In Section V-A, the joint M-GSP analysis introduced can be viewed as the bijection of GFT in the flattened MLG, with vertices indexed by both layers and entities. Beyond that, we flexibly provide order-wise spectral analysis based on tensor decompositions, which allow the order-wise analysis on layers and nodes. One can select suitable MLG-based tools depending on tasks. The joint spectral analysis can be implemented if we aim to explore layers and entities fully, whereas the order-wise spectral and singular analysis are more efficient in characterizing layers and entities separately.

VII-C2 Joint Time-Vertex Fourier Analysis

In [16], a joint time-vertex Fourier transform (JFT) is defined by implementing GFT and DFT consecutively. As discussed in Section VII-A1, the time sequences can be interpreted under a cyclic graph, and thus reside on a MLG structure. However, JFT assumes that all the layers have the same intra-layer connections, which limits the generalization of the MLG analysis. Differently, the tensor-based representation allows heterogeneous structures for the intra-layer connections, which makes M-GSP more general.

VII-C3 Multi-way Graph Signal Processing

In [18], MWGSP has been proposed to process high-dimensional signals. Given KKth-order high-dimensional signals, one can decompose the tensor signal in different orders, and construct one graph for each. Graph signal is said to reside on a high-dimensional product graph obtained by the product of all individual factor graphs. Although the MW-GFT is similar to M-GFT for K=2K=2, there still are notable differences in terms of spectrum. First, MWGSP can only process signals without exploiting a given structure since multiple graph spectra would arise from each order of the signals. For a multilayer graph with a given structure, such as physical networks with heterogeneous intralayer connections, MWGSP does not naturally process it efficiently and cohesively. The order-wise spectra come from factor graphs of each order in MWGSP while M-GSP spectra are calculated from the tensor representation of the whole MLG. Second, MWGSP assumes all the layers residing on a homogeneous factor graph and restricts the types of manageable MLG structure. For example, in a spatial temporal dataset, a product graph, formed by the product of spatial connections and temporal connections, assumes the same topology in each layer. However, many practical systems and datasets feature more complex geometric interactions. M-GSP provide a more intuitive and natural framework for such MLG. In summary, despite some shared similarities between MW-GFT and M-GFT in some scenarios, they serve different purposes and are suitable for different underlying data structures.

VII-D Comparison of Different Decomposition Methods

To compare recovery accuracy of representing tensor using different tensor decomposition methods, we examine eigen-tensor decomposition, HOSVD, optimal CP decomposition and Tucker decomposition in MLGs randomly generated from the Erdo¨sRe˙nyiErd\ddot{o}s-R\dot{e}nyi (ER) random graph ER(p,q,M,N)ER(p,q,M,N). Here MM is the number of layers with NN nodes in each layer, pp is the intralayer connection probability and qq is the interlayer connection probability. We apply different decomposition methods of similar complexity, and compute errors between decomposed and original tensors. From Table I, we can see that the eigen-tensor decomposition and HOSVD exhibit better overall accuracy. Generally, eigen-tensor decomposition is better suited for applications emphasizing joint features of layers and entities. On the other hand, HOSVD is effective at separating individual features of layers and entities. Note that, in addition to recovery accuracy, different decompositions may have different performance advantages when capturing different data features that can be better measured with different metrics.

TABLE I: Error of Decomposing the Representing Tensor
Graph Structure ER(0.3,0.3,6,5) ER(0.5,0.7,11,15) ER(0.8,0.7,6,15)
Eigen-tensor 8.3893e-15 1.6001e-13 3.8347e-13
HOSVD 1.011e-14 1.9563e-13 1.9056e-13
OPT-CP 9.22e-01 8.82e-01 9.24e-01
Tucker 9.37e-05 9.10e-05 9.40e-05

VIII Application Examples

We now provide some illustrative application examples within our M-GSP framework.

VIII-A Analysis of Cascading Failure in IoT Systems

Analysis of cascading failure in IoT systems based on the spreading processes in multilayer graphs has recently attracted significant interests [39]. Modeling the failure propagation in complex systems by shifting over multilayer graph, M-GSP spectrum theory can help the analysis of system stability. In this part, we introduce a M-GSP analysis for cascading failure over multilayer cyber-physical systems based on epidemic model [12]. Shown in Fig. 1, a cyber-physical system with MM layers and NN nodes in each layer can be intuitively modeled by a MLG with adjacency tensor 𝐀M×N×M×N\mathbf{A}\in\mathbb{R}^{M\times N\times M\times N}.

Here, we consider the susceptible-infectious-susceptible (SIS) model [40] for the failure propagation. In the SIS model, each node has two possible states: susceptible (not fail) or infectious (fail). At each time slot, the infectious node may cause failure to other nodes through directed links at certain infection rates, or it may heal itself spontaneously at a self-cure rate. The initial attack make several nodes infectious.

Since the nodes in the same layer correspond to the same functionality, e.g., power transmission, nodes on the same layer have the same self-cure rate and infection rate. The notations of the epidemic model are given as follows:

  • μα\mu_{\alpha}: self-cure rate for nodes on layer α\alpha;

  • θαβ\theta_{\alpha\beta}: infection rate describing failure propagation probability from nodes on layer β\beta to those on layer α\alpha;

  • Pi,α(t)P_{i,\alpha}(t): failure probability of the projected node of entity ii on layer α\alpha at time tt;

  • ϵi,α(t)\epsilon_{i,\alpha}(t): transition probability that the projected node of entity ii on layer α\alpha shifts from infectious state to susceptible state at time tt;

  • σi,α(t)\sigma_{i,\alpha}(t): transition probability that the projected node of entity ii on layer α\alpha remains susceptible at time tt.

Since an infectious node becomes susceptible if it cures itself without being infected by its neighbors, we have

ϵi,α(t)=μαj,β[1θαβAαiβjPj,β(t)]\epsilon_{i,\alpha}(t)=\mu_{\alpha}\prod_{j,\beta}[1-\theta_{\alpha\beta}A_{\alpha i\beta j}P_{j,\beta}(t)] (71)

Similarly, a susceptible node remains susceptible without being infected by its neighbors. Thus,

σi,α(t)=j,β[1θαβAαiβjPj,β(t)]\sigma_{i,\alpha}(t)=\prod_{j,\beta}[1-\theta_{\alpha\beta}A_{\alpha i\beta j}P_{j,\beta}(t)] (72)

The state transition forms a Markov chain, for which we derive the failure probability as

Pi,α(t)=1j,β[1TαiβjPj,β(t1)]\displaystyle P_{i,\alpha}(t)=1-\prod_{j,\beta}[1-T_{\alpha i\beta j}P_{j,\beta}(t-1)] (73)

where Tαiβj=(1μα)δαiβj+θαβAαiβjT_{\alpha i\beta j}=(1-\mu_{\alpha})\delta_{\alpha i\beta j}+\theta_{\alpha\beta}A_{\alpha i\beta j}, with δαiβj=1\delta_{\alpha i\beta j}=1 if (j,β)=(i,α)(j,\beta)=(i,\alpha); otherwise, δαiβj=0\delta_{\alpha i\beta j}=0.

We can define a transition tensor 𝐓M×N×M×N\mathbf{T}\in\mathbb{R}^{M\times N\times M\times N} with elements TαiβjT^{\beta j}_{\alpha i} to characterize the failure propagation in Eq. (73). In steady state, Pi,α(τ)=Pi,α(τ1)P_{i,\alpha}(\tau)=P_{i,\alpha}(\tau-1). Moreover,

P~i,α=1j,β[1TαiβjP~j,β],\tilde{P}_{i,\alpha}=1-\prod_{j,\beta}[1-T^{\beta j}_{\alpha i}\tilde{P}_{j,\beta}], (74)

where P~i,α\tilde{P}_{i,\alpha} is the failure probability of the projected node of entity ii on layer α\alpha in steady state. Following [12], we can arrive that P~i,α=0\tilde{P}_{i,\alpha}=0 if the spectral radius of the transition tensor ρ(𝐓)<1\rho(\mathbf{T})<1, which indicates no failed nodes in steady state. Thus, ρ(𝐓)\rho(\mathbf{T}) could serve as an indicator for system robustness. Here, to avoid being repetitive, we merely introduce a simple example of MLG-based cascading failure analysis. Interested readers may refer to our work [12] for more details. With a better understanding of M-GSP spatial shifting, one can develop more general analysis for various failure models in the multilayer IoT systems.

Refer to caption
(a) Original Image.
Refer to caption
(b) K-Means.
Refer to caption
(c) GSP.
Refer to caption
(d) MLG-SVD.
Refer to caption
(e) MLG-FZ.
Figure 8: Example of BCCD Datasets and Segmented Images: (a) the original image; (b)-(e) segmented images under different methods (WBCs are marked yellow, RBCs are marked green, and Platelet (P) is marked blue).

VIII-B Spectral Clustering

Clustering is a widely used tool in a variety of applications such as social network analysis, computer vision, and IoT. Spectral clustering is popular and effective among many variants. Modeling dataset by a normal graph before spectral clustering, significant performance improvement is possible in structured data [37]. In this part, we introduce M-GSP spectral clustering and demonstrate its application in RGB image segmentation.

To model an RGB image using MLG, we can directly treat its three colors as three layers. To reduce the number of nodes for computational efficiency, we first build NN superpixels for a given image and represent each superpixel as an entity in the multilayer graph. Here, we define the feature of a superpixel according to its RGB pixel values. For interlayer connections, each node connects with its counterparts in other layers. For intralayer connections on layer \ell, we calculate the Gaussian-based distance between two superpixels according to

Wij=exp(|𝐬i()𝐬j()|2δ2)W_{ij}=\exp\left(-\frac{|\mathbf{s}_{i}(\ell)-\mathbf{s}_{j}(\ell)|^{2}}{\delta_{\ell}^{2}}\right) (75)

if |𝐬i()𝐬j()|2t|\mathbf{s}_{i}(\ell)-\mathbf{s}_{j}(\ell)|^{2}\leq t_{\ell}; otherwise, Wij=0W_{ij}=0, where 𝐬i()\mathbf{s}_{i}(\ell) is the superpixel value on layer \ell, δ\delta_{\ell} is an adjustable parameter and tt_{\ell} is a predefined threshold. Different layers may have different structures. The threshold tt is set to be the mean of all pairwise distances. As such, an RGB image is modeled as multiplex graph with M=3M=3 and NN nodes.

We now consider MLG-based spectral clustering. For image segmentation, we focus on the properties of entities (i.e., superpixels), and implement spectral clustering over entity-wise spectrum by proposing Algorithm 1. The previous discussions have been summarized in steps 1-3. In Step 4, different schemes may be used to calculate spectrum, including spectral vector via tensor factorization in Eq. (29), and singular vector in Eq. (48). Step 5 determines KK based on the largest arithmetic gap in eigenvalues. Traditional clustering methods, such as kk-means clustering [37], can be carried out in Step 6.

TABLE II: Results of Image Segmentation in Image BSD300 and BSD500
BSD300(N=100, all) BSD300(N=300, all) BSD300(N=100, Coarse) BSD300(N=300, Coarse) BSD300(N=900,Coarse) BSD500(Coarse)
GSP 0.1237 0.1149 0.3225 0.3087 0.3067 0.3554
K-MEANS 0.1293 0.1252 0.3044 0.3105 0.3124 0.3154
MLG-SVD 0.1326 0.1366 0.3344 0.3394 0.3335 0.3743
MLG-CP 0.1321 0.1293 0.3195 0.3256 0.3243 0.3641
IIC 0.2071
GS 0.3658
BP 0.3239
DFC 0.3739
1:  Input: RGB Image 𝐈P×Q×3\mathbf{I}\in\mathbb{R}^{P\times Q\times 3};
2:  Build NN superpixels for the image 𝐈\mathbf{I} and calculate the value of superpixel based on the mean of all pixels inside that superpixel, i.e., 𝐬N×3\mathbf{s}\in\mathbb{R}^{N\times 3};
3:  Construct a multilayer graph 𝐀M×N×M×N\mathbf{A}\in\mathbb{R}^{M\times N\times M\times N};
4:  Find entity-wise spectrum 𝐄=[𝐞1,,𝐞N]N×N\mathbf{E}=[\mathbf{e}_{1},\cdots,\mathbf{e}_{N}]\in\mathbb{R}^{N\times N};
5:  Select the first KK important leading spectrum based on the eigenvalues (singular values) of 𝐄\mathbf{E} as 𝐂N×K\mathbf{C}\in\mathbb{R}^{N\times K};
6:  Cluster each row of 𝐂\mathbf{C}, and assign the iith superpixel into jjth cluster if the iith row of 𝐂\mathbf{C} is clustered into jjth group;
7:  Assign all pixels inside one superpixel to the cluster of that superpixel;
8:  Output: The segmented image.
Algorithm 1 MLG-based Unsupervised Image Segmentation

To test the proposed Algorithm 1, we first compare its results with those from GSP-based method and traditional kk-means clustering by using a public BCCD blood cell dataset shown in Fig. 8(a). In this dataset, there are mainly three types of objects, i.e., White Blood Cell (WBC) vs. Red Blood Cell (RBC) vs. Platelet (P). We set the number of clusters to 3 and N=1000N=1000. For GSP-based spectral clustering, we construct graphs based on the Gaussian model by using information from all 3 color values =13|𝐬i()𝐬j()|2\sum_{\ell=1}^{3}|\mathbf{s}_{i}(\ell)-\mathbf{s}_{j}(\ell)|^{2} to form edge connections in a single graph. There is only a single δ\delta_{\ell} and tt_{\ell} in the Gaussian model. For M-GSP, we use the MLG singular vectors (MLG-SVD), and tensor factorization (MLG-FZ) for spectral clustering, separately. Their respective results are compared in Fig 8. WBCs are marked yellow, and RBCs are marked green. Platelet (P) is marked blue. From the illustrative results, MLG methods exhibit better robustness and are better in detecting regions under noise. Comparing results from different MLG-based methods, we find MLG-FZ to be less stable than HOSVD, partly due to approximation algorithms used for tensor factorization. Overall, MLG-based methods shows reliable and robust performance over GSP-based method and kk-means.

In addition to visual inspection of results for such images, we are further interested to numerically evaluate the performance of the proposed methods against some state-of-art methods for several more complex datasets that contain more classes. For this purpose, we test our methods on the BSD300 and BSD500 datasets [41]. BSD300 contains 300 images with labels, and BSD500 contains 500 images with labels. We first cluster each image, and label each cluster with the best map of cluster orders against the ground truth. Numerically, we use mIOU (mean of Intersection-over-Union), also known as the mean Jaccard Distance, for all clusters in each image to measure the performance. The Jaccard Distance between two groups AA and BB is defined as

J(A,B)=|AB||AB|.J(A,B)=\frac{|A\cap B|}{|A\cup B|}. (76)

A larger mIOU indicates stronger performance. To better illustrate the results, we considered two setups of datasets, i.e., one containing fewer classes (coarse) and one containing all images (all). We compare our methods together with kk-means, GSP-based spectral clustering, invariant information clustering (IIC) [42], graph-based segmentation (GS) [43], back propagation (BP) [44] and differentiable feature clustering (DFC) [45]. The best performance is marked in bold. From the results of Table II, we can see that larger number of clusters in the first two columns generate worse performance. There are two natural reasons. First, the mapping of the best order of cluster labels is more difficult for more classes. Second, the graph-based spectral clustering is sensitive to the number of KK leading spectra and the structure of graphs. Regardless, MLG-based methods still demonstrate better performance. Moreover, even when we use the same total number of nodes in a single layer graph and multilayer graph for another fairness comparison in terms of complexity, i.e., N=300N=300 for graph and N=100N=100 for MLG, MLG-based methods still perform better than graph-based methods in this example application. MLG methods have competitive performances compared to the state-of-the-art methods. Note that, under proper training, neural network (NN)-based methods may give good results in cases with many clusters as suggested in [45].

VIII-C Semi-Supervised Classification

Semi-supervised classification is an important practical application for IoT intelligence. In this application, we apply MLG polynomial filters for semi-supervised classification. Traditional GSP defines adaptive filter as

f(𝐬)=iai𝐖i𝐬,f(\mathbf{s})=\sum_{i}a_{i}\mathbf{W}^{i}\mathbf{s}, (77)

where 𝐖\mathbf{W} is an adjacency matrix based on pairwise distance or a representing matrix constructed from the adjacency matrix. Here, signals are defined as labels or confidence values of nodes, i.e., 𝐬=[𝐬LT 0ULT]T\mathbf{s}=[\mathbf{s}_{L}^{\mathrm{T}}\>\mathbf{0}_{UL}^{\mathrm{T}}]^{\mathrm{T}} by setting unlabeled signals to zero. To estimate parameters aia_{i} of f()f(\cdot), Optimization can be formulated to minimize, e.g., the mean square error (MSE) from ground truth label 𝐲L\mathbf{y}_{L}

min𝐚M(f(𝐬))L𝐲L22,\min_{\mathbf{a}}\quad||M(f(\mathbf{s}))_{L}-\mathbf{y}_{L}||_{2}^{2}, (78)

where M()M(\cdot) is a mapping of filtered signals to their discrete labels. For example, in a {±1}\{\pm 1\} binary classification, one can assign a label to a filtered signal against a threshold (e.g. 0). Some other objective functions include labeling uncertainty, Laplacian regularization, and total variation. Using estimated parameters, we can filter the signal one more time to determine labels for some unlabeled data by following the same process.

Refer to caption
Figure 9: Results of Classification.

Similarly, in an MLG, we can also apply polynomial filters for label estimation. Given an arbitrary dataset 𝐗=[𝐱1,,𝐱N]K×N\mathbf{X}=[\mathbf{x}_{1},\cdots,\mathbf{x}_{N}]\in\mathbb{R}^{K\times N} with NN signal points and KK features for each node, we can construct a MLG by defining M=KM=K layers based on features and NN entities based on signal points. The inter- and intra- layer connections are calculated by the Gaussian distance with different parameters. Let its adjacency tensor 𝐀M×N×M×N\mathbf{A}\in\mathbb{R}^{M\times N\times M\times N}. A signal is defined by

𝐬=[𝐬L𝐬L𝟎UL𝟎UL]TM×N,\mathbf{s}=\begin{bmatrix}\mathbf{s}_{L}&\cdots&\mathbf{s}_{L}\\ \mathbf{0}_{UL}&\cdots&\mathbf{0}_{UL}\end{bmatrix}^{\mathrm{T}}\in\mathbb{R}^{M\times N}, (79)

which is an extended version of graph signal. Note that we do not necessarily need to order signals by placing zeros in the rear. We only write the signal as Eq. (79) for notational convenience. We now apply polynomial filters on signals, i.e.,

𝐬1=h1(𝐬)=iai𝐀[i]𝐬,\mathbf{s}_{1}=h_{1}(\mathbf{s})=\sum_{i}a_{i}\mathbf{A}^{[i]}\diamond\mathbf{s}, (80)

and

𝐬2=h2(𝐬)=𝐀[i]𝐬.\mathbf{s}_{2}=h_{2}(\mathbf{s})=\mathbf{A}^{[i]}\diamond\mathbf{s}. (81)

For a filtered signal 𝐬XM×N\mathbf{s}_{X}\in\mathbb{R}^{M\times N} (X=1,2X=1,2), we define a function to map 2-D signals into 1-D by calculated the column-wise mean of 𝐬X\mathbf{s}_{X}, i.e.,

𝐬¯X=meancol(𝐬X)1×N.\mathbf{\bar{s}}_{X}={\rm mean}_{col}(\mathbf{s}_{X})\in\mathbb{R}^{1\times N}. (82)
Refer to caption
Figure 10: Example of Transformed Signals in a Dynamic Point Cloud.

Next, we can define a function M()M(\cdot) on 𝐬¯X\mathbf{\bar{s}}_{X} and consider certain objective functions in filter design. To validate the efficacy of polynomial filtering in the MLG framework, we test h1()h_{1}(\cdot) and h2()h_{2}(\cdot) for the binary classification problem on the Cleveland Heart Disease Dataset. In this dataset, there are 297297 data points with 13 feature dimensions. We directly build a MLG with N=297N=297 nodes in each of the M=13M=13 layers. More specifically, we directly use the labels as 𝐬\mathbf{s}. For h1()h_{1}(\cdot) (AF), we set ai0a_{i}\neq 0 for at least one i>0i>0. Using MSE as objective function, we apply a greedy algorithm to estimate parameters {ai}\{a_{i}\}. We limit the highest polynomial order to 10{10}. For h2()h_{2}(\cdot) (APF), we estimate a classification threshold via the mean of 𝐬¯X\bar{\mathbf{s}}_{X} by setting the polynomial order i=10i=10.

We compare our methods with GSP-based method in similar setups as in aforementioned examples. The only difference is that we use 𝐬¯X\bar{\mathbf{s}}_{X} in M-GSP and use 𝐬=f([𝐬LT 0ULT]T)\mathbf{s}^{\prime}=f([\mathbf{s}_{L}^{T}\;\mathbf{0}_{\rm UL}^{T}]^{T}) in GSP for mapping and classification. We also present the results of label propagation and SVM for comparison. We randomly split the test and training data for 100 rounds. From the results shown in Fig. 9, GSP-based and M-GSP based methods exhibit better performance than traditional learning algorithms, particularly when the fraction of test samples is large. In general, M-GSP based methods demonstrate superior performance among all methods owing to its strength to extract ‘multilayer’ features, which could potentially benefit semi-supervised classification tasks in IoT systems.

VIII-D Dynamic Point Cloud Analysis

Refer to caption
Figure 11: Example of Filtered Signals in a Dynamic Point Cloud: entity 1 and entity 2 are legs; entity 3 is head; entity 4 and entity 5 are hands.

3D perception plays an important role in the high growth fields of IoT devices and cyber-physical systems, and continues to drive many progresses made in advanced point cloud processing [46]. Here, we propose a short time M-GST method to analyze dynamic point cloud. Given a dynamic point cloud with MM frames and at most NN points in each frame, we model it as a multilayer graph with MM layers and NN nodes in each layer. More specifically, we test the singular spectrum analysis over the motion sequences of subject 86 in the CMU database [47]. To implement the M-GST, we first divide the motion sequence into several shorter sequences with NfN_{f} frames. Next for each shorter sequence, we model interlayer connections by connecting points with the same label among successive frames. For points in the same frame, we connect two points based on the Gaussian-kernel within a Euclidean threshold τs\tau_{s} [6]. Let 𝐱i\mathbf{x}_{i} be the 3D coordinates of the iith point. We assign an edge weight between two points 𝐱i\mathbf{x}_{i} and 𝐱j\mathbf{x}_{j} as a nonzero Aij=exp(𝐱i𝐱j22/σ2)A_{ij}=\exp(-\|\mathbf{x}_{i}-\mathbf{x}_{j}\|^{2}_{2}/{\sigma^{2}}) only if 𝐱i𝐱j22τs\|\mathbf{x}_{i}-\mathbf{x}_{j}\|^{2}_{2}\leq\tau_{s}. Next, we estimate the spatial and temporal basis vectors of each shorter-term sequences by HOSVD in Eq. (48). Finally, we use the 3D coordinates of all points in each shorter-term sequences as signals and calculate their M-GST. To illustrate the results of M-GST, we examine the spectrogram similar to that of short-time Fourier transform (STFT) [48]. In Fig. 10, we transform the signal defined by the coordinates in ZZ dimension via M-GST and illustrate the transformation results for the divided frame sequence. From Fig. 10, one can easily identify different motions based on the MLG singular analysis.

To explore motions in dynamic point clouds, we can also apply the entity-wise MLG highpass filters described in Section VI-B to capture critical details of human bodies. More specifically, we select the first 140 frames in ‘walking’ and define the norm of three coordinates as signals. We select 5 body joints (entities) in each temporal frame (layers) shown as Fig. 11. From the results shown, entity 1 and entity 2 exhibit periodic patterns which are linked to the leg motion. Entity 3 (head) shows little movement relative to the main body. Entity 4 and entity 5 (hands) display more irregular patterns since they do not directly identify ‘walking’. To summarize, the MLG highpass filter can efficiently capture some key information of body movement and identify the meaning of nodes (entities). These and related information can assist further analysis of dynamic point clouds including compression and classification.

Our future works shall target more practical applications of point cloud on IoT devices, including point cloud compression, low-complex point cloud segmentation and robust denoising.

VIII-E Other Potential Applications in IoT Systems

Along with the widespread deployment of IoT technologies, system structures become increasingly complex. Traditional graph-based tools are less adept at modeling ‘multilayer’ graph interactions. The more general model of M-GSP provides additional opportunities for IoT applications. Here, we suggest several potential scenarios in IoT systems for M-GSP:

  • IoT networks with multilayer structure fit naturally to MLG, for which M-GSP can be designed for various tasks such as intrusion detection, resource management and state prediction;

  • Because of the dynamic nature in practical IoT networking, even signals on single-layer IoT systems naturally fit a spatial-temporal graph model, which can be also characterized by MLG. For such dynamic IoT systems, M-GSP tools, such as adaptive filters and MLG learning machines, can be developed for signal prediction and node classification.

Overall, the power of MLG in extracting underlying ‘multilayer/multi-level’ structures in the IoT systems makes M-GSP a potentially important tool in handling high-dimensional signal processing and learning tasks.

IX Conclusion

In this work, we present a novel tensor-based framework of multilayer graph signal processing (M-GSP) that naturally generalizes the traditional GSP to multilayer graphs. We first present the basic foundation and definitions of M-GSP including MLG signals, signal shifting, spectrum space, singular space, and filter design. We also provide interpretable discussion and physical insights through numerical results and examples to illustrate the strengths, general insights, and benefits of novel M-GSP framework. We further demonstrate exciting potentials of M-GSP in data processing applications through experimental results in several practical scenarios.

With recent advances in tensor algebra and multilayer graph theory, more opportunities are emerging to explore M-GSP and its applications. One such interesting problem is the efficient construction of multilayer graph, where M-GSP spectrum properties could improve the robustness of graph structure. Another promising direction is to develop multilayer graph neural networks based on the M-GSP spectral convolution. Additional future directions include the development of M-GSP sampling theory and fast M-GFT.

Unlike for undirected graphs, representing tensors of directed graphs is asymmetric, thereby making each layer or entity characterized by a pair of spectral vectors. To find the spectrum space of a directed multilayer graph, we also present two ways to compute:

  • Flattening analysis: Similar to the representing tensor of undirected graphs, we flatten the representing tensor as a second-order supra-matrix, and define spectrum space as the reshaped eigenvectors of the supra-matrix. The flattened matrix 𝐀FX\mathbf{A}_{FX} (or 𝐀FN\mathbf{A}_{FN}, 𝐀FL\mathbf{A}_{FL}) can be decomposed as

    𝐀FX𝐄Σ𝐄1,\mathbf{A}_{FX}\approx\mathbf{E}\Sigma\mathbf{E}^{-1}, (83)

    where 𝐄MN×MN\mathbf{E}\in\mathbb{R}^{MN\times MN} is the matrix of eigenvectors and Σ=diag(λi)\Sigma=diag(\lambda_{i}) is a diagonal matrix of eigenvalues. Then, we can reshape the eigenvectors, i.e., each column of 𝐄\mathbf{E} as 𝐕kM×N\mathbf{V}_{k}\in\mathbb{R}^{M\times N}, and reshape each row of 𝐄1\mathbf{E}^{-1} as 𝐔kM×N\mathbf{U}_{k}\in\mathbb{R}^{M\times N}. Consequently, the original tensor can be decomposed into

    𝐀k=1MNλk𝐕k𝐔k.\mathbf{A}\approx\sum_{k=1}^{MN}\lambda_{k}\mathbf{V}_{k}\circ\mathbf{U}_{k}. (84)
  • Tensor Factorization: We can also compute the spectrum from the tensor factorization based on CP-decomposition

    𝐀\displaystyle\mathbf{A} k=1Rλk𝐚k𝐛k𝐜k𝐝k\displaystyle\approx\sum_{k=1}^{R}\lambda_{k}\mathbf{a}_{k}\circ\mathbf{b}_{k}\circ\mathbf{c}_{k}\circ\mathbf{d}_{k} (85)
    =k=1Rλk𝐕k𝐔k.\displaystyle=\sum_{k=1}^{R}\lambda_{k}\mathbf{V}_{k}\circ\mathbf{U}_{k}. (86)

    where RR is the rank of tensor, 𝐚k,𝐜kM\mathbf{a}_{k},\mathbf{c}_{k}\in\mathbb{R}^{M} characterize layers, 𝐛k,𝐝kN\mathbf{b}_{k},\mathbf{d}_{k}\in\mathbb{R}^{N} characterize entities, and 𝐕k=𝐚k𝐛k,𝐔k=𝐜k𝐝kM×N\mathbf{V}_{k}=\mathbf{a}_{k}\circ\mathbf{b}_{k},\mathbf{U}_{k}=\mathbf{c}_{k}\circ\mathbf{d}_{k}\in\mathbb{R}^{M\times N} characterize the joint features. Since there are MNMN nodes, it is clear that RMNR\leq MN. Note that, for a single layer, Eq. (85) reduces to

    𝐀k=1Nλk𝐯k𝐮k.\displaystyle\mathbf{A}\approx\sum_{k=1}^{N}\lambda_{k}\mathbf{v}_{k}\circ\mathbf{u}_{k}. (87)

    Moreover, if 𝐕=(𝐯k)\mathbf{V}=(\mathbf{v}_{k}) and 𝐔=(𝐮kT)=𝐕1\mathbf{U}=(\mathbf{u}_{k}^{\mathrm{T}})=\mathbf{V}^{-1} are orthogonal bases, Eq. (87) is in a consistent form of the eigen-decomposition in a single-layer normal graph. In addition, Eq. (28) is also a special case of Eq. (85) if the multilayer graph is undirected.

Since tensor decomposition is less stable when exploring the factorization of a specific order or when extracting the separate features in the asymmetric tensors, we will defer more general analysis of directed networks to future works.

References

  • [1] A. Sandryhaila, and J. M. F. Moura, “Discrete signal processing on graphs” IEEE Transactions on Signal Processing, vol. 61, no. 7, pp. 1644-1656, Apr. 2013.
  • [2] D. I. Shuman, S. K. Narang, P. Frossard, A. Ortega and P. Vandergheynst, “The emerging field of signal processing on graphs: extending high-dimensional data analysis to networks and other irregular domains,” in IEEE Signal Processing Magazine, vol. 30, no. 3, pp. 83-98, May 2013.
  • [3] A. Ortega, P. Frossard, J. Kovačević, J. M. F. Moura and P. Vandergheynst, “Graph signal processing: overview, challenges, and applications,” in Proceedings of the IEEE, vol. 106, no. 5, pp. 808-828, May 2018.
  • [4] A. Sandryhaila and J. M. F. Moura, ”Discrete signal processing on graphs: frequency analysis,” in IEEE Transactions on Signal Processing, vol. 62, no. 12, pp. 3042-3054, Jun., 2014.
  • [5] S. Chen, A. Sandryhaila, J. M. F. Moura and J. Kovacevic, “Signal denoising on graphs via graph filtering,” 2014 IEEE GlobalSIP, Atlanta, GA, USA, Dec. 2014, pp. 872-876.
  • [6] S. Chen, D. Tian, C. Feng, A. Vetro and J. Kovačević, “Fast resampling of three-dimensional point clouds via graphs,” in IEEE Transactions on Signal Processing, vol. 66, no. 3, pp. 666-681, Feb., 2018.
  • [7] S. Chen, A. Sandryhaila, J. M. F. Moura and J. Kovačević, “Adaptive graph filtering: multiresolution classification on graphs,” 2013 IEEE Global Conf. on Signal and Info. Processing, Austin, TX, USA, Dec. 2013, pp. 427-430.
  • [8] S. Zhang, Z. Ding and S. Cui, “Introducing hypergraph signal processing: theoretical foundation and practical applications,” in IEEE Internet of Things Journal, vol. 7, no. 1, pp. 639-660, Jan. 2020.
  • [9] S. Barbarossa and S. Sardellitti, “Topological signal processing over simplicial complexes,” in IEEE Transactions on Signal Processing, vol. 68, pp. 2992-3007, Mar. 2020.
  • [10] M. Kivelä, A. Arenas, M. Barthelemy, J. P. Gleeson, Y. Moreno, and M. A. Porter, “Multilayer networks,” in Journal of complex networks, vol. 2, no. 3, pp. 203-271, Jul. 2014.
  • [11] M. De Domenico, A. Sole-Ribalta, E. Cozzo, M. Kivela, Y. Moreno, M. A. Porter, S. Gomez, and A. Arenas, “Mathematical formulation of multilayer networks,” Physical Review X, vol. 3, no. 4, p. 041022, 2013.
  • [12] S. Zhang, H. Zhang, H. Li and S. Cui, “Tensor-based spectral analysis of cascading failures over multilayer complex systems,” in Proc. 56th Allerton Conf. on Communication, Control, and Computing, Monticello, USA, Oct. 2018, pp. 997-1004.
  • [13] Z. Huang, C. Wang, M. Stojmenovic, and A. Nayak, “Characterization of cascading failures in interdependent cyberphysical systems,” IEEE Transactions on Computers, vol. 64, no. 8, pp. 2158-2168, Aug. 2015.
  • [14] S. V. Buldyrev, R. Parshani, G. Paul, H. E. Stanley, and S. Havlin, “Catastrophic cascade of failures in interdependent networks,” Nature, vol. 464, no. 7291, pp. 1025-1028, Apr. 2010.
  • [15] S. Gomez, A. Diaz-Guilera, J. Gomez-Gardenes, C. J. Perez-Vicente, Y. Moreno, and A. Arenas, “Diffusion dynamics on multiplex networks,”, Physical Review Letters, vol. 110, no. 2, p. 028701, Jan. 2013.
  • [16] F. Grassi, A. Loukas, N. Perraudin and B. Ricaud, “A time-vertex signal processing framework: scalable processing and meaningful representations for time-series on graphs,” in IEEE Trans. Signal Processing, vol. 66, no. 3, pp. 817-829, Feb. 2018.
  • [17] P. Das and A. Ortega, “Graph-based skeleton data compression,” 2020 IEEE 22nd International Workshop on Multimedia Signal Processing (MMSP), Tampere, Finland, Sep. 2020, pp. 1-6.
  • [18] J. S. Stanley, E. C. Chi and G. Mishne, “Multiway graph signal processing on tensors: integrative analysis of irregular geometries,” in IEEE Signal Processing Magazine, vol. 37, no. 6, pp. 160-173, 2020.
  • [19] M. A. Porter, “What is a multilayer network,” Notices of the AMS, vol. 65, no. 11, pp.1419- 1423, Dec. 2018.
  • [20] M. De Domenico, V. Nicosia, A. Arenas, and V. Latora, “Structural reducibility of multilayer networks,” Nature Communications, vol. 6, no. 6864, pp. 1-9, Apr. 2015.
  • [21] S. Chen, R. Varma, A. Sandryhaila and J. Kovačević, “Discrete Signal Processing on Graphs: Sampling Theory,” in IEEE Transactions on Signal Processing, vol. 63, no. 24, pp. 6510-6523, Dec.15, 2015.
  • [22] A. Sandryhaila, and J. M. F. Moura, “Discrete signal processing on graphs: graph filters,” in Proceedings of 2013 IEEE ICASSP, Vancouver, Canada, May 2013, pp. 6163-6166.
  • [23] T. G. Kolda and B. W. Bader, “Tensor decompositions and applications,” SIAM Review, vol. 51, no. 3, pp. 455-500, Aug. 2009.
  • [24] H. A. L. Kiers, “Towards a standardized notation and terminology in multiway analysis,” Journal of Chemometrics: A Journal of the Chemometrics Society, vol. 14, no. 3, pp. 105-122, Jan. 2000.
  • [25] L. De Lathauwer, B. De Moor, and J. Vandewalle, “A multilinear singular value decomposition,” SIAM Journal on Matrix Analysis and Applications, vol. 21, no. 4, pp. 1253-1278, Jan. 2000.
  • [26] A. Afshar, J. C. Ho, B. Dilkina, I. Perros, E. B. Khalil, L. Xiong, and V. Sunderam, “Cp-ortho: an orthogonal tensor factorization framework for spatio-temporal data,” Proc. 25th ACM SIGSPATIAL Intl. Conf. Advances in Geographic Info. Syst., Redondo Beach, USA, Jan. 2017, p. 67.
  • [27] I. V. Oseledets, “Tensor-train decomposition,” SIAM Journal on Scientific Computing, vol. 33, no. 5, pp. 2295-2317, Jun. 2011.
  • [28] A. Bogojeska, S. Filiposka, I. Mishkovski and L. Kocarev, ”On opinion formation and synchronization in multiplex networks,” 2013 21st TELFOR, Belgrade, Serbia, Nov. 2013, pp. 172-175.
  • [29] A. C. Kinsley, G. Rossi, M. J. Silk, and K. VanderWaal, “Multilayer and multiplex networks: an introduction to their use in veterinary epidemiology,” Frontiers in Veterinary Science, vol. 7, p. 596, Sep. 2020.
  • [30] M. De Domenico, C. Granell, M. A. Porter, and A. Arenas, “The physics of spreading processes in multilayer networks,” Nature Physics, vol. 12, no. 10, pp. 901-906, Aug. 2016.
  • [31] A. Loukas, “Graph reduction with spectral and cut guarantees,” Journal of Machine Learning Research, vol. 20, no. 116, pp. 1-42, Jun. 2019.
  • [32] Q. Shi, Y. -M. Cheung, Q. Zhao and H. Lu, “Feature extraction for incomplete data via low-rank tensor decomposition with feature regularization,” in IEEE Transactions on Neural Networks and Learning Systems, vol. 30, no. 6, pp. 1803-1817, Jun. 2019.
  • [33] M. Jouni, M. D. Mura and P. Comon, “Hyperspectral image classification using tensor cp decomposition,” IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, Jul. 2019, pp. 1164-1167.
  • [34] A. Sandryhaila, and J. M. F. Moura, “Discrete signal processing on graphs: graph filters,” in Proceedings of 2013 IEEE ICASSP, Vancouver, Canada, May 2013, pp. 6163-6166.
  • [35] S. Chen, F. Cerda, P. Rizzo, J. Bielak, J. H. Garrett and J. Kovačević, “Semi-supervised multiresolution classification using adaptive graph filtering with application to indirect bridge structural health monitoring,” in IEEE Trans. Signal Processing, 62(11):2879-2893, Jun. 2014.
  • [36] S. Mallat, A Wavelet Tour of Signal Processing, 3rd ed. New York, NY, USA: Academic, 2008.
  • [37] U. V. Luxburg, “A tutorial on spectral clustering.” Statistics and computing, vol .17, no. 4, pp. 395-416, 2007.
  • [38] H. Abdi, and L. J. Williams, “Principal component analysis,” Wiley Interdisciplinary Reviews: Computational Statistics, vol. 2, no. 4, pp. 433-459, 2010.
  • [39] M. Salehi, R. Sharma, M. Marzolla, M. Magnani, P. Siyari and D. Montesi, “Spreading processes in multilayer networks,” in IEEE Transactions on Network Science and Engineering, vol. 2, no. 2, pp. 65-83, 1 April-June 2015.
  • [40] M. J. Keeling and P. Rohani, Modeling Infectious Diseases in Humans and Animals.  Princeton, NJ, USA: Princeton University Press, 2011.
  • [41] D. Martin, C. Fowlkes, D. Tal and J. Malik, “A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics,” Proceedings Eighth IEEE ICCV, Vancouver, BC, Canada, Jul. 2001, pp. 416-423.
  • [42] X. Ji, J. F. Henriques, and A. Vedaldi, “Invariant information clustering for unsupervised image classification and segmentation,” in Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, Nov. 2019, pp. 9865– 9874.
  • [43] P. F. Felzenszwalb and D. P. Huttenlocher, “Efficient graph-based image segmentation,” International Journal of Computer Vision, vol. 59, no. 2, pp. 167–181, 2004.
  • [44] A. Kanezaki, “Unsupervised image segmentation by backpropagation,” in Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Calgary, AB, Canada, Apr. 2018, pp. 1543-1547.
  • [45] W. Kim, A. Kanezaki and M. Tanaka, “Unsupervised learning of image segmentation based on differentiable feature clustering,” in IEEE Transactions on Image Processing, vol. 29, pp. 8055-8068, Jul. 2020.
  • [46] Q. Deng, S. Zhang and Z. Ding, “An efficient hypergraph approach to robust point cloud resampling,” in IEEE Transactions on Image Processing, vol. 31, pp. 1924-1937, 2022.
  • [47] CMU, “Carnegie Mellon University Graphics Lab: Motion Capture Database,” accessed on Apr. 21, 2016. [Online].
  • [48] L. Durak, and O. Arikan, “Short-time fourier transform: two fundamental properties and an optimal implementation,” IEEE Transactions on Signal Processing, vol. 51, no. 5, pp. 1231-1242, May 2003.