This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Image Processing via Multilayer Graph Spectra

Songyang Zhang, Qinwen Deng, and Zhi Ding S. Zhang, Q. Deng, and Z. Ding are with the Department of Electrical and Computer Engineering, University of California at Davis, Davis, CA 95616 USA (e-mail: sydzhang, mrdeng, and [email protected]).
Abstract

Graph signal processing (GSP) has become an important tool in image processing because of its ability to reveal underlying data structures. Many real-life multimedia datasets, however, exhibit heterogeneous structures across frames. Multilayer graphs (MLG), instead of traditional single-layer graphs, provide better representation of these datasets such as videos and hyperspectral images. To generalize GSP to multilayer graph models and develop multilayer analysis for image processing, this work introduces a tensor-based framework of multilayer graph signal processing (M-GSP) and present useful M-GSP tools for image processing. We then present guidelines for applying M-GSP in image processing and introduce several applications, including RGB image compression, edge detection and hyperspectral image segmentation. Successful experimental results demonstrate the efficacy and promising futures of M-GSP in image processing.

Index Terms:
Image processing, multilayer graph signal processing, convolution, sampling theory

I Introduction

Signal processing over graphs has demonstrated successes in image processing [1] owing to its ability to uncover some hidden data features. Representing pixels (or superpixels) by nodes and their internal interactions by edges, graphs can be formed to describe the underlying image structures. In graph signal processing (GSP), graph Fourier space [2] and the corresponding spectral analysis have been applied in image compression [3] and denoising [4]. Nevertheless, single-layer graphs tend to be less efficient in processing high-dimensional images, where each frame corresponds to a distinct geometric structure. For example, hyperspectral images at different frequencies may exhibit heterogeneous graph structures, and do not lend themselves to traditional GSP tools. Similar scenarios include the color frames in RGB images and temporal frames in videos. On the other hand, since these frames are correlated, their inter-frame correlations should not be ignored in graph representations. Such complex intra- and inter- frame interactions among pixels can be represented by multilayer graphs (or multilayer networks) instead of a single-layer graph. Shown as Fig. 1, a hyperspectral image can be naturally modeled as a multilayer graph (MLG) with each pixel position as a node and each frequency frame as one layer. Of strong interest are questions of how to generalize single-layer GSP to multilayer graph signal processing (M-GSP) and how to extract additional benefits of M-GSP in problems such as image processing.

Graph signal processing over multilayer networks (graphs) has attracted increasing attentions recently. In [6], a two-step graph Fourier transform (GFT) is separately implemented spatially and temporally to compress spatial-temporal skeleton data. Similarly, authors of [7] defined a joint time-vertex Fourier transform (JFT) by applying GFT and DFT consecutively on time-varying datasets. However, both the two-step GFT and JFT assume that the spatial connections in all layers match the same underlying graph structure, thereby limiting the use of such spectral analysis on heterogeneous data frames. In addition to matrix-based GSP analysis, a tensor-based multi-way graph signal processing framework (MWGSP), proposed in [8], decomposes the high-dimensional tensor signal into individual orders and constructs factor graphs for each order. Although MWGSP allows more complex interlayer connections and is suitable for high-dimensional signals, it assumes the dataset to lie in a product graph from all factor graphs, therefore still requires all intralayer connections to exhibit homogeneous structures from the node-wise factor graph.

To capture the heterogeneous layer structures in MLG for generalized GSP, a framework of multilayer network signal processing (M-GSP) has been proposed by using tensor representations [9]. Defining MLG spectra via tensor decomposition, both joint and order-wise analysis for MLG can tackle various intralayer and complex interlayer connections. However, we note that existing M-GSP tools still lack some definitions of basic MLG spectral operations useful in image processing, such as sampling and convolution.

In this work, we investigate the use of M-GSP spectral analysis [9] and the use of M-GSP tools in image processing. The contributions can be summarized as follows:

  • We first introduce the MLG models and representations for images, and then define several M-GSP tools useful for image processing.

  • We provide a tutorial on applying M-GSP in image processing and present several practical cases, including image compression, edge detection, and hyperspectral image segmentation.

We compare each proposed M-GSP image application to benchmarks from traditional GSP and/or learning methods. Our test results demonstrate the strength of M-GSP and the proposed MLG spectral tools in processing images.

II Models and Spectrum of Images in M-GSP

Refer to caption
Figure 1: Example of Multilayer Graph Model for Hyperspectral Images: pixels are represented by nodes and their similarities are represented by edges.

In this section, we introduce the MLG model of images and outline fundamentals of M-GSP [9]. We refrain from using the term “network” because of its different meanings in various field ranging from communications to deep learning. From here onward, unless otherwise specified, we shall use the less ambiguous term of multilayer graph instead.

To model related image frames as multilayer graph (MLG), we can naturally define layers as frames and nodes as pixels (superpixels) in each frame. Constructing interactions based on feature similarity or physical connections, a multilayer graph with the same number of nodes in each layer can be constructed to represent the images. Moreover, viewing NN (super)pixels {x1,,xN}\{x_{1},\cdots,x_{N}\} as entities, the MLG structure can be viewed as projecting the physical entities into natural representations of MM frames {l1,,lM}\{l_{1},\cdots,l_{M}\}, e.g., spectrum band frames for hyperspectral images, color frames for RGB images, and temporal frames for time-varying images. Such multilayer graph structure with MM layers and NN nodes in each layer can be intuitively represented by a forth-order adjacency tensor 𝐀=(Aαiβj)M×N×M×N\mathbf{A}=(A_{\alpha i\beta j})\in\mathbb{R}^{M\times N\times M\times N} with 1α,βM,1i,jN1\leq\alpha,\beta\leq M,1\leq i,j\leq N. Here, each entry AαiβjA_{\alpha i\beta j} of the adjacency tensor 𝐀\mathbf{A} indicates the edge strength between entity jj’s projected node in layer β\beta and entity ii’s projected node in layer α\alpha. Similar to normal graphs, Laplacian tensor 𝐋M×N×M×N\mathbf{L}\in\mathbb{R}^{M\times N\times M\times N} can be defined as the alternative representation of MLG. Interested readers may refer to [9] for more details of the adjacency and Laplacian tensor. For convenience, we use a general notation 𝐅M×N×M×N\mathbf{F}\in\mathbb{R}^{M\times N\times M\times N} to represent the MLG in this work. In addition to the representation of MLG, signals over MLG can be intuitively defined as 𝐬=(sαi)M×N\mathbf{s}=({s_{\alpha i}})\in\mathbb{R}^{M\times N}, with sαis_{\alpha i} as the value of pixel ii in frame α\alpha.

In an undirected multilayer graph, the representing tensor (adjacency/Laplacian) 𝐅\mathbf{F} is partially symmetric between orders one and three, and between orders two and four, respectively. Then, it can be approximated via orthogonal CANDECOMP/PARAFAC (CP) decomposition [10] as

𝐅\displaystyle\mathbf{F} α=1Mi=1Nλαi𝐟α𝐞i𝐟α𝐞i,\displaystyle\approx\sum_{\alpha=1}^{M}\sum_{i=1}^{N}\lambda_{\alpha i}\cdot\mathbf{f}_{\alpha}\circ\mathbf{e}_{i}\circ\mathbf{f}_{\alpha}\circ\mathbf{e}_{i}, (1)

where \circ is the tensor outer product, 𝐟αM\mathbf{f}_{\alpha}\in\mathbb{R}^{M} are orthonormal, 𝐞iN\mathbf{e}_{i}\in\mathbb{R}^{N} are orthonormal and 𝐕~αi=𝐟α𝐞iM×N\tilde{\mathbf{V}}_{\alpha i}=\mathbf{f}_{\alpha}\circ\mathbf{e}_{i}\in\mathbb{R}^{M\times N}. Here, 𝐟α\mathbf{f}_{\alpha} and 𝐞i\mathbf{e}_{i} are named as the MLG spectral bases characterizing the features of layers and entities, respectively.

Besides MLG spectral space, the singular space is defined from HOSVD [11] as an alternative subspace of MLG, i.e.,

𝐅=𝐒×1𝐔(1)×2𝐔(2)×3𝐔(3)×4𝐔(4),\mathbf{F}=\mathbf{S}\times_{1}\mathbf{U}^{(1)}\times_{2}\mathbf{U}^{(2)}\times_{3}\mathbf{U}^{(3)}\times_{4}\mathbf{U}^{(4)}, (2)

where ×n\times_{n} denotes the nn-mode product [15] and 𝐔(n)\mathbf{U}^{(n)} is a unitary (In×In)(I_{n}\times I_{n}) matrix, with I1=I3=MI_{1}=I_{3}=M and I2=I4=NI_{2}=I_{4}=N. For an undirected multilayer graph, the adjacency tensor is symmetric for every 2-D combination. Thus, there are two modes of singular spectrum, i.e., (γα,𝐟α)(\gamma_{\alpha},\mathbf{f}_{\alpha}) for mode 1,31,3, and (σi,𝐞i)(\sigma_{i},\mathbf{e}_{i}) for mode 2,42,4. More specifically, 𝐔(1)=𝐔(3)=(𝐟α)\mathbf{U}^{(1)}=\mathbf{U}^{(3)}=(\mathbf{f}_{\alpha}) and 𝐔(2)=𝐔(4)=(𝐞i)\mathbf{U}^{(2)}=\mathbf{U}^{(4)}=(\mathbf{e}_{i}). Singular tensor analysis and spectral analysis are both efficient tools for image processing depending on specific tasks. We shall elaborate in Section IV.

Suppose that 𝐄f=[𝐟1,,𝐟M]\mathbf{E}_{f}=[\mathbf{f}_{1},\cdots,\mathbf{f}_{M}] and 𝐄e=[𝐞1,,𝐞N]\mathbf{E}_{e}=[\mathbf{e}_{1},\cdots,\mathbf{e}_{N}] include all the layer-wise (frame) and entity-wise (pixel) spectral (or singular) vectors. The MLG Fourier (or Singular) transform can be defined as 𝐬^=𝐄fT𝐬𝐄eM×N\hat{\mathbf{s}}=\mathbf{E}_{f}^{\mathrm{T}}\mathbf{s}\mathbf{E}_{e}\in\mathbb{R}^{M\times N}, and the inverse transform is given as 𝐬=𝐄f𝐬^𝐄eT\mathbf{s}^{\prime}=\mathbf{E}_{f}\hat{\mathbf{s}}\mathbf{E}_{e}^{\mathrm{T}}. Due to page limitation, here we only introduce fundamentals of MLG spectral and singular vectors for MLG analysis. More details, such as spectral interpretation, frequency analysis, and filter design can be found in [9].

III M-GSP Tools for Image Processing

In this section, we introduce important operations for M-GSP frequency analysis specially useful in image processing.

III-1 MLG Spectral Convolution

Graph-based spectral convolution is useful in applications such as feature extraction, classification, and graph convolutional networks [12]. Graph convolution can be implemented indirectly in the transform domain [13]. Define graph convolution as 𝐱𝐲=1((𝐱)(𝐲))\mathbf{x}\star\mathbf{y}=\mathcal{F}^{-1}(\mathcal{F}(\mathbf{x})*\mathcal{F}(\mathbf{y})), where * denotes Hadamard product [15], \mathcal{F} is the graph Fourier transform (GFT) and 1\mathcal{F}^{-1} denotes inverse GFT. This formulation exploits the property that graph convolution in vertex domain is equivalent to spectral domain product.

In [14], a more general definition of graph convolution take the filtering form, i.e., 𝐱𝐲=f(𝐱)=(𝐕diag(𝐲^)𝐕1)𝐱\mathbf{x}\star\mathbf{y}=f(\mathbf{x})=(\mathbf{V}diag(\mathbf{\hat{y}})\mathbf{V}^{-1})\cdot\mathbf{x}, where 𝐕\mathbf{V} consists of eigenvectors of the representing matrix as columns and 𝐲^\hat{\mathbf{y}} can be interpreted as filter parameters in graph spectral domain. Note that, these two definitions of graph convolution have the same mathematical formulation. For convenience, we define spectral convolution over MLG in a similar mathematical formulation as follows. In M-GSP, the spectral convolution is defined as

𝐱𝐲=MLG1(MLG(𝐱)MLG(𝐲)),\mathbf{x}\star\mathbf{y}=\mathcal{F}_{\rm MLG}^{-1}(\mathcal{F}_{\rm MLG}(\mathbf{x})*\mathcal{F}_{\rm MLG}(\mathbf{y})), (3)

where MLG\mathcal{F}_{\rm MLG} is M-GFT (or M-GST) and MLG1\mathcal{F}_{\rm MLG}^{-1} is the inverse M-GFT (or M-GST). Then, the MLG convolutional filter with parameters 𝐜\mathbf{c} on the signal 𝐬\mathbf{s} is defined by fc(𝐬)=𝐬𝐜f_{c}(\mathbf{s})=\mathbf{s}\star\mathbf{c}.

III-A Sampling and Interpolation

III-A1 Sampling in Vertex Domain

Within the M-GSP framework, sampling is a process to select a subset of signals to describe some global information. Consider the sampling of signal 𝐬M×N\mathbf{s}\in\mathbb{R}^{M\times N} over MLG into PP layers with index {p1,p2,,pP}\{p_{1},p_{2},\cdots,p_{P}\} and QQ entities with index {q1,q2,,qQ}\{q_{1},q_{2},\cdots,q_{Q}\}. Let ×n\times_{n} be the nn-mode product introduced in [15]. The sampling operation is defined by

𝐬D=𝐬×1𝐒P×2𝐒QP×Q\mathbf{s}_{D}=\mathbf{s}\times_{1}\mathbf{S}_{P}\times_{2}\mathbf{S}_{Q}\in\mathbb{R}^{P\times Q} (4)

where sampling operator 𝐒PP×M\mathbf{S}_{P}\in\mathbb{R}^{P\times M} consists of elements

[SP]ij={1j=pi0otherwise,\displaystyle[{S}_{P}]_{ij}=\left\{\begin{array}[]{rcl}1&&j=p_{i}\\ 0&&\mbox{otherwise}\end{array},\right. (7)

and 𝐒QQ×N\mathbf{S}_{Q}\in\mathbb{R}^{Q\times N} with elements calculated as

[SQ]ij={1j=qi0otherwise.\displaystyle[{S}_{Q}]_{ij}=\left\{\begin{array}[]{rcl}1&&j=q_{i}\\ 0&&\mbox{otherwise}\end{array}.\right. (10)

The interpolation operation is defined by

𝐬R=𝐬D×1𝐓M×2𝐓NM×N,\mathbf{s}_{R}=\mathbf{s}_{D}\times_{1}\mathbf{T}_{M}\times_{2}\mathbf{T}_{N}\in\mathbb{R}^{M\times N}, (11)

where 𝐓MM×P\mathbf{T}_{M}\in\mathbb{R}^{M\times P} and 𝐓NN×Q\mathbf{T}_{N}\in\mathbb{R}^{N\times Q}. Sampling in vertex domain can be viewed as a two-step sampling in layers and entities. M-GSP sampling and GSP sampling share many interesting properties. Readers may refer to [16] for more details.

Refer to caption
(a) Layer-Wise Sampling.
Refer to caption
(b) Entity-Wise Sampling.
Refer to caption
(c) Block-wise (Layer=3).
Refer to caption
(d) Block-wise (Layer=2).
Refer to caption
(e) Block-wise (Layer=1).
Figure 2: Results of Sampling in the Order of Transformed Value.

III-A2 Sampling in Spectral Domain

Beyond direct sampling in the vertex domain, it can be more convenient and straightforward to sample signals in spectral domain. By neglecting spectral domain zeros, we can easily approximately recover the full signal from samples by interpolating with zeros to the sampled signal transforms before recovering with the inverse transform. Sampling in spectral domain is intuitive for lossless sampling and recovery. Here we shall focus on the lossy sampling of tensor-based signals.

Generally, we can sample signals in spectral domain according to Algorithm 1. We reshape transformed signals in terms of eigenvalues or transformed values from top left to bottom right in descending order. When sampling a sub-block in 𝐬^\mathbf{\hat{s}}, there are three different directions shown in Fig. 3. Depending on the structure and features of the underlying dataset, different methods may be suitable for different tasks. We will show examples of data compression based on the MLG sampling theory in Section IV.

1:  Input: Signals 𝐬M×N\mathbf{s}\in\mathbb{R}^{M\times N};
2:  Transform the signal into spectral space as 𝐬^M×N\mathbf{\hat{s}}\in\mathbb{R}^{M\times N} via M-GFT or M-GST;
3:  Reorder the elements in 𝐬^\mathbf{\hat{s}} via certain rules;
4:  Sample transformed signal in the left top as 𝐬^DP×Q\mathbf{\hat{s}}_{D}\in\mathbb{R}^{P\times Q};
5:  Interpolate zeros in the right bottom of 𝐬^D\mathbf{\hat{s}}_{D} as 𝐬^RM×N\mathbf{\hat{s}}_{R}\in\mathbb{R}^{M\times N}
6:  Implement the inverse MLG transform on 𝐬^R\mathbf{\hat{s}}_{R} to obtain the recovered signal 𝐬RM×N\mathbf{s}_{R}\in\mathbb{R}^{M\times N}.
7:  Output: Sampled coefficients 𝐬^D\mathbf{\hat{s}}_{D} and recovered signal 𝐬R\mathbf{s}_{R}.
Algorithm 1 Sampling of Tensor Signal in Spectral Domain

IV Application Examples

IV-A Image Compression

Data compression through sampling may reduce dataset size for efficient storage and transmission. Moreover, projecting signals into a suitable manifold or space, one may reveal a sparse representation of the data. In M-GSP, we can also select a subset of transformed signal samples in spectral domain to approximate the full signal as suggested in Algorithm 1. Moreover, we can also integrate data encoding to further compress data size. In this application example, we mainly focus on sampling to demonstrate the use of MLG spectrum space to represent structured data.

We test MLG sampling on the RGB icon dataset [17]. This dataset contains several icon images of size 16×16×3\mathbb{R}^{16\times 16\times 3}. Our goal is to sample a subset of K=16×16×3=768K=16\times 16\times 3=768 transformed data points in spectral domain. We follow Algorithm 1 to sample the transformed data before using the sampled data to recover the original images. In this case, a 3-layer MLG is constructed where the intralayer connections are formed by connecting to the 4-neighborhood pixels and the interlayer connections connect the counterparts of the same pixel in all layers. We compare our methods with two GSP-based methods. The first method (GFT) builds a graph with 16×1616\times 16 nodes and implements GFT on signals in each layer separately. The second method (GFT2) builds a graph of size 16×1616\times 16 for pixels that are neighbor-4 connected and constructs a graph of size 33 for (fully connected) frames. Let 𝐄\mathbf{E} be the spectrum of frame-wise graph, and 𝐅\mathbf{F} be the spectrum of pixel-wise graph. We apply GFT on the 2-D signal 𝐬3×256\mathbf{s}\in\mathbb{R}^{3\times 256} to generate 𝐬^=𝐄T𝐬𝐅.\mathbf{\hat{s}}=\mathbf{E}^{\mathrm{T}}\mathbf{s}\mathbf{F}. For both GSP and M-GSP methods, we apply the Laplacian matrix/tensor. We order the transformed signal in descending order of the transformed values from top left to bottom right, i.e., keeping larger values in Fig. 2. As mentioned in Section III-A2, there are three directions for sampling node values. In layer-wise sampling, we remove samples layer-by-layer from bottom to up. More specifically, we remove samples from right to left when sampling each layer. Similarly, in entity-wise sampling, we remove samples from right to left. When sampling each column, we move from bottom to top. In block-wise sampling, we preserve samples in top-left blocks. Since there are only three layers, we present all results of block-wise sampling layer-by-layer in Fig. 2(c)-2(e).

Defining sampling fraction as the ratio of saved samples to total original signal samples, we measure errors between the recovered and original images. The results from are shown in Fig. 2, where MLN-EIG and MLN-HOSVD are from MLG Fourier and singular space respectively. These results confirm that the proposed MLG methods provide better performance than GSP.

IV-B Edge Detection

Convolutions are widely used for edge detection of images. In traditional processing, one may first transform RGB to gray-scale images for edge detection using operators such as Prewitt, Canny, and Sobel [18, 19]. Similarly, geometric convolution methods could also detect edges in gray-scale images [20]. Unlike traditional (graph) signal processing, M-GSP can model RGB images by using 3-D windows/blocks and define 3-D convolution kernels as shown in Fig. 4(b). To construct an MLG for edge detection, we also represent each RGB color by a layer and connect each pixel in one layer with its counterparts in other layers for interlayer connections. Each pixel is connected with its four adjacency neighbors for intralayer connections. The benefit of constructing such graph is to bypass the computation needed to regenerate spectrum for each window (block). For a 3×3×33\times 3\times 3 kernel, we construct a 3-layer MLG with 9 nodes in each layer.

Refer to caption
(a) Layer-Wise.
Refer to caption
(b) Entity-Wise.
Refer to caption
(c) Block-Wise.
Figure 3: Options of Sampling Directions for Tensor Signals.
Refer to caption
(a) 2-D.
Refer to caption
(b) 3-D.
Figure 4: Example of Convolution Kernels.
Refer to caption
Figure 5: Results of Edge Detection.

In general, one can first extract smooth image features, before applying edge detection on the difference between original and smoothed images. To define a graph localization kernel for smoothing, we start from the single-layer graph. Let KK be the number of nodes and the graph convolution kernel be 𝐜K\mathbf{c}\in\mathbb{R}^{K} where ci=1c_{i}=1 for i=k1,,kzi=k_{1},\cdots,k_{z} with zKz\leq K; otherwise, ci=0c_{i}=0. The convolution output for a signal 𝐬K\mathbf{s}\in\mathbb{R}^{K} is 𝐬=𝐕[(𝐕T𝐬)(𝐕T𝐜)]=j(𝐟j𝐟jT𝐬i[fj]ki)\mathbf{s^{\prime}}=\mathbf{V}[(\mathbf{V}^{\mathrm{T}}\mathbf{s})*(\mathbf{V}^{\mathrm{T}}\mathbf{c})]=\sum_{j}(\mathbf{f}_{j}\mathbf{f}_{j}^{\mathrm{T}}\mathbf{s}\sum_{i}[f_{j}]_{k_{i}}), where 𝐕=[𝐟1,,𝐟K]\mathbf{V}=[\mathbf{f}_{1},\cdots,\mathbf{f}_{K}] is the graph spectrum. Different from a pure shift, the convolution kernel enhances information around nodes {vki}\{v_{k_{i}}\}. With locally-enhanced gray-scale images, we can replace centroid of a window by the mean of convolution outputs inside that window to smooth signals further. By shifting the window over the entire image, one can obtain a locally-enhanced blurred gray-scale image.

Similarly, since our goal is to smooth signals within the 3-D window in MLG, we can simply localize the information around the center of the block. Here, we design two 3-D kernels with size 3×k×k3\times k\times k, i.e, 𝐜13×k2\mathbf{c}_{1}\in\mathbb{R}^{3\times k^{2}} with entry in node v2v_{2} as 1 and all other entries as zero, and 𝐜23×k2\mathbf{c}_{2}\in\mathbb{R}^{3\times k^{2}} with entries in node v1,v2,v3v_{1},v_{2},v_{3} as 1 and all other entries as zero, indexed as 4(b). We then obtain a smoothed gray-scale image from convolution output and determine its difference from the original gray-scale image generated from RGB. A threshold is designed here to detect edges based on the explicit difference between the smoothed image and its original. MLG spectra are applied here to implement the MLG convolution. Results from the MLG edge detection can be visually examined in Fig. 5, compared with results from several classic detection methods (e.g., Canny, Prewitt, Sobel) and a GSP-based method. The comparison shows that MLG-based methods yield more smooth edges and show clearer details.

Refer to caption
Figure 6: Results of HSI Segmentation.

IV-C Hyperspectral Image Segmentation

Spectral clustering [21] is a widely-used method for unsupervised image segmentation. In this part, we introduce unsupervised hyperspectral image (HSI) segmentation based on MLG singular space. To construct a MLG, we first cluster the multi-spectral frames into M=10M=10 clusters as layers and construct N=100N=100 superpixels as entities. The new signals in each layer is the mean of all frames in that cluster. By constructing intralayer and interlayer connections using Gaussian distance, we generate an MM-layer MLG with NN nodes in each layer. Implementing HOSVD on the adjacency tensor, we can obtain the entity-wise (superpixel) singular vectors in descending order of the entity-wise singular values, i.e., 𝐄e=[𝐞1𝐞N]N×N\mathbf{E}_{e}=[\mathbf{e}_{1}\cdots\mathbf{e}_{N}]\in\mathbb{R}^{N\times N}. Selecting the first PP singular tensors as 𝐏k=[𝐞1,,𝐞P]N×P\mathbf{P}_{k}=[\mathbf{e}_{1},\cdots,\mathbf{e}_{P}]\in\mathbb{R}^{N\times P} based on the largest singular gap, we can cluster the rows of 𝐏k\mathbf{P}_{k} into QQ groups based on kk-means clustering. The superpixel ii is clustered into group jj if the ii-th row of 𝐏k\mathbf{P}_{k} is in group jj. Labeling the pixels as the same cluster of its superpixel, we segment the HSI. Let us compare the M-GSP spectral clustering with kk-means and GSP spectra clustering in datasets named Indian Pine, Pavia University and Salinas. The visualization results of the segmented Salinas HSI are shown as Fig. 6, where M-GSP illustrates more robust results. In addition to visual results, we also consider quantitative metrics. We compute the edges of segmentation results from different methods, and compare them to the ground truth in Table I. M-GSP continues to show superior performance in all HSIs.

TABLE I: Accuracy of Segmentation Boundaries
Data kk-means GSP M-GSP
IndianP 0.8257 0.8298 0.8441
Salinas 0.9208 0.9285 0.9409
PaviaU 0.9070 0.9088 0.9255

V Conclusions

In this work, we introduce image modeling based on multilayer graphs. We present the fundamentals of M-GSP frequency analysis together with the spectral operations. We further provide several example applications based on M-GSP operations. Our test results demonstrate the efficacy and strong future potentials of applying M-GSP in image processing. Other applications of M-GSP may include dynamic point clouds and video signals.

References

  • [1] G. Cheung, E. Magli, Y. Tanaka and M. K. Ng, “Graph spectral image processing,” in Proceedings of the IEEE, vol. 106, no. 5, pp. 907-930, May 2018.
  • [2] A. Ortega, P. Frossard, J. Kovačević, J. M. F. Moura and P. Vandergheynst, ”Graph signal processing: overview, challenges, and applications,” in Proceedings of the IEEE, vol. 106, no. 5, pp. 808-828, May 2018.
  • [3] G. Fracastoro, D. Thanou and P. Frossard, “Graph transform optimization with application to image compression,” in IEEE Transactions on Image Processing, vol. 29, pp. 419-432, Aug. 2020.
  • [4] M. Onuki, S. Ono, M. Yamagishi and Y. Tanaka, “Graph signal denoising via trilateral filter on graph spectral domain,” in IEEE Transactions on Signal and Information Processing over Networks, vol. 2, no. 2, pp. 137-148, Jun. 2016.
  • [5] M. D. Domenico, A. Solé-Ribalta, E. Cozzo, M. Kivelä, Y. Moreno, M. A. Porter, S. Gómez, and A. Arenas, “Mathematical formulation of multilayer networks,” Physical Review X, vol. 3, no. 4, p. 041022, Dec. 2013.
  • [6] P. Das and A. Ortega, “Graph-based skeleton data compression,” 2020 IEEE 22nd International Workshop on Multimedia Signal Processing (MMSP), Tampere, Finland, Sep. 2020, pp. 1-6.
  • [7] F. Grassi, A. Loukas, N. Perraudin and B. Ricaud, “A time-vertex signal processing framework: scalable processing and meaningful representations for time-series on graphs,” in IEEE Transactions on Signal Processing, vol. 66, no. 3, pp. 817-829, Feb. 2018
  • [8] J. S. Stanley, E. C. Chi and G. Mishne, “Multiway graph signal processing on tensors: integrative analysis of irregular geometries,” in IEEE Signal Processing Magazine, vol. 37, no. 6, pp. 160-173, Oct. 2020.
  • [9] S. Zhang, Q. Deng, and Z. Ding. “Introducing Graph Signal Processing over Multilayer Networks: Theoretical Foundations and Frequency Analysis”, arXiv:2108.13638, 2022. [available online: https://github.com/zsy93/Introducing-Graph-Signal-Processing-over-Multilayer-Networks/blob/main/ms.pdf]
  • [10] A. Afshar, J. C. Ho, B. Dilkina, I. Perros, E. B. Khalil, L. Xiong, and V. Sunderam, “Cp-ortho: an orthogonal tensor factorization framework for spatio-temporal data,” in Proceedings of the 25th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, Redondo Beach, CA, USA, Jan. 2017, p. 67.
  • [11] L. De Lathauwer, B. De Moor, and J. Vandewalle, “A multilinear singular value decomposition,” SIAM Journal on Matrix Analysis and Applications, vol. 21, no. 4, pp. 1253-1278, Jan. 2000.
  • [12] T. N. Kipf and M. Welling, “Semi-supervised classification with graph convolutional networks,” ICLR, Toulon, France, Apr. 2017.
  • [13] D. I. Shuman, S. K. Narang, P. Frossard, A. Ortega and P. Vandergheynst, “The emerging field of signal processing on graphs: Extending high-dimensional data analysis to networks and other irregular domains,” in IEEE Signal Processing Magazine, vol. 30, no. 3, pp. 83-98, May 2013.
  • [14] J. Shi, and J. M. Moura, “Graph signal processing: Modulation, convolution, and sampling,” arXiv preprint arXiv:1912.06762, Dec. 2019.
  • [15] T. G. Kolda and B. W. Bader, “Tensor decompositions and applications,” SIAM Review, vol. 51, no. 3, pp. 455-500, Aug. 2009.
  • [16] S. Chen, R. Varma, A. Sandryhaila and J. Kovačević, “Discrete signal processing on graphs: sampling theory,” in IEEE Transactions on Signal Processing, vol. 63, no. 24, pp. 6510-6523, Dec. 2015.
  • [17] S. Zhang, Z. Ding and S. Cui, “Introducing hypergraph signal processing: theoretical foundation and practical applications,” in IEEE Internet of Things Journal, vol. 7, no. 1, pp. 639-660, Jan. 2020.
  • [18] A. K. Cherri, and M. A. Karim, “Optical symbolic substitution: edge detection using Prewitt, Sobel, and Roberts operators,” Applied Optics, vol. 28, no. 21, pp. 4644-4648, Nov. 1989.
  • [19] L. Ding, and A. Goshtasby, “On the Canny edge detector,” Pattern Recognition, vol. 34, no. 3, pp. 721-725, Mar. 2001.
  • [20] S. Zhang, S. Cui and Z. Ding, “Hypergraph-based image processing,” 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, Oct. 2020, pp. 216-220.
  • [21] U. Von Luxburg, “A tutorial on spectral clustering,” Statistics and computing, vol. 17, no. 4, pp. 395-416, Aug. 2007.