This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Fixed-Point Centrality for Networks

Shuang Gao Shuang Gao is with the Simons Institute for the Theory of Computing at University of California, Berkeley, CA, USA, 94720. Email: [email protected][email protected]. The author would like to thank Peter E. Caines and Minyi Huang for valuable discussions and feedback on this work, and thank Matthew O. Jackson for conversations about several important references related to this work and for sharing his insights.The author gratefully acknowledge the supported by the U.S. ARL and ARO grant W911NF1910110, the U.S. AFOSR grant FA9550-19-1-0138, and the Simons-Berkeley Research Fellowship.
Abstract

This paper proposes a family of network centralities called fixed-point centralities. This centrality family is defined via the fixed point of permutation equivariant mappings related to the underlying network. Such a centrality notion is immediately extended to define fixed-point centralities for infinite graphs characterized by graphons. Variation bounds of such centralities with respect to the variations of the underlying graphs and graphons under mild assumptions are established. Fixed-point centralities connect with a variety of different models on networks including graph neural networks, static and dynamic games on networks, and Markov decision processes.

I Introduction

Centrality which quantifies the “importance” or “influence” of nodes on networks is a useful concept in social network analysis [1, 2, 3] and it also finds applications in biological, technological and economics networks (see e.g. [4, 5, 6, 3]). Plenty of centralities with different properties are defined for different problems (see e.g. [7, 8]), such as, degree centrality, eigencentrality [9], Katz-Bonacich centrality [10, 11, 12], PageRank centrality [13], Shapley value [14], closeness centrality [15], betweenness centrality [16], diffusion centrality [17], among others. These centralities provide a collection of different quantitative measures for the “importance” or “influence” of nodes on networks associated with various application contexts. For instance, the quality of a website may be modelled by the PageRank centrality [13], the importance of individuals on social influence networks may be reflected by eigencentrality in [9], equilibrium actions of certain static network games correspond to Katz-Bonacich centrality [6], contribution values in a coalition game may be represented by Shapley values [14], and so on. Many social, technological and biological networks are growing and varying in terms of nodes and (or) connections and hence centrality values may vary accordingly. Properties of such variations of centrality values with respect to the variations of graphs (see [18]) motivate the current work. A second motivation is to identify a suitable centrality notion for dynamic game problems on networks and graphons ([19, 20, 21, 22]). A third motivation is the search of a class of new centralities for centrality-weighted opinion dynamics models proposed in [23].

I-A Related work

The formulation of fixed-point centrality in this paper follows the idea of the seminal work on graph neural network models ([24, 25]) in using fixed points of some underlying mappings associated with networks. Fixed-point characterizations find applications in many problems in data science, including graph neural networks ([24, 25]), implicit neural networks ([26, 27, 28, 29]), deep equilibrium models [30], among others [31]. A first salient feature of fixed-point centralities that distinguishes themselves from these models above is that permutation equivariance properties must be satisfied. Another salient feature of fixed-point centralities is that the values of fixed-point centralities are restricted to real numbers to allow natural rankings of the nodes and are restricted to non-negative numbers to allow interpretations (after normalizations) as probability distributions. Furthermore, the current paper focuses on variations of the fixed point (centralities) with respect to the variations of graph structures and weights, which differs from [24, 25, 26, 27, 28, 29, 30].

Centralities and graph neural networks are respectively generalized in [18] and [32] to those for infinite graphs characterized by graphons (developed in [33, 34, 35] to characterize dense graph sequences and their limits). The work [18] studies the eigencentrality, PageRank centrality, Katz-Bonacich centrality of symmetric graphs generated from graphons and establishes the rate of convergence of these centralities to the associated graphon centralities. The fixed-point centrality for graphon in the current paper provides a unified view towards these centralities. The graphon versions of graph neural networks as approximations or generalization models of graph neural networks are proposed and analyzed in [32]. One modelling difference is that the graphon neural networks are characterized by layered structures in [32] whereas in the current paper fixed-point equilibrium structures are employed.

I-B Contribution

We propose the “fixed-point centrality”, which is a class of centralities that can be constructed via a (permutation equivariant) fixed-point mapping associated with the underlying graph. This class of centralities unifies many different centralities (including PageRank centrality, eigencentrality, and Katz-Bonacich centrality) and furthermore it connects to a variety of different problems including graph neural networks [24], and LQG mean field games on networks [20]. In addition, fixed-point centralities are applicable to a broader class of graphs, whether they are undirected or directed, unweighted or weighted (with possibly negative weights), finite or infinite. Moreover, variation bounds of fixed-point centralities with respect to the variations of the underlying graphs are established under mild assumptions following a rather simple idea based on fixed-point analysis.

Notation: \mathds{R} and 0\mathds{R}_{\geq 0} denote respectively reals and non-negative reals. For An×nA\in\mathds{R}^{n\times n}, 𝒢(A)\mathcal{G}(A) denotes the graph with the adjacency matrix AA and the node set [n]{1,,n}[n]\triangleq\{1,...,n\}. 𝒢(V,E)\mathcal{G}(V,E) denotes the graph with vertex set VV and edge set EV×VE\subset V\times V. For a vector vnv\in\mathds{R}^{n}, span(v){αv:αn}\text{span}(v)\triangleq\{\alpha v:\alpha\in\mathds{R}^{n}\}. 𝟏n\mathbf{1}_{n} denotes the nn-dimensional column vector of 11s and 𝟏\mathbf{1} denotes the function defined over [0,1][0,1] with 𝟏α=1\mathbf{1}_{\alpha}=1 for all α[0,1]\alpha\in[0,1]. We use the word “network” to refer to an interconnected group or system where the connection structures along with weights can be characterized by some graph 𝒢(A)\mathcal{G}(A). For a vector vnv\in\mathds{R}^{n}, diag(v)\text{diag}(v) denotes the n×nn\times n diagonal matrix with the elements of vv on the main diagonal, [v]i[v]_{i} (or viv_{i}) denotes the iith element of vv, vp(i=1n|vi|p)1/p\left\|{v}\right\|_{p}\triangleq\left(\sum_{i=1}^{n}\left|v_{i}\right|^{p}\right)^{1/p} with 1p<1\leq p<\infty, and vmaxi[n]|vi|.\|v\|_{\infty}\triangleq\max_{i\in[n]}|v_{i}|. For a matrix An×nA\in\mathds{R}^{n\times n}, [A]ij[A]_{ij} (or aija_{ij}) denotes the ijijth element of AA and Apsupvn,v0Avpvp\|A\|_{p}\triangleq\sup_{v\in\mathds{R}^{n},v\neq 0}\frac{\|Av\|_{p}}{\|v\|_{p}} with 1p1\leq p\leq\infty. We note that A1=max1jni=1m|aij|andA2=λmax(AA).\|A\|_{1}=\max_{1\leq j\leq n}\sum_{i=1}^{m}|a_{ij}|~{}\text{and}~{}\|A\|_{2}={\sqrt{\lambda_{\max}\left(A^{*}A\right)}}.

II Preliminaries on Centralities

A centrality for a network characterized by a graph 𝒢(V,E)\mathcal{G}(V,E) is a mapping ρ:V0\rho:V\to\mathds{R}_{\geq 0} that provides a quantification of “importance” or “influence” of nodes on the network. It is worth emphasizing that the “importance” or “influence” of nodes on networks is defined differently under different application contexts. Hence for the same graph structure and graph weights, various centralities can be defined and may be very different from one another. The choice of the range 0\mathds{R}_{\geq 0} from centralities allows a natural ranking of nodes. The fundamental idea of centralities is to summarize the information about a two-variable function characterized by a graph (or a matrix) into a one-variable function characterized by a centrality (or a vector).

We review several centralities related to the current paper.

II-A Centralities for Finite Networks

Consider a graph 𝒢(A)\mathcal{G}(A) with non-negative adjacency matrix A=[aij]n×nA=[a_{ij}]\in\mathds{R}^{n\times n}. Depending on AA, the graph 𝒢(A)\mathcal{G}(A) may be directed or undirected, and weighted or unweighted.

  1. (E1)

    Eigencentrality (proposed in [9]): Assume the largest eigenvalue λ1\lambda_{1} of AA is simple (i.e. λ1\lambda_{1} has multiplicity 11). Then the eigencentrality of 𝒢(A)\mathcal{G}(A) is given by

    ρi=[limk(1λ1A)k𝟏n]i,i[n].\rho_{i}=\left[\lim_{k\to\infty}\Big{(}\frac{1}{\lambda_{1}}A^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}\Big{)}^{k}\mathbf{1}_{n}\right]_{i},\quad i\in[n].

    An equivalent form in terms of local connections is

    ρi=1λ1j=1Najiρj,i[n],i.e.ρ=1λ1Aρ.\rho_{i}=\frac{1}{\lambda_{1}}\sum_{j=1}^{N}a_{ji}\rho_{j},~{}~{}i\in[n],\quad\text{i.e.}\quad\rho=\frac{1}{\lambda_{1}}A^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}\rho.
  2. (E2)

    Katz-Bonacich centrality with α(0,1)\alpha\in(0,1) (proposed in [10] and generalized in [11, 12]): Let α<A21\alpha<\|A\|_{2}^{-1}. One (simplest) Katz-Bonacich centrality is given by

    ρi\displaystyle\rho_{i} =k=0j=1nαk[Ak]ji=[k=0αkAk𝟏n]i,i[n],\displaystyle=\sum_{k=0}^{\infty}\sum_{j=1}^{n}\alpha^{k}[A^{k}]_{ji}=\left[\sum_{k=0}^{\infty}\alpha^{k}A^{{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}^{k}}\mathbf{1}_{n}\right]_{i},\quad i\in[n],

    where the upper bound of α\alpha ensures the boundedness of the infinite series. An equivalent form in terms of local connections is given by

    ρi=αj=1najiρj+1,i[n],i.e.ρ=αAρ+𝟏n,\rho_{i}=\alpha\sum_{j=1}^{n}a_{ji}\rho_{j}+1,~{}i\in[n],\quad\text{i.e.}\quad\rho=\alpha A^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}\rho+\mathbf{1}_{n},

    and the equivalent explicit form is ρ=(1αA)1𝟏n.\rho=(1-\alpha A^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}})^{-1}\mathbf{1}_{n}.

  3. (E3)

    PageRank centrality (proposed in [13]): Consider a network of webpages, where each node represents a webpage, and aji=1a_{ji}=1 if there is a hyperlink from webpage jj to ii and aij=0a_{ij}=0 otherwise [13]. PageRank centrality with α(0,1)\alpha\in(0,1) is given by

    ρi=αj=1najiρjdj+1αn,dj=i=1naji,i[n],\rho_{i}=\alpha\sum_{j=1}^{n}a_{ji}{\frac{\rho_{j}}{d_{j}}}+{\frac{1-\alpha}{n}},\quad d_{j}=\sum_{i=1}^{n}a_{ji},\quad i\in[n],

    where αajidjρj+(1α)n2\alpha\frac{a_{ji}}{{d_{j}}}\rho_{j}+(1-\alpha)n^{-2} is the probability of jumping from node jj to node ii in the steady state of the associated random walk. In equivalent forms, PageRank centrality ρ\rho satisfies

    ρ=αAD1ρ+1αn𝟏n,withDdiag(d1,,dn)\rho=\alpha A^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}D^{-1}\rho+\frac{1-\alpha}{n}\mathbf{1}_{n},~{}~{}\text{with}~{}D\triangleq\textup{diag}(d_{1},...,d_{n})

    and the explicit computation form is given by

    ρ=1αn(IαAD1)1𝟏n.\rho=\frac{1-\alpha}{n}(I-\alpha A^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}D^{-1})^{-1}\mathbf{1}_{n}.

    PageRank centrality can be interpreted as the steady state distribution of random walks on the network.

II-B Centralities for Graphons

Graphons are defined as bounded symmetric measurable functions 𝐀:[0,1]2[0,1]\mathbf{A}:[0,1]^{2}\to[0,1], which, roughly speaking, can be viewed as the “adjacency matrix” of graphs with the vertex set [0,1][0,1] (see [36]). Let 𝒲0\mathcal{W}_{0} denote the set of graphons with the range [0,1][0,1]. A graphon 𝐀𝒲0\mathbf{A}\in\mathcal{W}_{0} can be interpreted as an integral operator (for instance from L2([0,1])L^{2}([0,1]) to L2([0,1])L^{2}([0,1])) as follows:

(𝐀𝐯)()=[0,1]𝐀(,α)𝐯α𝑑α,𝐯L2([0,1]).(\mathbf{A}\mathbf{v})(\cdot)=\int_{[0,1]}\mathbf{A}(\cdot,\alpha)\mathbf{v}_{\alpha}d\alpha,\quad\mathbf{v}\in L^{2}([0,1]).

The definitions of eigenvector, PageRank and Katz-Bonacich centralities for graphons in [18] are summarized below. Consider a graphon 𝐀𝒲0\mathbf{A}\in\mathcal{W}_{0}.

  1. (E4)

    The graphon eigencentrality for 𝐀\mathbf{A} is given by

    ρ=1λ1𝐀ρ,(𝐀ρ)()[0,1]𝐀(,α)ρα𝑑α,\rho=\frac{1}{\lambda_{1}}\mathbf{A}\rho,\quad(\mathbf{A}\rho)(\cdot)\triangleq\int_{[0,1]}\mathbf{A}(\cdot,\alpha)\rho_{\alpha}d\alpha,

    where ρ\rho denotes the eigenfunction in L2([0,1])L^{2}([0,1]) associated to the largest eigenvalue λ1\lambda_{1} of 𝐀\mathbf{A} and λ1\lambda_{1} is assumed to have multiplicity 11.

  2. (E5)

    The graphon Katz-Bonacich centrality with α(0,1)\alpha\in(0,1) for 𝐀\mathbf{A} is defined by one of the equivalent forms:

    ρ=k=1αk𝐀k𝟏,ρ=(Iα𝐀)1𝟏,or ρ=α𝐀ρ+𝟏\displaystyle\rho=\sum_{k=1}^{\infty}\alpha^{k}\mathbf{A}^{k}\mathbf{1},~{}\rho=(I-\alpha\mathbf{A})^{-1}\mathbf{1},\text{or }\rho=~{}\alpha\mathbf{A}\rho+\mathbf{1}

    where α<1λ1\alpha<\frac{1}{\lambda_{1}} and λ1\lambda_{1} is the largest eigenvalue of 𝐀\mathbf{A}.

  3. (E6)

    The graphon PageRank centrality with α(0,1)\alpha\in(0,1) for 𝐀\mathbf{A} is given by

    ρ=α𝐀𝐃1ρ+(1α)𝟏,𝐀𝒲0,\rho=\alpha\mathbf{A}\odot\mathbf{D}^{-1}\rho+({1-\alpha})\mathbf{1},\quad\mathbf{A}\in\mathcal{W}_{0}, (1)

    where 𝐃(x)=[0,1]𝐀(y,x)𝑑y,\mathbf{D}(x)=\int_{[0,1]}\mathbf{A}(y,x)dy, and (𝐀𝐃1)(x)=𝐀(x,y)𝐃(y)(\mathbf{A}\odot\mathbf{D}^{-1})(x)=\frac{\mathbf{A}(x,y)}{\mathbf{D}(y)} if 𝐃(y)0\mathbf{D}(y)\neq 0, and zero otherwise. Equivalent representation forms are as follows: ρ=(1α)(Iα𝐀𝐃1)1𝟏\rho=(1-\alpha)(I-\alpha\mathbf{A}\odot\mathbf{D}^{-1})^{-1}\mathbf{1} and ρ=(1α)k=0(α𝐀𝐃1)k𝟏.\rho=(1-\alpha)\sum_{k=0}^{\infty}\big{(}\alpha\mathbf{A}\odot\mathbf{D}^{-1}\big{)}^{k}\mathbf{1}.

Proposition 1

The graphon PageRank centrality ρ\rho with α(0,1)\alpha\in(0,1) is a probability density function over [0,1][0,1].

Proof

From the equivalent form ρ=(1α)k=0(α𝐀𝐃1)k𝟏\rho=(1-\alpha)\sum_{k=0}^{\infty}\big{(}\alpha\mathbf{A}\odot\mathbf{D}^{-1}\big{)}^{k}\mathbf{1}, we obtain that ρ(x)0\rho(x)\geq 0 for all x[0,1]x\in[0,1] for 𝐀𝒲0\mathbf{A}\in\mathcal{W}_{0}. Furthermore, based on (1), we verify that

𝟏,ρ=𝟏,α𝐀𝐃1ρ+(1α)𝟏,𝟏=1.\langle\mathbf{1},\rho\rangle=\langle\mathbf{1},\alpha\mathbf{A}\odot\mathbf{D}^{-1}\rho\rangle+(1-\alpha)\langle\mathbf{1},\mathbf{1}\rangle=1.

Thus ρ\rho is a probability density.

III Fixed-Point Centrality for Finite Networks

A permutation matrix is a square matrix that has exactly one element of 11 in every row and every column and 0s elsewhere. An n×nn\times n-dimensional permutation matrix PπP_{\pi} can be obtained by permuting the rows of an n×nn\times n identity matrix according to the permutation map π:[n][n]\pi:[n]\to[n]. For any permutation map π:[n][n]\pi:[n]\to[n], its associated permutation matrix PπP_{\pi} is orthonormal, that is, PπPπ=IP_{\pi}^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}P_{\pi}=I.

Definition 1 (Permutation Equivariance)

A mapping f(,):n×n×nnf(\cdot,\cdot):\mathds{R}^{n\times n}\times\mathds{R}^{n}\to\mathds{R}^{n} is permutation equivariant with respect to a permutation map π:[n][n]\pi:[n]\to[n] if

Pπf(A,ρ)=f(PπAPπ,Pπρ),ρn,An×nP_{\pi}f(A,\rho)=f(P_{\pi}AP^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}_{\pi},P_{\pi}\rho),\quad\forall\rho\in\mathds{R}^{n},~{}{\forall A\in\mathds{R}^{n\times n}}

where PπP_{\pi} is the permutation matrix corresponding to π\pi. A mapping f(,):n×n×nnf(\cdot,\cdot):\mathds{R}^{n\times n}\times\mathds{R}^{n}\to\mathds{R}^{n} is permutation equivariant if it is permutation equivariant with respect to all permutation maps π:[n][n]\pi:[n]\to[n].

Definition 2 (Permutation Invariance)

A mapping f(,):n×n×nnf(\cdot,\cdot):\mathds{R}^{n\times n}\times\mathds{R}^{n}\to\mathds{R}^{n} is permutation invariant with respect to permutation map π:[n][n]\pi:[n]\to[n] if

f(A,ρ)=f(PπAPπ,Pπρ),ρn,An×nf(A,\rho)=f(P_{\pi}AP^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}_{\pi},P_{\pi}\rho),~{}~{}{\forall\rho\in\mathds{R}^{n},~{}\forall A\in\mathds{R}^{n\times n}}

where PπP_{\pi} is the permutation matrix corresponding to π\pi. A mapping f(,):n×n×nnf(\cdot,\cdot):\mathds{R}^{n\times n}\times\mathds{R}^{n}\to\mathds{R}^{n} is permutation invariant if it is permutation invariant with respect to all permutation maps π:[n][n]\pi:[n]\to[n].

Similarly, a mapping g():nng(\cdot):\mathds{R}^{n}\to\mathds{R}^{n} is permutation equivariant (resp. permutation invariant) if Pπg(ρ)=g(Pπρ)P_{\pi}g(\rho)=g(P_{\pi}\rho) (resp. g(ρ)=g(Pπρ)g(\rho)=g(P_{\pi}\rho)) for all ρn\rho\in\mathds{R}^{n} and all permutation maps π:[n][n]\pi:[n]\to[n].

Permutation equivalence and permutation invariance are important properties of many functions associated with DeepSets [37] and graph neural networks [38]. Consider a network, the structure of which is characterized by a graph 𝒢(A)\mathcal{G}(A) with the adjacency matrix An×nA\in\mathds{R}^{n\times n} (which may have negative weights) and the node set [n][n]. SS denotes a set of nodal features and SnS^{n} denotes its nn-fold Cartesian product. Let SnS^{n} be associated with a metric dd.

Definition 3 (Fixed-Point Centrality)

A centrality ρ:[n]0\rho:[n]\to\mathds{R}_{\geq 0} is a fixed-point centrality for 𝒢(A)\mathcal{G}(A) associated with the feature space (Sn,d)(S^{n},d) if there exists a permutation equivariant mapping f(,):n×n×SnSnf(\cdot,\cdot):\mathds{R}^{n\times n}\times S^{n}\to S^{n}, a permutation equivariant mapping g():Sn0ng(\cdot):S^{n}\to\mathds{R}^{n}_{\geq 0}, and a unique xSnx\in S^{n} under the metric dd such that

x=f(A,x),xSn,\displaystyle x=f(A,x),~{}~{}x\in S^{n}, (2)
ρ=g(x),ρ0n.\displaystyle\rho=g(x),~{}~{}~{}~{}\rho\in\mathds{R}^{n}_{\geq 0}.

We note the (symmetric or asymmetric) adjacency matrix A=[aij]A=[a_{ij}] of 𝒢(A)\mathcal{G}(A) is allowed to have non-negative elements. The choices of ff and gg are contingent to the network application context and hence different fixed-point centralities may be associated with the same underlying graph 𝒢(A)\mathcal{G}(A).

The existence of the fixed-point feature is assumed in the definition of fixed-point centrality, which, with extra assumptions, can be established via various fixed-point theorems [39] (see, for instance, [40] based on Kakutani fixed-point theorem and [25] based on Banach fixed-point theorem). The uniqueness of the fixed-point feature depends on the properties of both AA and ff, as it is determined by f(A,):SnSnf(A,\cdot):S^{n}\to S^{n}. Thus, for the same permutation equivariant mapping f(,)f(\cdot,\cdot), a different AA may result in the non-uniqueness (or even non-existence) of the fixed-point feature. To enforce uniqueness of the fixed-point feature (when it exists), one way is to select a suitable feature set SS along with the metric dd for the product space SnS^{n} such that uniqueness is defined up to equivalent classes (see Remark 1 below for an example).

Remark 1 (Linear Case)

When SS is a vector space and f(A,)f(A,\cdot) is a linear function from SnS^{n} to SnS^{n} (that is, f(A,x1+x2)=f(A,x1)+f(A,x2)f(A,x_{1}+x_{2})=f(A,x_{1})+f(A,x_{2}) and f(A,αx1)=αf(A,x1)f(A,\alpha x_{1})=\alpha f(A,x_{1}) for any x1,x2Snx_{1},x_{2}\in S^{n}, α\alpha\in\mathds{R}) for any An×nA\in\mathds{R}^{n\times n}, the unique xSnx\in S^{n} in (2) should be interpreted as the unique 11-dimensional subspace, or in other words, xSnx\in S^{n} is unique up to its linear span; a formal way to have the uniqueness is to extend the feature vector space SnS^{n} to the Grassmannian Gr(1,Sn)Gr(1,S^{n}) (i.e. the space of 11-dimensional linear subspaces in SnS^{n}) and use a distance for Gr(1,Sn)Gr(1,S^{n}) (see e.g. [41]).

Fixed-point centralities can be viewed as a specialization of graph neural network models in [24, 25] to the case where outputs are characterized by non-negative reals and initial nodal labels there are homogenous. Centralities in 0\mathds{R}_{\geq 0} naturally allow ranking nodes according to their centrality values, whereas in general outputs of graph neural networks require extra constructions to allow such ranking.

III-A Examples of Fixed-Point Centrality

Proposition 2

Eigencentrality, Katz-Bonacich centrality, and PageRank centrality are fixed-point centralities.

Proof

The proof is by identifying the functions ff and gg following the definition of fixed-point centrality. The mapping gg in (2) is specialized to the identify mapping from nn\mathds{R}^{n}\to\mathds{R}^{n} (i.e. ρ=x\rho=x) for Katz-Bonacich centrality, PageRank centrality and eigencentrality. For Katz-Bonacich centrality, f(A,x)=αAx+𝟏n,α(0,A21).f(A,x)=\alpha A^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}x+\mathbf{1}_{n},~{}\alpha\in(0,~{}\|A\|_{2}^{-1}). We observe that αA2<1\|\alpha A^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}\|_{2}<1 and hence f(A,)f(A,\cdot) for Katz-Bonacich centrality is a contraction from n\mathds{R}^{n} to n\mathds{R}^{n} under the vector 2-norm. For PageRank centrality, α(0,1)\alpha\in(0,1), f(A,x)=αAdiag(A𝟏n)1x+1αn𝟏n.f(A,x)=\alpha A^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}\text{diag}(A^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}\mathbf{1}_{n})^{-1}x+\frac{1-\alpha}{n}\mathbf{1}_{n}. We note that αAdiag(A𝟏n)11=α<1\|\alpha A^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}\text{diag}(A^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}\mathbf{1}_{n})^{-1}\|_{1}=\alpha<1 and hence f(A,)f(A,\cdot) for PageRank centrality is a contraction under vector 1-norm from n\mathds{R}^{n} to n\mathds{R}^{n}. The existence and uniqueness of fixed-point features for these two cases above are immediate via Banach fixed-point theorem. For the case with eigencentrality, the largest eigenvalue λ1\lambda_{1} of AA is assumed to have multiplicity 11, and the permutation equivariant mapping is f(A,x)=1λ1Axf(A,x)=\frac{1}{\lambda_{1}}Ax. The fixed-point feature xx is unique up to its linear span as f(A,)f(A,\cdot) is a linear function from n\mathds{R}^{n} to n\mathds{R}^{n} (see Remark 1). The permutation equivariance properties of functions f(,)f(\cdot,\cdot) for these centralities can be easily verified.

Proposition 3

Any eigenvector corresponding to a nonzero simple eigenvalue of AA is a fixed-point centrality for 𝒢(A)\mathcal{G}(A).

Proofs are omitted as readers can readily verify the result.

We emphasize that the choice of SS in (2) can be very general: it can be a set of vectors, matrices, functions, probability distributions, strings, etc. Below we give an example where SS is the space of continuous functions from [0,T][0,T] to q\mathds{R}^{q} denoted by C([0,T];q)C([0,T];\mathds{R}^{q}) with q1q\geq 1.

Proposition 4

The equilibrium nodal cost of LQG Network Mean Field Games [21, Sec. IV-B] with homogenous initial conditions, if the unique equilibrium exists, is a fixed-point centrality.

Proof

Following [21, Prop. 1], the network mean field trajectory denoted by z=(z1,.,zn)z=(z_{1},....,z_{n})^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}} with zi(t)qz_{i}(t)\in\mathds{R}^{q} for t[0,T]t\in[0,T] on a network with nn nodes satisfies

z=Φ(A,z),z(C([0,T];q))n,z=\Phi(A,z),\quad z\in(C([0,T];\mathds{R}^{q}))^{n},

where C([0,T];q)C([0,T];\mathds{R}^{q}) denotes the space of continuous functions from [0,T][0,T] to q\mathds{R}^{q} (endowed with the sup norm) and the permutation equivariant mapping

Φ(,):n×n×(C([0,T];q))n(C([0,T];q))n\Phi(\cdot,\cdot):\mathds{R}^{n\times n}\times(C([0,T];\mathds{R}^{q}))^{n}\to(C([0,T];\mathds{R}^{q}))^{n}

is characterized by a forward-backward coupled differential equation pair [21, Prop. 1]. The equilibrium nodal cost is

ρi=J(zi),ziC([0,T];q),i[n]\rho_{i}=J(z_{i}),\quad z_{i}\in C([0,T];\mathds{R}^{q}),\quad i\in[n]

with the same J():C([0,T];q)+0J(\cdot):C([0,T];\mathds{R}^{q})\to\mathds{R}^{+}\cup 0 for all nodes. Hence it satisfies the definition of fixed-point centrality.

We omit the details of the problem formulation of LQG Network Mean Field Games which requires significant space. Interested readers are referred to [21, Sec. IV-B].

III-B Properties of Fixed-Point Centrality

An automorphism of a (directed or undirected) graph 𝒢(V,E)\mathcal{G}(V,E) is a permutation map π:VV\pi:V\to V that satisfies

(i,j)Eif and only if(π(i),π(j))E,i,jV.(i,j)\in E\quad\text{if and only if}\quad(\pi(i),\pi(j))\in E,~{}\forall i,j\in V.
Proposition 5

Any fixed-point centrality of a graph 𝒢(V,E)\mathcal{G}(V,E) is permutation invariant with respect to any automorphism map of 𝒢(V,E)\mathcal{G}(V,E).

Proof

Let AA denote the adjacency matrix of 𝒢(V,E)\mathcal{G}(V,E). Let AπPπAPπA_{\pi}\triangleq P^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}_{\pi}AP_{\pi} and xπPπxx_{\pi}\triangleq P_{\pi}x, where PπP_{\pi} is the permutation matrix corresponding to the permutation map π:[n][n]\pi:[n]\to[n]. By the definition of an automorphism π\pi, A=AπA=A_{\pi} (that is, the adjacency matrix does not change), and hence the fixed-point feature xx given by x=f(A,x)x=f(A,x) satisfies that

xπ=f(Aπ,xπ)=f(A,xπ).x_{\pi}=f(A_{\pi},x_{\pi})=f(A,x_{\pi}).

In the definition of fixed-point centrality, such fixed-point feature is assumed to be unique. Then x=xπx=x_{\pi}. That is, an automorphism does not change the fixed-point features and hence does not change the fixed-point centrality ρ=g(x)\rho=g(x).

A vertex transitive graph is a graph 𝒢\mathcal{G} satisfying that for any given node pair (i,j)(i,j), there exists some automorphism map ϕi,jΠ\phi^{i,j}\in\Pi such that ϕi,j(i)=j,\phi^{i,j}(i)=j, where Π\Pi denotes the set of permutation mappings π:[n][n]\pi:[n]\to[n]. See [42] for examples of vertex transitive graphs.

Proposition 6 (Vertex Transitive Graphs)

All nodes of a vertex transitive graph share the same fixed-point centrality value, that is, any fixed-point centrality for a vertex transitive graph is permutation invariant.

Proof

Following Prop. 5 and the definition of vertex transitive graphs, we obtain, for each i,j[n]i,j\in[n], there exists some ϕi,jΠ\phi^{i,j}\in\Pi111There may be one ϕi,j\phi^{i,j} for each node pair (i,j)(i,j) instead of one ϕ\phi for all node pairs., such that the fixed-point features satisfy xi=xϕi,j(i)=xjx_{i}=x_{\phi^{i,j}(i)}=x_{j}. This implies that xi=xjx_{i}=x_{j} for all i,j[n]i,j\in[n]. Finally, the permutation equivariance of g()g(\cdot) in (2) leads to the desired result.

Properties in Prop. 5 and Prop. 6 are general properties shared by all existing centralities that depend only on graph structures. These properties may not hold in general for the outputs of graph neural network models ([24, 25]).

III-C Centrality Variations with Respect to Graph Variations

Consider two graphs 𝒢(A)\mathcal{G}(A) and 𝒢(B)\mathcal{G}(B) with the same number of nodes. Let ρA\rho_{A} be a fixed-point centrality for 𝒢(A)\mathcal{G}(A) and ρB\rho_{B} that of 𝒢(B)\mathcal{G}(B) with the same function f(,)f(\cdot,\cdot), that is,

xA=f(A,xA),ρA=g(xA),\displaystyle x_{A}=f(A,x_{A}),\quad\rho_{A}=g(x_{A}), (3)
xB=f(B,xB),ρB=g(xB),\displaystyle x_{B}=f(B,x_{B}),\quad\rho_{B}=g(x_{B}),

where SnS^{n} is specialized to n\mathds{R}^{n}, and f(,):n×n×nnf(\cdot,\cdot):{\mathds{R}}^{n\times n}\times\mathds{R}^{n}\to{\mathds{R}}^{n} and g():nng(\cdot):\mathds{R}^{n}\to\mathds{R}^{n}. (The specialization of SnS^{n} to n\mathds{R}^{n} is for the simplicity of presentation, and it can be relaxed to any normed vector space). In this section we study the conditions under which ρA\rho_{A} and ρB\rho_{B} are close and establish upper bounds of their differences.

Let 𝒰fn\mathcal{U}_{f}\subset\mathds{R}^{n} denote the set of feasible fixed-point features with f(,)f(\cdot,\cdot) in (3). Consider the following assumption.
Assumption (A1): (a) There exists L1>0L_{1}>0 such that for all x𝒰fx\in\mathcal{U}_{f},

f(A,x)f(B,x)L1ABop,\|f(A,x)-f(B,x)\|\leq L_{1}\|A-B\|_{\textup{op}}, (4)

where the operator norm Aopsupvn,v0Avv\|A\|_{\textup{op}}\triangleq\sup_{v\in\mathds{R}^{n},v\neq 0}\frac{\|Av\|}{\|v\|};
(b) For any matrix AA and for any x𝒰fx\in\mathcal{U}_{f}, there exists L0(A,x)0L_{0}(A,x)\geq 0 such that

f(A,xA)f(A,x)L0(A,x)xAx\|f(A,x_{A})-f(A,x)\|\leq L_{0}(A,x)\|x_{A}-x\| (5)

where xA=f(A,xA)x_{A}=f(A,x_{A});
(c) For the given matrix AA,

L0(A)supx𝒰fL0(A,x)<1;L_{0}(A)\triangleq\sup_{x\in\mathcal{U}_{f}}L_{0}(A,x)<1;

(d) There exists Lg>0L_{g}>0 such that for all x,v𝒰fx,v\in\mathcal{U}_{f},

g(x)g(v)Lgxv.\|g(x)-g(v)\|\leq L_{g}\|x-v\|.

We call (A1)-(c) the Contraction Condition for Fixed-Point Centrality for 𝒢(A)\mathcal{G}(A); if, furthermore, 𝒰f\mathcal{U}_{f} is complete under the chosen norm \|\cdot\|, it then gives the existence of a unique fixed-point feature for f(A,)f(A,\cdot) following Banach fixed-point theorem, and one can simply apply fixed-point iterations to identify such fixed-point feature with the given graph 𝒢(A)\mathcal{G}(A).

Remark 2 (Different Choices of Norms)

We note that \|\cdot\| can take any vector p\|\cdot\|_{p} norm, 1p1\leq p\leq\infty, as long as the operator norm op\|\cdot\|_{\textup{op}} in (A1)-(a) is compatible with the chosen vector norm (that is, op=supvn,v0Avpvp\|\cdot\|_{\textup{op}}=\sup_{v\in\mathds{R}^{n},v\neq 0}\frac{\|Av\|_{p}}{\|v\|_{p}}).

For Katz-Bonacich centrality, we choose 22-norm and L0(A)=αA2<1L_{0}(A)=\alpha\|A\|_{\textup{2}}<1 if α<A21\alpha<\|A\|_{\textup{2}}^{-1}. For PageRank centrality, we choose 11-norm and L0(A)=αAdiag(A𝟏n)11<1L_{0}(A)=\alpha\|A^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}\text{diag}(A^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}\mathbf{1}_{n})^{-1}\|_{\textup{1}}<1 if α<Adiag(A𝟏n)111\alpha<\|A^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}\text{diag}(A^{\mathchoice{\raisebox{0.0pt}{$\displaystyle\intercal$}}{\raisebox{0.0pt}{$\textstyle\intercal$}}{\raisebox{0.0pt}{$\scriptstyle\intercal$}}{\raisebox{0.0pt}{$\scriptscriptstyle\intercal$}}}\mathbf{1}_{n})^{-1}\|_{\textup{1}}^{-1}.

Remark 3 (Fixed-Point Centrality in the Linear Case)

The condition L0(A)<1L_{0}(A)<1 in (A1) is not satisfied under 2\|\cdot\|_{2} norm for the (normalized) eigencentrality, as L0(A)=1L_{0}(A)=1 for eigencentrality. To establish the error bound, further spectral properties of the graphs are required (see e.g. [18] via rotation analysis of eigenvectors by perturbations [43]). In general, for fixed-point centralities where f(A,)f(A,\cdot) is a linear function, one should establish the difference of two 1-dimensional subspaces characterized by span(xA)\text{span}(x_{A}) and span(xB)\text{span}(x_{B}). Such difference can be characterized by the angular difference between the two subspaces as follows:

d(xA,xB)=|cos1(|xA,xB|xA2xB2)|{d}(x_{A},x_{B})=\left|\cos^{-1}{\left(\frac{\left|\langle x_{A},x_{B}\rangle\right|}{\|x_{A}\|_{2}\|x_{B}\|_{2}}\right)}\right|

which is a specialization of a Grassmann distance (see [41]) to 1-dimensional subspaces (i.e. Grassmannian Gr(1,n)\textup{Gr}(1,\mathds{R}^{n})). For characterizing such differences between xAx_{A} and xBx_{B} when BB differs from AA by a small perturbation, one may employ the error estimation results in [43].

Theorem 1

Under Assumption (A1) for the fixed-point centrality (3), the following holds

ρAρBL1Lg1L0(A)ABop.\|\rho_{A}-\rho_{B}\|\leq\frac{L_{1}L_{g}}{1-L_{0}(A)}\|A-B\|_{\textup{op}}. (6)

Proof

Following the definition of the fixed-point centrality and Assumption (A1)(a)-(A1)(c),

xA\displaystyle\|x_{A}- xB=f(A,xA)f(B,xB)\displaystyle x_{B}\|=\|f(A,x_{A})-f(B,x_{B})\|
f(A,xA)f(A,xB)+f(A,xB)f(B,xB)\displaystyle\leq\|f(A,x_{A})-f(A,x_{B})\|+\|f(A,x_{B})-f(B,x_{B})\|
=L0(A)xAxB+L1ABop.\displaystyle=L_{0}(A)\|x_{A}-x_{B}\|+L_{1}\|A-B\|_{\textup{op}}.

Hence subtracting L0(A)xAxBL_{0}(A)\|x_{A}-x_{B}\| and then dividing by (1L0(A))(1-L_{0}(A)) on both sides yield

ρAρBL11L0(A)ABop.\|\rho_{A}-\rho_{B}\|\leq\frac{L_{1}}{1-L_{0}(A)}\|A-B\|_{\textup{op}}. (7)

Then employing the condition in Assumption (A1)-(d) yields the desired result.

Remark 4

If f(A,ρ)f(A,\rho) does not depend on ρ\rho, then the fixed-point centrality is trivial and the centrality variation upper bounds above should be treated differently. Such examples include degree, closeness and betweenness centralities.

Centralities can be associated with probability distributions: PageRank centrality is the steady state distribution of random walks on the graph of hyperlinks [13], and degree centrality is used as the probability distribution for forming new connections [44]. To (uniquely) associate the fixed-point centrality with a probability (mass function), we consider the following assumption.
Assumption (A2): The fixed-point centralities are normalized with nonnegative entries, that is,

i[n]ρi=1,ρi0,i[n].\sum_{i\in[n]}\rho_{i}=1,\quad\rho_{i}\geq 0,\quad\forall i\in[n].

Clearly, this implies ρ1i=1n|ρi|=1\|\rho\|_{1}\triangleq\sum_{i=1}^{n}|\rho_{i}|=1.

Remark 5 (Normalization of Centralities)

If a centrality cc does not satisfy the condition (A2) above, it can be normalized via ρi=cii=1nci,i[n].{\rho}_{i}=\frac{c_{i}}{\sum_{i=1}^{n}c_{i}},~{}i\in[n]. This normalization is useful to associate any centrality with a probability distribution. For instance, the degree centrality with normalization ρi=dii=1ndi\rho_{i}=\frac{d_{i}}{\sum_{i=1}^{n}d_{i}} where di=j=1naijd_{i}=\sum_{j=1}^{n}a_{ij}, i[n],i\in[n], is used in scale-free network models [44] to represent the probability of forming new connections. In general, one can introduce a monotone function ϕ():0\phi(\cdot):\mathds{R}\to\mathds{R}_{\geq 0}, such that

ρi=ϕ(ci)i=1nϕ(ci),i[n].{\rho}_{i}=\frac{\phi(c_{i})}{\sum_{i=1}^{n}\phi(c_{i})},\quad i\in[n].

When ϕ()=exp()\phi(\cdot)=\exp(\cdot) (or ϕ()=exp()\phi(\cdot)=\exp(-\cdot)), it is then specialized to the softmax function (or the Boltzmann distribution that maximizes an associated entropy). Such normalizations can be incorporated into the permutation equivariant mapping g()g(\cdot) in the definition of fixed-point centrality in (2).

For a metric space (𝒳,d)(\mathcal{X},d) and p1p\geq 1, let Pp(𝒳){\displaystyle P_{p}(\mathcal{X})} denote the set of all probability measures on 𝒳\mathcal{X} with finite ppth moment. The pp-Wasserstein distance between two probability measures in Pp(𝒳){\displaystyle P_{p}(\mathcal{X})} is defined as follows:

Wp(ρA,ρB)=(infγΓ(ρA,ρB)𝒳×𝒳d(x,y)p𝑑γ(x,y))1pW_{p}(\rho_{A},\rho_{B})=\bigg{(}\inf_{\gamma\in\Gamma(\rho_{A},\rho_{B})}\int_{\mathcal{X}\times\mathcal{X}}d(x,y)^{p}d\gamma(x,y)\bigg{)}^{\frac{1}{p}}

where Γ(ρA,ρB)\Gamma(\rho_{A},\rho_{B}) denotes the set of probability measures on 𝒳×𝒳\mathcal{X}\times\mathcal{X} with marginals ρA\rho_{A} and ρB\rho_{B}.

Proposition 7

Under Assumptions (A1) and (A2), the following holds for the fixed-point centrality in (3):

Wp(ρA,ρB)L1Lg1L0(A)infπΠAπBop,p,W_{p}(\rho_{A},\rho_{B})\leq\frac{L_{1}L_{g}}{1-L_{0}(A)}\inf_{\pi\in\Pi}\|A^{\pi}-B\|_{\textup{op,p}}, (8)

where the matrix operator norm is Aop, pAp\|A\|_{\textup{op, p}}\triangleq\|A\|_{p}.

Proof

Recall from Theorem 1 that

ρAπρBpL1Lg1L0(A)AπBop,p,\|\rho_{A}^{\pi^{*}}-\rho_{B}\|_{p}\leq\frac{L_{1}L_{g}}{1-L_{0}(A)}\|A^{\pi^{*}}-B\|_{\textup{op,p}}, (9)

where π=argminπΠAπBop, p\pi^{*}=\arg\min_{\pi\in\Pi}\|A^{\pi}-B\|_{\text{op, p}}. Furthermore, one can verify that

Wp(ρA,ρB)ρAπρBp,W_{p}(\rho_{A},\rho_{B})\leq\|\rho_{A}^{\pi^{*}}-\rho_{B}\|_{p},

since π\pi^{*} is just a particular transport map. We obtain the desired result by combining the two inequalities above.

When p=2p=2, the operator norm Aop,2\|A\|_{\textup{op},2} is the maximum singular value of AA. Consider the matrix cut norm [45]

AmaxS×T[n]×[n]|iS,jTaij|,An×n\|A\|_{\Box}\triangleq\max_{S\times T\subset[n]\times[n]}\Big{|}\sum_{i\in S,j\in T}a_{ij}\Big{|},\quad A\in\mathds{R}^{n\times n} (10)

(without the scaling factor 1n2\frac{1}{n^{2}} used in [36, p.127]).

Lemma 1

The following inequality holds for any symmetric matrix A=[aij]A=[a_{ij}] with elements |aij|1|a_{ij}|\leq 1:

Aop,28A,An×n.\|A\|_{\textup{op},2}\leq\sqrt{8\|A\|_{\Box}},\quad A\in\mathds{R}^{n\times n}.

Proof

For any symmetric matrix AA, the norms Aop, 2\|A\|_{\textup{op, 2}} and A\|A\|_{\Box} scaled respectively by 1n\frac{1}{n} and 1n2\frac{1}{n^{2}} correspond to those graphon norms 𝐀op\|\mathbf{A}\|_{\textup{op}} and 𝐀\|\mathbf{A}\|_{\Box} in [36] for the stepfunction graphon 𝐀\mathbf{A} with uniform partitions associated with AA, where, with an abuse of the notation, 𝐀supS,T[0,1]|S×T𝐀(x,y)𝑑x𝑑y|\|\mathbf{A}\|_{\Box}\triangleq\sup_{S,T\subset[0,1]}\left|\int_{S\times T}\mathbf{A}(x,y)dxdy\right| and 𝐀opsup𝐯L2([0,1])𝐀𝐯2𝐯2\|\mathbf{A}\|_{\textup{op}}\triangleq\sup_{\mathbf{v}\in L^{2}([0,1])}{\frac{\|\mathbf{A}\mathbf{v}\|_{2}}{\|\mathbf{v}\|_{2}}}. Since 𝐀op8𝐀\|\mathbf{\mathbf{A}}\|_{\textup{op}}\leq\sqrt{8\|\mathbf{\mathbf{A}}\|_{\Box}} holds for any graphon in 𝐀𝒲1\mathbf{\mathbf{A}}\in\mathcal{W}_{1} (see [46, Lem. E.6 and Eqn. (4.4)] or [18, Lem. 7]), we obtained the desired result.

Replacing the operator norm by the cut norm in Prop. 7 via the inequality in Lemma 1 yields the following result.

Proposition 8

Consider two symmetric matrices AA and BB. Assume (A1) and (A2) for the fixed-point centrality (3) hold. If |aij|1|a_{ij}|\leq 1 and |bij|1|b_{ij}|\leq 1 for all i,j[n]i,j\in[n], then

W2(ρA,ρB)L1Lg1L0(A)8δ(A,B)W_{2}(\rho_{A},\rho_{B})\leq\frac{L_{1}L_{g}}{1-L_{0}(A)}\sqrt{8\delta_{\Box}(A,B)} (11)

where δ(A,B)infπΠAπB\delta_{\Box}(A,B)\triangleq\inf_{\pi\in\Pi}\|A^{\pi}-B\|_{\Box}, AmaxS×T[n]×[n]|iS,jTaij|\|A\|_{\Box}\triangleq\max_{S\times T\subset[n]\times[n]}\Big{|}\sum_{i\in S,j\in T}a_{ij}\Big{|} and Π\Pi denotes the set of all permutations from [n][n] to [n][n].

IV Fixed-point Centrality for Infinite Networks

Graphons are useful in characterizing and comparing graphs of different size and defining limits of (deterministic or random) graph sequences. This section extends the fixed-point centralities to those for graphons.

IV-A Fixed-Point Centrality for Graphons

Let 𝒲c\mathcal{W}_{c} denote the set of symmetric measurable functions 𝐖:[0,1]2[c,c]\mathbf{W}:[0,1]^{2}\to[-c,c] with c>0c>0. Let S[0,1]S^{[0,1]} denote the infinite Cartesian product of SS with the index set [0,1][0,1]. Let dd denotes the metric for S[0,1]S^{[0,1]}. Similar to the finite graph case, a centrality for a graphon with the vertex set [0,1][0,1] is defined as the mapping ρ:[0,1]0\rho:[0,1]\to\mathds{R}_{\geq 0} which characterizes the “importance” of nodes on the infinite network associated with the underlying graphon.

Definition 4 (Permutation Equivariant Operator)

An operator f(,):𝒲c×S[0,1]S[0,1]f(\cdot,\cdot):\mathcal{W}_{c}\times S^{[0,1]}\to S^{[0,1]} is permutation equivariant with respect to a measure preserving bijection π:[0,1][0,1]\pi:[0,1]\to[0,1] if

f(𝐀,ρ)π=f(𝐀π,ρπ),ρS[0,1],𝐀𝒲cf(\mathbf{A},\rho)^{\pi}=f(\mathbf{A}^{\pi},\rho^{\pi}),~{}~{}{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\forall\rho\in S^{[0,1]},~{}\forall\mathbf{A}\in\mathcal{W}_{c}} (12)

where 𝐀π(α,β)𝐀(π(α),π(β))\mathbf{A}^{\pi}(\alpha,\beta)\triangleq\mathbf{A}(\pi(\alpha),\pi(\beta)) and ρπ(α)ρ(π(α))\rho^{\pi}(\alpha)\triangleq\rho(\pi(\alpha)) for α,β[0,1]\alpha,\beta\in[0,1]. An operator f(,):𝒲c×S[0,1]S[0,1]f(\cdot,\cdot):\mathcal{W}_{c}\times S^{[0,1]}\to S^{[0,1]} is permutation equivariant if it is permutation equivariant with respect to all measure preserving bijections π:[0,1][0,1]\pi:[0,1]\to[0,1].

Definition 5 (Permutation Invariant Operator)

An operator f(,):𝒲c×S[0,1]S[0,1]f(\cdot,\cdot):\mathcal{W}_{c}\times S^{[0,1]}\to S^{[0,1]} is permutation invariant with respect to a measure preserving bijection π:[0,1][0,1]\pi:[0,1]\to[0,1] if

f(𝐀,ρ)=f(𝐀π,ρπ),ρS[0,1],𝐀𝒲c,f(\mathbf{A},\rho)=f(\mathbf{A}^{\pi},\rho^{\pi}),~{}{\color[rgb]{0,0,0}\definecolor[named]{pgfstrokecolor}{rgb}{0,0,0}\pgfsys@color@gray@stroke{0}\pgfsys@color@gray@fill{0}\forall\rho\in S^{[0,1]},~{}\forall\mathbf{A}\in\mathcal{W}_{c},} (13)

where 𝐀π(α,β)𝐀(π(α),π(β))\mathbf{A}^{\pi}(\alpha,\beta)\triangleq\mathbf{A}(\pi(\alpha),\pi(\beta)) and ρπ(α)ρ(π(α))\rho^{\pi}(\alpha)\triangleq\rho(\pi(\alpha)) for α,β[0,1]\alpha,\beta\in[0,1]. An operator f(,):𝒲c×S[0,1]S[0,1]f(\cdot,\cdot):\mathcal{W}_{c}\times S^{[0,1]}\to S^{[0,1]} is permutation invariant if it is permutation invariant with respect to all measure preserving bijections π:[0,1][0,1]\pi:[0,1]\to[0,1].

Similarly, a mapping g():S[0,1]S[0,1]g(\cdot):S^{[0,1]}\to S^{[0,1]} is permutation equivariant (resp. permutation invariant) if for all measure preserving bijections π:[0,1][0,1]\pi:[0,1]\to[0,1], g(ρ)π=g(ρπ)(resp.g(ρ)=g(ρπ)).g(\rho)^{\pi}=g(\rho^{\pi})~{}(\text{resp.}~{}g(\rho)=g(\rho^{\pi})).

Definition 6 (Graphon Fixed-Point Centrality)

A centrality ρ:[0,1]0\rho:[0,1]\to\mathds{R}_{\geq 0} is a fixed-point centrality for a graphon 𝐀𝒲c\mathbf{A}\in\mathcal{W}_{c} associated with the feature space (S[0,1],d)(S^{[0,1]},d) if there exists a permutation equivariant fixed-point mapping f(,):𝒲c×S[0,1]S[0,1]f(\cdot,\cdot):\mathcal{W}_{c}\times S^{[0,1]}\to S^{[0,1]}, a permutation equivariant mapping g():S[0,1]0g(\cdot):S^{[0,1]}\to\mathds{R}_{\geq 0}, and a unique function 𝐱S[0,1]\mathbf{x}\in S^{[0,1]} under the metric dd, such that

𝐱=f(𝐀,𝐱),\displaystyle\mathbf{x}=f(\mathbf{A},\mathbf{x}), (14)
ρ=g(𝐱),ρα0,α[0,1].\displaystyle\rho=g(\mathbf{x}),\quad\rho_{\alpha}\geq 0,~{}~{}\alpha\in[0,1].

We note that the “uniqueness” of 𝐱\mathbf{x} in the definition above depends on the choice of S[0,1]S^{[0,1]} and the underlying metric dd, and it could mean an equivalent class of functions. For example, if we choose S=S=\mathds{R} and let the set [0,1]\mathds{R}^{[0,1]} be endowed with Lp([0,1])L^{p}([0,1]) norm, then the unique 𝐱Lp([0,1])\mathbf{x}\in L^{p}([0,1]) is interpreted as the equivalent class up to discrepancies on sets with Lebesgue measure zero. Another such example, similar to the finite graph case, is that the “uniqueness” of 𝐱\mathbf{x} when f(𝐀,)f(\mathbf{A},\cdot) is a linear mapping shall be interpreted as the unique subspace spanned by 𝐱\mathbf{x} (see Remark 1).

Proposition 9

Graphon eigencentrality, graphon Katz-Bonacich centrality, and graphon PageRank centrality are graphon fixed-point centralities.

Proposition 10

Any eigenfunction of a graphon operator from L2([0,1])L^{2}([0,1]) to L2([0,1])L^{2}([0,1]) corresponding to a non-zero simple eigenvalue is a graphon fixed-point centrality.

Proposition 11

The equilibrium nodal cost of LQG Graphon Mean Field Games [21, Sec. IV-C] with homogenous initial conditions, if the unique equilibrium exists, is a graphon fixed-point centrality.

In LQG Graphon Mean Field Games [21, Sec. IV-C], the product set S[0,1]S^{[0,1]} is specialized to C([0,T];(L2([0,1]))q)C([0,T];(L^{2}([0,1]))^{q}), where q1q\geq 1 is the dimension of the local state of agents. Proofs of these propositions follow similar arguments as those in the finite network case, and hence are omitted.

IV-B Centrality Variations with Respect to Graphon Variations

Consider two graphons 𝐀\mathbf{A} and 𝐁\mathbf{B} in 𝒲c\mathcal{W}_{c}, and let ρ𝐀\rho_{\mathbf{A}} and ρ𝐁\rho_{\mathbf{B}} be respectively their fixed-point centralities as in (14), that is,

𝐱𝐀=f(𝐀,𝐱𝐀),ρA=g(𝐱𝐀),\displaystyle\mathbf{x}_{\mathbf{A}}=f(\mathbf{A},\mathbf{x}_{\mathbf{A}}),\quad\rho_{A}=g(\mathbf{x}_{\mathbf{A}}), (15)
𝐱𝐁=f(𝐁,𝐱𝐁),ρB=g(𝐱𝐁),\displaystyle\mathbf{x}_{\mathbf{B}}=f(\mathbf{B},\mathbf{x}_{\mathbf{B}}),\quad\rho_{B}=g(\mathbf{x}_{\mathbf{B}}),

where the feature space S[0,1]S^{[0,1]} is specialized to Lp([0,1])L^{p}([0,1]) with p1p\geq 1, and the operators f(,)f(\cdot,\cdot) and g()g(\cdot) are specialized to f(,):𝒲c×Lp([0,1])Lp([0,1])f(\cdot,\cdot):\mathcal{W}_{c}\times L^{p}([0,1])\to L^{p}([0,1]) and g():Lp([0,1])Lp([0,1])g(\cdot):L^{p}([0,1])\to L^{p}([0,1]). Let 𝒰fLp([0,1])\mathcal{U}_{f}\subset L^{p}([0,1]) denote the set of feasible fixed-point features associated with f(,)f(\cdot,\cdot) in (15). Consider the following assumption.
Assumption (A3): (a) There exists L1>0L_{1}>0 such that for all 𝐱𝒰f\mathbf{x}\in\mathcal{U}_{f},

f(𝐀,𝐱)f(𝐁,𝐱)L1𝐀𝐁op,\|f(\mathbf{A},\mathbf{x})-f(\mathbf{B},\mathbf{x})\|\leq L_{1}\|\mathbf{A}-\mathbf{B}\|_{\textup{op}}, (16)

where the operator norm 𝐀opsup𝐱0,𝐱<𝐀𝐱𝐱\|\mathbf{A}\|_{\textup{op}}\triangleq\sup_{\mathbf{x}\neq 0,\|\mathbf{x}\|<\infty}\frac{\|\mathbf{A}\mathbf{x}\|}{\|\mathbf{x}\|}.
(b) For any graphon 𝐀𝒲c\mathbf{A}\in\mathcal{W}_{c} and 𝐱𝒰f\mathbf{x}\in\mathcal{U}_{f}, there exists L0(𝐀,𝐱)0L_{0}(\mathbf{A},\mathbf{x})\geq 0 such that

f(𝐀,𝐱𝐀)f(𝐀,𝐱)L0(𝐀,𝐱)𝐱𝐀𝐱\|f(\mathbf{A},\mathbf{x}_{\mathbf{A}})-f(\mathbf{A},\mathbf{x})\|\leq L_{0}(\mathbf{A},\mathbf{x})\|\mathbf{x}_{\mathbf{A}}-\mathbf{x}\| (17)

where 𝐱A=f(A,𝐱A)\mathbf{x}_{A}=f(A,\mathbf{x}_{A}).
(c) For the given graphon 𝐀\mathbf{A},

L0(𝐀)sup𝐱𝒰fL0(𝐀,𝐱)<1.L_{0}(\mathbf{A})\triangleq\sup_{\mathbf{x}\in\mathcal{U}_{f}}L_{0}(\mathbf{A},\mathbf{x})<1.

(d) There exists Lg>0L_{g}>0 such that for all 𝐱,𝐯𝒰f\mathbf{x},\mathbf{v}\in\mathcal{U}_{f},

g(𝐱)g(𝐯)Lg𝐱𝐯.\|g(\mathbf{x})-g(\mathbf{v})\|\leq L_{g}\|\mathbf{x}-\mathbf{v}\|.
Theorem 2

Under Assumption (A3) for the graphon fixed-point centrality (15), the following holds

ρ𝐀ρ𝐁L1Lg1L0(𝐀)𝐀𝐁op.\|\rho_{\mathbf{A}}-\rho_{\mathbf{B}}\|\leq\frac{L_{1}L_{g}}{1-L_{0}(\mathbf{A})}\|\mathbf{A}-\mathbf{B}\|_{\textup{op}}. (18)

The proof essentially follows the same lines of arguments as those for Theorem 1.
Assumption (A4): The graphon fixed-point centrality ρ\rho satisfies

[0,1]ρα𝑑α=1andρα0,\int_{[0,1]}\rho_{\alpha}d\alpha=1\quad\text{and}\quad\rho_{\alpha}\geq 0, (19)

that is, ρ1[0,1]|ρα|𝑑α=1\|\rho\|_{1}\triangleq\int_{[0,1]}|\rho_{\alpha}|d\alpha=1.

Proposition 12

Under Assumptions (A3) and (A4), the following holds for the fixed-point centrality in (15):

Wp(ρ𝐀,ρ𝐁)L1Lg1L0(𝐀)infϕΦ𝐀ϕ𝐁op,p,W_{p}(\rho_{\mathbf{A}},\rho_{\mathbf{B}})\leq\frac{L_{1}L_{g}}{1-L_{0}(\mathbf{A})}\inf_{\phi\in\Phi}\|\mathbf{A}^{\phi}-\mathbf{B}\|_{\textup{op,p}}, (20)

where Φ\Phi denotes the set of all measure preserving bijections from [0,1][0,1] to [0,1][0,1] and the operator norm is 𝐀op, psup𝐱0,𝐱Lp([0,1])𝐀𝐱p𝐱p\|\mathbf{A}\|_{\textup{op, p}}\triangleq\sup_{\mathbf{x}\neq 0,\mathbf{x}\in L^{p}([0,1])}\frac{\|\mathbf{A}\mathbf{x}\|_{p}}{\|\mathbf{x}\|_{p}}.

Proposition 13

Consider two graphons 𝐀\mathbf{A} and 𝐁\mathbf{B} in 𝒲1\mathcal{W}_{1}. Assume (A3) and (A4) for the graphon fixed-point centrality (15) hold. Then the following holds

W2(ρ𝐀,ρ𝐁)L1Lg1L0(𝐀)8δ(𝐀,𝐁).W_{2}(\rho_{\mathbf{A}},\rho_{\mathbf{B}})\leq\frac{L_{1}L_{g}}{1-L_{0}(\mathbf{A})}\sqrt{8\delta_{\Box}(\mathbf{A},\mathbf{B})}. (21)

where δ(𝐀,𝐁)infϕΦ𝐀ϕ𝐁\delta_{\Box}(\mathbf{A},\mathbf{B})\triangleq\inf_{\phi\in\Phi}\|\mathbf{A}^{\phi}-\mathbf{B}\|_{\Box} and 𝐀supS,T[0,1]|S×T𝐀(x,y)𝑑x𝑑y|\|\mathbf{A}\|_{\Box}\triangleq\sup_{S,T\subset[0,1]}\left|\int_{S\times T}\mathbf{A}(x,y)dxdy\right|, and Φ\Phi denotes the set of all measure preserving bijections ϕ:[0,1][0,1]\phi:[0,1]\to[0,1].

Proofs follow similar arguments as those in Prop.7 and 8.

Remark 6

Any finite undirected graphs can be represented by stepfunction graphons [36, Chp.7.1] and hence the characterization of centrality variations applies to finite graphs as well. Moreover, finite graphs of different size can be compared via their graphon representations as well as the associated fixed-point centralities. Thus, for the undirected graph case, the results above in Prop. 12 and 13 generalize those in Prop. 7 and 8.

V Conclusion

The notion of the fixed-point centrality proposed in the current paper is useful in at least the following ways: (a) it helps identify properties for a large family of centralities and apply similar analysis techniques (e.g. in studying changes of centralities with respect to graph perturbations); (b) the well-established theoretical and numerical results of fixed-point analysis can be readily employed for such centralities; (c) learning and training methods can be readily applied to approximate fixed-point centralities due to its close connection with graph neural networks ([24, 25]).

The connection of fixed-point centrality with LQG mean field games on networks suggests collective multi-agent learning of centralities from the equilibrium cost for (dynamic or repeated) game problems. Fixed-point centralities are also related to certain Markov decision processes if each state is viewed as a node and the value function (typically characterized by the fixed-point of the Bellman operator) is then a mapping from the vertex set to non-negative real number. Details will be discussed in future extensions.

The representation of sparse graph sequences and limits requires extra concepts (e.g. graphings for bounded degree graphs [36] and LpL^{p} graphon for sparse WW-random graphs [47]). Future extensions should formulate fixed-point centralities for sparse graph limit models. Other important future directions include: (a) improving upper bounds for centrality variations by exploring further properties of the permutation equivariant mappings; (b) axoimatizing fixed-point centralities via extra properties of f(,)f(\cdot,\cdot), g()g(\cdot), and the feature space SS similar to that in [8]; (c) analyzing the change of the ranking properties of fixed-point centralities with respect to modification on networks; (d) exploring variational analysis of fixed-point centralities where the underlying graphs are characterized by vertexon-graphons [48].

References

  • [1] K. Faust, “Centrality in affiliation networks,” Social Networks, vol. 19, no. 2, pp. 157–191, 1997.
  • [2] A. Landherr, B. Friedl, and J. Heidemann, “A critical review of centrality measures in social networks,” Business & Information Systems Engineering, vol. 2, no. 6, pp. 371–385, 2010.
  • [3] M. O. Jackson, Social and economic networks.   Princeton university press, 2010.
  • [4] D. Koschützki and F. Schreiber, “Centrality analysis methods for biological networks and their application to gene regulatory networks,” Gene Regulation and Systems Biology, vol. 2, pp. GRSB–S702, 2008.
  • [5] Z. Wang, A. Scaglione, and R. J. Thomas, “Electrical centrality measures for electric power grid vulnerability analysis,” in Proceedings of the 49th IEEE Conference on Decision and Control (CDC), 2010, pp. 5792–5797.
  • [6] C. Ballester, A. Calvó-Armengol, and Y. Zenou, “Who’s who in networks. wanted: The key player,” Econometrica, vol. 74, no. 5, pp. 1403–1417, 2006.
  • [7] M. Newman, Networks.   Oxford university press, 2018.
  • [8] F. Bloch, M. O. Jackson, and P. Tebaldi, “Centrality measures in networks,” arXiv preprint arXiv:1608.05845, 2016.
  • [9] P. Bonacich, “Technique for analyzing overlapping memberships,” Sociological Methodology, vol. 4, pp. 176–185, 1972.
  • [10] L. Katz, “A new status index derived from sociometric analysis,” Psychometrika, vol. 18, no. 1, pp. 39–43, 1953.
  • [11] P. Bonacich, “Power and centrality: A family of measures,” American Journal of Sociology, vol. 92, no. 5, pp. 1170–1182, 1987.
  • [12] P. Bonacich and P. Lloyd, “Eigenvector-like measures of centrality for asymmetric relations,” Social Networks, vol. 23, no. 3, pp. 191–201, 2001.
  • [13] S. Brin and L. Page, “The anatomy of a large-scale hypertextual web search engine,” Computer networks and ISDN systems, vol. 30, no. 1-7, pp. 107–117, 1998.
  • [14] L. S. Shapley, “A value for n-person games, contributions to the theory of games, 2, 307–317,” 1953.
  • [15] A. Bavelas, “Communication patterns in task-oriented groups,” The Journal of the Acoustical Society of America, vol. 22, no. 6, pp. 725–730, 1950.
  • [16] L. C. Freeman, “A set of measures of centrality based on betweenness,” Sociometry, pp. 35–41, 1977.
  • [17] A. Banerjee, A. G. Chandrasekhar, E. Duflo, and M. O. Jackson, “The diffusion of microfinance,” Science, vol. 341, no. 6144, p. 1236498, 2013.
  • [18] M. Avella-Medina, F. Parise, M. T. Schaub, and S. Segarra, “Centrality measures for graphons: Accounting for uncertainty in networks,” IEEE Transactions on Network Science and Engineering, vol. 7, no. 1, pp. 520–537, 2018.
  • [19] T. Başar and G. J. Olsder, Dynamic noncooperative game theory.   SIAM, 1998.
  • [20] S. Gao, P. E. Caines, and M. Huang, “LQG graphon mean field games: Graphon invariant subspaces,” in Proceedings of the 60th IEEE Conference on Decision and Control, Austin, Texas, USA, December 2021, pp. 5253–5260.
  • [21] ——, “LQG graphon mean field games: Analysis via graphon invariant subpaces,” Conditionally accepted by IEEE Transactions on Automatic Control, 2021, arXiv preprint arXiv:2004.00679.
  • [22] P. E. Caines and M. Huang, “Graphon mean field games and their equations,” SIAM Journal on Control and Optimization, vol. 59, no. 6, pp. 4373–4399, 2021.
  • [23] S. Gao, “Centrality-weighted opinion dynamics: Disagreement and social network partition,” in Proceedings of the 60th IEEE Conference on Decision and Control, Austin, Texas, USA, December 2021, pp. 5496–5501.
  • [24] M. Gori, G. Monfardini, and F. Scarselli, “A new model for learning in graph domains,” in Proceedings of the 2005 IEEE International Joint Conference on Neural Networks, vol. 2, no. 2005, 2005, pp. 729–734.
  • [25] F. Scarselli, M. Gori, A. C. Tsoi, M. Hagenbuchner, and G. Monfardini, “The graph neural network model,” IEEE Transactions on Neural Networks, vol. 20, no. 1, pp. 61–80, 2009.
  • [26] F. Gu, H. Chang, W. Zhu, S. Sojoudi, and L. El Ghaoui, “Implicit graph neural networks,” Advances in Neural Information Processing Systems, vol. 33, pp. 11 984–11 995, 2020.
  • [27] L. El Ghaoui, F. Gu, B. Travacca, A. Askari, and A. Tsai, “Implicit deep learning,” SIAM Journal on Mathematics of Data Science, vol. 3, no. 3, pp. 930–958, 2021.
  • [28] S. Jafarpour, A. Davydov, A. Proskurnikov, and F. Bullo, “Robust implicit networks via non-euclidean contractions,” Advances in Neural Information Processing Systems, vol. 34, pp. 9857–9868, 2021.
  • [29] A. Davydov, A. V. Proskurnikov, and F. Bullo, “Non-euclidean contractivity of recurrent neural networks,” arXiv preprint arXiv:2110.08298, 2021.
  • [30] S. Bai, J. Z. Kolter, and V. Koltun, “Deep equilibrium models,” Advances in Neural Information Processing Systems, vol. 32, 2019.
  • [31] P. L. Combettes and J.-C. Pesquet, “Fixed point strategies in data science,” IEEE Transactions on Signal Processing, vol. 69, pp. 3878–3905, 2021.
  • [32] L. Ruiz, L. Chamon, and A. Ribeiro, “Graphon neural networks and the transferability of graph neural networks,” Advances in Neural Information Processing Systems, vol. 33, pp. 1702–1712, 2020.
  • [33] L. Lovász and B. Szegedy, “Limits of dense graph sequences,” Journal of Combinatorial Theory, Series B, vol. 96, no. 6, pp. 933–957, 2006.
  • [34] C. Borgs, J. T. Chayes, L. Lovász, V. T. Sós, and K. Vesztergombi, “Convergent sequences of dense graphs i: Subgraph frequencies, metric properties and testing,” Advances in Mathematics, vol. 219, no. 6, pp. 1801–1851, 2008.
  • [35] ——, “Convergent sequences of dense graphs ii. multiway cuts and statistical physics,” Annals of Mathematics, vol. 176, no. 1, pp. 151–219, 2012.
  • [36] L. Lovász, Large Networks and Graph Limits.   American Mathematical Soc., 2012, vol. 60.
  • [37] M. Zaheer, S. Kottur, S. Ravanbakhsh, B. Poczos, R. Salakhutdinov, and A. Smola, “Deep sets,” arXiv preprint arXiv:1703.06114, 2017.
  • [38] M. M. Bronstein, J. Bruna, T. Cohen, and P. Veličković, “Geometric deep learning: Grids, groups, graphs, geodesics, and gauges,” arXiv preprint arXiv:2104.13478, 2021.
  • [39] R. P. Agarwal, M. Meehan, and D. O’regan, Fixed point theory and applications.   Cambridge University Press, 2001, vol. 141.
  • [40] L. Lyu, B. Fain, K. Munagala, and K. Wang, “Centrality with diversity,” in Proceedings of the 14th ACM International Conference on Web Search and Data Mining, 2021, pp. 644–652.
  • [41] K. Ye and L.-H. Lim, “Schubert varieties and distances between subspaces of different dimensions,” SIAM Journal on Matrix Analysis and Applications, vol. 37, no. 3, pp. 1176–1197, 2016.
  • [42] E. W. Weisstein, “Vertex-Transitive Graph.” [Online]. Available: https://mathworld.wolfram.com/Vertex-TransitiveGraph.html
  • [43] C. Davis and W. M. Kahan, “The rotation of eigenvectors by a perturbation. iii,” SIAM Journal on Numerical Analysis, vol. 7, no. 1, pp. 1–46, 1970.
  • [44] A.-L. Barabási and R. Albert, “Emergence of scaling in random networks,” Science, vol. 286, no. 5439, pp. 509–512, 1999.
  • [45] A. Frieze and R. Kannan, “Quick approximation to matrices and applications,” Combinatorica, vol. 19, no. 2, pp. 175–220, 1999.
  • [46] S. Janson, “Graphons, cut norm and distance, couplings and rearrangements,” New York Journal of Mathematics, vol. 4, 2013.
  • [47] C. Borgs, J. T. Chayes, H. Cohn, and Y. Zhao, “An LpL^{p} theory of sparse graph convergence II: LD convergence, quotients and right convergence,” The Annals of Probability, vol. 46, no. 1, pp. 337–396, 2018.
  • [48] P. E. Caines, “Embedded vertexon-graphons and embedded GMFG systems,” Accepted for presentation at the 61th IEEE Conference on Decision and Control (CDC), 2022.