This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

GSSI, Italy and Dalhousie University, [email protected]://orcid.org/0000-0002-1402-5298 \CopyrightNicola Cotumaccio \ccsdesc[500]Theory of computation Graph algorithms analysis \ccsdesc[500]Theory of computation Pattern matching \fundingThis work was partially funded by Dante Labs.

Acknowledgements.
I thank Nicola Prezza for pointing out the paper [21]. \hideLIPIcs

Prefix Sorting DFAs: a Recursive Algorithm

Nicola Cotumaccio
Abstract

In the past thirty years, numerous algorithms for building the suffix array of a string have been proposed. In 2021, the notion of suffix array was extended from strings to DFAs, and it was shown that the resulting data structure can be built in O(m2+n5/2)O(m^{2}+n^{5/2}) time, where nn is the number of states and mm is the number of edges [SODA 2021]. Recently, algorithms running in O(mn)O(mn) and O(n2logn)O(n^{2}\log n) time have been described [CPM 2023].

In this paper, we improve the previous bounds by proposing an O(n2)O(n^{2}) recursive algorithm inspired by Farach’s algorithm for building a suffix tree [FOCS 1997]. To this end, we provide insight into the rich lexicographic and combinatorial structure of a graph, so contributing to the fascinating journey which might lead to solve the long-standing open problem of building the suffix tree of a graph.

keywords:
Suffix Array, Burrows-Wheeler Transform, FM-index, Recursive Algorithms, Graph Theory, Pattern Matching.
category:

1 Introduction

The suffix tree [30] of a string is a versatile data structure introduced by Weiner in 1973 which allows solving a myriad of combinatorial problems, such as determining whether a pattern occurs in the string, computing matching statistics, searching for regular expressions, computing the Lempel-Ziv decomposition of the string and finding palindromes. The book by Gusfield [17] devotes almost 150 pages to the applications of suffix trees, stressing the importance of these applications in bioinformatics. However, the massive increase of genomic data in the last decades requires space-efficient data structures able to efficiently support pattern matching queries, and the space consumption of suffix trees is too high. In 1990, Manber and Myers invented suffix arrays [24] as a space-efficient alternative to suffix trees. While suffix arrays do not have the full functionality of suffix trees, they still allow solving pattern matching queries. Suffix arrays started a new chapter in data compression, which culminated in the invention of data structures closely related to suffix arrays, notably, the Burrows-Wheeler Transform [4] and the FM-index [13, 15], which have heavily influenced sequence assembly [29].

The impact of suffix arrays has led to a big effort in the attempt of designing efficient algorithms to construct suffix arrays, where “efficient” refers to various metrics (worst-case running time, average running time, space, performance on real data and so on); see [27] for a comprehensive survey on the topic. Let us focus on worst-case running time. Manber and Myers build the suffix array of a string of length nn in O(nlogn)O(n\log n) by means of a prefix-doubling algorithm [24]. In 1997, Farach proposed a recursive algorithm to build the suffix tree of a string in linear time for integer alphabets [11]. In the following years, the recursive paradigm of Farach’s algorithm was used to developed a multitude of linear-time algorithms for building the suffix array [22, 19, 20, 25]. All these algorithms carefully exploit the lexicographic struture of the suffixes of a string, recursively reducing the problem of computing the suffix array of a string to the problem of computing the suffix array of a smaller string (induced sorting).

The problem of solving pattern matching queries not only on strings, but also on labeled graphs, is an active topic of research. Recently, Equi et al. showed that no algorithm can solve pattern matching queries on arbitrary graphs in O(m1ϵ|P|)O(m^{1-\epsilon}|P|) time or O(m|P|1ϵ)O(m|P|^{1-\epsilon}) (where mm is the number of edges, PP is the pattern and ϵ>0\epsilon>0), unless the Orthogonal Vectors hypothesis fails [10, 9]. On the other hand, over the years the idea of (lexicographically) sorting the suffixes of a string has been generalized to graphs, thus leading to compact data structures that are able to support pattern matching queries on graphs. The mechanism behind the suffix array, the Burrows-Wheeler Transform and the FM-index was first generalized to trees [12, 14]; later on, it was generalized to De Brujin graphs [3, 23] (which can be used for Eulerian sequence assembly [18]). Subsequently, it was extended to the so-called Wheeler graphs [16, 1], and finally to arbitrary graphs and automata [8, 6]. The idea of lexicographically sorting the strings reaching the states of an automaton has also deep theoretical consequences in automata theory: for example, it leads to a parametrization of the powerset construction, which implies fixed-parameter tractable algorithms for PSPACE-complete problems such as deciding the equivalence of two non-deterministic finite automata (NFAs) [7].

The case of deterministic finite automata (DFAs) is of particular interest, because in this case the notion of “suffix array” of a DFA has a simple interpretation in terms of strings. Assume that there is a fixed total order \preceq on the alphabet Σ\Sigma of a DFA 𝒜\mathcal{A}, and extend \preceq lexicographically to the set of all infinite strings on Σ\Sigma. Assume that each state has at least one incoming edge. If uu is a state, consider the set IuI_{u} of all infinite strings that can be read starting from uu and following edges in a backward fashion, and let minu\min_{u} and maxu\max_{u} be the lexicograpically smallest (largest, respectively) string in IuI_{u} (see Figure 1). Consider the partial order 𝒜\preceq_{\mathcal{A}} on the set of all states QQ such that for every u,vQu,v\in Q, with uvu\not=v, it holds u𝒜vu\prec_{\mathcal{A}}v if and only if maxuminv\max_{u}\preceq\min_{v}. Then, the partial order 𝒜\preceq_{\mathcal{A}} induces a (partial) permutation of the set of all states that plays the same role of the permutation of text positions induced by the suffix array of a string [8, 21], so 𝒜\preceq_{\mathcal{A}} is essentially the “suffix array” of the DFA. If we are able to compute the partial order 𝒜\preceq_{\mathcal{A}} (and a minimum-size partition of QQ into sets such that the restriction of 𝒜\preceq_{\mathcal{A}} to each set is a total order), then we can efficiently and compactly solve pattern matching queries on the DFA by means of techniques that extend the Burrows-Wheeler transform and the FM-index from strings to graphs. As a consequence, we now have to solve the problem of efficiently building the “suffix array” of a DFA, that is, the problem of computing 𝒜\preceq_{\mathcal{A}}.

11start223344556677aabbccccaaccaacccc#\#
ii mini\min_{i} maxi\max_{i}
11 #####\#\#\#\#\#\dots #####\#\#\#\#\#\dots
22 a####a\#\#\#\#\dots a####a\#\#\#\#\dots
33 b####b\#\#\#\#\dots b####b\#\#\#\#\dots
44 c####c\#\#\#\#\dots c####c\#\#\#\#\dots
55 ca###ca\#\#\#\dots cccccccccc\dots
66 ab###ab\#\#\#\dots ac###ac\#\#\#\dots
77 cc###cc\#\#\#\dots cccccccccc\dots
Figure 1: A DFA 𝒜\mathcal{A}, with the minimum and maximum string reaching each state (we assume #abc\#\prec a\prec b\prec c). The min/max partition is given by {(1,min),(1,max)}<{(2,min),(2,max)}<{(6,min)}<{(6,max)}<{(3,min),(3,max)}<{(4,min),(4,max)}<{(5,min)}<{(7,min)}<{(5,max),(7,max)}\{(1,\min),(1,\max)\}<\{(2,\min),(2,\max)\}<\{(6,\min)\}<\{(6,\max)\}<\{(3,\min),(3,\max)\}<\{(4,\min),(4,\max)\}<\{(5,\min)\}<\{(7,\min)\}<\{(5,\max),(7,\max)\}, meaning that min1=max1min2=max2min6max6min3=max3min4=max4min5min7max5=max7\min_{1}=\max_{1}\prec\min_{2}=\max_{2}\prec\min_{6}\prec\max_{6}\prec\min_{3}=\max_{3}\prec\min_{4}=\max_{4}\prec\min_{5}\prec\min_{7}\prec\max_{5}=\max_{7}.

The first paper on the topic [8] presents an algorithm that builds 𝒜\preceq_{\mathcal{A}} in O(m2+n5/2)O(m^{2}+n^{5/2}) time, where nn is the number of states and mm is the number of edges. In a recent paper [21], Kim et al. describe two algorithms running in O(mn)O(mn) and O(n2logn)O(n^{2}\log n) time. The key observation is that, if we build the min/max-partition of the set of all states, then we can determine the partial order 𝒜\preceq_{\mathcal{A}} in linear time by means of a reduction to the interval partitioning problem. Determining the min/max-partition of the set of all states means picking the string minu\min_{u} and maxu\max_{u} for every state uQu\in Q, and sorting these 2|Q|2|Q| strings (note that some of these strings may be equal, see Figure 1 for an example). As a consequence, we are only left with the problem of determining the min/max-partition efficiently.

The O(n2logn)O(n^{2}\log n) algorithm builds the min/max-partition by generalizing Manber and Myers’s O(nlogn)O(n\log n) algorithm from strings to DFAs. However, since it is possible to build the suffix array of a string in O(n)O(n) time, it is natural to wonder whether it is possible to determine the min/max-partition in O(n2)O(n^{2}) time.

In this paper, we show that, indeed, it is possible to build the min/max-partition in O(n2)O(n^{2}) time by adopting a recursive approach inspired by one of the linear-time algorithms that we have mentioned earlier, namely, Ko and Aluru’s algorithm [22]. As a consequence, our algorithm is asymptotically faster than all previous algorithms.

A long-standing open problem is whether it is possible to define a suffix tree of a graph. Some recent work [5] suggests that it is possible to define data structures that simulate the behavior of a suffix tree by carefully studying the lexicographic structure of the graph (the results in [5] only hold for Wheeler graphs, but we believe that they can be extended to arbitrary graphs). More specifically, it is reasonable to believe that it is possible to generalize the notion of compressed suffix tree of a string [28] to graphs. A compressed suffix tree is a compressed representation of a suffix tree which consists of some components, including a suffix array. We have already seen that 𝒜\preceq_{\mathcal{A}} generalizes the suffix array to a graph structure, and [5] suggests that the remaining components may also be generalized to graphs. The complements of the suffix tree of a string heavily exploit the lexicographic and combinatorial structure of a string. Since the algorithm that we present in this paper deeply relies on the richness of the lexicographic structure (which becomes even more challenging and surprising when switching from a string setting to a graph setting), we believe that our results also provide a solid conceptual contribution towards extending suffix tree functionality to graphs.

We remark that, a few days after we submitted this paper to arXiv, a new arXiv preprint showed how to determine the min/max partition in O(mlogn)O(m\log n) time, where nn is the number of states and mm is the number of edges (this new arXiv preprint was also accepted for publication [2]). If the graph underlying the DFA is sparse, then the algorithm in [2] improves our O(n2)O(n^{2}) algorithm. Since the O(mlogn)O(m\log n) algorithm uses different techniques (it is obtained by adapting Paige and Tarjan’s partition refinement algorithm [26]), we are left with the intriguing open problem of determining whether, by possibly combining the ideas behind our algorithm and the algorithm in [2], it is possible to build the min/max partition in O(m)O(m) time.

Due to space constraints, all proofs can be found in the appendix.

2 Preliminaries

2.1 Relation with Previous Work

In the setting of the previous works on the topic [1, 8, 21], the problem that we want to solve is defined as follows. Consider a deterministic finite automaton (DFA) such that (i) all edges entering the same state have the same label, (ii) each state is reachable from the initial state, (iii) each state is co-reachable, that is, it is either final or it allows reaching a final state, (iv) the initial state has no incoming edges. Then, determine the min/max partition of the set of states (see Section 2.2 for the formal definition of min/max partition, and see Figure 1 for an example).

  • Assumptions (ii) and (iii) are standard assumptions in automata theory, because all states that do not satisfy these assumptions can be removed without changing the accepted language.

  • In this setting, all the non-initial states have an incoming edge, but the initial state has no incoming edges. This implies for some state uu it may hold Iu=I_{u}=\emptyset (remember that IuI_{u} is the set of all infinite strings that can be read starting from uu and following edges in a backward fashion, see the introduction), so Kim et al. [21] need to perform a tedious case analysis which also takes finite strings into account in order to define the min/max-partition (in particular, the minimum and maximum strings reaching the initial state are both equal to the empty string). However, we can easily avoid this complication by means of the same trick used in [5]; we can add a self-loop to the initial state, and the label of the self-loop is a special character #\# smaller than any character in the alphabet. Intuitively, #\# plays the same role as the termination character in the Burrows-Wheeler transform of a string, and since #\# is the smallest character, adding this self-loop does not affect the min/max-partition (see [21] for details).

  • Notice that the initial state and the set of all final states play no role in the definition of the min/max partition; this explains why, more generally, it will be expedient to consider deterministic graphs rather than DFAs (otherwise we would need to artificially add an initial state and add a set of final states when we recursively build a graph in our algorithm). Equivalently, one may assume to work with semiautomata in which the transition function is not necessarily total. This justifies the assumptions that we will make in Section 2.2.

  • Some recent papers [6, 7] have shown that assumptions (i) and (iv) can be removed. The partial order 𝒜\preceq_{\mathcal{A}} is defined analogously, and all the algorithms for building 𝒜\preceq_{\mathcal{A}} that we have mentioned still work. Indeed, if a state uu is reached by edges with the distinct labels, we need to only consider all edges with the smallest label when computing minu\min_{u} and all edges with the largest label when computing maxu\max_{u}; moreover, we once again assume that the initial state has a self loop labeled #\#. The only difference is that assumption (i) implies that mn2m\leq n^{2} (nn being the number of states, mm being the number of edges) because each state can have at most nn incoming edges, but this is no longer true if we remove assumption (i). As a consequence, the running time of our algorithm is no longer O(n2)O(n^{2}) but O(m+n2)O(m+n^{2}) (and the running time of the O(n2logn)O(n^{2}\log n) algorithm in [21] becomes O(m+n2logn)O(m+n^{2}\log n)) because we still need to process all edges in the DFA.

To sum up, all the algorithms for computing 𝒜\preceq_{\mathcal{A}} work on arbitrary DFAs.

2.2 Notation and First Definitions

Let Σ\Sigma be a finite alphabet. We consider finite, edge-labeled graphs G=(V,E)G=(V,E), where VV is the set of all nodes and EV×V×ΣE\subseteq V\times V\times\Sigma is the set of all edges. Up to taking a subset of Σ\Sigma, we assume that all cΣc\in\Sigma label some edge in the graph. We assume that all nodes have at least one incoming edge, and all edges entering the same node uu have the same label λ(u)\lambda(u) (input consistency). This implies that an edge (u,v,a)E(u,v,a)\in E can be simply denoted as (u,v)(u,v), because it must be a=λ(v)a=\lambda(v). In particular, it must be |E||V|2|E|\leq|V|^{2} (and so an O(|E|)O(|E|) algorithm is also a O(|V|2)O(|V|^{2}) algorithm). If we do not know the λ(u)\lambda(u)’s, we can easily compute them by scanning all edges. In addition, we always assume that GG is deterministic, that is, for every uVu\in V and for every aΣa\in\Sigma there exists at most one vVv\in V such that (u,v)E(u,v)\in E and λ(v)=a\lambda(v)=a.

Let Σ\Sigma^{*} be the set of all finite strings on Σ\Sigma, and let Σω\Sigma^{\omega} be the set of all (countably) right-infinite strings on Σ\Sigma. If αΣΣω\alpha\in\Sigma^{*}\cup\Sigma^{\omega} and i1i\geq 1, we denote by α[i]Σ\alpha[i]\in\Sigma the ithi^{\text{th}} character of α\alpha (that is, α=α[1]α[2]α[3]\alpha=\alpha[1]\alpha[2]\alpha[3]\dots). If 1ij1\leq i\leq j, we define α[i,j]=α[i]α[i+1]α[j1]α[j]\alpha[i,j]=\alpha[i]\alpha[i+1]\dots\alpha[j-1]\alpha[j], and if j<ij<i, then α[i,j]\alpha[i,j] is the empty string ϵ\epsilon. If αΣ\alpha\in\Sigma^{*}, then |α||\alpha| is length of α\alpha; for every 0j|α|0\leq j\leq|\alpha| the string α[1,j]\alpha[1,j] is a prefix of α\alpha, and if 0j<|α|0\leq j<|\alpha| it is a strict prefix of α\alpha; analogously, one defines suffixes and strict suffixes of α\alpha. An occurrence of αΣ\alpha\in\Sigma^{*} starting at uVu\in V and ending at uVu^{\prime}\in V is a sequence of nodes u1,u2,,u|α|+1u_{1},u_{2},\dots,u_{|\alpha|+1} of VV such that (i) u1=uu_{1}=u, (ii) u|α|+1=uu_{|\alpha|+1}=u^{\prime}, (iii) (ui+1,ui)E(u_{i+1},u_{i})\in E for every 1i|α|1\leq i\leq|\alpha| and (iv) λ(ui)=α[i]\lambda(u_{i})=\alpha[i] for every 1i|α|1\leq i\leq|\alpha|. An occurrence of αΣω\alpha\in\Sigma^{\omega} starting at uVu\in V is a sequence of nodes (ui)i1(u_{i})_{i\geq 1} of VV such that (i) u1=uu_{1}=u, (ii) (ui+1,ui)E(u_{i+1},u_{i})\in E for every i1i\geq 1 and (iii) λ(ui)=α[i]\lambda(u_{i})=\alpha[i] for every i1i\geq 1. Intuitively, a string αΣΣω\alpha\in\Sigma^{*}\cup\Sigma^{\omega} has an occurrence starting at uVu\in V if we can read α\alpha on the graph starting from uu and following edges in a backward fashion.

In the paper, occurrences of strings in Σω\Sigma^{\omega} will play a key role, while occurrences of strings in Σ\Sigma^{*} will be used as a technical tool. For every uVu\in V, we denote by IuI_{u} the set of all strings in Σω\Sigma^{\omega} admitting an occurrence starting at uu. Since every node has at least one incoming edge, then IuI_{u}\not=\emptyset.

A total order \leq on a set VV if a reflexive, antisymmetric and transitive relation on VV. If u,vVu,v\in V, we write u<vu<v if uvu\leq v and uvu\not=v.

Let \preceq be a fixed total order on Σ\Sigma. We extend \preceq to ΣΣω\Sigma^{*}\cup\Sigma^{\omega} lexicographically. It is easy to show that in every IuI_{u} there is a lexicographically smallest string minu\min_{u} and a lexicographically largest string maxu\max_{u} (for example, it follows from [21, Observation 8]).

We will often use the following immediate observation. Let uVu\in V, and let (ui)i1(u_{i})_{i\geq 1} be an occurrence of minu\min_{u}. Fix i1i\geq 1. Then, (uj)ji(u_{j})_{j\geq i} is an occurrence of minui\min_{u_{i}}, and minu=minu[1,i1]minui\min_{u}=\min_{u}[1,i-1]\min_{u_{i}}.

Let VVV^{\prime}\subseteq V. Let 𝒜\mathcal{A} be the unique partition of VV^{\prime} and let \leq be the unique total order on 𝒜\mathcal{A} such that, for every I,J𝒜I,J\in\mathcal{A} and for every uIu\in I and vJv\in J, (i) if I=JI=J, then minu=minv\min_{u}=\min_{v} and (ii) if I<JI<J, then minuminv\min_{u}\prec\min_{v}. Then, we say that (𝒜,)(\mathcal{A},\leq), or more simply 𝒜\mathcal{A}, is the min-partition of VV^{\prime}. The max-partition of VV^{\prime} is defined analogously. Now, consider the set V×{min,max}V^{\prime}\times\{\min,\max\}, and define ρ((u,min))=minu\rho((u,\min))=\min_{u} and ρ((u,max))=maxu\rho((u,\max))=\max_{u} for every uVu\in V^{\prime}. Let \mathcal{B} be the unique partition of V×{min,max}V^{\prime}\times\{\min,\max\} and let \leq be the unique total order on \mathcal{B} such that, for every I,JI,J\in\mathcal{B} and for every xIx\in I and yJy\in J, (i) if I=JI=J, then ρ(x)=ρ(y)\rho(x)=\rho(y) and (ii) if I<JI<J, then ρ(x)ρ(y)\rho(x)\prec\rho(y). Then, we say that (,)(\mathcal{B},\leq), or more simply \mathcal{B}, is the min/max-partition of VV^{\prime}.

The main result of this paper will be proving that the min/max partition of VV can be determined in O(|V|2)O(|V|^{2}) time.

2.3 Our Approach

Let G=(V,E)G=(V,E) be a graph. We will first show how to build the min-partition of VV in O(n2)O(n^{2}) time, where n=|V|n=|V| (Section 4); then, we will show how the algorithm can be adapted so that it builds the min/max-partition in O(n2)O(n^{2}) time (Section 5).

In order to build a min-partition of VV, we will first classify all minima into three categories (Section 3), so that we can split VV into three pairwise-disjoint sets V1,V2,V3V_{1},V_{2},V_{3}. Then, we will show that in O(n2)O(n^{2}) time:

  • we can compute V1,V2,V3V_{1},V_{2},V_{3} (Section 4.1);

  • we can define a graph G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}) having |V3||V_{3}| nodes (Section 4.2);

  • assuming that we have already determined the min-partition of V¯\bar{V}, we can determine the min-partition of VV (Section 4.3).

Analogously, in O(n2)O(n^{2}) time we can reduce the problem of determining the min-partition of VV to the problem of determining the min-partition of the set of all nodes of a graph having |V1||V_{1}| (not |V3||V_{3}|) nodes (Section 4.4). As a consequence, since min{|V1|,|V3|}|V|/2=n/2\min\{|V_{1}|,|V_{3}|\}\leq|V|/2=n/2, we obtain a recursive algorithm whose running time is given by the recurrence:

T(n)=T(n/2)+O(n2)T(n)=T(n/2)+O(n^{2})

and we conclude that the running time of our algorithm is O(n2)O(n^{2}).

3 Classifying Strings

In [22], Ko and Aluru divide the suffixes of a string into two groups. Here we follow an approach purely based on stringology, without fixing a string or a graph from the start. We divide the strings of Σω\Sigma^{\omega} into three groups, which we call group 1, group 2 and group 3 (Corollary 3.3 provides the intuition behind this choice).

Definition 3.1.

Let αΣω\alpha\in\Sigma^{\omega}. Let aΣa\in\Sigma and αΣω\alpha^{\prime}\in\Sigma^{\omega} such that α=aα\alpha=a\alpha^{\prime}. Then, we define τ(α)\tau(\alpha) as follows:

  1. 1.

    τ(α)=1\tau(\alpha)=1 if αα\alpha^{\prime}\prec\alpha.

  2. 2.

    τ(α)=2\tau(\alpha)=2 if α=α\alpha^{\prime}=\alpha.

  3. 3.

    τ(α)=3\tau(\alpha)=3 if αα\alpha\prec\alpha^{\prime}.

We will constantly use the following characterization.

Lemma 3.2.

Let αΣω\alpha\in\Sigma^{\omega}. Let aΣa\in\Sigma and αΣω\alpha^{\prime}\in\Sigma^{\omega} such that α=aα\alpha=a\alpha^{\prime}. Then:

  1. 1.

    τ(α)=2\tau(\alpha)=2 if and only if α=aω\alpha^{\prime}=a^{\omega}, if and only if α=aω\alpha=a^{\omega}.

  2. 2.

    τ(α)2\tau(\alpha)\not=2 if and only if αaω\alpha^{\prime}\not=a^{\omega}, if and only if αaω\alpha\not=a^{\omega}.

Assume that τ(α)2\tau(\alpha)\not=2. Then, there exist unique cΣ{a}c\in\Sigma\setminus\{a\}, αΣω\alpha^{\prime\prime}\in\Sigma^{\omega} and i0i\geq 0 such that α=aicα\alpha^{\prime}=a^{i}c\alpha^{\prime\prime} (and so α=ai+1cα\alpha=a^{i+1}c\alpha^{\prime\prime}). Moreover:

  1. 1.

    τ(α)=1\tau(\alpha)=1 if and only if cac\prec a, if and only if αaω\alpha^{\prime}\prec a^{\omega}, if and only if αaω\alpha\prec a^{\omega}.

  2. 2.

    τ(α)=3\tau(\alpha)=3 if and only if aca\prec c, if and only if aωαa^{\omega}\prec\alpha^{\prime}, if and only if aωαa^{\omega}\prec\alpha,

The following corollary will be a key ingredient in our recursive approach.

Corollary 3.3.

Let α,βΣω\alpha,\beta\in\Sigma^{\omega}. Let a,bΣa,b\in\Sigma and α,βΣω\alpha^{\prime},\beta^{\prime}\in\Sigma^{\omega} such that α=aα\alpha=a\alpha^{\prime} and β=bβ\beta=b\beta^{\prime}. Then:

  1. 1.

    If a=ba=b and τ(α)=τ(β)=2\tau(\alpha)=\tau(\beta)=2, then α=β\alpha=\beta.

  2. 2.

    If a=ba=b and τ(α)<τ(β)\tau(\alpha)<\tau(\beta), then αβ\alpha\prec\beta. Equivalently, if a=ba=b and αβ\alpha\preceq\beta, then τ(α)τ(β)\tau(\alpha)\leq\tau(\beta).

4 Computing the min-partition

Let G=(V,E)G=(V,E) be a graph. We will prove that we can compute the min-partition of VV in O(|V|2)O(|V|^{2}) time. In this section, for every uVu\in V we define τ(u)=τ(minu)\tau(u)=\tau(\min_{u}) (see Figure 2).

11223344556677aabbccccaaccaacccc#\#
ii mini\min_{i} τ(i)(=τ(mini))\tau(i)\;(=\tau(\min_{i}))
11 #####\#\#\#\#\#\dots 22
22 a####a\#\#\#\#\dots 11
33 b####b\#\#\#\#\dots 11
44 c####c\#\#\#\#\dots 11
55 ca###ca\#\#\#\dots 11
66 ab###ab\#\#\#\dots 33
77 cc###cc\#\#\#\dots 11
Figure 2: The graph from Figure 1, with the values mini\min_{i}’s and τ(i)\tau(i)’s.

Let uVu\in V, and let (ui)i1(u_{i})_{i\geq 1} be an occurrence of minu\min_{u} starting at uu. It is immediate to realize that (i) if τ(u)=1\tau(u)=1, then λ(u2)λ(u1)\lambda(u_{2})\preceq\lambda(u_{1}), (ii) if τ(u)=2\tau(u)=2, then λ(uk)=λ(u1)\lambda(u_{k})=\lambda(u_{1}) for every k1k\geq 1 and (iii) if τ(u)=3\tau(u)=3, then λ(u1)λ(u2)\lambda(u_{1})\preceq\lambda(u_{2}).

As a first step, let us prove that without loss of generality we can remove some edges from GG without affecting the min/max-partition. This preprocessing will be helpful in Lemma 4.20.

Definition 4.1.

Let G=(V,E)G=(V,E) be a graph. We say that GG is trimmed if it contains no edge (u,v)E(u,v)\in E such that τ(v)=1\tau(v)=1 and λ(v)λ(u)\lambda(v)\prec\lambda(u).

In order to simplify the readability of our proofs, we will not directly remove some edges from G=(V,E)G=(V,E), but we will first build a copy of GG where every node uu is a mapped to a node uu^{*}, and then we will trim the graph. In this way, when we write minu\min_{u} and minu\min_{u^{*}} it will be always clear whether we refer to the original graph or the trimmed graph. We will use the same convention in Section 4.2 when we define the graph G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}) that we will use for the recursive step.

Lemma 4.2.

Let G=(V,E)G=(V,E) be a graph. Then, in O(|E|)O(|E|) time we can build a trimmed graph G=(V,E)G^{*}=(V^{*},E^{*}), with V={u|uV}V^{*}=\{u^{*}\;|\;u\in V\}, such that for every uVu\in V it holds minu=minu\min_{u^{*}}=\min_{u}. In particular, τ(u)=τ(u)\tau(u^{*})=\tau(u) for every uVu\in V.

4.1 Classifying Minima

Let us first show how to compute all uVu\in V such that τ(u)=1\tau(u)=1.

Lemma 4.3.

Let G=(V,E)G=(V,E) be a graph, and let u,vVu,v\in V.

  1. 1.

    If (u,v)E(u,v)\in E and λ(u)λ(v)\lambda(u)\prec\lambda(v), then τ(v)=1\tau(v)=1.

  2. 2.

    If (u,v)E(u,v)\in E, λ(u)=λ(v)\lambda(u)=\lambda(v) and τ(u)=1\tau(u)=1, then τ(v)=1\tau(v)=1.

Corollary 4.4.

Let G=(V,E)G=(V,E) be a graph, and let uVu\in V. Then, τ(u)=1\tau(u)=1 if and only if there exist k2k\geq 2 and z1,,zkVz_{1},\dots,z_{k}\in V such that (i) (zi,zi+1)E(z_{i},z_{i+1})\in E for every 1ik11\leq i\leq k-1, (ii) zk=uz_{k}=u, (iii) λ(z1)λ(z2)\lambda(z_{1})\prec\lambda(z_{2}) and (iv) λ(z2)=λ(z3)==λ(zk)\lambda(z_{2})=\lambda(z_{3})=\dots=\lambda(z_{k}).

Corollary 4.4 yields an algorithm to decide whether uVu\in V is such that τ(u)=1\tau(u)=1.

Corollary 4.5.

Let G=(V,E)G=(V,E) be a graph. We can determine all uVu\in V such that τ(u)=1\tau(u)=1 in time O(|E|)O(|E|).

Now, let us show how to determine all uVu\in V such that τ(u)=2\tau(u)=2. We can assume that we have already determined all uVu\in V such that τ(u)=1\tau(u)=1.

Lemma 4.6.

Let G=(V,E)G=(V,E) be a graph, and let uVu\in V such that τ(u)1\tau(u)\not=1. Then, we have τ(u)=2\tau(u)=2 if and only if there exist k2k\geq 2 and z1,,zkVz_{1},\dots,z_{k}\in V such that (i) (zi+1,zi)E(z_{i+1},z_{i})\in E for every 1ik11\leq i\leq k-1, (ii) z1=uz_{1}=u, (iii) zk=zjz_{k}=z_{j} for some 1jk11\leq j\leq k-1 and (iv) λ(z1)=λ(z2)==λ(zk)\lambda(z_{1})=\lambda(z_{2})=\dots=\lambda(z_{k}).

In particular, such z1,,zkVz_{1},\dots,z_{k}\in V must satisfy τ(zi)=2\tau(z_{i})=2 for every 1ik1\leq i\leq k.

Corollary 4.7.

Let G=(V,E)G=(V,E) be a graph. We can determine all uVu\in V such that τ(u)=2\tau(u)=2 in time O(|E|)O(|E|).

From Corollary 4.5 and Corollary 4.7 we immediately obtain the following result.

Corollary 4.8.

Let G=(V,E)G=(V,E) be a graph. Then, in time O(|E|)O(|E|) we can compute τ(u)\tau(u) for every uVu\in V.

4.2 Recursive Step

Let us sketch the general idea to build a smaller graph for the recursive step. We consider each uVu\in V such that τ(u)=3\tau(u)=3, and we follow edges in a backward fashion, aiming to determine a prefix of minu\min_{u}. As a consequence, we discard edges through which no occurrence of minu\min_{u} can go, and by Corollary 3.3 we can restrict our attention to the nodes vv such that τ(v)\tau(v) is minimal. We proceed like this until we encounter nodes vv^{\prime} such that τ(v)=3\tau(v^{\prime})=3.

Let us formalize our intuition. We will first present some properties that the occurrences of a string minu\min_{u} must satisfy.

Lemma 4.9.

Let G=(V,E)G=(V,E) be a graph. Let u,vVu,v\in V be such that minu=minv\min_{u}=\min_{v}. Let (ui)i1(u_{i})_{i\geq 1} be an occurrence of minu\min_{u} and let (vi)i1(v_{i})_{i\geq 1} be an occurrence of minv\min_{v}. Then:

  1. 1.

    λ(ui)=λ(vi)\lambda(u_{i})=\lambda(v_{i}) for every i1i\geq 1.

  2. 2.

    minui=minvi\min_{u_{i}}=\min_{v_{i}} for every i1i\geq 1.

  3. 3.

    τ(ui)=τ(vi)\tau(u_{i})=\tau(v_{i}) for every i1i\geq 1.

In particular, the previous results hold if u=vu=v and (ui)i1(u_{i})_{i\geq 1} and (vi)i1(v_{i})_{i\geq 1} are two distinct occurrences of minu\min_{u}.

Lemma 4.10.

Let G=(V,E)G=(V,E) be a graph. Let uVu\in V and let (ui)i1(u_{i})_{i\geq 1} an occurrence of minu\min_{u} starting at uu. Let k1k\geq 1 be such that τ(u1)=τ(u2)==τ(uk1)=τ(uk)2\tau(u_{1})=\tau(u_{2})=\dots=\tau(u_{k-1})=\tau(u_{k})\not=2. Then, u1,,uku_{1},\dots,u_{k} are pairwise distinct. In particular, k|V|k\leq|V|.

The previous results allow us to give the following definition.

Definition 4.11.

Let G=(V,E)G=(V,E) be a graph. Let uVu\in V such that τ(u)=3\tau(u)=3. Let u\ell_{u} to be the smallest integer k2k\geq 2 such that τ(uk)2\tau(u_{k})\geq 2, where (ui)i1(u_{i})_{i\geq 1} is an occurrence of minu\min_{u} starting at uu.

Note that u\ell_{u} is well-defined, because (i) it cannot hold τ(uk)=1\tau(u_{k})=1 for every k2k\geq 2 by Lemma 4.10 (indeed, if τ(u2)=1\tau(u_{2})=1, then (ui)i2(u_{i})_{i\geq 2} is an occurrence of minu2\min_{u_{2}} starting at u2u_{2}, and by Lemma 4.10 there exists 2k|V|+22\leq k\leq|V|+2 such that τ(uk)1\tau(u_{k})\not=1) and (ii) u\ell_{u} does not depend on the choice of (ui)i1(u_{i})_{i\geq 1} by Lemma 4.9. In particular, it must be u|V|+1\ell_{u}\leq|V|+1 because u1,u2,,uu1u_{1},u_{2},\dots,u_{\ell_{u}-1} are pairwise distinct (u1u_{1} is distinct from u2,,uu1u_{2},\dots,u_{\ell_{u}-1} because τ(u1)=3\tau(u_{1})=3 and τ(u2)=τ(u3)=τ(uu1)=1\tau(u_{2})=\tau(u_{3})=\dots\tau(u_{\ell_{u}-1})=1 by the minimality of u\ell_{u}).

Lemma 4.12.

Let G=(V,E)G=(V,E) be a graph. Let uVu\in V such that τ(u)=3\tau(u)=3. Then, minu[i+1]minu[i]\min_{u}[i+1]\preceq\min_{u}[i] for every 2iu12\leq i\leq\ell_{u}-1. In particular, if 2iju2\leq i\leq j\leq\ell_{u}, then minu[j]minu[i]\min_{u}[j]\preceq\min_{u}[i].

If RQR\subseteq Q is a nonempty set of nodes such that for every u,vRu,v\in R it holds λ(u)=λ(v)\lambda(u)=\lambda(v), we define λ(R)=λ(u)=λ(v)\lambda(R)=\lambda(u)=\lambda(v). If RQR\subseteq Q is a nonempty set of nodes such that for every u,vRu,v\in R it holds τ(u)=τ(v)\tau(u)=\tau(v), we define τ(R)=τ(u)=τ(v)\tau(R)=\tau(u)=\tau(v).

Let RQR\subseteq Q be a nonempty set of states. Let 𝙵(R)=argminuRτ(u)\mathtt{F}(R)=\operatorname*{arg\,min}_{u\in R^{\prime}}\tau(u), where R=argminvRλ(v)R^{\prime}=\operatorname*{arg\,min}_{v\in R}\lambda(v). Notice that 𝙵(R)\mathtt{F}(R) is nonempty, and both λ(𝙵(R))\lambda(\mathtt{F}(R)) and τ(𝙵(R))\tau(\mathtt{F}(R)) are well-defined. In other words, 𝙵(R)\mathtt{F}(R) is obtained by first considering the subset R𝙵(R)R^{\prime}\subseteq\mathtt{F}(R) of all nodes vv such that λ(v)\lambda(v) is as small as possible, and then considering the subset of RR^{\prime} of all nodes vv such that τ(v)\tau(v) is as small as possible. This is consistent with our intuition on how we should be looking for a prefix of minu\min_{u}.

Define:

Gi(u)={{u} if i=1;𝙵({vQ|(vGi1(u))((v,v)E)})j=2i1Gj(u) if 1<iu.G_{i}(u)=\begin{cases}\{u\}&\text{ if $i=1$;}\\ \mathtt{F}(\{v^{\prime}\in Q\;|\;(\exists v\in G_{i-1}(u))((v^{\prime},v)\in E)\})\setminus\bigcup_{j=2}^{i-1}G_{j}(u)&\text{ if $1<i\leq\ell_{u}$.}\\ \end{cases}

Notice that we also require that a node in Gi(u)G_{i}(u) has not been encountered before. Intuitively, this does not affect our search for a prefix of minu\min_{u} because, if we met the same node twice, then we would have a cycle where all edges are equally labeled (because by Lemma 4.12 labels can only decrease), and since τ(Gi(u))=1\tau(G_{i}(u))=1 for every 2iu12\leq i\leq\ell_{u}-1, then no occurrence of the minimum can go through the cycle because if we remove the cycle from the occurrence we obtain a smaller string by Lemma 3.2.

The following technical lemma is crucial to prove that our intuition is correct.

Lemma 4.13.

Let G=(V,E)G=(V,E) be a graph. Let uVu\in V such that τ(u)=3\tau(u)=3.

  1. 1.

    Gi(u)G_{i}(u) is well-defined and nonempty for every 1iu1\leq i\leq\ell_{u}.

  2. 2.

    Let (ui)i1(u_{i})_{i\geq 1} be an occurrence of minu\min_{u} starting at uu. Then, uiGi(u)u_{i}\in G_{i}(u) for every 1iu1\leq i\leq\ell_{u}. In particular, τ(ui)=τ(Gi(u))\tau(u_{i})=\tau(G_{i}(u)) and minu[i]=λ(ui)=λ(Gi(u))\min_{u}[i]=\lambda(u_{i})=\lambda(G_{i}(u)) for every 1iu1\leq i\leq\ell_{u}.

  3. 3.

    For every 1iu1\leq i\leq\ell_{u} and for every vGi(u)v\in G_{i}(u) there exists an occurrence of minu[1,i1]\min_{u}[1,i-1] starting at uu and ending at vv.

Let uVu\in V such that τ(u)=3\tau(u)=3. We define:

  • γu=minu[1,u]\gamma_{u}=\min_{u}[1,\ell_{u}];

  • 𝚝u=τ(Gu(u)){2,3}\mathtt{t}_{u}=\tau(G_{\ell_{u}}(u))\in\{2,3\}

Now, in order to define the smaller graph for the recursive step, we also need a new alphabet (Σ,)(\Sigma^{\prime},\preceq^{\prime}), which must be defined consistently with the mutual ordering of the minima. The next lemma yields all the information that we need.

Lemma 4.14.

Let G=(V,E)G=(V,E) be a graph. Let u,vVu,v\in V such that τ(u)=τ(v)=3\tau(u)=\tau(v)=3. Assume that one of the following statements is true:

  1. 1.

    γu\gamma_{u} is not a prefix of γv\gamma_{v} and γuγv\gamma_{u}\prec\gamma_{v}.

  2. 2.

    γu=γv\gamma_{u}=\gamma_{v}, 𝚝u=2\mathtt{t}_{u}=2 and 𝚝v=3\mathtt{t}_{v}=3.

  3. 3.

    γv\gamma_{v} is a strict prefix of γu\gamma_{u}.

Then, minuminv\min_{u}\prec\min_{v}.

Equivalently, if minuminv\min_{u}\preceq\min_{v}, then one the following is true: (i) γu\gamma_{u} is not a prefix of γv\gamma_{v} and γuγv\gamma_{u}\prec\gamma_{v}; (ii) γu=γv\gamma_{u}=\gamma_{v} and 𝚝u𝚝v\mathtt{t}_{u}\leq\mathtt{t}_{v}; (iii) γv\gamma_{v} is a strict prefix of γu\gamma_{u}.

Now, let Σ={(γu,𝚝u)|uV,τ(u)=3}\Sigma^{\prime}=\{(\gamma_{u},\mathtt{t}_{u})\;|\;u\in V,\tau(u)=3\}, and let \preceq^{\prime} be the total order on Σ\Sigma^{\prime} such that for every distinct (α,x),(β,y)Σ(\alpha,x),(\beta,y)\in\Sigma^{\prime}, it holds (α,x)(β,y)(\alpha,x)\prec^{\prime}(\beta,y) if and only if one of the following is true:

  1. 1.

    α\alpha is not a prefix of β\beta and αβ\alpha\prec\beta.

  2. 2.

    α=β\alpha=\beta, x=2x=2 and y=3y=3.

  3. 3.

    β\beta is a strict prefix of α\alpha.

It is immediate to verify that \preceq^{\prime} is a total order: indeed, \preceq^{\prime} is obtained (i) by first comparing the γu\gamma_{u}’s using the variant of the (total) lexicographic order on Σ\Sigma^{*} in which a string is smaller than every strict prefix of it and (ii) if the γu\gamma_{u}’s are equal by comparing the 𝚝u\mathtt{t}_{u}’s, which are elements in {2,3}\{2,3\}.

Starting from G=(V,E)G=(V,E), we define a new graph G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}) as follows:

  • V¯={u¯|uV,τ(u)=3}\bar{V}=\{\bar{u}\;|\;u\in V,\tau(u)=3\}.

  • The new totally-ordered alphabet is (Σ,)(\Sigma^{\prime},\preceq^{\prime}).

  • For every u¯V¯\bar{u}\in\bar{V}, we define λ(u¯)=(γu,𝚝u)\lambda(\bar{u})=(\gamma_{u},\mathtt{t}_{u}).

  • E¯={(v¯,v¯)|𝚝v=2}{(u¯,v¯)|𝚝v=3,uGv(v)}\bar{E}=\{(\bar{v},\bar{v})\;|\;\mathtt{t}_{v}=2\}\cup\{(\bar{u},\bar{v})\;|\;\mathtt{t}_{v}=3,u\in G_{\ell_{v}}(v)\}.

Note that for every v¯V¯\bar{v}\in\bar{V} such that 𝚝v=3\mathtt{t}_{v}=3 and for every uGv(v)u\in G_{\ell_{v}}(v) it holds τ(u)=τ(Gv(v))=𝚝v=3\tau(u)=\tau(G_{\ell_{v}}(v))=\mathtt{t}_{v}=3, so u¯V¯\bar{u}\in\bar{V} and (u¯,v¯)E(\bar{u},\bar{v})\in E. Moreover, G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}) satisfies all the assumptions about graphs that we use in this paper: (i) all edges entering the same node have the same label (by definition), (ii) every node has at least one incoming edge (because if v¯V¯\bar{v}\in\bar{V}, then Gv(v)G_{\ell_{v}}(v)\not=\emptyset by Lemma 4.13) and (iii) G¯\bar{G} is deterministic (because if (u¯,v¯),(u¯,v¯)E¯(\bar{u},\bar{v}),(\bar{u},\bar{v}^{\prime})\in\bar{E} and λ(v¯)=λ(v¯)\lambda(\bar{v})=\lambda(\bar{v}^{\prime}), then γv=γv\gamma_{v}=\gamma_{v^{\prime}} and 𝚝v=𝚝v\mathtt{t}_{v}=\mathtt{t}_{v^{\prime}}, so by the definition of E¯\bar{E} if 𝚝v=𝚝v=2\mathtt{t}_{v}=\mathtt{t}_{v^{\prime}}=2 we immediately obtain v¯=u¯=v¯\bar{v}=\bar{u}=\bar{v}^{\prime}, and if 𝚝v=𝚝v=3\mathtt{t}_{v}=\mathtt{t}_{v}^{\prime}=3 we obtain uGv(v)Gv(v)u\in G_{\ell_{v}}(v)\cap G_{\ell_{v^{\prime}}}(v^{\prime}); since by Lemma 4.13 there exist two occurrences of minv[1,v1]=γv[1,v1]=γv[1,v1]=minv[1,v1]\min_{v}[1,\ell_{v}-1]=\gamma_{v}[1,\ell_{v}-1]=\gamma_{v^{\prime}}[1,\ell_{v^{\prime}}-1]=\min_{v^{\prime}}[1,\ell_{v^{\prime}}-1] starting at vv and vv^{\prime} and both ending at uu, the determinism of GG implies v=vv=v^{\prime} and so v¯=v¯\bar{v}=\bar{v}^{\prime}).

Notice that if v¯V¯\bar{v}\in\bar{V} is such that 𝚝v=2\mathtt{t}_{v}=2, then Iv¯I_{\bar{v}} contains exactly one string, namely, λ(v¯)ω\lambda(\bar{v})^{\omega}; in particular, minv¯=maxv¯=λ(v¯)ω\min_{\bar{v}}=\max_{\bar{v}}=\lambda(\bar{v})^{\omega}.

When we implement G=(V,E)G=(V,E) and G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}), we use integer alphabets Σ={0,1,,|Σ|1}\Sigma=\{0,1,\dots,|\Sigma|-1\} and Σ={0,1,,|Σ|1}\Sigma^{\prime}=\{0,1,\dots,|\Sigma^{\prime}|-1\}; in particular, we will not store Σ\Sigma^{\prime} by means of pairs (γu,𝚝u)(\gamma_{u},\mathtt{t}_{u})’s, but we will remap Σ\Sigma^{\prime} to an integer alphabet consistently with the total order \preceq^{\prime} on Σ\Sigma^{\prime}, so that the mutual order of the minu¯\min_{\bar{u}}’s is not affected.

Let us prove that we can use G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}) for the recursive step. We will start with some preliminary results.

Lemma 4.15.

Let G=(V,E)G=(V,E) be a graph. Let u,vVu,v\in V be such that τ(u)=τ(v)=3\tau(u)=\tau(v)=3, γu=γv\gamma_{u}=\gamma_{v} and 𝚝u=𝚝v=2\mathtt{t}_{u}=\mathtt{t}_{v}=2. Then, minu=minv\min_{u}=\min_{v}.

Lemma 4.16.

Let G=(V,E)G=(V,E) be a graph. Let uVu\in V, and let (ui)i1(u_{i})_{i\geq 1} be an occurrence of minu\min_{u} starting at uu. Then, exactly one of the following holds true:

  1. 1.

    There exists i01i_{0}\geq 1 such that τ(ui)2\tau(u_{i})\not=2 for every 1i<i01\leq i<i_{0} and τ(ui)=2\tau(u_{i})=2 for every ii0i\geq i_{0}.

  2. 2.

    τ(ui)2\tau(u_{i})\not=2 for every i1i\geq 1, and both τ(ui)=1\tau(u_{i})=1 and τ(ui)=3\tau(u_{i})=3 are true for infinitely many ii’s.

Crucially, the next lemma establishes a correspondence between minima of nodes in G=(V,E)G=(V,E) and minima of nodes in G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}).

Lemma 4.17.

Let G=(V,E)G=(V,E) be a graph. Let uVu\in V such that τ(u)=3\tau(u)=3. Let (ui)i1(u_{i})_{i\geq 1} be an occurrence of minu\min_{u} starting at uu. Let (ui)i1(u^{\prime}_{i})_{i\geq 1} be the infinite sequence of nodes in VV obtained as follows. Consider L={k1|τ(uk)=3}L=\{k\geq 1\;|\;\tau(u_{k})=3\}, and for every i1i\geq 1, let ji1j_{i}\geq 1 be the ithi^{\text{th}} smallest element of LL, if it exists. For every i1i\geq 1 such that jij_{i} is defined, let ui=ujiu^{\prime}_{i}=u_{j_{i}}, and if i1i\geq 1 is such that jij_{i} is not defined (so LL is a finite set), let ui=u|L|u^{\prime}_{i}=u^{\prime}_{|L|}. Then, (u¯i)i1(\bar{u}^{\prime}_{i})_{i\geq 1} is an occurrence of minu¯\min_{\bar{u}} starting at u¯\bar{u} in G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}).

The following theorem shows that our reduction to G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}) is correct.

Theorem 4.18.

Let G=(V,E)G=(V,E) be a graph. Let u,vVu,v\in V be such that τ(u)=τ(v)=3\tau(u)=\tau(v)=3.

  1. 1.

    If minu=minv\min_{u}=\min_{v}, then minu¯=minv¯\min_{\bar{u}}=\min_{\bar{v}}.

  2. 2.

    If minuminv\min_{u}\prec\min_{v}, then minu¯minv¯\min_{\bar{u}}\prec^{\prime}\min_{\bar{v}}.

Since \preceq is a total order (so exactly one among minuminv\min_{u}\prec\min_{v}, minu=minv\min_{u}=\min_{v} and minvminu\min_{v}\prec\min_{u} holds true), from Theorem 4.18 we immediately obtain the following result.

Corollary 4.19.

Let G=(V,E)G=(V,E) be a graph. Let u,vVu,v\in V be such that τ(u)=τ(v)=3\tau(u)=\tau(v)=3.

  1. 1.

    It holds minu=minv\min_{u}=\min_{v} if and only if minu¯=minv¯\min_{\bar{u}}=\min_{\bar{v}}.

  2. 2.

    It holds minuminv\min_{u}\prec\min_{v} if and only if minu¯minv¯\min_{\bar{u}}\prec^{\prime}\min_{\bar{v}}.

In particular, if we have the min-partition of V¯\bar{V} (with respect to G¯\bar{G}), then we also have the min-partition of {uV|τ(u)=3}\{u\in V\;|\;\tau(u)=3\} (with respect to GG).

Lastly, we show that our reduction to G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}) can be computed within O(n2)O(n^{2}) time.

Lemma 4.20.

Let G=(V,E)G=(V,E) be a trimmed graph. Then, we can build G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}) in O(|V|2)O(|V|^{2}) time.

4.3 Merging

We want to determine the min-partition 𝒜\mathcal{A} of VV, assuming that we already have the min-partition \mathcal{B} of {uV|τ(u)=3}\{u\in V\;|\;\tau(u)=3\}.

First, note that we can easily build the min-partition \mathcal{B}^{\prime} of {uV|τ(u)=2}\{u\in V\;|\;\tau(u)=2\}. Indeed, if τ(u)=2\tau(u)=2, then minu=λ(u)ω\min_{u}=\lambda(u)^{\omega} by Lemma 3.2. As a consequence, if τ(u)=τ(v)=2\tau(u)=\tau(v)=2, then (i) minu=minv\min_{u}=\min_{v} if and only if λ(u)=λ(v)\lambda(u)=\lambda(v) and (ii) minuminv\min_{u}\prec\min_{v} if and only if λ(u)λ(v)\lambda(u)\prec\lambda(v), so we can build \mathcal{B}^{\prime} in O(|V|)O(|V|) time by using counting sort.

For every cΣc\in\Sigma and t{1,2,3}t\in\{1,2,3\}, let Vc,t={vV|λ(v)=c,τ(v)=t}V_{c,t}=\{v\in V\;|\;\lambda(v)=c,\tau(v)=t\}. Consider u,vVu,v\in V: (i) if λ(u)λ(v)\lambda(u)\prec\lambda(v), then minuminv\min_{u}\prec\min_{v} and (ii) if λ(u)=λ(v)\lambda(u)=\lambda(v) and τ(u)<τ(v)\tau(u)<\tau(v), then minuminv\min_{u}\prec\min_{v} by Corollary 3.3. As a consequence, in order to build 𝒜\mathcal{A}, we only have to build the min-partition 𝒜c,t\mathcal{A}_{c,t} of Vc,tV_{c,t}, for every cΣc\in\Sigma and every t{1,2,3}t\in\{1,2,3\}.

A possible way to implement each 𝒜c,t\mathcal{A}_{c,t} is by means of an array Ac,tA_{c,t} storing the elements of Vc,tV_{c,t}, where we also use a special character to delimit the border between consecutive elements of Ac,tA_{c,t}.

It is immediate to build incrementally 𝒜c,3\mathcal{A}_{c,3} for every cΣc\in\Sigma, from its smallest element to its largest element. At the beginning, Ac,3A_{c,3} is empty for every cΣc\in\Sigma. Then, scan the elements II in \mathcal{B} from smallest to largest, and add II to 𝒜c,3\mathcal{A}_{c,3}, where c=λ(u)c=\lambda(u) for any uIu\in I (the definition of cc does not depend on the choice of uu). We scan \mathcal{B} only once, so this step takes O(|V|)O(|V|) time. Analogously, we can build 𝒜c,2\mathcal{A}_{c,2} for every cΣc\in\Sigma by using \mathcal{B}^{\prime}.

We are only left with showing how to build 𝒜c,1\mathcal{A}_{c,1} for every cΣc\in\Sigma. At the beginning, each Ac,1A_{c,1} is empty, and we will build each 𝒜c,1\mathcal{A}_{c,1} from its smallest element to its largest element. During this step of the algorithm, we will gradually mark the nodes uVu\in V such that τ(u)=1\tau(u)=1. At the beginning of the step, no such node is marked, and at the end of the step all these nodes will be marked. Let Σ={c1,c2,,cσ}\Sigma=\{c_{1},c_{2},\dots,c_{\sigma}\}, with c1c2cσc_{1}\prec c_{2}\prec\dots\prec c_{\sigma}. Notice that it must be Vc1,1=V_{c_{1},1}=\emptyset, because if there existed uVc1,1u\in V_{c_{1},1}, then it would be minuc1ω\min_{u}\prec c_{1}^{\omega} by Lemma 3.2 and so c1c_{1} would not be the smallest character in Σ\Sigma. Now, consider Vc1,2V_{c_{1},2}; we have already fully computed Ac1,2A_{c_{1},2}. Process each II in Ac1,2A_{c_{1},2} from smallest to largest, and for every ckΣc_{k}\in\Sigma compute the set JkJ_{k} of all non-marked nodes vVv\in V such that τ(v)=1\tau(v)=1, λ(v)=ck\lambda(v)=c_{k}, and (u,v)E(u,v)\in E for some uIu\in I. Then, if JkJ_{k}\not=\emptyset add JkJ_{k} to Ack,1A_{c_{k},1} and mark the nodes in JkJ_{k}. After processing the elements in Ac1,2A_{c_{1},2}, we process the element in Ac1,3A_{c_{1},3}, Ac2,1A_{c_{2},1}, Ac2,2A_{c_{2},2}, Ac2,3A_{c_{2},3}, Ac3,1A_{c_{3},1} and so on, in this order. Each Aci,tA_{c_{i},t} is processed from its (current) smallest element to its (current) largest element. We never remove or modify elements in any Ac,tA_{c,t}, but we only add elements to the Ac,1A_{c,1}’s. More precisely, when we process II in Ac,tA_{c,t}, for every ckΣc_{k}\in\Sigma we compute the set JkJ_{k} of all non-marked nodes vVv\in V such that τ(v)=1\tau(v)=1, λ(v)=ck\lambda(v)=c_{k}, and (u,v)E(u,v)\in E for some uIu\in I and, if JkJ_{k}\not=\emptyset, then we add JkJ_{k} to Ack,1A_{c_{k},1} and we mark the nodes in JkJ_{k}.

The following lemma shows that our approach is correct. Let us give some intuition. A prefix of a min-partition 𝒞\mathcal{C} is a subset 𝒞\mathcal{C}^{\prime} of 𝒞\mathcal{C} such that, if I,J𝒞I,J\in\mathcal{C}, I<JI<J and J𝒞J\in\mathcal{C}^{\prime}, then I𝒞I\in\mathcal{C^{\prime}}. Notice that every prefix of 𝒜\mathcal{A} is obtained by taking the union of 𝒜c1,2\mathcal{A}_{c_{1},2}, 𝒜c1,3\mathcal{A}_{c_{1},3}, 𝒜c2,1\mathcal{A}_{c_{2},1}, 𝒜c2,2\mathcal{A}_{c_{2},2}, 𝒜c2,3\mathcal{A}_{c_{2},3}, 𝒜c3,1\mathcal{A}_{c_{3},1}, \dots in this order up to some element 𝒜c,t\mathcal{A}_{c,t}, where possibly we only pick a prefix of the last element 𝒜c,t\mathcal{A}_{c,t}. Then, we will show that, when we process II in Ac,tA_{c,t}, we have already built the prefix of 𝒜\mathcal{A} whose largest element is II. This means that, for every vJkv\in J_{k} and for any any occurrence (vi)i1(v_{i})_{i\geq 1} of minv\min_{v} starting at vv, it must hold that v2v_{2} is in II.

Lemma 4.21.

Let G=(V,E)G=(V,E) be a graph. If we know the min-partition of {uV|τ(u)=3}\{u\in V\;|\;\tau(u)=3\}, then we can build the min-partition of VV in O(|E|)O(|E|) time.

4.4 The Complementary Case

We have shown that in O(n2)O(n^{2}) time we can reduce the problem of determining the min-partition of VV to the problem of determining the min-partition of the set of all nodes of a graph having |{uV|τ(u)=3}||\{u\in V\;|\;\tau(u)=3\}| nodes. Now, we must show that (similarly) in O(n2)O(n^{2}) time we can reduce the problem of determining the min-partition of VV to the problem of determining the min-partition of the set of all nodes of a graph having |{uV|τ(u)=1}||\{u\in V\;|\;\tau(u)=1\}| nodes. This time, building the smaller graph G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}) will be simpler, because when searching for prefixes the minima we will never meet nodes that we have already met (so the the definition of the Gi(u)G_{i}(u)’s is simpler and we do not need to trim the graph to stay within the O(n2)O(n^{2}) bound). On the other hand, the merging step will be more complex, because the order in which we will process the 𝒜c,t\mathcal{A}_{c,t} will be from largest to smallest (𝒜cσ,2\mathcal{A}_{c_{\sigma},2}, 𝒜cσ,1\mathcal{A}_{c_{\sigma},1}, 𝒜cσ1,3\mathcal{A}_{c_{\sigma-1},3}, 𝒜cσ1,2\mathcal{A}_{c_{\sigma-1},2}, 𝒜cσ1,1\mathcal{A}_{c_{\sigma-1},1}, 𝒜cσ2,3\mathcal{A}_{c_{\sigma-2},3} and so on) so we will need to update some elements of some Ac,tA_{c,t}’s to include the information about minima that we may infer at a later stage of the algorithm. We provide the details in the appendix.

5 Computing the min/max-partition

Let G=(V,E)G=(V,E) be a graph. We can build the max-partition of VV by simply considering the transpose total order \preceq^{*} of \preceq (the one for which aba\preceq^{*}b if and only if bab\preceq a) and building the min-partition. As a consequence, the algorithm to build the max-partition is entirely symmetrical to the algorithm to build the min-partition.

Let G=(V,E)G=(V,E) be a graph. Let us show how we can build the min/max-partition of VV in O(|V|2)O(|V|^{2}) time. Assume that we have two graphs G1=(V1,E1)G_{1}=(V_{1},E_{1}) and G2=(V2,E2)G_{2}=(V_{2},E_{2}) on the same alphabet (Σ,)(\Sigma,\preceq), with V1V2=V_{1}\cap V_{2}=\emptyset (we allow G1G_{1} and G2G_{2} to possibly be the null graph, that is, the graph without vertices). Let V1V1V^{\prime}_{1}\subseteq V_{1}, V2V2V^{\prime}_{2}\subseteq V_{2}, W=V1V2W=V^{\prime}_{1}\cup V^{\prime}_{2}, and for every uWu\in W define ρ(u)=minu\rho(u)=\min_{u} if uV1u\in V^{\prime}_{1}, and ρ(u)=maxu\rho(u)=\max_{u} if uV2u\in V^{\prime}_{2}. Let 𝒜\mathcal{A} be the unique partition of WW and let \leq be the unique total order on 𝒜\mathcal{A} such that, for every I,J𝒜I,J\in\mathcal{A} and for every uIu\in I and uJu\in J, (i) if I=JI=J, then ρ(u)=ρ(u)\rho(u)=\rho(u) and (ii) if I<JI<J, then ρ(u)ρ(u)\rho(u)\prec\rho(u). Then, we say that (𝒜,)(\mathcal{A},\leq), or more simply 𝒜\mathcal{A}, is the min/max-partition of (V1,V2)(V^{\prime}_{1},V^{\prime}_{2}). We will show that we can compute the min/max partition of (V1,V2)(V_{1},V_{2}) in O((|V1|+|V2|)2)O((|V_{1}|+|V_{2}|)^{2}) time. In particular, if G1=(V1,E1)G_{1}=(V_{1},E_{1}) and G2=(V1,E2)G_{2}=(V_{1},E_{2}) are two (distinct) copies of the same graph G=(V,E)G=(V,E), then we can compute the min/max-partition of VV in O(|V|2)O(|V|^{2}) time.

We compute τ(minu)\tau(\min_{u}) for every uV1u\in V_{1} and we compute τ(maxu)\tau(\max_{u}) for every uV2u\in V_{2}. If the number of values equal to 33 is smaller than the number of values equal to 1, then (in time O(|V1|2+|V2|2)=O((|V1|+|V2|)2)O(|V_{1}|^{2}+|V_{2}|^{2})=O((|V_{1}|+|V_{2}|)^{2})) we build the graphs G¯1=(V¯1,E¯1)\bar{G}_{1}=(\bar{V}_{1},\bar{E}_{1}) and G¯2=(V¯2,E¯2)\bar{G}_{2}=(\bar{V}_{2},\bar{E}_{2}) as defined before, where V¯1={u¯|uV1,τ(minu)=3}\bar{V}_{1}=\{\bar{u}\;|\;u\in V_{1},\tau(\min_{u})=3\} and V¯2={u¯|uV2,τ(maxu)=3}\bar{V}_{2}=\{\bar{u}\;|\;u\in V_{2},\tau(\max_{u})=3\}, otherwise we consider the complementary case (which is symmetrical). When building G¯1=(V¯1,E¯1)\bar{G}_{1}=(\bar{V}_{1},\bar{E}_{1}) and G¯2=(V¯2,E¯2)\bar{G}_{2}=(\bar{V}_{2},\bar{E}_{2}), we define a unique alphabet (Σ,)(\Sigma^{\prime},\preceq^{\prime}) obtained by jointly sorting the (γminu,𝚝minu)(\gamma_{\min_{u}},\mathtt{t}_{\min_{u}})’s and the (γmaxu,𝚝maxu)(\gamma_{\max_{u}},\mathtt{t}_{\max_{u}})’s, which is possible because Lemma 4.14 also applies to maxima. Note that |V¯1|+|V¯2|(|V1|+|V2|)/2|\bar{V}_{1}|+|\bar{V}_{2}|\leq(|V_{1}|+|V_{2}|)/2.

Assume that we have recursively obtained the min/max-partition of (V¯1,V¯2)(\bar{V}_{1},\bar{V}_{2}) with respect to G¯1\bar{G}_{1} and G¯2\bar{G}_{2}. This yields the min/max-partition of ({uV1|τ(minu)=3},{uV2|τ(maxu)=3})(\{u\in V_{1}\;|\;\tau(\min_{u})=3\},\{u\in V_{2}\;|\;\tau(\max_{u})=3\}). Then, we can build the min/max-partition of (V1,V2)(V_{1},V_{2}) by jointly applying the merging step, which is possible because both the merging step for minima and the merging step for maxima require to build the 𝒜c,1\mathcal{A}_{c,1}’s by processing Ac1,2Ac1,3A_{c_{1},2}A_{c_{1},3}, Ac2,1A_{c_{2},1}, Ac2,2A_{c_{2},2}, Ac2,3A_{c_{2},3}, Ac3,1A_{c_{3},1} and so on in this order.

Since we obtain the same recursion as before, we conclude that we can compute the min/max partition of (V1,V2)(V_{1},V_{2}) in O((|V1|+|V2|)2)O((|V_{1}|+|V_{2}|)^{2}) time.

References

  • [1] Jarno Alanko, Giovanna D’Agostino, Alberto Policriti, and Nicola Prezza. Regular languages meet prefix sorting. In Shuchi Chawla, editor, Proc. of the 31st Symposium on Discrete Algorithms, (SODA’20), pages 911–930. SIAM, 2020. doi:10.1137/1.9781611975994.55.
  • [2] Ruben Becker, Manuel Cáceres, Davide Cenzato, Sung-Hwan Kim, Bojana Kodric, Francisco Olivares, and Nicola Prezza. Sorting Finite Automata via Partition Refinement. In Inge Li Gørtz, Martin Farach-Colton, Simon J. Puglisi, and Grzegorz Herman, editors, 31st Annual European Symposium on Algorithms (ESA 2023), volume 274 of Leibniz International Proceedings in Informatics (LIPIcs), pages 15:1–15:15, Dagstuhl, Germany, 2023. Schloss Dagstuhl – Leibniz-Zentrum für Informatik. URL: https://drops.dagstuhl.de/opus/volltexte/2023/18668, doi:10.4230/LIPIcs.ESA.2023.15.
  • [3] Alexander Bowe, Taku Onodera, Kunihiko Sadakane, and Tetsuo Shibuya. Succinct de Bruijn graphs. In Ben Raphael and Jijun Tang, editors, Algorithms in Bioinformatics, pages 225–235, Berlin, Heidelberg, 2012. Springer Berlin Heidelberg.
  • [4] M. Burrows and D. J. Wheeler. A block-sorting lossless data compression algorithm. Technical report, 1994.
  • [5] Alessio Conte, Nicola Cotumaccio, Travis Gagie, Giovanni Manzini, Nicola Prezza, and Marinella Sciortino. Computing matching statistics on Wheeler DFAs. In 2023 Data Compression Conference (DCC), pages 150–159, 2023. doi:10.1109/DCC55655.2023.00023.
  • [6] Nicola Cotumaccio. Graphs can be succinctly indexed for pattern matching in O(|E|2+|V|5/2)O(|E|^{2}+|V|^{5/2}) time. In 2022 Data Compression Conference (DCC), pages 272–281, 2022. doi:10.1109/DCC52660.2022.00035.
  • [7] Nicola Cotumaccio, Giovanna D’Agostino, Alberto Policriti, and Nicola Prezza. Co-lexicographically ordering automata and regular languages - part i. J. ACM, 70(4), aug 2023. doi:10.1145/3607471.
  • [8] Nicola Cotumaccio and Nicola Prezza. On indexing and compressing finite automata. In Dániel Marx, editor, Proc. of the 32nd Symposium on Discrete Algorithms, (SODA’21), pages 2585–2599. SIAM, 2021. doi:10.1137/1.9781611976465.153.
  • [9] Massimo Equi, Veli Mäkinen, and Alexandru I. Tomescu. Graphs cannot be indexed in polynomial time for sub-quadratic time string matching, unless SETH fails. In Tomáš Bureš, Riccardo Dondi, Johann Gamper, Giovanna Guerrini, Tomasz Jurdziński, Claus Pahl, Florian Sikora, and Prudence W.H. Wong, editors, SOFSEM 2021: Theory and Practice of Computer Science, pages 608–622, Cham, 2021. Springer International Publishing.
  • [10] Massimo Equi, Veli Mäkinen, Alexandru I. Tomescu, and Roberto Grossi. On the complexity of string matching for graphs. ACM Trans. Algorithms, 19(3), apr 2023. doi:10.1145/3588334.
  • [11] M. Farach. Optimal suffix tree construction with large alphabets. In Proceedings 38th Annual Symposium on Foundations of Computer Science, pages 137–143, 1997. doi:10.1109/SFCS.1997.646102.
  • [12] P. Ferragina, F. Luccio, G. Manzini, and S. Muthukrishnan. Structuring labeled trees for optimal succinctness, and beyond. In proc. 46th Annual IEEE Symposium on Foundations of Computer Science (FOCS’05), pages 184–193, 2005. doi:10.1109/SFCS.2005.69.
  • [13] P. Ferragina and G. Manzini. Opportunistic data structures with applications. In Proc. 41st Annual Symposium on Foundations of Computer Science (FOCS’00), pages 390–398, 2000. doi:10.1109/SFCS.2000.892127.
  • [14] Paolo Ferragina, Fabrizio Luccio, Giovanni Manzini, and S. Muthukrishnan. Compressing and indexing labeled trees, with applications. J. ACM, 57(1), nov 2009. doi:10.1145/1613676.1613680.
  • [15] Paolo Ferragina and Giovanni Manzini. Indexing compressed text. J. ACM, 52(4):552–581, jul 2005. doi:10.1145/1082036.1082039.
  • [16] Travis Gagie, Giovanni Manzini, and Jouni Sirén. Wheeler graphs: A framework for BWT-based data structures. Theoretical Computer Science, 698:67 – 78, 2017. Algorithms, Strings and Theoretical Approaches in the Big Data Era (In Honor of the 60th Birthday of Professor Raffaele Giancarlo). doi:10.1016/j.tcs.2017.06.016.
  • [17] Dan Gusfield. Algorithms on Strings, Trees, and Sequences: Computer Science and Computational Biology. Cambridge University Press, 1997. doi:10.1017/CBO9780511574931.
  • [18] Ramana M. Idury and Michael S. Waterman. A new algorithm for DNA sequence assembly. Journal of computational biology : a journal of computational molecular cell biology, 2 2:291–306, 1995.
  • [19] Juha Kärkkäinen, Peter Sanders, and Stefan Burkhardt. Linear work suffix array construction. J. ACM, 53(6):918–936, nov 2006. doi:10.1145/1217856.1217858.
  • [20] Dong Kyue Kim, Jeong Seop Sim, Heejin Park, and Kunsoo Park. Constructing suffix arrays in linear time. Journal of Discrete Algorithms, 3(2):126–142, 2005. Combinatorial Pattern Matching (CPM) Special Issue. URL: https://www.sciencedirect.com/science/article/pii/S1570866704000486, doi:https://doi.org/10.1016/j.jda.2004.08.019.
  • [21] Sung-Hwan Kim, Francisco Olivares, and Nicola Prezza. Faster prefix-sorting algorithms for deterministic finite automata. In Laurent Bulteau and Zsuzsanna Lipták, editors, 34th Annual Symposium on Combinatorial Pattern Matching, CPM 2023, June 26-28, 2023, Marne-la-Vallée, France, volume 259 of LIPIcs, pages 16:1–16:16. Schloss Dagstuhl - Leibniz-Zentrum für Informatik, 2023. doi:10.4230/LIPIcs.CPM.2023.16.
  • [22] Pang Ko and Srinivas Aluru. Space efficient linear time construction of suffix arrays. Journal of Discrete Algorithms, 3(2):143–156, 2005. Combinatorial Pattern Matching (CPM) Special Issue. URL: https://www.sciencedirect.com/science/article/pii/S1570866704000498, doi:https://doi.org/10.1016/j.jda.2004.08.002.
  • [23] Veli Mäkinen, Niko Välimäki, and Jouni Sirén. Indexing graphs for path queries with applications in genome research. IEEE/ACM Transactions on Computational Biology and Bioinformatics, 11:375–388, 2014.
  • [24] U. Manber and G. Myers. Suffix arrays: A new method for on-line string searches. SIAM J. Comput., 22(5):935–948, 1993. doi:10.1137/0222058.
  • [25] Joong Chae Na. Linear-time construction of compressed suffix arrays using o(n log n)-bit working space for large alphabets. In Alberto Apostolico, Maxime Crochemore, and Kunsoo Park, editors, Combinatorial Pattern Matching, pages 57–67, Berlin, Heidelberg, 2005. Springer Berlin Heidelberg.
  • [26] Robert Paige and Robert E. Tarjan. Three partition refinement algorithms. SIAM Journal on Computing, 16(6):973–989, 1987. arXiv:https://doi.org/10.1137/0216062, doi:10.1137/0216062.
  • [27] Simon J. Puglisi, W. F. Smyth, and Andrew H. Turpin. A taxonomy of suffix array construction algorithms. ACM Comput. Surv., 39(2):4–es, jul 2007. doi:10.1145/1242471.1242472.
  • [28] Kunihiko Sadakane. Compressed suffix trees with full functionality. Theory Comput. Syst., 41(4):589–607, 2007. doi:10.1007/s00224-006-1198-x.
  • [29] Jared T. Simpson and Richard Durbin. Efficient construction of an assembly string graph using the FM-index. Bioinformatics, 26(12):i367–i373, 06 2010. arXiv:https://academic.oup.com/bioinformatics/article-pdf/26/12/i367/48859513/bioinformatics\_26\_12\_i367.pdf, doi:10.1093/bioinformatics/btq217.
  • [30] P. Weiner. Linear pattern matching algorithms. In Proc. 14th IEEE Annual Symposium on Switching and Automata Theory, pages 1–11, 1973. doi:10.1109/SWAT.1973.13.

Appendix A Proofs from Section 3

Statement of Lemma 3.2. Let αΣω\alpha\in\Sigma^{\omega}. Let aΣa\in\Sigma and αΣω\alpha^{\prime}\in\Sigma^{\omega} such that α=aα\alpha=a\alpha^{\prime}. Then:

  1. 1.

    τ(α)=2\tau(\alpha)=2 if and only if α=aω\alpha^{\prime}=a^{\omega}, if and only if α=aω\alpha=a^{\omega}.

  2. 2.

    τ(α)2\tau(\alpha)\not=2 if and only if αaω\alpha^{\prime}\not=a^{\omega}, if and only if αaω\alpha\not=a^{\omega}.

Assume that τ(α)2\tau(\alpha)\not=2. Then, there exist unique cΣ{a}c\in\Sigma\setminus\{a\}, αΣω\alpha^{\prime\prime}\in\Sigma^{\omega} and i0i\geq 0 such that α=aicα\alpha^{\prime}=a^{i}c\alpha^{\prime\prime} (and so α=ai+1cα\alpha=a^{i+1}c\alpha^{\prime\prime}). Moreover:

  1. 1.

    τ(α)=1\tau(\alpha)=1 if and only if cac\prec a, if and only if αaω\alpha^{\prime}\prec a^{\omega}, if and only if αaω\alpha\prec a^{\omega}.

  2. 2.

    τ(α)=3\tau(\alpha)=3 if and only if aca\prec c, if and only if aωαa^{\omega}\prec\alpha^{\prime}, if and only if aωαa^{\omega}\prec\alpha,

Proof A.1.
  1. 1.

    Let us prove that τ(α)=2\tau(\alpha)=2 if and only if α=aω\alpha^{\prime}=a^{\omega}.

    ()(\Leftarrow) If α=aω\alpha^{\prime}=a^{\omega}, we have α=aα=aaω=aω=α\alpha=a\alpha^{\prime}=aa^{\omega}=a^{\omega}=\alpha^{\prime}, so τ(α)=2\tau(\alpha)=2.

    ()(\Rightarrow) Assume that τ(α)=2\tau(\alpha)=2 and so α=α\alpha=\alpha^{\prime}. In order to prove that α=aω\alpha^{\prime}=a^{\omega}, it will suffice to show that for every i1i\geq 1 it holds α=aiα\alpha^{\prime}=a^{i}\alpha^{\prime}. We proceed by induction on ii. Notice that α=α=aα\alpha^{\prime}=\alpha=a\alpha^{\prime}, which proves our claim for i=1i=1. Now, assume that i>1i>1. By the inductive hypothesis, we can assume α=ai1α\alpha^{\prime}=a^{i-1}\alpha^{\prime}. Hence, α=α=aα=a(ai1α)=aiα\alpha^{\prime}=\alpha=a\alpha^{\prime}=a(a^{i-1}\alpha^{\prime})=a^{i}\alpha^{\prime}.

    Let us prove that α=aω\alpha^{\prime}=a^{\omega} if and only if α=aω\alpha=a^{\omega}. We have α=aω\alpha^{\prime}=a^{\omega} if and only if aα=aaωa\alpha^{\prime}=aa^{\omega}, if and only if α=aω\alpha=a^{\omega}.

  2. 2.

    It follows by negating the first point.

Now assume that τ(α)2\tau(\alpha)\not=2. Since αaω\alpha^{\prime}\not=a^{\omega}, then there exist unique cΣ{a}c\in\Sigma\setminus\{a\}, αΣω\alpha^{\prime\prime}\in\Sigma^{\omega} and i0i\geq 0 such that α=aicα\alpha^{\prime}=a^{i}c\alpha^{\prime\prime}. Moreover:

  1. 1.

    We have τ(α)=1\tau(\alpha)=1 if and only if αα\alpha^{\prime}\prec\alpha, if and only if αaα\alpha^{\prime}\prec a\alpha^{\prime}, if and only if aicαa(aicα)a^{i}c\alpha^{\prime\prime}\prec a(a^{i}c\alpha^{\prime\prime}), if and only if aicαai(acα)a^{i}c\alpha^{\prime\prime}\prec a^{i}(ac\alpha^{\prime\prime}), if and only if cαacαc\alpha^{\prime\prime}\prec ac\alpha^{\prime\prime}, if and only if cac\prec a, if and only if aicαaωa^{i}c\alpha^{\prime\prime}\prec a^{\omega}, if and only if αaω\alpha^{\prime}\prec a^{\omega}, if and only if aαaaωa\alpha^{\prime}\prec aa^{\omega}, if and only if αaω\alpha\prec a^{\omega}.

  2. 2.

    It follows by negating the previous points.

Statement of Corollary 3.3. Let α,βΣω\alpha,\beta\in\Sigma^{\omega}. Let a,bΣa,b\in\Sigma and α,βΣω\alpha^{\prime},\beta^{\prime}\in\Sigma^{\omega} such that α=aα\alpha=a\alpha^{\prime} and β=bβ\beta=b\beta^{\prime}. Then:

  1. 1.

    If a=ba=b and τ(α)=τ(β)=2\tau(\alpha)=\tau(\beta)=2, then α=β\alpha=\beta.

  2. 2.

    If a=ba=b and τ(α)<τ(β)\tau(\alpha)<\tau(\beta), then αβ\alpha\prec\beta. Equivalently, if a=ba=b and αβ\alpha\preceq\beta, then τ(α)τ(β)\tau(\alpha)\leq\tau(\beta).

Proof A.2.
  1. 1.

    By Lemma 3.2 we have α=aω=bω=β\alpha=a^{\omega}=b^{\omega}=\beta.

  2. 2.

    Let us prove that, if a=ba=b and τ(α)<τ(β)\tau(\alpha)<\tau(\beta), then αβ\alpha\prec\beta. We distinguish three cases.

    1. (a)

      τ(α)=1\tau(\alpha)=1 and τ(β)=2\tau(\beta)=2. Then by Lemma 3.2 we have αaω=bω=β\alpha\prec a^{\omega}=b^{\omega}=\beta.

    2. (b)

      τ(α)=2\tau(\alpha)=2 and τ(β)=3\tau(\beta)=3. Then by Lemma 3.2 we have α=aω=bωβ\alpha=a^{\omega}=b^{\omega}\prec\beta.

    3. (c)

      τ(α)=1\tau(\alpha)=1 and τ(β)=3\tau(\beta)=3. Then by Lemma 3.2 we have αaω=bωβ\alpha\prec a^{\omega}=b^{\omega}\prec\beta.

    Lastly, if a=ba=b and αβ\alpha\preceq\beta, then τ(α)τ(β)\tau(\alpha)\leq\tau(\beta), because if it were τ(β)τ(α)\tau(\beta)\leq\tau(\alpha), then we would conclude βα\beta\prec\alpha, a contradiction.

Appendix B Proofs from Section 4

Statement of Lemma 4.2. Let G=(V,E)G=(V,E) be a graph. Then, in O(|E|)O(|E|) time we can build a trimmed graph G=(V,E)G^{*}=(V^{*},E^{*}), with V={u|uV}V^{*}=\{u^{*}\;|\;u\in V\}, such that for every uVu\in V it holds minu=minu\min_{u^{*}}=\min_{u}. In particular, τ(u)=τ(u)\tau(u^{*})=\tau(u) for every uVu\in V.

Proof B.1.

Define λ(u)=λ(u)\lambda(u^{*})=\lambda(u) for every uVu^{*}\in V^{*}, F={(u,v)E|τ(v)=1,λ(v)λ(u)}F=\{(u,v)\in E\;|\;\tau(v)=1,\lambda(v)\prec\lambda(u)\}, and E={(u,v)|(u,v)EF}E^{*}=\{(u^{*},v^{*})\;|\;(u,v)\in E\setminus F\}. Essentially, we define GG^{*} by removing from GG all edges that violate the definition of trimmed graph. It is easy to realize that GG^{*} still satisfies the assumptions on graphs that we make in this paper.

  1. 1.

    All edges entering the same node in VV^{*} have the same label by construction.

  2. 2.

    Every vGv^{*}\in G^{*} has an incoming edge in GG^{*}. Indeed, if τ(v)1\tau(v)\not=1, then, if uVu\in V is any node such that (u,v)E(u,v)\in E, then (u,v)F(u,v)\not\in F and so (u,v)E(u^{*},v^{*})\in E^{*}. If τ(v)=1\tau(v)=1, then there exists uVu\in V such that (u,v)E(u,v)\in E and λ(u)λ(v)\lambda(u)\preceq\lambda(v), so (u,v)F(u,v)\not\in F and (u,v)E(u^{*},v^{*})\in E^{*}.

  3. 3.

    GG^{*} is deterministic because it is essentially a subgraph of a deterministic graph. More precisely, assume that (u,v).(u,z)E(u^{*},v^{*}).(u^{*},z^{*})\in E^{*} are such that λ(v)=λ(z)\lambda(v^{*})=\lambda(z^{*}). We must prove that v=zv^{*}=z^{*}, or equivalently, v=zv=z. From (u,v),(u,z)E(u^{*},v^{*}),(u^{*},z^{*})\in E^{*} we obtain (u,v),(u,z)E(u,v),(u,z)\in E, so from λ(v)=λ(v)=λ(z)=λ(z)\lambda(v)=\lambda(v^{*})=\lambda(z^{*})=\lambda(z) we obtain v=zv=z because GG is deterministic.

Now, let us prove that for every uVu\in V it holds minu=minu\min_{u^{*}}=\min_{u}. To this end we have to prove that (i) minuIu\min_{u}\in I_{u^{*}} and (ii) if αIu\alpha\in I_{u^{*}}, then minuα\min_{u}\preceq\alpha.

  1. 1.

    Let us prove that minuIu\min_{u}\in I_{u^{*}}. To this end, it will suffice to prove that, if (ui)i1(u_{i})_{i\geq 1} is an occurrence of minu\min_{u} starting at uu, then (ui)i1(u^{*}_{i})_{i\geq 1} is an occurrence of minu\min_{u} starting at uu^{*}. For every i1i\geq 1, we have λ(ui)=λ(ui)=minu[i]\lambda(u^{*}_{i})=\lambda(u_{i})=\min_{u}[i]. We are only left with showing that (ui+1,ui)E(u^{*}_{i+1},u^{*}_{i})\in E^{*} for every i1i\geq 1. We know that (ui+1,ui)E(u_{i+1},u_{i})\in E, so we only have to prove that (ui+1,ui)F(u_{i+1},u_{i})\not\in F. Assume that τ(ui)=1\tau(u_{i})=1. We must prove that λ(ui+1)λ(ui)\lambda(u_{i+1})\preceq\lambda(u_{i}). Since (uj)ji(u_{j})_{j\geq i} is an occurrence of minui\min_{u_{i}} starting at uiu_{i}, this is equivalent to proving that minui[2]minui[1]\min_{u_{i}}[2]\preceq\min_{u_{i}}[1], which indeed follows from τ(ui)=1\tau(u_{i})=1.

  2. 2.

    Let us prove that if αIu\alpha\in I_{u^{*}}, then αIu\alpha\in I_{u} (which implies minuα\min_{u}\preceq\alpha). Let (vi)i1(v^{*}_{i})_{i\geq 1} be an occurrence of α\alpha starting at uu^{*}. It will suffice to prove that (vi)i1(v_{i})_{i\geq 1} is an occurrence of α\alpha starting at uu. For every i1i\geq 1 we have (vi+1,vi)E(v_{i+1},v_{i})\in E because (vi+1,vi)E(v^{*}_{i+1},v^{*}_{i})\in E, and λ(vi)=λ(vi)=α[i]\lambda(v_{i})=\lambda(v_{i})=\alpha[i], so the conclusion follows.

In particular, τ(u)=τ(u)\tau(u^{*})=\tau(u) for every uVu\in V, and G=(V,E)G^{*}=(V^{*},E^{*}) is a trimmed graph.

B.1 Proofs from Section 4.1

Statement of Lemma 4.3. Let G=(V,E)G=(V,E) be a graph, and let u,vVu,v\in V.

  1. 1.

    If (u,v)E(u,v)\in E and λ(u)λ(v)\lambda(u)\prec\lambda(v), then τ(v)=1\tau(v)=1.

  2. 2.

    If (u,v)E(u,v)\in E, λ(u)=λ(v)\lambda(u)=\lambda(v) and τ(u)=1\tau(u)=1, then τ(v)=1\tau(v)=1.

Proof B.2.
  1. 1.

    Since λ(u)λ(v)\lambda(u)\prec\lambda(v), we have minuλ(v)ω\min_{u}\prec\lambda(v)^{\omega}. Moreover, λ(v)minuIv\lambda(v)\min_{u}\in I_{v}, hence minvλ(v)minuλ(v)λ(v)ω=λ(v)ω\min_{v}\preceq\lambda(v)\min_{u}\prec\lambda(v)\lambda(v)^{\omega}=\lambda(v)^{\omega}, so τ(v)=1\tau(v)=1 by Lemma 3.2.

  2. 2.

    Since τ(u)=1\tau(u)=1, then minuλ(u)ω\min_{u}\prec\lambda(u)^{\omega} by Lemma 3.2. From (u,v)E(u,v)\in E we obtain λ(v)minuIv\lambda(v)\min_{u}\in I_{v}, hence minvλ(v)minuλ(v)λ(u)ω=λ(v)λ(v)ω=λ(v)ω\min_{v}\preceq\lambda(v)\min_{u}\prec\lambda(v)\lambda(u)^{\omega}=\lambda(v)\lambda(v)^{\omega}=\lambda(v)^{\omega}, so again by Lemma 3.2 we conclude τ(v)=1\tau(v)=1.

Statement of Corollary 4.4. Let G=(V,E)G=(V,E) be a graph, and let uVu\in V. Then, τ(u)=1\tau(u)=1 if and only if there exist k2k\geq 2 and z1,,zkVz_{1},\dots,z_{k}\in V such that (i) (zi,zi+1)E(z_{i},z_{i+1})\in E for every 1ik11\leq i\leq k-1, (ii) zk=uz_{k}=u, (iii) λ(z1)λ(z2)\lambda(z_{1})\prec\lambda(z_{2}) and (iv) λ(z2)=λ(z3)==λ(zk)\lambda(z_{2})=\lambda(z_{3})=\dots=\lambda(z_{k}).

Proof B.3.

()(\Leftarrow) Let k2k\geq 2 and z1,,zkVz_{1},\dots,z_{k}\in V be such that (i) (zi,zi+1)E(z_{i},z_{i+1})\in E for every 1ik11\leq i\leq k-1, (ii) zk=uz_{k}=u, (iii) λ(z1)λ(z2)\lambda(z_{1})\prec\lambda(z_{2}) and (iv) λ(z2)=λ(z3)==λ(zk)\lambda(z_{2})=\lambda(z_{3})=\dots=\lambda(z_{k}). We must prove that τ(u)=1\tau(u)=1. Let us prove by induction that for every 2ik2\leq i\leq k it holds τ(zi)=1\tau(z_{i})=1 (and in particular τ(u)=τ(zk)=1\tau(u)=\tau(z_{k})=1). If i=2i=2, then from (z1,z2)E(z_{1},z_{2})\in E and λ(z1)λ(z2)\lambda(z_{1})\prec\lambda(z_{2}) we obtain τ(z2)=1\tau(z_{2})=1 by Lemma 4.3. Now, assume that 3ik3\leq i\leq k. By the inductive hypothesis, we have τ(zi1)=1\tau(z_{i-1})=1. Since (zi1,zi)E(z_{i-1},z_{i})\in E and λ(zi1)=λ(zi)\lambda(z_{i-1})=\lambda(z_{i}), then by Lemma 4.3 we conclude τ(zi)=1\tau(z_{i})=1.

()(\Rightarrow) Assume that τ(u)=1\tau(u)=1. Then, by Lemma 3.2 there exist cΣ{λ(u)}c\in\Sigma\setminus\{\lambda(u)\}, γΣω\gamma^{\prime}\in\Sigma^{\omega} and j1j\geq 1 such that minu=λ(u)jcγ\min_{u}=\lambda(u)^{j}c\gamma^{\prime} and cλ(u)c\prec\lambda(u). Let (ui)i1(u_{i})_{i\geq 1} be an occurrence of minu\min_{u} starting at uu. Then, (i) u=u1u=u_{1}, (ii) (ui+1,ui)E(u_{i+1},u_{i})\in E for every i1i\geq 1, (iii) λ(ui)=λ(u)\lambda(u_{i})=\lambda(u) for 1ij1\leq i\leq j and (iv) λ(uj+1)=c\lambda(u_{j+1})=c. As a consequence, if k=j+1k=j+1 and zi=uk+1iz_{i}=u_{k+1-i} for every 1ik1\leq i\leq k, then (i) (zi,zi+1)=(uki+1,uki)E(z_{i},z_{i+1})=(u_{k-i+1},u_{k-i})\in E for every 1ik11\leq i\leq k-1, (ii) zk=u1=uz_{k}=u_{1}=u, (iii) λ(z1)=λ(uk)=λ(uj+1)=cλ(u)=λ(uj)=λ(uk1)=λ(z2)\lambda(z_{1})=\lambda(u_{k})=\lambda(u_{j+1})=c\prec\lambda(u)=\lambda(u_{j})=\lambda(u_{k-1})=\lambda(z_{2}) and (iv) λ(z2)=λ(z3)==λ(zk)\lambda(z_{2})=\lambda(z_{3})=\dots=\lambda(z_{k}) because λ(uj)=λ(uj1)==λ(u1)\lambda(u_{j})=\lambda(u_{j-1})=\dots=\lambda(u_{1}).

Statement of Corollary 4.5. Let G=(V,E)G=(V,E) be a graph. We can determine all uVu\in V such that τ(u)=1\tau(u)=1 in time O(|E|)O(|E|).

Proof B.4.

In our algorithm, we will mark exactly all nodes uu such that τ(u)=1\tau(u)=1 by means of a graph traversal. We will use an (initially empty) queue. Initially, no node is marked. First, process all edges (u,v)E(u,v)\in E and, if λ(u)λ(v)\lambda(u)\prec\lambda(v) and vv has not been marked before, then mark vv and add vv to the queue. This step takes O(m)O(m) time. Next, recursively pick an element uu from the queue, and consider the unique vQv\in Q such that (u,v)E(u,v)\in E and λ(u)=λ(v)\lambda(u)=\lambda(v) (if it exists). If vv has not been marked before, then mark vv and add vv to the queue. This step also takes O(|E|)O(|E|) time because each edge is considered at most once. At the end of the algorithm, by Corollary 4.4 a node uu has been marked if and only if τ(u)=1\tau(u)=1.

Statement of Lemma 4.6. Let G=(V,E)G=(V,E) be a graph, and let uVu\in V such that τ(u)1\tau(u)\not=1. Then, we have τ(u)=2\tau(u)=2 if and only if there exist k2k\geq 2 and z1,,zkVz_{1},\dots,z_{k}\in V such that (i) (zi+1,zi)E(z_{i+1},z_{i})\in E for every 1ik11\leq i\leq k-1, (ii) z1=uz_{1}=u, (iii) zk=zjz_{k}=z_{j} for some 1jk11\leq j\leq k-1 and (iv) λ(z1)=λ(z2)==λ(zk)\lambda(z_{1})=\lambda(z_{2})=\dots=\lambda(z_{k}).

In particular, such z1,,zkVz_{1},\dots,z_{k}\in V must satisfy τ(zi)=2\tau(z_{i})=2 for every 1ik1\leq i\leq k.

Proof B.5.

()(\Leftarrow) Let k2k\geq 2 and z1,,zkVz_{1},\dots,z_{k}\in V such that (i) (zi+1,zi)E(z_{i+1},z_{i})\in E for every 1ik11\leq i\leq k-1, (ii) z1=uz_{1}=u, (iii) zk=ziz_{k}=z_{i} for some 1jk11\leq j\leq k-1 and (iv) λ(z1)=λ(z2)==λ(zk)\lambda(z_{1})=\lambda(z_{2})=\dots=\lambda(z_{k}). We must prove that τ(u)=2\tau(u)=2. Notice that minu=minz1λ(z1)λ(z2)λ(zj1)minzj=λ(u)j1minzj\min_{u}=\min_{z_{1}}\preceq\lambda(z_{1})\lambda(z_{2})\dots\lambda(z_{j-1})\min_{z_{j}}=\lambda(u)^{j-1}\min_{z_{j}}. Similarly, we have minzjλ(zj)λ(zj+1)λ(zk1)minzk=λ(u)kjminzj\min_{z_{j}}\preceq\lambda(z_{j})\lambda(z_{j+1})\dots\lambda(z_{k-1})\min_{z_{k}}=\lambda(u)^{k-j}\min_{z_{j}}. Since minzjλ(u)kiminzj\min_{z_{j}}\preceq\lambda(u)^{k-i}\min_{z_{j}}, then by induction we obtain minzi(λ(u)kj)hminzi\min_{z_{i}}\preceq(\lambda(u)^{k-j})^{h}\min_{z_{i}} for every h1h\geq 1, and so minzjλ(u)ω\min_{z_{j}}\preceq\lambda(u)^{\omega}. As a consequence, minuλ(u)j1minzjλ(u)j1λ(u)ω=λ(u)ω\min_{u}\preceq\lambda(u)^{j-1}\min_{z_{j}}\preceq\lambda(u)^{j-1}\lambda(u)^{\omega}=\lambda(u)^{\omega}, and so τ(u)2\tau(u)\preceq 2 by Lemma 3.2. Since τ(u)1\tau(u)\not=1, we conclude τ(u)=2\tau(u)=2.

()(\Rightarrow) Assume that τ(u)=2\tau(u)=2. In particular, λ(u)ωIu\lambda(u)^{\omega}\in I_{u}, so there exists an occurrence (zi)i1(z_{i})_{i\geq 1} of λ(u)ω\lambda(u)^{\omega} starting at uu. This means that (i) (zi+1,zi)E(z_{i+1},z_{i})\in E for every i1i\geq 1, (ii) z1=uz_{1}=u and λ(zi)=λ(u)\lambda(z_{i})=\lambda(u) for every i1i\geq 1. Since VV is finite, there exist 1j<k1\leq j<k such that zj=zkz_{j}=z_{k}. Then, k2k\geq 2 and z1,,zkVz_{1},\dots,z_{k}\in V have the desired properties.

Now, let us prove that, if there exist z1,,zkVz_{1},\dots,z_{k}\in V with the desired properties, then τ(zi)=2\tau(z_{i})=2 for every 1ik1\leq i\leq k. Notice that it must be τ(zi)1\tau(z_{i})\not=1 for every 1ik1\leq i\leq k, because otherwise by Corollary 4.4 we would conclude τ(u)=1\tau(u)=1. As a consequence, it must be τ(zi)=2\tau(z_{i})=2 for every 1ik1\leq i\leq k by the characterization that we have just proved.

Statement of Corollary 4.7. Let G=(V,E)G=(V,E) be a graph. We can determine all uVu\in V such that τ(u)=2\tau(u)=2 in time O(|E|)O(|E|).

Proof B.6.

By Corollary 4.5, we can assume that we have already computed all uVu\in V such that τ(u)=1\tau(u)=1. In order to compute all uVu\in V such that τ(u)=2\tau(u)=2, we explore GG by using a breadth first search algorithm, with the following modifications: (i) we only start from nodes uVu\in V such that τ(u)1\tau(u)\not=1, (ii) we follow edges in a backward fashion, not in a forward fashion and (iii) if we start from uVu\in V, we explore the graph by only following edges labeled λ(u)\lambda(u) (that is, we first consider only all uVu^{\prime}\in V such that (u,u)E(u^{\prime},u)\in E and λ(u)=λ(u)\lambda(u^{\prime})=\lambda(u), then we repeat this step in BFS fashion). Note that during the search we can never encounter a node vv such that τ(v)=1\tau(v)=1, otherwise by Corollary 4.4 we would obtain τ(u)=1\tau(u)=1. During the search, we will infer the value τ(v)\tau(v) for every node vv that we encounter. Assume that we start the exploration from node uVu\in V. If during the exploration at uu at some point we encounter a node vv for which we have already concluded that τ(v)=3\tau(v)=3, then we do not consider the edges entering vv and we backtrack (because all the ziz_{i}’s in the characterization of Lemma 4.6 must satisfy τ(zi)=2\tau(z_{i})=2). If during the exploration at uu at some point we encounter a node vv for which we have already concluded that τ(v)=2\tau(v)=2, then we backtrack straight to uu and, by Lemma 4.6, we can also conclude that all zz that we encounter during the backtracking (including uu) are such that τ(z)=2\tau(z)=2. If during the exploration of uu at some point we encounter the same node twice, then we backtrack straight to uu and, by Lemma 4.6, we can also conclude that all zz that we encounter during the backtracking (including uu) are such that τ(z)=2\tau(z)=2. If during the exploration of uu we never encounter a node vv for which we have already determined τ(v)\tau(v) and we never encounter the same node twice (and so at some point we get stuck), by Lemma 4.6 we can conclude that all vv that we have encountered (including uu) are such that τ(v)=3\tau(v)=3. The algorithm runs in O(|E|)O(|E|) time because we never follow the same edge twice.

B.2 Proofs from Section 4.2

Statement of Lemma 4.9. Let G=(V,E)G=(V,E) be a graph. Let u,vVu,v\in V be such that minu=minv\min_{u}=\min_{v}. Let (ui)i1(u_{i})_{i\geq 1} be an occurrence of minu\min_{u} and let (vi)i1(v_{i})_{i\geq 1} be an occurrence of minv\min_{v}. Then:

  1. 1.

    λ(ui)=λ(vi)\lambda(u_{i})=\lambda(v_{i}) for every i1i\geq 1.

  2. 2.

    minui=minvi\min_{u_{i}}=\min_{v_{i}} for every i1i\geq 1.

  3. 3.

    τ(ui)=τ(vi)\tau(u_{i})=\tau(v_{i}) for every i1i\geq 1.

In particular, the previous results hold if u=vu=v and (ui)i1(u_{i})_{i\geq 1} and (vi)i1(v_{i})_{i\geq 1} are two distinct occurrences of minu\min_{u}.

Proof B.7.
  1. 1.

    For every i1i\geq 1 we have λ(ui)=minu[i]=minv[i]=λ(vi)\lambda(u_{i})=\min_{u}[i]=\min_{v}[i]=\lambda(v_{i}).

  2. 2.

    Fix i1i\geq 1. We have minu[1,i1]minui=minu=minv=minv[1,i1]minvi\min_{u}[1,i-1]\min_{u_{i}}=\min_{u}=\min_{v}=\min_{v}[1,i-1]\min_{v_{i}} and so minui=minvi\min_{u_{i}}=\min_{v_{i}}

  3. 3.

    Fix i1i\geq 1. By the previous points we have minui=minvi\min_{u_{i}}=\min_{v_{i}} and λ(ui)=λ(vi)\lambda(u_{i})=\lambda(v_{i}), so by Corollary 3.3 we conclude that it must necessarily be τ(ui)=τ(vi)\tau(u_{i})=\tau(v_{i}).

Statement of Lemma 4.10. Let G=(V,E)G=(V,E) be a graph. Let uVu\in V and let (ui)i1(u_{i})_{i\geq 1} an occurrence of minu\min_{u} starting at uu. Let k1k\geq 1 be such that τ(u1)=τ(u2)==τ(uk1)=τ(uk)2\tau(u_{1})=\tau(u_{2})=\dots=\tau(u_{k-1})=\tau(u_{k})\not=2. Then, u1,,uku_{1},\dots,u_{k} are pairwise distinct. In particular, k|V|k\leq|V|.

Proof B.8.

We assume that τ(u1)=τ(u2)==τ(uk1)=τ(uk)=1\tau(u_{1})=\tau(u_{2})=\dots=\tau(u_{k-1})=\tau(u_{k})=1, because the case τ(u1)=τ(u2)==τ(uk1)=τ(uk)=3\tau(u_{1})=\tau(u_{2})=\dots=\tau(u_{k-1})=\tau(u_{k})=3 is symmetrical. Notice that for every 1lk1\leq l\leq k we have that (ui)il(u_{i})_{i\geq l} is an occurrence of minul\min_{u_{l}} starting at ulu_{l} and τ(ul)=1\tau(u_{l})=1, so λ(ul+1)λ(ul)\lambda(u_{l+1})\preceq\lambda(u_{l}). In particular, for every 1rsk+11\leq r\leq s\leq k+1 we have λ(us)λ(ur)\lambda(u_{s})\preceq\lambda(u_{r}).

Suppose for sake of contradiction that there exist 1i<jk1\leq i<j\leq k such that ui=uju_{i}=u_{j}. Then, minui=minui[1,ji]minuj=minui[1,ji]minui\min_{u_{i}}=\min_{u_{i}}[1,j-i]\min_{u_{j}}=\min_{u_{i}}[1,j-i]\min_{u_{i}}. By induction, we obtain minui=(minui[1,ji])tminui\min_{u_{i}}=(\min_{u_{i}}[1,j-i])^{t}\min_{u_{i}} for every t1t\geq 1, and so minui=(minui[1,ji])ω\min_{u_{i}}=(\min_{u_{i}}[1,j-i])^{\omega}. For every 1hj1\leq h\leq j, we have λ(uj)λ(uh)λ(ui)\lambda(u_{j})\preceq\lambda(u_{h})\preceq\lambda(u_{i}). Since ui=uju_{i}=u_{j}, then λ(ui)=λ(uj)\lambda(u_{i})=\lambda(u_{j}), so λ(uh)=λ(ui)\lambda(u_{h})=\lambda(u_{i}) for every ihji\leq h\leq j, which implies minui[1,ji]=λ(ui)λ(ui+1)λ(uj1)=λ(ui)ji\min_{u_{i}}[1,j-i]=\lambda(u_{i})\lambda(u_{i+1})\dots\lambda(u_{j-1})=\lambda(u_{i})^{j-i} and so minui=(minui[1,ji])ω=λ(ui)ω\min_{u_{i}}=(\min_{u_{i}}[1,j-i])^{\omega}=\lambda(u_{i})^{\omega}. By Lemma 3.2 we conclude τ(ui)=2\tau(u_{i})=2, a contradiction.

Statement of Lemma 4.12. Let G=(V,E)G=(V,E) be a graph. Let uVu\in V such that τ(u)=3\tau(u)=3. Then, minu[i+1]minu[i]\min_{u}[i+1]\preceq\min_{u}[i] for every 2iu12\leq i\leq\ell_{u}-1. In particular, if 2iju2\leq i\leq j\leq\ell_{u}, then minu[j]minu[i]\min_{u}[j]\preceq\min_{u}[i].

Proof B.9.

Fix 2iu2\leq i\leq\ell_{u}. By the definition of u)\ell_{u}) we have τ(ui)=1\tau(u_{i})=1, so minu[i+1]=minui[2]minui[1]=minu[i]\min_{u}[i+1]=\min_{u_{i}}[2]\preceq\min_{u_{i}}[1]=\min_{u}[i].

Statement of Lemma 4.13. Let G=(V,E)G=(V,E) be a graph. Let uVu\in V such that τ(u)=3\tau(u)=3.

  1. 1.

    Gi(u)G_{i}(u) is well-defined and nonempty for every 1iu1\leq i\leq\ell_{u}.

  2. 2.

    Let (ui)i1(u_{i})_{i\geq 1} be an occurrence of minu\min_{u} starting at uu. Then, uiGi(u)u_{i}\in G_{i}(u) for every 1iu1\leq i\leq\ell_{u}. In particular, τ(ui)=τ(Gi(u))\tau(u_{i})=\tau(G_{i}(u)) and minu[i]=λ(ui)=λ(Gi(u))\min_{u}[i]=\lambda(u_{i})=\lambda(G_{i}(u)) for every 1iu1\leq i\leq\ell_{u}.

  3. 3.

    For every 1iu1\leq i\leq\ell_{u} and for every vGi(u)v\in G_{i}(u) there exists an occurrence of minu[1,i1]\min_{u}[1,i-1] starting at uu and ending at vv.

Proof B.10.

We proceed by induction on i1i\geq 1. If i=1i=1, then Gi(u)={u}G_{i}(u)=\{u\} is well-defined and nonempty, and u1=uGi(u)u_{1}=u\in G_{i}(u). Moreover, if vGi(u)v\in G_{i}(u), then v=uv=u, and there exists an occurrence of ϵ=min[1,0]\epsilon=\min[1,0] starting and ending at uu. Now, assume that 2iu2\leq i\leq\ell_{u}.

First, notice that 𝙵({vQ|(vGi1(u))((v,v)E)})\mathtt{F}(\{v^{\prime}\in Q\;|\;(\exists v\in G_{i-1}(u))((v^{\prime},v)\in E)\}) is well-defined and nonempty. Indeed, by the inductive hypothesis Gi1(u)G_{i-1}(u) is well-defined and nonempty, so {vQ|(vGi1(u))((v,v)E)}\{v^{\prime}\in Q\;|\;(\exists v\in G_{i-1}(u))((v^{\prime},v)\in E)\} is nonempty because every node has an incoming edge.

Let us prove that ui𝙵({vQ|(vGi1(u))((v,v)E)})u_{i}\in\mathtt{F}(\{v^{\prime}\in Q\;|\;(\exists v\in G_{i-1}(u))((v^{\prime},v)\in E)\}). Let R={vQ|(vGi1(u))((v,v)E)}R=\{v^{\prime}\in Q\;|\;(\exists v\in G_{i-1}(u))((v^{\prime},v)\in E)\} and R=argminvRλ(v)R^{\prime}=\operatorname*{arg\,min}_{v\in R}\lambda(v). We must prove that uiargminuRτ(u)u_{i}\in\operatorname*{arg\,min}_{u\in R^{\prime}}\tau(u).

  1. 1.

    Let us prove that uiRu_{i}\in R. By the inductive hypothesis, we know that ui1Gi1(u)u_{i-1}\in G_{i-1}(u). We know that (ui,ui1)E(u_{i},u_{i-1})\in E, so uiRu_{i}\in R.

  2. 2.

    Let us prove that uiRu_{i}\in R^{\prime}. Let vRv^{\prime}\in R. We must prove that λ(ui)λ(v)\lambda(u_{i})\preceq\lambda(v^{\prime}). Since vRv^{\prime}\in R, there exists vGi1(u)v\in G_{i-1}(u) such that (v,v)E(v^{\prime},v)\in E. From ui1,vGi1(u)u_{i-1},v\in G_{i-1}(u) we obtain λ(ui1)=λ(v)\lambda(u_{i-1})=\lambda(v). By the inductive hypothesis, there exists an occurrence of minu[1,i2]\min_{u}[1,i-2] starting at uu and ending at vv, so there exists an occurrence of minu[1,i2]λ(v)=minu[1,i2]λ(ui1)=minu[1,i1]\min_{u}[1,i-2]\lambda(v)=\min_{u}[1,i-2]\lambda(u_{i-1})=\min_{u}[1,i-1] starting at uu and ending at vv^{\prime}. Since minu=minu[1,i1]minui\min_{u}=\min_{u}[1,i-1]\min_{u_{i}}, it must be minuiminv\min_{u_{i}}\preceq\min_{v^{\prime}}, and in particular λ(ui)λ(v)\lambda(u_{i})\preceq\lambda(v^{\prime}).

  3. 3.

    Let us prove that uiargminuRτ(u)u_{i}\in\operatorname*{arg\,min}_{u\in R^{\prime}}\tau(u). Let vRv^{\prime}\in R^{\prime}. We must prove that τ(ui)τ(v)\tau(u_{i})\leq\tau(v). In particular, vRv^{\prime}\in R, so as before we obtain minuiminv\min_{u_{i}}\preceq\min_{v^{\prime}}. Moreover, from ui,vRu_{i},v^{\prime}\in R^{\prime} we obtain λ(ui)=λ(v)\lambda(u_{i})=\lambda(v^{\prime}). From Corollary 3.3 we conclude τ(ui)τ(v)\tau(u_{i})\leq\tau(v).

Let us prove that uij=2i1Gj(u)u_{i}\not\in\bigcup_{j=2}^{i-1}G_{j}(u). If i=ui=\ell_{u}, we have τ(ui)=τ(uu)2\tau(u_{i})=\tau(u_{\ell_{u}})\geq 2 and the conclusion follows because τ(Gj(u))=τ(uj)=1\tau(G_{j}(u))=\tau(u_{j})=1 for every 2ji12\leq j\leq i-1. As a consequence, in the following we can assume 2iu12\leq i\leq\ell_{u}-1. In particular, τ(ui)=1\tau(u_{i})=1, so by Lemma 3.2 there exists k1k\geq 1, cΣ{λ(ui)}c\in\Sigma\setminus\{\lambda(u_{i})\} and γΣω\gamma^{\prime}\in\Sigma^{\omega} such that minu=λ(u)kcγ\min_{u}=\lambda(u)^{k}c\gamma^{\prime} and cλ(ui)c\prec\lambda(u_{i}). Suppose for sake of contradiction that there exists 2ji1u22\leq j\leq i-1\leq\ell_{u}-2 such that uiGj(u)u_{i}\in G_{j}(u). We know that ujGj(u)u_{j}\in G_{j}(u), and in particular, λ(ui)=λ(Gj(u))=λ(uj)\lambda(u_{i})=\lambda(G_{j}(u))=\lambda(u_{j}). From Lemma 4.12 we obtain minu[i]minu[h]minu[j]\min_{u}[i]\preceq\min_{u}[h]\preceq\min_{u}[j] for every jhij\leq h\leq i, or equivalently λ(ui)λ(uh)λ(uj)\lambda(u_{i})\preceq\lambda(u_{h})\preceq\lambda(u_{j}) for every jhij\leq h\leq i. Since λ(uj)=λ(ui)\lambda(u_{j})=\lambda(u_{i}), we conclude λ(uh)=λ(ui)\lambda(u_{h})=\lambda(u_{i}) for every jhij\leq h\leq i. As a consequence, minu=minu[1,i1]minui=minu[1,j1]minu[j,i1]minui=minu[1,j1]λ(ui)ijλ(ui)kcγ=minu[1,j1]λ(ui)k+ijcγ\min_{u}=\min_{u}[1,i-1]\min_{u_{i}}=\min_{u}[1,j-1]\min_{u}[j,i-1]\min_{u_{i}}=\min_{u}[1,j-1]\lambda(u_{i})^{i-j}\lambda(u_{i})^{k}c\gamma^{\prime}=\min_{u}[1,j-1]\lambda(u_{i})^{k+i-j}c\gamma^{\prime}. On the other hand, since uiGj(u)u_{i}\in G_{j}(u) and j<ij<i, by the inductive hypothesis there exists an occurrence of minu[1,j1]\min_{u}[1,j-1] starting at uu and ending at uiu_{i}, so minu[1,j1]minuiIu\min_{u}[1,j-1]\min_{u_{i}}\in I_{u}. As a consequence, by the minimality of minu\min_{u} we obtain minuminu[1,j1]minui\min_{u}\preceq\min_{u}[1,j-1]\min_{u_{i}}, or equivalently, minu[1,j1]λ(ui)k+ijcγminu[1,j1]λ(ui)kcγ\min_{u}[1,j-1]\lambda(u_{i})^{k+i-j}c\gamma^{\prime}\preceq\min_{u}[1,j-1]\lambda(u_{i})^{k}c\gamma^{\prime}. Since j<ij<i, we obtain λ(ui)c\lambda(u_{i})\preceq c, a contradiction.

Let us prove that Gi(u)G_{i}(u) is well-defined and nonempty, and uiGi(u)u_{i}\in G_{i}(u). This follows from ui𝙵({vQ|(vGi1(u))((v,v)E)})u_{i}\in\mathtt{F}(\{v^{\prime}\in Q\;|\;(\exists v\in G_{i-1}(u))((v^{\prime},v)\in E)\}) and uij=2i1Gj(u)u_{i}\not\in\bigcup_{j=2}^{i-1}G_{j}(u).

Lastly, let us prove that if vGi(u)v^{\prime}\in G_{i}(u), then there exists an occurrence of min[1,i1]\min[1,i-1] starting at uu and ending at vv^{\prime}. Since vGi(u)v^{\prime}\in G_{i}(u), then there exists vGi1(u)v\in G_{i-1}(u) such that (v,v)E(v^{\prime},v)\in E. In particular, λ(v)=λ(Gi1(u))=λ(ui1)\lambda(v)=\lambda(G_{i-1}(u))=\lambda(u_{i-1}). By the inductive hypothesis, there exists an occurrence of minu[1,i2]\min_{u}[1,i-2] starting at uu and ending at vv, so there exists an occurrence of minu[1,i2]λ(v)=minu[1,i2]λ(ui1)=minu[1,i1]\min_{u}[1,i-2]\lambda(v)=\min_{u}[1,i-2]\lambda(u_{i-1})=\min_{u}[1,i-1] starting at uu and ending at vv^{\prime}.

Statement of Lemma 4.14. Let G=(V,E)G=(V,E) be a graph. Let u,vVu,v\in V such that τ(u)=τ(v)=3\tau(u)=\tau(v)=3. Assume that one of the following statements is true:

  1. 1.

    γu\gamma_{u} is not a prefix of γv\gamma_{v} and γuγv\gamma_{u}\prec\gamma_{v}.

  2. 2.

    γu=γv\gamma_{u}=\gamma_{v}, 𝚝u=2\mathtt{t}_{u}=2 and 𝚝v=3\mathtt{t}_{v}=3.

  3. 3.

    γv\gamma_{v} is a strict prefix of γu\gamma_{u}.

Then, minuminv\min_{u}\prec\min_{v}.

Equivalently, if minuminv\min_{u}\preceq\min_{v}, then one the following is true: (i) γu\gamma_{u} is not a prefix of γv\gamma_{v} and γuγv\gamma_{u}\prec\gamma_{v}; (ii) γu=γv\gamma_{u}=\gamma_{v} and 𝚝u𝚝v\mathtt{t}_{u}\leq\mathtt{t}_{v}; (iii) γv\gamma_{v} is a strict prefix of γu\gamma_{u}.

Proof B.11.

Let us prove that, if one of the three statements (1) - (3) is true, then minuminv\min_{u}\prec\min_{v}.

  1. 1.

    If minu[1,u]minv[1,v]\min_{u}[1,\ell_{u}]\prec\min_{v}[1,\ell_{v}] and minu[1,u]\min_{u}[1,\ell_{u}] is not a prefix of minv[1,v]\min_{v}[1,\ell_{v}], then minuminv\min_{u}\prec\min_{v}.

  2. 2.

    Assume that minu[1,u]=minv[1,v]\min_{u}[1,\ell_{u}]=\min_{v}[1,\ell_{v}] (so in particular u=v\ell_{u}=\ell_{v}), 𝚝u=2\mathtt{t}_{u}=2 and 𝚝v=3\mathtt{t}_{v}=3. Let (ui)i1(u_{i})_{i\geq 1} be an occurrence of uu, and let (vi)i1(v_{i})_{i\geq 1} be an occurrence of vv. It holds minu=minu[1,u1]minuu\min_{u}=\min_{u}[1,\ell_{u}-1]\min_{u_{\ell_{u}}} and minv=minv[1,v1]minvv\min_{v}=\min_{v}[1,\ell_{v}-1]\min_{v_{\ell_{v}}}. Since minu[1,u1]=minv[1,v1]\min_{u}[1,\ell_{u}-1]=\min_{v}[1,\ell_{v}-1], in order to prove that minuminv\min_{u}\prec\min_{v} we only have to show that minuuminvv\min_{u_{\ell_{u}}}\prec\min_{v_{\ell_{v}}}. By Lemma 4.13 we have uuGu(u)u_{\ell_{u}}\in G_{\ell_{u}}(u) and vvGv(v)v_{\ell_{v}}\in G_{\ell_{v}}(v), so λ(uu)=minu[u]=minv[v]=λ(vv)\lambda(u_{\ell_{u}})=\min_{u}[\ell_{u}]=\min_{v}[\ell_{v}]=\lambda(v_{\ell_{v}}). Since τ(uu)=τ(Gu(u))=𝚝u=2\tau(u_{\ell_{u}})=\tau(G_{\ell_{u}}(u))=\mathtt{t}_{u}=2 and τ(vv)=τ(Gv(v))=𝚝v=3\tau(v_{\ell_{v}})=\tau(G_{\ell_{v}}(v))=\mathtt{t}_{v}=3, we conclude minuuminvv\min_{u_{\ell_{u}}}\prec\min_{v_{\ell_{v}}} by Corollary 3.3.

  3. 3.

    Assume that γv\gamma_{v} is a strict prefix of γu\gamma_{u} (hence v<u\ell_{v}<\ell_{u}). In particular, minu[1,v]=minv[1,v]\min_{u}[1,\ell_{v}]=\min_{v}[1,\ell_{v}], so similarly to the previous case we only have to prove that minuvminvv\min_{u_{\ell_{v}}}\prec\min_{v_{\ell_{v}}}. Again, we obtain uvGv(u)u_{\ell_{v}}\in G_{\ell_{v}}(u), vvGu(v)v_{\ell_{v}}\in G_{\ell_{u}}(v), λ(uv)=λ(vv)\lambda(u_{\ell_{v}})=\lambda(v_{\ell_{v}}) and τ(vv)=3\tau(v_{\ell_{v}})=3. Since v<u\ell_{v}<\ell_{u}, the minimality of u\ell_{u} implies τ(uv)=τ(Gv(u))=1\tau(u_{\ell_{v}})=\tau(G_{\ell_{v}}(u))=1. By Corollary 3.3 we conclude minuvminvv\min_{u_{\ell_{v}}}\prec\min_{v_{\ell_{v}}}.

Now, let us prove that if minuminv\min_{u}\preceq\min_{v}, then one the following is true: (i) γu\gamma_{u} is not a prefix of γv\gamma_{v} and γuγv\gamma_{u}\prec\gamma_{v}; (ii) γu=γv\gamma_{u}=\gamma_{v} and 𝚝u𝚝v\mathtt{t}_{u}\leq\mathtt{t}_{v}; (iii) γv\gamma_{v} is a strict prefix of γu\gamma_{u}. If this were not true, then one of the following would be true: (1) γv\gamma_{v} is not a prefix of γu\gamma_{u} and γvγu\gamma_{v}\prec\gamma_{u}; (2) γv=γu\gamma_{v}=\gamma_{u}, 𝚝v=2\mathtt{t}_{v}=2 and 𝚝u=3\mathtt{t}_{u}=3; (3) γu\gamma_{u} is a strict prefix of γv\gamma_{v}. We would then conclude minvminu\min_{v}\prec\min_{u}, a contradiction.

Statement of Lemma 4.15. Let G=(V,E)G=(V,E) be a graph. Let u,vVu,v\in V be such that τ(u)=τ(v)=3\tau(u)=\tau(v)=3, γu=γv\gamma_{u}=\gamma_{v} and 𝚝u=𝚝v=2\mathtt{t}_{u}=\mathtt{t}_{v}=2. Then, minu=minv\min_{u}=\min_{v}.

Proof B.12.

First, note that γu=γv\gamma_{u}=\gamma_{v} implies u=v\ell_{u}=\ell_{v}, and it is equivalent to minu[1,u]=minv[1,v]\min_{u}[1,\ell_{u}]=\min_{v}[1,\ell_{v}]. Now, let (ui)i1(u_{i})_{i\geq 1} be an occurrence of minu\min_{u} starting at uu, and let (vi)i1(v_{i})_{i\geq 1} be an occurrence of minv\min_{v} starting at vv. We have minu=minu[1,u1]minuu\min_{u}=\min_{u}[1,\ell_{u}-1]\min_{u_{\ell_{u}}} and minv=minv[1,v1]minvv\min_{v}=\min_{v}[1,\ell_{v}-1]\min_{v_{\ell_{v}}}. Since minu[1,u1]=minv[1,v1]\min_{u}[1,\ell_{u}-1]=\min_{v}[1,\ell_{v}-1], we only have to show that minuu=minvv\min_{u_{\ell_{u}}}=\min_{v_{\ell_{v}}}. Notice that λ(uu)=minu[u]=minv[v]=λ(vv)\lambda(u_{\ell_{u}})=\min_{u}[\ell_{u}]=\min_{v}[\ell_{v}]=\lambda(v_{\ell_{v}}), and by Lemma 4.13 we have τ(uu)=τ(Gu(u))=𝚝u=2\tau(u_{\ell_{u}})=\tau(G_{\ell_{u}}(u))=\mathtt{t}_{u}=2 and τ(vv)=τ(Gv(v))=𝚝v=2\tau(v_{\ell_{v}})=\tau(G_{\ell_{v}}(v))=\mathtt{t}_{v}=2. By Corollary 3.3 we conclude minuu=minvv\min_{u_{\ell_{u}}}=\min_{v_{\ell_{v}}}.

Statement of Lemma 4.16. Let G=(V,E)G=(V,E) be a graph. Let uVu\in V, and let (ui)i1(u_{i})_{i\geq 1} be an occurrence of minu\min_{u} starting at uu. Then, exactly one of the following holds true:

  1. 1.

    There exists i01i_{0}\geq 1 such that τ(ui)2\tau(u_{i})\not=2 for every 1i<i01\leq i<i_{0} and τ(ui)=2\tau(u_{i})=2 for every ii0i\geq i_{0}.

  2. 2.

    τ(ui)2\tau(u_{i})\not=2 for every i1i\geq 1, and both τ(ui)=1\tau(u_{i})=1 and τ(ui)=3\tau(u_{i})=3 are true for infinitely many ii’s.

Proof B.13.

If τ(ui)2\tau(u_{i})\not=2 for every i1i\geq 1, then both τ(ui)=1\tau(u_{i})=1 and τ(ui)=3\tau(u_{i})=3 are true for infinitely many ii’s, because otherwise we could choose j1j\geq 1 such that either τ(ui)=1\tau(u_{i})=1 for every iji\geq j or τ(ui)=3\tau(u_{i})=3 for every iji\geq j, and in both cases Lemma 4.10 leads to a contradiction, because (uj)ji(u_{j})_{j\geq i} is an occurrence of minuj\min_{u_{j}} starting at uju_{j} and VV is a finite set.

Now, assume that there exists i01i_{0}\geq 1 such that τ(ui0)=2\tau(u_{i_{0}})=2; without loss of generality, we can assume that i0i_{0} is the smallest integer with this property. We only have to prove that if ii0i\geq i_{0}, then τ(ui)=2\tau(u_{i})=2. Since τ(ui0)=2\tau(u_{i_{0}})=2, then minui0=(λ(ui0))ω\min_{u_{i_{0}}}=(\lambda(u_{i_{0}}))^{\omega} by Lemma 3.2. Since (uj)ji0(u_{j})_{j\geq i_{0}} in an occurrence of minui0\min_{u_{i_{0}}} starting at i0i_{0}, then λ(uj)=λ(ui0)\lambda(u_{j})=\lambda(u_{i_{0}}) for every ji0j\geq i_{0}, so minui=(λ(ui))ω\min_{u_{i}}=(\lambda(u_{i}))^{\omega} for every ii0i\geq i_{0}, (uj)ji(u_{j})_{j\geq i} being an occurrence of minui\min_{u_{i}} starting at uiu_{i}. By Lemma 3.2, we conclude τ(ui)=2\tau(u_{i})=2 for every ii0i\geq i_{0}.

Statement of Lemma 4.17. Let G=(V,E)G=(V,E) be a graph. Let uVu\in V such that τ(u)=3\tau(u)=3. Let (ui)i1(u_{i})_{i\geq 1} be an occurrence of minu\min_{u} starting at uu. Let (ui)i1(u^{\prime}_{i})_{i\geq 1} be the infinite sequence of nodes in VV obtained as follows. Consider L={k1|τ(uk)=3}L=\{k\geq 1\;|\;\tau(u_{k})=3\}, and for every i1i\geq 1, let ji1j_{i}\geq 1 be the ithi^{\text{th}} smallest element of LL, if it exists. For every i1i\geq 1 such that jij_{i} is defined, let ui=ujiu^{\prime}_{i}=u_{j_{i}}, and if i1i\geq 1 is such that jij_{i} is not defined (so LL is a finite set), let ui=u|L|u^{\prime}_{i}=u^{\prime}_{|L|}. Then, (u¯i)i1(\bar{u}^{\prime}_{i})_{i\geq 1} is an occurrence of minu¯\min_{\bar{u}} starting at u¯\bar{u} in G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}).

Proof B.14.

First, notice that u¯iV¯\bar{u}^{\prime}_{i}\in\bar{V} for every i1i\geq 1 because τ(ui)=3\tau(u^{\prime}_{i})=3, and u1=u1=uu^{\prime}_{1}=u_{1}=u (so u¯1=u¯1=u¯\bar{u}^{\prime}_{1}=\bar{u}_{1}=\bar{u}). Now, let us prove that (u¯i+1,u¯i)E¯(\bar{u}^{\prime}_{i+1},\bar{u}^{\prime}_{i})\in\bar{E} for every i1i\geq 1. Fix i1i\geq 1. We distinguish three cases.

  1. 1.

    jij_{i} and ji+1j_{i+1} are both defined. This means that τ(ui)=τ(ui+1)=3\tau(u^{\prime}_{i})=\tau(u^{\prime}_{i+1})=3, and τ(uk)3\tau(u_{k})\not=3 for every ji<k<ji+1j_{i}<k<j_{i+1}. By Lemma 4.16, it must be τ(uk)=1\tau(u_{k})=1 for every ji<k<ji+1j_{i}<k<j_{i+1}. Since (uk)kji(u_{k})_{k\geq j_{i}} is an occurrence of minui\min_{u^{\prime}_{i}} starting at uiu^{\prime}_{i}, by Lemma 4.13 we obtain 𝚝ui=3\mathtt{t}_{u^{\prime}_{i}}=3 and ui+1Gui(ui)u^{\prime}_{i+1}\in G_{\ell_{u^{\prime}_{i}}}(u^{\prime}_{i}), so (u¯i+1,u¯i)E¯(\bar{u}^{\prime}_{i+1},\bar{u}^{\prime}_{i})\in\bar{E}.

  2. 2.

    jij_{i} is defined and ji+1j_{i+1} is not defined. This means that i=|L|i=|L|, τ(uj|L|)=3\tau(u_{j_{|L|}})=3, and by Lemma 4.16 there exists h>j|L|h>j_{|L|} such that τ(uh)=2\tau(u_{h})=2 and τ(uk)=1\tau(u_{k})=1 for every j|L|<k<hj_{|L|}<k<h. Since (uk)kj|L|(u_{k})_{k\geq j_{|L|}} is an occurrence of minu|L|\min_{u^{\prime}_{|L|}} starting at u|L|u^{\prime}_{|L|}, by Lemma 4.13 we obtain 𝚝u|L|=2\mathtt{t}_{u^{\prime}_{|L|}}=2, so (u¯i+1,u¯i)=(u¯|L|,u¯|L|)E¯(\bar{u}^{\prime}_{i+1},\bar{u}^{\prime}_{i})=(\bar{u}^{\prime}_{|L|},\bar{u}^{\prime}_{|L|})\in\bar{E}.

  3. 3.

    jij_{i} and ji+1j_{i+1} are both non-defined. Then, by the previous case we conclude (u¯i+1,u¯i)=(u¯|L|,u¯|L|)E¯(\bar{u}^{\prime}_{i+1},\bar{u}^{\prime}_{i})=(\bar{u}^{\prime}_{|L|},\bar{u}^{\prime}_{|L|})\in\bar{E}.

We are left with proving that minu¯[i]=λ(u¯i)\min_{\bar{u}}[i]=\lambda(\bar{u}^{\prime}_{i}) for every i1i\geq 1. Let α(Σ)ω\alpha\in(\Sigma^{\prime})^{\omega} be such that α[i]=λ(u¯i)\alpha[i]=\lambda(\bar{u}^{\prime}_{i}) for every i1i\geq 1. We have to prove that α=minu¯\alpha=\min_{\bar{u}}. Since (u¯i+1,u¯i)E¯(\bar{u}^{\prime}_{i+1},\bar{u}^{\prime}_{i})\in\bar{E} for every i1i\geq 1, we have αIu¯\alpha\in I_{\bar{u}}, and (u¯i)i1(\bar{u}^{\prime}_{i})_{i\geq 1} is an occurrence of α\alpha starting at u¯\bar{u}. The conclusion follows if we prove that for every βIu¯\beta\in I_{\bar{u}} we have αβ\alpha\preceq^{\prime}\beta. Fix βIu¯\beta\in I_{\bar{u}}; it will suffice to show that, for every k1k\geq 1, if α[1,k1]=β[1,k1]\alpha[1,k-1]=\beta[1,k-1], then α[k]β[k]\alpha[k]\preceq^{\prime}\beta[k]. Fix k1k\geq 1, and let (v¯i)i1(\bar{v}_{i})_{i\geq 1} be an occurrence of β\beta starting at u¯\bar{u}. Notice that for every 1hk11\leq h\leq k-1 we have α[h]=β[h]\alpha[h]=\beta[h], or equivalently, λ(u¯h)=λ(v¯h)\lambda(\bar{u}^{\prime}_{h})=\lambda(\bar{v}_{h}), or equivalently, γuh=γvh\gamma_{u^{\prime}_{h}}=\gamma_{v_{h}} (so uh=vh\ell_{u^{\prime}_{h}}=\ell_{v_{h}}) and 𝚝uh=𝚝vh\mathtt{t}_{u^{\prime}_{h}}=\mathtt{t}_{v_{h}}. Note that (v¯i+1,v¯i)E¯(\bar{v}_{i+1},\bar{v}_{i})\in\bar{E} for every i1i\geq 1. We distinguish two cases:

  1. 1.

    There exists 1hk11\leq h\leq k-1 such that 𝚝uh=𝚝vh=2\mathtt{t}_{u^{\prime}_{h}}=\mathtt{t}_{v_{h}}=2. In this case, the definition of uhu^{\prime}_{h} implies uh=uh+1=uh+2=u^{\prime}_{h}=u^{\prime}_{h+1}=u^{\prime}_{h+2}=\dots, and in particular uh=uku^{\prime}_{h}=u^{\prime}_{k}. Moreover, since 𝚝vh=2\mathtt{t}_{v_{h}}=2 and (v¯i+1,v¯i)E¯(\bar{v}_{i+1},\bar{v}_{i})\in\bar{E}, the definition of G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}) implies vh=vh+1=vh+2=v_{h}=v_{h+1}=v_{h+2}=\dots, and in particular vh=vkv_{h}=v_{k}. We conclude α[k]=λ(uk)=λ(uh)=λ(vh)=λ(vk)=β[k]\alpha[k]=\lambda(u^{\prime}_{k})=\lambda(u^{\prime}_{h})=\lambda(v_{h})=\lambda(v_{k})=\beta[k].

  2. 2.

    For every 1hk11\leq h\leq k-1 it holds 𝚝uh=𝚝vh=3\mathtt{t}_{u^{\prime}_{h}}=\mathtt{t}_{v_{h}}=3. In this case, for every 1hk11\leq h\leq k-1 we have 𝚝vh=3\mathtt{t}_{v_{h}}=3 and (v¯h+1,v¯h)E¯(\bar{v}_{h+1},\bar{v}_{h})\in\bar{E}, so vh+1Gvh(vh)v_{h+1}\in G_{\ell_{v_{h}}}(v_{h}) and by Lemma 4.13 there exists an occurrence of minvh[1,vh1]=γvh[1,vh1]=γuh[1,uh1]=minuh[1,uh1]\min_{v_{h}}[1,\ell_{v_{h}}-1]=\gamma_{v_{h}}[1,\ell_{v_{h}}-1]=\gamma_{u^{\prime}_{h}}[1,\ell_{u^{\prime}_{h}}-1]=\min_{u^{\prime}_{h}}[1,\ell_{u^{\prime}_{h}}-1] starting at vhv_{h} and ending at vh+1v_{h+1}. As a consequence, there is an occurrence of minu1[1,u11]minu2[1,u21]minuk1[1,uk11]=minu[1,jk1]\min_{u^{\prime}_{1}}[1,\ell_{u^{\prime}_{1}}-1]\min_{u^{\prime}_{2}}[1,\ell_{u^{\prime}_{2}}-1]\dots\min_{u^{\prime}_{k-1}}[1,\ell_{u^{\prime}_{k-1}}-1]=\min_{u}[1,j_{k}-1] starting at uu and ending at vkv_{k} and so, if β=minu[1,jk1]minvk\beta^{\prime}=\min_{u}[1,j_{k}-1]\min_{v_{k}}, then βIu\beta^{\prime}\in I_{u}. This implies minuβ\min_{u}\preceq\beta^{\prime}, so from minu=minu[1,jk1]minuk\min_{u}=\min_{u}[1,j_{k}-1]\min_{u^{\prime}_{k}} we obtain minukminvk\min_{u^{\prime}_{k}}\preceq\min_{v_{k}}. By Lemma 4.14 and minukminvk\min_{u^{\prime}_{k}}\preceq\min_{v_{k}}, we obtain that one of the following statements must be true:

    1. (a)

      γuk\gamma_{u^{\prime}_{k}} is not a prefix of γvk\gamma_{v_{k}}, γvk\gamma_{v_{k}} is not a prefix of γuk\gamma_{u^{\prime}_{k}} and γukγvk\gamma_{u^{\prime}_{k}}\prec\gamma_{v_{k}}.

    2. (b)

      γuk=γvk\gamma_{u^{\prime}_{k}}=\gamma_{v_{k}} and 𝚝uk𝚝vk\mathtt{t}_{u^{\prime}_{k}}\leq\mathtt{t}_{v_{k}}.

    3. (c)

      γvk\gamma_{v_{k}} is a strict prefix of γuk\gamma_{u^{\prime}_{k}}.

    In all three cases, we conclude (γuk,𝚝uk)(γvk,𝚝vk)(\gamma_{u^{\prime}_{k}},\mathtt{t}_{u^{\prime}_{k}})\preceq^{\prime}(\gamma_{v_{k}},\mathtt{t}_{v_{k}}), or equivalently, λ(uk)λ(vk)\lambda(u^{\prime}_{k})\preceq^{\prime}\lambda(v_{k}), or equivalently, α[k]β[k]\alpha[k]\preceq^{\prime}\beta[k].

Statement of Theorem 4.18. Let G=(V,E)G=(V,E) be a graph. Let u,vVu,v\in V be such that τ(u)=τ(v)=3\tau(u)=\tau(v)=3.

  1. 1.

    If minu=minv\min_{u}=\min_{v}, then minu¯=minv¯\min_{\bar{u}}=\min_{\bar{v}}.

  2. 2.

    If minuminv\min_{u}\prec\min_{v}, then minu¯minv¯\min_{\bar{u}}\prec^{\prime}\min_{\bar{v}}.

Proof B.15.

Let (ui)i1(u_{i})_{i\geq 1} be and occurrence of minu\min_{u} starting at uu, and let (vi)i1(v_{i})_{i\geq 1} be and occurrence of minv\min_{v} starting at vv . Let (u¯i)i1(\bar{u}^{\prime}_{i})_{i\geq 1} be the occurrence of minu¯\min_{\bar{u}} defined by means of (ui)i1(u_{i})_{i\geq 1} in Lemma 4.17, and let (v¯i)i1(\bar{v}^{\prime}_{i})_{i\geq 1} be the occurrence of minv¯\min_{\bar{v}} defined by means of (vi)i1(v_{i})_{i\geq 1} in Lemma 4.17. Let L={i1|τ(ui)=3}L=\{i\geq 1\;|\;\tau(u_{i})=3\}, and let ji1j_{i}\geq 1 be the ithi^{\text{th}} smallest element of LL, if it exists. Moreover, let M={i1|τ(vi)=3}M=\{i\geq 1\;|\;\tau(v_{i})=3\}, and let ki1k_{i}\geq 1 be the ithi^{\text{th}} smallest element of KK, if it exists. Notice that j1=k1=1j_{1}=k_{1}=1.

In the rest of the proof, we say that an integer h1h\geq 1 is nice if it satisfies the following properties:

  • j1,,jhj_{1},\dots,j_{h} and k1,,khk_{1},\dots,k_{h} are all defined, and ji=kij_{i}=k_{i} for every 1ih1\leq i\leq h.

  • γui=γvi\gamma_{u^{\prime}_{i}}=\gamma_{v^{\prime}_{i}} for every 1ih11\leq i\leq h-1.

Note the following properties:

  • 11 is always nice because j1=k1=1j_{1}=k_{1}=1.

  • If hh is nice, than every 1hh1\leq h^{\prime}\leq h is nice. This implies that either every h1h\geq 1 is nice, or there exists a unique h1h^{*}\geq 1 such that every hhh\leq h^{*} is nice and every h>hh>h^{*} is not nice.

  • If hh is nice, γuh=γvh\gamma_{u^{\prime}_{h}}=\gamma_{v^{\prime}_{h}} and 𝚝uh=𝚝vh=3\mathtt{t}_{u^{\prime}_{h}}=\mathtt{t}_{v^{\prime}_{h}}=3, then h+1h+1 is nice. Indeed, γuh=γvh\gamma_{u^{\prime}_{h}}=\gamma_{v^{\prime}_{h}} implies uh=vh\ell_{u^{\prime}_{h}}=\ell_{v^{\prime}_{h}}. Moreover, since (ui)ijh(u_{i})_{i\geq j_{h}} is an occurrence of minuh\min_{u^{\prime}_{h}} starting at uhu^{\prime}_{h}, then by the minimality of uh\ell_{u^{\prime}_{h}} and Lemma 4.13 we have τ(ujh+i)=τ(Gi+1(uh))=1\tau(u_{j_{h}+i})=\tau(G_{i+1}(u^{\prime}_{h}))=1 for every 1iuh21\leq i\leq\ell_{u^{\prime}_{h}}-2, and τ(ujh+uh1)=τ(Guh(uh))=𝚝uh=3\tau(u_{j_{h}+\ell_{u^{\prime}_{h}}-1})=\tau(G_{\ell_{u^{\prime}_{h}}}(u^{\prime}_{h}))=\mathtt{t}_{u^{\prime}_{h}}=3, which implies that jh+1j_{h+1} is defined. Analogously, one obtains τ(ukh+i)=1\tau(u_{k_{h}+i})=1 for every 1ivh21\leq i\leq\ell_{v^{\prime}_{h}}-2 and τ(vkh+vh1)=3\tau(v_{k_{h}+\ell_{v^{\prime}_{h}}-1})=3, so kh+1k_{h+1} is defined and jh+1=kh+1j_{h+1}=k_{h+1}, which implies that h+1h+1 is nice.

  • If hh is nice, then 𝚝ui=𝚝vi=3\mathtt{t}_{u^{\prime}_{i}}=\mathtt{t}_{v^{\prime}_{i}}=3 for every 1ih11\leq i\leq h-1. Indeed, assume for sake of contradiction that 𝚝ul=2\mathtt{t}_{u^{\prime}_{l}}=2 for some 1lh11\leq l\leq h-1 (the case 𝚝vl=2\mathtt{t}_{v^{\prime}_{l}}=2 is analogous). Since (ui)ijl(u_{i})_{i\geq j_{l}} is an occurrence of minul\min_{u^{\prime}_{l}} starting at ulu^{\prime}_{l}, then by Lemma 4.13 we have τ(ujl+i)=τ(Gi+1(ul))=1\tau(u_{j_{l}+i})=\tau(G_{i+1}(u^{\prime}_{l}))=1 for every 1iul21\leq i\leq\ell_{u^{\prime}_{l}}-2, and τ(ujl+ul1)=τ(Gul(ul))=𝚝ul=2\tau(u_{j_{l}+\ell_{u^{\prime}_{l}}-1})=\tau(G_{\ell_{u^{\prime}_{l}}}(u^{\prime}_{l}))=\mathtt{t}_{u^{\prime}_{l}}=2, which by Lemma 4.16 implies τ(ui)=2\tau(u_{i})=2 for every ijl+ul1i\geq j_{l}+\ell_{u^{\prime}_{l}}-1. This implies that jl+1j_{l+1} is not defined, which is a contradiction because 1lh11\leq l\leq h-1.

Let us prove that, if h1h\geq 1 is nice, then:

  • minu[1,jh1]=minv[1,kh1]\min_{u}[1,j_{h}-1]=\min_{v}[1,k_{h}-1].

  • minu¯[1,h1]=minv¯[1,h1]\min_{\bar{u}}[1,h-1]=\min_{\bar{v}}[1,h-1].

We will prove these two properties separately.

  • Let us prove that minu[1,jh1]=minv[1,jh1]\min_{u}[1,j_{h}-1]=\min_{v}[1,j_{h}-1]. Since hh is nice, then for every 1ih1\leq i\leq h we have that jij_{i} and kik_{i} are defined, and ji=kij_{i}=k_{i}. As a consequence, it will suffice to prove that minu[ji,ji+11]=minv[ki,ki+11]\min_{u}[j_{i},j_{i+1}-1]=\min_{v}[k_{i},k_{i+1}-1] for every 1ih11\leq i\leq h-1. This is equivalent to proving that minui[1,ui1]=minvi[1,vi1]\min_{u^{\prime}_{i}}[1,\ell_{u^{\prime}_{i}}-1]=\min_{v^{\prime}_{i}}[1,\ell_{v^{\prime}_{i}}-1] for every 1ih11\leq i\leq h-1. This follows from γui=γvi\gamma_{u^{\prime}_{i}}=\gamma_{v^{\prime}_{i}} for every 1ih11\leq i\leq h-1.

  • Let us prove that minu¯[1,h1]=minv¯[1,h1]\min_{\bar{u}}[1,h-1]=\min_{\bar{v}}[1,h-1]. Fix 1ih11\leq i\leq h-1. We must prove that minu¯[i]=minv¯[i]\min_{\bar{u}}[i]=\min_{\bar{v}}[i], or equivalently, λ(u¯i)=λ(v¯i)\lambda(\bar{u}^{\prime}_{i})=\lambda(\bar{v}^{\prime}_{i}). This follows from γ(ui)=γ(vi)\gamma(u^{\prime}_{i})=\gamma(v^{\prime}_{i}) and 𝚝ui=𝚝vi\mathtt{t}_{u^{\prime}_{i}}=\mathtt{t}_{v^{\prime}_{i}}.

We can now prove the two claims of the theorem. We will use that following observation: if every h1h\geq 1 is nice, then minu=minv\min_{u}=\min_{v} and minu¯=minv¯\min_{\bar{u}}=\min_{\bar{v}}, because minu[1,jh1]=minv[1,kh1]\min_{u}[1,j_{h}-1]=\min_{v}[1,k_{h}-1] and minu¯[1,h1]=minv¯[1,h1]\min_{\bar{u}}[1,h-1]=\min_{\bar{v}}[1,h-1] for every h1h\geq 1.

  1. 1.

    Assume that minu=minv\min_{u}=\min_{v}; we must prove that minu¯=minv¯\min_{\bar{u}}=\min_{\bar{v}}. If every h1h\geq 1 is nice, we are done, so we can assume that h1h^{*}\geq 1 is the largest nice integer. By Lemma 4.9, we have τ(ui)=τ(vi)\tau(u_{i})=\tau(v_{i}) for every i1i\geq 1, so L=ML=M. In particular, for every i1i\geq 1 we have that jij_{i} is defined if and only if kik_{i} is defined and, if so ji=kij_{i}=k_{i}. Since hh^{*} is nice, then jhj_{h^{*}} and khk_{h^{*}} are defined, jh=khj_{h^{*}}=k_{h^{*}} and minu¯[1,h1]=minv¯[1,h1]\min_{\bar{u}}[1,h^{*}-1]=\min_{\bar{v}}[1,h^{*}-1]. We know that minu¯=minu¯[1,h1]minu¯h\min_{\bar{u}}=\min_{\bar{u}}[1,h^{*}-1]\min_{\bar{u}^{\prime}_{h^{*}}} and minv¯=minv¯[1,h1]minv¯h\min_{\bar{v}}=\min_{\bar{v}}[1,h^{*}-1]\min_{\bar{v}^{\prime}_{h^{*}}}, so minu¯[1,h1]=minv¯[1,h1]\min_{\bar{u}}[1,h^{*}-1]=\min_{\bar{v}}[1,h^{*}-1] implies that in order to prove that minu¯=minv¯\min_{\bar{u}}=\min_{\bar{v}} we only have to prove that minu¯h=minv¯h\min_{\bar{u}^{\prime}_{h^{*}}}=\min_{\bar{v}^{\prime}_{h^{*}}}.

    By Lemma 4.9, we have minuh=minvh\min_{u^{\prime}_{h^{*}}}=\min_{v^{\prime}_{h^{*}}}. Since τ(ui)=τ(vi)\tau(u_{i})=\tau(v_{i}) for every i1i\geq 1, (ui)ijh(u_{i})_{i\geq j_{h^{*}}} is an occurrence of minuh\min_{u^{\prime}_{h^{*}}} starting at uhu^{\prime}_{h^{*}} and (vi)ijh(v_{i})_{i\geq j_{h^{*}}} is an occurrence of minvh\min_{v^{\prime}_{h^{*}}} starting at vhv^{\prime}_{h^{*}}, we obtain uh=vh\ell_{u^{\prime}_{h^{*}}}=\ell_{v^{\prime}_{h^{*}}}. As a consequence, γuh=minuh[1,uh]=minvh[1,vh]=γvh\gamma_{u^{\prime}_{h^{*}}}=\min_{u^{\prime}_{h^{*}}}[1,\ell_{u^{\prime}_{h^{*}}}]=\min_{v^{\prime}_{h^{*}}}[1,\ell_{v^{\prime}_{h^{*}}}]=\gamma_{v^{\prime}_{h^{*}}}. In particular, it must be 𝚝uh=𝚝vh\mathtt{t}_{u^{\prime}_{h^{*}}}=\mathtt{t}_{v^{\prime}_{h^{*}}}, so from γuh=γ(vh)\gamma_{u^{\prime}_{h^{*}}}=\gamma(v^{\prime}_{h^{*}}) we conclude λ(u¯h)=λ(v¯h)\lambda(\bar{u}^{\prime}_{h^{*}})=\lambda(\bar{v}^{\prime}_{h^{*}}). Moreover, it must be 𝚝uh=𝚝vh=2\mathtt{t}_{u^{\prime}_{h^{*}}}=\mathtt{t}_{v^{\prime}_{h^{*}}}=2, otherwise we would conclude that h+1h^{*}+1 is nice, which contradicts the maximality of hh^{*}. As a consequence, by the definition of G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}) we conclude that minu¯h=(λ(u¯h))ω=λ(v¯h))ω=minv¯h\min_{\bar{u}^{\prime}_{h^{*}}}=(\lambda(\bar{u}^{\prime}_{h^{*}}))^{\omega}=\lambda(\bar{v}^{\prime}_{h^{*}}))^{\omega}=\min_{\bar{v}^{\prime}_{h^{*}}}.

  2. 2.

    Assume that minuminv\min_{u}\prec\min_{v}; we must prove that minu¯minv¯\min_{\bar{u}}\prec^{\prime}\min_{\bar{v}}. Notice that it cannot happen that every h1h\geq 1 is nice, otherwise we would obtain minu=minv\min_{u}=\min_{v}, a contradiction. Let h1h^{*}\geq 1 be the biggest nice integer. In particular, jhj_{h^{*}} and khk_{h^{*}} are defined, jh=khj_{h^{*}}=k_{h^{*}}, minu[1,jh1]=minv[1,kh1]\min_{u}[1,j_{h^{*}}-1]=\min_{v}[1,k_{h^{*}}-1] and minu¯[1,h1]=minv¯[1,h1]\min_{\bar{u}}[1,h^{*}-1]=\min_{\bar{v}}[1,h^{*}-1]. We know that minu=minu[1,jh1]minuh\min_{u}=\min_{u}[1,j_{h^{*}}-1]\min_{u^{\prime}_{h^{*}}} and minv=minv[1,kh1]minvh\min_{v}=\min_{v}[1,k_{h^{*}}-1]\min_{v^{\prime}_{h^{*}}}, so from minuminv\min_{u}\prec\min_{v} we conclude minuhminvh\min_{u^{\prime}_{h^{*}}}\prec\min_{v^{\prime}_{h^{*}}}. Since minu¯=minu¯[1,h1]minu¯h\min_{\bar{u}}=\min_{\bar{u}}[1,h^{*}-1]\min_{\bar{u}^{\prime}_{h^{*}}}, minv¯=minv¯[1,h1]minv¯h\min_{\bar{v}}=\min_{\bar{v}}[1,h^{*}-1]\min_{\bar{v}^{\prime}_{h^{*}}} and minu¯[1,h1]=minv¯[1,h1]\min_{\bar{u}}[1,h^{*}-1]=\min_{\bar{v}}[1,h^{*}-1], in order to prove that minu¯minv¯\min_{\bar{u}}\prec^{\prime}\min_{\bar{v}} we only have to prove that minu¯hminv¯h\min_{\bar{u}^{\prime}_{h^{*}}}\prec^{\prime}\min_{\bar{v}^{\prime}_{h^{*}}}.

    Notice that it cannot be γuh=γvh\gamma_{u^{\prime}_{h^{*}}}=\gamma_{v^{\prime}_{h^{*}}} and τ(uh)=τ(vh)\tau(u^{\prime}_{h^{*}})=\tau(v^{\prime}_{h^{*}}), because:

    1. (a)

      If τ(uh)=τ(vh)=2\tau(u^{\prime}_{h^{*}})=\tau(v^{\prime}_{h^{*}})=2, then by Lemma 4.15 we would conclude minuh=minvh\min_{u^{\prime}_{h^{*}}}=\min_{v^{\prime}_{h^{*}}}, a contradiction.

    2. (b)

      If τ(uh)=τ(vh)=3\tau(u^{\prime}_{h^{*}})=\tau(v^{\prime}_{h^{*}})=3, then h+1h^{*}+1 would be a nice integer, which contradicts the maximality of hh^{*}.

    From this observation, Lemma 4.14 and minuhminvh\min_{u^{\prime}_{h^{*}}}\prec\min_{v^{\prime}_{h^{*}}} we conclude that one of the following must be true:

    1. (a)

      γuh\gamma_{u^{\prime}_{h^{*}}} is not a prefix of γvh\gamma_{v^{\prime}_{h^{*}}}, γvh\gamma_{v^{\prime}_{h^{*}}} is not a prefix of γuh\gamma_{u^{\prime}_{h^{*}}} and γuhγvh\gamma_{u^{\prime}_{h^{*}}}\prec\gamma_{v^{\prime}_{h^{*}}}.

    2. (b)

      γuh=γvh\gamma_{u^{\prime}_{h^{*}}}=\gamma_{v^{\prime}_{h^{*}}}, 𝚝uh=2\mathtt{t}_{u^{\prime}_{h^{*}}}=2 and 𝚝vh=3\mathtt{t}_{v^{\prime}_{h^{*}}}=3.

    3. (c)

      γuh\gamma_{u^{\prime}_{h^{*}}} is a strict prefix of γvh\gamma_{v^{\prime}_{h^{*}}}.

    In all three cases we conclude (γuh,𝚝uh)(γvh,𝚝vh)(\gamma_{u^{\prime}_{h^{*}}},\mathtt{t}_{u^{\prime}_{h^{*}}})\prec^{\prime}(\gamma_{v^{\prime}_{h^{*}}},\mathtt{t}_{v^{\prime}_{h^{*}}}), or equivalently, λ(u¯h)λ(v¯h)\lambda(\bar{u}^{\prime}_{h^{*}})\prec^{\prime}\lambda(\bar{v}^{\prime}_{h^{*}}), which implies minu¯hminv¯h\min_{\bar{u}^{\prime}_{h^{*}}}\prec^{\prime}\min_{\bar{v}^{\prime}_{h^{*}}}.

Statement of Lemma 4.20. Let G=(V,E)G=(V,E) be a trimmed graph. Then, we can build G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}) in O(|V|2)O(|V|^{2}) time.

Proof B.16.

The definition of G¯\bar{G} implies that in order to build G¯\bar{G} it is sufficient to compute γu\gamma_{u}, 𝚝u\mathtt{t}_{u} and Gu(u)G_{\ell_{u}}(u) for every uVu\in V such that τ(u)=3\tau(u)=3; moreover, we also need to compute the total order \preceq^{\prime}, so that we can remap Σ\Sigma^{\prime} to an integer alphabet consistently with the total order \preceq^{\prime} on Σ\Sigma^{\prime}.

Fix uVu\in V such that τ(u)=3\tau(u)=3. For every 1iu11\leq i\leq\ell_{u}-1, let Ei(u)E_{i}(u) be the set of all edges entering a node in Gi(u)G_{i}(u), and let mi(u)=|Ei(i)|m_{i}(u)=|E_{i}(i)|. Note that since the Gi(u)G_{i}(u)’s are pairwise distinct, then the Eu(i)E_{u}(i)’s are pairwise disjoint. Let us prove that i=1u1mi(u)O(n)\sum_{i=1}^{\ell_{u}-1}m_{i}(u)\in O(n). To this end, it will suffice to prove that for every vVv\in V there exist at most two edges in i=1u1Ei(u)\bigcup_{i=1}^{\ell_{u}-1}E_{i}(u) whose start nodes are equal to vv. Fix vVv\in V, and let 1iu11\leq i\leq\ell_{u}-1 be the smallest integer such that vv is the first state of an edge (v,v)Ei(u)(v,v^{\prime})\in E_{i}(u) for some vGi(u)v^{\prime}\in G_{i}(u), if such a vv^{\prime} exists. Notice that vv^{\prime} is uniquely determined because the value λ(z)\lambda(z) does not depend on the choice of zGi(u)z\in G_{i}(u), and GG is deterministic. This means that (v,v)Ei(u)(v,v^{\prime})\in E_{i}(u) is uniquely determined. In order to prove our claim, it will suffice to show that if (v,v)Ej(u)(v,v^{\prime\prime})\in E_{j}(u) for some i+1ju1i+1\leq j\leq\ell_{u}-1 and vGj(u)v^{\prime\prime}\in G_{j}(u), then λ(v)=λ(v)\lambda(v^{\prime\prime})=\lambda(v), because we will conclude that jj and vv^{\prime\prime} are uniquely determined, since GG is deterministic. Since vGi(u)v^{\prime}\in G_{i}(u) and (v,v)Ei(u)(v,v^{\prime})\in E_{i}(u), by the definition of Gi+1(u)G_{i+1}(u), we have λ(Gi+1(u))λ(v)\lambda(G_{i+1}(u))\preceq\lambda(v). Since 2i+1ju12\leq i+1\leq j\leq\ell_{u}-1, then by Lemma 4.12 we obtain minu[j]minu[i+1]\min_{u}[j]\preceq\min_{u}[i+1], which by Lemma 4.13 is equivalent to λ(Gj(u))λ(Gi+1(u))\lambda(G_{j}(u))\preceq\lambda(G_{i+1}(u)), and so we conclude λ(Gj(u))λ(v)\lambda(G_{j}(u))\preceq\lambda(v), or equivalently, λ(v)λ(v)\lambda(v^{\prime\prime})\preceq\lambda(v). Since vGj(u)v^{\prime\prime}\in G_{j}(u) and 2ju12\leq j\leq\ell_{u}-1, then τ(v)=τ(Gj(u))=1\tau(v^{\prime\prime})=\tau(G_{j}(u))=1, so from (v,v)E(v,v^{\prime\prime})\in E we conclude that it must be λ(v)λ(v)\lambda(v)\preceq\lambda(v^{\prime\prime}) because GG is trimmed, and so λ(v)=λ(v)\lambda(v^{\prime\prime})=\lambda(v).

Let us show that in O(|V|2)O(|V|^{2}) we can compute Gu(u)G_{\ell_{u}}(u), γu\gamma_{u} and 𝚝u\mathtt{t}_{u} for every uVu\in V such that τ(u)=3\tau(u)=3. It will suffice to show that, for a a fixed uVu\in V such that τ(u)=3\tau(u)=3, we can compute Gu(u)G_{\ell_{u}}(u), γu\gamma_{u} and 𝚝u\mathtt{t}_{u} in O(|V|)O(|V|) time.

Let us show how we recursively compute each set Gi(u)G_{i}(u), for 1iu1\leq i\leq\ell_{u}. During the algorithm, we will mark some nodes. Initially, no node is marked, and after step i1i\geq 1, a nodes of VV is marked if and only if it belongs to j=2i1Gj(u)\bigcup_{j=2}^{i-1}G_{j}(u). If i=1i=1, we just let G1(u)={u}G_{1}(u)=\{u\}, and we define c1=λ(u)c_{1}=\lambda(u). Now, assume that i2i\geq 2, and assume that we have computed Gi1(u)G_{i-1}(u). We first scan all edges in Ei1(u)E_{i-1}(u) and we collect the set RiR_{i} the set of all start nodes of these edges. Then, by scanning RiR_{i}, we first determine ci=min{λ(v)|vRi}c_{i}=\min\{\lambda(v)\;|\;v\in R_{i}\}, then we determine Ri={vRi|λ(v)=ci}R^{\prime}_{i}=\{v\in R_{i}\;|\;\lambda(v)=c_{i}\}. Next, we determine ti=min{τ(v)|vRi}t_{i}=\min\{\tau(v)\;|\;v\in R^{\prime}_{i}\} and Ri={vRi|λ(v)=ti}R^{\prime\prime}_{i}=\{v\in R^{\prime}_{i}\;|\;\lambda(v)=t_{i}\}. By checking which nodes in RiR^{\prime\prime}_{i} are marked, we determine Gi(u)G_{i}(u). Finally, we mark all nodes in Gi(u)G_{i}(u). We can perform all these steps in O(mi1(u))O(m_{i-1}(u)) time, where the hidden constant in the asymptotic notation does not depend on ii. Now, we check whether ti2t_{i}\geq 2. If ti2t_{i}\geq 2, then i=ui=\ell_{u} and we are done. Otherwise, we have i<ui<\ell_{u} and we proceed with step i+1i+1.

We conclude that we can determine Gu(u)G_{\ell_{u}}(u) in time i=1u1O(mi(u))=O(i=1u1mi(u))=O(n)\sum_{i=1}^{\ell_{u}-1}O(m_{i}(u))=O(\sum_{i=1}^{\ell_{u}-1}m_{i}(u))=O(n), where we have used that the hidden constant in O(mi(u))O(m_{i}(u)) does not depend on ii. In addition, we have γu=c1c2cu\gamma_{u}=c_{1}c_{2}\dots c_{\ell_{u}} and 𝚝u=tu\mathtt{t}_{u}=t_{\ell_{u}}.

We are only left with showing how to compute \preceq^{\prime}. Essentially, we only have to radix sort the strings γ(u)\gamma(u)’s by taking into account that \preceq^{\prime} is defined by considering a slight variation of the lexicographic order. More precisely, we proceed as follows. We know that |γ(u)||V|+1|\gamma(u)|\leq|V|+1 for each uVu\in V such that τ(u)=3\tau(u)=3, so we first pad the end of each γ(u)\gamma(u) with a special character larger than all characters in the alphabet until the length of each string is exactly |V|+1|V|+1. Next, we consider two extra special characters d2d_{2} and d3d_{3} such that d2d3d_{2}\prec^{\prime}d_{3}, and we append exactly one of this character to each γ(u)\gamma(u): we append d2d_{2} if 𝚝u=2\mathtt{t}_{u}=2, and we append d3d_{3} if 𝚝u=3\mathtt{t}_{u}=3. Now, we radix sort the (modified) γ(u)\gamma(u)’s in O(|V|2)O(|V|^{2}) time, so obtaining \preceq^{\prime}.

B.3 Proofs from Section 4.3

Statement of Lemma 4.21. Let G=(V,E)G=(V,E) be a graph. If we know the min-partition of {uV|τ(u)=3}\{u\in V\;|\;\tau(u)=3\}, then we can build the min-partition of VV in O(|E|)O(|E|) time.

Proof B.17.

We have seen that the algorithm correctly builds 𝒜c,3\mathcal{A}_{c,3} and 𝒜c,2\mathcal{A}_{c,2} for every cΣc\in\Sigma. Note that when the algorithm builds the 𝒜c,1\mathcal{A}_{c,1}’s, it never modifies the Ac,2A_{c,2}’s and the Ac,3A_{c,3}’s (because we only add elements to the Ac,1A_{c,1}’s).

Let us prove that, when we consider II in Aci,tA_{c_{i},t} (and before computing the JkJ_{k}’s), then:

  1. 1.

    II is an element of 𝒜\mathcal{A} and we have already built the prefix of 𝒜\mathcal{A} whose largest element is II.

  2. 2.

    every Ac,1A_{c,1} contains a prefix of 𝒜c,1\mathcal{A}_{c,1}.

At the beginning of the algorithm we consider II in Ac1,tA_{c_{1},t}, with t{2,3}t\in\{2,3\}, so our claim is true because we have already built 𝒜c1,2\mathcal{A}_{c_{1},2} and 𝒜c1,3\mathcal{A}_{c_{1},3}, and all the Ac,1A_{c,1}’s are empty.

Now, assume that our claim is true when we consider I𝒜ci,tI\in\mathcal{A}_{c_{i},t}. We want to prove that it is true when we process the next element (if it exists). When we consider II, we compute the nonempty JkJ_{k}’s, and we add each such JkJ_{k} to Ack,1A_{c_{k},1}. We now want to prove that JkJ_{k} is correctly identified as the next element in Ack,1A_{c_{k},1}.

  • First, let us prove that, if v1,v2Jkv_{1},v_{2}\in J_{k}, then minv1=minv2\min_{v_{1}}=\min_{v_{2}}. Since v1,v2Jkv_{1},v_{2}\in J_{k}, then there exist u1,u2Iu_{1},u_{2}\in I such that (u1,v1),(u2,v2)E(u_{1},v_{1}),(u_{2},v_{2})\in E. Since by the inductive hypothesis II is an element of 𝒜\mathcal{A} and we have already built the prefix of 𝒜\mathcal{A} whose largest element is II and since v1v_{1} and v2v_{2} were not marked before, then it must be minv1=ckminu1\min_{v_{1}}=c_{k}\min_{u_{1}}, minv2=ckminu2\min_{v_{2}}=c_{k}\min_{u_{2}} and minu1=minu2\min_{u_{1}}=\min_{u_{2}}, so we conclude minv1=minv2\min_{v_{1}}=\min_{v_{2}}.

  • We are only left with showing that if v1Jkv_{1}\in J_{k} and v2Vck,1v_{2}\in V_{c_{k},1} is not in some element of Ack,1A_{c_{k},1} after JkJ_{k} is added to Ack,1A_{c_{k},1}, then minv1minv2\min_{v_{1}}\prec\min_{v_{2}}. Let u1Iu_{1}\in I such that (u1,u2)E(u_{1},u_{2})\in E; we have shown that minv1=ckminu1\min_{v_{1}}=c_{k}\min_{u_{1}}. The definition of v2v_{2} implies that minv2=ckminu2\min_{v_{2}}=c_{k}\min_{u_{2}} for some u2u_{2} that we have not processed yet. Since by the inductive hypothesis II is an element of 𝒜\mathcal{A} and we have already built the prefix of 𝒜\mathcal{A} whose largest element is II, then it must be minu1minu2\min_{u_{1}}\prec\min_{u_{2}}, and we conclude minv1minv2\min_{v_{1}}\prec\min_{v_{2}}.

Now, let us prove that, if after considering II in Aci,tA_{c_{i},t}, the next element to be considered is not in Aci,tA_{c_{i},t}, then the construction of 𝒜c1,t\mathcal{A}_{c_{1},t} is complete. If t{2,3}t\in\{2,3\} then the conclusion is immediate because 𝒜ci,2\mathcal{A}_{c_{i},2} and 𝒜ci,3\mathcal{A}_{c_{i},3} had already been fully built. Now assume that t=1t=1. Suppose for sake of contradiction that there exists vVci,1v\in V_{c_{i},1} which has not been added to some element in Aci,1A_{c_{i},1}. Without loss of generality, we choose vv such that minv\min_{v} is as small as possible. Since τ(v)=1\tau(v)=1, then there exists uVu\in V such that (u,v)E(u,v)\in E and minuminv\min_{u}\prec\min_{v} (and so λ(u)λ(v)=ci\lambda(u)\preceq\lambda(v)=c_{i}). If λ(u)ci\lambda(u)\prec c_{i}, then we immediately obtain that uu has already been processed (because we have already built the prefix of 𝒜\mathcal{A} whose largest element is II) and so vv should have already been processed when uu was processed, a contradiction. If λ(u)=ci\lambda(u)=c_{i}, then by Corollary 3.3 we have τ(u)τ(v)\tau(u)\leq\tau(v), so τ(u)=1\tau(u)=1 and uVci,1u\in V_{c_{i},1}; the minimality of vv implies that uu was previously added to some element of Aci,1A_{c_{i},1} and so vv should have been processed, again a contradiction.

As a consequence, when we process the next element after II in Aci,tA_{c_{i},t}, then our claim will be true (independently of whether the next element is in Aci,tA_{c_{i},t} or not). In particular, if there is no next element, we conclude that all 𝒜c,1\mathcal{A}_{c,1}’s have been built,

Lastly, we can build each 𝒜c,1\mathcal{A}_{c,1} in O(|E|)O(|E|) time, because we only need to scan each node and each edge once.

Appendix C Details on the Complementary Case

Let us show how to reduce the problem of computing the min-partition of {uV|τ(u)=1}\{u\in V\;|\;\tau(u)=1\} to the problem of determining the min-partition of V¯\bar{V}, where G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}) is a graph such that |V¯|={uV|τ(u)=1}|\bar{V}|=\{u\in V\;|\;\tau(u)=1\}. We use the same notation that we used in the previous sections (with a complementary meaning) so that it will easy to follow our argument.

C.1 Recursive Step

For every uVu\in V such that τ(u)=1\tau(u)=1, let u\ell_{u} be the smallest integer k2k\geq 2 such that τ(uk)2\tau(u_{k})\leq 2, where (ui)i1(u_{i})_{i\geq 1} is an occurrence of minu\min_{u} starting at uu. For every 1iu1\leq i\leq\ell_{u}, we can define each Gi(u)G_{i}(u) as we did before; now, it will be τ(Gi(u))=3\tau(G_{i}(u))=3 for every 2iu12\leq i\leq\ell_{u}-1. In fact, when we explore the graph starting from uu in a backward fashion and we (implicitly) build a prefix of minu\min_{u}, at each step it is still true that we must consider the nodes with minimum value λ(v)\lambda(v) and, among those, the nodes with minimum value τ(v)\tau(v). This time, there is no chance that we include in Gi(u)G_{i}(u) a node already included before: at every step, λ(Gi(u))\lambda(G_{i}(u)) can only increase (because τ(Gi(u))=3\tau(G_{i}(u))=3), so if there were a cycle, then all edges of the cycles would have the same label and by Lemma 4.6 we would conclude τ(Gi(u))2\tau(G_{i}(u))\leq 2 for some 2iu12\leq i\leq\ell_{u}-1, a contradiction. In particular, this means that now we do not even need to somehow trim the graph to ensure that our algorithm runs in O(|V|2)O(|V|^{2}) time, because we cannot visit the same node twice. The γu\gamma_{u}’s and the 𝚝u\mathtt{t}_{u}’s are defined as before, and Lemma 4.14 still implies that Σ\Sigma^{\prime} and \preceq^{\prime} must be defined as before. When we define G¯=(V¯,E¯)\bar{G}=(\bar{V},\bar{E}), we define V¯={u¯|τ(u)=1}\bar{V}=\{\bar{u}\;|\;\tau(u)=1\} and we define E¯\bar{E} as before. It is easy to check that we can build G¯\bar{G} in O(|V|2)O(|V|^{2}) time as before, and if we have the min-partition of V¯\bar{V} (with respect to G¯\bar{G}), then we also have the min-partition of {uV|τ(u)=1}\{u\in V\;|\;\tau(u)=1\} (with respect to GG).

C.2 Merging

Let uVu\in V such that τ(u)=3\tau(u)=3. By Lemma 3.2, there exist k1k\geq 1, cΣc\in\Sigma and γΣω\gamma^{\prime}\in\Sigma^{\omega} such that minu=λ(u)kcγ\min_{u}=\lambda(u)^{k}c\gamma^{\prime} and λ(u)c\lambda(u)\prec c. Then, we define ψ(u)=k\psi(u)=k.

In the following, we will often use the following observation. Let vVv\in V such that τ(v)=3\tau(v)=3. Let uVu\in V such that (u,v)E(u,v)\in E and minv=λ(v)minu\min_{v}=\lambda(v)\min_{u}. Then, (i) λ(v)λ(u)\lambda(v)\preceq\lambda(u) and (ii) if λ(v)=λ(u)\lambda(v)=\lambda(u), then τ(u)=3\tau(u)=3 and ψ(v)=ψ(u)+1\psi(v)=\psi(u)+1.

Note that, if u,vVu,v\in V are such that τ(u)=τ(v)=3\tau(u)=\tau(v)=3, λ(u)=λ(v)\lambda(u)=\lambda(v) and ψ(u)ψ(v)\psi(u)\not=\psi(v), then minuminv\min_{u}\prec\min_{v} if and only if ψ(v)<ψ(u)\psi(v)<\psi(u).

It is easy to compute all ψ(u)\psi(u)’s in O(|E|)O(|E|) time. Start from each uVu\in V such that τ(u)=3\tau(u)=3, and explore the graph in a backward fashion by only following edges labeled λ(u)\lambda(u), until either we encounter a node for which we have already determined ψ(u)\psi(u), or we cannot explore the graph any longer. For all the nodes uu^{\prime} that we encounter it must be τ(u)=3\tau(u^{\prime})=3 (otherwise it would be τ(u)3\tau(u)\not=3), and if we cannot explore any longer from some uu^{\prime} that we encounter, it must be ψ(u)=1\psi(u^{\prime})=1. We cannot encounter the same node twice because GG is deterministic, and we cannot have cycles otherwise it would be τ(u)3\tau(u)\not=3 by Lemma 4.6. As a consequence, we have built a tree (where the root uu has no outgoing edges), and computing the ψ(u)\psi(u)’s is equivalent to computing the height of each node.

Let us show how we can use the ψ(u)\psi(u)’s to determine the min-partition 𝒜\mathcal{A} of VV, assuming that we already have the min-partition \mathcal{B} of {uV|τ(u)=1}\{u\in V\;|\;\tau(u)=1\}. As before we have the min-partition \mathcal{B}^{\prime} of {u𝒱|τ(u)=2}\{u\in\mathcal{V}\;|\;\tau(u)=2\}. For every cΣc\in\Sigma and for every t{1,2,3}t\in\{1,2,3\}, we define Vc,tV_{c,t}, 𝒜c,t\mathcal{A}_{c,t} and Ac,tA_{c,t} as before, and our problem reduces to compute each 𝒜c,t\mathcal{A}_{c,t}.

By using \mathcal{B} and \mathcal{B}^{\prime}, we can compute the 𝒜c,1\mathcal{A}_{c,1}’s and the 𝒜c,2\mathcal{A}_{c,2}’s as before, so the challenging part is to compute the 𝒜c,3\mathcal{A}_{c,3}’s. Let Σ={c1,c2,,cσ}\Sigma=\{c_{1},c_{2},\dots,c_{\sigma}\}, with c1c2cσc_{1}\prec c_{2}\prec\dots\prec c_{\sigma}. During the algorithm, we will assign a number ψ(I)1\psi(I)\geq 1 to every element II being in some Ac,3A_{c,3} at some point.

Notice that 𝒜cσ,3\mathcal{A}_{c_{\sigma},3} must be empty because otherwise we would conclude that cσc_{\sigma} is not largest character. This suggest that this time the order in which we process the 𝒜c,t\mathcal{A}_{c,t}’s must be 𝒜cσ,2\mathcal{A}_{c_{\sigma},2}, 𝒜cσ,1\mathcal{A}_{c_{\sigma},1}, 𝒜cσ1,3\mathcal{A}_{c_{\sigma-1},3}, 𝒜cσ1,2\mathcal{A}_{c_{\sigma-1},2}, 𝒜cσ1,1\mathcal{A}_{c_{\sigma-1},1}, 𝒜cσ2,3\mathcal{A}_{c_{\sigma-2},3} and so on. Moreover, we will build each 𝒜c,t\mathcal{A}_{c,t} incrementally from its largest element to its smallest element (so we will consider suffixes of min-partitions, not prefixes). This time we will not mark nodes, but we will mark entries of the Ac,tA_{c,t}’s to indicate that a node has been removed from an element in Ac,tA_{c,t}. Intuitively, this time we need to remove nodes because it will be true that when we process II in Aci,tA_{c_{i},t} then we have already built the suffix (not the prefix) of 𝒜\mathcal{A} whose largest element is II, but we are now building 𝒜\mathcal{A} from its largest element to its smallest element, so a node reached by an edge leaving a node uu in II may also be reached by an edge leaving a node uu^{\prime} that we have not processed, and for which it holds minuminu\min_{u^{\prime}}\prec\min_{u}.

Assume that we process II in Aci,tA_{c_{i},t}. Note that if vVv\in V is such that τ(v)=3\tau(v)=3, λ(v)=ck\lambda(v)=c_{k}, and (u,v)E(u,v)\in E for some uIu\in I, then it must be kik\leq i otherwise τ(v)=1\tau(v)=1; if II is an element in the 𝒜c,1\mathcal{A}_{c,1}’s or in the 𝒜c,2\mathcal{A}_{c,2}’s, it must always be k<ik<i because, if it were k=ik=i, then we would conclude τ(v)3\tau(v)\not=3. For every kik\leq i, define ψI,k=1\psi_{I,k}=1 if k<ik<i, and ψI,k=ψ(I)+1\psi_{I,k}=\psi(I)+1 if k=ik=i. We compute the set JkJ_{k} of all nodes vVv\in V such that τ(v)=3\tau(v)=3, λ(v)=ck\lambda(v)=c_{k}, ψ(v)=ψI,k\psi(v)=\psi_{I,k} and (u,v)E(u,v)\in E for some uIu\in I and, if JkJ_{k}\not=\emptyset, then (i) we mark the entries of Ack,3A_{c_{k},3} containing an element in JkJ_{k} and (ii) we add JkJ_{k} to Ack,3A_{c_{k},3}, letting ψ(Jk)=ψI,k\psi(J_{k})=\psi_{I,k}. We assume that we maintain an additional array that for every node in {u𝒱|τ(u)=3}\{u\in\mathcal{V}\;|\;\tau(u)=3\} already occurring in some Ac,3A_{c,3} stores its (unique) current position, so that operation (i) can be performed in constant time without affecting the running time of the algorithm (which is still O(|E|)O(|E|)).

Notice that at any time the following will be true in each Ac,3A_{c,3}: (a) if KK is in Ac,3A_{c,3}, then ψ(K)1\psi(K)\geq 1; (b) if Ac,3A_{c,3} is nonempty, then the first KK that we have added is such that ψ(K)=1\psi(K)=1; (c) If KK^{\prime} has been added to Ac,3A_{c,3} immediately after KK, then ψ(K)ψ(K)ψ(K)+1\psi(K)\leq\psi(K^{\prime})\leq\psi(K)+1; (d) If vKv\in K for some KK in Ac,3A_{c,3}, then ψ(v)=ψ(K)\psi(v)=\psi(K).

Let us prove that, when we consider II in Aci,tA_{c_{i},t} (and before computing the JkJ_{k}’s), then II is an element of 𝒜\mathcal{A} and we have already built the suffix of 𝒜\mathcal{A} whose largest element is II.

At the beginning of the algorithm we consider II in Acσ,tA_{c_{\sigma},t}, with t{1,2}t\in\{1,2\}, so our claim is true because we have already built 𝒜cσ,2\mathcal{A}_{c_{\sigma},2} and 𝒜cσ,1\mathcal{A}_{c_{\sigma},1}.

Now, assume that our claim is true when we consider II in Aci,tA_{c_{i},t}. Let us prove that we can consider the next element II^{\prime} in Aci,tA_{c_{i}^{\prime},t^{\prime}} (if it exists), then II^{\prime} is an element of 𝒜\mathcal{A} and we have already built the suffix of 𝒜\mathcal{A} whose largest element is II^{\prime}. Notice that either i=ii^{\prime}=i or i=i1i^{\prime}=i-1.

  1. 1.

    Assume that i=ii^{\prime}=i and t=tt^{\prime}=t. If t=t3t^{\prime}=t\not=3 we are done because we have already built the 𝒜c,2\mathcal{A}_{c,2}’s and the 𝒜c,1\mathcal{A}_{c,1}’s. Now assume that t=t=3t^{\prime}=t=3. This implies that ψ(I)+1ψ(I)ψ(I)\psi(I)+1\leq\psi(I^{\prime})\leq\psi(I).

    1. (a)

      Suppose that ψ(I)=ψ(I)\psi(I^{\prime})=\psi(I). First, let us prove that, if v1,v2Iv_{1},v_{2}\in I^{\prime}, then minv1=minv2\min_{v_{1}}=\min_{v_{2}}. Let u1,u2Vu_{1},u_{2}\in V be such that (u1,v1),(u2,v2)E(u_{1},v_{1}),(u_{2},v_{2})\in E, minv1=ciminu1\min_{v_{1}}=c_{i}\min_{u_{1}} and minv2=ciminu2\min_{v_{2}}=c_{i}\min_{u_{2}}. Since τ(v)=3\tau(v)=3 we obtain that either (i) ciλ(u1)c_{i}\prec\lambda(u_{1}), or (ii) λ(u1)=ci\lambda(u_{1})=c_{i}, τ(u1)=3\tau(u_{1})=3 and ψ(u1)=ψ(I)1=ψ(I)1\psi(u_{1})=\psi(I^{\prime})-1=\psi(I)-1. In both cases, we conclude that u1u_{1} is an element KK of the suffix of 𝒜\mathcal{A} whose largest element is II, which by the inductive hypothesis has been correctly built, and so v1v_{1} is added to an element of Aci,tA_{c_{i},t}; v1v_{1} is never removed from this element otherwise in Iv1I_{v_{1}} there would be a string smaller that minv1\min_{v_{1}}. As a consequence, it must also be u2Ku_{2}\in K, so by the inductive hypothesis minu1=minu2\min_{u_{1}}=\min_{u_{2}} and we conclude minv1=minv2\min_{v_{1}}=\min_{v_{2}}.

      We are only left with proving that, if v1v_{1} is neither in II^{\prime}, nor in the suffix of 𝒜\mathcal{A} whose largest element is II, and if v2Iv_{2}\in I^{\prime}, then minv1minv2\min_{v_{1}}\prec\min_{v_{2}}. The conclusion is immediate if τ(vi)3\tau(v_{i})\not=3, so we can assume τ(vi)=3\tau(v_{i})=3. It cannot be ci=λ(v2)λ(v1)c_{i}=\lambda(v_{2})\prec\lambda(v_{1}) otherwise v1v_{1} would be in the suffix of 𝒜\mathcal{A} whose largest element is II. Since τ(vi)=3\tau(v_{i})=3, it cannot be λ(v1)=ci\lambda(v_{1})=c_{i} and ψ(v1)<ψ(I)\psi(v_{1})<\psi(I), otherwise again v1v_{1} would be in the suffix of 𝒜\mathcal{A} whose largest element is II. As a consequence, it must λ(v1)ci\lambda(v_{1})\preceq c_{i} and, if λ(v1)=ci\lambda(v_{1})=c_{i}, then ψ(I)=ψ(I)ψ(v1)\psi(I^{\prime})=\psi(I)\leq\psi(v_{1}). If λ(v1)ci\lambda(v_{1})\prec c_{i}, or λ(v1)=ci\lambda(v_{1})=c_{i} and ψ(I)<ψ(v1)\psi(I^{\prime})<\psi(v_{1}), we immediately conclude minv1minv2\min_{v_{1}}\prec\min_{v_{2}}. Hence, we can assume λ(v1)=ci\lambda(v_{1})=c_{i}, τ(v1)=3\tau(v_{1})=3 and ψ(v1)=ψ(I)\psi(v_{1})=\psi(I^{\prime}). Let u1,u2Vu_{1},u_{2}\in V be such that (u1,v1),(u2,v2)E(u_{1},v_{1}),(u_{2},v_{2})\in E, minv1=ciminu1\min_{v_{1}}=c_{i}\min_{u_{1}} and minv2=ciminu2\min_{v_{2}}=c_{i}\min_{u_{2}}. As before, we must have already considered u1u_{1} and u2u_{2}, but since v1v_{1} is not in II^{\prime}, by construction it must be minu1minu2\min_{u_{1}}\prec\min_{u_{2}} and so we conclude minv1minv2\min_{v_{1}}\prec\min_{v_{2}}.

    2. (b)

      Suppose that ψ(I)=ψ(I)+1\psi(I^{\prime})=\psi(I)+1. Arguing as we did above we infer that all vVci,3v\in V_{c_{i},3} such that ψ(v)=ψ(I)\psi(v)=\psi(I) are already in some element of the suffix of 𝒜\mathcal{A} whose largest element is II. By using this information, the same proof of the previous case shows that II^{\prime} is correct.

  2. 2.

    Assume that i=ii^{\prime}=i and t2t^{\prime}\leq 2. Since we have already built the 𝒜c,2\mathcal{A}_{c,2}’s and the 𝒜c,1\mathcal{A}_{c,1}’s, we only have to prove that, if t=3t=3, then all vVci,3v\in V_{c_{i},3} are already in some element of the suffix of 𝒜\mathcal{A} whose largest element is II. Assume for sake of contradiction that this is not true for some vVci,3v\in V_{c_{i},3}, and choose vv such that ψ(v)\psi(v) is as small as possible. By arguing as before, we conclude (by the minimality of ψ(v)\psi(v)) that vv must be in some Aci,3A_{c_{i},3}; but since t2t^{\prime}\leq 2, we conclude that vv must be in some element of the suffix of 𝒜\mathcal{A} whose largest element is II, a contradiction.

  3. 3.

    Assume that i=i1i^{\prime}=i-1 and t=3t^{\prime}=3. In particular, ψ(I)=1\psi(I^{\prime})=1. As in the previous case, we obtain that if t=3t=3, then all vVci,3v\in V_{c_{i},3} are already in some element of the suffix of 𝒜\mathcal{A} whose largest element is II. Moreover, 𝒜ci,2\mathcal{A}_{c_{i},2} and 𝒜ci,1\mathcal{A}_{c_{i},1} must necessarily empty because we have already built them, and II^{\prime} is not contained in any of them. First, let us prove that, if v1,v2Iv_{1},v_{2}\in I^{\prime}, then minv1=minv2\min_{v_{1}}=\min_{v_{2}}. Let u1,u2Vu_{1},u_{2}\in V be such that (u1,v1),(u2,v2)E(u_{1},v_{1}),(u_{2},v_{2})\in E, minv1=ci1minu1\min_{v_{1}}=c_{i-1}\min_{u_{1}} and minv2=ci1minu2\min_{v_{2}}=c_{i-1}\min_{u_{2}}. Since τ(v)=3\tau(v)=3 we obtain that either (i) ci1λ(u1)c_{i-1}\prec\lambda(u_{1}), or (ii) λ(u1)=ci1\lambda(u_{1})=c_{i-1}, τ(u1)=3\tau(u_{1})=3 and ψ(u1)=ψ(I)1\psi(u_{1})=\psi(I^{\prime})-1. However, case (ii) cannot occur because ψ(I)=1\psi(I^{\prime})=1, so it must be ci1λ(u1)c_{i-1}\prec\lambda(u_{1}), and we conclude that u1u_{1} is an element KK of the suffix of 𝒜\mathcal{A} whose largest element is II. As before, we conclude that also u2u_{2} is in KK, minu1=minu2\min_{u_{1}}=\min_{u_{2}} and minv1=minv2\min_{v_{1}}=\min_{v_{2}}.

    We are only left with proving that, if v1v_{1} is neither in II^{\prime}, nor in the suffix of 𝒜\mathcal{A} whose largest element is II, and if v2Iv_{2}\in I^{\prime}, then minv1minv2\min_{v_{1}}\prec\min_{v_{2}}. The conclusion is immediate if τ(vi)3\tau(v_{i})\not=3, so we can assume τ(vi)=3\tau(v_{i})=3. It cannot be ci1=λ(v2)λ(v1)c_{i-1}=\lambda(v_{2})\prec\lambda(v_{1}) otherwise v1v_{1} would be in the suffix of 𝒜\mathcal{A} whose largest element is II. It cannot be λ(v1)=ci1\lambda(v_{1})=c_{i-1} and ψ(v1)<ψ(I)\psi(v_{1})<\psi(I^{\prime}), because ψ(I)=1\psi(I^{\prime})=1. As a consequence, it must hold λ(v1)ci1\lambda(v_{1})\preceq c_{i-1} and, if λ(v1)=ci1\lambda(v_{1})=c_{i-1}, then 1=ψ(I)ψ(v1)1=\psi(I^{\prime})\leq\psi(v_{1}). If λ(v1)ci1\lambda(v_{1})\prec c_{i-1}, or λ(v1)=ci1\lambda(v_{1})=c_{i-1} and 1=ψ(I)<ψ(v1)1=\psi(I^{\prime})<\psi(v_{1}), we immediately conclude minv1minv2\min_{v_{1}}\prec\min_{v_{2}}. Hence, we can assume λ(v1)=ci1\lambda(v_{1})=c_{i-1}, τ(v1)=3\tau(v_{1})=3 and ψ(v1)=ψ(I)=1\psi(v_{1})=\psi(I^{\prime})=1. Let u1,u2Vu_{1},u_{2}\in V be such that (u1,v1),(u2,v2)E(u_{1},v_{1}),(u_{2},v_{2})\in E, minv1=ci1minu1\min_{v_{1}}=c_{i-1}\min_{u_{1}} and minv2=ci1minu2\min_{v_{2}}=c_{i-1}\min_{u_{2}}. As before, we must have already considered u1u_{1} and u2u_{2}, but since v1v_{1} is not in II^{\prime}, by construction it must be minu1minu2\min_{u_{1}}\prec\min_{u_{2}} and so we conclude minv1minv2\min_{v_{1}}\prec\min_{v_{2}}.

  4. 4.

    Assume that i=i1i^{\prime}=i-1 and t2t^{\prime}\leq 2. As before, we obtain that if t=3t=3, then all vVci,3v\in V_{c_{i},3} are already in some element of the suffix of 𝒜\mathcal{A} whose largest element is II. Moreover, 𝒜ci,2\mathcal{A}_{c_{i},2} and 𝒜ci,1\mathcal{A}_{c_{i},1} must necessarily be empty because we have already built them, and II^{\prime} is not contained in any of them. Since we have already built 𝒜ci1,2\mathcal{A}_{c_{i-1},2} and 𝒜ci1,1\mathcal{A}_{c_{i-1},1}, we only have to prove that 𝒜ci1,3\mathcal{A}_{c_{i-1},3} is empty. If it were not empty, in particular there would exists vVci1,3v\in V_{c_{i-1},3} with ψ(v)=1\psi(v)=1. If uVu\in V is such that (u,v)E(u,v)\in E and minv=ci1minu\min_{v}=c_{i-1}\min_{u}, then we conclude that it must be ci1=λ(v)λ(u)c_{i-1}=\lambda(v)\prec\lambda(u). Then, uu must be in some element of the the suffix of 𝒜\mathcal{A} whose largest element is II, so vv would be in some element of Aci1,3A_{c_{i-1},3}; but this is a contradiction, because t2t^{\prime}\leq 2.

Lastly, if after considering II in Aci,tA_{c_{i},t} there is no element II^{\prime} to consider, we can check that every vVv\in V is in some element of the suffix of 𝒜\mathcal{A} whose largest element is II (and so have built 𝒜\mathcal{A}). Indeed, assume for sake of contradiction that this is not true. Consider the set SS of all nodes vv that do not satisfy these properties (it must be τ(v)=3\tau(v)=3); let SSS^{\prime}\subseteq S be the set of all nodes vv in SS for which λ(v)\lambda(v) is maximal, and let SSS^{\prime\prime}\subseteq S^{\prime} be the set of all nodes vv in SS^{\prime} for which ψ(v)\psi(v) is minimal. Pick any vSv\in S^{\prime\prime}, and let uVu\in V such that (u,v)E(u,v)\in E and minv=λ(v)minu\min_{v}=\lambda(v)\min_{u}. As before, we obtain that uu must be in some element of the suffix of 𝒜\mathcal{A} whose largest element is II, so vv should be in some Ac,3A_{c,3} and there would be an element II^{\prime} to consider, a contradiction.