Robustly Complete Finite-State Abstractions for Verification of Stochastic Systems
Abstract
In this paper, we focus on discrete-time stochastic systems modelled by nonlinear stochastic difference equations and propose robust abstractions for verifying probabilistic linear temporal specifications. The current literature focuses on developing sound abstraction techniques for stochastic dynamics without perturbations. However, soundness thus far has only been shown for preserving the satisfaction probability of certain types of temporal-logic specification. We present constructive finite-state abstractions for verifying probabilistic satisfaction of general -regular linear-time properties of more general nonlinear stochastic systems. Instead of imposing stability assumptions, we analyze the probabilistic properties from the topological view of metrizable space of probability measures. Such abstractions are both sound and approximately complete. That is, given a concrete discrete-time stochastic system and an arbitrarily small -perturbation of this system, there exists a family of finite-state Markov chains whose set of satisfaction probabilities contains that of the original system and meanwhile is contained by that of the slightly perturbed system. A direct consequence is that, given a probabilistic linear-time specification, initializing within the winning/losing region of the abstraction system can guarantee a satisfaction/dissatisfaction for the original system. We make an interesting observation that, unlike the deterministic case, point-mass (Dirac) perturbations cannot fulfill the purpose of robust completeness.
Keywords:
Verification of stochastic systems Finite-state abstraction Robustness Soundness Completeness -perturbation Linear temporal logic Metrizable space of probability measures1 Introduction
Formal verification is a rigorous mathematical technique for verifying system properties using formal analysis or model checking [4]. So far, abstraction-based formal verification for deterministic systems has gained its maturity [5]. Whilst bisimilar (equivalent) symbolic models exist for linear (control) systems [17, 31], sound and approximately complete finite abstractions can be achieved via stability assumptions [25, 14] or robustness (in terms of Dirac perturbations) [21, 20, 22].
There is a recent surge of interest in studying formal verification for stochastic systems. The verification of temporal logics for discrete-state homogeneous Markov chains can be solved by existing tools [4, 24, 6, 9].
In terms of verification for general discrete-time continuous-state Markov systems, a common theme is to construct abstractions to approximate the probability of satisfaction in proper ways. First attempts [26, 29, 30, 3, 2] were to relate the verification of trivial probabilistic computation tree logic (PCTL) formulas to the computation of corresponding value functions. The authors [32, 33] developed alternative techniques to deal with the potential error blow-up in infinite-horizon problems. The same authors [35] investigated the necessity of absorbing sets on the uniqueness of the solutions of corresponding Bellman equations. The related PCTL verification problem can be then precisely captured by finite-horizon ones. They also proposed abstractions for verifying general bounded linear-time (LT) properties [34], and extended them to infinite-horizon reach-avoid and repeated reachability problems [34, 36].
Markov set-chains are also constructive to be abstractions. The authors [1] showed that the error is finite under strong assumptions on stability (ergodicity). A closely related approach is to apply interval-valued Markov chains (IMCs), a family of finite-state Markov chains with uncertain transitions, as abstractions for the continuous-state Markov systems with certain transition kernel. The authors [18] argued without proof that for every PCTL formula, the probability of (path) satisfaction of the IMC abstractions forms a compact interval, which contains the real probability of the original system. They further developed ‘O’-maximizing/minimizing algorithms based on [15, 40] to obtain the upper/lower bound of the satisfaction probability of ‘next’, ‘bounded until’, and ‘until’ properties. The algorithm provides a fundamental view of computing the bounds of probability satisfaction given IMCs. However, the intuitive reasoning for soundness seems inaccurate based on our observation (readers who are interested in the details are referred to Remark 8 of this paper). Inspired by [18], the work in [7] formulated IMC abstraction for verifying bounded-LTL specifications; the work in [11, 12] constructed IMC abstractions for verifying general -regular properties of mixed-monotone systems, and provided a novel automata-based approach in obtaining the bounds of satisfaction probability. In [11, 12, Fact 1], the authors claimed the soundness of verifying general -regular properties using IMC abstractions, but a proof is not provided. In [10], the authors showed that IMC can be used to provide conservative estimates of the expected values for stochastic linear control system. The authors also remarked that their result can be extended to deal with omega-regular properties based on [15, 19] (where [19] is only regarding safety properties), but without any proofs. To the best knowledge of the authors, we currently lacks a general framework, as the one presented in the paper, for guaranteeing soundness of IMC abstractions for verifying -regular properties.
Motivated by these issues, our first contribution is to provide a formal mathematical proof of the soundness of IMC abstractions for verifying -regular linear-time properties. We show that, for any discrete-time stochastic dynamical systems modelled by a stochastic difference equation and any linear-time property, an IMC abstraction returns a compact interval of probability of (path) satisfaction which contains the satisfaction probability of the original system. A direct consequence is that starting within the winning/losing region computed by the abstraction can guarantee a satisfaction/dissatisfaction for the original system. The second contribution of this paper is to deal with stochastic systems with extra uncertain perturbations (due to, e.g., measurement errors or modeling uncertainties). Under mild assumptions, we show that, in verifying probabilistic satisfaction of general -regular linear-time properties, IMC abstractions that are both sound and approximately complete are constructible for nonlinear stochastic systems. That is, given a concrete discrete-time continuous-state Markov system , and an arbitrarily small -bounded perturbation of this system, there exists an IMC abstraction whose set of satisfaction probability contains that of , and meanwhile is contained by that of the slightly perturbed system. We argue in Section 4 that to make the IMC abstraction robustly complete, the perturbation is generally necessary to be -bounded rather than only bounded in terms of point mass. We analyze the probabilistic properties based on the topology of metrizable space of (uncertain) probability measures, and show that the technique proves more powerful than purely discussing the value of probabilities. We also would like to clarify that the main purpose of this paper is not on providing more efficient algorithms for computing abstractions. We aim to provide a theoretical foundation of IMC abstractions for verifying continuous-state stochastic systems with perturbations and hope to shed some light on designing more powerful robust verification algorithms.
The rest of the paper is organized as follows. Section 2 presents some preliminaries on probability spaces and Markov systems. Section 3 presents the soundness of abstractions in verifying -regular linear-time properties for discrete-time nonlinear stochastic systems. Section 4 presents the constructive robust abstractions with soundness and approximate completeness guarantees. We discuss the differences of robustness between deterministic and stochastic systems. The paper is concluded in Section 5.
Notation: We denote by the product of ordinary sets, spaces, or function values. Denote by the product of collections of sets, or sigma algebras, or measures. The -times repeated product of any kind is denoted by for simplification. Denote by the projection to the component. We denote by the Borel -algebra of a set.
Let denote the inifinity norm in , and let . We denote by the -norm for -valued random variables, and let . Given a matrix , we denote by its row and by its entry at row and column.
Given a general state space , we denote by the space of probability measures. The space of bounded and continuous functions on is denoted by . For any stochastic processes , we use the shorthand notation . For any stopped process , where is a stopping time, we use the shorthand notation .
2 Preliminaries
We consider as the discrete time index set, and a general Polish (complete and separable metric) space as the state space. For any discrete-time -valued stochastic process , we introduce some standard concepts as follows.
2.1 Canonical sample space
Given a stochastic process defined on some (most likely unknown) probability space . For and , we define and the coordinate process as associated with . Then () is a measurable map from to . In particular, , where is the collection of all finite-dimensional cylinder set of the following form:
The measure of the defined coordinate process is then uniquely determined and admits the probability law of the process on the product state space, i.e.,
(1) |
for any finite-dimensional cylinder set . We call the canonical space of and denote by the associated expectation operator.
Since we only care about the probabilistic behavior of trajectories in the state space, we prefer to work on the canonical probability spaces and regard events as sets of sample paths. To this end, we also do not distinguish the notation from due to their identicality in distribution, i.e., we use to denote its own coordinate process for simplicity.
In the context of discrete state space , we specifically use the boldface notation for the discrete canonical spaces of some discrete-state process.
Remark 1
We usually denote by the marginal distribution of at some . We can informally write the -dimensional distribution (on -dimensional cylinder set) as regardless of the dependence.
2.2 Markov transition systems
For any discrete-time stochastic process , we set to be the natural filtration.
Definition 1 (Markov process)
A stochastic process is said to be a Markov process if each is -adapted and, for any and , we have
(2) |
Correspondingly, for every , we define the transition probability as
(3) |
We denote as the family of transition probabilities at time . Note that homogeneous Markov processes are special cases such that for all .
We are interested in Markov processes with discrete observations of states, which is done by assigning abstract labels over a finite set of atomic propositions. We define an abstract family of labelled Markov processes as follows.
Definition 2 (Markov system)
A Markov system is a tuple , where
-
•
, where is a bounded working space, represents all the out-of-domain states;
-
•
is a collection of transition probabilities from which is chosen for every ;
-
•
is the finite set of atomic propositions;
-
•
is the (Borel-measurable) labelling function.
For with a.s., we denote by the law, and by its collection. Similarly, for any initial distribution , we define the law by , and denote by its collection. We denote by (resp. ) a sequence of (resp. ). We simply use (resp. ) if we do not emphasize the initial condition.
For a path , define by its trace. The space of infinite words is denoted by
A linear-time (LT) property is a subset of . We are only interested in LT properties such that , i.e., those are Borel-measurable.
Remark 2
A particular subclass of LT properties can be specified by linear temporal logic (LTL)111While we consider LTL due to our interest, it can be easily seen that all results of this paper in fact hold for any measurable LT property, including -regular specifications.. To connect with LTL specifications, we introduce the semantics of path satisfaction as well as probabilistic satisfaction as follows.
Definition 3
For the syntax of LTL formulae and the semantics of satisfaction of on infinite words, we refer readers to [21, Section 2.4].
For a given labelled Markov process from with initial distribution , we formulate the canonical space . For a path , we define the path satisfaction as
We denote by the events of path satisfaction. Given a specified probability , we define the probabilistic satisfaction of as
where .
2.3 Weak convergence and Prokhorov’s theorem
We consider the set of possible uncertain measures within the topological space of probability measures. The following concepts are frequently used later.
Definition 4 (Tightness of set of measures)
Let be any topological state space and be a set of probability measures on . We say that is tight if, for every there exists a compact set such that for every .
Definition 5 (Weak convergence)
A sequence is said to converge weakly to a probability measure , denoted by , if
(4) |
We frequently use the following alternative condition [8, Proposition 2.2]:
(5) |
Correspondingly, the weak equivalence of any two measures and on is such that
(6) |
Remark 3
Weak convergence describes the weak topology. The meaning of the weak topology is to extend the normal convergence in deterministic settings. Note that in is equivalent to the weak convergence of Dirac measures . It is interesting to note that (resp. ) in does not imply the strong convergence (resp. equivalence) of the associated Dirac measures. A classical counterexample is to let and , and we do not have in the strong sense since, i.e.,
To describe the convergence (in probability law) of more general random variables in , it is equivalent to investigate the weak convergence of their associated measures . It is also straightforward from Definition 5 that weak convergence also describes the convergence of probabilistic properties related to .
Theorem 1 (Prokhorov)
Let be a complete separable metric space. A family is relatively compact if an only if it is tight. Consequently, for each sequence of tight , there exists a and a subsequence such that .
Remark 4
The first part of Prokhorov’s theorem provides an alternative criterion for verifying the compactness of family of measures w.r.t. the corresponding metric space using tightness. On a compact metric space , every family of probability measures is tight.
2.4 Discrete-time continuous-state stochastic systems
We define Markov processes determined by the difference equation
(7) |
where the state for all , the stochastic inputs are i.i.d. Gaussian random variables with covariance without loss of generality. Mappings and are locally Lipschitz continuous. The memoryless perturbation are independent random variables with intensity and unknown distributions.
For , (7) defines a family of Markov processes . A special case of (7) is such that has Dirac (point-mass) distributions centered at some uncertain points within a unit ball.
Remark 5
The discrete-time stochastic dynamic is usually obtained from numerical schemes of stochastic differential equations driven by Brownian motions to simulate the probability laws at the observation times. Gaussian random variables are naturally selected to simulate Brownian motions at discrete times. Note that in [11], random variables are used with known unimodal symmetric density with an interval as support. Their choice is in favor of the mixed-monotone models to provide a more accurate approximation of transition probabilities. Other than the precision issue, such a choice does not bring us more of the other properties. Since we focus on formal analysis based on properties rather than providing accurate approximation, using Gaussian randomnesses as a realization does not lose any generality.
We only care about the behaviors in the bounded working space . By defining stopping time for each , we are able to study the probability law of the corresponding stopped (killed) process for any initial condition (resp. ), which coincides with (resp. ) on . To avoid any complexity, we use the same notation and (resp. ) to denote the stopped processes and the associated laws. Such processes driven by (7) can be written as a Markov system
(8) |
where for all , the transition probability should satisfy for all ; is the collection of transition probabilities. For having Dirac distributions, the transition is of the following form:
(9) |
Assumption 1
We assume that for any and . We can also include ‘always ’ in the specifications to observe sample paths for ‘inside-domain’ behaviors, which is equivalent to verifying .
2.5 Robust abstractions
We define a notion of abstraction between continuous-state and finite-state Markov systems via state-level relations and measure-level relations.
Definition 6
A (binary) relation from to is a subset of satisfying (i) for each , ; (ii) for each , ; (iii) for , ; (iv) and for , .
Definition 7
Given a continuous-state Markov system
and a finite-state Markov system
where and stands for a collection of stochastic matrices. A state-level relation is said to be an abstraction from to if (i) for all , there exists such that ; (ii) for all , .
A measure-level relation is said to be an abstraction from to if for all , all and all , there exists such that and that for all .
Similarly, is said to be an abstraction from to if for all , all and all , there exists such that and that for all .
If such relations and exist, we say that abstracts (resp. abstracts ) and write (resp. ).
Assumption 2
Without loss of generality, we assume that the labelling function is amenable to a rectangular partition222See e.g. [11, Definition 1].. In other words, a state-level abstraction can be obtained from a rectangular partition.
3 Soundness of Robust IMC Abstractions
IMCs333We omit the definition from this paper due to the limitation of space. For a formal definition see e.g. [18, Definition 3]. are quasi-Markov systems on a discrete state space with upper/under approximations (/) of the real transition matrices. To abstract the transition probabilities of continuous-state Markov systems (8), and are obtained from over/under approximations of based on the state space partition. Throughout this section, we assume that and have been correspondingly constructed.
Given an IMC, we recast it to a finite-state Markov system
(10) |
where
-
•
is the finite state-space partition with dimension containing , i.e., ;
- •
-
•
are as before.
To make an abstraction for (10), we need the approximation to be such that for all and , as well as . We further require that the partition should respect the boundaries induced by the labeling function, i.e., for any ,
Clearly, the above connections on the state and transition probabilities satisfy Definition 7.
The Markov system is understood as a family of ‘perturbed’ Markov chain generated by the uncertain choice of for each . The -step transition matrices are derived based on as
Given an initial distribution , the marginal probability measure at each forms a set
(12) |
If we do not emphasize the initial distribution , we also use to denote the marginals for short.
We aim to show the soundness of robust IMC abstractions in this section. The proofs in this section are completed in Appendix 0.A.
3.1 Weak compactness of marginal space of probabilities
The following lemma is rephrased from [39, Theorem 2] and shows the structure of the for each and any initial distribution .
Lemma 1
Let be a Markov system of the form (10) that is derived from an IMC. Then the set of all possible probability measures at each time is a convex polytope, and immediately is compact. The vertices of are of the form
(13) |
for some vertices of .
Example 1
Let and . The under/over estimations of transition matrices are given as
Then forms a convex set of stochastic matrices with vertices
Therefore, the vertices of are
Hence, . Similarly, the vertices of are
and
The calculation of the rest of follows the same procedure.
Now we introduce the total variation distance and see how (at each ) implies the weak topology.
Definition 8 (Total variation distance)
Given two probability measures and on , the total variation distance is defined as
(14) |
In particular, if is a discrete space,
(15) |
Corollary 1
Let be a Markov system of the form (10) that is derived from an IMC. Then at each time , for for each , there exists a and a subsequence such that . In addition, for each and , the set forms a convex and compact subset in .
Remark 6
The above shows that metricizes the weak topology of measures on . Note that since is bounded and finite, any metrizable family of measures on is compact. For example, let , and be a set of singular measures on . Then every sequence of the above set has a weakly convergent subsequence. However, these measures do not have a convex structure as . Hence the corresponding that is generated by only provides vertices in .
3.2 Weak compactness of probability laws of on infinite horizon
In this subsection, we focus on the case where a.s. for any . The cases for arbitrary initial distribution should be similar. We formally denote by the set of probability laws of every discrete-state Markov processes with initial state . We denote by the set of marginals at .
Proposition 1
For any , every sequence of has a weakly convergent subsequence.
Remark 7
The property is an extension of the marginal weak compactness relying on the (countable) product topology. We can also introduce proper product metrics to metricize, see e.g. [28]. Similar results hold under certain conditions for continuous time processes on continuous state spaces with uniform norms [27, Lemma 82.3 and 87.3].
Theorem 2
Let be a Markov system of the form (10) that is derived from an IMC. Then for any LTL formula , the set is a convex and compact subset in , i.e., a compact interval.
3.3 Soundness of IMC abstractions
Proposition 2
Let be a Markov system driven by (8). Then every sequence of has a weakly convergent subsequence. Consequently, for any LTL formula , the set is a compact subset in .
Lemma 2
Let be any Markov process driven by (8) and be the finite-state IMC abstraction of . Suppose the initial distribution of is such that . Then, there exists a unique law of some such that, for any LTL formula ,
Theorem 3
Assume the settings in Lemma 2. For any LTL formula , we have
Corollary 2
Let , its IMC abstraction , an LTL formula , and a constant be given. Suppose for all , we have for all with .
Remark 8
Note that we do not have since each is a discrete measure whereas is not. They only coincide when measuring Borel subset of . It would be more accurate to state that is a member of rather than say “the true distribution (the law as what we usually call) of the original system is a member of the distribution set represented by the abstraction model” [18].
Remark 9
We have seen that, in view of Lemma 2, the ‘post-transitional’ measures are automatically related only based on the relations between transition probabilities. We will see in the next section that such relations can be constructed to guarantee an approximate completeness of .
Proposition 3
Let . Then for each LTL formula , as , the length .
Remark 10
By Lemma 2, for each , there exists exactly one of some by which satisfaction probability equals to that of . The precision of and determines the size of . Once we are able to calculate the exact law of , the becomes a singleton by Proposition 3. For example, let each become , we have each reduced to a singleton automatically. The verification problem becomes checking whether given the partition . The probability of satisfaction is either or . Another example would be , where are linear matrices. We are certain about the exact law of this system, and there is no need to introduce IMC for approximations at the beginning. IMC abstractions prove more useful when coping with systems whose marginal distributions are uncertain or not readily computable.
4 Robust Completeness of IMC Abstractions
In this section, we are given a Markov system driven by (7) with point-mass perturbations of strength . Based on , we first construct an IMC abstraction . We then show that can be abstracted by a system with more general -bounded noise of any arbitrary strength .
Recalling the soundness analysis of IMC abstractions in Section 3, the relation of satisfaction probability is induced by a relation between the continuous and discrete transitions. To capture the probabilistic properties of stochastic processes, reachable set of probability measures is the analogue of the reachable set in deterministic cases. We rely on a similar technique in this section to discuss how transition probabilities of different uncertain Markov systems are related. To metricize sets of Gaussian measures and to connect them with discrete measures, we prefer to use Wasserstein metric.
Definition 9
Let for , the Wasserstein distance555This is formally termed as -Wasserstein metric. We choose -Wasserstein metric due to the convexity and nice property of test functions. is defined by , where the infimum is is taken over all joint distributions of the random variables and with marginals and respectively. We frequently use the following duality form of definition666 is the Lipschitz constant of such that .,
The discrete case, , is nothing but to change the integral to summation. Let . Given a set , we denote by the distance from to , and 777This is valid by definition. by the -neighborhood of .
Remark 11
Note that is dual to . For any , the associated random variable should satisfy , and vice versa.
The following well-known result estimates the Wasserstein distance between two Gaussians.
Proposition 4
Let and be two Gaussian measures on . Then
(16) |
where is the Frobenius norm.
On finite state spaces, total variation and Wasserstein distances manifest equivalence [13, Theorem 4]. We only show the following side of inequality in favor of the later proofs.
Proposition 5
For any on some discrete and finite space , we have
(17) |
Before proceeding, we define the set of transition probabilities of from any box as
and use the following lemma to approximate .
Lemma 3
Fix any , any box . For all , there exists a finitely terminated algorithm to compute an over-approximation of the set of (Gaussian) transition probabilities from , such that
where is the computed over-approximation set of Gaussian measures.
Remark 12
The proof is completed in Appendix 0.B. The lemma renders the inclusions with larger Wasserstein distance to ensure no missing information about the covariances.
Definition 10
For , we introduce the modified transition probabilities for based on (9). For all , let
(18) |
Correspondingly, let denote the collection. Likewise, we also use to denote the induced quantities of any other types w.r.t. such a modification.
Remark 13
We introduce the concept only for analysis. The above modification does not affect the law of the stopped processes since we do not care about the ‘out-of-domain’ transitions. We use a weighted point mass to represent the measures at the boundary, and the mean should remain the same. It can be easily shown that the Wasserstein distance between any two measures in is upper bounded by that of the non-modified ones.
Theorem 4
For any , we set , , where is perturbed by point masses with intensity , and is perturbed by general -perturbation with intensity . Then, under Assumption 2, there exists a rectangular partition (state-level relation ),a measure-level relation and a collection of transition matrices , such that the system abstracts and is abstracted by by the following relation:
(19) |
Proof We construct a finite-state IMC with partition and an inclusion of transition matrices as follows. By Assumption 2, we use uniform rectangular partition on and set , where is the floor function and is to be chosen later. Denote the number of discrete nodes by .
Note that any family of (modified) Gaussian measures is induced from and should contain its information. For any and ,
-
(i)
for all , store ;
-
(ii)
for each , define (implicitly, we need to compute ); compute for each ;
-
(iii)
for each , define ;
-
(iv)
compute and ;
-
(v)
construct ;
-
(vi)
Let be a relation between and the generated .
Repeat the above step for all , the relation is obtained. The rest of the proof falls in the following steps. For , we simply denote and .
Claim 1: For , let . Then the finite-state IMC with transition collection abstracts .
Indeed, for each and each , we have . We pick any modified Gaussian , there exists a such that (by Proposition 4) . We aim to find all discrete measures induced from (such that their probabilities match on discrete nodes as requirement by Definition 7). All such should satisfy888Note that we also have , but it is hard to connect with for general measures. This connection can be done if we only compare Dirac or discrete measures.,
(20) |
where the first term of line 2 is bounded by,
(21) |
and the third term of line 2 is bounded in a similar way.
By step (v)(vi) and Proposition 5,
all possible discrete measures induced from should be included in . Combining the above, for any and hence in , there exists a discrete measures in such that for all we have . This satisfies the definition of abstraction.
Claim 2: . This is to recover all possible (modified) measures from the constructed , such that their discrete probabilities coincide. Note that, the ‘ref’ information is recorded when computing in the inner parentheses. Therefore, for any there exists a within a total variation radius . We aim to find corresponding measure that matches by their probabilities on discrete nodes. All such should satisfy,
(22) |
where the bounds for the first and third terms are obtained in the same way as (21).
The second term is again by a rough comparison in Proposition 5. Note that is already recorded in . The inequality in (22) provides an upper bound of Wasserstein deviation between any possible satisfactory measure and some .
Claim 3: If we can choose and sufficiently small such that
(23) |
then .
Indeed, the is obtained by for each . By Claim 2 and Lemma 3, we have
for each . By the construction, we can verify that . The selection of makes , which completes the proof.
Remark 14
The relation (resp. ) provides a procedure to include all proper (continuous, discrete) measures that connect with the discrete probabilities. The key point is to record , , and the corresponding radius. These are nothing but finite coverings of the space of measures. This also explains the reason why we use ‘finite-state’ rather than ‘finite’ abstraction. The latter has a meaning of using finite numbers of representative measures to be the abstraction.
To guarantee a sufficient inclusion, conservative estimations, e.g. the bound in Claim 1 and the bound in Proposition 5, are made. This estimation can be done more accurately given more assumptions. For example, the deterministic systems (where becomes ) provide Dirac transition measures, the and hence the second term in (22) is .
Remark 15
Note that, to guarantee the second abstraction based on , we search all possible measures that has the same discrete probabilities as , not only those Gaussians with the same covariances as (or ). Such a set of measures provide a convex ball w.r.t. Wasserstein distance. This actually makes sense because in the forward step of creating , we have used both Wasserstein and total variation distance to find a convex inclusion of all Gaussian or Gaussian related measures. There ought to be some measures that are ‘non-recoverable’ to Gaussians, unless we extract some ‘Gaussian recoverable’ discrete measures in , but this loses the point of over-approximation. In this view, IMC abstractions provide unnecessarily larger inclusions than needed.
For the deterministic case, the above mentioned ‘extraction’ is possible, since the transition measures do not have diffusion, the convex inclusion becomes a collection of vertices themselves (also see Remark 6). Based on these vertices, we are able to use to find the measures within a convex ball w.r.t. Wasserstein distance.
In contrast to the above special case [21], where the uncertainties are bounded w.r.t. the infinity norm, we can only guarantee the approximated completeness via a robust -bounded perturbation with strictly larger intensity than the original point-mass perturbation. However, this indeed describes a general type of uncertainties for the stochastic systems to guarantee -related properties, including probabilistic properties. Unless higher-moment specifications are of interests, uncertain -random variables are what we need to be the analogue of perturbations in [21].
Corollary 3
Given an LTL formula , let () and , where the initial conditions are such that . Then all the above sets are compact and .
The proof in shown in Appendix 0.B.
5 Conclusion
In this paper, we constructed an IMC abstraction for continuous-state stochastic systems with possibly bounded point-mass (Dirac) perturbations. We showed that such abstractions are not only sound, in the sense that the set of satisfaction probability of linear-time properties contains that of the original system, but also approximately complete in the sense that the constructed IMC can be abstracted by another system with stronger but more general -bounded perturbations. Consequently, the winning set of the probabilistic specifications for a more perturbed continuous-state stochastic system contains that of the less Dirac perturbed system. Similar to most of the existing converse theorems, e.g. converse Lyapunov functions, the purpose is not to provide an efficient approach for finding them, but rather to characterize the theoretical possibilities of having such existence.
It is interesting to compare with robust deterministic systems, where no random variables are involved. In [21], both perturbed systems are w.r.t. bounded point masses. More heavily perturbed systems abstract less perturbed ones and hence preserve robust satisfaction of linear-time properties. However, when we try to obtain the approximated completeness via uncertainties in stochastic system, the uncertainties should be modelled by more general random variables. Note that the probabilistic properties of random variables is dual to the weak topology of measures, we study the measures and hence probability laws of processes instead of the state space per se. The state-space topology is not sufficient to quantify the regularity of IMC abstractions. In contrast, the uncertain random variables is a perfect analogue of the uncertain point masses (in ) for deterministic systems. If we insist on using point masses as the only type of uncertainties for stochastic systems, the IMC type abstractions would possibly fail to guarantee the completeness. For example, suppose the point-mass perturbations represents less precision of deterministic control inputs [23, Definition 2.3], the winning set decided by the -precision stationary policies is not enough to cover that of the IMC abstraction, which fails to ensure an approximated bi-similarity of IMCs compared to [21].
For future work, it would be useful to extend the current approach to robust stochastic control systems. It would be interesting to design algorithms to construct IMC (resp. bounded-parameter Markov decision processes) abstractions for more general robust stochastic (resp. control) systems with perturbations based on metrizable space of measures and weak topology. The size of state discretization can be refined given more specific assumptions on system dynamics and linear-time objectives. For verification or control synthesis w.r.t. probabilistic safety or reachability problems, comparisons can be made with stochastic Lyapunov-barrier function approaches.
References
- [1] Abate, A., D’Innocenzo, A., Di Benedetto, M.D., Sastry, S.S.: Markov set-chains as abstractions of stochastic hybrid systems. In: Proc. of Hybrid Systems: Computation and Control (HSCC). pp. 1–15. Springer (2008)
- [2] Abate, A., Katoen, J.P., Mereacre, A.: Quantitative automata model checking of autonomous stochastic hybrid systems. In: Proc. of Hybrid Systems: Computation and Control (HSCC). pp. 83–92 (2011)
- [3] Abate, A., Prandini, M., Lygeros, J., Sastry, S.: Probabilistic reachability and safety for controlled discrete time stochastic hybrid systems. Automatica 44(11), 2724–2734 (2008)
- [4] Baier, C., Katoen, J.P.: Principles of Model Checking. MIT Press (2008)
- [5] Belta, C., Yordanov, B., Gol, E.A.: Formal Methods for Discrete-time Dynamical Systems, vol. 89. Springer (2017)
- [6] Bustan, D., Rubin, S., Vardi, M.Y.: Verifying -regular properties of markov chains. In: International Conference on Computer Aided Verification. pp. 189–201. Springer (2004)
- [7] Cauchi, N., Laurenti, L., Lahijanian, M., Abate, A., Kwiatkowska, M., Cardelli, L.: Efficiency through uncertainty: Scalable formal synthesis for stochastic hybrid systems. In: Proc. of Hybrid Systems: Computation and Control (HSCC). pp. 240–251 (2019)
- [8] Da Prato, G., Zabczyk, J.: Stochastic equations in infinite dimensions. Cambridge University Press (2014)
- [9] Dehnert, C., Junges, S., Katoen, J.P., Volk, M.: A storm is coming: A modern probabilistic model checker. In: International Conference on Computer Aided Verification. pp. 592–600. Springer (2017)
- [10] Delimpaltadakis, G., Laurenti, L., Mazo Jr, M.: Abstracting the sampling behaviour of stochastic linear periodic event-triggered control systems. arXiv preprint arXiv:2103.13839 (2021)
- [11] Dutreix, M., Coogan, S.: Specification-guided verification and abstraction refinement of mixed monotone stochastic systems. IEEE Transactions on Automatic Control (2020)
- [12] Dutreix, M.D.H.: Verification and synthesis for stochastic systems with temporal logic specifications. Ph.D. thesis, Georgia Institute of Technology (2020)
- [13] Gibbs, A.L., Su, F.E.: On choosing and bounding probability metrics. International statistical review 70(3), 419–435 (2002)
- [14] Girard, A., Pola, G., Tabuada, P.: Approximately bisimilar symbolic models for incrementally stable switched systems. IEEE Transactions on Automatic Control 55(1), 116–126 (2009)
- [15] Givan, R., Leach, S., Dean, T.: Bounded-parameter markov decision processes. Artificial Intelligence 122(1-2), 71–109 (2000)
- [16] Hartfiel, D.J.: Markov Set-Chains. Springer (2006)
- [17] Kloetzer, M., Belta, C.: A fully automated framework for control of linear systems from temporal logic specifications. IEEE Transactions on Automatic Control 53(1), 287–297 (2008)
- [18] Lahijanian, M., Andersson, S.B., Belta, C.: Formal verification and synthesis for discrete-time stochastic systems. IEEE Transactions on Automatic Control 60(8), 2031–2045 (2015)
- [19] Laurenti, L., Lahijanian, M., Abate, A., Cardelli, L., Kwiatkowska, M.: Formal and efficient synthesis for continuous-time linear stochastic hybrid processes. IEEE Transactions on Automatic Control 66(1), 17–32 (2020)
- [20] Li, Y., Liu, J.: Robustly complete synthesis of memoryless controllers for nonlinear systems with reach-and-stay specifications. IEEE Transactions on Automatic Control (2020)
- [21] Liu, J.: Robust abstractions for control synthesis: Completeness via robustness for linear-time properties. In: Proc. of Hybrid Systems: Computation and Control (HSCC). pp. 101–110 (2017)
- [22] Liu, J.: Closing the gap between discrete abstractions and continuous control: Completeness via robustness and controllability. In: International Conference on Formal Modeling and Analysis of Timed Systems. pp. 67–83. Springer (2021)
- [23] Majumdar, R., Mallik, K., Soudjani, S.: Symbolic controller synthesis for büchi specifications on stochastic systems. In: Proceedings of the 23rd International Conference on Hybrid Systems: Computation and Control. pp. 1–11 (2020)
- [24] Parker, D.: Verification of probabilistic real-time systems. Proc. 2013 Real-time Systems Summer School (ETR’13) (2013)
- [25] Pola, G., Girard, A., Tabuada, P.: Approximately bisimilar symbolic models for nonlinear control systems. Automatica 44(10), 2508–2516 (2008)
- [26] Ramponi, F., Chatterjee, D., Summers, S., Lygeros, J.: On the connections between pctl and dynamic programming. In: Proc. of Hybrid Systems: Computation and Control (HSCC). pp. 253–262 (2010)
- [27] Rogers, L.C.G., Williams, D.: Diffusions, markov processes and martingales, volume 1: Foundations. Cambridge Mathematical Library, (2000)
- [28] Sagar, G., Ravi, D.: Compactness of any countable product of compact metric spaces in product topology without using tychonoff’s theorem. arXiv preprint arXiv:2111.02904 (2021)
- [29] Soudjani, S.E.Z., Abate, A.: Adaptive gridding for abstraction and verification of stochastic hybrid systems. In: 2011 Eighth International Conference on Quantitative Evaluation of SysTems. pp. 59–68. IEEE (2011)
- [30] Summers, S., Lygeros, J.: Verification of discrete time stochastic hybrid systems: A stochastic reach-avoid decision problem. Automatica 46(12), 1951–1961 (2010)
- [31] Tabuada, P., Pappas, G.J.: Linear time logic control of discrete-time linear systems. IEEE Transactions on Automatic Control 51(12), 1862–1877 (2006)
- [32] Tkachev, I., Abate, A.: On infinite-horizon probabilistic properties and stochastic bisimulation functions. In: 2011 50th IEEE Conference on Decision and Control and European Control Conference. pp. 526–531. IEEE (2011)
- [33] Tkachev, I., Abate, A.: Regularization of bellman equations for infinite-horizon probabilistic properties. In: Proc. of Hybrid Systems: Computation and Control (HSCC). pp. 227–236 (2012)
- [34] Tkachev, I., Abate, A.: Formula-free finite abstractions for linear temporal verification of stochastic hybrid systems. In: Proc. of Hybrid Systems: Computation and Control (HSCC). pp. 283–292 (2013)
- [35] Tkachev, I., Abate, A.: Characterization and computation of infinite-horizon specifications over markov processes. Theoretical Computer Science 515, 1–18 (2014)
- [36] Tkachev, I., Mereacre, A., Katoen, J.P., Abate, A.: Quantitative model-checking of controlled discrete-time markov processes. Information and Computation 253, 1–35 (2017)
- [37] Valby, L.V.: A category of polytopes. Ph.D. thesis, Reed College (2006)
- [38] Vardi, M.Y.: Automatic verification of probabilistic concurrent finite state programs. In: 26th Annual Symposium on Foundations of Computer Science (FOCS). pp. 327–338. IEEE (1985)
- [39] Vassiliou, P.C.: Non-homogeneous markov set systems. Mathematics 9(5), 471 (2021)
- [40] Wu, D., Koutsoukos, X.: Reachability analysis of uncertain systems using bounded-parameter markov decision processes. Artificial Intelligence 172(8-9), 945–954 (2008)
Appendix 0.A Proofs of Section 3
Proof of Corollary 1.
Proof It is clear that under discrete metric is complete and separable. In addition, for each , the space is complete and separable. By Lemma 1, each is also compact. For any sequence , a quick application of Theorem 1 leads to the existence of a weakly convergent subsequence and a weak limit point in . By the definition of weak convergence and the discrete structure of , it is clear that for each and , we have
in a strong sense, which concludes the compactness of . Now we choose , then for all . Therefore,
for all . This shows the convex structure of .
Proof of Proposition 1.
Proof We make a bit abuse of notation and define as the projection onto the finite product space of up to time . Since we do no emphasize the initial conditions, we also use , and for short. By Tychonoff theorem, any product of is also compact w.r.t. the product topology. Therefore, any family of measures on is tight and hence compact. By Remark 1, for every , we have (recall Remark 1) for some , and forms a compact set. Hence, every sequence with any finite contains a weakly convergent subsequence. We construct the convergent subsequence of in the following way.
We initialize the procedure by setting . Then is compact, and there exists a weakly convergent subsequence . Based on , we are able to see that it contains a weakly convergent subsequence, denoted by , such that weakly converges. By induction, we have for each . Repeating this argument and picking the diagonal subsequence , then has the property that is weakly convergent for each . We denote the weak limit point of each by . By the way of construction, we have
By Kolmogorov’s extension theorem, there exists a unique on such that for each .
We have seen that for each , the constructed subsequence satisfies , which concludes the claim.
Proof of Theorem 2
Proof Since we do not emphasize the initial conditions, we simply drop the superscripts for short. Given with any initial condition, the corresponding canonical space is . By Proposition 1, every sequence has a weakly convergent subsequence, denoted by , to a of some . Note that for any , the measurable set is the same due to the identical labelling function. It is important to notice that due to the discrete topology of , every Borel measurable set is such that . By Definition 5 we have for all . The compactness of follows immediately. The convexity of the set of laws is based on the tensor product of convex polytopes [37]. To show the convexity of , we notice that, for any and ,
and hence forms a convex set. Immediately, the convexity holds for
for any cylinder set . By a standard monotone class argument, is also convex for any Borel measurable set , which implies the convexity of in the statement.
Proof of Proposition 2
Proof
Note that the laws are associated with with , which actually means the stopped process (Recall notations in Section 2.4). Now that for each , the state space of is compact, so is the countably infinite product. By a similar argument as Proposition 1, we can conclude the first part of the statement. Note that by assumption, the partition respects the boundary of the labelling function. Hence, for all formula , the boundary of has measure . The second part can be concluded directly by Definition 5.
Proof of Lemma 2
Proof Note that is on and is on . We first show the case when for any . That is, for a.s. with any , there exists a unique law of some such that for any .
Let denote the marginal distribution of at each . Let denote the set of marginal distributions of . Now, at , for all . Suppose is the element of , by the construction of IMC, we have
Since , by letting , we have automatically by definition. Note that is unique w.r.t. , and has the property that for each .
Similarly, at , we have
where may not be the same as that of . Therefore, for any ,
and there exists a such that , which means (by (12)) . In addition, there also exists a such that its one-dimensional marginals up to admit and , and satisfies
Repeating this procedure, there exists a unique w.r.t. for each , such that for each . It is also clear that for each given and each , the selected satisfies
By Kolmogrov extension theorem, there exists a unique law of some such that each -dimensional distribution coincides with , and, for each given , for all Due to the assumption that for all and , we have
for all LTL formula , which implies by definition. Thus, given , the above should satisfy .
Based on the above conclusion, as well as the definition of and the convexity of (recall Theorem 2), the result for more general initial distribution with can be obtained.
Proof of Proposition 3.
Proof Let for each , and be any stochastic matrices generated by . Then, for each , we have
(24) |
This implies that the total deviation of any is bounded by
Note that at , . Hence, at each , as ,
By the product topology and Kolmogrov extension theorem, the set is reduced to a singleton. The conclusion follows after this.
Appendix 0.B Proofs of Section 4
Proof of Lemma 3.
Proof It can be proved, for example, using inclusion functions. Let denote the set of all boxes in . Let be a convergent inclusion function of satisfying (i) for all ; (ii) , where denote the width. Similarly, let be a convergent inclusion matrix of and satisfy (i) for all ; (ii) , where .
Without loss of generality, we assume that . Due to the Lipschitz continuity of and , we can find inclusions such that for any subintervals of , and similarly . For each such interval , we can obtain the interval and . Let denote the collection of Gaussian measures with mean and covariance of all such intervals ( and ), and be its union. Then satisfies the requirement. Indeed, we have . For the second part of inclusion, we have for any and ,
(25) |
where we are able to choose arbitrarily small such that . The second part of inclusion can be completed by such a choice of .
Proof of Corollary 3
Proof The first part of the proof is provided in Section 3. The second inclusion is done in a similar way as Lemma 2 and Theorem 3. Indeed, by the definition of abstraction, for any , there exists a marginal measure of some such that their probabilities match on discrete nodes. By the same induction as Lemma 2, we have that for any law of some , there exists a of some such that the probabilities of any match. The second inclusion follows after this. The compactness also follows a similar way as Proposition 2. Note that, may not be convex, but and are (also see details in Remark 15).