probabilistic verification of learning-enabled cyber ...tems [13], [10]. the major difference...

8
Probabilistic Verification of Learning-Enabled Cyber-Physical Systems using Conformal Regression Chuchu Fan, Xin Qin and Jyotirmoy Deshmukh Abstract— The use of machine learning components has posed significant challenges for the verification of cyber-physical systems due to its complexity, nonlinearity, and large space of parameters. In this work, we propose a novel probabilistic verification framework for learning-enabled CPS which can search over the entire (infinite) space of parameters, to figure out the ones that lead to satisfaction or violation of specification that are captured by Signal Temporal Logic (STL) formulas. Our technique is based on conformal regression, a technique for constructing prediction intervals with marginal coverage guarantees using finite samples, without making assumptions on the distribution and regression model. Our verification frame- work, using conformal regression, can predict the quantitative satisfaction values of the system’s trajectories over different sets of the parameters and use those values to quantify how well/bad the system with the parameters can satisfy/violate the given STL property. We use three case studies of learning- enabled CPS applications to demonstrate that our technique can be successfully applied to partition the parameter space and provide the needed level of assurance. I. I NTRODUCTION There is pervasive interest in using machine learning com- ponents such as advanced driving assist systems (ADAS), self-driving controllers, and neural network-based controllers for robotics. As these are safety-critical cyber-physical sys- tems, there is urgent need for effective and sound verification algorithms [7], [16], [18], [33], [35], [8], [15], [40]. Such systems have several uncertainties arising from exogenuous inputs from the environment, initial system configurations, and other design parameters. We call the space of these uncertainties the parameter space, and would like to use verification techniques to find the range of parameters where we can assure that the learning-enabled CPS exhibits safe behavior with high confidence. Such systems are unfortunately not well-suited to cur- rent verification and testing methods that either conduct exhaustive state-space search and are therefore face the curse of dimensionality [35], [8], or use best-effort approaches without explicit coverage guarantees [14], [26], [39]. In this paper, we fill this gap by developing a verification framework that can search over the entire parameter space Θ of the system to find the set of parameters that lead to the satisfaction or violation of given properties, and provide probabilistic coverage guarantees. We treat learning-enabled CPS as black-box dynamical models for which we only have executable access: once the parameter θ Θ is fixed, we have a deterministic trajectory Chuchu Fan is with the Department of Computing and Mathemati- cal Sciences, California Institute of Technology, Los Angeles CA, USA [email protected] Xin Qin and Jyotirmoy Deshmukh are with the Department of Com- puter Science, University of Southern California, Los Angeles CA, USA {xinqin,jyotirmoy.deshmukh}@usc.edu ξ θ which represents how the state of the system evolves. We consider system property that are specified using Signal temporal logic (STL) [6], which a formalism that enables specifications over real-valued continuous trajectories and have been widely applied to rigorously specify requirements in CPS [40]. STL is naturally equipped with a quantitative semantics called robust satisfaction value (RSV) ρ, which is a function mapping a property ϕ and a trajectory ξ to a real number ρ(ϕ, ξ ). A positive value suggests that ξ (t) satisfies ϕ and a negative value indicates that ξ (t) violates of ϕ. Moreover, the larger (smaller) the RSV is, more easily ξ satisfies (violates) the specification ϕ. Suppose the parameters follow a probabilistic distribution D θ over Θ. Given an STL formula ϕ, θ and the correspond- ing RSV ρ(ϕ, ξ θ ) follow a joint distribution D θ,ρ . Although we do not know the concrete form of D θ,ρ , we can always sample from this distribution by sampling θ ∼D θ first then compute the corresponding RSV. Then we use confor- mal regression [22], [21], [29], a generic tool to construct distribution-free prediction sets, over the joint distribution D θρ to predict the range of the RSV ρ for different sets of parameters. Conformal regression techniques can construct prediction intervals that attain valid coverage in finite sam- ples, without making assumptions on the distribution and the underlying regression algorithm. Therefore, our verification algorithm (Algorithm 2) is able to use finite samples on D θρ to partition the parameter space Θ into Θ + , Θ - , and Θ U according to the predicted ranges of ρ. Given a probability threshold 1 - α, trajectories with parameters in Θ + (or Θ - ) are guaranteed to satisfy (or violate) the specification ϕ with probability greater than 1 - α (Theorem 8 and Corollary 9), while Θ U contains the remaining portion Θ \{Θ + Θ - } where a conclusion cannot be made before the algorithm reaches a user-provided resolution on the partitions. Related works Verifying learning-enabled CPS has at- tracted significant attention in recent years. Each technique focuses on different models of the CPS and provided differ- ent levels of guarantees. Techniques have been developed to look at the accurate models of the neural networks (NN) or NN-controlled dy- namical systems to obtain safety guarantees, using barrier functions [31], [36], reachability analysis [35], [15], [16], [11], Satisfiability Modulo Theory (SMT) based methods or mixed-integer linear program (MILP) optimizer [18], [33], [8]. Most of these methods provide deterministic guarantees but face scalability problem and have additional restrictions on the class of models they can handle (e.g. ReLU activation functions). Our method treats learning-enabled CPS as black- boxes that can produce simulations given fixed parameters and therefore can be used on a very broader class of CPS. Moreover, we are only reasoning over the joint distribution of arXiv:2004.00279v1 [eess.SY] 1 Apr 2020

Upload: others

Post on 29-Oct-2020

2 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Probabilistic Verification of Learning-Enabled Cyber ...tems [13], [10]. The major difference between our work with respect to SMC techniques and PAC-learning techniques is that our

Probabilistic Verification of Learning-Enabled Cyber-Physical Systemsusing Conformal Regression

Chuchu Fan, Xin Qin and Jyotirmoy Deshmukh

Abstract— The use of machine learning components hasposed significant challenges for the verification of cyber-physicalsystems due to its complexity, nonlinearity, and large spaceof parameters. In this work, we propose a novel probabilisticverification framework for learning-enabled CPS which cansearch over the entire (infinite) space of parameters, to figureout the ones that lead to satisfaction or violation of specificationthat are captured by Signal Temporal Logic (STL) formulas.Our technique is based on conformal regression, a techniquefor constructing prediction intervals with marginal coverageguarantees using finite samples, without making assumptions onthe distribution and regression model. Our verification frame-work, using conformal regression, can predict the quantitativesatisfaction values of the system’s trajectories over differentsets of the parameters and use those values to quantify howwell/bad the system with the parameters can satisfy/violate thegiven STL property. We use three case studies of learning-enabled CPS applications to demonstrate that our techniquecan be successfully applied to partition the parameter spaceand provide the needed level of assurance.

I. INTRODUCTION

There is pervasive interest in using machine learning com-ponents such as advanced driving assist systems (ADAS),self-driving controllers, and neural network-based controllersfor robotics. As these are safety-critical cyber-physical sys-tems, there is urgent need for effective and sound verificationalgorithms [7], [16], [18], [33], [35], [8], [15], [40]. Suchsystems have several uncertainties arising from exogenuousinputs from the environment, initial system configurations,and other design parameters. We call the space of theseuncertainties the parameter space, and would like to useverification techniques to find the range of parameters wherewe can assure that the learning-enabled CPS exhibits safebehavior with high confidence.

Such systems are unfortunately not well-suited to cur-rent verification and testing methods that either conductexhaustive state-space search and are therefore face the curseof dimensionality [35], [8], or use best-effort approacheswithout explicit coverage guarantees [14], [26], [39]. Inthis paper, we fill this gap by developing a verificationframework that can search over the entire parameter spaceΘ of the system to find the set of parameters that lead tothe satisfaction or violation of given properties, and provideprobabilistic coverage guarantees.

We treat learning-enabled CPS as black-box dynamicalmodels for which we only have executable access: once theparameter θ ∈ Θ is fixed, we have a deterministic trajectory

Chuchu Fan is with the Department of Computing and Mathemati-cal Sciences, California Institute of Technology, Los Angeles CA, [email protected]

Xin Qin and Jyotirmoy Deshmukh are with the Department of Com-puter Science, University of Southern California, Los Angeles CA, USA{xinqin,jyotirmoy.deshmukh}@usc.edu

ξθ which represents how the state of the system evolves.We consider system property that are specified using Signaltemporal logic (STL) [6], which a formalism that enablesspecifications over real-valued continuous trajectories andhave been widely applied to rigorously specify requirementsin CPS [40]. STL is naturally equipped with a quantitativesemantics called robust satisfaction value (RSV) ρ, whichis a function mapping a property ϕ and a trajectory ξ toa real number ρ(ϕ, ξ). A positive value suggests that ξ(t)satisfies ϕ and a negative value indicates that ξ(t) violatesof ϕ. Moreover, the larger (smaller) the RSV is, more easilyξ satisfies (violates) the specification ϕ.

Suppose the parameters follow a probabilistic distributionDθ over Θ. Given an STL formula ϕ, θ and the correspond-ing RSV ρ(ϕ, ξθ) follow a joint distribution Dθ,ρ. Althoughwe do not know the concrete form of Dθ,ρ, we can alwayssample from this distribution by sampling θ ∼ Dθ firstthen compute the corresponding RSV. Then we use confor-mal regression [22], [21], [29], a generic tool to constructdistribution-free prediction sets, over the joint distributionDθρ to predict the range of the RSV ρ for different sets ofparameters. Conformal regression techniques can constructprediction intervals that attain valid coverage in finite sam-ples, without making assumptions on the distribution and theunderlying regression algorithm. Therefore, our verificationalgorithm (Algorithm 2) is able to use finite samples on Dθρto partition the parameter space Θ into Θ+, Θ−, and ΘU

according to the predicted ranges of ρ. Given a probabilitythreshold 1−α, trajectories with parameters in Θ+ (or Θ−)are guaranteed to satisfy (or violate) the specification ϕ withprobability greater than 1−α (Theorem 8 and Corollary 9),while ΘU contains the remaining portion Θ \ {Θ+ ∪ Θ−}where a conclusion cannot be made before the algorithmreaches a user-provided resolution on the partitions.

Related works Verifying learning-enabled CPS has at-tracted significant attention in recent years. Each techniquefocuses on different models of the CPS and provided differ-ent levels of guarantees.

Techniques have been developed to look at the accuratemodels of the neural networks (NN) or NN-controlled dy-namical systems to obtain safety guarantees, using barrierfunctions [31], [36], reachability analysis [35], [15], [16],[11], Satisfiability Modulo Theory (SMT) based methods ormixed-integer linear program (MILP) optimizer [18], [33],[8]. Most of these methods provide deterministic guaranteesbut face scalability problem and have additional restrictionson the class of models they can handle (e.g. ReLU activationfunctions). Our method treats learning-enabled CPS as black-boxes that can produce simulations given fixed parametersand therefore can be used on a very broader class of CPS.Moreover, we are only reasoning over the joint distribution of

arX

iv:2

004.

0027

9v1

[ee

ss.S

Y]

1 A

pr 2

020

Page 2: Probabilistic Verification of Learning-Enabled Cyber ...tems [13], [10]. The major difference between our work with respect to SMC techniques and PAC-learning techniques is that our

the parameters and corresponding robust value with respectto STL properties. Therefore our algorithm’s complexity isindependent of the model complexity (except for gettingsimulations from those models) and can scale to complexreal-world learning-enabled CPS.

Methods based on Statiscial Model Checking (SMC) [19],[20], [41] can overcome the hurdles like scalability andnonlinearity and provide probabilistic guarantees [40], [30],[38], [4], [1]. These methods are based on stochasticalinference methods like sequential probability ratio tests [19],[30], [32], [4], Bayesian statistics [41], and Clopper-Pearsonbounds [40]. Another line of works use Probably Approx-imately Correct (PAC) learning theory to give probabilisticbounds for Markov decision processes and black-box sys-tems [13], [10].

The major difference between our work with respect toSMC techniques and PAC-learning techniques is that ourmethod can provide the needed level of probabilistic guaran-tees with any number of samples. The guarantee provided bymethods based on SMC and PAC, instead, is a function of thenumber of samples. This is because we build a guaranteedregression model from the system parameters to the robustsatisfaction value of the corresponding STL properties. If theregression model is poor with less samples, using the cali-bration step in conformal regression, the predicted (therebypoor) interval can still have the same level of guarantee.The main advantage of using conformal regression is thatthe tradeoff is between how well the regression model fitsthe data and the width of the interval for which we havehigh-confidence property satisfaction, and not the level ofthe guarantee itself.

II. PRELIMINARIES AND BACKGROUND

A. System, Trajectories, and SimulatorsA trajectory of the system defines how the value of a set of

continuous variables evolve over time. Given X ⊆ Rn, whichis the set of possible values for the continuous variables, atrajectory ξ over X is defined as a function ξ : dom → X ,where dom is the time domain of evolution, and it is either[0, T ] for some T > 0, or it is [0,∞). The domain of ξis referred to as ξ.dom. The state of the system along thetrajectory at time t ∈ ξ.dom is ξ(t).

In this paper, the system is defined as a set of parameter-ized deterministic trajectories Ξ := {ξθ}θ∈Θ, where Θ is theparameters space, which include uncertainties from the input,initial configuration, and other parameters of the systemmodel. Once θ ∈ Θ is fixed, there will be a deterministictrajectory ξθ ∈ Ξ. We also assume that θ ∈ Θ is a randomvariable and follows a distribution Dθ with probability den-sity function (PDF) p(θ) and ∀θ /∈ Θ, p(θ) = 0. If we onlywish to draw samples from a subset S ⊆ Θ (by droppingsamples from Θ \ S), the corresponding distribution of thesamples is denoted by Dθ ↓ S and follows the PDF of

p′(θ) =

{p(θ)∫

τ∈S p(τ)dτif θ ∈ S

0 otherwise.(1)

Instead of a closed form description of ξθ such as ODEs,we assume that there is a simulator that can generate sam-pled data points on individual trajectories. We will developtechniques that avoid over-reliance on the models generating

the trajectories and instead work with sampled data of ξθ(·)generated from the simulators. Of course, in order to obtainsafety guarantees we will need to make assumptions aboutthe underlying system generating the data.

Definition 1. A simulator for a (deterministic) set Ξ oftrajectories is a function (or a program) sim that takesas input a parameter θ ∈ Θ, and a finite sequence oftime points t1, . . . , tk, and returns a sequence of statessim(θ, t1), . . . , sim(θ, tk) such that for each i ∈ {1, . . . , k},sim(θ, ti) = ξθ(ti).

The actual trajectory ξθ can be reconstructed using someform of interpolation of these discrete-time simulation datasim(θ, t1), . . . , sim(θ, tk). For the rest of paper, without lossof generality, we assume that the time sequence is fixed foreach system and let IplSim(θ) to denote the reconstructedtrajectory by interpolating the simulation data. For simplicity,we assume that the IplSim(θ) is perfect (i.e., equals to ξθ).In other words, the formal probabilistic guarantees is validonly on IplSim(θ).

B. Signal Temporal LogicSignal Temporal Logic (STL) [23] is a powerful formalism

that can rigorously specify the requirements of systems. STLformulas are defined over predicates of the form f(ξ(·)) ≥ 0,where ξ(·) is a trajectory and f : Rn → R is a function. STLformulas are written using the following grammar:

I := (a, b) | (a, b] | [a, b) | [a, b]

ϕ,ψ := true | f(ξ(·)) > 0 | ¬ϕ | ϕ ∧ ψ | ϕ ∨ ψ| ♦Iϕ | �Iϕ | ϕ UI ψ

where f(ξ(·)) ≥ 0 is a predicate, and the logical operators(¬,∧,∨) have their typical meanings. In addition, ♦ (eventu-ally), � (always), and U (until) are temporal operators. Thetemporal operators have an associated time interval I where0 ≤ a < b, a, b ∈ R. Given t ∈ R≥0 and I = [a, b] (or otheropen intervals I), t+ I := [t+ a, t+ b].

Given a trajectory ξ and a time t, we use (ξ, t) |= ϕto denote that ξ satisfies ϕ at time t. STL is equippedwith a quantitative semantics as the robust satisfaction value(RSV), which is a function mapping a property ϕ and atrajectory ξ(t) to a real number [9], [5]. While there aremany competing definitions for robust satisfaction value inthe literature [2], [28], [17], we use the original definitionsfrom [5] in this paper. The RSV can provide a quantitativemetric on how good (bad) a trajectory satisfies (violates) thecorresponding STL formula ϕ.

Definition 2. The robust satisfaction value is a function ρmapping ϕ, the trajectory ξ, and a time t ∈ ξ.dom as follows:

ρ(f(ξ) > 0, ξ, t) = f(ξ(t))ρ(¬ϕ, ξ, t) = −ρ(ϕ, ξ, t)

ρ(ϕ ∧ ψ, ξ, t) = min(ρ(ϕ, ξ, t), ρ(ψ, ξ, t))ρ(ϕ ∨ ψ, ξ, t) = max(ρ(ϕ, ξ, t), ρ(ψ, ξ, t))ρ(�Iϕ, ξ, t) = inf

t′∈t+Iρ(ϕ, ξ, t′)

ρ(♦Iϕ, ξ, t) = supt′∈t+I

ρ(ϕ, ξ, t′)

ρ(ϕ UI ψ) = supt1∈t+I

min(ρ(ψ, ξ, t1), inft2∈[t,t1]

ρ(ϕ, ξ, t2))

The translation from the quantitative semantics defined byRSV to the Boolean satisfaction semantics is that a trajectory

Page 3: Probabilistic Verification of Learning-Enabled Cyber ...tems [13], [10]. The major difference between our work with respect to SMC techniques and PAC-learning techniques is that our

(ξ, t) |= ϕ iff the RSV ρ(ϕ, ξ, t) ≥ 0. When t = 0, we dropt and simply say that ξ |= ϕ iff ρ(ϕ, ξ) = ρ(ϕ, ξ, 0) ≥ 0.

Example 3. Consider two STL formulas for a one-dimensional trajectory x(t): ϕ1 = �[0,10](x(t) < 0.5) meansfor any time t ∈ [0, 10], the value of the trajectory x(t)should be less than 0.5; ϕ2 = ♦[0,2]�[0,8](‖x(t)‖ < 0.3)means from some time within the first 2 time units, x(t)settles in the region [−0.3, 0.3] for 8 time units. Figure 1shows example trajectories with respect to ϕ1 (Left) andϕ2 (Right). The trajectories are the x1(t) from a inversevanderpol dynamics x1 = −x2, x2 = 4(x2

1 − 1)x2 + x1. Inboth figures, the top red trajectories violate the correspondingSTL formulas, while the below blue trajectories satisfy theSTL formulas.

å

!

"

(a) ϕ1 = �[0,10](x(t) < 0.5)

#$ #%

"

(b) ϕ2 = ♦[0,2]�[0,8](‖x(t)‖ < 0.3)

Fig. 1: Example of trajectories satisfying (blue) and violating(red) the STL formulas.

C. Conformal Regression

Conformal Regression [21], [22] provides a framework toquantify the accuracy of the predictive inference in regressionusing conformal inference [37]. Conformal regression canprovide valid coverage in finite samples, without makingassumptions on the distribution.

Consider i.i.d. regression data Z1, · · · , Zm drawn froman arbitrary joint DXY , where each Zi = (Xi, Yi) is arandom variable in Rn × R, comprised of a n-dimensionalfeature vectors Xi and a response variable Yi. Conformalregression problem is to predict a new response Ym+1 froma new feature value Xm+1, with no assumptions on DXY .Formally, given a positive value α ∈ (0, 1), conformalregression techniques can construct a prediction band C ⊆Rn × R based on Z1, · · · , Zn with the property that

P(Ym+1 ∈ C(Xm+1)) ≥ 1− α, (2)

where the probability is taken over the m + 1 i.i.d. drawsZ1, · · · , Zm, Zm+1 ∼ DXY , and for a point x ∈ Rn wedenote C(x) = {y ∈ R : (x, y) ∈ C}. Such α is called themiscoverage level and 1− α is the probability threshold.

Letµ(x) = E(Y | X = x), x ∈ Rn

denote the regression function. The regression problem is toestimate such conditional mean of the test response Ym+1

given the test feature Xm+1 = x. Common regressionmethods use a regression model g(x, η) and minimize thesum of squared residuals of such model on the m trainingregression data Z1, · · · , Zm, where η are the parameters of

the regression model. The estimator for u is given by

µ(x) = g(x, η),

where η = argminη1m

∑mi=1(Yi − g(Xi, η))2 + R(η)

and R(η) is a potential regularizer. Common regressionmodel g(x, η) includes linear, polynomial, and neural net-works [12], [3].

We use a split conformal methods from [21] to constructprediction intervals that satisfy the finite-sample coverageguarantees as in Equation (2). The procedure is describedin Algorithm 1 as a function ConfInt which takes asinput the i.i.d. training data {(Xi, Yi)}mi=1, miscoverage levelα and any regression algorithm Reg. Algorithm 1 beginsby splitting the training data into two equal-sized disjointsubsets. Then a regression estimator µ is fit to the trainingset {(Xi, Yi)} : i ∈ I1) using the regression algorithm Reg(Line 2). Then the algorithm compute the absolute residualsRis on the calibration set {(Xi, Yi)} : i ∈ I2) (Line 3). For agiven miscoverage level α, the algorithm rank {Ri : i ∈ I2}and take the d(n/2+1)(1−α)eth one as the confidence ranged. Finally, it can be proved that the prediction interval at anew point Xm+1 is given by such µ and d as in Theorem 4.

Algorithm 1: Conformal regression algorithmConfInt({(Xi, Yi)}mi=1, α, Reg)

input : Data {(Xi, Yi)}mi=1, miscoverage level α,regression algorithm Reg

output: Regression estimator µ, confidence range d1 Randomly split {1, · · · , n} into two equal-sized subsetsIi, I2;

2 µ = Reg((Xi, Yi) : i ∈ I1) ;3 Ri = |Yi − µ(Xi)|, i ∈ I2 ;4 d = the kth smallest value in {Ri : i ∈ I2}, where

k = d(n/2 + 1)(1− α)e ;5 return µ, d

Theorem 4 (Theorem 2.1 in [21]). If (Xi, Yi), i = 1, · · · ,mare i.i.d., then for an new i.i.d. draw (Xm+1, Ym+1), usingµ and d constructed in Algorithm 1, we have that

P(Ym+1 ∈ [µ(Xm+1)− d, µ(Xm+1) + d]) ≥ 1− α.

Moreover, if we assume additionally that the residuals {Ri :i ∈ I2} have a continuous joint distribution, then

P(Ym+1 ∈ [µ(Xm+1)−d, µ(Xm+1)+d]) ≤ 1−α+2

m+ 2.

Generally speaking, as we improve our estimator Regof the underlying regression function µ, the resulting con-formal prediction interval decreases in length. Intuitively,this happens because a more accurate µ leads to smallerresiduals, and conformal intervals are essentially defined bythe quantiles of the (augmented) residual distribution.

Note that Theorem 4 assert marginal coverage guarantees,which should be distinguished with the conditional coverageguarantee P(Ym+1 ∈ C(x) | Xm+1 = x) ≥ 1 − α for allx ∈ Rn. The latter one is a a much stronger property andhard to be achieved without assumptions on DXY .

Page 4: Probabilistic Verification of Learning-Enabled Cyber ...tems [13], [10]. The major difference between our work with respect to SMC techniques and PAC-learning techniques is that our

III. PARAMETER SEARCHING AND PARTITION WITHPROBABILISTIC COVERAGE GUARANTEES

In this section, we discuss how to search and partition theparameter space Θ with the goal of satisfying or falsifyingthe given property ϕ over system behaviors. To be specific,we want to partition Θ to subsets Θ+, Θ− and ΘU , such thatevery possible parameter in Θ+ (or Θ−) leads to satisfaction(or falsification) the given property ϕ with probabilisticguarantees, and ΘU contains the remaining part that is notdecided (ideally ΘU should be as close to an empty set aspossible).

Assume that we know the distribution Dθ over Θ. Letθ ∈ Θ be the feature vector and the robust satisfactionvalue ρ(ϕ, ξθ) be the response, where ξθ is the deterministictrajectory once θ is fixed. The key idea of our approachis to construct a regression model for the joint distributionDθρ of θ and ρ(ϕ, ξθ). We use conformal regression (Algo-rithm 1) to train and calibrate such regression estimator, andprovide prediction intervals that have probabilistic coverageguarantees (Lemma 5). If the predicted intervals for the RSVρ(ϕ, ξθ) of the entire parameter space θ ∈ Θ contain onlypositive (or negative) values, we conclude that Θ+ = Θ (orΘ− = Θ). Otherwise we partition Θ and repeat the processiteratively until the subsets belong either to Θ+ or Θ−, oruntil the subset is smaller than a user-provided threshold(Algorithm 2). We prove that new samples from Θ+ (or Θ−)indeed have the probabilistic guarantee on the correspondingsimulation trajectory satisfying (violating) ϕ.

A. Conformal Regression of Parameters to RSV

Consider a system with a distribution Dθ of the pa-rameters over Θ. Given an STL formula ϕ to describe adesired property of the system, then for each IplSim(θ),we can use Definition 2 to compute the correspondingRSV ρ(ϕ, IplSim(θ)). θ and ρ(ϕ, IplSim(θ)) follow a jointdistribution Dθ,ρ. We want to find a constant predictioninterval [`, u], such that for any new sample from Dθρ, theprobability of the RSV fall in [`, u] is bounded.

We first draw m i.i.d samples θ1, · · · , θm from Dθ andcompute the RSV ρi = ρ(ϕ, IplSim(θi) for each interpo-lated simulation with parameter θi. Let Algorithm 1 take thepairs (θi, ρi), i = 1, · · · ,m ∼ Dθρ, along with a regressionalgorithm Reg and miscoverage level α, and return anestimator µ and confidence range d. In our implementation inSection IV we choose both polynomial and neural networksas the regression model for Reg.

From Theorem 4 we know that a new i.i.d. sample fromDθρ satisfies P(ρm+1 ∈ [µ(θm+1)−d, µ(θm+1)+d]) ≥ 1−α.θm+1 can only be drawn from Θ (as defined in Section II-A, p(θ) = 0 if θ /∈ Θ). Take the minimum value of µ(θ)over Θ as v∗min, we know that µ(θm+1) ≥ v∗min. Similarly,µ(θm+1) ≤ v∗max where v∗max is the maximum value of µ(θ)over Θ. Therefore, we know that with probability of at least1 − α, ρm+1 is contained in the constant interval [v∗min −d, v∗max + d]. This gives us the following lemma:

Lemma 5. Assume that (θi, ρi), i = 1, · · · ,m are i.i.d.samples draw from a joint distribution Dθ,ρ of θ ∈ Θ and ρ.Given a regression algorithm Reg and miscoverage level α,and µ and d are returned by Algorithm 1, then for an new

i.i.d. draw (θm+1, ρm+1) ∼ Dθρ,

P(ρm+1 ∈ [v∗min − d, v∗max + d]) ≥ 1− α, (3)

where v∗min = minθ∈Θ

µ(θ) and v∗max = maxθ∈Θ

µ(θ).

Note that although Equation (3) does not depend on thenumber of training (and calibration ) samples m for theconformal regression algorithm, Theorem 4 also indicatesthat the more samples we have for training, the more accuratethe probabilistic threshold is. In the case where µ(θ) isa non-convex function and the chosen over-approximationalgorithms cannot compute the perfect optimal value v∗min (orv∗max), but can only give conservative estimates of the opti-mal value, we can update the predicted interval in Lemma 5as follows.

Corollary 6. In Lemma 5, if v∗min and v∗max can be esti-mated with vmin and vmax respectively using optimizationalgorithms, and if v∗min > vmin and v∗max < vmax, then foran new i.i.d. draw (θm+1, ρm+1) ∼ Dθρ,

P (ρm+1 ∈ [vmin − d, vmax + d]) ≥ (1− α). (4)

The vmin and vmax in Corollary 6 can be computed usingoptimization solvers, SMT sovlers, or reachablity tools forneural networks [35], [8].

B. Overall Algorithm

In this section, we present the overall algorithm, whichutilizes Lemma 5 (or Corollary 6) to partition the parameterspace so the predicted RSV for each subset has the samesign. From Section III-A, we can compute a constant pre-diction interval [`, u], which provides probabilistic coverageguarantee for new samples from Dθ,ρ with θ ∈ Θ. That is, fornew samples (θm+1, ρm+1 = ρ(ϕ, IplSim(θm+1)) ∼ Dθρ,P(ρm+1 ∈ [`, u]) ≥ 1 − α. If such interval contains onlynon-negative values (` ≥ 0), from Section II-B it means thatthat the (interpolated simulation) trajectory satisfies the givenproperty ϕ : IplSim(θm+1) |= ϕ. Similarly, if such intervalonly contains non-positive values (u ≤ 0), we can say thatIplSim(θm+1) 6|= ϕ. This gives us the following lemma:

Lemma 7. Following Lemma 5, if v∗min − d ≥ 0, thenP(IplSim(θm+1) |= ϕ) ≥ 1 − α; If v∗max + d ≤ 0, thenP(IplSim(θm+1) |= ϕ) < α.

If v∗min−d < 0 and v∗max+d > 0, we cannot conclude andwill have to partition Θ and repeat the interval computationfor each subset after the partition. Note that the probabilityin Lemma 5 (and Theorem 4 ) is marginal, being taken overall the i.i.d. samples {θi, ρi}m+1

i=1 from Dθρ. Therefore, whenwe work on each subset S ⊆ Θ after the partitions, we willhave to restrict θ to be in S (according to Equation (1)) toensure that the Lemma 5 is valid. We abuse the notationand denote the joint distribution of θ and ρ(ϕ, ξθ) when werestrict θ to be sampled from S ⊆ Θ by Dθρ ↓ S.

Algorithm 2 implements the above procedure. It searchesover the entire parameter space Θ and partition it to Θ+, Θ−,and ΘU , along with the prediction intervals for RSV. The newsamples from Dθρ ↓ S where (S, ·) ∈ Θ+ (or Dθρ ↓ S where(S, ·)Θ−) have probabilistic guarantee that the correspondingtrajectory satisfies (violates) ϕ (At Line 10 and 12). When

Page 5: Probabilistic Verification of Learning-Enabled Cyber ...tems [13], [10]. The major difference between our work with respect to SMC techniques and PAC-learning techniques is that our

Algorithm 2: Parameter space partition with respect toSTL formulas using conformal regression.

input : Parameter space Θ and correspondingdistribution Dθ, simulator sim and interpolationmethod to provide IplSim, miscoverage levelα, regression algorithm Reg, an STL formulaϕ, the minimum radius of subsets allowed δmin

output: Parameter set Θ+ that lead to satisfaction of ϕ,Θ− that lead to violation of ϕ, and the restparameter set ΘU that is undecided

1 Θ+,Θ−,ΘU ← ∅, Θr ← {Θ};2 while Θr 6= ∅ do3 S ← Pop(Θr) ;4 θ1, · · · , θm ← IID Sample(Dθ ↓ S) ;5 for i = 1, · · · ,m do6 ρi ← ρ(ϕ, IplSim(θi)) ;7 µ, d← ConfInt({(θi, ρi)}mi=1, α, Reg);8 vmax ← maxθ∈S µ(θ), vmin ← minθ∈S µ(θ);9 if vmin − d ≥ 0 then

10 Θ+ ← Θ+ ∪ (S, [vmin − d, vmax + d]) ;11 else if vmax + d ≤ 0 then12 Θ− ← Θ− ∪ (S, [vmin − d, vmax + d]) ;13 else if Radius(S) < δmin then14 ΘU ← ΘU ∪ (S, [vmin − d, vmax + d]) ;15 else16 Θr.Push(Partition(S)) ;17 return Θ+,Θ−,ΘU ;

Algorithm 2 cannot decide whether a subset S belongs toΘ+ or Θ− because the corresponding predicted RSV intervalcontains both negative and positive, our algorithm workswith any heuristic partitioning algorithm to further partitionS (Line 16). In our implementation, in order to keep thenumber of subsets to be explored bounded, we randomly picka dimension in the parameter space, and split the parameterspace into two equal subsets along that dimension. Notethat the partitioning can be accelerated by using parallelcomputation, but we leave that for future exploration.

To ensure that the algorithm can terminate in finite time,we also require that the subsets have radius no less thanδmin. If for a subset S with radius less than δmin but thecorresponding prediction interval contains both positive andnegative value, we will assign S to ΘU (Line 14). Foreach subset S, Algorithm 2 also gives the correspondingprediction interval, which indicates how good (or bad) thetrajectories satisfy (or violate) ϕ.

Theorem 8. In Algorithm 1, for any pair (S, [vmin −d, vmax + d]) ∈ Θ+ and for an new i.i.d pair(θm+1, ρ(ϕ, IplSim(θm+1))) drawn from ρθρ ↓ S,

P(IplSim(θm+1) |= ϕ) ≥ 1− α,

Similarly, for any pair (S, [vmin−d, vmax+d]) ∈ Θ− and foran new i.i.d pair (θm+1, ρ(ϕ, IplSim(θm+1))) drawn fromρθρ ↓ S,

P(IplSim(θm+1) |= ϕ) < α.

Theorem 8 directly follows Lemma 5 and Lemma 7.Theorem 8 provides probabilistic guarantees separately foreach subset. If we combine the subsets in Θ+ and Θ−, wealso have a similar level of guarantee.

Corollary 9. In Algorithm 1, for an new i.i.d pair(θm+1, ρ(ϕ, IplSim(θm+1))) drawn from Dθρ ↓

⋃(S,·)∈Θ+

S,

P(IplSim(θm+1) |= ϕ) ≥ 1− α,

Similarly, for an new i.i.d pair (θm+1, ρ(ϕ, IplSim(θm+1)))drawn from Dθρ ↓

⋃(S,·)∈Θ−

S,

P(IplSim(θm+1) |= ϕ) < α.

Corollary 9 follows Theorem 8 using total probabilitytheorem. Figure 2 shows the performance of Algorithm 2 on

x0.5 0.0 0.5

y0.50.0

0.5

0.10.20.30.4

(a) Training data points

x1.0 0.5 0.0 0.5 1.0

y1.0

0.50.0

0.51.0

0.30.20.1

0.00.10.20.3

(b) Polynomial regressionresult with order three

x1.0 0.5 0.0 0.5 1.0

y1.0

0.50.0

0.51.0

0.10.00.10.20.3

(c) NN regression result (d) Partition result: greenregions are Θ+, red regionsare Θ− with yellow points ascounter-examples, and greyare ΘU .

Fig. 2: Regression and Algorithm 2 on inverse vanderpolwith respect to ε2 in Example 3.

the inverse vanderpol system over a initial set of [−0.8, 0.8]×[−0.8, 0.8] with respect to ϕ2 in Example 3. Figure 2a is theinitial training data before any partition happens, Figure 2band 2c shows different regression algorithms on the trainingdata and Figure 2d shows the partition results.

IV. EXPERIMENTAL EVALUATION

In this section, we benchmark various aspects of ourproposed algorithm through three different case studies oflearning-enabled CPS applications. In these applications, weassume that there is a learning-enabled component (LEC)being used to control the underlying physical system. Twoof the LECs are controllers trained using deep reinforcementlearning (RL) [34], [24], [25] using the RL toolbox inMatlab R©/Simulink R©. The third controller is a nonlinear au-toregressive moving average (NARMA) neural network con-troller, trained using the neural network toolbox in Matlab R©.In a small deviation from Algorithm 2, we call a region

Page 6: Probabilistic Verification of Learning-Enabled Cyber ...tems [13], [10]. The major difference between our work with respect to SMC techniques and PAC-learning techniques is that our

unsafe if it contains a concrete counterexample. This is inaddition to the case where the upper bound of the predictedinterval for the RSVs is negative. We call a region unknownif its predicted interval contains 0 and it has no knowncounterexamples.

A. Mountain Car

Problem Description. The mountain car problem is a classi-cal problem in the RL literature. It models an under-poweredcar attempting to drive up a hill. The hill is too steep for thecar, and it can succeed only by building potential energygoing in the opposite direction and using its momentum toclimb the hill. We obtained this example from [40]. In thispaper, the authors describe the dynamics of the mountaincar using x = v, v = −mg cos(3x) + F

mu − µx, where x,v, m = 0.2 kg, g = 9.8ms−2, f = 0.2 N, and µ = 0.5are respectively the position (in meters), velocity (in ms−1),mass, acceleration due to gravity, force produced by the carengine, and the friction factor respectively. The input fromthe controller is denoted by u(t). The parameter space isdefined by the initial position xinit and velocity vinit of thecar. In Fig. 3, we show the results from [40], where theauthors used statistical model checking approaches to obtainconfidence bounds on the system reaching its goals. Here,the green dots indicate initial conditions for which the RLcontroller was able to power the car to the top of the hill, andthe red dots indicate counterexamples. The purpose of ourapproach is to identify regions of space where we can iden-tify regions satisfying or violating the property (of reachingthe goal). We choose two sub-regions (shown in Fig. 3) fromthe larger set indicated in 3 in such a way that the number ofcounterexamples in those regions was small (and thus hardto find). The first region Θ1 is defined as (xinit, vinit) ∈[−0.7,−0.5]×[0.2, 0.5], and the second region Θ2 is definedas (xinit, vinit) ∈ [−0.5,−0.2] × [−0.6,−0.2]. We considerthe initial parameter setting satisfying if it satisfies thefollowing STL formula: ϕMC = ♦[0,10](x(t) > 0.45).

Fig. 3: Regions of interest that we analyze using Algo. 2

Results. In Fig. 4, we show the results of applying Algo-rithm 2 to regions Θ1 and Θ2. In the algorithm, we useda neural network based regression to approximate the mapfrom parameters to the robustness values of property ϕMC.The neural network was a fitnet neural network, which isa feedforward neural network with two hidden layers with 10neurons each. The activation function used was tansig —a custom implementation of the tanh function in Matlab R©.

(a) Results of runningAlgorithm 2 on Θ1

(b) Results of runningAlgorithm 2 on Θ2

Fig. 4: Mountain Car Results

The algorithm is able to partition Θ1 and Θ2 into severalsafe regions (shown in green) and unsafe regions (shown inred). The algorithm does not refine the safe regions further,but refines the regions for which the predicted RSV intervalcontains both positive and negative RSV, or if the regionhas concrete counterexamples. For both cases, the algorithmhalts upon reaching a user-provided δ, which in this case isa region smaller than 1

100 times the size of the Θis. Furthercomputational benchmarks are indicated in Table I.

B. NARMA controller for magnetic levitation

Problem Description. The magnetic levitation model inSimulink R© captures the control of the magnet in a powertransformer with dynamics y = −g+ αi2

My −βM y, where y is

distance from magnet to electromagnet, i is current flowingin the electromagnet, g = 9.8ms−2, α, β, M are respectivelythe gravitational acceleration of earth, field strength, frictioncoefficient, and mass of the magnet. We used a versionof this example provided in [40]. Here, the authors usea NARMA-L2 (nonlinear auto regressive moving average)neural network controller to control the magnet. The the NN-controller contains seven hidden layers, five nodes each layerand uses tansig activation functions. We are interested inthe same property (ϕNARMA = ♦[5,5](IAE < 1.0)) as [40],which captures the integrated absolute error (IAE ) in theposition of the magnet with respect to an externally suppliedreference yref .Results. In the region Θ3 = {yinit | yinit ∈ [0.5, 4]}. Thesetting in [40] is to apply statistical model checking (SMC)ϕNARMA assuming that the external reference is a normallydistributed quantity with mean µ = 2 and standard deviationof σ = 0.7. We conduct a differente experiment, wherewe consider a uniformly distributed yref over Θ3. In [40],the authors show the system satisfies ϕNARMA with highprobability (under a normally distributed yref ). Our algorithmconcludes that we get a 95%-confident positive RSV intervalonly in the region [0.5, 1.4]. In the region [1.4, 2.7], whichhas a big overlap with the region [µ − σ, µ + σ] from[40], our tool reports that the 95%-confidence RSV intervalhas a negative upper bound. As the two techniques usedare dramatically different (SMC vs. conformal regression),we investigated this apparent discrepancy a bit further, andfound that the relationship between the parameter value yref

and the robustness value is highly nonlinear and nonmono-tonic. Under an unfortunate sampling regime, SMC couldaltogether miss violations. This demonstrations that using acombination of conformal regression and prediction can give

Page 7: Probabilistic Verification of Learning-Enabled Cyber ...tems [13], [10]. The major difference between our work with respect to SMC techniques and PAC-learning techniques is that our

better guarantees than SMC.

C. Reinforcement Learning Lane Keep Assist

Problem Description. Lane-keep assist (LKA) is an auto-mated driver assistance technique used in semi-autonomousvehicles to keep the ego vehicle traveling along the centerlineof a lane. The recent reinforcement learning toolbox fromMatlab R© introduces a Deep Q Network (DQN)-based rein-forcement learning agent that seeks to keep the ego vehiclecentered. The inputs to the RL agent are the lateral deviatione1, relative yaw angle (i.e. yaw error) e2, their derivativesand their integrals. Specifics of the DQN-based RL agentand its training can be found in [27]. The parameter spacefor this model is the initial values for e1 and e2. We areinterested in the signals corresponding to the lateral deviationand yaw error. We are interested in checking the control-theoretic properties such as overshoot/undershoot bounds andthe settling time for these signals. In this experiment, weconsider two properties characterizing bounds on e2 andsettling time for e1: ϕLKA,settle = �[2,15](|e1| < 0.025) andϕLKA,bounds = �[0,15](e2 < 0.6).Results. Our first experiment used initial values of e1 and e2

that were both in the interval [−0.1, 0.1]. For this region, wewere able to prove that both the above properties are satisfiedwith 95%-confidence over the region without requiring anyfurther region splitting. For the second experiment, we usedthe following region: e1 ∈ [−0.2,−0.1], and e2 ∈ [0.1, 0.2].For this region, there were sub-regions that were shownsafe, others that were unsafe, and a few regions that wereunknown (as they did not contain counterexamples, but ouralgorithm reached the user-provided δ). However, what wasinteresting was the non-monotonicity in the parameter space.For example, for the property ϕLKA,bounds, the sub-regions[−0.2,−0.15]× [0.1, 0.175] and [−0.125,−0.1]× [0.1, 0.15]were both safe with 95% confidence, but the intermediateregion [−0.15,−0.125] × [0.1, 0.15] was unknown as thepredicted interval for this region contained both positive andnegative values.

D. DiscussionIn Table I, we provide the runtimes for each component

of our algorithm. We remark that the total times taken areinflated because of time required to plot regression surfaces,which is not essential when benchmarking for runtime effi-ciency. The key point to note is that regression is a relativelycheap step, while black-box optimization (to obtain upperand lower bounds on the RSV intervals) consumes mostof the runtime. In the future, we will investigate fasterconservative techniques to approximate the RSV intervalsusing techniques such as [8], [35]. In the table, we usePoly(2) to denote polynomial regression with a degree 2polynomial, and fitnet(10, 10) to denote a feedforwardneural network with 2 hidden layers with 10 neurons each.We used the Neural Network library from Matlab R© for thisneural network.

In table I, fitnet is a feed forward neural network with twohidden layers, ten nodes at each layer.

V. CONCLUSION AND FUTURE WORK

In this paper, we proposed a verification framework thatcan search the parameter space to find the regions that

lead to satisfaction or violation of given specification withprobabilistic coverage guarantees. There are a couple ofdirections we aim to explore as future work: 1) We useda very basic version of conformal regression in Algorithm 1,which gives a constant confidence range d across all X .Techniques based on quantile regression [29] and locally-weighed conformal [21] can make d a function of X and givemuch shorter prediction intervals. 2) Our current partitionmethod chooses a random dimension of the parameter spaceto split. We plan to investigate more strategic partitionmethods that can conclude faster.

REFERENCES

[1] Houssam Abbas, Bardh Hoxha, Georgios Fainekos, and Koichi Ueda.Robustness-guided temporal logic testing and verification for stochas-tic cyber-physical systems. In The 4th Annual IEEE InternationalConference on Cyber Technology in Automation, Control and Intelli-gent, pages 1–6. IEEE, 2014.

[2] Takumi Akazaki and Ichiro Hasuo. Time Robustness in MTL andExpressivity in Hybrid System Falsification. In Daniel Kroening andCorina S. Puasuareanu, editors, Computer Aided Verification, LectureNotes in Computer Science, pages 356–374. Springer InternationalPublishing, 2015.

[3] Andrew R Barron. Universal approximation bounds for superpositionsof a sigmoidal function. IEEE Transactions on Information theory,39(3):930–945, 1993.

[4] Edmund M Clarke, James R Faeder, Christopher J Langmead,Leonard A Harris, Sumit Kumar Jha, and Axel Legay. Statisticalmodel checking in biolab: Applications to the automated analysisof t-cell receptor signaling pathway. In International Conference onComputational Methods in Systems Biology, pages 231–250. Springer,2008.

[5] Alexandre Donze and Oded Maler. Robust satisfaction of temporallogic over real-valued signals. In International Conference on FormalModeling and Analysis of Timed Systems, pages 92–106. Springer,2010.

[6] Alexandre Donze, Oded Maler, Ezio Bartocci, Dejan Nickovic, RaduGrosu, and Scott Smolka. On temporal logic and signal processing.In International Symposium on Automated Technology for Verificationand Analysis, pages 92–106. Springer, 2012.

[7] Tommaso Dreossi, Daniel J. Fremont, Shromona Ghosh, Edward Kim,Hadi Ravanbakhsh, Marcell Vazquez-Chanlatte, and Sanjit A. Seshia.Verifai: A toolkit for the formal design and analysis of artificialintelligence-based systems. In Isil Dillig and Serdar Tasiran, editors,Computer Aided Verification, pages 432–442, Cham, 2019. SpringerInternational Publishing.

[8] Souradeep Dutta, Susmit Jha, Sriram Sankaranarayanan, and AshishTiwari. Learning and verification of feedback control systems usingfeedforward neural networks. IFAC-PapersOnLine, 51(16):151–156,2018.

[9] Georgios E. Fainekos and George J. Pappas. Robustness of temporallogic specifications for continuous-time signals. Theoretical ComputerScience, 410(42):4262–4291, 2009.

[10] Chuchu Fan, Bolun Qi, Sayan Mitra, and Mahesh Viswanathan. Dryvr:Data-driven verification and compositional reasoning for automotivesystems. In 29th International Conference on Computer AidedVerification (CAV), volume 10426 LNCS, pages 441–461, Heidelberg,Germany, 2017. Springer, Cham.

[11] Mahyar Fazlyab, Alexander Robey, Hamed Hassani, Manfred Morari,and George Pappas. Efficient and accurate estimation of lipschitzconstants for deep neural networks. In Advances in Neural InformationProcessing Systems, pages 11423–11434, 2019.

[12] Jerome Friedman, Trevor Hastie, and Robert Tibshirani. The elementsof statistical learning, volume 1. Springer series in statistics NewYork, 2001.

[13] Jie Fu and Ufuk Topcu. Probably approximately correct mdp learn-ing and control with temporal logic constraints. arXiv preprintarXiv:1404.7073, 2014.

[14] Bardh Hoxha, Hoang Bach, Houssam Abbas, Adel Dokhanchi, Yoshi-hiro Kobayashi, and Georgios Fainekos. Towards formal specificationvisualization for testing and monitoring of cyber-physical systems. InInt. Workshop on Design and Implementation of Formal Tools andSystems. sn, 2014.

[15] Xiaowei Huang, Marta Kwiatkowska, Sen Wang, and Min Wu. Safetyverification of deep neural networks. In International Conference onComputer Aided Verification, pages 3–29. Springer, 2017.

Page 8: Probabilistic Verification of Learning-Enabled Cyber ...tems [13], [10]. The major difference between our work with respect to SMC techniques and PAC-learning techniques is that our

Case Study Total Time taken Simulations Regression Spec.Regression Optimization Total per region Method

Mountain Car 1 1.01 1860 3541 100 fitnet(10, 10) ϕMC

Mountain Car 2 3.53 583 3850 100 fitnet(10, 10) ϕMC

NARMA 0.065 16 1075 50 Poly(2) ϕNARMA

Lane Keep Assist 1 1.04 44.9 178 40 fitnet(10, 10) ϕLKA,bounds

Lane Keep Assist 1 1.14 43.9 186 40 fitnet(10, 10) ϕLKA,settle

Lane Keep Assist 2 0.45 287 4634 50 fitnet(10, 10) ϕLKA,bounds

Lane Keep Assist 2 0.51 324 5167 50 fitnet(10, 10) ϕLKA,settle

TABLE I: Benchmarking the runtimes of individual components of our technique. All times provided above are in seconds.

[16] Radoslav Ivanov, James Weimer, Rajeev Alur, George J. Pappas, andInsup Lee. Verisig: Verifying safety properties of hybrid systems withneural network controllers. In Proceedings of the 22nd ACM Inter-national Conference on Hybrid Systems: Computation and Control,HSCC ’19, page 169–178, New York, NY, USA, 2019. Associationfor Computing Machinery.

[17] Stefan Jaksic, Ezio Bartocci, Radu Grosu, Thang Nguyen, and DejanNickovic. Quantitative monitoring of STL with edit distance. FormalMethods in System Design, 53(1):83–112, August 2018.

[18] Guy Katz, Clark Barrett, David L. Dill, Kyle Julian, and Mykel J.Kochenderfer. Reluplex: An efficient smt solver for verifying deepneural networks. In Rupak Majumdar and Viktor Kuncak, editors,Computer Aided Verification, pages 97–117, Cham, 2017. SpringerInternational Publishing.

[19] Axel Legay, Benoıt Delahaye, and Saddek Bensalem. Statistical modelchecking: An overview. In Runtime Verification, pages 122–135,Berlin, Heidelberg, 2010. Springer Berlin Heidelberg.

[20] Axel Legay and Mahesh Viswanathan. Statistical model checking:challenges and perspectives, 2015.

[21] Jing Lei, Max G’Sell, Alessandro Rinaldo, Ryan J Tibshirani, andLarry Wasserman. Distribution-free predictive inference for regression.Journal of the American Statistical Association, 113(523):1094–1111,2018.

[22] Jing Lei and Larry Wasserman. Distribution-free prediction bands fornon-parametric regression. Journal of the Royal Statistical Society:Series B (Statistical Methodology), 76(1):71–96, 2014.

[23] Oded Maler and Dejan Nickovic. Monitoring temporal properties ofcontinuous signals. In Formal Techniques, Modelling and Analysis ofTimed and Fault-Tolerant Systems, pages 152–166. Springer, 2004.

[24] Volodymyr Mnih, Adria Puigdomenech Badia, Mehdi Mirza, AlexGraves, Timothy P. Lillicrap, Tim Harley, David Silver, and KorayKavukcuoglu. Asynchronous Methods for Deep Reinforcement Learn-ing. arXiv:1602.01783 [cs], February 2016.

[25] Volodymyr Mnih, Koray Kavukcuoglu, David Silver, Andrei A Rusu,Joel Veness, Marc G Bellemare, Alex Graves, Martin Riedmiller,Andreas K Fidjeland, Georg Ostrovski, et al. Human-level controlthrough deep reinforcement learning. Nature, 518(7540):529–533,2015.

[26] Christopher Z Mooney. Monte carlo simulation, volume 116. Sagepublications, 1997.

[27] Mathworks R2020a. Train dqn agent for lane keep assist.https://www.mathworks.com/help/reinforcement-learning/ug/train-dqn-agent-for-lane-keeping-assist.html. The Mathworks.

[28] Alena Rodionova, Ezio Bartocci, Dejan Nickovic, and Radu Grosu.Temporal Logic as Filtering. Proceedings of the 19th InternationalConference on Hybrid Systems: Computation and Control - HSCC’16, pages 11–20, 2016.

[29] Yaniv Romano, Evan Patterson, and Emmanuel Candes. Confor-malized quantile regression. In Advances in Neural InformationProcessing Systems, pages 3538–3548, 2019.

[30] Nima Roohi, Yu Wang, Matthew West, Geir E. Dullerud, and MaheshViswanathan. Statistical verification of the toyota powertrain controlverification benchmark. In Proceedings of the 20th InternationalConference on Hybrid Systems: Computation and Control, HSCC ’17,page 65–70, New York, NY, USA, 2017. Association for ComputingMachinery.

[31] Vicenc Rubies Royo, David Fridovich-Keil, Sylvia L. Herbert, andClaire J. Tomlin. Classification-based approximate reachability withguarantees applied to safe trajectory tracking. ArXiv, abs/1803.03237,2018.

[32] Koushik Sen, Mahesh Viswanathan, and Gul Agha. Statistical model

checking of black-box probabilistic systems. In Rajeev Alur andDoron A. Peled, editors, Computer Aided Verification, pages 202–215,Berlin, Heidelberg, 2004. Springer Berlin Heidelberg.

[33] Xiaowu Sun, Haitham Khedr, and Yasser Shoukry. Formal verificationof neural network controlled autonomous systems. In Proceedings ofthe 22nd ACM International Conference on Hybrid Systems: Compu-tation and Control, HSCC ’19, page 147–156, New York, NY, USA,2019. Association for Computing Machinery.

[34] Richard S Sutton and Andrew G Barto. Reinforcement learning: Anintroduction. MIT press, 2018.

[35] Hoang-Dung Tran, Diago Manzanas Lopez, Patrick Musau, XiaodongYang, Luan Viet Nguyen, Weiming Xiang, and Taylor T. Johnson. Star-based reachability analysis of deep neural networks. In Maurice H.ter Beek, Annabelle McIver, and Jose N. Oliveira, editors, FormalMethods – The Next 30 Years, pages 670–686, Cham, 2019. SpringerInternational Publishing.

[36] Cumhur Erkan Tuncali, James Kapinski, Hisahiro Ito, and Jyotir-moy V. Deshmukh. Reasoning about safety of learning-enabledcomponents in autonomous cyber-physical systems. In Proceedingsof the 55th Annual Design Automation Conference, DAC ’18, NewYork, NY, USA, 2018. Association for Computing Machinery.

[37] Vladimir Vovk, Alex Gammerman, and Glenn Shafer. Algorithmiclearning in a random world. Springer Science & Business Media,2005.

[38] Yu Wang, Mojtaba Zarei, Borzoo Bonakdarpour, and Miroslav Pajic.Statistical verification of hyperproperties for cyber-physical systems.ACM Trans. Embed. Comput. Syst., 18(5s), October 2019.

[39] Shakiba Yaghoubi and Georgios Fainekos. Gray-box adversarialtesting for control systems with machine learning components. InProceedings of the 22nd ACM International Conference on HybridSystems: Computation and Control, pages 179–184, 2019.

[40] Mojtaba Zarei, Yu Wang, and Miroslav Pajic. Statistical verificationof learning-based cyber-physical systems. In Proceedings of the 23ndACM International Conference on Hybrid Systems: Computation andControl, 2020.

[41] Paolo Zuliani, Andre Platzer, and Edmund M Clarke. Bayesian statisti-cal model checking with application to simulink/stateflow verification.In Proceedings of the 13th ACM international conference on Hybridsystems: computation and control, pages 243–252, 2010.