interval analysis scott ferson, [email protected] 11 september 2007, stony brook university, mar 550,...
TRANSCRIPT
![Page 1: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/1.jpg)
Interval Analysis
Scott Ferson, [email protected] September 2007, Stony Brook University, MAR 550, Challenger 165
![Page 2: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/2.jpg)
Outline of Interval Analysis• What they are and why they’re used• Forward calculations
– Elementary formulas– Outward-directed rounding– Example: contaminant plume travel time– Repeated parameters– Tricks and subinterval reconstitution– Propagating uncertainty through black boxes – Dependencies among intervals
• Specifying intervals• Updating
– Aggregation– Constraint propagation
• Backcalculation• Exercises
![Page 3: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/3.jpg)
An interval x = [x1, x2]
• Used when upper and lower bounds are known reliably• Sometimes absolutely sure, sometimes just a judgment
range
x1 x2
![Page 4: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/4.jpg)
Why intervals are a good idea
• Intervals are easy to understand
• Calculations with them are easy
• They work with even the crappiest data
• Often sufficient for a decision
![Page 5: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/5.jpg)
Rigorousness
• The computations are guaranteed to enclose the true results (so long as the inputs do)
• “Automatically verified calculations”
• You can still be wrong, but the method won’t be the reason if you are
![Page 6: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/6.jpg)
General interval arithmetic
x + y = [x1 + y1, x2 + y2]x y = [x1 y2, x2 y1]x y = [min(x1y1, x1y2, x2y1, x2y2), max(x1y1, x1y2, x2y1, x2y2)]x y = x [1/y2, 1/y1], if y doesn’t include zero
Examples: x = [3, 5]y = [1, 2]x + y = [2, 7]x y = [1, 6]x y = [5, 10]
![Page 7: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/7.jpg)
Other simple arithmetic operations
min(x, y) = [min(x1, y1), min(x2, y2)]
max(x, y) = [max(x1, y1), max(x2, y2)]
env(x, y) = [min(x1, y1), max(x2, y2)]
intersect(x, y) = [max(x1, y1), min(x2, y2)], if x and y overlap
Examples: x = [0.3, 0.5], y = [0.1, 0.2]
min(x,y) = [0.1, 0.2]
env(x, y) = [0.1, 0.5]
intersect(x, y) doesn’t exist
![Page 8: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/8.jpg)
Other functions
• If the function is monotone (e.g., exp, log, sqrt) just apply it to the two endpoints
• Trickier if it isn’t monotone (e.g., abs, sin, cos)
• Note that generally non-monotone functions may be monotone over a particular interval
Examples: x = [0.3, 0.5]exp(x) = [1.34, 1.65]sin(x) = [0.295, 0.480]abs(x) = [0.3, 0.5]
![Page 9: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/9.jpg)
Outward-directed rounding
• To maintain the rigor of interval calculations, any necessary truncation of digits should be done with outward-directed rounding
Example:
[2,4] / 3 = [ 0.66666666666, 1.33333333333]
would be written as [0.66, 1.34]
even though 2/3 ordinarily rounds to 0.67,
and 4/3 ordinarily rounds to 1.33
_ _
![Page 10: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/10.jpg)
Example: contaminant plume
• Hydrocarbon in groundwater near some wells• Constant, one-dimensional, uniform Darcian flow• Homogeneous properties (e.g., no pipes, conduits,
barriers or differential permeability among layers)• Linear retardation• No dispersion (just Darcian advection)
• How long before the contaminant reaches the wells?
![Page 11: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/11.jpg)
Darcy’s Law
• Statistical description of fluid flow in porous media • Expresses conservation of momentum• Flow rate is proportional to
– cross-sectional area
– hydraulic conductivity and
– gradient in the hydraulic head
• Analogous to– Fick’s Law in diffusion
– Ohm’s Law (I=V/R)
• Derivable from Navier-Stokes by homogenization
![Page 12: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/12.jpg)
Travel time (Lobascio)
iK
LKocfocBDnT
ParameterL source-receptor distancei hydraulic gradientK hydraulic conductivityn effective soil porosityBD soil bulk densityfoc fraction organic carbonKoc organic partition coefficient
Min800.00033000.215000.00015
Max1200.000830000.3517500.00520
Mean1000.0005510000.2516500.0025510
Stdv11.550.00014437500.051000.0014153
Unitsmm/mm/yr-kg/m3
-m3/kg
![Page 13: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/13.jpg)
Results
• The most optimistic outcome would be that the plume takes ___ years to reach the well
• But, in the worst case, it could take __ years to reach the well
Answers: 233800 years, 31.6 years
![Page 14: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/14.jpg)
[31.6, 233800] years
• Is ‘6’ the last decimal digit of the lower bound?
• Did you check that the units balance?
• Did you include units in the answer?
![Page 15: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/15.jpg)
How to understand this result
• Highly reliable result, given the assumptions– Can’t get any worse
• Represents parametric uncertainty– Neglects (possibly big) model uncertainty
• Expresses only best and worst cases– How likely is 32 years? 50 years? 100 years?
![Page 16: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/16.jpg)
Lobascio’s original formulation
Kd= foc Koc = [ 0.0005, 0.1] m3 kg-1
R = 1 + BD Kd / n = [ 3.143, 876] V = K i / (n R) = [ 0.000293, 3.82] m yr-1
T = L/V = [ 20.95, 408800] yr
Quickest plume reaches the well = 20.95 yrLongest plume reaches the well = 408,800 yr
What explains the difference? (hint: n is repeated above)
![Page 17: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/17.jpg)
Repeated parameters
a = [1,2]b = [2,3]c = [2, 5]
z = a × (b + c) zz = a × b + a × c
b + c = [0, 8] a × b = [2, 6]z = [0, 16] a × c = [4, 10] zz = [2, 16]
inflated uncertainty
![Page 18: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/18.jpg)
What to do about repeated parameters
• Always rigorous, but maybe not best possible when uncertain parameters are repeated
• Inconsequential if all are non-negative and all operations are increasing (+, ×, but not – or ÷)
• Use cancellation to reduce repetitions, e.g., caia/m + cwiw/m + cdid/m = (caia + cwiw + cdid)/m
• Cancellation not always possible, e.g., (a + b) / (a + c) = ??
![Page 19: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/19.jpg)
If you can’t cancel
• Use tricks with algebrae.g., a² + a = (a +½)² – ¼
• Employ subinterval reconstitutionA brute-force and computationally intensive strategy
Workable if there aren’t too many repeated parameters
• Live with the suboptimalityDecisions may not require perfect precision
![Page 20: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/20.jpg)
TricksBasic identities
u + 0 = u
u – 0 = u
0 – u = –u
u 0 = 0
u 1 = u
u / 1 = u
u0 = 1
u1 = u
u & 1 = u
u | 1 = 1
u & 0 = 0
u | 0 = u
u & u = u
u | u = u
u & not(u) = 0
u | not(u) = 1
(u&a) | (u&b) = u&(a | b)
(u | a)&(u | b) = u | (a&b)
etc.
One repetitionu + u = 2uu – u = 0u u = u2
u / u = 1(1+u) / u = 1/u + 1(1+u)/(1–u) = (1/tan(acos(u)/2))2
au + bu = u(a + b)au – bu = u(a – b)a/u + b/u = (a + b) / ua/u – b/u = (a – b) / uu/a + u/b = u(b + a)/(ab)u/a – u/b = u(b – a)/(ab)aub + cub = (a + c) ub
aub cud = a c u(b + d)
au bu = exp(u (ln(a) + ln(b)))u2 + u = (u + ½)2 – ¼u2 – u = –¼ sin(2 asin(sqrt(u)))2
u2 + au = (u + a/(2))2 – a2/4etc.
u, v, etc. represent the uncertain numbers a, b, etc. represent arbitrary expressions
Two repetitionsu + v – uv = 1 – (1 – u) (1 – v)(u + v) / (1 – uv) = tan(atan(u) + atan(v))(u – v) / (1 + uv) = tan(atan(u) – atan(v))(1 + uv) / (u – v) = 1 / tan(atan(u) – atan(v))(1 – uv) / (u + v) = 1 / tan(atan(u) + atan(v))(uv – 1) / (u + v) = –1 / tan(atan(u) + atan(v))u sqrt(1 – v2) + v sqrt(1 – u2) = sin(asin(u) + asin(v))u sqrt(1 – v2) – v sqrt(1 – u2) = sin(asin(u) – asin(v))u v + sqrt(1 – u2) sqrt(1 – v2) = cos(acos(u) – acos(v))u v – sqrt((1 – u2) (1 – v2)) = cos(acos(u) + acos(v))u v – sqrt(1 – u2 – v2 + u2 v2) = cos(acos(u) + acos(v))sin(u) sqrt(1 – sin(v)2) + sin(v) sqrt(1 – sin(u)2) = sin(u + v)cos(u) cos(v) – sin(u) sin(v) = cos(u + v)sin(u) cos(v) – cos(u) sin(v) = sin(u – v)sqrt((1 + u) / (1 – u)) = 1 / tan(acos(u)/2)etc.
![Page 21: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/21.jpg)
Subinterval reconstitution
• Partition each repeated interval into subintervals• Compute the function for every subinterval• The union of all the results contains the true range
i j k
kji zyxwvufzyxwvuf ),...,,,,...,,(...),...,,,...,,(
i j k
kji wwvvuu ...;;;
where u, v, …, w are repeated intervals and x, y,…, z are other interval and scalar inputs, and
![Page 22: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/22.jpg)
Example: (a + b)a, a = [0.1, 1], b = [0,1]
Partition the repeated uncertain a:ai = [(i 1)w/m + a, iw/m + a] where i = 1,2,…, m, and m is the number of subintervals, w is thewidth of a, and a is its lower bound
m U(ai+b)ai
1 [ 0.1, 2]2 [ 0.282, 2]3 [ 0.398, 2]4 [ 0.473, 2]5 [ 0.525, 2]10 [ 0.624, 2]100 [ 0.686, 2]1,000 [ 0.692, 2]10,000 [ 0.692, 2]a
0 0.5 1
m U(ai+b)ai
1 [ 0.1, 2]2 [ 0.282, 2]3 [ 0.398, 2]4 [ 0.473, 2]5 [ 0.525, 2]10 [ 0.624, 2]100 [ 0.686, 2]1,000 [ 0.692, 2]10,000 [ 0.692, 2]
![Page 23: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/23.jpg)
Cauchy-deviate method
• Propagates intervals through black-box model– Don’t need to know, but have to be able to query it
• “Sample” from around interval – Points not necessarily inside the interval!
• Scale results to get an asymptotically correct estimate of the interval uncertainty of the output
(Trejo and Kreinovich 2001)
![Page 24: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/24.jpg)
Cauchy-deviate method
• Depends on the number of samples, not inputs– Works just as well for 2000 variables as 20– Similar in performance to Monte Carlo
• Need about 200 samples to obtain 20% relative accuracy of half-width of output range– With fewer samples, we’d get lower accuracy, but
we can compensate for this by scaling by N, which works under the linearity assumption
![Page 25: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/25.jpg)
Limitations of the method
• Asymptotically correct, but not rigorous
• Intervals narrow relative to the nonlinearity– Function almost linear OR uncertainties small– Could combine with subinterval reconstitution
• Most efficient when dimensionality is high
• Only handles interval uncertainty
![Page 26: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/26.jpg)
Dependence among intervals
![Page 27: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/27.jpg)
Dependence in interval analysis
• Interval analysis automatically accounts for all possible dependencies
• Information about dependencies isn’t usually used to tighten results, but it could be
![Page 28: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/28.jpg)
Dependencies
• Any restriction on the possible pairings of inputs (any subset of the unit square) May also require each value of u to match with at least v, and vice versa
• The null restriction is the full unit square“Nondependence” or “noninteractivity” rather than “independence”
• D denotes the set of all possible dependencies Set of all subsets (power set) of the unit square D = 2[0,1][0,1]
![Page 29: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/29.jpg)
Three special cases
0 u 1 0 u 1 0 u 1
1
v
0
1
v
0
1
v
0
Perfect(comonotonic)
Nondependent(the Fréchet case)
Opposite(countermonotonic)
![Page 30: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/30.jpg)
Correlation
• A model of dependence that’s parameterized by a (scalar) value called the “correlation coefficient”
: [1, +1] D = 2[0,1][0,1]
• A correlation model is called “complete” if
(1) = , (0) = , (+1) =
![Page 31: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/31.jpg)
r = 1 r = 0 r = +1
Corner-shaving dependence
D(r) = { (u,v) : max(0, ur, u1+r) v min(1, u+1r, u+2+r)} u [0,1], v [0,1]
f (A, B) = { c : c = f (u (a2 – a1) + a1, v (b2 – b1) + b1), (u,v) D }A+B = [env(w(A, r)+b1, a1+w(B,r)), env(a2+w(B,1+r),w(A,1+r)+b2)]
a1 if p < 0w([a1,a2], p) = a2 if 1 < p p(a2a1)+a1 otherwise
![Page 32: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/32.jpg)
Other complete correlation familiesr = 1 r = 0 r = +1
![Page 33: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/33.jpg)
Elliptic dependence
• Not complete (because r = 0 isn’t nondependence)
A+B = [d, d] + (a1 + a2 + b1 + b2) / 2
d = (d12 + d2
2 + r d1 d2), d1 = (a2 a1)/2, d2 = (b2 b1)/2
r = 1 r = 0 r = +1
A+B = [d, d] + (a1 + a2 + b1 + b2) / 2
d = (d12 + d2
2 + r d1 d2), d1 = (a2 a1)/2, d2 = (b2 b1)/2
![Page 34: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/34.jpg)
[ 5, 14] Perfect
[ 8, 11] Opposite
[ 7.1, 11.9] Corner-shaving (r = 0.7)
[ 7.27, 11.73] Elliptic (r = 0.7)
[ 5, 14] Upper, left
[ 5, 11] Lower, left
[ 8, 14] Upper, right
[ 5, 14] Lower, right
[ 6.5, 12.5] Diamond
[ 5, 14] Nondependent
A + B
A = [2,5]B = [3,9]
![Page 35: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/35.jpg)
Implications
• Intervals aren’t more widely used is because they’re thought to be (and often are) too wide to be helpful
• They can be much tighter if we used all available information, such as about intervariable dependence
• (Other information, like about moments, could also be used to tighten interval calculations)
![Page 36: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/36.jpg)
Computing
• Sequence of binary operations– Need to deduce dependencies of intermediate results
with each other and the original inputs– Different calculation order can give different results
(which should be intersected)
• Do all at once in one multivariate calculation – Can be much more difficult computationally– Can produce much better tightening
![Page 37: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/37.jpg)
Specifying input intervals
![Page 38: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/38.jpg)
Interval uncertainty
• Statisticians often ignore this uncertainty
• “Interval uncertainty doesn’t exist in real life”(Tony O’Hagan et
al.)
Hammer salesmen saying screws don’t exist?
![Page 39: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/39.jpg)
When do intervals arise?
• Periodic observationsWhen did the fish in my aquarium die during the night?
• Plus-or-minus measurement uncertaintiesCoarse measurements, measurements from digital readouts
• Non-detects and data censoringChemical detection limits, studies prematurely terminated
• Privacy requirementsEpidemiological or medical information, census data
• Theoretical constraintsConcentrations, solubilities, probabilities, survival rates
• Bounding studies Presumed or hypothetical limits in what-if calculations
![Page 40: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/40.jpg)
Ways to characterize intervals
• Theoretical constraints
• Modeled from other intervals
• Expert assertions
• Discounting (widening) intervals (Shlyakhter)
• Confidence procedures (Grosof)– But 95% confidence isn’t the same as surety– Use in interval calculations requires an assumption
![Page 41: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/41.jpg)
Problems with confidence intervals
• Cannot be combined in arithmetic or logical operations without an assumption
• Don’t measure epistemic belief anyway
Example (Walley): For instance, a 90% confidence interval could have zero chance of containing the value. For example, suppose X ~ normal(, 1), where 0 < . If the sample mean happens by chance to be 1.3, the 90% confidence interval on the mean is the empty set (Walley, P. 1991, Statistical Reasoning with Imprecise Probabilities, Chapman & Hall, page 379).
![Page 42: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/42.jpg)
Why we have to be careful
• Interval analysis yields contingent results• Results are contingent on assumptions that model inputs are within
their respective intervals• But all analysis results are contingent on similar assumptions that the
models they came from are true
• Naïve elicitation has big problems• Intervals are usually unrealistically narrow• People make incoherent statements
• Can’t mix together different kinds• Not clear how to translate data into intervals
![Page 43: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/43.jpg)
Determining endpoints
• The largest observed may not be the largest possible (and it usually isn’t)
• Sampling theory >> theory of extremes
• Rigor of analysis is contingent on inputs
• If you’re nervous, just widen the bounds
![Page 44: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/44.jpg)
Point sample data
Range (envelope)
Extreme value model
Tolerance interval
Confidence interval
Central value and width
P-box
Level cutcut
Model
Output rangesimulation, etc.Plus-minus interval
Shlyakhter widening
Envelope
Interval function
Distribution
Support
Percentile range
Trejo-KreinovichCauchy deviates
Tolerance solutionbackcalculation
Certain and tenable ranges
Intersection
Credibility interval
Prediction interval
Percentile range
Support
![Page 45: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/45.jpg)
Eliciting dependence
• As hard as getting intervals (maybe a bit worse)
• Theoretical or “physics-based” arguments
• Inference from empirical data – Risk of loss of rigor at this step (just as there is
when we try to infer intervals from data)
![Page 46: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/46.jpg)
Updating
![Page 47: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/47.jpg)
Aggregation (updating)
• How do you combine different sources?
• If you trust them all, take the intersection– [max(x1, y1, z1, …), min(x2, y2, z2, …)]– What if there is no intersection (right<left)?
• If you’re not sure which is right, use the envelope– [min(x1, y1, z1, …), max(x2, y2, z2, …)]– But are you sure this is wide enough?
![Page 48: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/48.jpg)
Example
• Suppose we have two rigorous interval estimates of the same quantity: [1,7] & [4,10]
• Their intersection [4,7] is also a rigorous interval for the quantity
AB
0 5 10
![Page 49: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/49.jpg)
Constraint analysis (updating)
• Using knowledge of how variables are related to tighten their estimates
• Removes internal inconsistency and explicates unrecognized knowledge
• Also called ‘constraint updating’ or ‘editing’
• Also called ‘natural extension’
![Page 50: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/50.jpg)
Example
• Suppose we know
W = [23, 33] m
H = [112, 150] m
A = [2000, 3200] m2
• Does knowing W H = A let us to say any more?
![Page 51: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/51.jpg)
Answer
• Yes! We can infer that
W = [23, 28.57]
H = [112, 139.13]
A = [2576, 3200]
• The formulas are just W = intersect(W, A/H), etc.
To get the largest possible W, for instance, let A be as large as possible and H as small as possible, and solve for W =A/H.
![Page 52: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/52.jpg)
Updating with p-boxes
2000 3000 40000
1
20 30 400
1
120 140 1600
1 A W H
![Page 53: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/53.jpg)
2000 3000 40000
1
20 30 400
1
120 140 1600
1
Answers
intersect(W, A/H) intersect(H, A/W) intersect(A, WH)
A W H
![Page 54: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/54.jpg)
Bayesian strategy
20003200
])3200,2000[(
112150
])150,112[(
2333
])33,23[(),,Pr(
AIHIWI
AHW
)(),,|( HWAAHWHWAL
),,Pr()()|,,( AHWHWAHWAAHWf
Prior
Likelihood
Posterior
![Page 55: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/55.jpg)
Bayes’ rule
• Concentrates mass onto the manifold of feasible combinations of W, H, and A
• Answers have the same supports as intervals
• Computationally complex• Needs specification of priors• Yields distributions that are not justified
(coming from the choice of priors)• Expresses less uncertainty than is present
![Page 56: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/56.jpg)
Backcalculation
![Page 57: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/57.jpg)
Backcalculation
• Needed for cleanup and remediation planning
• Untangles an equation in uncertain numbers when we know all but one of the variables
• For instance, backcalculation finds B such that A+B = C, from estimates for A and C
![Page 58: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/58.jpg)
Can’t just invert the equation
Dose = Concentration × Intake
Concentration = Dose / Intake
When concentration is put back into the forward equation, the resulting dose is wider than planned
prescribed knownunknown
![Page 59: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/59.jpg)
Exampledose = [0, 2] milligram per kilogramintake = [1, 2.5] litermass = [60, 96] kilogram
conc = dose * mass / intake [ 0, 192] milligram liter-1
dose = conc * intake / mass [ 0, 8] milligram kilogram-1
Doses four times larger than tolerable levels we planned
![Page 60: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/60.jpg)
Untangling backcalculations
• Solving for B given A + B = C
B = backcalc(A, C) = [C1 A1, C2 A2]
• Solving for B given A B = C
B = factor(A, C) = [C1 / A1, C2 / A2]
• Sometimes called “tolerance solutions”
![Page 61: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/61.jpg)
1
2
3
4
5
6
1 1.5 2A
Bsh
ell
Kernal versus shell
Given A [1,2] C [2,6] C A BThere are two different ways to solve for B
kern
el
Shell (united solution)B C A [0,5]
Kernel (tolerance solution)B backcalc(A,C) [1,4]
![Page 62: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/62.jpg)
When you need for
A + B C
A – B C
A B C
A / B C
A ^ B C
2A C
A² C
And you have estimates for
A, BA, CB ,CA, BA, CB ,CA, BA, CB ,CA, BA, CB ,CA, BA, CB ,C
ACAC
Use this formulato find the unknownC = A + BB = backcalc(A,C)A = backcalc (B,C)C = A – BB = –backcalc(A,C)A = backcalc(–B,C)C = A * BB = factor(A,C)A = factor(B,C)C = A / BB = 1/factor(A,C)A = factor(1/B,C)C = A ^ BB = factor(log A, log C)A = exp(factor(B, log C))C = 2 * AA = C / 2C = A ^ 2A = sqrt(C)
![Page 63: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/63.jpg)
Interval algebra
• Commutativity a+b=b+a, a×b=b×a• Associativity (a+b)+c=a+(b+c), (a×b)×c=a×(b×c)• Neutral elements a+0=0+a=a, a×1=1×a=a
• Subdistributivity a×(b+c) a×b+a×c• Subcancellation a a+bb, a a×b/b• No inverse elements a+( a) 0, a×(1/a) 1
![Page 64: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/64.jpg)
Conclusions
• Interval analysis is a worst case analysis (that also includes the best case)
• Repeated uncertain parameters can cause unnecessary inflation of uncertainy
• Results will always be rigorous, but might not be best possible
• Moving a uncertain parameter to the other side of an equal sign often requires backcalculation
![Page 65: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/65.jpg)
Exercises
1. Do the inputs in the travel time example seem dependent?2. What does subinterval reconstitution with m=100 on the
original Lobascio formulation give for the travel time?3. What contaminant concentrations C in water will lead to
doses D no larger than 6 mg per kg per day if it comes from both drinking and eating fish as D = (Iwater C) / W + (Ifish B C) / W, where
Iwater = [1.5, 2.5] liters per day // water intake
Ifish = [0, 8] g per day // dietary ingestion of fishB = [0.9, 2.1] liters per mg // bioaccumulation factorW = [60, 90] kg // receptor biomass
How do you check the solution?4. Is there a Bayesian analog of backcalculation?
////////////////////////////////////////////////////////////////////////////////////////////////////////////////////Answer to #1
L source-receptor distancei hydraulic gradientK hydraulic conductivityn effective soil porosityBD soil bulk densityfoc fraction organic carbonKoc organic partition coefficient
Porosity and conductivity would likely be positively relatedPorosity and density would likely be inversely relatedThus, conductivity and density should be inversely related tooFraction organic carbon and the partition coefficient would likely be positively related
////////////////////////////////////////////////////////////////////////////////////////////////////////////////////Answer to #2
L = [80,120] m // source-receptor distancei = [0.0003,0.0008] m per m // hydraulic gradientK = [300,3000] m per yr // hydraulic conductivityn = [0.2,0.35] // effective soil porosityBD = [1500,1750] kg per m{3} // soil bulk densityfoc = [0.0001,0.005] // fraction organic carbonKoc = [5 ,20] m{3} per kg // organic partition coefficient Kd = foc * KocR = 1 + BD * Kd / n V = K * i / (n * R)T = L/V T [ 20.952380952, 408800] yr
func TT() return L/(K * i / ($1 * (1 + BD * foc * Koc / $1)))
b = left(n)e = TT(b)w = width(n)/100for k=1 to 100 do begin s = [ (k-1), k] * w + b e = env(e, TT(s)) ende [ 31.480562448, 235352] yr
////////////////////////////////////////////////////////////////////////////////////////////////////////////////////Answer to #3
Iwater = [1.5, 2.5] liters per day // water intakeIfish = [0, 8] g per day // dietary ingestion of fish tissueB = [0.9, 2.1] liters per g // bioaccumulation factorW = [60, 90] kg // receptor biomass
D = [0, 6] mg per kg per day // tolerable dose
// forward equation// D = (Iwater * C) / W + (Ifish * B * C) / W
// condense repeated parameters// D = C * (Iwater + Ifish * B) / W
// solve for C by factoring the quotient (Iwater+Ifish*B)/W out of DC = factor((Iwater + Ifish * B) / W, D) C [ 0, 18.652849741] mg liters-1
// put C back into original forward equation (and ensure correct units)d = (Iwater * C) / W + (Ifish * C * B) / W + 0 mg per kg per day
d [ 0, 6] mg kg-1 day-1
![Page 66: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/66.jpg)
(break)
![Page 67: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/67.jpg)
Interval probability
![Page 68: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/68.jpg)
Outline for Interval Probability
• Logical operations under independence• Dependence among events
– Perfect and opposite dependence– Frank and Pearson correlation– Fréchet inequalities– Signed dependence
• Example: rupture of a pump system• Exercises• Conclusions• References
![Page 69: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/69.jpg)
Interval arithmetic for probabilities
x + y = [x1, x2] + [y1, y2] = [x1 + y1, x2 + y2]
x y = [x1, x2] [y1, y2] = [x1 y2, x2 y1]
x y = [x1, x2] [y1, y2] = [x1 y1, x2 y2]
x y = [x1, x2] [y1, y2] = [x1 y2, x2 y1]
Examples: x = [0.3, 0.5]y = [0.1, 0.2]x + y = [0.4, 0.7]x y = [0.1, 0.4]
Rules are simpler because intervals confined to [0,1]
![Page 70: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/70.jpg)
Probabilistic logic
• Conjunctions (and)
• Disjunctions (or)
• Negations (not)
• Exclusive disjunctions (xor)
• Modus ponens (if-then)
• etc.
![Page 71: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/71.jpg)
Conjunction (and)
P(A |&| B) = P(A) P(B)
Example: P(A) = [0.3, 0.5]
P(B) = [0.1, 0.2]
A and B are independent
P(A |&| B) = [0.03, 0.1]
![Page 72: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/72.jpg)
Disjunction (or)
P(A || B) = = 1 (1 P(A))(1 P(B))
Example: P(A) = [0.3, 0.5]
P(B) = [0.1, 0.2]
A and B are independent
P(A || B) = [0.37, 0.6]
![Page 73: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/73.jpg)
Disjunction (or)
P(A || B) = P(A) + P(B) P(A) P(B)
Example: P(A) = [0.3, 0.5]
P(B) = [0.1, 0.2]
A and B are independent
P(A || B) = [0.3, 0.67]
repeatedparameters!
![Page 74: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/74.jpg)
Negation (not)
P(not A) = 1 P(A)
Example: P(A) = [0.3, 0.5]
P(not A) = [0.5, 0.7]
![Page 75: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/75.jpg)
1 2 3 4 5
Dependence
Independent
Probabilities are depicted hereas the areas in Venn diagrams
Perfect Opposite
![Page 76: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/76.jpg)
Perfect dependence
P(A /&/ B) = min(P(A), P(B)) P(A // B) = max(P(A), P(B))
Examples: P(A) = [0.3, 0.5]P(B) = [0.1, 0.2]A and B are perfectly dependentP(A /&/ B) = [0.1, 0.2]P(A // B) = [0.3, 0.5]
![Page 77: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/77.jpg)
Opposite dependence
P(A \&\ B) = max(P(A) + P(B) 1, 0)
P(A \\ B) = min(1, P(A) + P(B))
Examples: P(A) = [0.3, 0.5]
P(B) = [0.1, 0.2]
A and B are oppositely dependent
P(A \&\ B) = 0
P(A \\ B) = [0.4, 0.7]certain
![Page 78: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/78.jpg)
Correlated events (Frank)
where r is the correlation, s = tan((1r)/4), a = P(A), b = P(B)
otherwise
1 if
0 if
1 if
)1/()1)(1(1log
)0,1max(
),min(
),,(and)&P(
s
Frank
r
r
r
sss
ba
ab
ba
rbaBA
ba
)&P( BAr
otherwise.
1 if
0 if
1 if
)1/()1)(1(1log-1
)1,min(
)1)(1(1
),max(
),,(or) P(
11s
Frank
r
r
r
sss
ba
ba
ba
rbaBA
ba
) ∨P( BA r
![Page 79: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/79.jpg)
0
0.1
0.2
0.3
0.4
0.5
0.6
1 0.75 0.5 0.25 0 0.25 0.5 0.75 1
Frank correlation coefficient, r
Pro
babi
lity
orFrank(0.29, 0.22, r)
andFrank(0.29, 0.22, r)
Examples of Frank correlations
![Page 80: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/80.jpg)
Other correlation models
• Pearson correlation (on indicator functions)
–
– doesn’t allow all possible correlation coefficients– permits negative and larger-than-one probabilities
• But any complete copula would work fine– Frank, Clayton, Mardia, etc.– Nelsen (1999) is the essential reference on copulas
![Page 81: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/81.jpg)
P(A & B) = [max(0,P(A)+P(B)–1), min(P(A), P(B))]
P(A or B) = [max(P(A), P(B)), min(1, P(A)+ P(B))]
• Makes no assumption about the dependence
• Rigorous (guaranteed to enclose true value)
• Best possible (cannot be any tighter)
Fréchet case
![Page 82: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/82.jpg)
The proofs are elementaryP(A B) = P(A) + P(B) P(A & B) implies P(A) + P(B) P(A B) = P(A & B).
P(A B) 1, since probabilities are no bigger than 1, so P(A) + P(B) 1 P(A & B).
0 P(A & B), since probabilities are positive, so max(0, P(A) + P(B) 1) P(A & B).
This gives the lower bound on the conjunction.
To get the upper bound, recall that P(A & B) = P(A|B) P(B) = P(B|A) P(A).
P(A|B) 1 and P(B|A) 1, so P(A & B) P(A) and P(A & B) P(B).
Therefore, P(A & B) min(P(A), P(B)), which is the upper bound.
The best-possible nature of these bounds follows from observing that
they are realized by some dependency between the events A and B.
Comparable bounds on the disjunction are similarly derived.
![Page 83: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/83.jpg)
Fréchet examples
Examples: P(A) = [0.3, 0.5]P(B) = [0.1, 0.2]P(A & B) = [0, 0.2]P(A B) = [0.3, 0.7]
P(C) = 0.29P(D) = 0.22P(C & D) = [0, 0.22]P(C D) = [0.29, 0.51]
}certain
}uncertain
![Page 84: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/84.jpg)
Sign of the dependence
P(A &+ B) = [ab, min(a,b)]
P(A + B) = [1(1a)(1b), max(a, b)]
P(A & B) = [max(a+b1, 0), ab]
P(A B) = [1(1a)(1b), min(1, a+b)]
where a = P(A), b = P(B)
![Page 85: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/85.jpg)
RelayK2
Pressureswitch S
Timerrelay R
Pump
Motor
RelayK1
Fromreservoir
Pressure tank T Outletvalve
SwitchS1
S tartupSw itc h S1 c lo sedR e lay K1 c lo sedR e lay K2 c lo sedT im e r re la y c lo sedP re ssu re sw it ch c lo sed
Start pum pR e lay K2 c lo sedT im e r re la y s ta r tsP re ssu re sw it ch c lo sed
ReadyS w itch S1 o p enR e lay K1 c lo sedR ela y K2 o p enT im e r re la y c lo sedPre ssure sw itch o p en
System offS w itch S1 o p enR ela y K1 o p enR ela y K2 o p enT im e r re la y c lo sedP re ssu re sw it ch c lo sed
Shut downR ela y K1 o p enR ela y K2 o p enT im er re la y o p enP re ssu re sw it ch c lo sed
P um pingS w itch S1 o p enR e lay K1 c lo sedR e lay K2 c lo sedT im e r re la y c lo sedP re ssu re sw it ch c lo sed
S top pum pR ela y K2 o p enT im er re lay re se tsPre ssure sw itch o p en
Example: pump system
What’s the chance thetank ruptures under pumping?
![Page 86: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/86.jpg)
E1 = “tank rupturing under pressurization”
Fault tree
Vesely et al. (1981)
E1
T
K2
E2
E3
S
E4
S1
E5
K1
or and
or
or
or
R
![Page 87: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/87.jpg)
Boolean expression
E1 = T (K2 (S & (S1 (K1 R))))
ComponentPressure tank T Relay K2 Pressure switch S Relay K1 Timer relay R On-switch S1
Probability5 106
3 105
1 104
3 105
1 104
3 105
![Page 88: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/88.jpg)
Different dependency models
Vesely et al. (all variables precise, independent)E1 = T || (K2 || (S |&| (S1 || (K1 || R))))
Mixed dependencies
E1 = T || (K2 (S &r (S1 || (K1 // R))))Frank to Fréchet
E1 = T || (K2 (S & (S1 || (K1 // R))))All Fréchet
E1 = T (K2 (S & (S1 (K1 R))))
![Page 89: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/89.jpg)
Interval probabilities
ComponentPressure tank T Relay K2 Pressure switch S Relay K1 Timer relay R On-switch S1
Probability interval[4.5 106, 5.5 106][2.5 105, 3.5 105][0.5 104, 1.5 104][2.5 105, 3.5 105][0.5 104, 1.5 104][2.5 105, 3.5 105]
![Page 90: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/90.jpg)
105 104 103
Probability of tank rupturing under pumping
Mixed dependencies
Frank to Fréchet
All Fréchet
Mixed dependence with intervals
Vesely et al. (all independent)
Results
3.5105
[3.499105, 3.504105]
[3.50105, 1.35104]
[3105, 1.4104]
[2.9105, 4.1105]
![Page 91: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/91.jpg)
Strategies to handle repetitions
• Interval analysis is always rigorous, but maybe not best possible when parameters are repeated
• Use cancellation to reduce repetitions, e.g.,(A & B) (A & C) (A & D) = A & (B C D)
• When cancellation is not possible, mathematical programming may be necessary to get best possible result
![Page 92: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/92.jpg)
Subtle dependencies
• May also require mathematical programming to obtain the best possible result
• But rigorous bounds are always easy to get with the artful use of the Fréchet case
![Page 93: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/93.jpg)
Advances in practices
“It practice, independence among events and event probabilities is often assumed for the sake of convenience because it makes otherwise intractable problems easier to solve.” (Vick 2002, page 412)
“…Independent and conditional relationships [among events] could not be established without the subjective elements of one’s state of knowledge, making it otherwise all but impossible to apply the [probability] calculus to most situations of practical importance.” (ibid.)
![Page 94: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/94.jpg)
Conclusions
• Easy to compute rigorous bounds
• Mathematical programming may be needed to get answers that are also best possible
• Rigor of analysis is contingent on inputs
• If you’re nervous, just widen the bounds
![Page 95: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/95.jpg)
Exercises
5. Calculate the probability of tank rupture under pumping that assumes the interval inputs and makes no assumption about the dependencies among the events.
6. Develop an fault tree for establishment of snake populations on a Hawaiian island (or a star exploding).
7. Compute the probability of the conjunction of two events having probabilities 0.29 and 0.22, assuming a Pearson correlation of 1.0. Compare the result to the Fréchet range for such probabilities. What’s going on?
8. Derive an algorithm to compute the probability that n of k events occur, given intervals for the probability of each event, assuming they’re independent. Derive an analogous algorithm for the Fréchet case.
Vesely et al. (all independent)3.5105
[3.499105, 3.504105]
[3.50105, 1.35104]
[3105, 1.4104]
[2.9105, 4.1105]
[ 2.5105, 1.905 104]
![Page 96: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/96.jpg)
Snakes on a plane (or boat)Introduction
Stowaway on plane OR Stowaway on ship OR Arrives on natural raft OR Brought as pet OR Brought for zoo
AND Escape and ReproductionPregnant snake escapes OR Male and female escape AND Successful reproduction OR Female escapes AND Parthenogenic reproduction
AND Survives predationDomestic and feral dogs ANDFeral cats ANDWild pigs
AND Prey is accessibleGround nesting bird eggsOR Rodents
AND Survives eradication programEludes detection by biologistsOR Reproduction too quickOR Dispersal too quick
![Page 97: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/97.jpg)
Overall conclusions
![Page 98: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/98.jpg)
Why bounding?
• Often sufficient to specify a decision • Possible even when estimates are impossible• Usually easy to compute and simple to combine• Rigorous, rather than an approximation
(after N.C. Rowe 1988)
![Page 99: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/99.jpg)
Reasons to use interval analysis
• Requires very little data• Applicable to all kinds of uncertainty• Can be comprehensive• Fast and easy to compute answers• Conservative when correlations unknown • Can be made “best possible”• Backcalculations easy• Updating relatively easy
![Page 100: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/100.jpg)
Reasons not to use it
• Same thing as worst case analysis
• Doesn't say how likely extreme event is
• Repeated parameters are cumbersome
• Not optimal when there’s a lot of data
• Can't use distribution information
• Doesn't use correlation information
![Page 101: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/101.jpg)
Interval (worst case ) analysisHow?
– bound inputs, a = [a1, a2], where a1 a2
– addition: [a1, a2] + [b1, b2] = [a1+b1, a2+b2]– subtraction: [a1, a2] – [b1, b2] = [a1–b2, a2–b1]– multiplication, division, etc. are a little more complex
Why?– natural for scientists and easy to explain to others– works no matter where uncertainty comes from
Why not?– paradoxical: can’t give exact value but can give exact bounds– ranges could grow quickly, yielding very wide results– doesn’t give probabilities of extreme outcomes (tail risks)
![Page 102: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/102.jpg)
Interval probabilityHow?
– bound event probabilities, p = [p1, p2], where 0 p1 p2 1
– evaluate event trees as composition of ANDs, ORs, etc.
– standard probabilistic rules if events are independent
– Fréchet rules if their dependence is unknown
– other dependency relations can also be represented
Why?
– can capture incertitude about event probabilities
Why not?
– paradoxical: can’t give exact value but can give exact bounds
– ranges can grow quickly, especially without independence
![Page 103: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/103.jpg)
ReferencesDwyer, P. 1951. Linear Computations. John Wiley & Sons, New York.Ferson, S. 2002. RAMAS Risk Calc 4.0: Risk Assessment with Uncertain Numbers. Lewis Publishers, Boca Raton.Grosof, B.N. 1986. An inequality paradigm for probabilistic knowledge: the logic of conditional probability intervals.
Uncertainty in Artificial Intelligence. L.N. Kanal and J.F. Lemmer (eds.), Elsevier Science Publishers, Amsterdam.Hailperin, T. 1986. Boole’s Logic and Probability. North-Holland, Amsterdam.Kyburg, H.E., Jr. 1998. “Interval Valued Probabilities,” http://ippserv.rug.ac.be/documentation/interval_prob/interval_prob.html,
The Imprecise Probabilities Project, edited by G. de Cooman and P. Walley, http://ippserv.rug.ac.be/home/ipp.html.Lobascio, M.C. 1993. Uncertainty analysis tools for environmental modeling: application of Crystal Ball® to predict
groundwater plume traveling times. ENVIRONews 1: 6-10.Loui, R.P. 1986. Interval based decisions for reasoning systems. Uncertainty in Artificial Intelligence. L.N. Kanal and
J.F. Lemmer (eds.), Elsevier Science Publishers, Amsterdam.Moore, R.E. 1966. Interval Analysis. Prentice-Hall, Englewood Cliffs, New Jersey.Moore, R. 1979. Methods and Applications of Interval Analysis. SIAM, Philadelphia.Rowe, N.C. 1988. Absolute bounds on the mean and standard deviation of transformed data for constant-sign-derivative
transformations. SIAM Journal of Scientific Statistical Computing 9: 1098–1113. Shlyakhter A. 1994. Improved framework for uncertainty analysis: accounting for unsuspected errors. Risk Analysis
14(4):441-447.Tessem, B. 1992. Interval probability propagation. International Journal of Approximate Reasoning 7: 95-120.Trejo, R. and V. Kreinovich. 2001. Error estimations for indirect measurements: randomized vs. deterministic
algorithms for ‘black-box’ programs. Handbook on Randomized Computing, S. Rajasekaran, P. Pardalos, J. Reif, and J. Rolim (eds.), Kluwer, 673–729. http://www.cs.utep.edu/vladik/2000/tr00-17.pdf
Vesely, W.E., F.F. Goldberg, N.H. Roberts, D.F. Haasl. 1981. Fault Tree Handbook. Nuclear Regulatory Commission, Washington, DC.
Vick, S.G. 2002. Degrees of Belief: Subjective Probability and Engineering Judgment. ASCE Press, Reston, Virginia.
![Page 104: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/104.jpg)
End
![Page 105: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/105.jpg)
![Page 106: Interval Analysis Scott Ferson, scott@ramas.com 11 September 2007, Stony Brook University, MAR 550, Challenger 165](https://reader030.vdocuments.net/reader030/viewer/2022032722/56649f505503460f94c736a3/html5/thumbnails/106.jpg)
For next time
• Do exercise #1 and either #2 or #3
• Try to do one of the exercises #5, #6 or #7