Download presentation
Presentation is loading. Please wait.
1
The Infeasibility of Quantifying the Reliability of Life-Critical Real-Time Software
2
introduction -The availability of enormous computing power at a low cost has led to expanded use of digital computers in current applications and their introduction into many new applications. -Increased performance at a minimal hardware cost. -Software systems which contain more errors.
3
Terminology: Failure rate per hour: Ultra reliability = < 10 -7 Moderate reliability= 10 -3 to 10 -7 Low reliability= > 10 -3 -Software errors behaves like a stochastic point process. -In a real-time system, the software is periodically scheduled- the probability of software failure is given by the binomial distribution : Software Reliability
4
p(Sn = k) = p k (1- p) n-k P(sn > 0) = 1 – (1-p) n = 1 – (1 – p) kt k – number of inputs per unit time. To simplify: P(Sn > n) = 1- e -ktp
5
Analyzing Software as a Black Box 1. Testing with replacement - D t = y 0 * (r/n) 2. Testing without replacement- D t = y 0* Y 0 - mean failure time of a test specimen. For probability of failure of 10 –9 for a 10 hour mission: y 0 = 10 / -ln(1- 10 –9 ) 10 10
6
110 10 hours = 1141550 years 1010 9 hours = 114155 years 100108 hours = 11415 years 100010 6 hours = 114 years (r = 1) No. of replicates (n) Expected Test Duration Dt
7
Reliability Growth Models -The software design involves a repetitive cycle of testing and repairing a program. The result is a sequence of programs : p 1, …. p n and a sequence of failure times, t 1,…. t n.. -The goal is the predict the reliability of the p n.. Experiment performed by Nagel and Skrivan: Program A1: number of bugs Removed failure probability per input 10.9803 20.1068 30.002602 40.002104 50.001176 60.0007659
9
Calculation the requirements per input : p = -ln(1- p aye ) / Kt P aye = 10 -9 for a 10 hour mission, k = 10/sec then: P = 2.78 * 10 -15 Extrapolation to predict when ultra reliability will be reached
10
-To get a rate of 2.78*10 -15 you need about 24 bugs. -Bug 23 will have a failure rate of about 9.38*10 -15, the expected number of test cases until observing a binomial event of probability 9.38*10 -15 is 1.07*10 -14. -If each test case would require 0.10 sec then the expected time to discover bug 23 alone would be 1.07*10 13 sec or 3.4*10 5 years.
11
ProgramslopeLast bugTest time A1 - 1.145 233.4 *10 5 years B1 - 1.3358 253.3 *10 5 years A2 - 1.998 171.5 *10 5 years B2 - 3.628 94.5 *10 5 years A3 -. 54526 586.8 *10 5 years b3 - 1.3138 255.3 *10 5 years Results for 5 different programs:
12
Low Sample Rate Systems and Accelerated Testing R = test time per input 1/p = number of inputs until the next bug appears D t = R/p Therefore D t = RKt / -ln(1 - p aye ). K = number of inputs per unit time.
13
K (R = 0.1) Expected Test Time, D t 10/sec1.14*10 6 years 1/sec1.14*10 5 years 1/minute1.9*10 3 years 1/hour31.7 years 1/day1.32 years 1/month16 days
14
Reliability Growth Models and Accelerated Testing If the sample rate is 1 input per minute then the failure rate per input must be less than 10 -9 /60 = 1.67*10 -11 bug failure rate per input -The removal of the last bug alone would take approximately 2.2*10 10 test cases. Even if the testing process were 60/1000 sec testing would take 42 years 161.87*10 -10 174.55*10 -11 181.1*10 -11
15
Summary for all the programs: Test Time To Remove the Last Bug to Obtain Ultra reliability ProgramSlopeLast bugTest time A1 - 1.41 1742 years B1-1.33581966 years A2-1.9981331 years B2-3.623719 years A3-.545264266 years B3-1.31381932 years
16
Models of Software Fault Tolerance -independence assumption enables quantification in the ultra reliable region -Quantification of fault-tolerant software reliability is unlikely without the independence assumption -independence assumption cannot be experimentally justified for ultra reliable region
17
E i,k = The event that the I version fails on its k execution. P i,k = The probability that version I fails during the k execution. -The probability that two or more versions fail on the k th execution : P aye,k = P( (E 1,k ^E 2,k ) or (E 1,k ^E 3,k )or (E 2,k ^E 3,k ) or (E 1,k ^ E 2,k ^E 3,k )) = P(E 1,k ^E 2,k ) + P (E 1,k ^E 3,k )+ P(E 2,k ^E 3,k ) - 2P(E 1,k ^ E 2,k ^E 3,k ). = P(E 1,k )P(E 2,k ) + P(E 1,k )P(E 3,k ) + P(E 2,k )P(E 3,k ) – 2P(E 1,k )P(E 2,k )P(E 3,k ) P aye,k = 3p 2 - 2 p 3 3p 2 Independence enables quantification of ultra reliability
18
P aye (T) = 1- e (-3p^2*KT) 3p 2 KT If T = 1,k = 3600 (1 execution per second) and P(E 1,k ) = 10 -6 then we get P aye (T) = 1.08*10 -8
19
P aye = P(E 1 ^E 2 ) + P (E 1 ^E 3 )+ P(E 2 ^E 3 ) - 2P(E 1 ^ E 2 ^E 3 ). = P(E 1 )P(E 2 ) + P(E 1 )P(E 3 )+P(E 2 )P(E 3 )-2P(E 1 )P(E 2 )P(E 3 ) +[P(E 1 ^ E 2 ) - P(E 1 )P(E 2 )] +[P(E 1 ^ E 3 ) - P(E 1 )P(E 3 )] +[P(E 2 ^ E 3 ) - P(E 2 )P(E 3 )] -2[P(E 1 ^ E 2 ^ E 3 ) - P(E 1 ) P(E 2 )P(E 3 )] - P(E 1 ^ E 2 ^ E 3 ) < P(E i ^ E j ) therefore P(E i ^ E j ) < P aye Ultra reliable Quantification Is Infeasible Without Independence
20
Example1: E 1 = E 2 = E 3 = 10 -5 If independent then p(E i ^Ej) = 10 -10 -If p(E i ^Ej) = 10 -7 /hour one could test for a 100 years and not seen even one coincident error. Example2: E 1 = E 2 = E 3 = 10 -4 -If p(E i ^Ej) = 10 -4 /hour one could test for a one years and not likely see even one coincident error!! Danger Of Extrapolation to the Ultra reliability Region
21
-In the second case if the erroneous assumption of independence would be made then it would allow the assignment of a 3*10 -8 probability of failure to the system when in reality the system is no better than 10 -5. -In order to get probability of failure to be less than 10 -9 at 1 hour we need p(E i ^Ej) to be less then 10 -12
22
There are two kinds of models: 1.The model includes terms which cannot be measured within feasible amounts of time. 2.The model includes only parameters which can be measured within feasible amounts of time. -A general model must provide a mechanism that makes the interaction terms negligibly small. - There is little hope of deriving the interaction terms from fundamental Laws, since the error process occurs in the human mind. Feasibility of a General Model For Coincident Errors
23
The Coincident-Error Experiments Experiment that was performed by Knight and Leveson: -27 versions of a program were produced and subjected to 1,000,000 input cases. -The observed average failure rate per input was 0.0007. -independence model was rejected. -In order to observe the errors the error rate must be in the low to moderate reliability region. Future experiments will have one of the following results :
24
1.Demonstration that the independence assumption does not hold for the low reliability system 2. Demonstration that the independence assumption does hold from systems for the low reliability system. 3. No coincident errors were seen but the test time was insufficient to demonstrate independence for the potentially ultra reliable system.
25
Conclusions The potential performance advantages of using computers over their analog predecessors have created an atmosphere where serious safety concerns about digital hardware and software are not adequately addressed. Practical methods to prevent design errors have not been found.
26
Life testing of ultra reliable software is infeasible. (i.e. to quantify 10 -8 /hour failure rate requires more than 10 8 hours o testing). The assumption of independence is not reasonable for software and can not be tested for software or for hardware in the ultra reliable region. It is possible that models which are inferior to other models in the moderate region are superior in the ultra reliable region – but this cannot be demonstrated.
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.