Chapter 8 Continuous Time Markov Chains. Markov Availability Model.

Slides:



Advertisements
Similar presentations
A. BobbioReggio Emilia, June 17-18, Dependability & Maintainability Theory and Methods 3. Reliability Block Diagrams Andrea Bobbio Dipartimento di.
Advertisements

A. BobbioReggio Emilia, June 17-18, Dependability & Maintainability Theory and Methods 5. Markov Models Andrea Bobbio Dipartimento di Informatica.
COE 444 – Internetwork Design & Management Dr. Marwan Abu-Amara Computer Engineering Department King Fahd University of Petroleum and Minerals.
Lecture 6  Calculating P n – how do we raise a matrix to the n th power?  Ergodicity in Markov Chains.  When does a chain have equilibrium probabilities?
SECURITY AND DEPENDABILITY OF COMPUTER SYSTEMS
Fault Tree Analysis Part 12 – Redundant Structure and Standby Units.
6. Reliability Modeling Reliable System Design 2010 by: Amir M. Rahmani.
1 Chapter 5 Continuous time Markov Chains Learning objectives : Introduce continuous time Markov Chain Model manufacturing systems using Markov Chain Able.
Reliability Block Diagrams A reliability block diagram is a success-oriented network describing the function of the system. If the system has more than.
Markov Analysis Jørn Vatn NTNU.
Lecture 12 – Discrete-Time Markov Chains
Solutions Markov Chains 1
Oct. 2007State-Space ModelingSlide 1 Fault-Tolerant Computing Motivation, Background, and Tools.
Probability and Statistics with Reliability, Queuing and Computer Science Applications: Chapter 8 on Continuous-Time Markov Chains Kishor Trivedi.
Markov Reward Models By H. Momeni Supervisor: Dr. Abdollahi Azgomi.
NETE4631:Capacity Planning (3)- Private Cloud Lecture 11 Suronapee Phoomvuthisarn, Ph.D. /
Reliable System Design 2011 by: Amir M. Rahmani
Lecture 13 – Continuous-Time Markov Chains
Oct State-Space Modeling Slide 1 Fault-Tolerant Computing Motivation, Background, and Tools.
1 Software Testing and Quality Assurance Lecture 36 – Software Quality Assurance.
A. BobbioBertinoro, March 10-14, Dependability Theory and Methods 5. Markov Models Andrea Bobbio Dipartimento di Informatica Università del Piemonte.
While there is a generally accepted precise definition for the term "first order differential equation'', this is not the case for the term "Bifurcation''.
Ch 5.5: Series Solutions Near a Regular Singular Point, Part I We now consider solving the general second order linear equation in the neighborhood of.
A. BobbioReggio Emilia, June 17-18, Dependability & Maintainability Theory and Methods Part 2: Repairable systems: Availability Andrea Bobbio Dipartimento.
Lecture #8 Circuits with Capacitors
Sample Space Probability implies random experiments.
Chapter 6 Time dependent reliability of components and system.
CDA6530: Performance Models of Computers and Networks Examples of Stochastic Process, Markov Chain, M/M/* Queue TexPoint fonts used in EMF. Read the TexPoint.
Chapter 7. First and second order transient circuits
Transient Analysis - First Order Circuits
Chapter 2 Machine Interference Model Long Run Analysis Deterministic Model Markov Model.
A. BobbioBertinoro, March 10-14, Dependability Theory and Methods 2. Reliability Block Diagrams Andrea Bobbio Dipartimento di Informatica Università.
Lecture 2: Combinatorial Modeling CS 7040 Trustworthy System Design, Implementation, and Analysis Spring 2015, Dr. Rozier Adapted from slides by WHS at.
By Behzad Akbari Tarbiat Modares University Spring 2009
Solutions Markov Chains 2 3) Given the following one-step transition matrix of a Markov chain, determine the classes of the Markov chain and whether they.
Copyright © Cengage Learning. All rights reserved. 14 Partial Derivatives.
Chapter 61 Continuous Time Markov Chains Birth and Death Processes,Transition Probability Function, Kolmogorov Equations, Limiting Probabilities, Uniformization.
Courtesy of J. Akinpelu, Anis Koubâa, Y. Wexler, & D. Geiger
1 Component reliability Jørn Vatn. 2 The state of a component is either “up” or “down” T 1, T 2 and T 3 are ”Uptimes” D 1 and D 2 are “Downtimes”
Fault-Tolerant Computing Systems #4 Reliability and Availability
Approximate Three-Stage Model: Active Learning – Module 3 Dr. Cesar Malave Texas A & M University.
Copyright © Cengage Learning. All rights reserved The Chain Rule.
NETE4631: Network Information System Capacity Planning (2) Suronapee Phoomvuthisarn, Ph.D. /
Continuous Time Markov Chains
Meaning of Markov Chain Markov Chain is a stochastic model describing a sequence of possible events in which the probability of each event depends only.
Fault Tree Analysis Part 11 – Markov Model. State Space Method Example: parallel structure of two components Possible System States: 0 (both components.
Part.2.1 In The Name of GOD FAULT TOLERANT SYSTEMS Part 2 – Canonical Structures Chapter 2 – Hardware Fault Tolerance.
Reliability of Disk Systems. Reliability So far, we looked at ways to improve the performance of disk systems. Next, we will look at ways to improve the.
16111 presentation by Arindam Das System Reliability Prediction A presentation by Arindam Das.
Reliability Engineering
Copyright 2007 Koren & Krishna, Morgan-Kaufman Part.4.1 FAULT TOLERANT SYSTEMS Part 4 – Analysis Methods Chapter 2 – HW Fault Tolerance.
Expectations of Random Variables, Functions of Random Variables
Discrete and Continuous Randonm Variables; Group Activity Solution
Discrete-time Markov chain (DTMC) State space distribution
Chapter 12 Design via State Space <<<4.1>>>
Discrete-time markov chain (continuation)
Much More About Markov Chains
Fault Tolerance & Reliability CDA 5140 Spring 2006
Fault-Tolerant Computing Systems #5 Reliability and Availability2
Software Reliability: 2 Alternate Definitions
T305: Digital Communications
Dept. of Electrical & Computer engineering
Solutions Markov Chains 1
Reliability Engineering
Measurements of Inductance & Capacitance
Problem Sheet 5 Question 4
Sample Space Probability implies random experiments.
Chapter 6 Time dependent reliability of components and system
Reliable Web Services: Methodology, Experiment and Modeling International Conference on Web Services (ICWS 2007) Pat. P. W. Chan, Michael R. Lyu Department.
Chapter 2 Machine Interference Model
Presentation transcript:

Chapter 8 Continuous Time Markov Chains

Markov Availability Model

2-State Markov Availability Model 1) Steady-state balance equations for each state: –Rate of flow IN = rate of flow OUT State1: State0: 2 unknowns, 2 equations, but there is only one independent equation. UP 1 DN 0

2-State Markov Availability Model (Continued) 1) Need an additional equation: Downtime in minutes per year = * 8760*60

2-State Markov Availability Model (Continued) 2) Transient Availability for each state: –Rate of buildup = rate of flow IN - rate of flow OUT This equation can be solved to obtain assuming P 1 (0)=1

2-State Markov Availability Model (Continued) 3) 4) Steady State Availability:

Assume we have a two-component parallel redundant system with repair rate . Assume that the failure rate of both the components is. When both the components have failed, the system is considered to have failed. Markov availability model

Markov availability model (Continued) Let the number of properly functioning components be the state of the system. The state space is {0,1,2} where 0 is the system down state. We wish to examine effects of shared vs. non-shared repair.

Non-shared (independent) repair Shared repair Markov availability model (Continued)

Note: Non-shared case can be modeled & solved using a RBD or a FTREE but shared case needs the use of Markov chains. Markov availability model (Continued)

Steady-state balance equations For any state: Rate of flow in = Rate of flow out Consider the shared case  i : steady state probability that system is in state i

Steady-state balance equations (Continued) Hence Since We have or

Steady-state balance equations (Continued) Steady-state unavailability =  0 = 1 - A shared Similarly for non-shared case, steady-state unavailability = 1 - A non-shared Downtime in minutes per year = (1 - A)* 8760*60

Steady-state balance equations

Homework 5: Return to the 2 control and 3 voice channels example and assume that the control channel failure rate is c, voice channel failure rate is v. Repair rates are  c and  v, respectively. Assuming a single shared repair facility and control channel having preemptive repair priority over voice channels, draw the state diagram of a Markov availability model. Using SHARPE GUI, solve the Markov chain for steady-state and instantaneous availability.

Markov Reliability Model

Consider the 2-component parallel system but disallow repair from system down state Note that state 0 is now an absorbing state. The state diagram is given in the following figure. This reliability model with repair cannot be modeled using a reliability block diagram or a fault tree. We need to resort to Markov chains. (This is a form of dependency since in order to repair a component you need to know the status of the other component). Markov reliability model with repair

Markov chain has an absorbing state. In the steady-state, system will be in state 0 with probability 1. Hence transient analysis is of interest. States 1 and 2 are transient states. Markov reliability model with repair (Continued) Absorbing state

Assume that the initial state of the Markov chain is 2, that is, P 2 (0) = 1, P k (0) = 0 for k = 0, 1. Then the system of differential Equations is written based on: rate of buildup = rate of flow in - rate of flow out for each state Markov reliability model with repair (Continued)

After solving these equations, we get R(t) = P 2 (t) +P 1 (t) Recalling that, we get: Markov reliability model with repair (Continued)

Note that the MTTF of the two component parallel redundant system, in the absence of a repair facility (i.e.,  = 0), would have been equal to the first term, 3 / ( 2* ), in the above expression. Therefore, the effect of a repair facility is to increase the mean life by  / (2* 2 ), or by a factor Markov reliability model with repair (Continued)

Markov Reliability Model with Imperfect Coverage

Markov model with imperfect coverage Next consider a modification of the above example proposed by Arnold as a model of duplex processors of an electronic switching system. We assume that not all faults are recoverable and that c is the coverage factor which denotes the conditional probability that the system recovers given that a fault has occurred. The state diagram is now given by the following picture:

Now allow for Imperfect coverage c

Markov model with imperfect coverage (Continued) Assume that the initial state is 2 so that: Then the system of differential equations are:

Markov model with imperfect coverage (Continued) After solving the differential equations we obtain: R(t)=P 2 (t) + P 1 (t) From R(t), we can system MTTF: It should be clear that the system MTTF and system reliability are critically dependent on the coverage factor.

SOURCES OF COVERAGE DATA Measurement Data from an Operational system: Large amount of data needed; Improved Instrumentation Needed Fault/Error Injection Experiments Costly yet badly needed: tools from CMU, Illinois, Toulouse

SOURCES OF COVERAGE DATA (Continued) A Fault/Error Handling Submodel Phases of FEHM: Detection, Location, Retry, Reconfig, Reboot Estimate Duration & Prob. of success of each phase IBM(EDFI), HARP(FEHM), Draper(FDIR)

Homework 6: Modify the Markov model with imperfect coverage to allow for finite time to detect as well as imperfect detection. You will need to add an extra state, say D. The rate at which detection occurs is . Draw the state diagram and using SHARPE GUI investigate the effects of detection delay on system reliability and mean time to failure.