Survival Data survival time examples: –time a cancer patient is in remission –time til a disease-free person has a heart attack –time til death of a healthy.

Slides:



Advertisements
Similar presentations
Time to failure Probability, Survival,  the Hazard rate, and the Conditional Failure Probability.
Advertisements

Lecture (7) Random Variables and Distribution Functions.
1 Set #3: Discrete Probability Functions Define: Random Variable – numerical measure of the outcome of a probability experiment Value determined by chance.
Continuous Random Variables. L. Wang, Department of Statistics University of South Carolina; Slide 2 Continuous Random Variable A continuous random variable.
Part 9: Normal Distribution 9-1/47 Statistics and Data Analysis Professor William Greene Stern School of Business IOMS Department Department of Economics.
Probability Densities
Chapter 6 Introduction to Sampling Distributions
Probability Distributions Finite Random Variables.
Simulation Modeling and Analysis
SUMS OF RANDOM VARIABLES Changfei Chen. Sums of Random Variables Let be a sequence of random variables, and let be their sum:
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 6 Introduction to Sampling Distributions.
Discrete Probability Distributions
1 Review Definition: Reliability is the probability that a component or system will perform a required function for a given period of time when used under.
Continuous Random Variables and Probability Distributions
Inferences About Process Quality
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 4 Continuous Random Variables and Probability Distributions.
Discrete Random Variables and Probability Distributions
Chapter 21 Random Variables Discrete: Bernoulli, Binomial, Geometric, Poisson Continuous: Uniform, Exponential, Gamma, Normal Expectation & Variance, Joint.
1 Chapter 20 Two Categorical Variables: The Chi-Square Test.
Exponential Distribution & Poisson Process
1 Exponential Distribution & Poisson Process Memorylessness & other exponential distribution properties; Poisson process and compound P.P.’s.
1 Performance Evaluation of Computer Systems By Behzad Akbari Tarbiat Modares University Spring 2009 Introduction to Probabilities: Discrete Random Variables.
1 Reliability Application Dr. Jerrell T. Stracener, SAE Fellow Leadership in Engineering EMIS 7370/5370 STAT 5340 : PROBABILITY AND STATISTICS FOR SCIENTISTS.
-Exponential Distribution -Weibull Distribution
Simulation Output Analysis
NASSER DAVARZANI DEPARTMENT OF KNOWLEDGE ENGINEERING MAASTRICHT UNIVERSITY, 6200 MAASTRICHT, THE NETHERLANDS 22 OCTOBER 2012 Introduction to Survival Analysis.
The Poisson Process. A stochastic process { N ( t ), t ≥ 0} is said to be a counting process if N ( t ) represents the total number of “events” that occur.
Exponential and Chi-Square Random Variables
Statistical Distributions
Tch-prob1 Chap 3. Random Variables The outcome of a random experiment need not be a number. However, we are usually interested in some measurement or numeric.
Random Sampling, Point Estimation and Maximum Likelihood.
1 Normal Random Variables In the class of continuous random variables, we are primarily interested in NORMAL random variables. In the class of continuous.
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
More Continuous Distributions
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
Censoring an observation of a survival r.v. is censored if we don’t know the survival time exactly. usually there are 3 possible reasons for censoring.
STA347 - week 31 Random Variables Example: We roll a fair die 6 times. Suppose we are interested in the number of 5’s in the 6 rolls. Let X = number of.
The two way frequency table The  2 statistic Techniques for examining dependence amongst two categorical variables.
Random Variables Presentation 6.. Random Variables A random variable assigns a number (or symbol) to each outcome of a random circumstance. A random variable.
MATH 4030 – 4B CONTINUOUS RANDOM VARIABLES Density Function PDF and CDF Mean and Variance Uniform Distribution Normal Distribution.
Stats Probability Theory Summary. The sample Space, S The sample space, S, for a random phenomena is the set of all possible outcomes.
7 sum of RVs. 7-1: variance of Z Find the variance of Z = X+Y by using Var(X), Var(Y), and Cov(X,Y)
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Probability Distributions: Binomial & Normal Ginger Holmes Rowell, PhD MSP Workshop June 2006.
A random variable is a variable whose values are numerical outcomes of a random experiment. That is, we consider all the outcomes in a sample space S and.
Random Variable The outcome of an experiment need not be a number, for example, the outcome when a coin is tossed can be 'heads' or 'tails'. However, we.
Some Common Discrete Random Variables. Binomial Random Variables.
Probability –classical approach P(event E) = N e /N, where N = total number of possible outcomes, N e = total number of outcomes in event E assumes equally.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
1 Probability and Statistical Inference (9th Edition) Chapter 5 (Part 2/2) Distributions of Functions of Random Variables November 25, 2015.
We’ll now look at the relationship between a survival variable Y and an explanatory variable X; e.g., Y could be remission time in a leukemia study and.
Chapter 4 Continuous Random Variables and Probability Distributions  Probability Density Functions.2 - Cumulative Distribution Functions and E Expected.
Numerical parameters of a Random Variable Remember when we were studying sets of data of numbers. We found some numbers useful, namely The spread The.
A random variable is a variable whose values are numerical outcomes of a random experiment. That is, we consider all the outcomes in a sample space S and.
Computer Performance Modeling Dirk Grunwald Prelude to Jain, Chapter 12 Laws of Large Numbers and The normal distribution.
Virtual University of Pakistan Lecture No. 26 Statistics and Probability Miss Saleha Naghmi Habibullah.
Topic 19: Survival Analysis T = Time until an event occurs. Events are, e.g., death, disease recurrence or relapse, infection, pregnancy.
Chapter 4 Continuous Random Variables and Probability Distributions  Probability Density Functions.2 - Cumulative Distribution Functions and E Expected.
Copyright © Cengage Learning. All rights reserved. 4 Continuous Random Variables and Probability Distributions.
Sums of Random Variables and Long-Term Averages Sums of R.V. ‘s S n = X 1 + X X n of course.
Expectations of Random Variables, Functions of Random Variables
Chapter 4 Continuous Random Variables and Probability Distributions
ASV Chapters 1 - Sample Spaces and Probabilities
STT : Biostatistics Analysis Dr. Cuixian Chen
The Exponential and Gamma Distributions
Chapter 4 Continuous Random Variables and Probability Distributions
Continuous Random Variables
Review of survival models:
T305: Digital Communications
The hazard function The hazard function gives the so-called “instantaneous” risk of death (or failure) at time t, assuming survival up to time t. Estimate.
Presentation transcript:

Survival Data survival time examples: –time a cancer patient is in remission –time til a disease-free person has a heart attack –time til death of a healthy mouse –time til a computer component fails –time til a paroled prisoner gets rearrested –time til death of a liver transplant patient –time til a cell phone customer switches carrier –time til recovery after surgery all are "time til some event occurs" - longer times are better in all but the last…

estimate the survival function compare survival functions (e.g., across levels of a categorical variable - treatment vs. placebo) understand the relationship of the survival function to explanatory variables ( e.g., is survival time different for various values of an explanatory variable?) Three goals of survival analysis

The survival function S(y)=P(Y>y) can be estimated by the empirical survival function, which essentially gets the relative frequency of the number of Y’s > y… Look at Definition 1.3 on p.5: Y 1, …,Y n are i.i.d. (independent and identically distributed) survival variables. Then S n (y) =empirical survival function at y = (# of the Y’s > y)/n = estimate of S(y). Note that where I is the indicator function…

Review of Bernoulli & Binomial RVs: –Show that the expected value of a Bernoulli rv Z with parameter p (i.e., P(Z=1)=p) is p and that the variance of Z is p(1-p) –Then knowing that the sum of n iid (independent and identically distributed) Bernoullis is a Binomial rv with parameters n and p, show on the next slide that the empirical survivor function S n (y) is an unbiased estimator of S(y)

Note that and as such nS n has B(n,p) where p=P(Y>y)=S(y). Also note that for a fixed y * so S n is unbiased as an estimator of S What is the Var(S n )? (see 1.6 and on p.6 where the confidence interval is computed…) Try this for Example 1.3, p.6

Example 1.4 on page 8 shows that it is sometimes difficult to compare survival curves since they can cross each other… (what makes one survival curve “better” than another?) One way of comparing two survival curves is by comparing their MTTF (mean time til failure) values. Let’s try to use R to draw the two curves given in Ex. 1.4: S 1 (y)=exp(-y/2) and S 2 (y)=exp(-y 2 /4)… see the handout R#1.

Note that the MTTF of a survival rv Y is just its expected value E(Y). We can also show (Theorem 1.2) that (Math & Stat majors: Show this is true using integration by parts and l’Hospital’s rule…!) So suppose we have an exponential survival function: (btw, can you show this satisfies the properties of a survival function?)

Then the MTTF for this variable is  - show this… And for any two such survival functions, S 1 (y)=exp(-y/    and S 2 (y)=exp(-y/    one is “better” than the other if the corresponding beta is “better”… HW: Use R to plot on the same axes at least two such survival functions with different values of beta and show this result.

The hazard function The hazard function gives the so-called “instantaneous” risk of death (or failure) at time t. Recall that for continuous rvs, the probability of occurrence at time t is 0 for all t. So we think about the probability in a “small” interval around t, given that we’ve survived to t, and then let the small interval go to zero (in the limit). The result is given on page 9 as the hazard rate or hazard function…

Definition of hazard function: notes –the hazard function is conditional on the individual having already survived to time y –the numerator is a non-decreasing function of  y (it is more likely that Y will occur in a longer interval) so we divide by the length of the interval to compensate –we take the limit as the length of the interval gets smaller to get the risk at exactly y - “instantaneous risk”

–we can show (see p.9) that the hazard function is equal to –use f(y)=-d/dy(S(y)) and the above to show that –so all three of f, h, and S are representations that can be found from the others and are used in various situations…

more notes on the hazard function: –hazard is in the form of a rate - hazard is not a probability because it can be >1, but the hazard must be > 0; so the graph of h(y) does not have to look at all like that of a survivor function –in order to understand the hazard function, it must be estimated. –think of the hazard h(y) as the instantaneous risk the event will occur per unit time, given that the event has not occurred up to time y. –note that for given y, a larger S(y) corresponds to a smaller h(y) and vice versa…

life expectancy at age t: –if Y=survival time and we know that Y>t, then Y-t=residual lifetime at age t and the mean residual lifetime at age t is the conditional expectation E(Y-t|Y>t) = r(t) –it can be shown that –note that when t=0, r(0)=MTTF –we define the mean life expectancy at age t as E(Y|Y>t) = t +r(t) –go over Example 1.6 on page 11…