Critical Appraisal Course for Emergency Medicine Trainees Module 5 Evaluation of a Diagnostic Test.

Slides:



Advertisements
Similar presentations
Likelihood ratios Why these are the most thrilling statistics in Emergency Medicine.
Advertisements

Lecture 3 Validity of screening and diagnostic tests
Step 3: Critically Appraising the Evidence: Statistics for Diagnosis.
TESTING A TEST Ian McDowell Department of Epidemiology & Community Medicine November, 2004.
Is it True? Evaluating Research about Diagnostic Tests
Critically Evaluating the Evidence: diagnosis, prognosis, and screening Elizabeth Crabtree, MPH, PhD (c) Director of Evidence-Based Practice, Quality Management.
Azita Kheiltash Social Medicine Specialist Tehran University of Medical Sciences Diagnostic Tests Evaluation.
Evaluation of Diagnostic Test Studies
GerstmanChapter 41 Epidemiology Kept Simple Chapter 4 Screening for Disease.
Baye’s Rule and Medical Screening Tests. Baye’s Rule Baye’s Rule is used in medicine and epidemiology to calculate the probability that an individual.
By Dr. Ahmed Mostafa Assist. Prof. of anesthesia & I.C.U. Evidence-based medicine.
Evidence Based Medicine for the Athletic Trainer: What is It?
(Medical) Diagnostic Testing. The situation Patient presents with symptoms, and is suspected of having some disease. Patient either has the disease or.
Screening and Early Detection Epidemiological Basis for Disease Control – Fall 2001 Joel L. Weissfeld, M.D. M.P.H.
Statistics in Screening/Diagnosis
Diagnosis Articles Much Thanks to: Rob Hayward & Tanya Voth, CCHE.
DEB BYNUM, MD AUGUST 2010 Evidence Based Medicine: Review of the basics.
Diagnostic Cases. Goals & Objectives Highlight Bayesian and Boolean processes used in classic diagnosis Demonstrate use/misuse of tests for screening.
Lecture 4: Assessing Diagnostic and Screening Tests
Diagnostic Testing Ethan Cowan, MD, MS Department of Emergency Medicine Jacobi Medical Center Department of Epidemiology and Population Health Albert Einstein.
Vanderbilt Sports Medicine How to practice and teach EBM Chapter 3 May 3, 2006.
Evidence Based Medicine Workshop Diagnosis March 18, 2010.
EBCP. Random vs Systemic error Random error: errors in measurement that lead to measured values being inconsistent when repeated measures are taken. Ie:
Evaluation of Diagnostic Tests
EVIDENCE ABOUT DIAGNOSTIC TESTS Min H. Huang, PT, PhD, NCS.
Diagnosis: EBM Approach Michael Brown MD Grand Rapids MERC/ Michigan State University.
1 Epidemiological Measures I Screening for Disease.
Evaluating Diagnostic Tests Payam Kabiri, MD. PhD. Clinical Epidemiologist Tehran University of Medical Sciences.
Appraising A Diagnostic Test
Likelihood 2005/5/22. Likelihood  probability I am likelihood I am probability.
1. Statistics Objectives: 1.Try to differentiate between the P value and alpha value 2.When to perform a test 3.Limitations of different tests and how.
Prediction statistics Prediction generally True and false, positives and negatives Quality of a prediction Usefulness of a prediction Prediction goes Bayesian.
Positive Predictive Value and Negative Predictive Value
Diagnostic Test Characteristics: What does this result mean
Screening.  “...the identification of unrecognized disease or defect by the application of tests, examinations or other procedures...”  “...sort out.
Evidence based medicine Diagnostic tests Ross Lawrenson.
1 Medical Epidemiology Interpreting Medical Tests and Other Evidence.
/ 101 Saudi Diploma in Family Medicine Center of Post Graduate Studies in Family Medicine EBM Diagnostic Tests Dr. Zekeriya Aktürk
10 May Understanding diagnostic tests Evan Sergeant AusVet Animal Health Services.
Laboratory Medicine: Basic QC Concepts M. Desmond Burke, MD.
Diagnosis Examination(MMSE) in detecting dementia among elderly patients living in the community. Excel.
Timothy Wiemken, PhD MPH Assistant Professor Division of Infectious Diseases Diagnostic Tests.
Diagnosis:Testing the Test Verma Walker Kathy Davies.
Is suicide predictable? Paul St John-Smith Short Courses in Psychiatry 15/10/2008.
Accuracy and usefulness of a clinical prediction rule and D-dimer testing in excluding deep vein thrombosis in cancer patients Thrombosis Research (2008)
GP ST2 Group, 28/9/11 Tom Gamble
Diagnostic studies Adrian Boyle.
Performance of a diagnostic test Tunisia, 31 Oct 2014
DR.FATIMA ALKHALEDY M.B.Ch.B;F.I.C.M.S/C.M
Diagnostic Test Studies
When is the post-test probability sufficient for decision-making?
Sensitivity and Specificity
Evidence-Based Medicine
M. Lee Chambliss MD MSPH Associate Professor
Comparing Accuracy of Ultrasound to Chest X-ray in Determining Pulmonary Edema Matt Wooten, DO PGY 2 5/4/2017.
Class session 7 Screening, validity, reliability
EPIDEMIOLOGICAL METHOD TO DETERMINE UTILITY OF A DIAGNOSTIC TEST
بسم الله الرحمن الرحيم Clinical Epidemiology
What is Screening? Basic Public Health Concepts Sheila West, Ph.D.
How do we delay disease progress once it has started?
Diagnosis II Dr. Brent E. Faught, Ph.D. Assistant Professor
Hint: Numerator Denominator. Vascular Technology Lecture 34: Test Validation (Statistical Profile and Correlation) HHHoldorf.
Refining Probability Test Informations Vahid Ashoorion MD. ,MSc,
Diagnosis General Guidelines:
Figure 1. Table for calculating the accuracy of a diagnostic test.
Patricia Butterfield & Naomi Chaytor October 18th, 2017
Critical Appraisal วิจารณญาณ
Evidence Based Diagnosis
Clinical Effectiveness: sensitivity, specificity and PPV
Presentation transcript:

Critical Appraisal Course for Emergency Medicine Trainees Module 5 Evaluation of a Diagnostic Test

Evaluating a diagnostic test The reference standard Independence of the reference standard Blinding Patient spectrum Inter-observer error Sensitivity, specificity, PPV, NPV Likelihood ratios

The reference (gold) standard Used to determine whether the patient has the disease or not Possible reference standards: A single accurate diagnostic test A combination of tests Tests with follow-up for negatives Often involves trade-off between accuracy and feasibility of reference standard

Independent reference standard Ideally all patients should get the same reference standard test If not, the results of the test under investigation should not determine which reference standard test is used Work-up bias: Patients with +ve test get one reference standard, those with –ve test get another Incorporation bias: The test under investigation is incorporated into the reference standard test

Blinding Was the test under investigation interpreted blind to the results of the reference standard? Was the reference standard interpreted blind to the results of the test under investigation? Awareness either way may influence interpretation

Patient spectrum Study population should be representative of the population in which the test will be used How were patients selected? What was disease prevalence? High prevalence suggests a selected population

Inter-observer error (reliability) Does the test give the same results when interpreted by different people? Measurement of reliability needs to take into account agreement due to chance Usually measured by Kappa Kappa=0: any agreement was due to chance Kappa=1: perfect agreement

What do the following mean? Case positive Case negative Test positive Test negative Prevalence

What do the following mean? Case positive: person with disease (positive reference standard) Case negative: person without disease (negative reference standard) Test positive: person with positive test under investigation Test negative: person with negative test under investigation Prevalence: proportion with disease

What do the following mean? True positive True negative False positive False negative

What do the following mean? True positive: positive test result & has disease True negative: negative test result & does not have disease False positive: positive test result but does not have disease False negative: negative test result but has disease

What do the following mean? Sensitivity Specificity Positive predictive value Negative predictive value

What do the following mean? Sensitivity: proportion of patients with disease who have a positive test Specificity: proportion of patients without disease who have a negative test Positive predictive value: proportion of patients with a positive test who have disease Negative predictive value: proportion of patients with a negative test who do not have disease

How are parameters used? Sensitivity: If a sensitive test is negative disease is ruled out (SnOut) Specificity: If a specific test is positive disease is ruled in (SpIn) These are useful, simple rules but not strictly true Likelihood ratios are better for decision- making

Prevalence Sensitivity and specificity are mathematically independent of prevalence Positive predictive value increases when prevalence increases Negative predictive value decreases when prevalence increases

Likelihood ratio Applies to a piece of diagnostic information, such as an observation, a clinical finding or a test result Tells you how diagnostically useful that piece of information is Is a number between zero and infinity If greater than one, indicates that the information increases the likelihood of the suspected diagnosis If less than one, indicates that the information decreases the likelihood of the suspected diagnosis

Diagnostic value of likelihood ratios 1None at all 0.5 to 2Little clinical significance 2 to 5Moderately increases likelihood of disease, but does not rule-in 0.2 to 0.5Moderately decreases likelihood of disease, but does not rule-out 5 to 10Markedly increases likelihood of disease, may rule-in 0.1 to 0.2Markedly decreases likelihood of disease, may rule-out >10Rules in <0.1Rules out

Examples: Clinical assessment for DVT FeatureLikelihood ratio of positive finding Likelihood ratio of negative finding Past history VTE Malignancy Immobilisation Recent surgery Difference in calf diameter Homan’s sign Oedema

Calculating likelihood ratios LR of +ve test = sensitivity / (1-specificity) LR of –ve test = (1-sensitivity) / specificity

Summary What reference standard was used? Was it independent of the test under investigation? Was the test under investigation interpreted blind? Was the reference standard interpreted blind? Was an appropriate spectrum of patients included? Was inter-observer error measured? What were the sensitivity & specificity? What were the likelihood ratios?

Any questions or comments?