Studying the Impact of Tests

Slides:



Advertisements
Similar presentations
Assessing quality of asthma care in your practice-- The Asthma APGAR
Advertisements

Workshop C – Evaluation Rod Taylor Complex Interventions Research Framework Masterclass 2010.
THE 2004 LIVING CONDITIONS MONITORING SURVEY : ZAMBIA EXTENT TO WHICH GENDER WAS INCORPORATED presented at the Global Forum on Gender Statistics, Accra.
Stephen Ryan, PwSI Lead, Newquay, 2 nd March 2005 GPwSIs – National Perspective.
Learning from managed care in mental health Dr Richard Ford Director.
Development and Implementation of a Recovery-Based System: Comparison of Instruments for Assessing Recovery Jeanette M. Jerrell, Ph.D. Professor of Neuropsychiatry,
Emma Frew Introduction to health economics, MSc HEHP, October 2012 Outcomes: part II.
Is research working for you? A self-assessment tool and discussion guide Maria Judd, MSc Senior Program Officer Knowledge Transfer and Exchange.
THE RATIONAL USE OF ANTIDEPRESSANTS COMBINED WITH BENZODIAZEPINES Izabela Fulone Silvio Barberato Filho Luciane Cruz Lopes THIRD INTERNACIONAL CONFERENCE.
ROBERT D. KERNS, PH.D. NATIONAL PROGRAM DIRECTOR FOR PAIN MANAGEMENT
Rod Jackson EPIQ group University of Auckland, NZ
Utilization of Integrated Hepatitis Services Among People with a History of Injection Drug Use in an STD Clinic 2004 National STD Prevention Conference.
USE OF EVIDENCE IN DECISION MODELS: An appraisal of health technology assessments in the UK Nicola Cooper Centre for Biostatistics & Genetic Epidemiology,
USE OF EVIDENCE IN DECISION MODELS: An appraisal of health technology assessments in the UK Nicola Cooper Centre for Biostatistics & Genetic Epidemiology,
How Clinician-Patient Communication Can Improve Health Outcomes Richard L. Street, Jr. Texas A&M University June 8, 2010.
Understanding p-values Annie Herbert Medical Statistician Research and Development Support Unit
IAPT-SMI Demonstration Site for Psychosis Professor Philippa Garety
Lecture 3 Validity of screening and diagnostic tests
Knowledge for Knowledge Translation Jeremy Grimshaw MD, PhD Clinical Epidemiology Program, OHRI Department of Medicine, University of Ottawa Canada Research.
Mark Jit Modelling and Economics Unit Health Protection Agency, London Case study: Structural uncertainty in human papillomavirus vaccination modelling.
Challenges in evaluating social interventions: would an RCT design have been the answer to all our problems? Lyndal Bond, Kathryn Skivington, Gerry McCartney,
PROCESS vs. WA State SCS Study A Comparison of Study Design, Patient Population, and Outcomes August 29,2007.
The geometric mean is the positive number x such that: A X X B.
ORGANIC Peptic ulcer Reflux Cholelithiasis Pancreatic Malignancy Musculoskeletal Drugs, medication and rarities ischaemic metabolic.
Patient Preference and Comprehensive Cohort Designs.
NIHR Research Design Service London Enabling Better Research Forming a research team Victoria Cornelius, PhD Senior Lecturer in Medical Statistics Deputy.
Supporting people in Dorset to lead healthier lives Commissioning the Dorset Community Persistent Pain Management Service Why is it so Painful to Commission.
Copyright 2011 Right Care Integrating Pathway Hub: A Prime Contractor Model Dr. Alan Nye Director Pennine MSK Partnership Ltd January 2014 Online Learning.
Journal Club Usha Niranjan PICU. Rationale 2 x cases of severe dehydration with metabolic acidosis –requesting for HDU management –as given 40mls/kg fluid.
Evidence Based Low Back Pain – Concord Hospital Pilot Project Diane Olimpio, PT, MS October 21, 2014.
Clare A Mackie Centre for Partnerships in Medicines for Health Economic Evaluation of a RCT of a ‘Medication Review Clinic’ in Patients Receiving Repeat.
Journal Club Alcohol and Health: Current Evidence July–August 2005.
The Cost-Effectiveness and Value of Information Associated with Biologic Drugs for the Treatment of Psoriatic Arthritis Y Bravo Vergel, N Hawkins, C Asseburg,
What type of information do service providers/ commissioners need? – good quality evidence to underpin service delivery/commissioning Screening Matched.
Self-Report Measures of Functional Status and Quality of Life: Adults Sandra A. Mitchell, CRNP, M.Sc.N., AOCN National Institutes of Health Bethesda, MD.
The Royal Stoke MS Centre - hub and spoke model
How to use Clinical Evidence to inform clinical decision making A case presentation using the CE review on acne.
The importance of musculoskeletal health problems Up to 30% of all GP consultations relate to musculoskeletal problems, and this area accounts for a £230m.
DISCUSSION Alex Sutton Centre for Biostatistics & Genetic Epidemiology, University of Leicester.
School of Healthcare FACULTY OF MEDICINE AND HEALTH The long road to influencing evidence-based NHS decision making; the NICE low back pain story Kate.
Complementary and integrative Medicine; George Lewith – Professor of Health Research School for Primary Care Research The.
Monitoring and evaluation: making your work count Dr Liz Weekes Consultant Dietitian & NIHR Clinical Lecturer Department of Nutrition & Dietetics Guy’s.
Evaluation Trials and Studies Coordinating Centre 5 July 2013 NIHR Programmes and topic identification Alison Ford, Senior Programme Manager.
Invention for Innovation (i4i) Bev Luchmun Industry Lead NISCHR
EVIDENCE BASED MEDICINE Health economics Ross Lawrenson.
Basic Economic Analysis David Epstein, Centre for Health Economics, York.
What makes difficult asthma difficult? SCH Journal Club Nicki Barker 2012 June 2012 Dysfunctional breathing in children1.
This study has been supported by Psychotherapy for traumatised refugees – a randomised clinical trial Jessica Carlsson, M.D., PhD Charlotte Sonne, M.D.,PhD-student.
Sifting through the evidence Sarah Fradsham. Types of Evidence Primary Literature Observational studies Case Report Case Series Case Control Study Cohort.
What we’re working on.. Defining the role of different types of evidence in decision-making. Develop tools to go from evidence to decisions about clinical.
Journal Club Alcohol, Other Drugs, and Health: Current Evidence November-December 2012.
Randomised controlled trials in primary care: case study Doctor Sue Wilson University of Birmingham United Kingdom.
Research Unit for General Practice University of Aarhus Treatment of functional somatic symptoms in general practice Marianne Rosendal,
HTA Efficient Study Designs Peter Davidson Head of HTA at NETSCC.
STICH CABG Objective To determine whether surgery plus medical management or medical management alone improve outcomes in CABG patients Study Design 1212.
Hierachy of Plans Karen Stubbs Corelli Consulting Limited.
SECCN/SPACeR Critical Care Conference Introduction to the NIHR Research Design Service Research Design Service- South East Dr Bernadette Egan University.
School of Allied Health Professions & School of Medicine, Health Policy & Practice LAMP A pragmatic unblinded randomised controlled trial and economic.
Post-It Notes to Improve Questionnaire Response Rates in RCTs Findings from a Randomised Sub-Study Ada Keding 1, Helen Lewis 2, Kate Bosanquet 2, Simon.
‘PhysioDirect’ for patients with MSK problems ECONOMIC MODEL OVERVIEW.
Providing Occupational Health Triage remotely
Introduction.
Choice – 6 Steps, 6 Actions, 6 Weeks
Evidence-based Medicine
The use of regular text messaging over one year to collect primary outcome data in an RCT Reuben Ogollah1,2, Martyn Lewis1,2, Kika Konstantinou1 Sarah.
Evaluating the effectiveness and cost effectiveness of BSL IAPT
Theory of Family-Centred Care
How to apply successfully to the NIHR HTA Board?
The Research Question Background: Question:
Presentation transcript:

Studying the Impact of Tests Studying the impact of tests Prof Jon Deeks University of Birmingham Studying the Impact of Tests Jon Deeks Professor of Health Statistics University of Birmingham Work supported by a DOH NCC RCD Senior Research Scientist in Evidence Synthesis Award Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Answering policy decisions about the use of diagnostic tests Studying the impact of tests Prof Jon Deeks University of Birmingham Answering policy decisions about the use of diagnostic tests Should GPs refer patients with low back pain for X-ray and/or MRI? Should patients with dyspeptic symptoms receive serology tests for H.pylori, endoscopy, or empirical therapy? Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Standard hierarchy for HTA of tests (Fryback and Thornton 1991) Studying the impact of tests Prof Jon Deeks University of Birmingham Standard hierarchy for HTA of tests (Fryback and Thornton 1991) Technical quality of the test Diagnostic accuracy Change in diagnostic thinking Change in patient management Change in patient outcomes Societal costs and benefits Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studies on the Diagnostic Evaluation Pathway Studying the impact of tests Prof Jon Deeks University of Birmingham Studies on the Diagnostic Evaluation Pathway Analytical validity Reliability (repeatability and reproducibility) Measurement accuracy Diagnostic validity Diagnostic accuracy Comparative/incremental diagnostic accuracy Impact Change in diagnostic yield Change in management Change in patient outcomes Economic evaluation Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

HTA policy on evaluating tests (up until 2004) Studying the impact of tests Prof Jon Deeks University of Birmingham HTA policy on evaluating tests (up until 2004) “the emphasis of the HTA programme is to assess the effect on patient management and outcomes … improvements in diagnostic accuracy, whilst relevant, are not the primary interest of this commissioned research programme” Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studies on the Diagnostic Evaluation Pathway Studying the impact of tests Prof Jon Deeks University of Birmingham Studies on the Diagnostic Evaluation Pathway Analytical validity Reliability (repeatability and reproducibility) Measurement accuracy Diagnostic validity Diagnostic accuracy Incremental diagnostic accuracy Impact Change in diagnostic yield Change in management Change in patient outcomes Economic evaluation Focus of HTA programme Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham Outline of talk Trials of diagnostic evaluations Problems What is being evaluated? Statistical power Study validity Outcomes Pragmatic suggestions When are trials really needed? Alternative trial designs Alternative of assessing comparative accuracy More research is needed Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

RCT to assess patient outcomes Studying the impact of tests Prof Jon Deeks University of Birmingham RCT to assess patient outcomes Active Outcome Population Sample Randomise Control Outcome Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham Diagnostic RCT TEST Outcome Population Sample Randomise Control Outcome Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham RCT 1: X-ray at first GP presentation for low back pain. HTA 2000(4): 20 Randomise GP attendees aged 16-64 yrs with LBP. Excluded if ‘flu or previous consultation for LBP in last 4 weeks Referred for X-ray N=73 Sample N=153 No X-ray referral N=80 Outcome 6 weeks N=59 Outcome 1 year N=50 PRIMARY Roland score HADS SF-36 EuroQol SECONDARY time off work therapists medication satisfaction Outcome 6 weeks N=67 Outcome 1 year N=58 Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham RCT 1: X-ray at first GP presentation for low back pain. HTA 2000(4): 20 Randomise GP attendees aged 16-64 yrs with LBP. Excluded if ‘flu or previous consultation for LBP in last 4 weeks Referred for X-ray N=73 Sample N=153 No X-ray referral N=80 Outcome 6 weeks N=59 Outcome 1 year N=50 RESULTS At 6 weeks  SF-36 mental health and vitality subscales (P<.05) At 12 months  SF-36 mental health subscale (P<.05) Outcome 6 weeks N=67 Outcome 1 year N=58 Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham RCT 2: X-ray for GP presentation for low back pain >6 weeks. HTA 2001(5): 30 Randomise GP attendees aged 20-55. 1st episode of LBP between 6 weeks and 6 months duration. Excluded if ‘red flags’ Referred for X-ray N=210 Sample N=421 No X-ray referral N=211 Outcome 3 months N=199 Outcome 9 months N=195 PRIMARY Roland score SECONDARY pain (VAS) EuroQol pain (diary) satisfaction pain (any) belief in X-ray time off work therapists medication consultations Outcome 3 months N=203 Outcome 9 months N=199 Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham RCT 2: X-ray for GP presentation for low back pain >6 weeks. HTA 2001(5): 30 Randomise GP attendees aged 20-55. 1st episode of LBP between 6 weeks and 6 months duration. Excluded if ‘red flags’ Referred for X-ray N=210 Sample N=421 No X-ray referral N=211 Outcome 3 months N=199 Outcome 9 months N=195 RESULTS At 3 months  proportion reporting LBP (P<.05) At 9 months None Outcome 3 months N=203 Outcome 9 months N=199 Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

What is being evaluated? Studying the impact of tests Prof Jon Deeks University of Birmingham What is being evaluated? Medical Test Information Diagnostic accuracy Diagnostic yield Management RCT combines effects Test harms and placebo effects Decision Action Patient Outcome Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

What is being evaluated? Studying the impact of tests Prof Jon Deeks University of Birmingham What is being evaluated? Conditions for a test to be of diagnostic benefit Test is more accurate Interpretation of test results is rational and consistent Management is rational and consistent Treatment is effective Conditions for a trial to be informative Rules for interpretation of test results are described Management protocol is described No descriptions given in example trials Applying the results requires faith that the behaviour of your patients and clinicians is the same as the trial Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

What is being evaluated? Studying the impact of tests Prof Jon Deeks University of Birmingham What is being evaluated? If no difference is observed … Is the test no more accurate? Are clinicians not correctly interpreting test results? Are management decisions inconsistent or inappropriate? Is the treatment ineffective? None of these questions can be answered If one element changes, the results of the trial become redundant Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham Statistical Power RCT 1: Reduction in proportion with pain at 2 weeks from 40% to 30% could be detected with 300 patients with 80% power at 5% significance RCT 2: Difference of 1.5 on Roland score could be detected with 388 patients with 90% power and 5% significance sd=4.5, standardised difference=1.5/4.5=0.33 These sample size calculations are suitable for a trial of treatment vs placebo, not a trial of test+treatment Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Diagnostic Accuracy of Clinical Judgement Studying the impact of tests Prof Jon Deeks University of Birmingham Diagnostic Accuracy of Clinical Judgement Serious (requires intervention) Minor (requires no intervention) TP FN FP TN Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Diagnostic Accuracy of Clinical Judgement + X-ray Studying the impact of tests Prof Jon Deeks University of Birmingham Diagnostic Accuracy of Clinical Judgement + X-ray Serious (requires intervention) Minor (requires no intervention) TP FN FP TN Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Comparison of Diagnostic Accuracy Studying the impact of tests Prof Jon Deeks University of Birmingham Comparison of Diagnostic Accuracy Serious (requires intervention) Minor (requires no intervention) All TP Discrepant A All FN All FP Discrepant B All TN Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Benefit can only occur in those whose diagnosis changes Studying the impact of tests Prof Jon Deeks University of Birmingham Benefit can only occur in those whose diagnosis changes Where can differences arise? Discrepant A could benefit if intervention effective Discrepant B could benefit if intervention harmful All others have no benefit as no change in their intervention Sample size must take into account Prevalence of treatable condition Detection rate (sensitivity) with control test Detection rate (sensitivity) with new test Treatment rate if control test negative (assume zero) Treatment rate if new test positive (assume 100%) Outcome for treatable condition if untreated Treatment effect Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Sample size for detecting treatment effects Studying the impact of tests Prof Jon Deeks University of Birmingham Sample size for detecting treatment effects Sample size for treatment vs control 300-400. Sample size must be adjusted according to the proportion in discrepant cells (particularly A). If 20% have serious disease and sensitivity 20% there will be 4% in Discrepant A  increase N 25-fold (N=7,500-10,000) If 10% have serious disease and sensitivity 10% there will be 1% in Discrepant A  increase N 100-fold (N=30,000-40,000) Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Sample size for detecting differences in accuracy Studying the impact of tests Prof Jon Deeks University of Birmingham Sample size for detecting differences in accuracy Sample size depends on whether the sample all receive both tests, or are randomised to tests Sample sizes for difference in sensitivity If 20% have serious disease to detect sensitivity 20% from 70% to 90% (80% power, alpha 0.05)  paired cohort design N=116 [68-136]  parallel cohort design N=232 If 10% have serious disease to detect sensitivity 10% from 80% to 90% (80% power, alpha 0.05)  paired cohort design N=706 [271-814]  parallel cohort design N=1411 Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Sample size for detecting differences in diagnoses and management Studying the impact of tests Prof Jon Deeks University of Birmingham Sample size for detecting differences in diagnoses and management Sample size based on accuracy sample size inflated according to: For diagnostic impact diagnosis rate if control test negative diagnosis rate if new test positive* For therapeutic impact treatment rate if control test negative treatment rate if new test positive* * subject to “learning effects” Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham Validity Concerns Blinding Participants and outcome assessors are rarely blind in diagnostic trials Trials may be more susceptible to measuring preconceived notions of participants and expectations of trialists Drop-out Lack of blinding can induce differential drop-out There are more stages at which drop-out occurs Compliance Lack of blinding and complexity in strategies can reduce compliance Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham What outcomes? The problem is multi-multi-factorial Assessing the effect of a single intervention for a single disease requires multiple outcomes Tests are used to differentiate between multiple diseases and disease states A trial should assess all the important outcomes for the multiple diseases within the differential diagnosis But trials usually have a focus on one condition Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham Summary of problems Diagnostic trials are … Rarely done Assess effects of “test+treatment package” Uninformative about the value of the test Often underpowered At risk of bias May not assess all relevant outcomes May be more likely to detect “placebo” effects than benefits of better diagnoses May not represent future impact on treatment and diagnostic decisions Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham Key issues Trials only need be done in limited circumstances Only patients in the discrepant cell are informative Audit and feedback studies are better for assessing and changing clinicians’ behaviour than trials More good comparative studies of test accuracy are required Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham When is measuring sensitivity and specificity sufficient to evaluate a new test? Lord et al. Ann Int Med 2006; 144: 850-5 Categories of test attributes: The new test is safer or is less costly The new test is more specific (excludes more cases of non-disease) The new test is more sensitive (detects more cases of disease) If an RCT of treatments exists, when do we still need to undertake an RCT of test+treatment? Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham Trial evidence versus linked evidence of test accuracy and treatment efficacy Lord, S. J. et. al. Ann Intern Med 2006;144:850-855 Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham Assessing new tests using evidence of test accuracy, given that treatment is effective for cases detected by the old test Lord, S. J. et. al. Ann Intern Med 2006;144:850-855 Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham When is measuring sensitivity and specificity sufficient to evaluate a new test? Lord et al. Ann Int Med 2006; 144: 850-5 If the new test has similar sensitivity Trials of test+treatment are not required Reductions in harm or cost are benefits Improved specificity can only be a benefit Decision models can be used to analyse trade-offs between positive and negative benefits Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham When is measuring sensitivity and specificity sufficient to evaluate a new test? Lord et al. Ann Int Med 2006; 144: 850-5 If the new test has improved sensitivity Value of using the test depends on treatment response in the extra cases detected A trial is still not needed if Inclusion in the treatment trial was based on the reference standard for assessing test accuracy The test is evaluated in a treatment trial as a predictor of response The new cases represent the same spectrum or subtype of disease Treatment response is known to be similar across the spectrum or subtype of disease Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Alternative Diagnostic RCT Studying the impact of tests Prof Jon Deeks University of Birmingham Alternative Diagnostic RCT Intervene Outcome Serious Minor X-ray Outcome Intervene Population Sample Serious Minor Randomise Clinical diagnosis Outcome Do not intervene Do not intervene Outcome Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Alternative Diagnostic RCT Studying the impact of tests Prof Jon Deeks University of Birmingham Alternative Diagnostic RCT Serious Minor X-ray Outcome Intervene Population Sample Serious Minor Randomise Compare Clinical diagnosis Outcome Do not intervene Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Alternative Diagnostic RCT Studying the impact of tests Prof Jon Deeks University of Birmingham Alternative Diagnostic RCT Everybody gets all tests, randomise only those with discrepant results Benefits Assess diagnostic yield and resultant patient outcomes Less follow-up required Include a reference standard for a random sample and comparative diagnostic accuracy can also be assessed Downsides More tests undertaken Problems when test material is limited Does not assess test harms or other direct effects May not be ethical to randomise treatment Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Assessing clinicians’ behaviours Studying the impact of tests Prof Jon Deeks University of Birmingham Assessing clinicians’ behaviours Informative trials require documentation and standardisation of decision-making Particularly difficult when the comparison group is standard practice Assessing behaviour observed in a trial may not be representative Future behaviour will depend on the trial results Learning curves may affect compliance Becoming acquainted with a test Ascertaining how best to use it Gaining confidence in its findings Allowing it to replace other investigations Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Diagnostic Before-and-After Studies Studying the impact of tests Prof Jon Deeks University of Birmingham Diagnostic Before-and-After Studies Design Doctors’ assessments of diagnostic, prognostic and required management decisions recorded Result of new test made available Doctors’ changes in diagnostic, prognostic and required management decisions noted (Reference standard applied) Application Assessment of an Additional Test only Assessment of Diagnostic Yield and Management Concerns New test assessed independent of other tests Doctors’ processes may not reflect standard clinical practice Learning effects Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham Conclusions We have much to learn about the best way of studying diagnostic tests Test+treatment trials are difficult to undertake, are prone to bias, and often require unattainable sample sizes. Good comparative studies of test accuracy combined in decision models with evidence from trials of treatments may in many circumstances provide the necessary evidence for policy decisions Good comparative studies of test accuracy should be commissioned more readily Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Situations when test accuracy is likely to be adequate Studying the impact of tests Prof Jon Deeks University of Birmingham Situations when test accuracy is likely to be adequate Effective treatment exists Reference standard similar to trial entry criteria All tests detect disease at the same stage All tests lead to the same treatment options Evidence of test accuracy is from similar populations to the evidence of effectiveness Studies of diagnostic accuracy need not be large Good studies that compare tests head-to-head need to be done Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham Research Questions Do the limitations of HTA evaluations of patient outcomes bias towards showing no difference? How often are trials appropriately powered? How often are the criteria for not needing RCT evidence met? What efficiencies can be made with other designs? How are economic analyses affected by these issues? Would money be better spent getting good evidence of the comparative accuracy of alternative diagnostic tests? Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Studying the impact of tests Prof Jon Deeks University of Birmingham New Diagnostic Test Assessment Framework and Examples Lord, S. J. et. al. Ann Intern Med 2006;144:850-855 Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

When are trials most needed? Studying the impact of tests Prof Jon Deeks University of Birmingham When are trials most needed? When tests detect disease earlier introducing different treatment options (screening) When interventions have harmful effects so treating some non-diseased (FP) may outweigh benefits of treating diseased (TP) When the test itself has a harmful effect When diagnostic accuracy cannot be assessed as a reference standard does not exist (although there still need to be indications for therapy – which could be used as a ref std) Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Destination worth reaching Studying the impact of tests Prof Jon Deeks University of Birmingham Destination worth reaching Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

The Role of Randomised Trials in Evaluating Diagnostic Tests Studying the impact of tests Prof Jon Deeks University of Birmingham The Role of Randomised Trials in Evaluating Diagnostic Tests Jon Deeks Professor of Health Statistics University of Birmingham Work supported by a DOH NCC RCD Senior Research Scientist in Evidence Synthesis Award Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006

Defects and Disasters in Evaluations of the Impact of Diagnostic Tests Studying the impact of tests Prof Jon Deeks University of Birmingham Defects and Disasters in Evaluations of the Impact of Diagnostic Tests Jon Deeks Professor of Health Statistics University of Birmingham Work supported by a DOH NCC RCD Senior Research Scientist in Evidence Synthesis Award Oxford CEBM Workshop in Evidence-based Diagnostics 17th-19th July 2006