AGA Practice Guidelines Committee Meeting, Chicago May 31, 2009 Yngve Falck-Ytter, M.D. Assistant Professor of Medicine Case Western Reserve University.

Slides:



Advertisements
Similar presentations
Evidence-based Dental Practice Developing guidelines or clinical recommendations Slide #1 This lecture follows the previous online lecture on evidence.
Advertisements

Introduction to the User’s Guide for Developing a Protocol for Observational Comparative Effectiveness Research Prepared for: Agency for Healthcare Research.
Grading evidence and recommendations 1 February 2005.
Doug Altman Centre for Statistics in Medicine, Oxford, UK
Holger Schünemann, MD, PhD From Evidence to EMS Practice: Building the National Model Washington, September 4,
The Science of Guidelines The 7th ACCP Conference on Antithrombotic and Thrombolytic Therapy: Evidence-Based Guidelines Holger Schünemann, MD, PhD Italian.
Critically Evaluating the Evidence: Tools for Appraisal Elizabeth A. Crabtree, MPH, PhD (c) Director of Evidence-Based Practice, Quality Management Assistant.
Summarising findings about the likely impacts of options Judgements about the quality of evidence Preparing summary of findings tables Plain language summaries.
Grading of Recommendations Assessment, Development and Evaluation (GRADE) Methodology.
Michael Rawlins Chairman, National Institute for Health and Clinical Excellence, London Emeritus Professor, University of Newcastle upon Tyne Honorary.
Chapter 7. Getting Closer: Grading the Literature and Evaluating the Strength of the Evidence.
By Dr. Ahmed Mostafa Assist. Prof. of anesthesia & I.C.U. Evidence-based medicine.
Introduction to evidence based medicine
Critical Appraisal of Clinical Practice Guidelines
Illustrating the GRADE Methodology: The Cather Associated-UTI Case Study TEACH Level II Workshop 5 NYAM August 9 th, 2013 Craig A Umscheid, MD, MSCE, FACP.
Grading of Recommendations Assessment, Development, and Evaluation (GRADE) Working Group
AHRQ Annual Meeting 2009: "Research to Reform: Achieving Health System Change" September 14, 2009 Yngve Falck-Ytter, M.D. Case Western Reserve University,
Holger Sch ü nemann Yngve Falck-Ytter NICE, London December 11, 2006 GRADE – An introduction and workshop.
Department of O UTCOMES R ESEARCH. Daniel I. Sessler, M.D. Michael Cudahy Professor and Chair Department of O UTCOMES R ESEARCH The Cleveland Clinic Clinical.
Holger Schünemann, MD, PhD Professor Utrecht, NL September ,
AASLD Practice Guidelines Committee Meeting, Chicago 1 May 2009 Yngve Falck-Ytter, M.D. Case Western Reserve University.
Grading evidence and recommendations The GRADE initiative Holger Schünemann, MD, PhD Associate Professor Italian National Cancer Institute Regina Elena,
AHRQ Annual Meeting 2009: "Research to Reform: Achieving Health System Change" September 14, 2009 Yngve Falck-Ytter, M.D. Case Western Reserve University,
Grading evidence and recommendations The GRADE approach Holger Schünemann, MD, PhD for the GRADE Working Group.
The New York Academy of Medicine Teaching Evidence Assimilation for Collaborative Healthcare New York, August 8, 2013 Yngve Falck-Ytter, MD, AGAF for the.
AHRQ Annual Meeting 2010: “Better Care, Better Health: Delivering on Quality for All Americans" September 28, 2010 Yngve Falck-Ytter, M.D. Associate Professor.
Holger Schünemann, MD, PhD Chair, Department of Clinical Epidemiology & Biostatistics Professor of Clinical Epidemiology, Biostatistics and Medicine McMaster.
Canadian Task Force on Preventive Health Care:
Evidence Based Medicine
Grading evidence and recommendations The GRADE approach
Brief summary of the GRADE framework Holger Schünemann, MD, PhD Chair and Professor, Department of Clinical Epidemiology & Biostatistics Professor of Medicine.
Cost-Effectiveness Thresholds Professor of Health Economics
Research Techniques Made Simple: Evaluating the Strength of Clinical Recommendations in the Medical Literature: GRADE, SORT, and AGREE Mayra Buainain de.
Harmonizing levels of evidence: The Grading of Recommendations Assessment, Development and Evaluation (GRADE) working group Holger Schünemann, Andy Oxman,
Grading evidence and recommendations The GRADE initiative Holger Schünemann, MD, PhD Associate Professor Italian National Cancer Institute Regina Elena,
Grading Strength of Evidence Prepared for: The Agency for Healthcare Research and Quality (AHRQ) Training Modules for Systematic Reviews Methods Guide.
Evidence-Based Public Health Nancy Allee, MLS, MPH University of Michigan November 6, 2004.
Grading evidence and recommendations The GRADE approach Holger Schünemann, MD, PhD for the GRADE Working Group.
Grading evidence and recommendations Holger Schünemann Gunn Vist Gordon Guyatt for the GRADE Working Group.
The New York Academy of Medicine Teaching Evidence Assimilation for Collaborative Healthcare New York, August 10, 2011 Yngve Falck-Ytter, MD, AGAF for.
AGA Clinical Practice & Quality Management Committee Teleconference 17 Oct 2008 Yngve Falck-Ytter, M.D. Assistant Professor of Medicine Case Western Reserve.
Clinical Writing for Interventional Cardiologists.
Evidence-Based Medicine: What does it really mean? Sports Medicine Rounds November 7, 2007.
Grading the quality of evidence
Grading evidence and recommendations Holger Schünemann Andy Oxman for the GRADE Working Group.
Stakeholder Summit on Using Quality Systematic Reviews to Inform Evidence-based Guidelines US Cochrane Center June 4 and 5, 2009 Yngve Falck-Ytter, M.D.
Introduction to Healthcare and Public Health in the US The Evolution and Reform of Healthcare in the US Lecture b This material (Comp1_Unit9b) was developed.
WHO GUIDANCE FOR THE DEVELOPMENT OF EVIDENCE-BASED VACCINE RELATED RECOMMENDATIONS August 2011.
EBM --- Journal Reading Presenter :呂宥達 Date : 2005/10/27.
Anne Matthews, Health & Society, School of Nursing and Human Sciences, DCU The paradox of ‘low quality evidence; strong recommendation’: An analysis of.
Developing evidence-based guidelines at WHO. Evidence-based guidelines at WHO | January 17, |2 |
EVALUATING u After retrieving the literature, you have to evaluate or critically appraise the evidence for its validity and applicability to your patient.
Grading evidence and recommendations Workshop W-069 Congress Hall ABEF Oct
BEST PRACTICE PORTAL BEST PRACTICE PORTAL project presentation to the Scientific Committee Ferri et al Lisbon, 16th July 2010.
Holger Schünemann Yngve Falck-Ytter AHRQ Annual Conference Washington, September 8,
The New York Academy of Medicine Teaching Evidence Assimilation for Collaborative Healthcare New York, August 8, 2012 Yngve Falck-Ytter, MD, AGAF for the.
GDG Meeting Wednesday November 9, :30 – 11:30 am.
GRADE Grading of Recommendations Assessment, Development and Evaluation British Association of Dermatologists April 2014.
Clinical Practice Guidelines: Can we fix Babel? Eddy Lang Department Chair, Emergency Alberta Health Services Associate Professor University of Calgary.
Evidence-Based Mental Health PSYC 377. Structure of the Presentation 1. Describe EBP issues 2. Categorize EBP issues 3. Assess the quality of ‘evidence’
Ghada Aboheimed, Msc. Review the principles of an evidence based approach to clinical practice. Appreciate the value of EBM Describe the 5 steps of evidence.
From evidence to Policy: Paediatric guideline development in Kenya Mercy Mulaku.
Approach to guideline development
Why this talk? you will be seeing a lot of GRADE
Conflicts of interest Major role in development of GRADE
Overview of the GRADE approach – selected slides
Randomized Trials: A Brief Overview
STROBE Statement revision
WHO Guideline development
GRADE – An introduction and workshop
Presentation transcript:

AGA Practice Guidelines Committee Meeting, Chicago May 31, 2009 Yngve Falck-Ytter, M.D. Assistant Professor of Medicine Case Western Reserve University Director of Hepatology, VA Medical Center

Disclosure In the past 5 years, Dr. Falck-Ytter received no personal payments for services from industry. His research group received research grants from Three Rivers, Valeant and Roche that were deposited into non-profit research accounts. He is a member of the GRADE working group which has received funding from various governmental entities in the US and Europe. Some of the GRADE work he has done is supported in part by grant # 1 R13 HS from the Agency for Healthcare Research and Quality (AHRQ).

Content Part 1  Why revisiting guideline methodology?  GRADE approach  Quality of evidence  Strength of recommendations  Why societies have adopted GRADE

Content (continued) Part 2 – practical consideration  Ideal vs. practical ad hoc approaches  Funding guideline work

Reassessment of clinical practice guidelines  Editorial by Shaneyfelt and Centor (JAMA 2009)  “Too many current guidelines have become marketing and opinion-based pieces…”  “AHA CPG: 48% of recommendations are based on level C = expert opinion…”  “…clinicians do not use CPG […] greater concern […] some CPG are turned into performance measures…”  “Time has come for CPG development to again be centralized, e.g., AHQR…”

Evidence-based clinical decisions Research evidence Patient values and preferences Clinical state and circumstances Expertise Equal for all Haynes et al. 2002

Confidence in evidence  There always is evidence  “When there is a question there is evidence”  Evidence alone is never sufficient to make a clinical decision  Better research  greater confidence in the evidence and decisions

Hierarchy of evidence STUDY DESIGN Randomized Controlled Trials Cohort Studies and Case Control Studies Case Reports and Case Series, Non-systematic observations BIAS Expert Opinion

Reasons for grading evidence?  People draw conclusions about the  quality of evidence and strength of recommendations  Systematic and explicit approaches can help to  protect against errors, resolve disagreements  communicate information and fulfill needs  be transparent about the process  Change practitioner behavior  However, wide variation in approaches GRADE working group. BMJ & 2008

10

Which grading system? P: In patients with acute hepatitis C … I : Should anti-viral treatment be used … C: Compared to no treatment … O: To achieve viral clearance? EvidenceRecommendationOrganization BClass IAASLD (2009) VA (2006)II-1-/-SIGN (2006)1+AAGA (2006)-/-“Most authorities…”

Scenario (2) Should patients with risk factors for viral hepatitis be screened with a hepatitis C antibody (ELISA) test to identify patients with past hepatitis C exposure?

13 Level of evidence in GI CPGs AASLD AGA ACGASGE AMultiple RCTs or meta-analysis Good Consistent, well-designed, well conducted studies […] 1. Multiple published, well-controlled (?) randomized trials or a well designed systemic (?) meta- analysis A. RCTs BSingle randomized trial, or non- randomized studies C Only consensus opinion of experts, case studies, or standard-of-care FairLimited by the number, quality or consistency of individual studies […] Poor… important flaws, gaps in chain of evidence… 2. One quality- published (?) RCT, published well- designed cohort/ case-control studies 3. Consensus of authoritative (?) expert opinions based on clinical evidence or from well designed, but uncontrolled or non-rand. clin. trials B. RCT with important limitations C. Obser- vational studies D. Expert opinion

What to do? 14

Limitations of existing systems  Confuse quality of evidence with strength of recommendations  Lack well-articulated conceptual framework  Criteria not comprehensive or transparent  GRADE unique  breadth, intensity of development process  wide endorsement and use  conceptual framework  comprehensive, transparent criteria  Focus on all important outcomes related to a specific question and overall quality

 G rades of R ecommendation A ssessment, D evelopment and E valuation

GRADE Working Group  David Atkins, chief medical officer a  Dana Best, assistant professor b  Martin Eccles, professor d  Francoise Cluzeau, lecturer x  Yngve Falck-Ytter, associate director e  Signe Flottorp, researcher f  Gordon H Guyatt, professor g  Robin T Harbour, quality and information director h  Margaret C Haugh, methodologist i  David Henry, professor j  Suzanne Hill, senior lecturer j  Roman Jaeschke, clinical professor k  Regina Kunx, Associate Professor  Gillian Leng, guidelines programme director l  Alessandro Liberati, professor m  Nicola Magrini, director n  James Mason, professor d  Philippa Middleton, honorary research fellow o  Jacek Mrukowicz, executive director p  Dianne O ’ Connell, senior epidemiologist q  Andrew D Oxman, director f  Bob Phillips, associate fellow r  Holger J Sch ü nemann, professor g,s  Tessa Tan-Torres Edejer, medical officer t  David Tovey, Editor y  Jane Thomas, Lecturer, UK  Helena Varonen, associate editor u  Gunn E Vist, researcher f  John W Williams Jr, professor v  Stephanie Zaza, project director w  a) Agency for Healthcare Research and Quality, USA  b) Children's National Medical Center, USA  c) Centers for Disease Control and Prevention, USA  d) University of Newcastle upon Tyne, UK  e) German Cochrane Centre, Germany  f) Norwegian Centre for Health Services, Norway  g) McMaster University, Canada  h) Scottish Intercollegiate Guidelines Network, UK  i) F é d é ration Nationale des Centres de Lutte Contre le Cancer, France  j) University of Newcastle, Australia  k) McMaster University, Canada  l) National Institute for Clinical Excellence, UK  m) Universit à di Modena e Reggio Emilia, Italy  n) Centro per la Valutazione della Efficacia della Assistenza Sanitaria, Italy  o) Australasian Cochrane Centre, Australia  p) Polish Institute for Evidence Based Medicine, Poland  q) The Cancer Council, Australia  r) Centre for Evidence-based Medicine, UK  s) National Cancer Institute, Italy  t) World Health Organisation, Switzerland  u) Finnish Medical Society Duodecim, Finland  v) Duke University Medical Center, USA  w) Centers for Disease Control and Prevention, USA  x) University of London, UK  Y) BMJ Clinical Evidence, UK

GRADE uptake

Where GRADE fits in Prioritize problems, establish panel Systematic review Searches, selection of studies, data collection and analysis Assess the relative importance of outcomes Prepare evidence profile: Quality of evidence for each outcome and summary of findings Assess overall quality of evidence Decide direction and strength of recommendation Draft guideline Consult with stakeholders and / or external peer reviewer Disseminate guideline Implement the guideline and evaluate GRADE

20 GRADE: Quality of evidence The extent to which our confidence in an estimate of the treatment effect is adequate to support particular recommendation. Although the degree of confidence is a continuum, we suggest using four categories:  High  Moderate  Low  Very low

I B IIVIII Quality of evidence across studies Outcome #1 Outcome #2 Outcome #3 Quality: High Quality: Moderate Quality: Low

Determinants of quality  RCTs start high  Observational studies start low  What lowers quality of evidence? 5 factors:  Detailed design and execution  Inconsistency of results  Indirectness of evidence  Imprecision  Publication bias

1. Design and execution  Study limitations (risk of bias) For RCTs:  Lack of allocation concealment  No true intention to treat principle  Inadequate blinding  Loss to follow-up  Early stopping for benefit For observational studies:  Selection  Comparability  Exposure/outcome

Schulz KF et al. JAMA Allocation concealment 250 RCTs out of 33 meta-analyses Allocation concealment:Effect (Ratio of OR) adequate1.00(Ref.) unclear0.67 [0.60 – 0.75] not adequate0.59 [0.48 – 0.73] * * significant

Cochrane Risk of bias graph in RevMan 5 25

2. Consistency of results  Look for explanation for inconsistency  patients, intervention, comparator, outcome, methods  Judgment  variation in size of effect  overlap in confidence intervals  statistical significance of heterogeneity I2I2

Pagliaro L et al. Ann Intern Med 1992;117: Heterogeneity

3. Directness of Evidence  Indirect comparisons  Interested in head-to-head comparison  Drug A versus drug B  Tenofovir versus entecavir in hepatitis B treatment  Differences in  patients (early cirrhosis vs end-stage cirrhosis)  interventions (CRC screening: flex. sig. vs colonoscopy)  comparator (e.g., differences in dose)  outcomes (non-steroidal safety: ulcer on endoscopy vs symptomatic ulcer complications)

4. Imprecision Small sample size  small number of events  wide confidence intervals  uncertainty about magnitude of effect

Imprecision RR appreciable benefit appreciable harm impreciseprecise

5. Reporting Bias (Publication Bias)  Reporting of studies  publication bias  number of small studies  Reporting of outcomes

32 Quality assessment criteria Lower if… Quality of evidence High (4) Moderate (3) Low (2) Very low (1) Study limitations (design and execution) Inconsistency Indirectness Imprecision Publication bias Observational study Study design Randomized trial Higher if… What can raise the quality of evidence?

BMJ 2003;327:1459–61 33

34

35 Quality assessment criteria Lower if…Higher if… Quality of evidence High (4) Moderate (3) Low (2) Very low (1) Study design Randomized trial Observational study Study limitations Inconsistency Indirectness Imprecision Publication bias Large effect (e.g., RR 0.5) Very large effect (e.g., RR 0.2) Evidence of dose-response gradient All plausible confounding would reduce a demonstrated effect

36 Categories of quality Further research is very unlikely to change our confidence in the estimate of effect High Low Further research is very likely to have an important impact on our confidence in the estimate of effect and is likely to change the estimate Moderate Further research is likely to have an important impact on our confidence in the estimate of effect and may change the estimate Very lowAny estimate of effect is very uncertain

37 Judgments about the overall quality of evidence  Most systems not explicit  Options:  Benefits  Primary outcome  Highest  Lowest  Beyond the scope of a systematic review  GRADE: Based on lowest of all the critical outcomes

GRADE evidence profile

Going from evidence to recommendations  Deliberate separation of quality of evidence from strength of recommendation  No automatic one-to-one connection as in other grading systems  Example: What if there is high quality evidence, but the balance between benefit and risks are finely balanced? 39

Strength of recommendation “The strength of a recommendation reflects the extent to which we can, across the range of patients for whom the recommendations are intended, be confident that desirable effects of a management strategy outweigh undesirable effects.” Although the strength of recommendation is a continuum, we suggest using two categories : “Strong” and “Weak”

Desirable and undesirable effects  Desirable effects  Mortality reduction  Improvement in quality of life, fewer hospitalizations/infections  Reduction in the burden of treatment  Reduced resource expenditure  Undesirable effects  Deleterious impact on morbidity, mortality or quality of life, increased resource expenditure

4 determinants of the strength of recommendation Factors that can weaken the strength of a recommendation Explanation  Lower quality evidenceThe higher the quality of evidence, the more likely is a strong recommendation.  Uncertainty about the balance of benefits versus harms and burdens The larger the difference between the desirable and undesirable consequences, the more likely a strong recommendation warranted. The smaller the net benefit and the lower certainty for that benefit, the more likely is a weak recommendation warranted.  Uncertainty or differences in values The greater the variability in values and preferences, or uncertainty in values and preferences, the more likely weak recommendation warranted.  Uncertainty about whether the net benefits are worth the costs The higher the costs of an intervention – that is, the more resources consumed – the less likely is a strong recommendation warranted.

Developing recommendations

Implications of a strong recommendation  Patients: Most people in this situation would want the recommended course of action and only a small proportion would not  Clinicians: Most patients should receive the recommended course of action  Policy makers: The recommendation can be adapted as a policy in most situations

Implications of a weak recommendation  Patients: The majority of people in this situation would want the recommended course of action, but many would not  Clinicians: Be prepared to help patients to make a decision that is consistent with their own values/decision aids and shared decision making  Policy makers: There is a need for substantial debate and involvement of stakeholders

6 main misconceptions 1. Isn’t GRADE expensive to realize? 2. Isn’t GRADE more complicated, takes longer and requires more resources? 3. Isn’t GRADE eliminating the expert ? 4. But what about mechanism of disease, diagnosis, cost? 5. But GRADE does not have an “insufficient evidence to make recommendation” category! (or: the “optional” category), no? 6. But we only “recommend” – we can’t possibly give weak recommendations!

Systematic review Guideline development PICOPICO Outcome Formulate question Rate importance Critical Important Critical Not important Create evidence profile with GRADEpro Summary of findings & estimate of effect for each outcome Rate overall quality of evidence across outcomes based on lowest quality of critical outcomes RCT start high, obs. data start low 1.Risk of bias 2.Inconsistency 3.Indirectness 4.Imprecision 5.Publication bias Grade down Grade up 1.Large effect 2.Dose response 3.Confounders Rate quality of evidence for each outcome Select outcomes Very low Low Moderate High Formulate recommendations: For or against (direction) Strong or weak (strength) By considering:  Quality of evidence  Balance benefits/harms  Values and preferences Revise if necessary by considering:  Resource use (cost) “We recommend using…” “We suggest using…” “We recommend against using…” “We suggest against using…” Outcomes across studies

Summary, and Why institutions adopt GRADE 1. GRADE is gaining acceptance as international standard 2. GRADE has criteria for evidence assessment across a range of questions and outcomes 3. Criteria for moving from evidence to recommendations 4. Simple, transparent, systematic 5. Balance between simplicity and methodological rigor

Ideal vs. practical ad hoc GRADE approaches StageElementsAdvantageComment IdealSystematic review GRADE eTables Qual. of evidence Strength of rec. Follows int. standards Methodolog. most rigorous Easily maintainable Fully transparent process Access to methodologist Access to evidence centers Initially more resource intensive, long-term savings Inter- mediary Ad hoc review GRADE eTables Qual. of evidence Strength of rec. Still retaining major advantages of the of the “ideal approach” Risk of bias higher Access methodologist rec. Only minimal addl. cost Initiation Ad hoc review GRADE eTables Qual. of evidence Strength of rec. Option to fully “upgrade” to an “ideal approach” Foundation of a methodo- logically sound system Risk of bias higher Access methodologist prn No additional cost