Instrument Development and Psychometric Evaluation: Scientific Standards May 2012 Dynamic Tools to Measure Health Outcomes from the Patient Perspective.

Slides:



Advertisements
Similar presentations
Performance Assessment
Advertisements

Critical Reading Strategies: Overview of Research Process
Study Objectives and Questions for Observational Comparative Effectiveness Research Prepared for: Agency for Healthcare Research and Quality (AHRQ)
Cross Cultural Research
Conceptualization and Measurement
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
PROMIS: The Right Place at the Right Time? David Cella, Ph.D. Department of Medical Social Sciences Northwestern University Chair, PROMIS Steering Committee.
Issues of Technical Adequacy in Measuring Student Growth for Educator Effectiveness Stanley Rabinowitz, Ph.D. Director, Assessment & Standards Development.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Reviewing and Critiquing Research
Chapter 4 Validity.
Transcultural Research Prof. N.J.Mathers Institute of primary and community care University of Sheffield.UK 06/Nov/2007.
Chapter Three Research Design.
Proposal Writing.
Ch 6 Validity of Instrument
Dr Keith Meadows, DHP Research & Consultancy Ltd.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 12 Undertaking Research for Specific Purposes.
Standardization and Test Development Nisrin Alqatarneh MSc. Occupational therapy.
Program Evaluation. Program evaluation Methodological techniques of the social sciences social policy public welfare administration.
Evaluating a Research Report
CCSSO Criteria for High-Quality Assessments Technical Issues and Practical Application of Assessment Quality Criteria.
Validity Is the Test Appropriate, Useful, and Meaningful?
ScWk 242 Course Overview and Review of ScWk 240 Concepts ScWk 242 Session 1 Slides.
SURVEY RESEARCH.  Purposes and general principles Survey research as a general approach for collecting descriptive data Surveys as data collection methods.
VALIDITY AND VALIDATION: AN INTRODUCTION Note: I have included explanatory notes for each slide. To access these, you will probably have to save the file.
MEASUREMENT: SCALE DEVELOPMENT Lu Ann Aday, Ph.D. The University of Texas School of Public Health.
Research Design – Where to Begin…. Purpose of EDD 9300 Provide guidance and help you to: Select a topic Conduct a Preliminary Literature Review Design.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 15 Developing and Testing Self-Report Scales.
3-1 Copyright © 2010 Pearson Education, Inc. Chapter Three Research Design.
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
Chapter Eight: Quantitative Methods
PFF Teal = MAIN COLORS PFF Green = Light Green = Red = HIGHLIGHT COLORS Light Grey = Dark Grey =
CoRPS London 26 & 27 October 2010 Center of Research on Psychology in Somatic diseases Understanding PRO in hematological disorders: Do we have a consensus?
ESTABLISHING RELIABILITY AND VALIDITY OF RESEARCH TOOLS Prof. HCL Rawat Principal UCON,BFUHS Faridkot.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 25 Critiquing Assessments Sherrilene Classen, Craig A. Velozo.
HCS 465 OUTLET Experience Tradition /hcs465outlet.com FOR MORE CLASSES VISIT
Copyright © 2009 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 47 Critiquing Assessments.
HCS 465 Week 3 Individual Applying the Results and Conclusion of the Research Process to Problems in Health Care To purchase this material click below.
Introduction to Marketing Research
EVALUATING EPP-CREATED ASSESSMENTS
Chapter Three MaxIT WiMax.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
Assist. Prof. Merve Topcu Department of Psychology, Çankaya University
Writing a sound proposal
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
Introduction to the Validation Phase
Constructing hypotheses & research design
3 Research Design Formulation
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
WHO The World Health Survey General Introduction
QUESTIONNAIRE DESIGN AND VALIDATION
Test Blueprints for Adaptive Assessments
Reliability and Validity in Research
Test Design & Construction
Test Validity.
METHODOLOGY AND MEASUREMENT ASPECTS
Performance Improvement Project Validation Process Outcome Focused Scoring Methodology and Critical Analysis Presenter: Christi Melendez, RN, CPHQ Associate.
Week 3 Class Discussion.
پرسشنامه کارگاه.
Chapter Three Research Design.
Chapter Eight: Quantitative Methods
Lesson 1 Foundations of measurement in Psychology
Spanish and English Neuropsychological Assessment Scales - Guiding Principles and Evolution Friday Harbor Psychometrics Workshop 2005.
Measurement Concepts and scale evaluation
Presenter: Kate Bell, MA PIP Reviewer
TESTING AND EVALUATION IN EDUCATION GA 3113 lecture 1
Analyzing Reliability and Validity in Outcomes Assessment
AACC Mini Conference June 8-9, 2011
Managerial Decision Making and Evaluating Research
Chapter 3: How Standardized Test….
Presentation transcript:

Instrument Development and Psychometric Evaluation: Scientific Standards May 2012 Dynamic Tools to Measure Health Outcomes from the Patient Perspective

Click to edit Master title style PROMIS ® Standards document Patient-Reported Outcome Measurement Information System (PROMIS ® ) provides clinicians and researchers access to efficient, valid and responsive self-reported measures in health, including symptoms, function and well- being. Instrument Development and Psychometric Evaluation: Scientific Standards describes a set of standards that serve as the scientific foundation and guidance for the development of PROMIS evaluation of PROMIS item banks and instruments.

Click to edit Master title style PROMIS ® Standards document Practices are based on measurement science, experience of PROMIS ® investigators, published literature on methodology of PROMIS ® and other PRO measurement development Standards are operationalized by series of guidelines that provide detailed guidance for item bank development and psychometric evaluation

Click to edit Master title style PROMIS ® Scientific Standards Address: 1)Defining target concept and conceptual model 2)Generating and design of individual items 3)Constructing item pool 4)Determining item bank properties 5)Field testing and instrument formats 6)Validity 7)Reliability 8)Interpretability 9)Language translation and cultural adaptation

Click to edit Master title style 1.Defining the Target Concept and Conceptual Model  Conceptual model incorporating target concept(s) should be defined and based on extant literature with input from patients, content and measurement experts, clinicians, end users and stakeholders Placement of the instrument within PROMIS ® framework should be defined

Click to edit Master title style 2. Composing Individual Items I ndividual items should be refined through cognitive interviewing to ensure: –The meaning is understood as intended –The item is clear and contains one concept Items also should be reviewed for: –Translatability –Literacy –Readability

Click to edit Master title style 2. Composing Individual Items, cont’d. Level of life course and cultural harmonization should be addressed Existing PROMIS ® item formats should be considered and utilized as appropriate The language used should be clear to patients and consistent with results of formative research

Click to edit Master title style 3. Constructing Item Pool Item pool should cover a full breath of target construct Consideration should be paid to overlap of facets already covered by extant PROMIS ® domains Consideration should be paid to intellectual property issues

Click to edit Master title style 4. Determining Item Bank Properties Psychometric characteristics of items within an item bank should be determined based on a representative sample of respondents drawn from a relevant and interpretable population Item banks should have good measurement characteristics: –Well-characterized and modeled dimensionality –High degree of information and low standard error –Model fit, item and scale properties Differential item functioning (DIF) should be assessed for key groups and its impact on measurement properties identified

Click to edit Master title style 5. Testing and Instrument Formats Instrument formats should be appropriately defined based on intended use and item bank properties including: –CATS –Fixed length short-forms –Profiles –Screeners Adequate scale properties and performance should be demonstrated and include assessment of respondent burden Instruments that use different modes and methods of administration should demonstrate: –comparability of scale properties and performance –assessment of respondent burden for each mode

Click to edit Master title style 6. Validity Validity for construct, content and criterion should be addressed relative to a priori hypothesized relationships with related measures – Description of methods and sample used to evaluate validity, including hypotheses tested and rationale for choice of “gold standard” and measures should be provided Final instrument should be re-reviewed by experts and end-users to assess consistency with or identify differences between original definitions and final product

Click to edit Master title style 6. Validity, cont’d. If an instrument is purported to be responsive, relevant anchor-based methods in representative populations should be provided Longitudinal data collected will compare a group expected to change with a group expected to remain stable Rationale should be provided for the external anchors used to document change Rationale should be provided for the time intervals used for assessment

Click to edit Master title style 7. Reliability Reliability of an instrument should be described, including methods used to collect data and estimate reliability Internal consistency reliability estimates may consist of: –Information and standard errors at different locations of the scale (item response theory) –Reliability estimates and standard errors for all score elements (classical test theory)

Click to edit Master title style 7. Reliability, cont’d. Reproducibility of the measure should be described, providing –Rationale to support the design of the study –The interval between initial and subsequent administration to support the assumption that the population is stable

Click to edit Master title style 8. Interpretability The degree to which one can assign easily understood meaning to the instrument’s quantitative scores should be described Rationale should be provided for the external anchors used to facilitate interpretability of scores Information should be provided on the ways in which data from an instrument should be reported or displayed

Click to edit Master title style 8. Interpretability, cont’d. Availability of comparative data from the general population and/or group-specific scores should be described Guidance should be provided on the meaningfulness of scores and changes in scores for use by researchers and clinicians

Click to edit Master title style 9. Translation and Cultural Adaptation Translation of items and instruments should include both forward and backward translations of all items, response choices, and instructions Translation of items, response choices, and instructions should be obtained through –Iterative process of forward and back translation –Bilingual expert review –Pre-testing with cognitive debriefing

Click to edit Master title style 9. Translation and Cultural Adaptation, cont’d Harmonization across all languages and a universal approach to translation should guide the process

Click to edit Master title style Appendix Each standard refers to guideline documents in the appendix for further description of processes for performing recommended practices Appendix also includes the PROMIS ® Instrument Maturity Model that describes the stages of instrument development from conceptualization through evidence of psychometric properties in multiple diverse populations.

Click to edit Master title style PROMIS Instrument Maturity Model Stages A1B2A2B3A3B4A4B5 Item Pool Preliminary Item Bank Calibrated Item Bank Item Bank, Profile or Global Health Measure - Prelim Validated Instruments Validated Instruments - Prelim responsive- ness Maturing Instruments Item Bank Expansion Fully Mature Instruments: Score Interpretations ConceptualizedReady for Calibration Dimension-ality Assessed & Calibrated Validity (Construct & Concurrent - limited) Cross- Sectional, Population Specific Prelim responsive- ness Extensive validity & responsiveness in general and pertinent population samples Item bank modifications - population specific or expansion/ refinement How scores can be used to understand and respond to health care needs and differences in health is determined and documented PROMIS Instrument Maturity Model Stages Stage Descriptions CriteriaApplies To QUALITATIVE: Conceptual documentation and evidence supporting content validityAll stages Item response theory (IRT): Dimensionality Specified; Item calibration; information and DIF analysesAll but stage 1A Classical test theory (CTT): Evidence supporting dimensionality, reliability and validity (e.g. concurrent validity with legacy) All but stage 1A POPULATION: Sample variability reflects variability in constructStages 2, 3, and 4 FORMAT: CAT and short form measures; Computer, paper forms Stages 3A, 3B, and 4A Continued Documentation of Relevance of Item Content and Generalizability as neededStages 1 and 2 Continued Documentation of Relevance of Item Content and GeneralizabilityStages 4B and 5 Internal Psychometric Criteria CriteriaApplies To IRT: DIF analyses by different disease conditions and relevant population characteristics (e.g. age, sex, etc.)Stages 3 and 4 CTT: Evidence supporting responsiveness and interpretation guidelines (MID, responder criteria)Stages 3 and 4 POPULATION: General population and multiple disease conditions; Relevant language translationsStages 3 and 4 FORMAT: CAT, short form, and study tailored formsStages 3 and 4 MODE: Evidence supporting multiple modes of administration (CAT, paper, IVRS, computer)Stages 3 and 4 External Psychometric Criteria