Reliability or Validity Reliability gets more attention: n n Easier to understand n n Easier to measure n n More formulas (like stats!) n n Base for validity.

Slides:



Advertisements
Similar presentations
Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
Advertisements

 Degree to which inferences made using data are justified or supported by evidence  Some types of validity ◦ Criterion-related ◦ Content ◦ Construct.
1 COMM 301: Empirical Research in Communication Kwan M Lee Lect4_1.
© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.
VALIDITY AND RELIABILITY
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 6 Validity.
Chapter 4A Validity and Test Development. Basic Concepts of Validity Validity must be built into the test from the outset rather than being limited to.
Assessment: Reliability, Validity, and Absence of bias
Reliability, Validity, Trustworthiness If a research says it must be right, then it must be right,… right??
Chapter 4 Validity.
VALIDITY.
Validity Does test measure what it says it does? Is the test useful? Can a test be reliable, but not valid? Can a test be valid, but not reliable?
Validity of Selection. Objectives Define Validity Relation between Reliability and Validity Types of Validity Strategies.
Research Methods in MIS
Chapter 7 Evaluating What a Test Really Measures
Classroom Assessment A Practical Guide for Educators by Craig A
Validity Lecture Overview Overview of the concept Different types of validity Threats to validity and strategies for handling them Examples of validity.
Understanding Validity for Teachers
Questions to check whether or not the test is well designed: 1. How do you know if a test is effective? 2. Can it be given within appropriate administrative.
Validity and Reliability Neither Valid nor Reliable Reliable but not Valid Valid & Reliable Fairly Valid but not very Reliable Think in terms of ‘the purpose.
Presented to: Dr. Ava Clare Marie O. Robles Class Schedule: TFr /1:00-2:30 pm Presented by: Ierine Joy L. Caserial Republic of the Philippines MINDANAO.
VALIDITY. Validity is an important characteristic of a scientific instrument. The term validity denotes the scientific utility of a measuring instrument,
Measurement Concepts & Interpretation. Scores on tests can be interpreted: By comparing a client to a peer in the norm group to determine how different.
Measurement and Data Quality
Validity and Reliability
PhD Research Seminar Series: Reliability and Validity in Tests and Measures Dr. K. A. Korb University of Jos.
Reliability and Validity what is measured and how well.
Instrumentation.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
Technical Adequacy Session One Part Three.
Validity. Face Validity  The extent to which items on a test appear to be meaningful and relevant to the construct being measured.
Validity Is the Test Appropriate, Useful, and Meaningful?
EDU 8603 Day 6. What do the following numbers mean?
Week 5 Lecture 4. Lecture’s objectives  Understand the principles of language assessment.  Use language assessment principles to evaluate existing tests.
6. Evaluation of measuring tools: validity Psychometrics. 2012/13. Group A (English)
Chapter 8 Validity and Reliability. Validity How well can you defend the measure? –Face V –Content V –Criterion-related V –Construct V.
Validity and Reliability Neither Valid nor Reliable Reliable but not Valid Valid & Reliable Fairly Valid but not very Reliable Think in terms of ‘the purpose.
Concurrent Validity Pages By: Davida R. Molina October 23, 2006.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Presented By Dr / Said Said Elshama  Distinguish between validity and reliability.  Describe different evidences of validity.  Describe methods of.
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
Validity and Item Analysis Chapter 4. Validity Concerns what the instrument measures and how well it does that task Not something an instrument has or.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
Measurement and Scaling
Week 4 Slides. Conscientiousness was most highly voted for construct We will also give other measures – protestant work ethic and turnover intentions.
Nurhayati, M.Pd Indraprasta University Jakarta.  Validity : Does it measure what it is supposed to measure?  Reliability: How the representative is.
RESEARCH METHODS IN INDUSTRIAL PSYCHOLOGY & ORGANIZATION Pertemuan Matakuliah: D Sosiologi dan Psikologi Industri Tahun: Sep-2009.
Chapter 6 - Standardized Measurement and Assessment
Classroom Assessment Chapters 4 and 5 ELED 4050 Summer 2007.
Michigan Assessment Consortium Common Assessment Development Series Module 16 – Validity.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
RELIABILITY AND VALIDITY Dr. Rehab F. Gwada. Control of Measurement Reliabilityvalidity.
Copyright © Springer Publishing Company, LLC. All Rights Reserved. DEVELOPING AND USING TESTS – Chapter 11 –
Copyright © 2009 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 47 Critiquing Assessments.
Principles of Language Assessment
Ch. 5 Measurement Concepts.
VALIDITY by Barli Tambunan/
Lecture 5 Validity and Reliability
QUESTIONNAIRE DESIGN AND VALIDATION
Reliability and Validity in Research
Concept of Test Validity
Validity and Reliability
Validity and Reliability
Week 3 Class Discussion.
Week 10 Slides.
پرسشنامه کارگاه.
PSY 614 Instructor: Emily Bullock Yowell, Ph.D.
Reliability and Validity of Measurement
Chapter 8 VALIDITY AND RELIABILITY
Presentation transcript:

Reliability or Validity Reliability gets more attention: n n Easier to understand n n Easier to measure n n More formulas (like stats!) n n Base for validity

Need for validity n Does test measure what it claims? n Can test be used to make decisions?

Validity Reliability is a necessary, but not a sufficient condition for validity.

Validity: a definition “A test is valid to the extent that inferences made from it are appropriate, meaningful, and useful” Standards for Educational and Psychological Testing, 1999

“Face Validity” “looks good to me”!!!!!!!

Trinitarian view of Validity n Content (meaning) n Construct (meaning) n Criterion (use)

1) Content Validity “How adequately a test samples behaviors representative of the universe of behaviors the test was designed to measure.”

Determining Content Validity describe the domain describe the domain specify areas to be measured specify areas to be measured compare test to domain compare test to domain

Content Validity Ratio (CVR) Agreement among raters if item is:  Essential  Useful but not essential  Not necessary

2) Construct validity “A theoretical intangible” “An informed, scientific idea” -- how well the test measures that construct

Determining Construct validity behaviors related to constructs behaviors related to constructs related/unrelated constructs related/unrelated constructs identify relationships identify relationships multi trait/multi method multi trait/multi method

Multitrait-Multimethod Matrix n Correlate scores from 2 (or more tests) n Correlate scores obtained from 2 (or more) methods

Evidence of Construct Validity n Upholds theoretical predictions u Changes (?) over time, gender, training n Homogeneity of questions u (internal consistency, factor or item analysis) n Convergent/discriminant u Multitrait-multimethod matrix

Decision Making How well the test can be used to help in decision making about a particular criterion.

Decision Theory n Base rate n Hit rate n Miss rate n False positive n False negative

3) Criterion Validity “The relationship between performance on the test and on some other criterion.”

Validity coefficient Correlation between test score and score on criterion measure.

Two ways to establish Criterion Validity A) Concurrent validity B) Predictive validity

Determining Concurrent validity n Assess individuals on construct n Administer test to lo/hi on construct n Correlate test scores to prior identification  Use test later to make decisions

Determining Predictive validity n Give test to group of people n Follow up group n Assess later n Review test scores  If correlate with behavior later can use later to make decisions

Incremental validity  Value of including more than one predictor  Based on multiple regression  What is added to prediction not present with previous measures?

Expectancy data n Taylor-Russell Table n Naylor-Shine Tables n Too vague, outdated, biased

Unified Validity - Messick “Validity is not a property of the test, but rather the meaning of the scores.” Value implications Relevance and utility

Unitarian considerations n Content n Construct n Criterion n Consequences

Threats to validity n Construct underrepresentation (too narrow) n Construct-irrelevant variance (too broad) construct-irrelevant difficulty construct-irrelevant difficulty construct-irrelevant easiness construct-irrelevant easiness

Example 1 Dr. Heidi considers using the Scranton Depression Inventory to help identify severity of depression and especially to distinguish depression from anxiety. What evidence should Dr. Heidi use to determine if the test does what she hopes it will do?

Example 2 The newly published Diagnostic Wonder Test promises to identify children with a mathematics learning disability. How will we know whether the test does so or is simply a slickly packaged general ability test?

Example 3 Ivy College uses the Western Admissions Test (WTA) to select applicants who should be successful in their studies. What type of evidence should we seek to determine if the WAT satisfies its purpose?

Example 4 Mike is reviewing a narrative report of his scores on the Nifty Personality Questionnaire (NPQ). The report says he is exceptionally introverted and unusually curious about the world around him. Can Mike have any confidence in these statements or should they be dismissed as equivalent to palm readings at the county fair?

Example 5 A school system wants to use an achievement battery that will measure the extent to which students are learning the curriculum specified by the school. How should the school system proceed in reviewing the available achievement tests?

Project homework question n What content or construct is your measure assessing? (explain your answer) n What do you think congruent and discriminate constructs would be to the one in your measure? n How would you determine the content or construct validity of your measure? n How would you determine the criterion validity of your measure? n Why would you use those approaches?

Project homework question n Select a standardized instrument from MMY to use as a comparison for your measure? n Copy the relevant data. n Why did you select that instrument? n How would you use it to help standardize your measure?