Nurhayati, M.Pd Indraprasta University Jakarta.  Validity : Does it measure what it is supposed to measure?  Reliability: How the representative is.

Slides:



Advertisements
Similar presentations
Standardized Scales.
Advertisements

Chapter 8 Flashcards.
Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
The Research Consumer Evaluates Measurement Reliability and Validity
VALIDITY AND RELIABILITY
Part II Sigma Freud & Descriptive Statistics
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
RESEARCH METHODS Lecture 18
Chapter 4 Validity.
VALIDITY.
Concept of Measurement
Beginning the Research Design
VALIDITY & RELIABILITY Raja C. Bandaranayake. QUALITIES OF MEASUREMENT DEVICES  Validity Does it measure what it is supposed to measure?  Reliability.
Research Methods in MIS
Classroom Assessment A Practical Guide for Educators by Craig A
Questions to check whether or not the test is well designed: 1. How do you know if a test is effective? 2. Can it be given within appropriate administrative.
 Rosseni Din  Muhammad Faisal Kamarul Zaman  Nurainshah Abdul Mutalib  Universiti Kebangsaan Malaysia.
VALIDITY. Validity is an important characteristic of a scientific instrument. The term validity denotes the scientific utility of a measuring instrument,
Measurement and Data Quality
Validity and Reliability
Instrumentation.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Unanswered Questions in Typical Literature Review 1. Thoroughness – How thorough was the literature search? – Did it include a computer search and a hand.
Analyzing Reliability and Validity in Outcomes Assessment (Part 1) Robert W. Lingard and Deborah K. van Alphen California State University, Northridge.
Technical Adequacy Session One Part Three.
Principles in language testing What is a good test?
WELNS 670: Wellness Research Design Chapter 5: Planning Your Research Design.
Student assessment AH Mehrparvar,MD Occupational Medicine department Yazd University of Medical Sciences.
The Basics of Experimentation Ch7 – Reliability and Validity.
Chap. 2 Principles of Language Assessment
Week 5 Lecture 4. Lecture’s objectives  Understand the principles of language assessment.  Use language assessment principles to evaluate existing tests.
Formal Assessment Week 6 & 7. Formal Assessment Formal assessment is typical in the form of paper-pencil assessment or computer based. These tests are.
Measurement Validity.
Advanced Research Methods Unit 3 Reliability and Validity.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Presented By Dr / Said Said Elshama  Distinguish between validity and reliability.  Describe different evidences of validity.  Describe methods of.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
Measurement and Scaling
Chapter 7 Measuring of data Reliability of measuring instruments The reliability* of instrument is the consistency with which it measures the target attribute.
 A test is said to be valid if it measures accurately what it is supposed to measure and nothing else.  For Example; “Is photography an art or a science?
Reliability and Validity Themes in Psychology. Reliability Reliability of measurement instrument: the extent to which it gives consistent measurements.
Chapter 6 - Standardized Measurement and Assessment
VALIDITY, RELIABILITY & PRACTICALITY Prof. Rosynella Cardozo Prof. Jonathan Magdalena.
Validity & Reliability. OBJECTIVES Define validity and reliability Understand the purpose for needing valid and reliable measures Know the most utilized.
TEST SCORES INTERPRETATION - is a process of assigning meaning and usefulness to the scores obtained from classroom test. - This is necessary because.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
RELIABILITY AND VALIDITY Dr. Rehab F. Gwada. Control of Measurement Reliabilityvalidity.
Measurement and Scaling Concepts
ESTABLISHING RELIABILITY AND VALIDITY OF RESEARCH TOOLS Prof. HCL Rawat Principal UCON,BFUHS Faridkot.
Reliability and Validity
Ch. 5 Measurement Concepts.
Lecture 5 Validity and Reliability
Concept of Test Validity
Selecting Employees – Validation
HRM – UNIT 10 Elspeth Woods 9 May 2013
Tests and Measurements: Reliability
Journalism 614: Reliability and Validity
Reliability & Validity
Outline the steps خطوات in the selection اختيار process عملية
Human Resource Management By Dr. Debashish Sengupta
پرسشنامه کارگاه.
Reliability and Validity of Measurement
Analyzing Reliability and Validity in Outcomes Assessment Part 1
VALIDITY Ceren Çınar.
RESEARCH METHODS Lecture 18
Chapter 4 Characteristics of a Good Test
Measurement Concepts and scale evaluation
Analyzing Reliability and Validity in Outcomes Assessment
Chapter 8 VALIDITY AND RELIABILITY
Presentation transcript:

Nurhayati, M.Pd Indraprasta University Jakarta

 Validity : Does it measure what it is supposed to measure?  Reliability: How the representative is the measurement?  Practicality: Is it easy to construct, administer, score and interpret?  Objectivity: Do independent scorers agree?

 Validity refers to whether or not a test measures what it intends to measure.  A test with high validity has items closely linked to thetest’s intended focus. A test with poor validity does not measure the content and competencies it ought to.

 Content: related to objectives and their sampling.  Construct: reffering to the theoryunderlying the target  Criterion: related to concrete criteria in the real word. It can be concurrent or predictive  Face: related to the test overall appearance

Content validity refers to the connections between the test items and the subject, related tasks. The test should evaluate only the content related to the field of study in a manner sufficiently representative, relevant, and comprehensible.

It implies using the construct (concept, ideas, notion) in accordance to the state of the art in the filed construct validity seeks agreement between updated subject-matter theories and the specific measuring components of the test. Example; a test of intelligence nowadays must include measures of multiple intelligences, rather than just logical- mathematical and linguistic ability measures.

Also refered to as instrumental validity. It is used to demostrate the accuracy of a measure or procedure by comparing it with another process or method which has has been demostrated to be valid. Example: imagine a hands-on driving test has been proved to be an accurate test of driving skills. A written test can be validated by using a criterion related strategy in which the hands on driving test is compared to it

 Concurrent validity uses statistical methods of correlation to other measures.  Examinees who are known to be eiher masters or non masters on the content measured are identified before the test is administered. Once the tests have been scored, the relationship between the examinees’ status as either masters or non masters and their performance (pass or fail)is estimated based on the test.

 Predictive validity estimates the relationship of test scores to an examinee’s future performance master or non master. Predictive validity considered the question, “ how well does the test predict examinees’ future status as master or non master?”  For this type of validity, the correlation that is computed is based on the test results and the examinee’s later performances. This type of validity is especially useful for test purposes such as selection or admissions

 Face Validity is determined by a review of the items and not through the use of statistical analyses.unlike content validity, face validity is not investigated through formal procedures. Instead, anyone who look over the test, including examinees, may develop an informal opinion as to whether or not the test is measuring what it is supposed to measure.

Realiability is the extent to which an experiment, test, or any measuring procedure shows the same result an repeated trials.

 Equivalency: related to the co-occurrence of two items.  Stability: related to time consistency.  Internal: related to the instruments.  Interrater: related to the examiners’ criterion.

Equivalency reliability is the extent to which two items measure identical concepts at an identical level of difficulty. Equivalency reliability is determined by relating two set of test scores to one another to highlight the degree of relationship or association.

Stability reliability (sometimes called retest reliability) is the agreement of measuring instruments over time. To determine stability a measure or test is repeated onthe same subjects at a future date. Results arecompared and correlated withthe initial test to give a measure of stability. Instruments with a high stability reliability arethermometers, compasses, measuring cups.

Internal consistency is the extent to which tests or procedures assess the same characteristic, skill or quality. It is a measure of the precision between the measuring instrument used in a study. This type of reliability often helps researchers interpret data and predict the value of scores and the limits of the relationship among variables.

Interrrater reliability is the extent to which more individuals agree. Example: when two or more teachers use a rating scale with which they are rating the students’ oral responses in an interview (1 being most negative, 5 being most positive). If one researcher gives a “1” to a student response, while another researcher give a “5” obviously the interrater reliability would be inconsistent

Validity and reliability are closely related. A test can not be considered valid unless the measurements resulting from it are reliable. Likewise, results from a test can be reliable and not necessarily valid.

Backwash (also known as washback) effect is related to the potentially positive and negative effects of test design and content on the form and content of English Language training courseware.