Download presentation
Presentation is loading. Please wait.
Published byGarry McKinney Modified over 8 years ago
1
Educational Research Chapter 5 Selecting Measuring Instruments Gay and Airasian
2
Topics Discussed in this Chapter Data collection Measuring instruments Terminology Interpreting data Types of instruments Technical issues Validity Reliability Selection of a test
3
Data Collection Scientific and disciplined inquiry requires the collection, analysis, and interpretation of data Data – the pieces of information that are collected to examine the research topic Issues related to the collection of this information are the focus of this chapter
4
Data Collection Terminology related to data Constructs – abstractions that cannot be observed directly but are helpful when trying to explain behavior Intelligence Teacher effectiveness Self esteem
5
Data Collection Data terminology (continued) Operational definition – the ways by which constructs are observed and measured Weschler IQ test Virgilio Teacher Effectiveness Inventory Tennessee Self-Concept Scale Variable – a construct that has been operationalized and has two or more values
6
Data Collection Measurement scales Nominal – categories Gender, ethnicity, etc. Ordinal – ordered categories Rank in class, order of finish, etc. Interval – equal intervals Test scores, attitude scores, etc. Ratio – absolute zero Time, height, weight, etc.
7
Data Collection Types of variables Categorical or continuous Categorical variables reflect nominal scales Continuous variables reflect ordinal, interval or ratio scales Independent or dependent Independent variables are the purported causes Dependent variables are the purported effects
8
Measurement Instruments Important terms Instrument – a tool used to collect data Test – a formal, systematic procedure for gathering information Assessment – the general process of collecting, synthesizing, and interpreting information Measurement – the process of quantifying or scoring a subject’s performance
9
Measurement Instruments Important terms (continued) Cognitive tests – examining subject’s thoughts and thought processes Affective tests – examining subject’s feelings, interests, attitudes, beliefs, etc. Standardized tests – tests that are administered, scored, and interpreted in a consistent manner
10
Measurement Instruments Important terms (continued) Selected response item format– respondents select answers from a set of alternatives Supply response item format – respondents construct answers Individual tests – tests administered on an individual basis Group tests – tests administered to a group of subjects at the same time Performance assessments – assessments that focus on subject’s processes or products that have been created
11
Measurement Instruments Interpreting data Raw scores – the actual score made on a test Standard scores – statistical transformations of raw scores Percentiles Stanines Normal Curve Equivalents
12
Measurement Instruments Interpreting data (continued) Norm-referenced – scores are interpreted relative to the scores of others taking the test Criterion-referenced – scores are interpreted relative to a predetermined level of performance Self-referenced – scores are interpreted relative to changes over time
13
Measurement Instruments Types of instruments Cognitive – measuring intellectual processes such as thinking, memorizing, problem solving, analyzing, or reasoning Aptitude – measuring general mental ability usually for predicting future performance
14
Measurement Instruments Types of instruments (continued) Affective – assessing individual feelings, values, attitudes, beliefs, etc. Typical affective characteristics of interest Values – deeply held beliefs about ideas, persons, or objects Attitudes – dispositions to favorable or unfavorable toward things Interests – inclinations to seek out or participate in particular activities, objects, ideas, etc. Personality – characteristics that represent a person’s typical behaviors
15
Measurement Instruments Types of instruments (continued) Affective (continued) Scales used for responding to items on affective tests Likert Semantic differential Thurstone Guttman Rating scales
16
Measurement Instruments Issues for cognitive, aptitude, or affective tests Bias – distortions of a respondent’s performance or responses based on ethnicity, race, gender, language, etc. Responses to affective test items Socially acceptable responses Accuracy of responses Response sets Problems inherent in the use of self-report measures and the use of projective tests
17
Technical Issues Validity – extent to which interpretations made from a test score are appropriate Characteristics The most important technical characteristic Situation specific Does not refer to the instrument but to the interpretations of scores on the instrument Best thought of in terms of degree
18
Technical Issues Validity (continued) Four types Content – to what extent does the test measure what it is supposed to measure Item validity Sampling validity Determined by expert judgment
19
Technical Issues Validity (continued) Criterion-related Predictive – to what extent does the test predict a future performance Concurrent - to what extent does the test predict a performance measured at the same time Estimated by correlations between two tests Construct – the extent to which a test measures the construct it represents Underlying difficulty defining constructs Estimated in many ways
20
Technical Issues Validity (continued) Consequential – to what extent are the consequences that occur from the test harmful Estimated by empirical and expert judgment Factors affecting validity Unclear test directions Confusing and ambiguous test items Vocabulary that is too difficult for test takers
21
Technical Issues Validity (continued) Factors affecting validity Overly difficult and complex sentence structure Inconsistent and subjective scoring Untaught items Failure to follow standardized administration procedures Cheating by the participants or someone teaching to the test items
22
Technical Issues Reliability – the degree to which a test consistently measures whatever it is measuring Characteristics Expressed as a coefficient ranging from 0 to 1 A necessary but not sufficient characteristic of a test
23
Technical Issues Reliability (continued) Six reliability coefficients Stability – consistency over time with the same instrument Test – retest Estimated by a correlation between the two administrations of the same test Equivalence – consistency with two parallel tests administered at the same time Parallel forms Estimated by a correlation between the parallel tests
24
Technical Issues Reliability (continued) Six reliability coefficients (continued) Equivalence and stability – consistency over time with parallel forms of the test Combines attributes of stability and equivalence Estimated by a correlation between the parallel forms Internal consistency – artificially splitting the test into halves Several coefficients – split halves, KR 20, KR 21, Cronbach alpha All coefficients provide estimates ranging from 0 to 1
25
Technical Issues Reliability (continued) Six reliability coefficients Scorer/rater – consistency of observations between raters Inter-judge – two observers Intra-judge – one judge over two occasions Estimated by percent agreement between observations
26
Technical Issues Reliability (continued) Six reliability coefficients (continued) Standard error of measurement (SEM) – an estimate of how much difference there is between a person’s obtained score and his or her true score Function of the variation of the test and the reliability coefficient (e.g., KR 20, Cronbach alpha, etc.) Estimated by specifying an interval rather than a point estimate of a person’s score
27
Selection of a Test Sources of test information Mental Measurement Yearbooks (MMY) Provides factual information on all known tests Provides objective test reviews Comprehensive bibliography for specific tests Indices: titles, acronyms, subject, publishers, developers Buros Institute
28
Selection of a Test Sources (continued) Tests in Print Bibliography of all known commercially produced tests currently available Very useful to determine availability Tests in Print
29
Selection of a Test Sources (continued) ETS Test Collection Published and unpublished tests Includes test title, author, publication date, target population, publisher, and description of purpose Annotated bibliographies on achievement, aptitude, attitude and interests, personality, sensory motor, special populations, vocational/occupational, and miscellaneous ETS Test Collection
30
Selection of a Test Sources (continued) ERIC/AE Test Locator Search for citations about a particular instrument Search for names and addresses of test publishers ERIC/AE Test Locator
31
Selection of a Test Sources (continued) Professional journals Test publishers and distributors Issues to consider when selecting tests Psychometric properties Validity Reliability Length of test Scoring and score interpretation
32
Selection of a Test Issues to consider when selecting tests Non-psychometric issues Cost Administrative time Objections to content by parents or others Duplication of testing
33
Selection of a Test Designing you own tests Get help from others with experience developing tests Item writing guidelines Avoid ambiguous and confusing wording and sentence structure Use appropriate vocabulary Write items that have only one correct answer Give information about the nature of the desired answer Do not provide clues to the correct answer See Writing Multiple Choice ItemsWriting Multiple Choice Items
34
Selection of a Test Test administration guidelines Plan ahead Be certain that there is consistency across testing sessions Be familiar with any and all procedures necessary to administer a test
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.