Download presentation
Presentation is loading. Please wait.
Published byKellie Benson Modified over 9 years ago
1
Chapter 6 - Standardized Measurement and Assessment http://www.offthemarkcartoons.com/cartoons/2005-08-21.gif
2
What is measurement? the act of measuring assigning symbols or numbers to something according to a specific set of rules
3
What are the four different levels or scales of measurement? N ominal Scale O rdinal Scale I nterval Scale R atio Scale
4
What is essential to know about the Nominal Scale? it’s the simplest form of measurement it uses symbols, such as words or numbers it measures categorical variables LABEL IDENTIFY CLASSIFY
5
What is essential to know about the Ordinal Scale? it’s a rank-order scale it doesn’t indicate how much greater one ranking is over another
6
What is essential to know about the Interval Scale? it’s also a rank-order scale includes equal distances or intervals between adjacent numbers the absence of a zero points means you cannot make “ratio statements”
7
What is essential to know about the Ratio Scale? it’s the highest level of quantitative measure it has all the properties of the nominal, ordinal, and interval scales plus it has a true zero point it is not often used in educational research
9
How do we define testing? the measurement of variables
10
How do we define assessment? gathering data to make evaluations
11
How do we define error? the difference between true scores and observed scores
12
How do we define traits? distinguishable, enduring ways in which one individual differs from another
13
How do we define states? distinguishable but less enduring ways in which individuals vary
14
What are the twelve assumptions underlying testing and measurement? psychological traits & states exist psychological traits & states can be measured various approaches to measurement can be useful assessments can answer some of life’s most important questions 1 1 3 3 2 2 4 4
15
What are the twelve assumptions underlying testing and measurement? assessment can pinpoint phenomena that require further study various sources of data enrich & are part of the assessment process various sources of error are always part of the assessment process measurement techniques have strengths & weaknesses 5 5 7 7 6 6 8 8
16
What are the twelve assumptions underlying testing and measurement? test-related behavior predicts non-test related behavior testing & assessment can be done in a fair and unbiased way present-day sampling predicts future behavior testing & assessment benefit society 9 9 10 11 12
17
What is the difference between reliability and validity? Reliability refers to the consistency of a score Validity refers to the accuracy of the interpretations you make from the scores If you want validity, you must have reliability.
18
What is a reliability coefficient? a correlation coefficient that is used as an index of reliability researchers want reliability coefficients to be as close to +1.00 as possible
19
What are four different ways of assessing reliability? 1.Test-Retest Reliability 2.Equivalent Forms Reliability 3.Internal Consistency Reliability 4.Interscorer Reliability
20
What is test-retest reliability? a measure of the consistency of scores over time the time interval can have an effect on test- retest reliability because people change over time
21
What is equivalent forms reliability? the consistency of a group of individuals’ scores on two equivalent forms of a test measuring the same thing the success of this method depends on the ability to construct two equivalent forms of the same test
22
What is internal consistency reliability? the consistency with which the items on a test measure a single construct
23
What is split-half reliability? splitting a test into two equivalent halves and then assessing the consistency of the scores across the two halves of the test each half needs to be equal to the other in format, style, content, and other aspects
24
What is coefficient alpha? a formula that provides an estimate of the reliability of a homogeneous test or an estimate of the reliability of each dimension in a multidimensional test tells you the degree to which the items are interrelated need to consider the number of items; don’t just assume that because the coefficient alpha is large, the items are strongly related
25
What is interscorer reliability? the degree of agreement or consistency between two or more scorers, judges, or raters some degree of training and practice for the scorers is advised to improve the reliability of an evaluation
26
What is the definition of validity? the accuracy of the inferences, interpretations, or actions made on the basis of test scores to make sure that our test is measuring what we intended it to measure for the particular people in a particular context and that the interpretations we make on the basis of the test scores are correct we want our inferences to be accurate and our actions to be appropriate
27
What is the definition of validity evidence? the empirical evidence and theoretical rationales that support the inferences or interpretations made from the test scores
28
What is the definition of validation? the process of gathering evidence that supports inferences made on the basis of test scores the best rule is to collect multiple sources of evidence validation should be viewed as a never- ending process
29
What are the characteristics of the different ways of obtaining validity evidence? Evidence based on content Evidence based on internal structure Evidence based on relations to other variables
30
What are the characteristics of evidence based on content? content-related evidence is when you make a judgment of the degree to which the evidence suggests that the items, tasks, or questions on the test adequately represent the domain of interest it’s based on item content, but it is also based on the formatting, working, administration, and storing of the test
31
What are the characteristics of evidence based on internal structure? Factor Analysis – a statistical procedure that analyzes the relationships among items to determine whether a test is unidimensional or multidimensional Homogeneity – refers to how well the different items in a test measure the same construct or trait
32
What are the characteristics of evidence based on relations to other variables? Criterion Criterion-Related Evidence Validity coefficient Concurrent evidence Predictive evidence Convergent evidence Discriminant evidence Known groups evidence
33
Define criterion. the standard or benchmark that you want to predict accurately on the basis of the test scores
34
Define criterion-related evidence. validity evidence based on the extent to which scores from a test can be used to predict or infer performance on some criterion such as a test or future performance
35
Define validity coefficient. a correlation coefficient that is computed to provide validity evidence, such as the correlation between test scores and criterion scores
36
Define concurrent evidence. validity evidence based on the relationship between test scores and criterion scores obtained at the same time
37
Define predictive evidence. validity evidence based on the relationship between test scores collected at one point in time and criterion scores obtained at a later time
38
Define convergent evidence. validity evidence based on the relationship between the focal test scores and independent measure of the same construct
39
Define discriminant evidence. evidence that the scores on your focal test are not highly related to the scores from other tests that are designed to measure theoretically different constructs
40
Define known groups evidence. evidence that groups that are known to differ on the construct do differ on the test in the hypothesized direction
41
What are the characteristics of the different ways of obtaining validity evidence? norming group - the specific group for whaich the test publishers or researcher provides evidence for test validity and reliability it is not wise to rely solely on previously reported reliability and validity information the characteristics of you participants should closely match the characteristics of the norming group
42
What are some different types of psychological tests? intelligence tests - the ability to think abstractly and to learn readily from experience personality tests - patterns that characterize and classify people –self-report - participants rate themselves –performance measures - participants perform some real-life, observable behavior –projective measures - participants provide responses to ambiguous stimuli
43
What are some different types of educational assessment tests? preschool assessment tests - assess the various behaviors and cognitive skills of young children achievement tests - designed to measure the degree of learning that has taken place after a person is exposed to a specific learning experience aptitude tests - focuses on the information acquired through the informal learning that goes on through life diagnostic tests - identify where a student is having difficulty
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.