Appraisal and Its Application to Counseling COUN 550 Saint Joseph College For Week # 4 Copyright © 2004 by R. Halstead. All rights reserved.

Slides:



Advertisements
Similar presentations
Test Development.
Advertisements

Standardized Scales.
Measurement Concepts Operational Definition: is the definition of a variable in terms of the actual procedures used by the researcher to measure and/or.
Psychometrics William P. Wattles, Ph.D. Francis Marion University.
What is a Good Test Validity: Does test measure what it is supposed to measure? Reliability: Are the results consistent? Objectivity: Can two or more.
MGT-491 QUANTITATIVE ANALYSIS AND RESEARCH FOR MANAGEMENT
General Information --- What is the purpose of the test? For what population is the designed? Is this population relevant to the people who will take your.
Chapter 4A Validity and Test Development. Basic Concepts of Validity Validity must be built into the test from the outset rather than being limited to.
Chapter Fifteen Understanding and Using Standardized Tests.
Research Methods for Counselors COUN 597 University of Saint Joseph Class # 8 Copyright © 2015 by R. Halstead. All rights reserved.
Assessment: Reliability, Validity, and Absence of bias
Chapter 4 Validity.
VALIDITY.
Concept of Measurement
Copyright 2001 by Allyn and Bacon Standardized Testing Chapter 14.
Data Analysis Statistics. Levels of Measurement Nominal – Categorical; no implied rankings among the categories. Also includes written observations and.
Chapter 9 Flashcards. measurement method that uses uniform procedures to collect, score, interpret, and report numerical results; usually has norms and.
INTELLIGENCE AND PSYCHOLOGICAL TESTING. KEY CONCEPTS IN PSYCHOLOGICAL TESTING Psychological test: a standardized measure of a sample of a person’s behavior.
Chapter 7 Evaluating What a Test Really Measures
Classroom Assessment A Practical Guide for Educators by Craig A
Standardized Test Scores Common Representations for Parents and Students.
Understanding Validity for Teachers
Chapter 4. Validity: Does the test cover what we are told (or believe)
What is Intelligence? Definition: 3 main characteristics 1) 2) 3)
Measurement Concepts & Interpretation. Scores on tests can be interpreted: By comparing a client to a peer in the norm group to determine how different.
Chapter 14 Understanding and Using Standardized Tests Viewing recommendations for Windows: Use the Arial TrueType font and set your screen area to at least.
Reading Assessments for Elementary Schools Tracey E. Hall Center for Applied Special Technology Marley W. Watkins Pennsylvania State University Frank.
Ch 6 Validity of Instrument
Instrument Validity & Reliability. Why do we use instruments? Reliance upon our senses for empirical evidence Senses are unreliable Senses are imprecise.
Instrumentation.
Analyzing Reliability and Validity in Outcomes Assessment (Part 1) Robert W. Lingard and Deborah K. van Alphen California State University, Northridge.
Technical Adequacy Session One Part Three.
Psychometrics William P. Wattles, Ph.D. Francis Marion University.
Research Methods for Counselors COUN 597 University of Saint Joseph Class # 6 Copyright © 2015 by R. Halstead. All rights reserved.
Standardization and Test Development Nisrin Alqatarneh MSc. Occupational therapy.
Principles of Test Construction
Chapter 3 Understanding Test Scores Robert J. Drummond and Karyn Dayle Jones Assessment Procedures for Counselors and Helping Professionals, 6 th edition.
Instrumentation (cont.) February 28 Note: Measurement Plan Due Next Week.
Teaching Today: An Introduction to Education 8th edition
Chapter 4: Test administration. z scores Standard score expressed in terms of standard deviation units which indicates distance raw score is from mean.
Reliability & Validity
Validity Is the Test Appropriate, Useful, and Meaningful?
Counseling Research: Quantitative, Qualitative, and Mixed Methods, 1e © 2010 Pearson Education, Inc. All rights reserved. Basic Statistical Concepts Sang.
Measurement Validity.
Appraisal and Its Application to Counseling COUN 550 Saint Joseph College For Class # 3 Copyright © 2005 by R. Halstead. All rights reserved.
Assessing Learners with Special Needs: An Applied Approach, 6e © 2009 Pearson Education, Inc. All rights reserved. Chapter 5: Introduction to Norm- Referenced.
Assessment Information from multiple sources that describes a student’s level of achievement Used to make educational decisions about students Gives feedback.
Chapter 4 Validity Robert J. Drummond and Karyn Dayle Jones Assessment Procedures for Counselors and Helping Professionals, 6 th edition Copyright ©2006.
Validity and Item Analysis Chapter 4. Validity Concerns what the instrument measures and how well it does that task Not something an instrument has or.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
McGraw-Hill/Irwin © 2012 The McGraw-Hill Companies, Inc. All rights reserved. Obtaining Valid and Reliable Classroom Evidence Chapter 4:
Chapter 9 Correlation, Validity and Reliability. Nature of Correlation Association – an attempt to describe or understand Not causal –However, many people.
Appraisal and Its Application to Counseling COUN 550 Saint Joseph College Ability, Intelligence, Aptitude and Achievement Testing For Class #12 Copyright.
Psychometrics. Goals of statistics Describe what is happening now –DESCRIPTIVE STATISTICS Determine what is probably happening or what might happen in.
Criteria for selection of a data collection instrument. 1.Practicality of the instrument: -Concerns its cost and appropriateness for the study population.
Measurement MANA 4328 Dr. Jeanne Michalski
Chapter 7 Measuring of data Reliability of measuring instruments The reliability* of instrument is the consistency with which it measures the target attribute.
Chapter 3 Selection of Assessment Tools. Council of Exceptional Children’s Professional Standards All special educators should possess a common core of.
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 16 Communicating Assessment.
Educational Research Chapter 8. Tools of Research Scales and instruments – measure complex characteristics such as intelligence and achievement Scales.
WHS AP Psychology Unit 7: Intelligence (Cognition) Essential Task 7-3:Explain how psychologists design tests, including standardization strategies and.
Unit 8: Intelligence (Cognition)
Concept of Test Validity
Reliability & Validity
Week 3 Class Discussion.
پرسشنامه کارگاه.
Understanding and Using Standardized Tests
Chapter 4 Standardized Tests.
EDUC 2130 Quiz #10 W. Huitt.
Assessment Chapter 3.
Presentation transcript:

Appraisal and Its Application to Counseling COUN 550 Saint Joseph College For Week # 4 Copyright © 2004 by R. Halstead. All rights reserved.

Class Topics Validity Administering Tests & Communicating Results Scoring & Interpreting Tests Norms Standard Scores

I was just looking at the tree across the street. The distance variance of the fallen leaves from the trunk of the tree looks like they are normally distributed. Seeing that phenomenon is like looking into the eyes of Dog!!

Test Validity Validity - a means of expressing the degree to which a certain inference drawn from a test is appropriate and meaningful. Stated more simply, validity helps us answer the following question: Does a test measure what it purports to measure?

The Multidimensional Nature of Validity Validity - a means of expressing the degree to which a certain inference drawn from a test is appropriate and meaningful. There are two different types of inferences drawn from tests: Descriptive inferences and Predictive inferences Because validity, in a sense, is multidimensional in nature, one must use a variety of methods to establish if a test is valid.

Categories of Validity Most test manuals refer to three categories of validity Content Validity Criterion Validity Construct Validity

Categories of Validity Content Validity - The items of a test represent that which they are supposed to measure. Content Validity is often supported by expert judgment. Face Validity is also a term used in some test manuals to express that a panel of judges has held that test appears to measure what it purports to measure.

Categories of Validity Criterion-Related Validity - A test score is related to outcome criteria. There are two elements addressed when considering the Criterion Validity Concurrent Validity - the test score agrees (concurs) with other valid measures of the same construct. Predictive Validity - the test score is able to accurately predict performance within some domain that it purports to measure.

Criterion-Related Validity Concurrent Validity Concurrent Validity is established through the use of correlation. “The scores on the Aggressiveness Scale correlated.70 with the rating of teachers of students in their classes.” “The scores on the Beck Hopelessness Scale correlated.81 with the Beck Depression Inventory.”

Criterion-Related Validity Predictive Validity If a test is used to estimate criterion scores at some point in the future, predictive validity must be established. Predictive Validity - the test score is able to accurately predict performance within some domain that is purports to measure. The SAT score, for example, that one establishes in high school is used to predict performance level during one’s first year performance level in college.

Criterion-Related Validity Predictive Validity Some Examples: The SAT Verbal scores have been shown to correlate (.40) with first year students’ grade point average at the end of the first semester at SJC. The Spatial Relations Scale correlated.70 with success in the mental fabricating training for technical high school students.

Criterion-Related Validity Predictive Validity Aptitude and intelligence tests offer evidence regarding predictive validity. When one looks at these tests it is a good idea to examine the utility of predictive tests. Below is a model for considering predictive tests False Positive Positive Negative False Negative

Criterion-Related Validity Concurrent and Predictive Validity Concurrent and Predictive Validity To establish either concurrent or predictive validity it is critically important to make certain that valid, reliable, and relevant measures of the criterion are used.

Construct Validity Construct - is a mental construction of some grouping of variables or behaviors (e.g. anxiety, locus of control, cognitive ability). Construct Validity - establishes that a test score expresses and accurate measure the construct in question.

Construct Validity Because constructs are mental constructions we must derive Construct Validity This is done by examining sets of evidence to build a case for establishing the test's validity. Convergent Validity - high positive correlation with other measures of the same construct Discriminate Validity - low correlation with other measures of a different construct

Multitrait-multimethod Approach to Construct Validity Multitrait-multimethod Approach 1) compute r for same measure of same trait same method 2) compute r for same measure of same trait different method 3) compute r for measure of different trait same method 4) compute r for same measure of different trait different method

Multitrait-multimethod Approach to Construct Validity - An Example Lets suppose we want to establish construct validity for a new depression scale 1) compute r for measure with one that measures the same trait - Beck Depression Scale 2) compute r for measure with a different measurement method same trait - Clinical Interview 3) compute r for measure with one that measures a different trait - Snyder’s Hope Scale 4) compute r for measure with a different trait (Hope) and different method - Clinical Interview

Awareness and Orientation Codes of ethical standards remind test administrators of their responsibility for the orientation of takers. Orientation should describe the purpose of testing, content areas measured, method of administration as well as reporting and use of scores

7 topics to cover in orientation Purpose Criteria used for selecting the test Conditions under which test to be taken Range of skills or domains to be measured Administrative procedures and concerns Types of questions on test and overview Type of scoring, method, + schedule for reporting

Administering the Test Important to deliver instructions in test manual as they are stated. Follow the sequence and timing of the instructions so that your client gets that same information as those in the group on which the test was that normed. Goal for the results to give a valid picture of attributes measured. Examiner should record critical Incidents that deviate from normal conditions.

Posttesting Procedures Recording Test Behavior – checklists/behaviors Major Issues and Problems in Test Administration awareness/orientation phase very important in eliminating probs Examiner and Bias – communication skills/attitudes + expectations/gender/competence/test ethics = examiner bias Feedback on Test and Test Administration – Ask test taker.. how would you rate physical environment, performance, comfort with the administration, time allotted adequate, orientation helpful, fair test?

Communicating Results to Clients, Parents, and Professionals Guidelines & Standards for Communicating Test Results Know test manual, limits of test, informed consent procedures, protect rights Postinterview opportunity to deal w/ interpretation and its use in planning and decision making

Methods for Reporting Test Results 5 major methods Individual Sessions: discussion Group Sessions Written reports Interactive Approaches Video Approaches: current trend

Areas of Consideration Acceptance: goal of feedback session is acceptance of test results and incorporate info into decision making Readiness of client: critical factor in acceptance = client readiness Negative results Flat Profiles: look at other things, interests, values, goals, etc. Motivation and Attitude: more significant when motivated

Scoring Tests Primary models for scoring tests Cumulative – Number of items endorsed Class – Serves to categorize or describe the person Ipsative – How a person performed on a set of variables Alternate & Authentic Assessment Holistic Scoring –individual judgment with model answers

Interpreting Tests Criterion-Referenced Tests describes the specific types of skills, tasks, or knowledge that the test taker can demonstrate. Norm-Referenced Tests compare each person’s score with that of the norm established for that test.

Norms What establishes a norm? Answer: Norms are based on the occurrence of the majority, or bulk, of values within a distribution obtained from some defined sample of individuals (Think of the Normal or Bell Curve). Norms are used to give information about performance relative to what has been observed within a sample.

Age-Related Norms Some normative groups have been established for particular ages. Once age-related norms have been established tracking is possible. Tracking is a process by which, with some level of confidence, one can expect to see specific characteristics in a sample of interest. Examples: Developmental activities of infants, Grade based reading levels, Age based capacity for task of increasing cognitive complexity

Within-Group Norms Establishing norms is very easy to do. It is important to keep mindful that the science of testing is not perfect. One must be cautious, therefore, about assuming that a test is a valid measure for everyone taking it. If a test provides an accurate measure for only some members of a population accuracy of within-group norms must be questioned.

Percentile, Percentile Ranks, Quartiles, and Deciles Percentiles - divide the total frequency for a set of observations into hundredths Percentile Rank - establishes that point below which a certain percent of scores fall Quartile - divide the total frequency for a set of observations into quarters Q1 = 25% Q2 = 50% Q3 = 75% Q4 = 100% Deciles - divide the total frequency for a set of observations into tenths

Standard Scores One of the problems with establishing a mean and standard deviation for a specific distribution is the limitations to the meaning of such measures. Standard scores are a means of presenting the relative position of an individual on a test that are tied to the normal curve. There are two major forms of standard scores that you will encounter. The Z Score and the T Score

Z Scores Z Scores are based on an mean of 0 and a standard deviation of 1. Z Score Table - Appendix 1 - Salkind x - X Z = s Z = =

Z Scores and the Normal Curve Z Score Table - Appendix 1 in Salkind Z Scores give us a opportunity to establish a better understanding of where individual scores occur relative to standardized norms.

T Score T Scores - are standard scores using a fixed mean and standard deviation in units that eliminate the need for decimals and signs. On many tests the arbitrary mean or fixed mean is 50 and the arbitrary standard deviation is 10. T = s(z) + { X} T = 10(-.50) + 50 T = T = 45