© 2007 Thomson Brooks/Cole, a division of Thomson Learning Chapter 12: Testing and Assessment Chapter 13: Research and Evaluation 1.

Slides:



Advertisements
Similar presentations
Chapter 6 Process and Procedures of Testing
Advertisements

Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 5 Reliability.
VALIDITY AND RELIABILITY
Evaluation of the Iowa Algebra Aptitude Test Terri Martin Doug Glasshoff Mini-project 1 June 17, 2002.
General Information --- What is the purpose of the test? For what population is the designed? Is this population relevant to the people who will take your.
Chapter Fifteen Understanding and Using Standardized Tests.
Reliability and Validity of Research Instruments
Chapter 4 Validity.
1 School Counseling PowerPoint produced by Melinda Haley, M.S., New Mexico State University. “This multimedia product and its contents are protected under.
Test Validity: What it is, and why we care.
Chapter 5 Instrument Selection, Administration, Scoring, and Communicating Results.
Legal and Ethical Issues in Assessment Ignorance is no defense.
Copyright 2001 by Allyn and Bacon Standardized Testing Chapter 14.
SELECTION & ASSESSMENT SESSION THREE: MEASURING THE EFFECTIVENESS OF SELECTION METHODS.
1 Chapter 6: Assessment of Educational Ability Survey Battery Diagnostic Readiness Cognitive Ability Tests.
Assessing Intelligence
Standardized Test Scores Common Representations for Parents and Students.
AP Unit 11 Testing and Individual Differences pt. 1
Test Validity S-005. Validity of measurement Reliability refers to consistency –Are we getting something stable over time? –Internally consistent? Validity.
Measurement Concepts & Interpretation. Scores on tests can be interpreted: By comparing a client to a peer in the norm group to determine how different.
Chapter 14 Understanding and Using Standardized Tests Viewing recommendations for Windows: Use the Arial TrueType font and set your screen area to at least.
Test Worthiness Chapter 3.
Reading Assessments for Elementary Schools Tracey E. Hall Center for Applied Special Technology Marley W. Watkins Pennsylvania State University Frank.
Standardized Testing (1) EDU 330: Educational Psychology Daniel Moos.
COPYRIGHT © 2014 Brooks/Cole*Wadsworth Publishing Company A division of Cengage Inc. 1 Applying Career Development Theory to Counseling, 6e by Richard.
Instrument Validity & Reliability. Why do we use instruments? Reliance upon our senses for empirical evidence Senses are unreliable Senses are imprecise.
Instrumentation.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
MEASUREMENT CHARACTERISTICS Error & Confidence Reliability, Validity, & Usability.
Psychometrics William P. Wattles, Ph.D. Francis Marion University.
Chapter 7 Selection Group 7 August 24, Employee Selection Selection is the process of choosing from a group of applicants those individuals best.
Instrumentation (cont.) February 28 Note: Measurement Plan Due Next Week.
Appraisal and Its Application to Counseling COUN 550 Saint Joseph College For Week # 4 Copyright © 2004 by R. Halstead. All rights reserved.
Welcome to COU 660: Assessment & Evaluation in Counseling Introductory Material Ethical & Legal Issues.
Assessment Procedures for Counselors and Helping Professionals, 7e © 2010 Pearson Education, Inc. All rights reserved. Chapter 7 Selecting, Administering,
Reliability & Validity
Counseling Research: Quantitative, Qualitative, and Mixed Methods, 1e © 2010 Pearson Education, Inc. All rights reserved. Basic Statistical Concepts Sang.
Assessment in Counseling Chapter 1. 2 What is assessment? Needs to be systematic and objective Sample(s) of behavior from which we make inferences Measurement.
EDU 8603 Day 6. What do the following numbers mean?
Educational Psychology: Theory and Practice Chapter 14 Standardized Tests This multimedia product and its contents are protected under copyright law. The.
Formal Assessment Week 6 & 7. Formal Assessment Formal assessment is typical in the form of paper-pencil assessment or computer based. These tests are.
Session 7 Standardized Assessment. Standardized Tests Assess students’ under uniform conditions: a) Structured directions for administration b) Procedures.
ETHICS AND LAW FOR SCHOOL PSYCHOLOGISTS Chapters 6 and 8.
Chapter 8 Validity and Reliability. Validity How well can you defend the measure? –Face V –Content V –Criterion-related V –Construct V.
1 Chapter 1 History of Testing and Assessment 2 Testing Vs. Assessment Assessment includes a broad array of evaluation procedures, such as: The Clinical.
Evaluating Survey Items and Scales Bonnie L. Halpern-Felsher, Ph.D. Professor University of California, San Francisco.
Copyright © Allyn & Bacon 2008 Intelligent Consumer Chapter 14 This multimedia product and its contents are protected under copyright law. The following.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
Chapter 9 Correlation, Validity and Reliability. Nature of Correlation Association – an attempt to describe or understand Not causal –However, many people.
Appraisal and Its Application to Counseling COUN 550 Saint Joseph College Ability, Intelligence, Aptitude and Achievement Testing For Class #12 Copyright.
PSY 231: Introduction to Industrial and Organizational Psychology
Measurement MANA 4328 Dr. Jeanne Michalski
COPYRIGHT © 2014 Brooks/Cole*Wadsworth Publishing Company A division of Cengage Inc. 1 Applying Career Development Theory to Counseling, 6e by Richard.
Assessment in Counseling Chapter 1 Levi Armstrong, Psy.D.
CHAPTER 6 Selecting Employees and Placing Them in Jobs
Chapter 6 - Standardized Measurement and Assessment
Chapter 3 Selection of Assessment Tools. Council of Exceptional Children’s Professional Standards All special educators should possess a common core of.
Educational Research Chapter 8. Tools of Research Scales and instruments – measure complex characteristics such as intelligence and achievement Scales.
Research, Evaluation, and Assessment Chapter Nine.
Chapter 2 Norms and Reliability. The essential objective of test standardization is to determine the distribution of raw scores in the norm group so that.
Chapter 9 Intelligence. Objectives 9.1 The Nature of Intelligence Define intelligence from an adaptation perspective. Compare and contrast theories of.
XI. Testing and Individual Differences
Unit 2: Test Worthiness and Making Meaning out of Raw Scores
6 Selecting Employees and Placing Them in Jobs
Assessment in Counseling
Chapter 3: Legal, Ethical, and Diversity Foundations and Perspectives in Assessment ONLINE MODULE.
AP Unit 11 Testing and Individual Differences pt. 1
Reliability & Validity
پرسشنامه کارگاه.
Understanding and Using Standardized Tests
Presentation transcript:

© 2007 Thomson Brooks/Cole, a division of Thomson Learning Chapter 12: Testing and Assessment Chapter 13: Research and Evaluation 1

© 2007 Thomson Brooks/Cole, a division of Thomson Learning Testing and Assessment 2

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Testing: a subset of assessment  Assessment includes:  Informal Assessment  Personality Testing  Ability Testing  The Clinical Interview  See Figure 12.1, p

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  You will be administering and interpreting assessment instruments  You may consult with others on their proper use  You may use them in program evaluation and research  You will read about them in the professional literature  School counselors: Sometimes the only expert on assessment in the schools  Other counselors: Will likely be using them in your setting and consulting with others who use them  Why testing? Why not testing? Testing is an additional method of gaining information about your client 4

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  2200 BCE: Chinese developed essay type test for civil service employees  Darwin, set the stage for modern science and the examination of differences  Wundt, Fechner: 1 st experimental labs to examine differences in people  Binet: Hired by Ministry of Public Education in France to develop intelligence test  Binet test, later became “Stanford Binet”—revised by Terman 5

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Spread of testing at beginning of 20 th century:  Psychoanalysis spurred on development of objective and projective personality tests  Industrial Revolution and need for vocational assessment  WWI: Ability and personality tests used to determine placements of recruits  1940s and 1950s: advances in statistics led to better test construction  1980s and on: Personal computers make tests easier to develop, analyze, use, administer, and interpret 6

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Ability Testing (Testing in the Cognitive Domain) (see Figure 12.2, p. 399)  Two types ▪ Achievement Testing (What one has learned) ▪ Aptitude Testing (What one is capable of learning)  Achievement Testing ▪ Survey Battery Tests ▪ Diagnostic Tests (see Box 12.1, p. 400: PL ) ▪ Readiness Tests 7

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Ability Testing (Testing in the Cognitive Domain) (see Figure 12.2, p. 399) (Cont’d)  Aptitude Tests (What one is capable of learning) ▪ Intellectual and Cognitive Functioning Testing ▪ Intelligence Tests ▪ Neuropsychological Assessment ▪ Cognitive Ability Tests ▪ Special Aptitude Tests ▪ Multiple Aptitude Tests 8

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Personality Assessment (Testing in the Affective Domain; see Figure 12.3, p. 399)  Objective Tests  Projective Tests  Interest Inventories  Informal Assessment (see Figure 12.4, p. 399)  Observation  Rating Scales (see Box 12.2, p. 404)  Classification Systems (see Box 12.3)  Environmental Assessment  Records and Personal Documents  Performance-Based Assessment 9

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  The Clinical Interview  Sets a tone for the types of information that will be covered during the assessment process  Allows client to become desensitized to information that can be very intimate and personal  Allows examiner to assess nonverbals of client while he or she is talking about sensitive information  Allows examiner to learn problem areas firsthand  Gives client and examiner opportunity to study other’s personality style to assure they can work together 10

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Norm-referenced Tests  Your results are compared to your peer group  Criterion-referenced Tests:  Preset learning goals are established  Examinee has increased time to meet educational goals  Often used for individuals with learning disabilities  Norm-Referenced and Criterion Tests Can Be Standardized or Non- Standardized  Standardized: Given exactly the same way each time  Non-Standardized: Vary in how administered. Generally not as rigidly researched as standardized tests (e.g., teacher made tests)  See Table 12.1, p

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Relativity and Meaningfulness of Scores  Raw scores don’t hold much meaning unless you do something to them  By comparing raw scores to those of an individual’s peer group, you are able to: ▪ See how the individual did in comparison to similar people ▪ Allow test takers who took the same test, but are in different norm groups to compare their results ▪ Allow an individual to compare his or her results on two different tests 12

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Some statistics help us make meaning of test scores  Measures of Central Tendency ▪ Mean ▪ Median ▪ Mode  Measures of Variability ▪ Range ▪ Interquartile Range ▪ Standard Deviation ▪ See Figure 12.5, page 409 ▪ See Figures 12.6 and 12.7; page 410 and

© 2007 Thomson Brooks/Cole, a division of Thomson Learning TYPES OF DERIVED SCORES  Percentile Rank  T-Scores  Deviation IQ  SAT/GRE Type Scores  ACT Scores TYPES OF DERIVED SCORES  Normal Curve Equivalents (NCEs)  Stanines  Sten Scores  Grade Equivalent Scores  Idiosyncratic Publisher-Derived Scores 14

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  A basic statistic not directly related to interpretation of test but crucial in test construction  Ranges from -1.0 to +1.0  The closer to -1.0 and +1.0 the strong the relationship between variables  Positive correlation: tendency for two sets of scores to be related in same direction  Negative correlation: tendency for two sets of scores to be related in opposite direction  0 = no relationship between variables  See Figure 12.8, p

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Four Types  Validity: Is the test measuring what it’s supposed to measure?  Reliability: Is the test accurate (consistent) in its measurement?  Practicality: Is this a practical test to use?  Cross-Cultural Fairness: Has the test been shown to be fair across different cultures? 16

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Three types 1. Content 2. Criterion-Related ▪ Concurrent ▪ Predictive 3. Construct ▪ Experimental ▪ Convergent ▪ Discriminant ▪ Factor Analysis  Face validity  Not a “real” type of validity. Does the test, on the surface, seem to measure what it’s supposed to measure  Some tests may be valid, but may not seem to be measuring what it’s supposed to measure 17

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Is bias removed—as best as possible?  Does it predict well for all cultural groups?  Griggs v. Duke Power Company: Tests must show that they can predict for job performance  A number of ethical and legal issues have been addressed (see later under “Ethical, Professional, and Legal Issues”)  See Table 12.2, p.417: Summary of Types of Validity and Reliability 18

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Four Types: 1. Test-Retest 2. Alternate (Parallel; Equivalent) Forms 3. Split-Half (Odd-Even) 4. Internal Consistency ▪ Cronbach’s Coefficient Alpha ▪ Kuder-Richardson 19

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Is this a realistic test to give?  Based on:  Cost  Time to administer  Ease of administration  Format of test  Readability of test  Ease of interpretation 20

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Over 4000 assessment procedures  How do you find them:  Publisher resource catalogs  Journals  Source Books and On-Line Source “Book” Information ▪ Buros Mental Measurement Yearbook ▪ Tests in Print  Books on Testing and Assessment  Experts  The Internet 21

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Info usually included: 1. Demographic information 2. Reason for referral 3. Family background 4. Other relevant information (e.g., legal, medical, vocational) 5. Behavioral observations 6. Mental status 7. Test results 8. Diagnosis 9. Recommendations 10. Summary 22

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Usually a few pages long  Problems with:  Overuse of jargon  Focusing on assessment procedures & downplaying person  Focusing on person and downplaying assessment results  Poor organization  Poor writing skills  Failure to take a position  Demographics 23

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Caution in Using Assessment Procedures  Cultural bias continues to exist in testing  Standards and ethical codes have been developed to help us: ▪ Understand the cultural bias inherent in tests ▪ Know when a test should not be used due to bias ▪ Know what to do with test results when a test does not predict well for minorities  Standards for effective use of assessment instruments  Association for Assessment in Counseling’s Standards for Multicultural Assessment  Code of Fair Testing in Education  ACA Ethics Code 24

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Take A Stand—Do Something!  Our duty and moral responsibility to do something when ▪ Tests have been administered improperly ▪ Tests are culturally biased and the bias is not addressed ▪ Cheating has taken place ▪ Tests were used with limited validity or reliability 25

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Ethics  Guidelines for use of assessment instruments(see bottom p. 420)  Informed consent  Invasion of privacy and confidentiality  Competence in the use of Tests ▪ Levels A, B, and C  Technology and Assessment ▪ Sometimes, counselor not used with computer-generated reports ▪ Issues of confidentiality and privacy ▪ Knowing laws relative to the impact of on-line technology ▪ Adequate training in technology 26

© 2007 Thomson Brooks/Cole, a division of Thomson Learning ETHICAL ISSUES  Ethics (Other Issues)  Proper release of test results  Selecting Tests  Administering, Scoring and Interpreting Tests  Keeping Tests Secure  Picking up-to-date tests  Proper Test Construction PROFESSIONAL ISSUES  Professional Issues  Computer-Driven Assessment Reports ▪ Can be very good ▪ Make sure they reflect “you”  Professional Association ▪ Assoc. for Assessment in Counseling and Education (AACE) ▪ Adivision of ACA 27

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Legal Issues  Americans with Disabilities Act: Accommodations must be made when taking tests for employment  (FERPA) Buckley Amendment: Right to access school records, including test records  Carl Perkins Act (PL98-524): Right to vocational assessment, counseling, and placement for disadvantaged  Civil Rights Act (‘64) & Amendments: Tests must be shown to be valid for the job 28

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Legal Issues (Cont’d)  Freedom of Information Act: Right to access federal records, including test records  PL and IDEIA: Right of students to be tested, at school’s expense, for a suspected disability that interferes with learning  Section 504 of Rehabilitation Act: Instruments must measure person’s ability, not be a reflection of his or her disability  HIPAA: Right of privacy of records, including test records 29

© 2007 Thomson Brooks/Cole, a division of Thomson Learning  Assessment of clients is not just giving a test  Use multiple methods and be wise  Remember, people can and will change over time  Don’t view them as “stagnant” and always the same 30