Testing 101: Quantitative Approaches to Assessment CTE – November 2, 2005 Noelle Griffin, PhD LMU Office of Assessment at Data Analysis.

Slides:



Advertisements
Similar presentations
Test Development.
Advertisements

Standardized Scales.
Bridging Research, Information and Culture An Initiative of the Research and Planning Group for California Community Colleges Your Name Your Institution.
Validity and Reliability
Item Writing Techniques KNR 279. TYPES OF QUESTIONS Closed ended  Checking yes/no, multiple choice, etc.  Puts answers in categories  Easy to score.
Chapter 4: Conceptualization & Measurement
MEASUREMENT the process of determining the value or level of a particular CONSTRUCT for every unit of analysis or subject involves OPERATIONALIZATION –translating.
Developing General Education Course Assessment Measures Anthony R. Napoli, PhD Lanette A. Raymond, MA Office of Institutional Research & Assessment Suffolk.
SOWK 6003 Social Work Research Week 4 Research process, variables, hypothesis, and research designs By Dr. Paul Wong.
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Alignment Inclusive Assessment Seminar Brian Gong Claudia.
Cognitive and Academic Assessment
1 Classroom-Based Research: How to Be a Researcher in Your Classroom Basic Skills Initiative Teaching and Learning Workshop October 2009 Darla M. Cooper.
EDRS6208 Lecture Three Instruments and Instrumentation Data Collection.
Assessment Report Department of Psychology School of Science & Mathematics D. Abwender, Chair J. Witnauer, Assessment Coordinator Spring, 2013.
PhD Research Seminar Series: Reliability and Validity in Tests and Measures Dr. K. A. Korb University of Jos.
Annual Pupil Profile. 11 September 2003Review of Post Primary Education2 Ideas behind the Assessment Proposals Shift the emphasis to assessment for learning.
Iowa Collaborative Assessment Modules (ICAM) Heartland Area Education Agency.
OUTCOMES ASSESSMENT VIA RUBRICS: A PILOT STUDY IN AN MIS COURSE AS A PRECOURSOR TO A MULTIPLE MEASURE APPROACH By W. R. Eddins, York College of Pennsylvania.
Striving for Quality Using continuous improvement strategies to increase program quality, implementation fidelity and durability Steve Goodman Director.
Establishing MME and MEAP Cut Scores Consistent with College and Career Readiness A study conducted by the Michigan Department of Education (MDE) and ACT,
Measurement Neuman and Robson Ch. 6. What is it? The process of creating measurable concrete variables from abstract concepts Extends the senses (empirical)
ASSESSMENT OF STUDENT LEARNING Manal bait Gharim.
Understanding the TerraNova Test Testing Dates: May Kindergarten to Grade 2.
The Basics of Experimentation Ch7 – Reliability and Validity.
Validity Is the Test Appropriate, Useful, and Meaningful?
Slides to accompany Weathington, Cunningham & Pittenger (2010), Chapter 3: The Foundations of Research 1.
URBDP 591 I Lecture 3: Research Process Objectives What are the major steps in the research process? What is an operational definition of variables? What.
Reliability vs. Validity.  Reliability  the consistency of your measurement, or the degree to which an instrument measures the same way each time it.
Classroom Diagnostic Tools. Pre-Formative Assessment of Current CDT Knowledge.
Measurement Validity.
Research: Conceptualization and Measurement Conceptualization Steps in measuring a variable Operational definitions Confounding Criteria for measurement.
Research methods in clinical psychology: An introduction for students and practitioners Chris Barker, Nancy Pistrang, and Robert Elliott CHAPTER 4 Foundations.
Research: Conceptualization and Measurement Conceptualization Steps in measuring a variable Operational definitions Confounding Criteria for measurement.
SHOW US YOUR RUBRICS A FACULTY DEVELOPMENT WORKSHOP SERIES Material for this workshop comes from the Schreyer Institute for Innovation in Learning.
Validity Validity: A generic term used to define the degree to which the test measures what it claims to measure.
Evaluating Survey Items and Scales Bonnie L. Halpern-Felsher, Ph.D. Professor University of California, San Francisco.
Measurement and Scaling
W W W. C E S. C L E M S O N. E D U / G E / Planning Engineering Education Research Facilitator: Matthew W. Ohland.
Building Exams Dennis Duncan University of Georgia.
 A test is said to be valid if it measures accurately what it is supposed to measure and nothing else.  For Example; “Is photography an art or a science?
MEASUREMENT: PART 1. Overview  Background  Scales of Measurement  Reliability  Validity (next time)
Program Evaluation Principles and Applications PAS 2010.
Scales and Indices While trying to capture the complexity of a phenomenon We try to seek multiple indicators, regardless of the methodology we use: Qualitative.
Validity and Reliability in Instrumentation : Research I: Basics Dr. Leonard February 24, 2010.
RESEARCH METHODS IN INDUSTRIAL PSYCHOLOGY & ORGANIZATION Pertemuan Matakuliah: D Sosiologi dan Psikologi Industri Tahun: Sep-2009.
DENT 514: Research Methods
1 Announcement Movie topics up a couple of days –Discuss Chapter 4 on Feb. 4 th –[ch.3 is on central tendency: mean, median, mode]
Michigan Assessment Consortium Common Assessment Development Series Module 16 – Validity.
Standards-Based Tests A measure of student achievement in which a student’s score is compared to a standard of performance.
Selecting Employees Strategically: What Managers Need to Know Today.
Measurement Chapter 6. Measuring Variables Measurement Classifying units of analysis by categories to represent variable concepts.
THE RESEARCH PROCESS How do Sociologists study society? STUDYING SOCIETY TOPIC Glue the Learning Journey into your book.
Overview of Types of Measures Margaret Kasimatis, PhD VP for Academic Planning & Effectiveness.
Designing Quality Assessment and Rubrics
Instrument Development and Psychometric Evaluation: Scientific Standards May 2012 Dynamic Tools to Measure Health Outcomes from the Patient Perspective.
From Gray and Guppy – Tips for Creating Indicators:
Introduction to the Validation Phase
Concept of Test Validity
Assessment Theory and Models Part II
Measurement: Part 1.
Associated with quantitative studies
Tests and Measurements: Reliability
پرسشنامه کارگاه.
Measuring Social Life: How Many? How Much? What Type?
WP4. Development of Evaluation framework
From Gray and Guppy – Tips for Creating Indicators:
TESTING AND EVALUATION IN EDUCATION GA 3113 lecture 1
EDUC 2130 Quiz #10 W. Huitt.
First Hour - How can one measure intelligence?
From Gray and Guppy – Tips for Creating Indicators:
Presentation transcript:

Testing 101: Quantitative Approaches to Assessment CTE – November 2, 2005 Noelle Griffin, PhD LMU Office of Assessment at Data Analysis

Test Design  Generally applies to more quantified approaches to assessment  Multiple choice or short answer question  More objective (vs. subjective) approaches to scoring than qualitative/performance-based assessment

Approaches to Assessment Through Quantitative Testing  Standardized/externally developed tests  Locally designed tests

Standardized Tests  Examples: ETS content area tests, GREs, FE exam  Benefits Statistical properties established Less draw on faculty time Comparison data available  Drawbacks $$ Comparability of content Timing

Locally Designed Tests  Benefits Content linked directly to Los Local control over scope/focus Adaptable to curricular changes  Drawback Lack of outside comparison data No established reliability/validity

Steps in Test Design  Identify “constructs”  Develop items  Pilot  Scoring  Tracking/benchmarking

Identifying Constructs  What are the general areas of knowledge or skill that you will be assessing in the test?  Will form “scales” or groups of items  Example: Learning Outcome = students will be able to identify and define the primary theories in psychology; Constructs = Psychoanalytic theory, behavioral theory, cognitive theory

Role of Constructs Learning Outcome Construct Multiple Items Construct Multiple Items

Drafting Items  For each construct, what are the specific concepts/information points central to that construct?  Draft items that address each of these concepts  All items addressing a specific construct = scale

Item Drafting “Tips”  Avoid “dual pronged” items (asking two questions at once)  Avoid confounding vocabulary or jargon with the concept you want to assess (e.g., select the answer that best represents operationalization of the primacy effect)  “Multiple response” items  Including adequate detractors

Test Quality  Pilot testing (trying out the test with a smaller group of students before full-scale implementation)  Issues of reliability and validity

Reliability  Does the test measure what it purports to measure consistently?  Most applicable: internal consistency, or how well the items “hold together”  Emergence of “scales”  Looking for “outliers”  Assessment office as resource

Validity  Is the test assessing what you say it is assessing?  Conceptually: Face, construct  Empirically (if possible): Criterion  Issue of student intent: Do students actually address the test with serious effort?  Ideas for making test meaningful

Approaches to Scoring/Analysis  Overall vs. scale scores  Percentage correct  Average score  Criterion-based/% meeting standard

Comparison Considerations  Cross-sectional (comparing two groups at same time)  Cohort/time series (comparing different cohorts across time) Consideration of cohort effects  Longitudinal (looking at same group over time) Importance of matching pre/post

Additional Resources  Test Development (2001). P.W. Miller & H.E. Erickson. Miller & Associates.  Introduction to Test Construction in the Social and Behavioral Sciences (2003). J.A. Fishman & T. Galguera. Rowman and Littlefield Publishers.