AERA 2010 Robert L. Linn Lecture Slide 1 May 1, 2010 Integrating Measurement and Sociocognitive Perspectives in Educational Assessment Robert J. Mislevy.

Slides:



Advertisements
Similar presentations
National Accessible Reading Assessment Projects Defining Reading Proficiency for Accessible Large Scale Assessments Principles and Issues Paper American.
Advertisements

Traits of an Effective Reader. What does a good reader do when they read? When good readers read, they read critically, they read deeply, they apply information.
Elliott / October Understanding the Construct to be Assessed Stephen N. Elliott, PhD Learning Science Institute & Dept. of Special Education Vanderbilt.
1 Content-based Interpretations of Test Scores Michael Kane National Conference of Bar Examiners Maryland Assessment Research Center for Education Success.
Theories of Second language Acquisition
Part II Knowing How to Assess Chapter 5 Minimizing Error p115 Review of Appl 644 – Measurement Theory – Reliability – Validity Assessment is broader term.
Show Me an Evidential Approach to Assessment Design Michael Rosenfeld F. Jay Breyer David M. Williamson Barbara Showers.
Robert J. Mislevy & Min Liu University of Maryland Geneva Haertel SRI International Robert J. Mislevy & Min Liu University of Maryland Geneva Haertel SRI.
SRI Technology Evaluation WorkshopSlide 1RJM 2/23/00 Leverage Points for Improving Educational Assessment Robert J. Mislevy, Linda S. Steinberg, and Russell.
Beginning the Research Design
MSc Applied Psychology PYM403 Research Methods Validity and Reliability in Research.
University of Maryland Slide 1 May 2, 2001 ECD as KR * Robert J. Mislevy, University of Maryland Roy Levy, University of Maryland Eric G. Hansen, Educational.
University of Maryland Slide 1 July 6, 2005 Presented at Invited Symposium K3, “Assessment Engineering: An Emerging Discipline” at the annual meeting of.
Ethnography. In ethnography, the researcher  Participates in people's daily lives for an extended period of time  Watches everyday happenings  Listens.
SLRF 2010 Slide 1 Oct 16, 2010 What is the construct in task-based language assessment? Robert J. Mislevy Professor, Measurement, Statistics and Evaluation.
1 Some Key Points for Test Evaluators and Developers Scott Marion Center for Assessment Eighth Annual MARCES Conference University of Maryland October.
Consistency/Reliability
CILVR 2006 Slide 1 May 18, 2006 A Bayesian Perspective on Structured Mixtures of IRT Models Robert Mislevy, Roy Levy, Marc Kroopnick, and Daisy Wise University.
LTRC 2007 Messick Address Slide 1 June 9, 2007 Toward a Test Theory for the Interactionalist Era Robert J. Mislevy University of Maryland Samuel J. Messick.
Inference & Culture Slide 1 April 29, 2003 Argument Substance and Argument Structure in Educational Assessment Robert J. Mislevy Department of Measurement,
(1) If Language is a Complex Adaptive System, What is Language Assessment? Presented at “Language as a Complex Adaptive System”, an invited conference.
Maryland Validity ConferenceSlide 1October 10, 2008 Validity from the Perspective of Model-Based Reasoning Robert J. Mislevy Measurement, Statistics and.
FERA 2001 Slide 1 November 6, 2001 Making Sense of Data from Complex Assessments Robert J. Mislevy University of Maryland Linda S. Steinberg & Russell.
Historical Thinking Skills
Chapter 9 Flashcards. measurement method that uses uniform procedures to collect, score, interpret, and report numerical results; usually has norms and.
Case Study Research By Kenneth Medley.
Classroom Assessment A Practical Guide for Educators by Craig A
Validity Lecture Overview Overview of the concept Different types of validity Threats to validity and strategies for handling them Examples of validity.
Science and Engineering Practices
Noynay, Kelvin G. BSED-ENGLISH Educational Technology 1.
ADL Slide 1 December 15, 2009 Evidence-Centered Design and Cisco’s Packet Tracer Simulation-Based Assessment Robert J. Mislevy Professor, Measurement &
© 2013 Cengage Learning. Outline  Types of Cross-Cultural Research  Method validation studies  Indigenous cultural studies  Cross-cultural comparisons.
Terry Vendlinski Geneva Haertel SRI International
Some Implications of Expertise Research for Educational Assessment Robert J. Mislevy University of Maryland National Center for Research on Evaluation,
TEA Science Workshop #3 October 1, 2012 Kim Lott Utah State University.
Experimental Research Methods in Language Learning Chapter 2 Experimental Research Basics.
Validating Assessment Centers Kevin R. Murphy Department of Psychology Pennsylvania State University, USA.
ELA Common Core Shifts. Shift 1 Balancing Informational & Literary Text.
Learning Progressions: Some Thoughts About What we do With and About Them Jim Pellegrino University of Illinois at Chicago.
1 Diagnostic Measurement and Reporting on Concept Inventories Lou DiBello and Jim Pellegrino DRK-12 PI Meeting Washington, DC December 3, 2010.
ATTRIBUTEDESCRIPTION Focal Knowledge, Skills, Abilities The primary knowledge / skills / abilities (KSAs) targeted by this design pattern. RationaleHow/why.
APUSH ‘themes’ (B.A.G.P.I.P.E.)
Historical Thinking Skills
Inductive Generalizations Induction is the basis for our commonsense beliefs about the world. In the most general sense, inductive reasoning, is that in.
Developing Assessments for and of Deeper Learning [Day 2b-afternoon session] Santa Clara County Office of Education June 25, 2014 Karin K. Hess, Ed.D.
URBDP 591 I Lecture 3: Research Process Objectives What are the major steps in the research process? What is an operational definition of variables? What.
Introduction to Validity
Historical Thinking Skills
Qualitative Research January 19, Selecting A Topic Trying to be original while balancing need to be realistic—so you can master a reasonable amount.
Qualitative Research Topic : Analysis is Ongoing (P.272~290) MA1C0109 Owen 楊勝雄.
VALUE/Multi-State Collaborative (MSC) to Advance Learning Outcomes Assessment Pilot Year Study Findings and Summary These slides summarize results from.
Historical Thinking Skills A.P. World History Mr. Schabo Crestwood High School All info care of College Board:
Research Methodology and Methods of Social Inquiry Nov 8, 2011 Assessing Measurement Reliability & Validity.
Robert J. Mislevy University of Maryland National Center for Research on Evaluation, Standards, and Student Testing (CRESST) NCME San Diego, CA April 15,
Spring 2015 Kyle Stephenson
7 Themes. Chronological Reasoning 1. Historical Causation: relationships among multiple historical causes and effects, distinguishing between those that.
Knowing What Students Know: The Science and Design of Educational Assessment Committee on the Foundations of Assessment Board on Testing and Assessment,
Reliability performance on language tests is also affected by factors other than communicative language ability. (1) test method facets They are systematic.
Writing A Review Sources Preliminary Primary Secondary.
VALIDITY, RELIABILITY & PRACTICALITY Prof. Rosynella Cardozo Prof. Jonathan Magdalena.
Common Core State Standards in English/Language Arts What science teachers need to know.
Argumentative Writing Grades College and Career Readiness Standards for Writing Text Types and Purposes arguments 1.Write arguments to support a.
Thinking Critically in Psychology Introduction to Psychology Simon Fraser University.
Discourse Analysis Week 10 Riggenbach (1999) Chapter 1 - Quotes.
Discourse analysis May 2012 Carina Jahani
New ELA Guidelines Shifts in ELA Common Core  Rise in Nonfiction Texts.  Content Area Literacy Close and careful reading of text  Increase Complexity.
Dr Anie Attan 26 April 2017 Language Academy UTMJB
EDU 385 Session 8 Writing Selection items
Chapter Six Training Evaluation.
Historical Thinking Skills
Presentation transcript:

AERA 2010 Robert L. Linn Lecture Slide 1 May 1, 2010 Integrating Measurement and Sociocognitive Perspectives in Educational Assessment Robert J. Mislevy University of Maryland Robert L. Linn Distinguished Address Sponsored by AERA Division D. Presented at the Annual Meeting of the American Educational Research Association, Denver, CO, May 1, This work was supported by a grant from the Spencer Foundation.

AERA 2010 Robert L. Linn Lecture Slide 2 May 1, 2010 Messick, 1994 [W]hat complex of knowledge, skills, or other attribute should be assessed... Next, what behaviors or performances should reveal those constructs, and what tasks or situations should elicit those behaviors?

AERA 2010 Robert L. Linn Lecture Slide 3 May 1, 2010 Snow & Lohman, 1989 Summary test scores, and factors based on them, have often been though of as “signs” indicating the presence of underlying, latent traits. … An alternative interpretation of test scores as samples of cognitive processes and contents … is equally justifiable and could be theoretically more useful.

AERA 2010 Robert L. Linn Lecture Slide 4 May 1, 2010 Roadmap Rationale Model-based reasoning A sociocognitive perspective Assessment arguments Measurement models & concepts Why are these issues important? Conclusion

AERA 2010 Robert L. Linn Lecture Slide 5 May 1, 2010 Measurement frame Sociocognitive frame Rationale

AERA 2010 Robert L. Linn Lecture Slide 6 May 1, 2010 Rationale An articulated way to think about assessment: Understand task & use situations in “emic” sociocognitive terms. Identify the shift in to “etic” terms in task-level assessment arguments. Examine the synthesis of evidence across tasks in terms of model-based reasoning. Reconceive measurement concepts. Draw implications for assessment practice.

AERA 2010 Robert L. Linn Lecture Slide 7 May 1, 2010 Model-Based Reasoning

Entities and relationships Representational Form B Representational Form A y=ax+b (y-b)/a=x Mappings among representational systems Real-World Situation Reconceived Real-World Situation Measurement concepts Measurement models

Representational Form B Representational Form A y=ax+b (y-b)/a=x Entities and relationships in lower-level model Reconceived Entities and relationships in higher-level model Mappings among representational systems Real-World SituationReconceived Real-World Situation Measurement concepts Measurement models Sociocognitive concepts

AERA 2010 Robert L. Linn Lecture Slide 10 May 1, 2010 A Sociocognitive Perspective

AERA 2010 Robert L. Linn Lecture Slide 11 May 1, 2010 Some Foundations Themes from, e.g., cog psych, linguistics, neuroscience, anthropology: »Connectionist metaphor, associative memory, complex systems (variation, stability, attractors) Situated cognition & information processing »E.g., Kintsch’s Construction-Integration (CI) theory of comprehension; diSessa’s “knowledge in pieces” Intrapersonal & Extrapersonal patterns

AERA 2010 Robert L. Linn Lecture Slide 12 May 1, 2010 Some Foundations Extrapersonal patterns: »Linguistic: Grammar, conventions, constructions »Cultural models: What ‘being sick’ means, restaurant script, apology situations »Substantive: F=MA, genres, plumbing, etc. Intrapersonal resources: »Connectionist metaphor for learning »Patterns from experience at many levels

AERA 2010 Robert L. Linn Lecture Slide 13 May 1, 2010 B Inside B Inside AA observablenot observable

AERA 2010 Robert L. Linn Lecture Slide 14 May 1, 2010 B Inside B Inside A Context A A la Kintsch: Propositional content of text / speech… and internal and external aspects of context …

AERA 2010 Robert L. Linn Lecture Slide 15 May 1, 2010 B Inside B Inside A Context A The C in CI theory is Construction: Activation of both relevant and irrelevant bits from LTM, past experience. All L/C/S levels involved. Example: Chemistry problems in German. The C in CI theory is Construction: Activation of both relevant and irrelevant bits from LTM, past experience. All L/C/S levels involved. Example: Chemistry problems in German. If a pattern hasn’t been developed in past experience, it can’t be activated (although it may get constructed in the interaction). A relevant pattern from LTM may be activated in some contexts but not others (e.g., physics models). If a pattern hasn’t been developed in past experience, it can’t be activated (although it may get constructed in the interaction). A relevant pattern from LTM may be activated in some contexts but not others (e.g., physics models).

AERA 2010 Robert L. Linn Lecture Slide 16 May 1, 2010 B Inside B Inside A Context A The I in CI theory, Integration: Situation model: synthesis of coherent / reinforced activated L/C/S patterns The I in CI theory, Integration: Situation model: synthesis of coherent / reinforced activated L/C/S patterns

AERA 2010 Robert L. Linn Lecture Slide 17 May 1, 2010 B Inside B Inside A Context A Situation model is also the basis of planning and action.

AERA 2010 Robert L. Linn Lecture Slide 18 May 1, 2010 B Inside B Inside A Context A

AERA 2010 Robert L. Linn Lecture Slide 19 May 1, 2010 B Inside B Inside A Context A Ideally, activation of relevant and compatible intrapersonal patterns…

AERA 2010 Robert L. Linn Lecture Slide 20 May 1, 2010 B Inside B Inside A Context A to lead to (sufficiently) shared understanding; i.e., co-constructed meaning. Persons’ capabilities, situations, and performances are intertwined – Meaning co-determined, through L/C/S patterns Persons’ capabilities, situations, and performances are intertwined – Meaning co-determined, through L/C/S patterns

AERA 2010 Robert L. Linn Lecture Slide 21 May 1, 2010 What can we say about individuals? Use of resources in appropriate contexts in appropriate ways; i.e., Attunement to targeted L/C/S patterns: Recognize markers of externally-viewed patterns? Construct internal meanings in their light? Act in ways appropriate to targeted L/C/S patterns? What is the range and circumstances of activation? (variation of performance across contexts)

AERA 2010 Robert L. Linn Lecture Slide 22 May 1, 2010 Assessment Arguments

AERA 2010 Robert L. Linn Lecture Slide 23 May 1, 2010 Messick, 1994 [W]hat complex of knowledge, skills, or other attribute should be assessed... Next, what behaviors or performances should reveal those constructs, and what tasks or situations should elicit those behaviors?

AERA 2010 Robert L. Linn Lecture Slide 24 May 1, 2010 Toulmin’s Argument Claim Backing unless since Warrant Alternative explanation so Data Structure

Student acting in assessment situation Alternative explanations unless on account of Backing concerning assessment situation Warrant concerning assessment since Warrant concerning task design since Other information concerning student vis a vis assessment situation so Claim about student Data concerning student performance Data concerning task situation Warrant concerning evaluation since Concerns features of (possibly evolving) context as seen from the view of the assessor – in particular, those seen as relevant to targets of inference. Evaluation of performance seeks evidence of attunement to features of targeted L/C/S patterns. Note the move from the emic to the etic! Choice in light of assessment purpose and conception of capabilities. Note the move from the emic to the etic! Choice in light of assessment purpose and conception of capabilities. Depends on contextual features implicitly, since evaluated in light of targeted L/C/S patterns.

Student acting in assessment situation on account of Backing concerning assessment situation Alternative explanations unless Warrant concerning assessment since Warrant concerning evaluation since Warrant concerning task design since Other information concerning student vis a vis assessment situation so Claim about student Data concerning student performance Data concerning task situation “Hidden” aspects of context—not in test theory model but essential to argument: What attunements to linguistic cultural / substantive patterns can be presumed or arranged for among examinees, to condition inference re targeted l/c/s patterns? “Hidden” aspects of context—not in test theory model but essential to argument: What attunements to linguistic cultural / substantive patterns can be presumed or arranged for among examinees, to condition inference re targeted l/c/s patterns? Fundamental to situated meaning of student variables in measurement models; Both critical and implicit. Fundamental to situated meaning of student variables in measurement models; Both critical and implicit.

Student acting in assessment situation on account of Backing concerning assessment situation Alternative explanations unless Warrant concerning assessment since Warrant concerning evaluation since Warrant concerning task design since Other information concerning student vis a vis assessment situation so Claim about student Data concerning student performance Data concerning task situation Macro features of performance Micro features of performance Unfolding situated performance Micro features of situation as it evolves Macro features of situation Time Features of context arise over time as student acts / interacts. Features of performance evaluated in light of emerging context. Especially important in simulation, game, and extended performance contexts (e.g., Shute)

Student acting in assessment situation on account of Backing concerning assessment situation Alternative explanations unless Warrant concerning assessment since Warrant concerning evaluation since Warrant concerning task design since Other information concerning student vis a vis assessment situation so Claim about student Data concerning student performance Data concerning task situation Design Argument

Claim about student in use situation Other information concerning student vis a vis use situation Warrant concerning use situation since on account of Alternative explanations unless Design Argument Use Argument Data concerning use situation Student acting in assessment situation on account of Backing concerning assessment situation Alternative explanations unless Warrant concerning assessment since Warrant concerning evaluation since Warrant concerning task design since Other information concerning student vis a vis assessment situation so Claim about student Data concerning student performance Data concerning task situation Backing concerning use situation (Bachman)

Claim about student in use situation Other information concerning student vis a vis use situation Warrant concerning use situation since on account of Alternative explanations unless Design Argument Use Argument Data concerning use situation Student acting in assessment situation on account of Backing concerning assessment situation Alternative explanations unless Warrant concerning assessment since Warrant concerning evaluation since Warrant concerning task design since Other information concerning student vis a vis assessment situation so Claim about student Data concerning student performance Data concerning task situation Backing concerning use situation (Bachman) Claim about student is output of the assessment argument, input to the use argument. How it is cast depends on psychological perspective and intended use. When measurement models are used, the claim is an etic synthesis of evidence, expressed as values of student-model variable(s).

Claim about student in use situation Other information concerning student vis a vis use situation Warrant concerning use situation since on account of Alternative explanations unless Design Argument Use Argument Data concerning use situation Student acting in assessment situation on account of Backing concerning assessment situation Alternative explanations unless Warrant concerning assessment since Warrant concerning evaluation since Warrant concerning task design since Other information concerning student vis a vis assessment situation so Claim about student Data concerning student performance Data concerning task situation Backing concerning use situation

Claim about student in use situation Other information concerning student vis a vis use situation Warrant concerning use situation since on account of Alternative explanations unless Design Argument Use Argument Data concerning use situation Student acting in assessment situation on account of Backing concerning assessment situation Alternative explanations unless Warrant concerning assessment since Warrant concerning evaluation since Warrant concerning task design since Other information concerning student vis a vis assessment situation so Claim about student Data concerning student performance Data concerning task situation Backing concerning use situation

Claim about student in use situation Other information concerning student vis a vis use situation Warrant concerning use situation since on account of Alternative explanations unless Design Argument Use Argument Data concerning use situation Student acting in assessment situation on account of Backing concerning assessment situation Alternative explanations unless Warrant concerning assessment since Warrant concerning evaluation since Warrant concerning task design since Other information concerning student vis a vis assessment situation so Claim about student Data concerning student performance Data concerning task situation Backing concerning use situation

Claim about student in use situation Other information concerning student vis a vis use situation Warrant concerning use situation since on account of Alternative explanations unless Design Argument Use Argument Data concerning use situation Student acting in assessment situation on account of Backing concerning assessment situation Alternative explanations unless Warrant concerning assessment since Warrant concerning evaluation since Warrant concerning task design since Other information concerning student vis a vis assessment situation so Claim about student Data concerning student performance Data concerning task situation Backing concerning use situation Warrant for inference: Increased likelihood of activation in use situation if was activated in task situations. What features do tasks and use situations share? Implicit in trait arguments Explicit in sociocognitive arguments What features do tasks and use situations share? Implicit in trait arguments Explicit in sociocognitive arguments Empirical question: Degrees of stability, ranges and conditions of variability (Chalhoub-Deville)

Claim about student in use situation Other information concerning student vis a vis use situation Warrant concerning use situation since on account of Alternative explanations unless Design Argument Use Argument Data concerning use situation Student acting in assessment situation on account of Backing concerning assessment situation Alternative explanations unless Warrant concerning assessment since Warrant concerning evaluation since Warrant concerning task design since Other information concerning student vis a vis assessment situation so Claim about student Data concerning student performance Data concerning task situation Backing concerning use situation What features do tasks and use situations not have in common? Use situation features call for other L/C/S patterns that weren’t in task and may or may not be in examinee’s resources. Target patterns activated in task but not use context. Target patterns activated in use but not task context. Issues of validity & generalizability e.g., “method factors” Use situation features call for other L/C/S patterns that weren’t in task and may or may not be in examinee’s resources. Target patterns activated in task but not use context. Target patterns activated in use but not task context. Issues of validity & generalizability e.g., “method factors” Knowing about relation of target examinees and use situations strengthen inferences “bias for the best” (Swain, 1985) Knowing about relation of target examinees and use situations strengthen inferences “bias for the best” (Swain, 1985)

AERA 2010 Robert L. Linn Lecture Slide 36 May 1, 2010 Multiple Tasks Claim about student … D p1 OI 1 A1A1 D s1 D p1 OI 2 A2A2 D s2 D p2 D p1 OI n AnAn D sn D pn Synthesize evidence from multiple tasks, in terms of proficiency variables in a measurement model Snow & Lohman’s sampling What accumulates? L/C/S patterns, but variation What is similar from analyst’s perspective need not be from examinee’s.

AERA 2010 Robert L. Linn Lecture Slide 37 May 1, 2010 AS IF Tendencies for certain kinds of performance in certain kinds of situations expressed as student model variables . Probability models for individual performances (X) modeled as probabilistic functions of  – variability. Probability models permit sophisticated reasoning about evidentiary relationships in complex and subtle situations, BUT they are models, with all the limitations implied! Measurement Models & Concepts

AERA 2010 Robert L. Linn Lecture Slide 38 May 1, 2010 Xs result from particular persons calling upon resources in particular contexts (or not, or how) Mechanically  s simply accumulate info across situations Our choosing situations and what to observe drives their situated meaning. Situated meaning of  s are tendencies toward these actions in these situations that call for certain interactional resources, via L/C/S patterns. Measurement Models & Concepts

AERA 2010 Robert L. Linn Lecture Slide 39 May 1, 2010 Classical Test Theory Probability model: “true score” = stability along implied dimension, “error” = variation Situated meaning from task features & evaluation Can organize around traits, task features, or both, depending on task sets and performance features. Profile differences unaddressed Claim about student … D p1 OI 1 A1A1 D s1 D p1 OI 2 A2A2 D s2 D p2 D p1 OI n AnAn D sn D pn  X

AERA 2010 Robert L. Linn Lecture Slide 40 May 1, 2010 Item Response Theory  = propensity to act in targeted way, b j =typical evocation, IRT function = typical variation Situated meaning from task features & evaluation Task features still implicit Profile differences / misfit highlights where the narrative doesn’t fit – for sociocognitive reasons Claim about student … D p1 OI 1 A1A1 D s1 D p1 OI 2 A2A2 D s2 D p2 D p1 OI n AnAn D sn D pn  X1X1 X2X2 XnXn … Complex systems concepts: Attractors & stability  regularities in response patterns, quantified in parameters; Typical variation  prob model Complex systems concepts: Attractors & stability  regularities in response patterns, quantified in parameters; Typical variation  prob model Will work best when most nontargeted L/C/S patterns are familiar… Item-parameter invariance vs Population dependence (Tatsuoka, Linn, Tatsuoka, & Yamamoto, 1988) Will work best when most nontargeted L/C/S patterns are familiar… Item-parameter invariance vs Population dependence (Tatsuoka, Linn, Tatsuoka, & Yamamoto, 1988)

AERA 2010 Robert L. Linn Lecture Slide 41 May 1, 2010 Multivariate Item Response Theory (MIRT)  s = propensities to act in targeted ways in situations with different mixes of L/C/S demands. Good for controlled mixes of situations

AERA 2010 Robert L. Linn Lecture Slide 42 May 1, 2010 Structured Item Response Theory Explicitly model task situations in terms of L/C/S demands. Links TD with sociocognitive view. Work explicitly with features in controlled and evolved situations (design / agents) Can use with MIRT; Cognitive diagnosis models Claim about student … D p1 OI 1 A1A1 D s1 D p1 OI 2 A2A2 D s2 D p2 D p1 OI n AnAn D sn D pn  X1X1 X2X2 XnXn … q1q1 v i1 q2q2 v i2 qnqn v in

AERA 2010 Robert L. Linn Lecture Slide 43 May 1, 2010 Mixtures of IRT Models Different IRT models for different unobserved groups of people Modeling different attractor states Can be theory driven or discovered in data Claim about student … D p1 OI 1 A1A1 D s1 D p1 OI 2 A2A2 D s2 D p2 D p1 OI n AnAn D sn D pn  X1X1 X2X2 XnXn … Claim about student … D p1 OI 1 A1A1 D s1 D p1 OI 2 A2A2 D s2 D p2 D p1 OI n AnAn D sn D pn  X1X1 X2X2 XnXn … OR

AERA 2010 Robert L. Linn Lecture Slide 44 May 1, 2010 Measurement Concepts Validity »Soundness of model for local inferences »Breadth of scope is an empirical question »Construct representation in L/C/S terms »Construct irrelevant sources of variation in L/C/S terms Reliability »Through model, strength of evidence for inferences about tendencies, given variabilities … or about characterizations of variability.

AERA 2010 Robert L. Linn Lecture Slide 45 May 1, 2010 Measurement Concepts Method Effects »What accumulates in terms of L/C/S patterns in assessment situations but not use situations Generalizability Theory (Cronbach) »Watershed in emphasizing evidentiary reasoning rather than simply measurement »Focus on external features of context; can be recast in L/C/S terms, & attend to correlates of variability

AERA 2010 Robert L. Linn Lecture Slide 46 May 1, 2010 Why are these issues important? Connect assessment/measurement with current psychological research »Connect assessment with learning Appropriate constraints on interpreting large scale assessments Inference in complex assessments »Games, simulations, performances »Assessment modifications & accommodations »Individualized yet comparable assessments

AERA 2010 Robert L. Linn Lecture Slide 47 May 1, 2010 Measurement frame Sociocognitive frame Conclusion Communication at the interface We have work we need to do, together.