A real-time assessment of students’ mental models of sound propagation

Slides:



Advertisements
Similar presentations
Questionnaire Development
Advertisements

Ed-D 420 Inclusion of Exceptional Learners. CAT time Learner-Centered - Learner-centered techniques focus on strategies and approaches to improve learning.
1 COMM 301: Empirical Research in Communication Kwan M Lee Lect4_1.
© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.
Reliability for Teachers Kansas State Department of Education ASSESSMENT LITERACY PROJECT1 Reliability = Consistency.
Assessment: Reliability, Validity, and Absence of bias
1 ETR 520 Introduction to Educational Research Dr. M C. Smith.
Principles of High Quality Assessment
Scaling and Attitude Measurement in Travel and Hospitality Research Research Methodologies CHAPTER 11.
Assessing Student Learning
Validity Lecture Overview Overview of the concept Different types of validity Threats to validity and strategies for handling them Examples of validity.
Measurement and Data Quality
Validity and Reliability
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 14 Measurement and Data Quality.
Principles of Test Construction
The Genetics Concept Assessment: a new concept inventory for genetics Michelle K. Smith, William B. Wood, and Jennifer K. Knight Science Education Initiative.
Chapter 11: Qualitative and Mixed-Method Research Design
WELNS 670: Wellness Research Design Chapter 5: Planning Your Research Design.
The Analysis of the quality of learning achievement of the students enrolled in Introduction to Programming with Visual Basic 2010 Present By Thitima Chuangchai.
1 Chapter 6 Selection and Placement. 2 Introduction Why Selection is Important?
Reliability & Validity
For ABA Importance of Individual Subjects Enables applied behavior analysts to discover and refine effective interventions for socially significant behaviors.
Question paper 1997.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Assessing Measurement Quality in Quantitative Studies.
MEASUREMENT. MeasurementThe assignment of numbers to observed phenomena according to certain rules. Rules of CorrespondenceDefines measurement in a given.
Experimental Research Methods in Language Learning Chapter 5 Validity in Experimental Research.
McGraw-Hill/Irwin © 2012 The McGraw-Hill Companies, Inc. All rights reserved. Obtaining Valid and Reliable Classroom Evidence Chapter 4:
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 11 Measurement and Data Quality.
WHS AP Psychology Unit 7: Intelligence (Cognition) Essential Task 7-3:Explain how psychologists design tests, including standardization strategies and.
VALIDITY What is validity? What are the types of validity? How do you assess validity? How do you improve validity?
ESTABLISHING RELIABILITY AND VALIDITY OF RESEARCH TOOLS Prof. HCL Rawat Principal UCON,BFUHS Faridkot.
Copyright © 2014 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 25 Critiquing Assessments Sherrilene Classen, Craig A. Velozo.
Zdeslav Hrepic Dean A. Zollman N. Sanjay Rebello Supported by NSF ROLE Grant # REC Fort Hays State University Kansas State University AAPT, Sacramento.
Action Research for School Leaders by Dr. Paul A. Rodríguez.
Introduction to Marketing Research
Inquiry-Based Instruction
Identifying Students' Mental Models of Sound Propagation
Evaluating Student-Teachers Using Student Outcomes
Unit 8: Intelligence (Cognition)
EXPERIMENTAL RESEARCH
QUESTIONNAIRE DESIGN AND VALIDATION
Improving Student Engagement Through Audience Response Systems
Reliability and Validity in Research
Concept of Test Validity
Zdeslav Hrepic Dean A. Zollman N. Sanjay Rebello
Zdeslav Hrepic Dean A. Zollman N. Sanjay Rebello
RELIABILITY OF QUANTITATIVE & QUALITATIVE RESEARCH TOOLS
Validity and Reliability
ASSESSMENT OF STUDENT LEARNING
CHAPTER 2 Research Methods in Industrial/Organizational Psychology
Classroom Assessment Validity And Bias in Assessment.
What to include in your Portfolio؟
Human Resource Management By Dr. Debashish Sengupta
Week 3 Class Discussion.
پرسشنامه کارگاه.
Chapter Six Training Evaluation.
Chapter Three Research Design.
Prepared by: Toni Joy Thurs Atayoc, RMT
TFI Wordle This presentation is intended to introduce the PBISApps site and the types of data teams will be working with. The teams will take their first.
Research Methods.
Multi-Mode Data Collection Approach
THE RELATIONSHIP BETWEEN PRE-SERVICE TEACHERS’ PERCEPTIONS TOWARD ACTIVE LEARNING IN STATISTIC 2 COURSE AND THEIR ACADEMIC ACHIEVEMENT Vanny Septia Efendi.
Unit 7: Instructional Communication and Technology
Zdeslav Hrepic Dean A. Zollman N. Sanjay Rebello
A Multi-disciplinary Perspective on Decision-making and Creativity:
Assessment Literacy: Test Purpose and Use
TESTING AND EVALUATION IN EDUCATION GA 3113 lecture 1
Chapter 8 VALIDITY AND RELIABILITY
EDUC 2130 Quiz #10 W. Huitt.
Presentation transcript:

A real-time assessment of students’ mental models of sound propagation Zdeslav Hrepic Dissertation Defense Kansas State University Physics Education Research Group Supported by NSF ROLE Grant # REC-0087788

Outline Rationale: Why use in-class, real-time assessment? Previous research: Mental models of sound propagation. Hybrid mental models and their role. Test construction and validation Results Using the test Further study

Real time, in class assessment Uses some form of Class Response System Enables quick collection and immediate analysis of students responses in the classroom.

Benefits of class assessment Engages students. Facilitates interactive learning and peer instruction (especially in large enrolment classes). Gives immediate feedback to the teacher. Enables the teacher to adjust the teaching before the exam rather than after it and according to specific needs of his/her students. Allows a post lecture detailed analysis.

Goal of the study To create a multiple choice test… …that can elicit students’ mental models of sound propagation… …during the lecture… …using a class response system and appropriate software.

Mental model definition Mental model is: an internal (mental) representation analogous to the physical world situations or processes that it represents and that serves to explain and predict the physical world behavior (Greca & Moreira, 2002) Mental model has: spatial configuration of identifiable kinds of things (a few) principles of how system works and (certain) predictive power (diSessa, 2002) Mental model state: Is defined by student’s consistency (Pure & Mixed)

Research questions Main question: (Some of) Sub questions: What is the optimal multiple choice test that can elicit students’ mental models of sound propagation in a real time, during the instruction? (Some of) Sub questions: Is model analysis the optimal analytical tool for analysis of students’ responses in this test? How do we represent data so the display provides a variety of instruction guiding information? How reliable is the test? How valid is the test?

Starting point in test creation: Identifying mental models of sound propagation Hrepic, Z., Zollman, D., & Rebello, S. (2002). Identifying students' models of sound propagation. Paper presented at the 2002 Physics Education Research Conference, Boise ID.

4 basic models - mechanisms of propagation Human characters = Air particles Footballs = Sound entities

4 basic models - mechanisms of propagation Wave Model Scientifically Accepted Model (+) Ear Born Sound Propagating Air Hybrid Models Dependent Entity Independent Entity Dominant Alternative Model

Implications of hybrid mental models Implications for analysis of our test Hybrid models cause overlaps in multiple choice questionnaires – more than one model corresponds to the same choice (E.g.) Model analysis requires one on one match of model and answer choice Implications for teaching A student can give a variety of correct answers on standard questions using a hybrid (wrong) model (E.g.)

Constructing the test Four steps of test construction and validation: Pilot testing Pre-survey testing Survey testing Post Survey testing

Pilot testing Did we miss anything in terms of mental models? Open-ended questionnaire on a large sample Did we miss anything in terms of productive questions to determine students mental models? Battery of semi-structured conceptual questions related to sound as a wave phenomena in variety of situations

How does sound propagate in this situation? Test Contexts 1. Air How does sound propagate in this situation?

How does sound propagate in this situation? Test Contexts 2. Wall How does sound propagate in this situation?

Test Contexts 1a, 2a - Vacuum What happens without the medium (air or wall)?

Pre-survey testing 5 option multiple choice test needed Does our choice selection match students’ “needs”? Trial with: None of the above More than one of the above Validation through expert reviews Probing and refining the test through students interviews

Survey testing Surveying - to determine: Stability of results… …across different institutions at equivalent educational levels …across different course levels at same institutions Instructional sensitivity of the test Correlations between response items Model distributions at different levels - for future use Interviewing – to determine: To validate new test version To inform and make sense of survey findings

Test questions - paraphrased What is the mechanism of sound propagation in the air/wall? How do particles of the medium vibrate, if at all, while the sound propagates? How do particles of the medium travel, if at all, while the sound propagates? What does this motion have to do with sound propagation – cause and effect relationship? What does this motion have to do with sound propagation – time relationship? What happens with sound propagation in the vacuum?

Displaying the test results Several representations of students’ state of understanding Available in real time and in post instruction analysis Consistency: Consistent – a student uses one model (Pure model state) Inconsistent – a student uses more than one model (Mixed model state)

Using a particular model Pre Instruction; Calculus based; University; NY Inconsistently Consistently N = 100

Using a particular model at least once Pre Instruction; Calculus based; University; NY Inconsistently Consistently N = 100

Movements of particles of the medium Pre Instruction; Calculus based; University; NY (+) Random Travel (+) Travel Away From The source Vibration on the Spot N = 100

Model states Pre Instruction; Calculus based; University; NY Mixed Any Pure Other Mixed Entity Pure Wave Mixed Ear-Wave N = 100

Correctness Pre Instruction; Calculus based; University; NY

Survey participants

Survey Results Results stable? Differences meaningful? Comparing consistency and correctness Different levels; Pre- and post-instruction

Comparing correctness and consistency Different levels; Pre- and post-instruction

Comparing correctness and consistency Different levels; Pre- and post-instruction

Comparing model distribution Different educational levels

Comparing model distribution Grouped models; Different educational levels

Comparing model distribution Grouped models; Different educational levels

Comparing model distribution Grouped models; Different Educational Levels

Comparing model distribution Different course levels

Comparing differences in model distribution Variability within different educational levels

Pre-Post instruction difference *Gain (G) = (post-test) – (pre-test) **Normalized gain (h) = gain / (maximum possible gain) (Hake, 1997).

Using a particular model Pre Instruction; Calculus based; University; NY Inconsistently Consistently N = 100

Using a particular model Post Instruction; Calculus based; University; NY Inconsistently Consistently N = 95

Movements of particles of the medium Pre Instruction; Calculus based; University; NY (+) Random Travel (+) Travel Away From The source Vibration on the Spot N = 100

Movements of particles of the medium Post Instruction; Calculus based; University; NY (+) Random Travel (+) Travel Away From The source Vibration on the Spot N = 95

Correctness Pre Instruction; Calculus based; University; NY

Correctness Post Instruction; Calculus based; University; NY

Correlation analysis of answer choices

Validity interviews 17 x 4 probes in the interviewed sample. The invalid display of a model would have occurred in 6 instances 8.8% of the probes 3 instances because of 5a (+ another 3 that did not cause invalid probe)

Post-Survey Testing Expert review: To validate post survey version Few minor items improved Surveying: To determine correlations between response items and see if changes made the desired effect. Problems fixed Role playing validation: To validate new test version in an additional way Perfect score

Test Reliability Reliability pertains to the degree to which a test consistently measures what it is supposed to measure. (Oosterhof, 2001) Content sampling error Occasion sampling error Examiner Error Scorer Error

Reliability addressed Content sampling error Occurs because students may be more or less lucky with how test items correspond to things they know. To reduce: Test more content To measure: Need parallel form Issues: No parallel form, Context dependence Reduced by probing a single model multiple times Addressed by showing meaningful* correlations between the answer choices: Not neg. if related to same model (pos. and frequently sig.) Not sig. pos. if related to different models (Except Dependent/Independent entity models - continuum) *Pertain only to secondary and tertiary levels but not to primary

Occasion sampling error Occurs because students can be more or less lucky with respect to time when the test was administered. To reduce: Test more often To measure: Need multiple administrations of the same test Issues: Problematic for instructors and students, Economy Did not probe time stability Addressed by showing : Stable results across institutions at the same level Meaningful differences between educational levels Meaningful differences between course levels Meaningful differences between pre- and post-instruction

Examiner error & Scorer error Examiner error occurs because of the differences in examiners. Not measurable Was reduced through the standard introduction to the test (verbal and written) Scorer error occurs if students’ scores depend on who happened to mark their work. Not an issue - computerized analysis of results. All four of the treats to the reliability well addressed Gives a ground for the statement that the test is a reliable instrument.

Validity addressed Test Validity: The extent to which a test measures what it is supposed to measure and nothing else. (Oosterhof, 2001) “Validity concerns the appropriateness of inferences and actions that are based on a test’s scores”. (Hanna, 1993, p. 8) Validity is not an attribute of the test, but “of the interaction of a test with a situation in which the test is used to make decisions”. (Hanna, 1993 p. 382) Content-related evidence of validity Criterion-related evidence of validity Construct-related evidence of validity

Content-related evidence of validity Indicates how well the content of a test corresponds to the student performance that we want to observe. (Oosterhof, 2001) Addressed by: Experts’ review of the content and correctness of the answer choices Demonstrated instructional sensitivity Table of (content) specifications

Criterion-related evidence of validity Indicates how well performance on a test correlates with performance on relevant criterion measures external to the test” (Oosterhof, 2001, p.55) concurrent validation (compares the test results a parallel, substitute measure). predictive validation (compares the test results with follow up testing) Addressed by: Validation through the interviews Think aloud interview protocols Comparisons of students free answers in interview setting with their results on the test Role playing validation Correlation analysis of answer choices

Construct-related evidence of validity “Establishes a link between the underlying psychological construct we wish to measure and the visible performance we choose to observe”. (Oosterhof, 2001, p.46) Addressed by: Building the case on previous research Table of (construct) specifications

Prospective uses of test, test questions Formative assessment combined with any instructional method/approach “traditional” “progressive” “misconception oriented” Model – cause Misconception – symptom As peer instruction questions (not model defining) Not recommended as a summative assessment Online package related to test and analysis of data available at: http://web.phys.ksu.edu/role/sound/

Limitations Common to multiple choice tests Answer options do affect students understanding / models Test taking strategies may obscure results Test projects no model state as mixed model state and possibly pure model state.

Future research Unique approach - Wide themes opened Applicability of the approach in other domains of physics: Is the approach “hybrid model-(in)dependent”? Applicability in domains of other natural sciences? How effectively teachers can implement the real-time aspects of this testing approach? Instructional utility of this type of testing: Will addressing of the underlying models in real time help students learn? Possibility of individualized addressing of student’s models in real time? Applicability of the testing approach in eliciting non-cognitive psychological constructs: Personality tests: Would it provide information that current tests in that field do not? Reduction of items when compared to Likert scale

Future research Specific issues opened Optimal using of the test in combination with online homework Saving of time Any classroom benefit counterbalance? How applicable is this test at the middle school level? How would a branched version of the test look, and would it have any advantages with respect to this one? Improved simplicity and validity of the test

More Information / Feedback zhrepic@phys.ksu.edu http://www.phys.ksu.edu/~zhrepic/ Thank You!