1 An Investigation of The Response Time for Maths Items in A Computer Adaptive Test C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK Chris Wheadon.

Slides:



Advertisements
Similar presentations
Auditing Subject Knowledge in Initial Teacher Training using Online Methods.
Advertisements

Implications and Extensions of Rasch Measurement.
The MidYIS Test.
Measures of Academic Progress (MAP). What is MAP? MAP (Measures of Academic Progress) are achievement tests delivered by computer to students.
Introduction to CEM Pre-16 Computer Adaptive Assessments Glasgow 13 th February 2013
What is a CAT?. Introduction COMPUTER ADAPTIVE TEST + performance task.
DIF Analysis Galina Larina of March, 2012 University of Ostrava.
Introduction to CEM Primary Dr Chris Jellis Research Associate CEM.
An Introduction to CEM Secondary Monitoring Systems Assessment for Excellence S1/S2 Baseline Assessment (MidYIS) & S2 Curriculum-based.
1 Scaling of the Cognitive Data and Use of Student Performance Estimates Guide to the PISA Data Analysis ManualPISA Data Analysis Manual.
Yr 12 Parents Forum: ALIS data West Island School November 2012.
Interactive Computerised Assessment System
The challenges of equating tests between Russia and the UK Higher School of Economics, Moscow, Alina Ivanova, Elena Kardanova (HSE, Moscow,
Modified Achievement Tests for Students with Disabilities: Basic Psychometrics and Group Analyses Ryan J. Kettler Vanderbilt University CCSSO’s National.
PSAE Practice Session Science Mr. Johns Room 2012.
STAR Basics.
Scaled Scoring.
© AJC /18 Extended Matching Sets Questions for Numeracy Assessments: A Case Study Alan J. Cann Department of Microbiology & Immunology University.
The results from international assessments of adult literacy and numeracy skills Juliette Mendelovits CEET 17th Annual National Conference Friday 1 November.
I want to test a wound treatment or educational program but I have no funding or resources, How do I do it? Implementing & evaluating wound research conducted.
Chris Castillo Comer Director of Science Texas Education Agency The Texas Science Diagnostic System TSDS.
1 The Development of A Fraction Testing and Tutoring System Q. He & P. Brna Qingping He (CEM Centre, University of Durham, UK) & Paul Brna (SCRE Centre,
Assessment Tomorrow Conference Edinburgh, 22nd November 2012
Evaluating Creative Partnerships Presentation to the Cultural Competencies Conference, Warsaw, July 2011 Caroline Sharp, National Foundation for Educational.
Modern Test Theory Item Response Theory (IRT). Limitations of classical test theory An examinee’s ability is defined in terms of a particular test The.
Introduction to CEM and Computer Adaptive Assessments
An innovative learning model for computation in first year mathematics Birgit Loch Department of Mathematics and Computing, USQ Elliot Tonkes CS Energy,
SATs What are SATs? SATs are a series of assessments carried out at the end of a Key Stage to assess progress within the National Curriculum The.
1 Assessment and Monitoring
Parents’ basic skills and children’s test scores Augustin De Coulon, Elena Meschi and Anna Vignoles.
Yr 7 Parents Forum: Introduction to MidYIS West Island School October 2013.
1 The Development of A Computer Assisted Design, Analysis and Testing System for Analysing Students’ Performance Q. He & P. Tymms, CEM CENTRE, UNIVERSITY.
EAssessment Colin Milligan Heriot-Watt University.
Yr 12 Parents Forum: ALIS data West Island School October 2013 Mike Williams (Data Development Leader)
ASSESMENT IN OPEN AND DISTANCE LEARNING Objectives: 1.To explain the characteristic of assessment in ODL 2.To identify the problem and solution of assessment.
1 Assessment and Monitoring in the Primary Years
Multiple Perspectives on CAT for K-12 Assessments: Possibilities and Realities Alan Nicewander Pacific Metrics 1.
Pearson Copyright 2010 Some Perspectives on CAT for K-12 Assessments Denny Way, Ph.D. Presented at the 2010 National Conference on Student Assessment June.
The role of CAA in Helping Engineering undergraduates to Learn Mathematics David Green, David Pidcock and Aruna Palipana HELM Project Mathematics Education.
University of Ostrava Czech republic 26-31, March, 2012.
HESI RN CAT Exam.
Estimation. The Model Probability The Model for N Items — 1 The vector probability takes this form if we assume independence.
Understanding the 2015 Smarter Balanced Assessment Results Assessment Services.
Summary of Bayesian Estimation in the Rasch Model H. Swaminathan and J. Gifford Journal of Educational Statistics (1982)
Stages of Test Development By Lily Novita
Smarter Balanced Assessment Consortium (SBAC) Fairfield Public Schools Elementary Presentation.
 The introduction of the new assessment framework in line with the new curriculum now that levels have gone.  Help parents understand how their children.
The challenges of equating tests between Russia and Scotland Higher School of Economics, Moscow, Alina Ivanova, Elena Kardanova, Irina.
Making the most of Assessment Data in the Secondary Years Dr Robert Clark.
Key Stage 3 Student Assessment English Language. 2 Student Assessment Introduction  Relationship to English Curriculum  Criteria for Assessment Tasks.
11 PIRLS The Trinidad and Tobago Experience Regional Policy Dialogue on Education 2-3 December 2008 Harrilal Seecharan Ministry of Education Trinidad.
Using Simulation to evaluate Rasch Models John Little CEM, Durham University
The influence of forgetting rate on complex span and academic performance Debbora Hall, Chris Jarrold, John Towse and Amy Zarandi.
5. Evaluation of measuring tools: reliability Psychometrics. 2011/12. Group A (English)
The Power of e-Assessment Robert Coe Qualifications Wales, Cardiff, 14 March 2016.
Daniel Muijs Saad Chahine
Assessment Research Centre Online Testing System (ARCOTS)
Classical Test Theory Margaret Wu.
Booklet Design and Equating
Introduction to Assessment and Monitoring
Survey What? It's a way of asking group or community members what they see as the most important needs of that group or community is. The results of the.
SATs for this academic year take place from May 14th-17th
Groups within Society Chapter 3 Section 4.
Aligned to Common Core State Standards
SATs 2018 Red Hill Primary School 15th March 2018
Mohamed Dirir, Norma Sinclair, and Erin Strauts
Investigating item difficulty change by item positions under the Rasch model Luc Le & Van Nguyen 17th International meeting of the Psychometric Society,
The MidYIS Test.
Target Setting and Monitoring
Presentation transcript:

1 An Investigation of The Response Time for Maths Items in A Computer Adaptive Test C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK Chris Wheadon and Qingping He (CEM Centre, Durham University, UK)

2 Background The CEM Centre’s Baseline Assessments on Primary and Secondary School Students The CEM Baseline Tests for Secondary Schools (for students aged 11-18: three separate projects) Paper based, Efficiency, Cost Use of CTT and IRT in Computerised Testing and to Interpret Test Data To Develop A Computerised Test to Replace the Existing Paper-based Baseline Tests for the Three Secondary Projects C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK

3 CABT comprises a calibrated item bank and a question display system The adaptive test contains an adaptive maths test and an adaptive English vocabulary test Testing is delivered through the Web or from the school’s local area network CABT-The CEM Computer Adaptive Baseline Test C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK

4 Establishing A Calibrated Item Bank C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK Testing of items by school students from different year groups through the administration of a series of tests containing common items Calibration of items in each test using the Rasch model (items meeting the Rasch requirements are used) Setting of the reference test and equating of different tests using the Rasch model and common items Further new item calibration through the embedding of new items in the adaptive test

5 Developing The Question Display System - Conducting Adaptive Testing C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK Realisation of adaptivity is achieved through the implementation in the question display system of the Rasch model for ability estimation (MLE method is used) and question selection Rules for stopping test: minimum number of question; maximum number of questions; and convergence value Variable starting difficulty for the first question for different year groups.

6 C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK Preliminary Results from the Adaptive Maths Test There are over 500 items in the maths item bank Effort has been made to make the items as curriculum- free as possible Items are content-independent to each other Item types include MCQ, short free text entry questions and interactive questions Items cover a wide range of difficulties in order for the three secondary projects (for three different year groups) to use the same adaptive test as their baseline test

7 C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK Individual Student: The Distribution of Ability and the Distribution of Item Response Time

8 C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK Individual Student: The Distribution of Ability and the Distribution of Item Response Time

9 C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK All Items: The Effect of Item Difficulty on Response Time

10 C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK A Selection of Items: The Effect of Age of Test Takers on Item Response Time

11 C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK A Selection of Items: The Relationship between Response Time for Correct Answers and Response Time for Incorrect Answers

12 C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK Individual Item: The Effect of Ability and Age on Response Time (Difficulty: -0.6 logits) – Easy Item

13 C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK

14 C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK Ability Bands within a Year Group for the Specific Item (Year 7)

15 C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK Individual Item: The Effect of Ability and Age on Response Time (Difficulty: 1.1 logits) – Medium- difficulty Item

16 C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK

17 C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK Ability Bands within a Year Group for the Specific Item (Year 10)

18 C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK Individual Item: The Effect of Ability and Age on Response Time (Difficulty: 2.9 logits) – Difficult Item

19 C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK

20 C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK Ability Bands within a Year Group for the Specific Item (Year 12)

21 Conclusions and Further Work Response time for all the maths items in the test generally increases with item difficulty but shows great variability Item difficulty levels and the age and ability of test takers have significant influence on item response time The information obtained in this study can be used for constructing more efficient tests Further work will involve investigating the relationship between CABT results and students’ subsequent academic performance C. Wheadon & Q. He, CEM CENTRE, DURHAM UNIVERSITY, UK