EXPLORING THE VALIDITY OF AUSTRALIA’S NATIONAL ASSESSMENT SYSTEM

Slides:



Advertisements
Similar presentations
Site-Based Decision Making Campus Planning. Restructuring A process through which a district or school alters the pattern of its structures (vision, rules,
Advertisements

Elliott / October Understanding the Construct to be Assessed Stephen N. Elliott, PhD Learning Science Institute & Dept. of Special Education Vanderbilt.
What is moderation and why should we moderate?
Developing consistency of teacher judgment Module 2.
Program Evaluation. Lecture Overview  Program evaluation and program development  Logic of program evaluation (Program theory)  Four-Step Model  Comprehensive.
NCTM’s Focus in High School Mathematics: Reasoning and Sense Making.
New Hampshire Enhanced Assessment Initiative: Technical Documentation for Alternate Assessments Consequential Validity Inclusive Assessment Seminar Elizabeth.
National Center on Educational Outcomes N C E O What the heck does proficiency mean for students with significant cognitive disabilities? Nancy Arnold,
Using Standards and Assessment for Learning Val Klenowski Queensland University of Technology 1 August 2014.
Higher Physical Education
Dimension of Learning 4 SYSTEMS ANALYSIS NAPLAN: How does systematic assessment and reporting impact on student success?
Enhancing assessment capacity For teachers of Authority and Authority-registered subjects.
1 A proposed skills framework for all 11- to 19-year-olds.
Overall Teacher Judgements
National Council of Teachers of English Dr. Beverly Ann Chin November 21, 2014 Washington, DC.
Including Quality Assurance Within The Theory of Action Presented to: CCSSO 2012 National Conference on Student Assessment June 27, 2012.
Alaska Staff Development Network – Follow-Up Webinar Emerging Trends and issues in Teacher Evaluation: Implications for Alaska April 17, :45 – 5:15.
REVISIONS TO GENERAL EDUCATION STUDENT LEARNING OUTCOMES Auburn University Senate Information Item, August 2014.
Assessment for learning
“A Truthful Evaluation Of Yourself Gives Feedback For Growth and Success” Brenda Johnson Padgett Brenda Johnson Padgett.
Evaluating Consequential Validity of AA-AAS Presented at OSEP Conference January 15, 2008 by Marianne Perie Center for Assessment.
VALUE/Multi-State Collaborative (MSC) to Advance Learning Outcomes Assessment Pilot Year Study Findings and Summary These slides summarize results from.
TEAN Cameo What changes hearts, minds and practice concerning children and young people with SEND?
The Literacy and Numeracy Secretariat Le Secrétariat de la littératie et de la numératie October – octobre 2007 The School Effectiveness Framework A Collegial.
Curriculum K-12 Directorate. A period of public consultation, with the opportunity to provide feedback on the draft Australian Curriculum in English,
Knowing What Students Know: The Science and Design of Educational Assessment Committee on the Foundations of Assessment Board on Testing and Assessment,
Curriculum K-12 Directorate October  Why an Australian Curriculum?  The NSW context  Shape of the Australian Curriculum  Challenges and opportunities.
C R E S S T / U C L A Validity Issues for Accountability Systems Eva L. Baker AERA April 2002 UCLA Graduate School of Education & Information Studies.
Fitz-Albert R. Russell, PhD JP Establishing Standards for Students’ Assessment.
Implementing the Australian Curriculum, History Helen Hennessy Education Officer, History, BCE
Foundations of American Education: Perspectives on Education in a Changing World, 15e © 2011 Pearson Education, Inc. All rights reserved. Chapter 11 Standards,
Greenbush. An informed citizen possesses the knowledge needed to understand contemporary political, economic, and social issues. A thoughtful citizen.
Building an Interim Assessment System: A Workbook for School Districts CCSSO National Conference on Student Assessment Detroit, MI June 22, 2010.
Conducting a research project. Clarify Aims and Research Questions Conduct Literature Review Describe methodology Design Research Collect DataAnalyse.
Dr. Aidah Abu Elsoud Alkaissi An-Najah National University Employ evidence-based practice: key elements.
Educational contributions to cohesion and well-being in European social and institutional life.
Common Core.  Find your group assignment.  As a group, read over the descriptors for mastery of this standard. (The writing standards apply to more.
Jeanette Gurrola Psychology Department School of Behavioral & Organizational Sciences Claremont Graduate University American Evaluation.
Stages of Research and Development
Understanding essay titles and thesis construction
Accounting 9706.
Writing a sound proposal
Chapter 11: Quasi-Experimental and Single Case Experimental Designs
Program Evaluation ED 740 Study Team Project Program Evaluation
What is moderation and why should we moderate?
How do early career teachers integrate high-level literacy strategies in high-poverty secondary schools? Gayle Thieman, Portland State University Susan.
Who we are We work for well run evidence based public care
AIS(ACT) Numeracy Coaching
Director of Policy Analysis and Research
HEALTH IN POLICIES TRAINING
CENTER FOR COORDINATION OF RESEARCH INTERNATIONAL FEDERATION OF CATHOLIC UNIVERSITIES  (IFCU) Continuing of Education for Disadvantaged Adolescents in.
Validating Interim Assessments
A new “pre-graduation expectation” for graduating seniors
Topic Principles and Theories in Curriculum Development
Understanding Standards Modern Studies
The Mid Tudors AS Evaluation and enquiry questions
A LEVEL Paper Three– Section A
What to do with your data?
Validity and Reliability II: The Basics
Research Methodology BE-5305
Principles of Assessment
Written and Presented by: David Marlow
Performance Management and Appraisal
Zimbabwe 2008 Critical Thinking.
Effect size = .40 (Leithwood, 2011)
Inquiry and IBL pedagogies
Monitoring and Evaluating FGM/C abandonment programs
Author Department of general internal medicine Conference Date
AICE General Paper What IS this class?.
Table 1. Conceptual Framework Learning Outcomes
Presentation transcript:

EXPLORING THE VALIDITY OF AUSTRALIA’S NATIONAL ASSESSMENT SYSTEM Assoc Prof Greg Thompson (QUT) Dr Lenore Adie (ACU) Prof Val Klenowski (QUT) BERA, Leeds, 2016

Validity of NAPLAN Validity: the most fundamental consideration in developing and evaluating tests (AERA, APA & NCME, 2014, p. 11) Lack of agreement about the best way to use the term ‘validity’ (Newton & Shaw, 2015) Validity is not the property of a test, but of the use of that test Each purpose of an assessment needs to be assessed for its own validity (Stobart, 2009) Lack of systematic analysis of the validity of NAPLAN

History of validity

Definition of validity The extent to which the proposed interpretations and uses of test scores are justified. The justification requires conceptual analysis of the coherence and completeness of the claims and empirical analyses of the inferences and assumptions inherent in the claims (Kane, 2015, p.1).

Argument-based approach to validity Two-step argument-based approach: Specifying intended interpretation and use of the test as an interpretation/use argument (IUA) which includes “inferences and assumptions leading from test performances to conclusions and decisions based on the test scores” (Kane, 2015, p. 4).   Develop the argument based on the whether the interpretations and uses of the test are supported by appropriate evidence.

Establishing an argument for the valid use of NAPLAN data “actual interpretations and uses of test scores in context are far more varied and nuanced than an intended interpretation might imply” (Moss, 2015: 1) NAPLAN Aims 1. to help drive improvements in student outcomes 2. provide increased accountability for the community (ACARA, 2011). In practice, these aims have become reinterpreted and reinscribed by jurisdictions, reports and policymakers

Two Illustrative Examples In these two contexts the proposed uses and interpretations of the NAPLAN scores have a negative impact on validity. 2014 writing test “Choose a rule or a law that you think needs to change and give your reasons” The ambiguity and lack of clarity of the task compromised the interpretations and uses of the results For what is being tested is not the students’ knowledge, skills and understanding pertinent to writing “but their ability to interpret the questions” (Baird, Ahmed, Hopfenbeck, Brown, & Elliott, 2013, p. 15) and “how well candidates cope with those surprises….” Ofqual, 2008, as cited by Baird et al., 2013, p. 15) The task was changed so that in 2015 Years 3 and 5 completed a different task to Years 7 and 9.

2. The uneven nature of participation across the population, and therefore the reasonableness of comparisons like MySchool A comparison between the average scores of the states must take into account the likelihood that what is being compared is as much different rates of participation as it is different levels of student achievement.

Implications for the validity of NAPLAN The use of test scores should be matched with accountability for those decisions. Arguments about the validity of the use of test data is mediated by the context and the stakes associated with the consequences. Disentangle the accountability and learning aims of NAPLAN Create a set of arguments for use and score interpretation targeted at specific groups such as teachers, principals, media, system authorities and policymakers as a means to assist them in understanding the limits of data - what data can and cannot be used for - to improve the assessment literacy of the population.

References ACARA. (2011). Why NAP. Retrieved March 11, 2014, from ACARA: Australian Curriculum and Reporting Authority: http://www.nap.edu.au/about/why-nap.html AERA, APA & NCME (2014). Standards for educational and psychological testing. Washington DC: AERA. Australian Government. (2014). Australian Government response to the Senate Education and Employment References Committee Report: Effectiveness of the National Assessment Program – Literacy and Numeracy. Accessed 6 November, 2015 https://docs.education.gov.au/system/files/doc/other/response_to_naplan_senate_report_0.docx Coburn, C.E. & Turner, E.O. (2011). Research on data use: A framework and analysis. Measurement: Interdisciplinary Research and Perspectives, 9(4), 173-206, DOI: 10.1080/15366367.2011.626729 Department of Education, Training and Employment. (2013). Great teachers = Great results. A direct action plan for Queensland schools. Queensland Government. Accessed 25 November, 2015 http://deta.qld.gov.au/about/government-responses/pdf/great- teachers-great-results-action-plan.pdf Kane, M.T. (2015). Explicating validity. Assessment in Education: Principles, Policy & Practice, DOI: 10.1080/0969594X.2015.1060192 Madaus, G., Russell, M., & Higgins, J. (2009). The Paradoxes of High Staks Testing: How they affect students, their parents, teachers, principals, schools. and society. Charlotte, NC: Information Age Publishing. Newton, P., & Shaw, S. (2015). Disagreement over the best way to use the word 'validity' and optios for reaching consensus. Assessment in Education: Principles, Policy and Practice, DOI: 10.1080/0969594X.2015.1037241 Nichols, S., & Berliner, D. (2007). Collateral Damage: How high-stakes testing corrupts America's schools. Harvard Education Press. Stobart, G. (2009). Determining validity in national curriculum assessments. Educational Research, 51(2), 161-179. Storey, A. (2000). A leap of faith? Performance pay for teachers. Journal of Education Policy, 15:5, 509-523, DOI: 10.1080/026809300750001667 Wu, M. (2010). The inappropriate use of NAPLAN data. Professional Voice, 8(1), 21-25.