Lessons Learned about Assessing Quantitative Literacy MAA PREP Workshop: Creating and Strengthening Interdisciplinary Programs in Quantitative Literacy.

Slides:



Advertisements
Similar presentations
Mastery Learning is a style of assessment in which the student must demonstrate mastery of the assignment by correctly answering a certain number of problems.
Advertisements

What is a CAT?. Introduction COMPUTER ADAPTIVE TEST + performance task.
Reliability for Teachers Kansas State Department of Education ASSESSMENT LITERACY PROJECT1 Reliability = Consistency.
Introduction: Introduction: Children often have a hard time understanding objects used as symbols. One reason is that they must think of the object both.
Mathematics in the MYP.
Minnesota State Community and Technical College Critical Thinking Assignment Example and Assessment.
A Quantitative Literacy Assessment Rubric Development & Lessons Learned - Stuart Boersma, Central Washington Univ. - Caren Diefenderfer, Hollins University.
Merrill Area United Way Outcomes Training (Session #1) Art Lersch Associate Professor, Community Resource Development Educator University of Wisconsin.
Fall 2006 – Fundamentals of Business Statistics 1 Chapter 13 Introduction to Linear Regression and Correlation Analysis.
Linear Regression and Correlation Analysis
Chapter 13 Introduction to Linear Regression and Correlation Analysis
A Quantitative Literacy Assessment Rubric Development & Lessons Learned - Stuart Boersma, Central Washington Univ. - Caren Diefenderfer, Hollins University.
Dr. Robert Mayes University of Wyoming Science and Mathematics Teaching Center
Improving Mathematics Problem Solving Through Written Explanations Report on action research project by Janet Schlattmann.
Today Concepts underlying inferential statistics
Math Attitudinal Survey Assessment Results The Math You Need When You Need It Incorporating quantitative skills to improve student.
Info for this website Taken from: From: Compass Test Information Presentation.
SAT PREP (AND ACT BASICS). WHEN TO GUESS Correct = 1 pt Omitted = 0 pt Wrong = Negative Pts Guess when you can positively eliminate 1 answer Exception:
Evaluation of Student Learning, Department of Mathematics and Computer Science, Westmont College, Santa Barbara, CA Interpretation The.
P.E.R.T. Diagnostic Learning Pathways Math, Reading, Writing.
Mean, Median, Mode, and Range using Frequency Tables Standard: 6-SP: Develop understanding of statistical variability. Understand that a set of data collected.
-SLO Development Progress -SLO Activity -Assessment Progress -Support Needs.
SCORING. INTRODUCTION & PURPOSE Define what SCORING means for the purpose of these modules Explain how and why you should use well-designed tools, such.
Mathematics Grade 4.
Reading Your Child’s Cognitive Abilities Test Profile Bonnie O’Regan Advanced Learning Program Facilitator Greenwich Public Schools February 5, 2014 (snowdate:
Department of Mathematical Sciences School of Science and Technology B.A. in Mathematics CIP Code: Program Code: Program Quality Improvement.
Pierce College CSUN-Pierce Paths Project Outcomes Report 2013.
Testing Information Session SAGE Testing Information 1 Information for Parents and Schools.
Completion, Short-Answer, and True-False Items
River Valley Primary School – Strive for the Best WELCOME Presentation slides can be downloaded from:
BOARD ENDS POLICY REVIEW E-2 Students will demonstrate a strong foundation in academic skills by working toward the Kansas Standards of Excellence in reading,
Classroom Assessment A Practical Guide for Educators by Craig A
4/16/07 Assessment of the Core – Science Charlyne L. Walker Director of Educational Research and Evaluation, Arts and Sciences.
The Genetics Concept Assessment: a new concept inventory for genetics Michelle K. Smith, William B. Wood, and Jennifer K. Knight Science Education Initiative.
College Algebra: An Overview of Program Change Dr. Laura J. Pyzdrowski Dr. Anthony S. Pyzdrowski Dr. Melanie Butler Vennessa Walker.
THE RELATIONSHIP BETWEEN PRE-SERVICE TEACHERS’ PERCEPTIONS TOWARD ACTIVE LEARNING IN STATISTIC 2 COURSE AND THEIR ACADEMIC ACHIEVEMENT Vanny Septia Efendi.
Smarter Balanced Assessment System March 11, 2013.
Quantitative SOTL Research Methods Krista Trinder, College of Medicine Brad Wuetherick, GMCTE October 28, 2010.
Dr. John D. Barge, State School Superintendent “Making Education Work for All Georgians” Assessment for Learning Series Module 2: Understanding.
Student Preferences For Learning College Algebra in a Web Enhanced Environment Dr. Laura J. Pyzdrowski, Pre-Collegiate Mathematics Coordinator Institute.
Using Common Core State Standards of Seventh Grade Mathematics in the Application of NXT LEGO® Robotics for CReSIS Middle School Students.
8 Strategies for the Multiple Choice Portion of the AP Literature and Composition Exam.
Warm Up- Get a tracking sheet and handout on your way in. Sit in your usual seat and begin working on the warm up. Solve for x Solve for x 3. 4.
VALUE/Multi-State Collaborative (MSC) to Advance Learning Outcomes Assessment Pilot Year Study Findings and Summary These slides summarize results from.
Health Literacy is an individual’s ability to access, understand, and use information to achieve good health. LA County has a vital interest in health.
Psychometrics. Goals of statistics Describe what is happening now –DESCRIPTIVE STATISTICS Determine what is probably happening or what might happen in.
Reliability performance on language tests is also affected by factors other than communicative language ability. (1) test method facets They are systematic.
Early Identification of Introductory Major's Biology Students for Inclusion in an Academic Support Program BETHANY V. BOWLING and E. DAVID THOMPSON Department.
Challenges of Quantitative Reasoning Assessment Donna L. Sundre Center for Assessment and Research Studies
Student Growth Goals for Coaching Conversations. Requirements for Student Growth Goals Time line Reference to an Enduring Skill Proficiency and Growth.
Quantitative Literacy Assessment At Kennedy King College Fall 2013 DRAFT - FOR DISCUSSION ONLY 1 Prepared by Robert Rollings, spring 2014.
Using Rubrics for Information Literacy Skill Assessment: A Case Study for Business Education ABSTRACT Information literacy—the ability to find, evaluate.
Evaluation and Assessment Evaluation is a broad term which involves the systematic way of gathering reliable and relevant information for the purpose.
Spring 2015 Verona PARCC Results: Year One Wednesday, March 16 7:00-8:00 p.m. VHS Learning Commons.
Laboratory Science and Quantitative Core Requirements.
Quantitative Literacy at WOU. LEAP In November, Faculty Senate recommended that WOU adopt LEAP's (Liberal Education & America's Promise) learning outcomes.
Quantitative Literacy Assessment
Computer aided teaching of statistics: advantages and disadvantages
What is a CAT? What is a CAT?.
CRITICAL CORE: Straight Talk.
Preliminary Review of the 2012 Math SOL Results
Classroom Assessment A Practical Guide for Educators by Craig A
Change Score Analysis versus ANCOVA in Pretest/Posttest Designs:
THE RELATIONSHIP BETWEEN PRE-SERVICE TEACHERS’ PERCEPTIONS TOWARD ACTIVE LEARNING IN STATISTIC 2 COURSE AND THEIR ACADEMIC ACHIEVEMENT Vanny Septia Efendi.
Business Administration Programs School of Business and Liberal Arts Fall 2016 Assessment Report
Foundational Studies CARS – Summer 2018.
Teaching Quantitative Reasoning
Educational Testing Service
Presentation transcript:

Lessons Learned about Assessing Quantitative Literacy MAA PREP Workshop: Creating and Strengthening Interdisciplinary Programs in Quantitative Literacy June 14–18, 2005, Macalester College, St. Paul, Minnesota Jack Bookman Mathematics Department Duke University

Assessing QL presents the same challenges of assessing other types of learning but QL presents some particular challenges for assessment: Assessment items must be set in a real world context. The problems and the contexts must be familiar to all the students but neither so routine that they require very little thought nor so non-routine that students cannot solve them. The assessment instruments must not take too much time to administer, must include a multitude of problems and situations, and must be designed to allow for reliable scoring.

The “lessons learned” have come from work on two evaluation projects: Macalester College’s Quantitative Methods for Public Policy the goal, that students can "demonstrate basic understanding of the mathematical language of change and comparison" will be measured using pre-and post-tests An NSF project where we examined some existing QL tests used at two liberal arts colleges item analyses of the placement tests pre-test and post-test results from the QL course

Making up good test items is harder than we thought In the first attempt at making up questions for QM4PP assessment, we discovered a problem of false negatives – a student who may know the quantitative implications of a question may not address those issues in his or her answer. Example – For the statement below, what additional thing(s) would you need to know to make a decision? Because immigrants pay taxes, we should allow anyone into the US as an immigrant.

Making up good test items is harder than we thought Too little guidance should be avoided as seen in the previous question. Too much guidance should be avoided.

Making up good test items is harder than we thought Students may surprise you with their answers For example: The government will allow people age 55 to prepay their retirement taxes --- the taxes that they need to pay on the retirement savings that they withdraw or on social security payments. For every $1000 that someone would pay at age 65, the potential retiree can pre-pay, at age 55, at a discount: $500. Does this make financial sense for the government to do this? Briefly say why.

Coming up with a good grading rubric was harder than we thought We used the following general rubric for scoring: 4: Top score. The three C’s need to be there: Correct, Complete, Clear. 3: Close but not quite. Generally missing one of the three C’s earned a 3. 2: Partial. Best possible for an incorrect answer, worst possible for a correct answer. Often a good answer with no support or a wrong answer with good but misguided support. 1: Attempt was made. 0: Blank or a variant of “I have no idea” with no actual try. A more specific rubric was then formed for each of the fourteen questions. In other more standard math tests of concepts or problem solving, this type of rubric has been reasonably effective with high inter-rater reliability. We found this harder to do with the more nuanced questions in our QL tests.

Reaching proficiency may take more than a one semester course For each of the two schools in the NSF study, we examined the pre and post test results for those students who did not pass the proficiency test and were placed into a quantitative literacy course. We found (at each of the institutions) statistically significant growth during the one semester QL course, even after scores were adjusted to consider regression toward the mean. However, in spite of this growth, the average post-test score was still below the competency level as defined for the pre-test. This is consistent with results reported elsewhere (at Virginia Commonwealth for example). We believe that this should be expected since there is only so much growth that one can expect from one course. We believe that this gives some preliminary evidence for the assertion that in order to create a large positive effect concerning QL, colleges and universities must have a sustained effort that continues past a one course requirement.

There are some reasonably easy statistical tests that can be used to examine how well individual items serve the goal of the test. In particular, if the goal of the test is to be a proficiency test it is possible to examine how well a question discriminates at the passing level by plotting the scores on each question vs. the total score on the test.

In order to examine the ability of the questions to discriminate at the critical score of 30, we plotted the scores on each question (adjusted for difficulty level, that is, the score on the question/maximum possible score for that question) vs. the total score on the test and fit a cubic equation to that scatterplot. The greater the slope at the critical score of 30, the better the question discriminates across students at that criterion. For example, compare the graphs in Figure 1 below for questions 4 and 5 and notice how much better question 4 discriminates at QR = 30.

One can look at the difficulty level for each of the questions Note that 1, 5, 7, and 11 below are relatively easy questions and are therefore less likely to discriminate between stronger and weaker students.

We also computed the correlation coefficient for each question with the total score less the score on that question. Note that questions 5 and 7 below correlate rather poorly with the other questions.

Session 2: Thursday This will be a hands-on workshop where participants will work in pairs in a computer lab working with some of the data discussed in today's session. In particular, we will use two software packages (Excel and SPSS) to perform item analyses on some of the data used for placement in order to see how effective the items are in separating proficient from non-proficient students. Session 3: Friday In the first part of this session, we will work in small groups either designing an evaluation plan or developing assessment instruments. In the second part, each of the groups will present their designs for discussion by the whole group.