Assessing the association between quantitative maturity and student performance in simulation-based and non-simulation based introductory statistics Nathan.

Slides:



Advertisements
Similar presentations
Copyright© 2004 Avaya Inc. All rights reserved Comments on Draft Guidelines for Assessment and Instruction in Statistics Education (GAISE) Jim Landwehr.
Advertisements

 Small liberal arts college: 1350 undergraduate students  Statistician within Department of Math, Stat and CS  Class size: Stat 131 (30 students),
Implementation and Order of Topics at Hope College.
A. John Bailer Statistics and Statistical Modeling in The First Two Years of College Math.
Is College Success Associated With High School Performance? Elizabeth Fisk, Dr. Kathryn Hamilton (Advisor), University of Wisconsin - Stout Introduction.
QUANTITATIVE EVIDENCE FOR THE USE OF SIMULATION AND RANDOMIZATION IN THE INTRODUCTORY STATISTICS COURSE Nathan Tintle Associate Professor of Statistics.
Christine Bastedo Robert Magyar Spring,  Determine if students are meeting learning objectives across all sections of PSY 121, Methods and Tools.
Robert delMas (Univ. of Minnesota, USA) Ann Ooms (Kingston College, UK) Joan Garfield (Univ. of Minnesota, USA) Beth Chance (Cal Poly State Univ., USA)
The Hybrid Mathematics Class: The best of both worlds, or No Man’s Land? Mary Hudachek-Buswell Catherine Matos Clayton State University 1.
Student Achievement and Predictors of Student Achievement in a State Level Agricultural Mechanics Career Development Event Edward Franklin Glen Miller.
Hands on Statistics: Active learning in College Guojing Wang, Maika Yanagishita, Andrew Hwang, James Florczak Introduction As learning assistants for Statistics.
Ismaila Odogba Department of Geography and Geology The Impact of Collaborative Testing on Student Learning in an Introductory Level Geography Course INTRODUCTION.
Developing teachers’ mathematics knowledge for teaching Challenges in the implementation and sustainability of a new MSP Dr. Tara Stevens Department of.
The Genetics Concept Assessment: a new concept inventory for genetics Michelle K. Smith, William B. Wood, and Jennifer K. Knight Science Education Initiative.
McMillan Educational Research: Fundamentals for the Consumer, 6e © 2012 Pearson Education, Inc. All rights reserved. Educational Research: Fundamentals.
Effectiveness of Instructional Techniques to Create Retainable Quantitative Skills Kathy Baughman Assistant Professor of Accounting Wei-Chung Wang Assistant.
The Impact of Robot Projects on Girls' Attitudes Toward Science and Engineering Jerry Weinberg, Associate Professor Dept. of Computer Science Susan Thomas,
1 Self-Regulation and Ability Predictors of Academic Success during College Anastasia Kitsantas, Faye Huie, and Adam Winsler George Mason University.
This material is based upon work supported by the National Science Foundation under Grant No and Any opinions, findings, and conclusions.
Early Identification of Introductory Major's Biology Students for Inclusion in an Academic Support Program BETHANY V. BOWLING and E. DAVID THOMPSON Department.
Quantitative Literacy Assessment At Kennedy King College Fall 2013 DRAFT - FOR DISCUSSION ONLY 1 Prepared by Robert Rollings, spring 2014.
Crystal Reinhart, PhD & Beth Welbes, MSPH Center for Prevention Research and Development, University of Illinois at Urbana-Champaign Social Norms Theory.
Title of Study : Preliminary findings from “An evaluation of the impact of the inclusion of a "Health and Well Being Module" in the undergraduate curriculum.
Assessment in First Courses of Statistics: Today and Tomorrow Ann Ooms University of Minnesota.
Classroom Network Technology as a Support for Systemic Mathematics Reform: Examining the Effects of Texas Instruments’ MathForward Program on Student Achievement.
Improving Reading Compliance and Quiz Scores through the Use of Reading Guides Trent W. Maurer & Judith Longfield.
“Excuse Me Sir, Here’s Your Change”
Katie Dell and Cindy Ghent
Analysis of AP Exam Scores
Discussant Webster West.
Inferential Statistics
Response to Intervention & Positive Behavioral Intervention & Support
Using a Strategy Project to Promote Self-Regulated Learning
Module Example: Influence of Race and Gender on Income1
Teaching Undergraduate Statistics without Math
Individualized research consultations in academic libraries: Useful or useless? Let the evidence speak for itself Karine Fournier Lindsey Sikora Health.
Evaluating Student-Teachers Using Student Outcomes
Preliminary Data Analyses
Indiana University – Purdue University Indianapolis
Assessing Students' Understanding of the Scientific Process Amy Marion, Department of Biology, New Mexico State University Abstract The primary goal of.
Improving Student Engagement Through Audience Response Systems
Hybrid Mathematics 140 Course (INTRODUCTORY STATISTICS) using Carnegie Mellon’s Open Learning Initiative.
Brotherson, S., Kranzler, B., & Zehnacker, G.
Longitudinal Designs.
Carrie O’Reilly, Ph.D., M.S.N., RN Touro University Nevada
Assessing the association between quantitative maturity and student performance in simulation-based and non-simulation based introductory statistics Nathan.
2015 PARCC Results for R.I: Work to do, focus on teaching and learning
What are the SATS tests? The end of KS2 assessments are sometimes informally referred to as ‘SATS’. SATS week across the country begins on 14th May 2018.
Derek Herrmann & Ryan Smith University Assessment Services
Chapter Eight: Quantitative Methods
Conclusion & Implications
CORE Academic Growth Model: Results Interpretation
Social Change Implications
Jennifer Bryer PhD, RN, CNE Virginia Peterson-Graziose DNP, RN, CNE
Multiple Measures Susan Barbitta Associate Director, Special Projects
PLCs Professional Learning Communities Staff PD
Indiana University School of Social Work
What are the SATS tests? The end of KS2 assessments are sometimes informally referred to as ‘SATS’. SATS week across the country begins on 13th May 2019.
What are the SATS tests? The end of KS2 assessments are sometimes informally referred to as ‘SATS’. SATS week across the country begins on 13th May 2019.
Student Equity Planning August 28, rd Meeting
EPAS Educational Planning and Assessment System By: Cindy Beals
John Symons, Department of Philosophy
Conclusions/ Future Directions
Assessment, Attitudes and ARTIST
What are the SATS tests? The end of KS2 assessments are sometimes informally referred to as ‘SATS’. SATS week across the country begins on 13th May 2019.
EBPS Year 6 SATs evening.
Split-Block Class Schedule at Yorktown High School
Educational Testing Service
CLASSROOM ENVIRONMENT AND THE STRATIFICATION OF SENIOR HIGH SCHOOL STUDENT’S MATHEMATICS ABILITY PERCEPTIONS Nelda a. nacion 5th international scholars’
KATHY ZOGHBY University of Richmond Richmond, Virginia
Presentation transcript:

Assessing the association between quantitative maturity and student performance in simulation-based and non-simulation based introductory statistics Nathan Tintle Associate Professor of Statistics Dordt College, Sioux Center, Iowa

Background What is simulation-based inference and why are people using it? Preliminary evidence is positive in aggregate and on particular subscales Getting questions from workshop participants and users Will this work for my students who are particularly weak (or strong) or…

Literature Some prior studies have examined student quantitative maturity and performance in introductory statistics courses Johnson and Kuenen (2006) – prior math skills (ACT and basic math skills test) strongly predicted student performance in intro stats Green et al. (2009) – prior college math courses taken and order they were taken strong predictors of student performance intro stats (business school) Rochelle and Dotterweich (2007) – Performance in algebra course and college GPA strong predictors of student performance (business school) Li et al. (2012) and Wang et al. (2007) – college GPA and ACT both strong predictors of student performance Gnaldi (2006) – Poor math training in high school leads to challenges in learning statistics in college

Literature Some prior studies have examined student quantitative maturity and performance in introductory statistics courses Lester (2007) – Age and algebra skills strong predictors of student performance Dupuis et al. (2011) – More math coursework in high school associated with better performance in college stats (multi-institutional) Cherney and Cooney (2005) and Silvia et al. (2008) – mathematical skills and student attitudes were both significant predictors of student performance

Overall themes To date Associations found between prior abilities and performance and statistics course performance Tend to be single institution Tend to focus on general mathematical and quantitative reasoning skills (ACT score, algebra skills) Tend to focus on course grade as a measure of performance

Gaps in research to date Course grade is not normalized across institutions E.g., not nationally normed assessment tests of student abilities; no sense of how much memorization/rehearsed algebra vs. student conceptual understanding Course grade does not look at change over time (pre-course to post-course) E.g., Are students actually learning anything in the course or did they know it already? Primarily focused on single curricula Similar results across institutions? Similar results across curricula? Simulation-based curriculum vs. Traditional curriculum

Results addressing these gaps: Part 1 Two ‘before and after’ stories Two Midwestern liberal arts colleges Using ‘traditional’ Stat 101 curricula (Moore at one; Agresti and Franklin at the other) Changed to early versions of the ISI curriculum

Results addressing these gaps: Part 1 Two ‘before and after’ stories Traditional curriculum 289 students Two semesters Two institutions Multiple sections and instructors (average size of 25-30 students p/section) Comprehensive Assessment of Outcomes in Statistics (CAOS) First week of class During finals week Online, incentive for taking, not for performance Response rate over 85% p/section

Results addressing these gaps: Part 1 Two ‘before and after’ stories Early SBI curriculum 366 students Three semesters Two institutions Multiple sections and instructors (average size of 25-30 students p/section) Comprehensive Assessment of Outcomes in Statistics (CAOS) First week of class During finals week Online, incentive for taking, not for performance Response rate over 85% p/section

Results addressing these gaps: Part 1 Screened out some bad data (e.g., too quick taking the assessment) Similar demographics between the two (before and after, within institution)

Table 1. Pre- and post-course CAOS scores stratified by pre-course performance and curriculum Pre-test score group Curriculum Pre-test mean % correct (SD) Post-test mean % correct (SD) Change in mean % correct (SD)1 Difference in mean change by curriculum2 Low (≤40%) Consensus (n=80) 35.2 (4.9) 48.1 (8.8) 12.9 (9.4)*** 1.9 Early-SBI (n=141) 35.2 (4.8) 50.1 (9.8) 14.9 (10.6)*** Middle (Between 40 and 50%) Consensus (n=77) 45.1 (2.1) 52.0 (10.2) 6.9 (10.2)*** 3.1* Early-SBI (n=108) 44.9 (2.0) 54.9 (10.8) 10.0 (10.4)*** High (≥50%) Consensus (n=129) 57.1 (6.8) 62.3 (6.8) 5.2 (9.1)*** 2.1 Early-SBI (n=117) 55.8 (5.6) 63.0 (11.3) 7.2 (10.5)*** Overall Consensus (n=289) 47.7 (10.7) 55.5 (11.8) 7.8 (10.0)*** 3.3*** Early-SBI (n=366) 44.6 (9.7) 55.6 (11.9) 11.0 (11.0)*** *p<0.05; **p<0.01; ***p<0.001 1. Significance is indicated by asterisks and reported based on results from paired t-tests comparing the pre-test and post-test scores 2. From a linear model predicting the change in score by curriculum and adjusted for institution. Institution was not significant in any of the four models (p-values of 0.62-low; 0.22- middle; 0.72- high; 0.38- overall).

Take homes from Table 1 All groups significantly better on both curricula Some evidence of improved performance within each strata of pre-test score with early SBI curriculum (not always statistically significant)

Table 2. Pre- and post-course CAOS scores stratified by ACT score and curriculum1 1. Only for students with ACT scores available (all students with available ACT scores were from one of the two colleges evaluated in Table 1) 2. From a paired t-test comparing the pre-test and post-test scores 3. These values indicate how different the two curricula are with regards to changing student scores. For example, 8.2 means that the Early-SBI curriculum shows an improvement in percent correct which is 8.2 percentage points higher than the Consensus curriculum. A test to see whether there was evidence that the difference in mean changes were different by ACT group (e.g., 8.2 vs. 4.7 vs. 6.0) did not yield evidence of a significant (p=0.15; ANOVA comparison of whether a model predicting post-test scores by pre-test, curriculum used and ACT score group was significantly different than a model which predicted post-test scores by pre-test and curriculum only). ACT Group Curriculum Pre-test mean % correct (SD) Post-test mean % correct (SD) Change in mean % correct (SD)2 Difference in mean change by curriculum3 Low (≤22) Consensus (n=21) 41.7 (10.3) 46.3 (10.1) 4.0 (11.7) 8.2*** Early-SBI (n=55) 42.7 (10.1) 54.9 (11.9) 12.2 (10.5)*** Middle (23-26) Consensus (n=34) 46.0 (8.2) 52.4 (10.3) 6.5 (9.2)*** 4.7* Early-SBI (n=48) 44.0 (10.0) 55.1 (10.8) 11.2 (11.4)*** High (≥27) Consensus (n=36) 51.3 (7.7) 57.1 (7.7) 5.8 (9.2)** 6.0* Early-SBI (n=49) 47.8 (9.8) 59.5 (12.0) 11.8 (10.1)*** Overall Consensus (n=91) 46.4 (9.3) 52.0 (11.0) 5.6 (9.8) 6.0*** Early-SBI (n=152) 44.9 (10.1) 56.5 (11.6) 11.6 (10.7)

Take homes from Table 2 All groups show significant improvement except The lowest ACT score group with the consensus curriculum All groups significantly better with early-SBI curriculum Limitation: Only one of the two institutions had ACT scores available, however, no significant evidence of institutional differences in the models for Table 1.

Part 1: Results by subscale Nine CAOS subscales Graphical representations Boxplots Data collection and design Descriptive statistics Tests of significance Bivariate relationships Confidence Intervals Sampling Variability Probability/Simulation

Results by subscale Lowest pre-test group (<40%) 3 subscales significant improvement (p<0.05) SBI vs. traditional Data collection and design (9.4 points more improvement for SBI), Tests of significance (8.4 points), Probability/Simulation (15.8 points) 6 no significant change (4 of 6 improved SBI vs. traditional) Middle pre-test group (40-50%) 2 subscales significant improvement (p<0.05) SBI vs. traditional Data collection and design (11.5 points) and Probability/Simulation (14.4 poitns) 6 no change (5 of 6 improved SBI vs. traditional) 1 significantly worse (10 point decrease on descriptive statistics) Highest pre-test group (50%+) 1 subscale significant improvement (Tests of significance 8.2 points) 8 no change (5 of 8 improved)

Take homes from subscale results SBI better across most subscales and most subgroups of students with most gains on tests of significance, probability/simulation and design *Note that decrease for descriptive statistics went away in later versions of the curriculum

Part 2. Multi-institution analysis 1078 students 34 instructor sections 13 institutions (1 CC, 1 private university, 2 high school AP courses, four liberal arts colleges and 4 public universities) Modified CAOS test Elimination or modification to questions commonly missed on post-test or commonly correct pre-test Generally similar administration (online, outside of class, first week/finals week, incentive to participate) across sections College GPA via self-report All sections used an SBI curriculum (the ISI curriculum)

Part 2. Multi-institution analysis How grouped Grouping Pre-test Mean (SD) Post-test Mean (SD) Change Mean (SD)1 Pre-test concept score Low (<40% correct; n=291) 35.0 (5.0) 50.2 (12.0) 15.2 (12.3)*** Middle (40-55%; n=422) 48.1 (3.8) 56.2 (12.1) 8.1 (11.9)*** High (55%+; n=365) 64.1 (7.3) 68.1 (12.8) 4.0 (10.8)*** Self-reported college GPA Low (B or worse; n=193) 45.6 (12.3) 52.9 (13.1) 7.3 (11.8)*** Middle (B+ to A-; n=654) 50.0 (12.0) 58.1 (13.6) 8.1 (12.6)*** High (A; n=231) 53.8 (13.6) 64.9 (14.7) 11.1 (12.2)***   Overall 50.0 (12.6) 58.6 (14.3) 8.6 (12.5)*** 1 From a paired t-test comparing the pre-test and post-test scores

Part 2. Multi-institution analysis Take home messages All groups show significant improvement Regression to mean effects noted when stratifying by pre-test score Improvement comparable across groups

Part 2. Multi-institution analysis by subscale Seven subscales with new instrument; stratified by GPA Table shows pre-test to post-test improvement Low Middle High Graphical rep’s 6.5* 6.4*** 8.9*** Data collection and design -2.6 -0.0 7.5*** Desc. Stats 4.4 9.7*** 5.7 Tests of sig 10.7*** 10.8*** 14.8*** Conf Intervals 10.3*** 14.0*** Sampling Variability 12.6*** 3.2* 5.5 Probability/Simulation 10.9*** 10.6*** 11.4***

Take homes Fairly consistent improvement across scales Fairly consistent improvement across subgroups stratified by GPA Some modest evidence of differences for data collection/design and sampling variability

Overall take home messages SBI showed consistent improvement at two institutions vs. traditional curriculum Across student ability groups (pre-test on stat conceptual understanding and ACT) Particular gains in conceptual areas emphasized by SBI; do-no-harm in other areas SBI showed improvement across conceptual subscales across institutions and across student ability groups as measured by conceptual pre-test and self-reported college GPA

Limitations Self-reported GPA; Limited ACT data Limited institutions (pedagogical effects; institutional effects, etc.) Limited abilities to draw cross-curricular comparisons at additional institutions Factors other than GPA/ACT/pre-test score worth looking at (e.g., SES; race/ethnicity) More sophisticated statistical modelling possibilities (hierarchical models; relative gain vs. absolute) Incorporating student attitudes These limitations are being addressed in current work with an expanded set of data being gathered and in ongoing statistical analysis of current data; Similar results so far.

Conclusions SBI continues to show promise overall and across student ability groups, especially in areas of emphasis of SBI curricula (design, tests of significance, probability/simulation) Future work is needed to ensure transferability of results to broader groups of students and additional SBI and non-SBI curricula and consider impacts of student demographics and attitudes towards statistics

Acknowledgments Beth Chance, Cindy Nederhoff and others on the ISI development and assessment team Support from the National Science Foundation under(Grant DUE- 1140629) and (Grant DUE-1323210) Class testers and students!

References Cherney, I. D., & Cooney, R. R. (2005). The Mathematics and Statistics Perception Scale. Transactions of the Nebraska Academy of Sciences, 30, 1–8. Dupuis, D. N., Medhanie, A., Harwell, M., Lebeau, B., Monson, D., & Post, T. R. (2011). A multi-institutional study of the relationship between high school mathematics achievement and performance in introductory college statistics. Statistics Education Research Journal, 11(1), 4–20. Gnaldi, M. (2006). The relationship between poor numerical abilities and subsequent difficulty in accumulating statistical knowledge. Teaching Statistics, 28(2), 49–53. Green, J. J., Stone, C. C., Zegeye, A., & Charles, T. A. (2009). How much math do students need to succeed in business and economics statistics? An ordered probit analysis. Journal of Statistics Education, 17(3). Johnson, M., & Kuennen, E. (2006). Basic math skills and performance in an introductory statistics course. Journal of Statistics Education, 14(2). Retrieved from http://www.amstat.org/publications/jse/v14n2/johnson.html Lester, D. (2007). Predicting performance in a psychological statistics course. Psychological Reports, 101, 334. Li, K., Uvah, J., & Amin, R. (2012). Predicting Students’ Performance in Elements of Statistics. US-China Review, 10, 875–884. Retrieved from http://files.eric.ed.gov/fulltext/ED537981.pdf Malone, C., Gabrosek, J., Curtiss, P., & Race, M. (2010). Resequencing topics in an introductory applied statistics course. The American Statistician, 64(1), 52– 8. Rochelle, C. F., & Dotterweich, D. (2007). Student Success in Business Statistics. Journal of Economics and Finance Education, 6(1). Scheaffer, R. (1997). Discussion to new pedagogy and new content: The case of statistics. International Statistics Review, 65(2), 156–8. Silvia, G., Matteo, C., Francesca, C., & Caterina, P. (2008). Who failed the introductory statistics examination? A study on a sample of psychology students. International Conference on Mathematics Education. Wang, J.-T., Tu, S.-Y., & Shieh, Y.-Y. (2007). A study on student performance in the college introductory statistics course. AMATYC Review, 29(1), 54–62.