Cluster 5 Spring 2005 Assessment Results Sociocultural Domain.

Slides:



Advertisements
Similar presentations
Ed-D 420 Inclusion of Exceptional Learners. CAT time Learner-Centered - Learner-centered techniques focus on strategies and approaches to improve learning.
Advertisements

FACULTY DEVELOPMENT PROFESSIONAL SERIES OFFICE OF MEDICAL EDUCATION TULANE UNIVERSITY SCHOOL OF MEDICINE Using Statistics to Evaluate Multiple Choice.
USING THE CLA TO INFORM CAMPUS PLANNING Anne L. Hafner Campus WASC Faculty Coordinator Winter 2008.
© 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill Validity and Reliability Chapter Eight.
Increasing your confidence that you really found what you think you found. Reliability and Validity.
State of Texas Assessment of Academic Readiness--EOC Tim Walker Nathaniel Session.
School Report Cards 2004– The Bottom Line More schools are making Adequate Yearly Progress. Fewer students show serious academic problems (Level.
Item Analysis: A Crash Course Lou Ann Cooper, PhD Master Educator Fellowship Program January 10, 2008.
Analyzing & Interpreting Data Assessment Institute Summer 2005.
OVERVIEW OF CHANGES TO EDUCATORS’ EVALUATION IN THE COMMONWEALTH Compiled by the MOU Evaluation Subcommittee September, 2011 The DESE oversees the educators’
Statistical Issues in Research Planning and Evaluation
Effects of MMM on Performance, Attitudes and Efficacy of Students in Precalculus Brad Bailey June 3, 2011.
-生醫統計期末報告- Reliability 學生 : 劉佩昀 學號 : 授課老師 : 蔡章仁.
4/16/07 Assessment of the Core – Social Inquiry Charlyne L. Walker Director of Educational Research and Evaluation, Arts and Sciences.
Chapter 4 Validity.
Lessons Learned about Assessing Quantitative Literacy MAA PREP Workshop: Creating and Strengthening Interdisciplinary Programs in Quantitative Literacy.
Getting Started with Hypothesis Testing The Single Sample.
Reliability of Selection Measures. Reliability Defined The degree of dependability, consistency, or stability of scores on measures used in selection.
+ Results of a National Assessment of Information Literacy Skills.
S-CAPE Testing for Higher Proficiency Levels and Other Factors that Influence Placement at Brigham Young University Elizabeth Robinson.
EngageNY.org Scoring the Regents Examination in Algebra I (Common Core)
1 Core Assessment: Attributes of a Successful Program at James Madison University TAIR Professional Development Workshop September 2004 Dr. Dena Pastor.
METHODS Study Population Study Population: 224 students enrolled in a 3-credit hour, undergraduate, clinical pharmacology course in Fall 2005 and Spring.
Assessment Report Department of Psychology School of Science & Mathematics D. Abwender, Chair J. Witnauer, Assessment Coordinator Spring, 2013.
Ismaila Odogba Department of Geography and Geology The Impact of Collaborative Testing on Student Learning in an Introductory Level Geography Course INTRODUCTION.
Chris Evans, University of Winchester Dr Paul Redford, UWE Chris Evans, University of Winchester Dr Paul Redford, UWE Self-Efficacy and Academic Performance:
Chemistry B.S. Degree Program Assessment Plan Dr. Glenn Cunningham Professor and Chair University of Central Florida April 21, 2004.
Interpreting Assessment Results using Benchmarks Program Information & Improvement Service Mohawk Regional Information Center Madison-Oneida BOCES.
McMillan Educational Research: Fundamentals for the Consumer, 6e © 2012 Pearson Education, Inc. All rights reserved. Educational Research: Fundamentals.
Technical Adequacy Session One Part Three.
Are there “Hidden Variables” in Students’ Initial Knowledge State Which Correlate with Learning Gains? David E. Meltzer Department of Physics and Astronomy.
4/16/07 Assessment of the Core – Science Charlyne L. Walker Director of Educational Research and Evaluation, Arts and Sciences.
The Genetics Concept Assessment: a new concept inventory for genetics Michelle K. Smith, William B. Wood, and Jennifer K. Knight Science Education Initiative.
What Was Learned from a Second Year of Implementation IES Research Conference Washington, DC June 8, 2009 William Corrin, Senior Research Associate MDRC.
Final Update on the New Faculty Course Evaluation & Online System November, 2003.
Assessment in General Education: A Case Study in Scientific and Quantitative Reasoning B.J. Miller & Donna L. Sundre Center for Assessment and Research.
By: Amani Albraikan.  Pearson r  Spearman rho  Linearity  Range restrictions  Outliers  Beware of spurious correlations….take care in interpretation.
Multiple Perspectives on CAT for K-12 Assessments: Possibilities and Realities Alan Nicewander Pacific Metrics 1.
Hypothesis Testing An understanding of the method of hypothesis testing is essential for understanding how both the natural and social sciences advance.
Comparing Senior And Sophomore Knowledge and Confidence Concerning Academic Advising Anecdotal evidence suggested that a discrepancy existed between what.
Validity and Item Analysis Chapter 4. Validity Concerns what the instrument measures and how well it does that task Not something an instrument has or.
Validity and Item Analysis Chapter 4.  Concerns what instrument measures and how well it does so  Not something instrument “has” or “does not have”
Criteria for selection of a data collection instrument. 1.Practicality of the instrument: -Concerns its cost and appropriateness for the study population.
Fidelity of Implementation A tool designed to provide descriptions of facets of a coherent whole school literacy initiative. A tool designed to provide.
Solution to Questions 1 and 2: Students' Use of Rule-based Reasoning in the Context of Calorimetry and Thermal Phenomena* Ngoc-Loan P. Nguyen, Warren M.
Challenges of Quantitative Reasoning Assessment Donna L. Sundre Center for Assessment and Research Studies
Faculty Forum: Evaluation of Teaching Sponsored by the Faculty Senate November 10, 2006.
LISA A. KELLER UNIVERSITY OF MASSACHUSETTS AMHERST Statistical Issues in Growth Modeling.
MT ENGAGE Student Learning Outcomes and Assessment April 27, 2015.
1 Main achievement outcomes continued.... Performance on mathematics and reading (minor domains) in PISA 2006, including performance by gender Performance.
STANDARDIZED TESTING Understanding the Vocabulary.
The Process The Results The Repository of Assessment Documents (ROAD) Project Sample Characteristics (“All” refers to all students enrolled in ENGL 1551)
What is a CAT? What is a CAT?.
Assessments for Monitoring and Improving the Quality of Education
Assessing Students' Understanding of the Scientific Process Amy Marion, Department of Biology, New Mexico State University Abstract The primary goal of.
Reliability and Validity in Research
Concept of Test Validity
Business Assessment Test Results
Validity and Reliability
Reliability & Validity
The General Education CLAS Core
Module 8 Statistical Reasoning in Everyday Life
Understanding Your PSAT/NMSQT Results
Access Center Assessment Report
Statistical Inference for Managers
Student Satisfaction Results
Assessment Literacy: Test Purpose and Use
TEST FOR RANDOMNESS: THE RUNS TEST
Chapter 8 VALIDITY AND RELIABILITY
Tests are given for 4 primary reasons.
Presentation transcript:

Cluster 5 Spring 2005 Assessment Results Sociocultural Domain

Cluster 5 Course Requirement  Complete 1 course in each domain Wellness Domain GHTH 100. Personal Wellness GKIN 100. Lifetime Fitness and Wellness GEIC 101. Individuals in the Human Community: The Individual Perspective Sociocultural Domain GPSYC 101. General Psychology GPSYC 160. Life Span Human Development GSOCI 240. Individual in Society GEIC 102. Individuals in the Human Community: The Community Perspective

Assessment Day Data Collection Scheme: Pre/Post COHORT 1 COHORT 2 COHORT 3 Students in each cohort are tested twice on the same instrument – once as incoming freshmen (0 credit hours) and again in after they’ve completed credit hours. Cohort 2: First time for Cluster 5 to have pre/post data since new objectives/instruments.

Sociocultural Domain Instruments  Sociocultural Domain Assessment (SDA) –3 different versions of test  Scenario-based multiple-choice exam created by faculty Administration error led to SDA1 & SDA2 being administered in Spring 2005 – because of this only 27 common items on both forms used to create total score

Results  First: examine results for nd semester sophomores in Spring 2005 –Overall results –By JMU vs. Transfer/AP –By domain “completers” vs. “non-completers” –By which course completed –Rel. between course grade and SDA score  Second: examine pre/post results for the 264 students who took the SDA in both Fall 2003 and Spring 2005 –Overall pre/post difference –pre/post difference by domain “completers” vs. “non- completers” –pre/post difference by which course completed –Pre/post difference by item for “completers”

Overall SDA Results: 2 nd Semester Sophomores (N = 398) On average, students are answering almost 70% of the items correctly. Most scores are between 17 & 23.

JMU (n = 369) vs. Transfer/AP Students (n = 29)  Are students who obtained transfer or AP credit for their courses scoring differently than students who completed their courses at JMU? –No.  JMU students: M = 18.63, SD = 3.83  AP/transfer students: M = 18.55, SD = 3.68  Difference was not statistically, t(335)=.11, p =.92, nor practically significant (d =.02).

Domain “Completers” vs. “Non-Completers” More courses completed, the higher the mean. Differences among the means, however, are of small practical significance (d’s between )

SDA Means by Course Completed “Completers” Only There were minor differences among the SDA means (d’s ~ ) GEIC101 not reported due to small N

Relationship between SDA and Course Grade r =.31r =.23 r =.69r =.36 w/o outlier Positive and moderate relationships …as anticipated.

Pre/Post Results (N = 264)  Spring 2005: M = 18.78, SD = 3.66  Fall 2003: M = 16.62, SD = 3.38  The difference between these means is of medium to large practical significance (d =.64) Average increase of 8% in sociocultural domain knowledge

Pre/Post for “Non-Completers” (n = 33) vs. “Completers” (n = 199) Non-completersCompleters “Completers” start off with higher means than “Non- completers” – but they both make similar gains over time (~2 pt. increase, d ~.55) Larger gain (~3 pts, d = 1.02) for the 32 students who completed 2 – 3 sociocultural domain courses.

Pre/Post by Sociocultural Course Completed Spring 2005 average for GSOCI240 students may seem superior, keep in mind that these students were more knowledgeable coming into the course and that the averages for GSOCI240 are based on only 10 students. The magnitude of the change in knowledge is very similar for the three different courses (d ~.75). N = 153 N = 98N = 10

Pre/Post Difference in Items for “Completers” 7 items in Fall 2003 where less than half of the students are obtained the correct answer; there are only 2 such items in Spring 2005 Most of the gains are made for items of moderate to large difficulty.

Items with Large Pre/Post Gains 7% or more students answered these items correctly in Spring 2005 than in Fall 2003 Majority of these items (67%) were written to assess goals 2 and 3. Goal 2:Identify implications of taking action regarding social/behavioral issue Goal 3:Use evidence to develop and evaluate positions regarding social/behavioral issue

Items with Small (or No) Pre/Post Gains Items with no change, or change in the wrong direction are shown below for items that are of medium to extreme difficulty. 67% of these items are assessing either goals 1 and 4, which indicates that: 1) either more focus should be placed these goals in sociocultural domain courses or 2) that these items are of poor quality and either need to be revised or replaced Goal 1: Make plausible interpretations about behavior in social contexts Goal 4: Discriminate between ethical and nonethical practices in the social/behavioral sciences

Conclusions 2 nd Semester Sophomore Data  2 nd semester sophomores scoring on average ~70%  AP/Transfer students are scoring no differently than JMU students  “Non-completers” of the sociocultural domain requirement are scoring no differently than “Completers”  Same average score regardless of what course student took to complete sociocultural domain requirement  There is a relationship between course grade and score, indicating that test content and course content are related

Conclusions Pre/Post Data(1)  Average increase of 8% in sociocultural domain knowledge between 0 credit hours and credit hours (large effect) –62%  70%  “Completers” and “Non-completers” both make similar gains over time (~2 pt. increase, d ~.55) –Might be evidence that score gains are a result of maturation, not a result of taking sociocultural course –Note that “completers” come in with more knowledge. –Larger gains (~3 pts, d = 1.02)are seen for students completing 2 – 3 courses. –May need to collect larger sample of “non-completers” in future assessments

Conclusions Pre/Post Data(2)  Same gains are being made on the test for students who fulfilled their requirement by taking different sociocultural courses –Students who take SOCI240 may be more knowledge coming into the course than students who choose to take other courses  7 items where <50% of students obtaining correct response in Fall 2003 compared to 2 in Spring 2005  Not seeing gains on easy items  Most gains are for items assessing Goals 2 & 3  No gains (or gains in the wrong direction) for items assessing Goals 1 & 4

Next Steps  Take a closer look at change in % of students correctly answering items over time –Use information to make inferences about program OR possibly to revise items  Set standard score on test in order to obtain richer information

Setting Standards  May want to use standard setting procedures to establish “cut-offs” for proficiency on the test  Could be that students are gaining knowledge/skills over time, but are they gaining enough?