Presentation is loading. Please wait.

Presentation is loading. Please wait.

Frequently asked questions about development, interpretation and use of rubrics on campuses 1.

Similar presentations


Presentation on theme: "Frequently asked questions about development, interpretation and use of rubrics on campuses 1."— Presentation transcript:

1 Frequently asked questions about development, interpretation and use of rubrics on campuses 1

2  Terry Rhodes ◦ Vice President, Office of Quality, Curriculum and Assessment – AAC&U  Ashley Finley ◦ Senior Director of Assessment and Research – AAC&U finley@aacu.org 2

3  Created with funding from FIPSE & State Farm  In response to Spellings Commission report on need for greater accountability in higher educ.  AAC&U guided development of 16 rubric teams ◦ Consisting of faculty (primarily), student affairs, institutional researchers, national scholars ◦ Inter-disciplinary, inter-institutional  Each rubric underwent 3 rounds of testing on campuses with samples of student work, feedback went back to teams for revision  Intended to be modified

4 4

5 5 What Types of Institutions are Accessing the Rubrics? Representation also includes: All US states and territories Higher education consortia International institutions K-12 schools and systems Also: The Voluntary System of Accountability (VSA) approved use of rubrics for campus reporting in 2011

6 6  Assessment of students’ demonstrated performance & capacity for improvement  Faculty-owned & institutionally shared  Used for students’ self-assessment of learning  Increase transparency of what matters to institutions for student learning

7  Knowledge of Human Cultures & the Physical & Natural Worlds ◦ Content Areas  No Rubrics  Intellectual and Practical Skills ◦ Inquiry & Analysis ◦ Critical Thinking ◦ Creative Thinking ◦ Written Communication ◦ Oral Communication ◦ Reading ◦ Quantitative Literacy ◦ Information Literacy ◦ Teamwork ◦ Problem-solving  Personal & Social Responsibility ◦ Civic Knowledge & Engagement ◦ Intercultural Knowledge & Competence ◦ Ethical Reasoning ◦ Foundations & Skills for Lifelong Learning ◦ Global Learning  Integrative & Applied Learning ◦ Integrative & Applied Learning

8

9 Criteria The Anatomy of a VALUE Rubric Levels Performance Descriptors

10  Do the performance levels correspond to year in school? (no) What about to letter grades? (no)  Can an assignment be scored by someone whose area of expertise is not the same as the content covered in the assignment? (yes)  “I’m a philosophy professor so how can I score a paper from a biology class?”  Can the rubrics be used by two year institutions? (yes)  Can the rubrics be used for graduate programs? (yes)  Can VALUE rubrics be used for course-level assessment? (yes, with modification or adaptation to include content areas) 10

11 Most people consider baking a skill. If a novice baker is given a recipe and a picture, can he/she recreate the dish? Chances are what is created will look more like this…but over time will get better with practice.

12  Calibration (norming) sessions  Assignment design workshops  Rubric modification workshops, specifically for adaptation of rubrics for program-level or course-level assessment  Data-centered events framed around interpretation of institutional data, celebration of areas of success and opportunity to gather recommendations for improvement

13 Dimension% of students who scored 2 or higher % of students who scored 3 of higher Explanation of Issues68.335.5 Interpreting & Analysis65.028.2 Influence of Context and Assumptions 48.821.2 Student’s position54.524.0 Conclusions and related outcomes 47.717.0 From: UNC-Wilmington, Critical Thinking Rubric

14 Percent of Ratings Critical Thinking: Issues, Analysis, and Conclusions Inter-rater reliability = >.8

15 Percent of Ratings Critical Thinking: Evaluation of Sources and Evidence

16 Percent of Ratings “VALUE added” for 4 years - writing

17  Crit. Lit. (CT, Rdg, Writing): 1,072 samples = gain of 0.88 bet. lower & higher credit students.  Research & Info. Literacy: 318 samples = gain of 1.49. The interdisc. scoring team found that programmatic definitions & practices around citation of researched info. varied widely, difficult to consistently score for plagiarism.  Oral Comm: 875 samples = gain of 0.14. 39% of samples not related to rubric. Samples had wide range of quality & other tech. limitations.  Quant. Reas.: 322 samples = gain of 0.97. Scoring team found 30% of samples were not related to rubric…

18  Using the VALUE Rubrics for Improvement of Learning and Authentic Assessment  12 Case Studies  Frequently asked questions http://www.aacu.org/value/casestudies/

19 19  Value Rubrics June 17, 3-4pm; June 18, 4-5pm (eastern)  Coding and Uploading Artifacts Using Taskstream Date and Time TBD Webinars will be recorded and posted to: http://www.sheeo.org/msc Webinars already posted:  Welcome to the MSC Questions?  Pilot Study Overview  Assignment Design Webinar  Sampling for the Pilot Study  IRB & Student Consent Additional MSC Webinars


Download ppt "Frequently asked questions about development, interpretation and use of rubrics on campuses 1."

Similar presentations


Ads by Google