Interpreting IDEA Results: Getting the Most from Course Evaluations

Slides:



Advertisements
Similar presentations
Standardized Scales.
Advertisements

 Goals: 1. To help faculty respond critically to student generated course evaluation; 2. To help faculty improve student learning.  Outcomes: Faculty.
IDEA Student Ratings of Instruction: A Diagnostic Guide for Improvement Dr. Kristi Roberson-Scott.
Using the IDEA Student Ratings System: An Introduction University of Saint Thomas Fall
Understanding and Using Your IDEA Evaluation Results Nina Campanicki & Ernie Linsay Faculty Development Day March 20, 2010.
IDEA Student Ratings of Instruction Update Carrie Ahern and Lynette Molstad Selected slides reproduced with permission of Dr. Amy Gross from The IDEA Center.
IDEA What it is and How to Implement the System Texas A & M, February 2013 Shelley A. Chapman, PhD Senior Educational Consultant.
IDEA Student Ratings of Instruction Insight Improvement Impact ® Loyola University of Chicago April 10-11, 2013 Steve Benton, PhD Senior Research Officer.
Performance Appraisal System Update
Interpreting Student Evaluations Heather McGovern, Fall 2011.
Interpreting IDEA Heather McGovern Director of the Institute for Faculty Development November 2009 add info on criterion vs. normed scores Add info on.
Writing Program Assessment Report Fall 2002 through Spring 2004 Laurence Musgrove Writing Program Director Department of English and Foreign Languages.
Evaluation of Teaching Excellence, a Guide for Administrators Heather McGovern, PhD Director of the Institute for Faculty Development Associate Professor.
Analyzing and Improving College Teaching: Here’s an IDEA Alan C. Lacy, Associate Dean College of Applied Science and Technology Illinois State University.
Interpreting IDEA reports Diagnostic Form Short Form
Assessment Overview Drake CPHS. Overview Overview of IDEA Data Assessing college-wide teaching goal Advising Results Q&A.
Office of Institutional Research, Planning and Assessment January 24, 2011 UNDERSTANDING THE DIAGNOSTIC GUIDE.
Student Engagement Survey Results and Analysis June 2011.
IDEA Student Ratings of Instruction Shelley A. Chapman, PhD Insight Improvement Impact ® University of Alabama Birmingham September 11, 2012.
The role of course evals in faculty reviews: One perspective John Petraitis CAFE March 25, 2011.
Data-Informed Faculty Development Planning Howard Community College August 17, 2012 Shelley A. Chapman, PhD.
University Teaching Symposium January 9,  Individual Development and Educational Assessment  Kansas State University  A diagnostic to.
 Shelley A. Chapman, PhD Texas A & M University February 2013.
My Professors Just Don’t Care! Carl Burns Director, Counseling Center Tammy Pratt Coordinator, Academic Support Programs.
IDEA STUDENT EVALUATION REPORTS Insight Improvement Impact ® Using IDEA as a Tool for Reflection about Your Teaching Raritan Valley Community College January.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
Summary of CPHS Course Evaluations: AY and Chuck Phillips August 2009.
Using IDEA for Assessment, Program Review, and Accreditation Texas A & M University November 8, 2012 Shelley A. Chapman, PhD.
DEVELOPED BY MARY BETH FURST ASSOCIATE PROFESSOR, BUCO DIVISION AMY CHASE MARTIN DIRECTOR OF FACULTY DEVELOPMENT AND INSTRUCTIONAL MEDIA UNDERSTANDING.
USING STUDENT EVALUATIONS AT STOCKTON Heather McGovern Director of the Institute for Faculty Development Associate Professor of Writing January 2012.
Starfish Training Minot State University
Advanced Writing Requirement Proposal
Good teaching for diverse learners
“Excuse Me Sir, Here’s Your Change”
Making an Old Enemy Your Friend
Portfolio Development
Evaluation Of and For Learning
Accreditation Survey Results
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
PeerWise Student Instructions
Performance Review Tool Updates College of Engineering
Online Course Design: Is the Conversation Over?
Culture Survey This document provides examples of how we analyze and report our clients’ culture survey data. It includes data from several clients in.
Internal Assessment 2016 IB Chemistry Year 2 HL.
Teaching Evaluations at TTU Using the IDEA Instrument
Implementing the NHS KSF Action Planning and Surgery Session
The University of Texas-Pan American
Internal assessment criteria
ASSESSMENT OF STUDENT LEARNING
Conducting the performance appraisal
Multi Rater Feedback Surveys FAQs for Participants
Multi Rater Feedback Surveys FAQs for Participants
Conducting the performance appraisal
School Climate Data Workshop
Online Probation Workshop
Derek Herrmann & Ryan Smith University Assessment Services
Starfish Faculty Training How to Raise Flags and Kudos
EVAAS Overview.
Your Institutional Report Step by Step
Implementing the Specialized Service Professional State Model Evaluation System for Measures of Student Outcomes.
Portfolio Development
North Carolina Positive Behavior Support Initiative
Leanne Havis, Ph.D., Neumann University
Parent-Teacher Partnerships for Student Success
Assessing Academic Programs at IPFW
2018 Great Colleges Survey for Champlain College
IDEA Student Ratings of Instruction
Learning Community II Survey
EDUC 2130 Quiz #10 W. Huitt.
Portfolio Development
Presentation transcript:

Interpreting IDEA Results: Getting the Most from Course Evaluations Formative Evaluation = Information for Faculty Improvement (what can I do better as an instructor) There are two sumative questions that can be informed through the new IDEA course evaluation system: How did the students in the class do? (achievement of learning objectives) How did the instructor do? Out discussion this morning will focus on the second question, specifically, the use of the IDEA results as a source of evidence in the USU Faculty Evaluation Process

Outline for Today Overview: IDEA Results Sample Size & Response Rate Avg. Score vs. Converted Avg. Understanding Learning Objectives Raw Scores vs. Adjusted IDEA’s use in Summative Evaluation IDEA Results & Teaching Practice There are two questions that can be informed through the new IDEA course evaluation system: How did the students in the class do (achievement of learning objectives)? How did the instructor do? Out discussion this morning will focus on the second question, specifically, the use of the IDEA results as a source of evidence in the USU Faculty Evaluation Process The IDEA system, and student ratings in general, can have a positive impact if they are not overemphasized in the evaluation process. Because student ratings do not perfectly measure teaching effectiveness, using them for more than 30-50% of the overall evaluation of teaching can be seen as punitive. Using them appropriately will encourage faculty trust in the process, which will facilitate effective use of the information for teaching improvement.

Overview: IDEA Results Reports will be mailed each semester Results are also available online: http://usu.edu/aaa/evaluations_all.cfm You will need to log-in to view this page Comments will be mailed/emailed, comments are not available online. Frequently Asked Questions are here: http://usu.edu/aaa/idea_faculty_faq.cfm IDEA always ships paper reports (168 pounds last fall). We will forward these to the departments each semester, along with a copy of the “Departmental Summary.” Two copies of the printed report are supplied for each class, so there is a departmental hard copy if desired. Only a single printed copy of students comments is provided. My office keeps an electronic copy of all materials. IDEA Results are available in a variety of formats online [show them the options by slowly walking them through the web pages] Search box provides access to evaluation results for all types of evaluations (new = IDEA, old: paper-and-pencil, old: Blackboard/Canvas online). You can search by faculty name (first, last, both); course name and/or number; term, year, or both; or any combination of the above The summary list includes only a few fields (there are screen limitations) These could be modified. Clicking on an entry in the list opens a new view, with much more detail [show them] From the detail view, you can directly access the full .pdf of the report These new IDEA .pdf files are also available in a searchable list: http://usu.edu/aaa/idea_results_frame.cfm Comments will not be posted online, but one printed copy will be mailed, and an electronic copy will be emailed Each department has (at least) one staff person trained on the IDEA system, and they will be copied on all mails and emails relating to IDEA

Sample Size/Response Rate 11/13/2018 Sample Size/Response Rate SHOULD WE EVEN BE EVALUATING CLASSES WITH ENROLLMENT OF 5 STUDENTS OR LESS? [Discuss] Large class sizes yield greater reliability (consistency). Response rate affects how representative your results are. The number of student respondents affects reliability. In this context, reliability refers to consistency, interrater reliability. Fewer than ten students are unreliable—evaluators should pay scant attention to numerical data from classes with fewer than ten respondents. IDEA reports the following median rates: 10 raters .69 reliability 15 raters .83 reliability 20 raters .83 reliability 30 raters .88 reliability 40 raters .91 reliability Reliability ratings below .70 are highly suspect. Response Rates at USU for Fall 2011: Total Response Rate: Logan Campus                      73.2% (Avg. of Avgs.)     70.4%  (Total Responses/Total Enrollment) Total Response Rate: RCDE                                   60.9% (Avg. of Avgs.)     60.4% (Total Responses/Total Enrollment)                Total Response Rate: USUE                                   36.0% (Avg. of Avgs.)     37.3% (Total Responses/Total Enrollment) Total Response Rate: Logan + RCDE                      70.1% (Avg. Of Avgs.)     68.3% (Total Responses/Total Enrollment) Total Response Rate: ALL (including USUE)           65.0% (Avg. of Avgs.)     65.8% (Total Responses/Total Enrollment)   Top Five Departments (based on Total Responses/Total Enrollment): Biological Engineering                                             90.9%                                    84.4% Accounting                                                             83.9%                                    84.1% Special Education & Rehabilitation                            88.3%                                    83.6% Civil & Environmental Engineering                            85.6%                                    83.2% Nutrition, Dietetics & Food Science                           80.1%                                    80.4% Sixteen departments achieved response rates of 75% or better (Total Responses/Total Enrollment). Forty “departments” (including Honors as a department) achieved response rates of 65% or better (Total Responses/Total Enrollment). In summary: USUE is going to need quite a bit of additional work, but we’re looking pretty good (on average) for Logan and RCDE. 11/13/2018 IDEA Train the Trainer Workshop - Day 2

Questions & Discussion 11/13/2018

Avg. Score vs Converted Avg. Left Hand Side = Average Scores (Scale = 1-5) Used for “Criterion Reference” administrative evaluation Right Hand Side = Converted Average Scores, or “T Scores.” Scale is 20-80, with 50 as the mean (average) “USU has historically used a comparative process”

Converted Average (T Score) Compares Your score to the average score of a normative group Standardizes scores with different averages and standard deviations Average = 50 Std. Deviation = 10 Walk them through the converted average. Large rectangle is “IDEA Database.” Smaller rectangle is “Discipline Database.” Note different mid-point/average, and impact on the T Score 4.2 4.2+3.7+2.5+4.7+3.4+4.1+4.0+3.7+4.3+2.7+3.5+3.1+4.3+4.2+4.1+4.6+3.9+4.6+4.1+4.2+4.4+4.1+3.8+3.9+4.0+4.3+3.6+4.1+3.6+4.3+2.7+3.9+4.0+4.2+4.8+2.9+3.1+4.6+4.0+3.1 / 40 4.0 (Avg) = 50 4.1 (Avg) = 50 = 52 = 54

Converted Average Walk through the animation and describe

Questions & Discussion 11/13/2018

Learning Objectives NA -- can’t combine those average scores because each item has a different mean and standard deviation. Adjustments - even the playing field - which does mean the you can have a lower score - especially in small classes, highly motivated students. It is clear from this table that “Work Habits” (WH, mean of Item 43) was generally the most potent predictor, followed by “Course Motivation” (CM, mean of Item 39). Classes that contained students who typically worked hard on their studies and/or were highly motivated to take the course regardless of who taught it were expected to receive favorable ratings; unless ratings were adjusted, the instructors of such classes would have an unfair advantage over colleagues with less motivated and dedicated students.   The regression coefficient for “Enrollment” (N) was not always statistically significant; but when it was, it was always negative. The larger the class, the lower the predicted (expected) rating. Those teaching small classes have an advantage over those teaching large classes; hence, in the interest of fairness, ratings should be adjusted to take this into account. Except for the first two criterion ratings, the regression coefficient for DN was always negative. Generally, if the discipline was perceived as difficult (after taking into account the impact of the instructor on perceived difficulty), an attenuated outcome can be expected. This was especially apparent in progress ratings on “Creative capacities” and “Communication skills” where high difficulty was strongly associated with low progress ratings. The two exceptions, where “disciplinary difficulty” had a positive effect on the predicted outcome, were for the progress ratings concerned with basic cognitive development (“Factual knowledge” and “Principles and theories”). In these instances, conventional wisdom (high difficulty = low ratings) was solidly refuted. In most cases, student effort in the class (adjusted for the instructor’s influence on effort) was also negatively related to predicted ratings. Classes containing an unusually large number of students who worked harder than the instructor’s approach required ended up with lower progress ratings. The reason for this is unclear. Perhaps those who found it necessary to put in extra effort were those whose backgrounds did not prepare them well for the class. They may also be students who lack self-confidence and, for this reason, under-achieve (or under-estimate their progress in a self-abasing manner). Decide which of 4 areas is most important - if #1, need to stress careful selection of the objectives

Pg 2: Progress on Objectives 4.5+4.5 4.4+4.4 4.2+4.2 +3.2 7 The 5-point averages of progress ratings on “Essential” or “Important” objectives vary across objective. For instance, the average for “gaining factual knowledge” is 4.00, while that for “gaining a broader understanding and appreciation for intellectual/cultural activity is 3.69. Essential weighted twice, important weighted once. Questions: On which objectives are students reporting the most progress? Which objectives might this instructor want to think about and why? What objectives are students reporting the most progress? What objectives might this instructor want to think about and why? Now…look at your own report and answer those two questions.

Too Many Objectives Select only 3 to 5 Objectives 11/13/2018 Too Many Objectives Select only 3 to 5 Objectives When selected from the list of 12 Learning Objectives, both Important & Essential count Fall 2011: 145 Courses with all 12 learning objectives selected More than 800 classes had more than five learning objectives selected as either “Important,” or “Essential.” 11/13/2018 IDEA Train the Trainer Workshop - Day 2

Statistical Detail

Questions & Discussion 11/13/2018

What is “Summary Evaluation” 11/13/2018 What is “Summary Evaluation” Default Weights 50% 25% 25% 11/13/2018 IDEA Train the Trainer Workshop - Day 2

IDEA: Adjusted Scores NA -- can’t combine those average scores because each item has a different mean and standard deviation. Converted scores are T-scores (Mean = 50, SD = 10). Adjustments - even the playing field - which does mean the you can have a lower score - especially in small classes, highly motivated students. It is clear from this table that “Work Habits” (WH, mean of Item 43) was generally the most potent predictor, followed by “Course Motivation” (CM, mean of Item 39). Classes that contained students who typically worked hard on their studies and/or were highly motivated to take the course regardless of who taught it were expected to receive favorable ratings; unless ratings were adjusted, the instructors of such classes would have an unfair advantage over colleagues with less motivated and dedicated students.   The regression coefficient for “Enrollment” (N) was not always statistically significant; but when it was, it was always negative. The larger the class, the lower the predicted (expected) rating. Those teaching small classes have an advantage over those teaching large classes; hence, in the interest of fairness, ratings should be adjusted to take this into account. Except for the first two criterion ratings, the regression coefficient for DN was always negative. Generally, if the discipline was perceived as difficult (after taking into account the impact of the instructor on perceived difficulty), an attenuated outcome can be expected. This was especially apparent in progress ratings on “Creative capacities” and “Communication skills” where high difficulty was strongly associated with low progress ratings. The two exceptions, where “disciplinary difficulty” had a positive effect on the predicted outcome, were for the progress ratings concerned with basic cognitive development (“Factual knowledge” and “Principles and theories”). In these instances, conventional wisdom (high difficulty = low ratings) was solidly refuted. In most cases, student effort in the class (adjusted for the instructor’s influence on effort) was also negatively related to predicted ratings. Classes containing an unusually large number of students who worked harder than the instructor’s approach required ended up with lower progress ratings. The reason for this is unclear. Perhaps those who found it necessary to put in extra effort were those whose backgrounds did not prepare them well for the class. They may also be students who lack self-confidence and, for this reason, under-achieve (or under-estimate their progress in a self-abasing manner). Decide which of 4 areas is most important - if #1, need to stress careful selection of the objectives

Impact of Extraneous Factors Gaining Factual Knowledge – Average Progress Ratings Work Habits (Item 43) Student Motivation (Item 39)   High Avg. Low 4.48 4.38 4.28 4.13 4.04 High Avg. 4.29 4.14 3.96 3.76 Average 4.01 3.83 3.64 Low Avg. 4.15 4.05 3.88 3.70 3.51 4.11 3.78 3.58 3.38 Technical Report 12, page 40

Understanding Adjusted Scores This class has average lower motivation to take course and higher typical effort (self-perceptions).

When to Use Adjusted Scores What is the focus? Do unadjusted (raw) scores meet or exceed ex- pectations?* Are adjusted scores lower or higher than unadjusted scores? Use scores (raw) scores Student outcomes Instructor contributions Lower Yes Higher No *Expectations defined by your unit.

Adjusted or Unadjusted Scores What Should be Used? 11/13/2018 Adjusted or Unadjusted Scores What Should be Used? Did faculty meet expectations for effective teaching based on raw scores OR adjusted scores? Adjusted ratings – special consideration If unadjusted ratings “meet expectations,” and adjusted ratings are lower, use unadjusted scores Cutoff determined by each institution Don’t want to penalize a faculty member for having highly motivated students with good work habits If you choose to use adjusted scores, there is one situation that merits special consideration. If you have a class whose raw scores are above 4.0 or 4.2, and their adjusted scores are lower, we recommend that you consider using the unadjusted scores. In these cases, the lower adjustments are typically due to classes with students who are motivated to take them and have good work habits. Instructors typically cannot select the students enrolling in their courses and you probably don’t want to penalize them for this. Your campus needs to determine the specific cut-off scores, but this is an important consideration. IDEA Train the Trainer Workshop - Day 2

Questions & Discussion 11/13/2018

IDEA Recommendations for Administrative Use of Scores Results are not over-interpreted 3-5 categories (Similar, Lower, Higher, Much Lower, Much Higher) Review a “set” of data (6-8 classes) More data if small samples (n<10) IDEA recommends that evaluation scores be used for no more than 30%-50% of summative evaluation The IDEA system, and student ratings in general, can have a positive impact if they are not overemphasized in the evaluation process. Because student ratings do not perfectly measure teaching effectiveness, using them for more than 30-50% of the overall evaluation of teaching can be seen as punitive. Using them appropriately will encourage faculty trust in the process, which will facilitate effective use of the information for teaching improvement.

Questions & Discussion 11/13/2018

Reflective Practice with IDEA Collect Feedback Interpret Results Read & Learn Reflect & Discuss Improve IDEA Student Performance Peers/Mentors Other Sources IDEA Knowledge Center POD-IDEA Center Notes IDEA Papers Other Resources

IDEA: Diagnostic

Final Discussion 11/13/2018