Potential Biases in Student Ratings as a Measure of Teaching Effectiveness Kam-Por Kwan EDU Tel: 2766 6287 E-mail: etkpkwan.

Slides:



Advertisements
Similar presentations
Stephen C. Court Presented at
Advertisements

Christine A. Draper Michelle Reidel Georgia Southern University.
Developing a feedback questionnaire: Principles and steps Workshop for NHS staff 28 Dec 1999 (Tuesday) Kam-Por Kwan, EDU
Wynne Harlen. What do you mean by assessment? Is there assessment when: 1. A teacher asks pupils questions to find out what ideas they have about a topic.
Action Research Not traditional educational research often research tests theory not practical Teacher research in classrooms and/or schools/districts.
Victorian Curriculum and Assessment Authority
Alternative Strategies for Evaluating Teaching How many have used end-of-semester student evaluations? How many have used an alternative approach? My comments.
Perceptions of English 110 Christina DeLapp, Michael Jobb & Shevaun Watson English Department, University of Wisconsin-Eau Claire “I’d like more meaningful.
Copyright © 2011 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 12 Measures of Association.
Student Consensus on RateMyProfessors.com April Bleske-Rechek, Amber Fritsch, and Brittany Henn University of Wisconsin Eau Claire Background Method Discussion.
Making the Most of Teacher Evaluation, Charlotte Danielson 1 Making the Most of Teacher Evaluation Charlotte Danielson
Middle Level Education Kyrene School District November 2007.
1 Peer Review of Teaching K.P. Kwan & J. Jones, EDU etkpkwan etjjones x 6287 x6320 EDU Seminar 7 May 1999 (Friday)
Principles of High Quality Assessment
Research Methods Chapter 1. Behavioral Research Behavioral Medicine Communication Criminology Human Development Education Psychology Sociology.
Inspiration Day 2007 To Evaluate or not Evaluate (Our Teaching), That is the Question.
Chapter 7 Correlational Research Gay, Mills, and Airasian
Grade 12 Subject Specific Ministry Training Sessions
ISD 200 Week 5 CRITICAL THINKING IMMIGRATION INSTRUCTORS: Mari Acob-Nash Ryan Packard.
Introduction to NYC Teacher Data Initiative Training for Schools Fall 2008.
Decide whether the following statements are true or false.
Analytical Thinking.
International Conference on Enhancement and Innovation in Higher Education Crowne Plaza Hotel, Glasgow 9-11 June 2015 Welcome.
July 28, 2011 Nancy Nation & Mirna Whidden.  What do you associate with the word assessment??
(Business Research Methods)
METHODS Study Population Study Population: 224 students enrolled in a 3-credit hour, undergraduate, clinical pharmacology course in Fall 2005 and Spring.
LOGO Teacher evaluation Dr Kia Karavas Session 5 Evaluation and testing in language education.
RATING THE PROFESSOR TA & PFF Workshop– Spring 2011 Artze-Vega.
Classroom Assessment A Practical Guide for Educators by Craig A
Office of Institutional Research, Planning and Assessment January 24, 2011 UNDERSTANDING THE DIAGNOSTIC GUIDE.
AFT 7/12/04 Marywood University Using Data for Decision Support and Planning.
Student Engagement Survey Results and Analysis June 2011.
What’s in the news right now related to science???? Flesh eating bacteria.
Classroom Assessment A Practical Guide for Educators by Craig A
Scientific Inquiry & Skills
S.T.E.M. Science Technology Engineering Mathematics What does S.T.E.M. stand for?
A Preliminary Investigation of Student Perceptions of Online Education Angela M. Clark University of South Alabama Presented at ISECON 2003 San Diego,
Main components of effective teaching. SRI validity A measuring instrument is valid if it measures what it is supposed to measure SRIs measure: Student.
Background Method Discussion References Acknowledgments Results  RateMyProfessors.com (RMP.com) launched in 1999 as an avenue for students to offer public.
Teacher Engagement Survey Results and Analysis June 2011.
SINTEF Telecom and Informatics EuroSPI’99 Workshop on Data Analysis Popular Pitfalls of Data Analysis Tore Dybå, M.Sc. Research Scientist, SINTEF.
Performance-Based Assessment HPHE 3150 Dr. Ayers.
C R E S S T / U C L A UCLA Graduate School of Education & Information Studies Center for the Study of Evaluation National Center for Research on Evaluation,
© 2008 Brigham Young University–Idaho Course Evaluations at BYU-Idaho 1.
Sampling and Probability Chapter 5. Sampling & Elections >Problems with predicting elections: Sample sizes are too small Samples are biased (also tied.
JS Mrunalini Lecturer RAKMHSU Data Collection Considerations: Validity, Reliability, Generalizability, and Ethics.
Using Journals to Build Information Literacy Skills Violet H. Harada University of Hawaii ACEI Annual International Conference April.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 4 Designing Studies 4.2Experiments.
BACKWARD DESIGN AND ASSESSMENT Course Design Intensive ~ Dr. Catherine D. Rawn June 2014 This work is licensed under the Creative Commons Attribution-NonCommercial.
Spring 2012 Writing 20:Ocean Acidification February 21, 2011 researching & developing a claim for MP2 Much of this material is compiled from:
Producing Data: Experiments BPS - 5th Ed. Chapter 9 1.
DEVELOPED BY MARY BETH FURST ASSOCIATE PROFESSOR, BUCO DIVISION AMY CHASE MARTIN DIRECTOR OF FACULTY DEVELOPMENT AND INSTRUCTIONAL MEDIA UNDERSTANDING.
© 2001 Laura Snodgrass, Ph.D.1 Experimental Psychology Introduction.
Ethics in Sports and Physical Education: A Source of Role Conflict.
Educating teachers about values in mathematics education: Alan J. Bishop Faculty of Education Monash University Melbourne Australia
TA & PFF Workshop– Spring 2011 Artze-Vega
Classroom Observations
Assessing Musical Behavior
Classroom Grading: A Summative Evaluation
Do My Grading Practices Support Learning?
Chapter Statistics and Probability
Derek Herrmann & Ryan Smith University Assessment Services
Classroom Observations
The Efficacy of Student Evaluations of Teaching Effectiveness
Nature of Science Dr. Charles Ophardt EDU 370.
Significance Tests: The Basics
Unit 7: Instructional Communication and Technology
Validity and Reliability II: The Basics
Student Evaluations of Teaching (SETs)
EDUC 2130 Quiz #10 W. Huitt.
Presentation transcript:

Potential Biases in Student Ratings as a Measure of Teaching Effectiveness Kam-Por Kwan EDU Tel: etkpkwan

Your beliefs about student ratings Answer the 10 questions on the worksheet by stating whether you think each of the statements is ‘True’ or ‘False’. (Click here for the questions in PDF format)here There is no right or wrong answer. The main purpose is to find out more about your own beliefs and views about student ratings of teaching.

Uses of student ratings As a measure of student perceptions or satisfaction about the teaching As a measure of teaching effectiveness of the instruction As a measure of teaching effectiveness of the teacher

Validity concerns about ratings Do ratings accurately reflect the perceptions and satisfaction of the students? Do ratings accurately reflect the teaching effectiveness of the instructional context? Do ratings accurately reflect the teaching effectiveness of the teacher?

Ratings as student perceptions Biases exist if the ratings fail to measure accurately what students really feel about the teaching u Little dispute in the validity of ratings for this use

Ratings as TE of the instruction Biases exist if ratings are influenced by factors unrelated to teaching effectiveness u Lower ratings for larger classes  no bias u Lower ratings for ‘better-looking’ teachers  existence of bias

Ratings as TE of the teacher Biases exist if ratings influenced by factors beyond the control of the instructor u Higher ratings for more emphatic teachers  possibly no bias u Higher ratings for teachers of smaller class  potential bias or unfairness

Criticisms of student ratings Students cannot make consistent judgments about teaching Student ratings are popularity contests Students cannot make accurate judgments until they have graduated Student ratings are unrelated to amount of learning Staff and students disagree on what constitutes good teaching

Some more criticisms Student ratings are influenced by u time and day of the teaching u class size u level of the course u rank of instructor u nature of the course: required or elective u difficulty level of course / assignments u expected grades u disciplinary differences

What research evidence is there? Over 70 years of research More than 2000 studies A huge body of research evidence Some of the criticisms are valid, some are not

Ratings are reasonably reliable Well-constructed student rating forms are highly reliable (alphas in the 0.8 / 0.9 range) Ratings are stable over time (r > 0.8) High correlation exists between ratings of same instructor and course (r = 0.7 to 0.89)

Ratings are reasonably valid Staff and students generally agree on the important dimensions of good teaching Student ratings are moderately correlated with achievement (r = about 0.5) Student ratings correlate moderately with alumni ratings, classroom observations, and self-evaluation by staff Distinguishable ‘profiles’ of teaching can be revealed from student ratings

Course and teacher effects Small associations are found between student ratings and the following factors: u class size (ratings > for smaller classes) u level of course (ratings > for higher-level courses) u nature (ratings > for elective courses) u discipline (languages & art > social sciences > engineering & science) u rank (ratings > for higher-rank teachers)

Course and teacher effects (2) Inconsistent / no associations are found between student ratings and: u gender of instructor u day and time of the course

Student factors Small associations found between student ratings and: u prior subject interest (ratings > for higher prior interest) u students’ major (ratings > for major) u perceived workload / difficulty (ratings > for higher workload / more difficult courses) u expected grades (ratings > for higher grade)

Overall effect About 15 to 20 percent of the variation in ratings can be explained by the combined effects of the background (course, teacher and student) variables

Ratings are basically unbiased The existence of course, teacher and student effects on ratings generally support rather than refute the validity of student ratings as a measure of teaching effectiveness In most cases, the effects are quite small

Ratings can be unfair Teacher evaluation based on raw student ratings can be ‘unfair’ to individual teachers because: u a lot of factors affecting teaching effectiveness are outside the control of the teacher (although the effects are quite small) u there are big differences in the context in which different teachers operate

Conclusion Student ratings are generally valid as a measure of student perceptions/ satisfaction Student ratings are reasonably valid as a measure of teaching effectiveness of instruction Student ratings can be unfair when they are used for making personnel decisions concerning individual teachers

Implications Student ratings are useful for u understanding student satisfaction and perceptions u improving teaching Student ratings must be interpreted and used cautiously and in context when used for making judgements and personnel decisions about individual teachers

References Aleamoni, L. M. (1987). Student rating myths versus research facts. Journal of Personnel Evaluation in Education, 1, Feldman, K.A. (1996). Identifying exemplary teaching: using data from course and teacher evaluations. New Directions for Teaching & Learning, 65. Jossey-Bass Publishers. Marsh, H.W. (1987). Students’ evaluation of university teaching: research findings, methodological issues, and directions for future research. Int. J. Edu Res., 11,

Answers to T/F Questions 1. F 2. F 3. F 4. T 5. F 6. F 7. T 8. F 9. T 10. T