Time to Go Online!
Why go online? Paper vs. Online
Our current paper process costs about $25,000 per year, plus staff time. An online process would be virtually free— no processing costs and very little staff time after first iteration.
The timing of paper evaluations is not optimal, but must be conducted in mid- November to allow processing time. Online evaluations could be conducted anytime during the last two weeks of class.
Our current paper process takes considerable staff time and effort: generating course lists soliciting and approving two sections per faculty member printing and sorting forms distributing, collecting, and logging forms. An online process would eliminate nearly all processing time.
Current paper process is not available for evaluation of part-time faculty—time and cost prohibitive. An online evaluation process would be available to both full- and part-time faculty and would allow us to use the same instrument for both.
Current paper process allows for evaluation of only two sections per full- time faculty member per year—a sample size of about 20%. An online process would allow for evaluation of every class, every faculty member, every semester—a sample size of about 53% given average return rates.
Results from paper evaluations are slow and distribution of results is unwieldy. Online evaluation results would be available as soon as grades are submitted and directly accessible by faculty.
Our paper process uses about 95,000 sheets of paper per year—not to mention a whole bunch of large envelopes. An online process is environmentally responsible.
Myth #1: Return rates on online course evaluations are abysmally low. Fact: The average return rate at other universities is about 53%. That compares to an average paper return rate of 78% and a current MC sample rate of less than 20%.
Myth #2: If evaluations are online, only the outliers (the very satisfied or very dissatisfied) will respond. Fact: Other institutions report that even when response rates are lower, the faculty member’s average rating does not change.
Myth #3: If evaluations are online, students will rush through them even more than they do with paper evaluations. Fact: Research shows that, when provided a free response section, students write longer comments on electronic forms than on paper forms.
Myth #4: Many of our students don’t have access to computers for completing their evaluations. Fact: A 2003 report indicated 92% of Maryland households had computer access to the Net. Plus, students can always use a lab on campus.
Myth #5: MC will have to use draconian measures, such as withholding final grades, to get students to complete the online evaluations. Fact: Many schools are getting very respectable return rates with awareness campaigns, friendly competitions, incentives, and other positive strategies.
What is the best timeline for implementing this change? Would it be possible to reduce the number of evaluation questions from 25 to 10? If so, what is the best mechanism for selecting the 10 most important questions? What strategies should we use to increase return rates? What should we avoid?
What quality safeguards or comparative data would make faculty more comfortable with this change? What are the best venues for soliciting faculty suggestion? What other question or concerns should we be addressing?
Please contact your dean with ideas, concerns, or other input!
Corragio, James, and Magaly Tymms. “Transiting to an Online Course Evaluation Model: The Online Student Survey of Instruction.” St. Petersburg College. St. Petersburg, FL. February “Facilitating Response Rates in IDEA Online.” IDEA Center. Manhattan, KS. August %20Online%20Response%20Rate.pdf %20Online%20Response%20Rate.pdf “Information on University-wide Course Evaluations.” University of Maryland. College Park, MD
Miller, Mary Helen. “Online Evaluations Show Same Results, Lower Response Rates.” The Chronicle of Higher Education. 6 May “Online Faculty and Course Evaluation FAQ.” Ball State University. Muncie, IN cms.bsu.edu/About/AdministrativeOffices/Provost/FacResources/CrseResp onseFAQs.aspx “Pilot Study for Assessing the Viability of Using Online Course Evaluations at California State University Sacramento.” Sacramento, CA. 8 October
Sorenson, Lynn, and Trav Johnson. “Online Student Ratings of Instruction.” Brigham Young University. Salt Lake City, UT. 12 April Thorpe, Stephen W. “Online Student Evaluation of Instruction: An Investigation of Non-Response Bias.” Paper presented at the 42nd Annual Forum of the Association for Institutional Research. Toronto, Canada. June,
Question or Concerns?