“Excuse Me Sir, Here’s Your Change”

Slides:



Advertisements
Similar presentations
TWS Aid for Supervisors & Mentor Teachers Background on the TWS.
Advertisements

The Teacher Work Sample
A GUIDE TO CREATING QUALITY ONLINE LEARNING DOING DISTANCE EDUCATION WELL.
The Research Consumer Evaluates Measurement Reliability and Validity
Designs to Estimate Impacts of MSP Projects with Confidence. Ellen Bobronnikov March 29, 2010.
Where Have We Been? Where Are We Going? Using Student Surveys to Assess and Improve Literature Courses Kelly Douglass, PhD Asst. Professor, English Riverside.
Learning Objectives, Performance Tasks and Rubrics: Demonstrating Understanding and Defining What Good Is Brenda Lyseng Minnesota State Colleges.
MSP Evaluation Rubric and Working Definitions Xiaodong Zhang, PhD, Westat Annual State Coordinators Meeting Washington, DC, June 10-12, 2008.
Supporting Quality of Student Learning Online: Using Quality Matters to Strengthen Online Teaching and Learning Valencia College - Orlando, Florida Charles.
The Academic Assessment Process
Quantitative Research
Making a difference? Measuring the impact of an information literacy programme Ann Craig
NCCSAD Advisory Board1 Research Objective Two Alignment Methodologies Diane M. Browder, PhD Claudia Flowers, PhD University of North Carolina at Charlotte.
The Effect of Quality Matters™ on Faculty’s Online Self-efficacy DLA Conference 2010 Jim Wright, Ed.S. June 9, 2010.
OUTCOMES ASSESSMENT VIA RUBRICS: A PILOT STUDY IN AN MIS COURSE AS A PRECOURSOR TO A MULTIPLE MEASURE APPROACH By W. R. Eddins, York College of Pennsylvania.
Comp 20 - Training & Instructional Design Unit 6 - Assessment This material was developed by Columbia University, funded by the Department of Health and.
TWS Aids for Student Teachers & Interns Overview of TWS.
The Use of Distance Learning Technology by Business Educators for Credentialing and Instruction Christal C. Pritchett, Ed.D. NABTE Research Session Anaheim,
Information Systems Use Among Ohio Registered Nurses: Testing Validity and Reliability of Nursing Informatics Measurements Amany A. Abdrbo, RN, MSN, PhD.
+ General Education Assessment Spring 2014 Quantitative Literacy.
Evaluating Impacts of MSP Grants Hilary Rhodes, PhD Ellen Bobronnikov February 22, 2010 Common Issues and Recommendations.
Evaluating Impacts of MSP Grants Ellen Bobronnikov Hilary Rhodes January 11, 2010 Common Issues and Recommendations.
Outcomes for Mathematical Literacy: Do Attitudes About Math Change?
Evaluating Impacts of MSP Grants Ellen Bobronnikov January 6, 2009 Common Issues and Potential Solutions.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
Qatar Comprehensive Educational Assessment (QCEA) 2008: Discussion Session For QCEA Support.
1 Information Systems Use Among Ohio Registered Nurses: Testing Validity and Reliability of Nursing Informatics Measurements Amany A. Abdrbo, RN, MSN,
COUNSELOR EDUCATION PEDAGOGY TRAINING Session One: Significant Learning and Counselor Education.
How can I use a digital library to support my teaching? Find good resources to enhance existing curriculum  Search special collections aimed at your interests.
Course Work 2: Critical Reflection GERALDINE DORAN B
The Basics of Social Science Research Methods
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov March 23, 2011.
Presenter: E. John Shinsky, Ph.D. Grand Valley State University
Office of Planning, Research, and Institutional Effectiveness
Reliability Analysis.
A community of learners improving our world
Cari-Ana, Alexis, Sean, Matt
The effects of physical activity on third grade math scores
Prebriefing: The Final Frontier
Phyllis Lynch, PhD Director, Instruction, Assessment and Curriculum
An Overview of the EQuIP Rubric for Lessons & Units
Evaluation of An Urban Natural Science Initiative
Concept of Test Validity
Personal Assessment of the College Environment (PACE)
Herding Cats: The Initial Roundup
Governor’s Teacher Network
Dr. Nina O’Brien Department of Management
Welcome to the CSBM workshop: Managing office services
2015 PARCC Results for R.I: Work to do, focus on teaching and learning
Chapter Six Training Evaluation.
Overview of the Teacher Work Sample (TWS)
Finding Answers through Data Collection
Conclusion & Implications
An Overview of the EQuIP Rubric for Lessons & Units
This presentation will include:
Providing Customized Training on Quality Online Design and Delivery
Jennifer Bryer PhD, RN, CNE Virginia Peterson-Graziose DNP, RN, CNE
Necessary but Not Sufficient: Why Strategic Implementation Climate and Molar Organizational Climate Both Matter for EBP Implementation Nate Williams, PhD,
Indiana University School of Social Work
Reliability Analysis.
Leanne Havis, Ph.D., Neumann University
The impact of small-group EBP education programme: barriers and facilitators for EBP allied health champions to share learning with peers.
Using Data to Assess High Quality Professional Development Findings From Interviews With Massachusetts Educators May 2015 Massachusetts Department.
What to do with your data?
Cuyamaca College Library
Evaluating Life Skills Development Through 4-H Challenge Course Outcomes VanderWey, S. Cooper, R. North American Association for Environmental Education.
Deanna L. Morgan The College Board
Misc Internal Validity Scenarios External Validity Construct Validity
Doug Glasshoff Greg Gaden Darin Kelberlau
The Teacher Work Sample: An Authentic Assessment
Presentation transcript:

“Excuse Me Sir, Here’s Your Change” Cognitive and Affective Effects of the Applying the Quality Matters Rubric Workshop René E. Mercer Ph.D (c) Instructional Design Specialist

Learning Objectives Discuss and evaluate an instrument used to measure faculty’s knowledge of best practices in online course design; assess faculty’s perception of online course quality and willingness to use the QM rubric. Identify methods used to maximize reliability of scores and provide evidence of content validity Identify cognitive and affective outcomes of faculty who completed the APPQMR workshop Interpret results to identify evidence of which of the QM standards the APPQMR training could provide the largest learning gains Use outcomes to inform future institutional plans for providing professional development for online faculty. Suggest options for future refinement and implementation of the instrument

Professional Development in Online Learning A large portion of faculty must design and deliver the courses they teach (Wright, 2012). Most online teachers are self-taught (Ray, 2009; Weaver, Robbie, & Borland, 2008) Professional Development increases confidence in online teaching (Powell, 2010; Wright, 2011) critical thinking skills in assessing online course quality (Reilly et al., 2012) Faculty’s desire to improve online course quality (Reilly et al., 2012) More generalizable research is needed for professional development trainings focused on quality in online course design (Ellis & Phelps, 2000)

QM Focused Research Improved learning outcomes (Swan & Matthews, 2012) Increased self-efficacy (Wright, 2012) Improved quality of courses (Effken et al., 2009; Pollacia & McAllister, 2009; Swan & Matthews, 2012) Professional growth (Bento & White, 2010; Swan & Matthews, 2012) Increased student evaluation (Bento & White, 2010) QM Rubric is a useful tool (Effken et al., 2009; Reif, 2009)

Research Problem and Solution Problem: More studies are needed to determine the extent to which completing professional development workshops such as Quality Matters increase faculty knowledge of best practices in online course design, perception of online course quality, and willingness to use the QM rubric. Solution: Measure the effectiveness of one professional development workshop called Applying the Quality Matters Rubric (APPQMR) using criterion-based and self-report measures

Why APPQMR? Quality Matters is recognized nationally and world-wide Quality Matters’ flagship workshop Focuses solely on course design via the QM Rubric Quality in online learning begins with course design (Pollacia, Russell, & Russell, 2009; Swan & Matthews, 2012) Covers all 8 QM General Review Standards Workshop objective: Apply specific review standards of the QM rubric to an online course and make decisions as to whether the course meets the standards. Hands-on activities covering 21 specific review standards.

Research Question What are the effects of the Applying the Quality Matters™ Rubric (APPQMR) workshop on faculty’s: knowledge of best practice in online course design perception of the quality of their own online course, and willingness to use the Quality Matters Rubric to redesign or design an online course?

Context RU – VH Institution Faculty who design and deliver online courses Free workshop in online course design Face-to-Face (two-days) Campus wide Sponsored by College of Education & Human Development May 2013 Part of a larger Mixed- Methods study

Methods & Procedures One group Pre-Post Online Pre-Questionnaire Faculty who design and deliver online courses Online Pre-Questionnaire Available One week prior to workshop session Facilitate four sessions of the APPQMR workshop over 2 week period Post-Questionnaire Online After each second day session

Participants Sample characteristics Roles n = 25 Gender Race 16% Male 84% Female Race 8% Asian 12% African American 16% Hispanic 64% White Roles Tenure/non-tenure track Adjunct Lecturers Staff

Instrumentation Three Part Questionnaire Part 1: Knowledge of best practice in online course design (KBP) Nine theorized constructs 8 General Standards Alignment *28 questions – 4 item multiple choice Part 2: Perception of online course quality (IPQ) *8 questions – 4 point Likert scale Scale: Strongly disagree, disagree, agree, strongly agree, not sure (not scored) Part 3: Willingness to use QM Rubric (W) *6 questions – 6 point Likert scale Scale: Strongly disagree, disagree, somewhat disagree, somewhat agree, agree, strongly agree Three Part Questionnaire

Sample Item Which of the following learning objectives is not written according to standards of best practice? Students will be able to understand the differences between theoretical research and applied research Students will write sentences that demonstrate correct use of commas, semicolons, and periods The student will be able to distinguish between characteristics of a square, a rhombus, and a rectangle Students will be able to describe the events that led to the American Civil War Part 1, General Review Standard 2

More Sample Items My course demonstrates a commitment to accessibility for all students Part 2, General Review Standard 8 If given adequate resources, I would be willing to work to use the Quality Matters rubric to ensure my online course met standards of quality assurance. Part 3, Willingness

Score Validity & Reliability Part 1: Knowledge of Best Practice in Online Course Design Content validity Three Quality Matters Experts provided formative content evaluation Expertise included varying levels of QM experience Score reliability Reliability of .690 (Cronbach’s Alpha)

Data Analysis Computed average scores for dependent variables Accounted for missing data and “not sure” Syntax: Conducted Paired t-tests Calculated statistical significance at p < .01 Calculated Effect size (η2) when appropriate

Results See handout for full results KBP Paired T-test was statistically significant p < .0001 η2 = .43 IPQ & W Paired T-tests were not statistically significant p = .970 (IPQ) P = .555 (W) See handout for full results

Results per General Standard Note. Full table provided in handout. * Statistically significant result at p < .01.

Conclusions Providing professional development such as the APPQMR workshop can increase knowledge of best practices in online course design. Providing one professional development course such as APPQMR may not be sufficient in helping faculty assess the quality of their own courses Faculty who took the APPQMR workshop may have already been willing to use the QM rubric because they were not required to participate.

Recommendations Increase sample size Use multiple institutions Focus on Knowledge of Best Practice Use multiple institutions Use experimental or quasi- experimental designs Solomon four-group Refine the questions Avoid “not” and “except” Include more input from experts Repeat testing over multiple professional development offerings in Quality Matters

Questions?