The Good The Bad & The Ugly Real-Life Examples of the SLO Assessment Report Form With Tips on How to Complete It August 21, 2012.

Slides:



Advertisements
Similar presentations
Using Embedded Assessment to Collect Evidence of Student Learning Using Embedded Assessment to Collect Evidence of Student Learning.
Advertisements

What is Assessment? The university, with the support and encouragement of President Hitt and the Vice President team, engages in the institutional effectiveness.
Developing an Outcomes Assessment Plan. Part One: Asking a Meaningful Question OA is not hard science as we are doing it. Data that you collect is only.
How to Document the Assessment. Student Learning Outcomes Assessment Report Once the SLO has been selected and assessment has taken place, faculty must.
Outcomes Assessment- Full Implementation Meeting Fall 2009.
Using Embedded Assessment to Collect Evidence of Student Learning Using Embedded Assessment to Collect Evidence of Student Learning.
Freehold Borough Teacher Evaluation System Freehold Intermediate School Friday – February 15, 2013 Rich Pepe Director of Curriculum & Instruction.
Student Learning Targets (SLT)
Student Learning Targets (SLT) You Can Do This! Getting Ready for the School Year.
Bill Zannini Business Programs Coordinator October 27, 2008.
STUDENT LEARNING OUTCOMES ASSESSMENT. Cycle of Assessment Course Goals/ Intended Outcomes Means Of Assessment And Criteria For Success Summary of Data.
Student Growth Objective (SGO) Evaluating SGO Quality
Making Your Assessments More Meaningful Flex Day 2015.
Annual Professional performance review (APPR overview) Wappingers CSD.
MATH ASSESSMENT TEST OCMA May, HISTORY OF MAT Test originally developed in late 60’s.
My Policies and Some Advice for Doing Well in this Course.
GOAL SETTING CONFERENCES BRIDGEPORT, CT SEPTEMBER 2-3,
How to Use Data to Improve Student Learning Training Conducted at Campus-Based SLO Summit Spring 2014 How to Use Data to Improve Student Learning.
1 Student Success Plans Regional Meeting February 9, 2007 Youngstown State University Office of Assessment Sharon Stringer
Assessment Report School of The Arts, Humanities and Social Sciences________________ Department: Political Science and International Studies.
Student Learning targets
2012 Secondary Curriculum Teacher In-Service
Closing the Loop From Analysis to Use of Results SLO Winter 2011 Workshop.
How to Administer Constructive and Effective Feedback to Online Students Denielle R. Vazquez, M.S.Ed – 2014 Teaching and Learning.
CCI Bootcamp Back to School Staff Development Rotation.
Assessment Report School of The Arts, Humanities and Social Sciences Department: Theatre and Music Studies Chair: P. Gibson Ralph Assessment.
The Genetics Concept Assessment: a new concept inventory for genetics Michelle K. Smith, William B. Wood, and Jennifer K. Knight Science Education Initiative.
Information for school leaders and teachers regarding the process of creating Student Learning Targets. Student Learning targets.
Lack of Learning or Lack of Studying? An Inquiry into Low Exam Scores Katherine M. Sauer Metropolitan State College of Denver February.
Closing the Loop From Analysis to Use of Results SLO Winter 2011 Workshop.
Using a Rubric CCC Assessment-Fall 2007 Marla Larimore and Don Lind Learning Assessment Coordinators.
Assessment Report School of The Arts, Humanities, and Social Sciences Department: Sociology Chair: Julie M. Ford Assessment Coordinator: Elliot.
WHAT IS THE WPE? Writing Assessment Office Office of Undergraduate Studies Campus Center 1st Floor, Room
My Course Was Selected For LAE Assessment Reporting. What Do I Do Now??? LAE Assessment.
Assessing Student Learning to Improve Teaching Jeff Bell and Jim Bidlack.
Evaluation of Shoreline Science Jia Wang & Joan Herman UCLA Graduate School of Education & Information Studies National Center for Research on Evaluation,
Richard Beinecke, Professor and Chair Suffolk University Institute for Public Service.
Fall 2015 Professional Development Days C. Cruz-Johnson & R. Gamez August 28, Walking with Integrity Towards Student Success.
Department of Secondary Education Program Assessment Report What We Assessed: Student Learning Outcomes (SLOs) and CA State Teaching Performance.
Assessed: 2007, 2010, 2011,  PHIL 101 (Introduction to Philosophy: Ethics)  GE elective choice  BA 300 (Ethical Decision Making in Business)
Grading and Analysis Report For Clinical Portfolio 1.
THE SLO PROCESS #1 create/update SLO’s, rubrics, & assessment methods #2 assess all students in all sections of all courses #3 maintain SLO assessment.
Life Science online SLO Assessment. Click on the course that you want to assess, e.g. Anthro 101.
Evaluation Requirements for MSP and Characteristics of Designs to Estimate Impacts with Confidence Ellen Bobronnikov February 16, 2011.
Welcome to NU 304!!! I am so excited about this class!! I have identified some items in this presentation to help orient you to the course! Please do not.
Student Growth Goals for Coaching Conversations. Requirements for Student Growth Goals Time line Reference to an Enduring Skill Proficiency and Growth.
Writing Across the Curriculum at Kennedy-King College Some myths and facts…
January, 2010 Betsy Ott, Charlotte Creason Co-Chairs, Assessment Committee Molly McCoy Assistant Director, Institutional Effectiveness 1.
Instructional Plan | Slide 1 AET/515 Instructional Plan For Associate’s Degree in Library Skills (Donna Roy)
AP Lang by the Numbers. Scoring Systems -When we talk about scores, there are two separate scoring systems that matter to you. What is my grade in class?
Foundational Studies CARF – Fall/Spring
GENERAL EDUCATION COURSE-BASED ASSESSMENT – IT’S EASIER THAN YOU THINK! C. Griffin.
GENERAL EDUCATION COURSE-BASED ASSESSMENT – IT’S EASIER THAN YOU THINK! S. Berlin.
Consider Your Audience
Structure of Homework Assignments
LAE Assessment My Course Was Selected For LAE Assessment Reporting.
Reporting the Course level RWR Assessment data
Director of Institutional Effectiveness
The Good The Bad & The Ugly Real-Life Examples of the SLO Assessment Report Form With Tips on How to Complete It August 21, 2012.
Practical and Painless Outcomes Reporting Through Canvas
Institutional Effectiveness USF System Office of Decision Support
Functional Area Assessment Process
Collecting Data to Assess Capstone and Culminating Courses
Assessing Academic Programs at IPFW
GENERAL EDUCATION COURSE-BASED ASSESSMENT – IT’S EASIER THAN YOU THINK!
What to do with your data?
Curriculum Coordinator: Marela Fiacco Date : February 29, 2015
Department Chair: Liz Brown Date of Presentation: 01/19/18
Life Science online SLO Assessment
GENERAL EDUCATION COURSE-BASED ASSESSMENT – IT’S EASIER THAN YOU THINK!
Presentation transcript:

The Good The Bad & The Ugly Real-Life Examples of the SLO Assessment Report Form With Tips on How to Complete It August 21, 2012

Critical Dates SEPTEMBER 14, 2012 All missing parts of assessment report forms to be completed and submitted. OCTOBER 15, 2012 ACCJC REPORT DUE

To report assessments, please use the five- column reporting form shown in the following slide. When completed, the report should be sent to the Department Chair who aggregates the data and then sends one form for each course with the aggregated data to the Office of Institutional Effectiveness and Assessments.

COURSE SLO ASSESSMENT REPORT, SCC Department: _______________ Course: _________________ Year: __________ Semester:___________ Faculty Member: _____________________________ 1) Outcome to be assessed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next 1) Outcome to be assessed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data5) Plan of action/what to do next COURSE SLO ASSESSMENT REPORT, SCC Department: _______________ Course:______________ Year: __________ Semester:___________ Faculty Member:

Five-column report form Correspond to the five steps in the SLO Assessment Cycle Identify/Revise Student Learning Outcomes Identify/Revise Means of Assessment Collect Data Column 4: Analyze Data Plans/Changes

COURSE SLO ASSESSMENT REPORT, SCC Department: _______________ Course: _________________ Year: __________ Semester:___________ Faculty Member: _____________________________ 1) Outcome to be assessed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next 1) Outcome to be assessed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data5) Plan of action/what to do next COURSE SLO ASSESSMENT REPORT, SCC Department: _______________ Course:______________ Year: __________ Semester:___________ Faculty Member: Prior to Fall 2012Fall 2012 End of Fall 2012 Flex week Jan 2013

The next few slides will explain what is to be included in each of the five columns.

Column One 1) Outcome to be assessed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data5) Plan of action/what to do next Student Learning Outcomes: Verbatim from CurricUnet List SLOs assessed (partial or all) Common error: SLOs do not match official SLOs in course outline

Column Two 1) Outcom e to be assessed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next Means of Assessment: C learly-identified and specific type of assessment Specific criteria of success Only assessment measuring the relevant SLO

Column Two - Examples 1) Outc ome to be asses sed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next “ Exam” THIS IS INSUFFICIENT AS MEANS OF ASSESSMENT BECAUSE IT IS NOT SPECIFIC ENOUGH. ALSO, THERE IS NO CRITERIA OF SUCCESS LISTED. Means of Assessment: C learly-identified and specific type of assessment Specific criteria of success Only assessment measuring the relevant SLO

Column Two - Examples 1) Outc ome to be asses sed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next “Ten embedded multiple-choice questions in the final exam” THIS IS A MORE SPECIFIC ASSESSMENT, BUT IT STILL DOES NOT GIVE A BENCHMARK FOR SUCCESS. Means of Assessment: C learly-identified and specific type of assessment Specific criteria of success Only assessment measuring the relevant SLO

Column Two - Examples 1) Outc ome to be asses sed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next “An internally-developed means of assessment to be given” THIS IS TOO VAGUE AND NOT COMPLETE (NO SPECIFIC TYPE OF ASSESSMENT, NO CRITERIA OF SUCCESS) Means of Assessment: C learly-identified and specific type of assessment Specific criteria of success Only assessment measuring the relevant SLO

Column Two - Examples 1) Outc ome to be asses sed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next “All exams, quizzes, homework, papers, projects, attendance and class participation with grading scale as follows: A=90-100%, B=80-89%, C=70-79%, D=60-69%, F=below 60%.” THIS ONE IS TOO BROAD. BASICALLY IT IS THE GRADE FOR THE COURSE AND NOT SPECIFICALLY TIED TO AN SLO Means of Assessment: C learly-identified and specific type of assessment Specific criteria of success Only assessment measuring the relevant SLO

Column Two - Examples 1) Outc ome to be asses sed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next “Twenty-two embedded m/c questions in the final exam, with achieving 70% correct as meeting standard.” THIS IS A GOOD EXAMPLE (WITH SPECIFIC TYPE OF ASSESSMENT, SPECIFIC MINIMUM CONSIDERED SUCCESSFUL). Means of Assessment: C learly-identified and specific type of assessment Specific criteria of success Only assessment measuring the relevant SLO

Column Two - Examples 1) Outc ome to be asses sed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next “A research paper to be evaluated using a 4- point rubric; a 3 or above considered meeting the standard. We expect 80% of students will meet the standard” THIS IS A GREAT EXAMPLE, WITH SPECIFIC TYPE OF ASSESSMENT, SPECIFIC CRITERIA OF SUCCESS, AND AN EXPECTATION / BENCHMARK OF % OF STUDENTS MEETING THE LEARNING OUTCOME. IF POSSIBLE, ATTACH A COPY OF THE RUBRIC USED FOR EVALUATION. Means of Assessment: C learly-identified and specific type of assessment Specific criteria of success Only assessment measuring the relevant SLO

Column Three 1) Outco me to be assesse d 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next Summary of Data: Meaningful and provide context Success rate, distribution, rubric scores, improvement scores # of students assessed (n) % of students meeting SLO

Column Three - Examples 1) Outcome to be assessed 2) Means of assessme nt and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next ”Avg: 4.33” NOT ENOUGH INFORMATION IN THIS REPORT (NO CONTEXT) Summary of Data: Meaningful and provide context Success rate, distribution, rubric scores, improvement scores # of students assessed (n) % of students meeting SLO

Column Three - Examples 1) Outcome to be assessed 2) Means of assessme nt and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next “30 students took the exam. 24 got at least 80% correct.” THIS IS BETTER BUT NOT COMPLETE. WHAT IS THE PERCENTAGE WHO REACHED THE BENCHMARK (MET THE SLO STANDARD)? Summary of Data: Meaningful and provide context Success rate, distribution, rubric scores, improvement scores # of students assessed (n) % of students meeting SLO

Column Three - Examples 1) Outcome to be assessed 2) Means of assessme nt and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next “26 students took the exam. Results are as follows: 10 got 90% or higher 10 got 80-89% 5 got 70-79% 2 got 60-69% 1 got below 60%” THIS IS THE RESULT OF THE TEST. WHAT WAS THE BENCHMARK AND WHAT PERCENT MET IT? Summary of Data: Meaningful and provide context Success rate, distribution, rubric scores, improvement scores # of students assessed (n) % of students meeting SLO

Column Three - Examples 1) Outcome to be assessed 2) Means of assessme nt and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next “Students did really well in the post test especially on the take home portion, because they had more time to think the questions through. The results are as expected.” WHAT DOES REALLY WELL MEAN? THERE IS NO DATA (NUMBERS, PERCENTAGES) HERE. Summary of Data: Meaningful and provide context Success rate, distribution, rubric scores, improvement scores # of students assessed (n) % of students meeting SLO

Column Three - Examples 1) Outcome to be assessed 2) Means of assessme nt and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next “32 students took pre-test, while 28 took post-test. Pre- test: 10 correct = 7 students 9 correct = 16 students 8 correct = 4 students 7 correct = 0 students 6 correct = 3 students 0-5 correct = 0 students Post-test 10 correct = 10 students 9 correct = 15 students 8 correct = 2 students 7 correct = 1 student THE SCORES ARE HERE, BUT WHAT %OF THE STUDENTS MET THE IMPROVEMENT BENCHMARK? ALSO, WE DON’T KNOW HOW MANY STUDENTS IMPROVED OR REGRESSED? Summary of Data: Meaningful and provide context Success rate, distribution, rubric scores, improvement scores # of students assessed (n) % of students meeting SLO

Column Three - Examples 1) Outcome to be assessed 2) Means of assessme nt and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next “30 students turned in research papers evaluated using a 4 point rubric. Results: 4 pts = 10 students 3 pts = 8 students 2 pts = 8 students 1 pt = 4 students Average: students (60%) meet SLO standard of achieving a 3 or above. This is below my expectations of at least 80% of students meeting the standard….” EXCELLENT, COMPLETE EXAMPLE Summary of Data: Meaningful and provide context Success rate, distribution, rubric scores, improvement scores # of students assessed (n) % of students meeting SLO

Column Four 1) Outcome to be assessed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/ what to do next Analysis of Data: Strengths and Weaknesses Compare results with expectations Possible explanations for outperformance or shortfall Statistical significance

Column Four - Examples 1) Outcome to be assessed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/ what to do next “Exceeds expectation” INCOMPLETE REPORT WITH VERY LITTLE INSIGHT THAT IS USEFUL FOR FUTURE PLANNING. Analysis of Data: Strengths and Weaknesses Compare results with expectations Possible explanations for outperformance or shortfall Statistical significance

Column Four - Examples 1) Outcome to be assessed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/ what to do next “The data showed there was a 25% improvement from pre to post-test, which we consider very successful. Our future goal is to increase the number of students achieving 65-75% score in the post-test” THIS ANALYSIS DOESN’T SHOW HOW MANY STUDENTS IMPROVED AND WHAT % MEET THE SLO. FUTURE GOAL SHOULD BE WRITTEN IN COLUMN FIVE Analysis of Data: Strengths and Weaknesses Compare results with expectations Possible explanations for outperformance or shortfall Statistical significance

Column Four - Examples 1) Outcome to be assessed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/ what to do next Analysis of Data: Strengths and Weaknesses Compare results with expectations Possible explanations for outperformance or shortfall Statistical significance THIS ANALYSIS DOESN’T ANALYZE THE SLO DATA COLLECTED, BUT INSTEAD DWELLS TOO MUCH ON STUDENTS’ CLASS GRADES AND THE ASSESSMENT INSTRUMENT.

Column Four - Examples 1) Outcome to be assessed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/ what to do next “ All operations of all types of equations assessed had a success rate over 70% except for two. 1) Quadratic formula had a success rate of only 54.27%. 2) Dividing Polynomial by a monomial 67.68% *)Factoring a 4-term by grouping was borderline at 70.73% There were two types of problems that had high success rates (>90%): 1) Solving Linear Equations 92.07% 2)Dividing a monomial by a monomial (exponent rules) 98.48%” The quadratic formula is generally taught during the last week. May not be enough time for a student to “master” it by the final. GOOD EXAMPLE. SEE NEXT SLIDE FOR WHY Analysis of Data: Strengths and Weaknesses Compare results with expectations Possible explanations for outperformance or shortfall Statistical significance

The previous report is good because: it analyzes what parts are strong and what parts are weak it compares results to expectation (70+%) it postulates a possible explanation for weaknesses.

Column Five 1) Outcome to be assessed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next Plans: Changes in teaching method, emphasis, time devoted to topics, and other pedagogical questions Changes to SLO Changes to assessment instrument. Discuss with whole department

Column Five - Examples 1) Outcome to be assessed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next “ Nothing ” IN THE UNLIKELY EVENT THAT 100% ACHIEVED THE SLO, THIS MIGHT BE THE START TO AN ANSWER TO THIS COLUMN. IT IS SIMPLY NOT ENOUGH. IDENTIFY THE TEACHING METHODS THAT PRODUCED THIS EXCELLENT RESULT OR RECONSIDER THE APPROPRIATENESS OF ASSESSMENT INSTRUMENT AND BENCHMARK FOR SUCCESS. Plans: Changes in teaching method, emphasis, time devoted to topics, and other pedagogical questions Changes to SLO Changes to assessment instrument. Discuss with whole department

Column Five - Examples 1) Outcome to be assessed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next “Everything seems to work just fine. Students who did not meet the standard were students who did not do the work and assignments. I will continue to do the same thing next semester and will use the same assessment.” A MORE THOUGHTFUL ANALYSIS OF WHAT COULD BE DONE TO IMPROVE THE RESULTS IS NECESSARY HERE. (SEE ADDITIONAL COMMENTS ON PREVIOUS SLIDE) Plans: Changes in teaching method, emphasis, time devoted to topics, and other pedagogical questions Changes to SLO Changes to assessment instrument. Discuss with whole department

Column Five - Examples 1) Outcome to be assessed 2) Means of assessment and criteria of success 3) Summary of data collected 4) Analysis of data 5) Plan of action/what to do next Plans: Changes in teaching method, emphasis, time devoted to topics, and other pedagogical questions Changes to SLO / assessment Expectation to improve Discuss with whole department THIS IS A GOOD THOUGHTFUL EXAMPLE, INCLUDING CHANGES TO ASSESSMENT TOOL, TEACHING STRATEGY, AND EXPECTED FUTURE IMPROVEMENT GOALS

Desirable Standard Practices Correct SLOs: – Listed in course outline in CurricUnet Means of Assessment: – Specific type of assessment with criteria of success – Specifically targeted to measure stated SLO Meaningful Data: – Number of students, Success Rate (% of students who met the standard), Improvement Scores Thoughtful Analysis: – Slice and dice the data. Evaluate strengths and weaknesses. Plans/Changes: – Possible changes in teaching method, means of assessment, SLO? – Continuous Quality Improvement – “Closing the loop”

Common Errors Unmatched SLOs – Different from SLOs in CurricUnet Vague or unmatched means of assessment – SLO and assessment don’t match or align – Not specifically-stated “Minimalism” – Lacks analysis, no data, no plans or changes Placing items in the wrong columns Copy & paste error Wrong or missing year, semesters, course numbers Multiple sections of the same course submitted separately (not aggregated)

Logistics. – Outcomes assessment website – SLOARC website – outcomes assessment office for a copy – Your department chair/SLO coordinator – As soon as form is completed (after full discussion with department) – Send to department chair/SLO coordinator, who will aggregate data for each course and forward the report to the Outcomes Assessment office. When and to whom do I turn in the completed form? Where can I get copy of the form?

QUESTIONS? SCC Outcomes Assessment Office A-213 Kay Powell Rudy Tjiptahadi THANK YOU!