Presentation is loading. Please wait.

Presentation is loading. Please wait.

Categorization of MSP-RETA tools via Guide to PD Program Quality Rubric, Section II: Evaluation from CCSSO Draft modifications for IMSP Evaluation Outcome.

Similar presentations


Presentation on theme: "Categorization of MSP-RETA tools via Guide to PD Program Quality Rubric, Section II: Evaluation from CCSSO Draft modifications for IMSP Evaluation Outcome."— Presentation transcript:

1 Categorization of MSP-RETA tools via Guide to PD Program Quality Rubric, Section II: Evaluation from CCSSO Draft modifications for IMSP Evaluation Outcome AreaQuestionsMeasureAdvice A. Quality of PD Activity Review the definition of professional development from your RFP, Appendix B—How can your design for the continuum of professional development meet these characteristics, regardless of the setting. (copy provided)  Go beyond the traditional:  Did they like it?  Was the room the right temperature?  Were there enough diet drinks?  Was their time well-spent? But how will you know this?  Were the activities consistent with goal? Were the goals clearly ‘explained’?  Did the material make sense?  Will it be useful?  Was the leader knowledgeable and helpful?  Was the content appropriately and clearly linked to standards—both for teachers (CAS) primarily, and students (ILS)  Was the course taught at the appropriate level? Innovatively?  Was the course linked to teachers’ classrooms? Innovatively  Was the content presented with effective pedagogy?  How were the best practices for instruction modeled? Which ones? How well? Innovatively?  How were the challenges to quality identified/addressed/resolved? How quickly?  What was the quality of the institute organization and presentation? (CPS); To what extent are teachers’ experiences in the program’s professional development opportunities of high quality? (PNNM); How could these experiences be improved? (PNNM)  In what ways could the professional development opportunities offered through this program be enhanced to further address teachers’ needs and/or provide them with support? (PNNM)  Was an effective learning environment provided? (CPS) 1. CCSSO Evaluation Tools for Professional Development: http://www.ccsso.org/project s/improving%5Fevaluation% 5Fof%5Fprofessional%5Fde velopment/Cross%5FState% 5FStudy/ http://www.ccsso.org/project s/improving%5Fevaluation% 5Fof%5Fprofessional%5Fde velopment/Cross%5FState% 5FStudy 2. CCSSO PD Program Quality Rubric: http://www.ccsso.org/content /PDFs/Revised%20Rubric% 20%2D%20rev022406.xls http://www.ccsso.org/content /PDFs/Revised%20Rubric% 20%2D%20rev022406.xls …and guide: http://www.ccsso.org/content /PDFs/RubricGuide_rev0329 06.doc http://www.ccsso.org/content /PDFs/RubricGuide_rev0329 06.doc 3. Horizon Research, Inc. Local Systemic Change Instruments: http://www.horizon- research.com/instruments/ http://www.horizon- research.com/instruments/ 4. Professional Development Activity Log (PDAL): http://www.pdal.net/ http://www.pdal.net/ See especially these reports on the use of the PDAL: http://www.pdal.net/inc/docs /MSP_PD%20effects_report. pdf http://www.pdal.net/inc/docs /MSP_PDStudy.pdf 5. Course Evaluations* Check the professional development standards in the National Science Education Standards (http://www.nap.edu/readingroom/books/ nses/) and the NCTM-Principles and Standards for School Mathematics (http://standards.nctm.org/) as useful references for evaluating professional development.http://www.nap.edu/readingroom/books/ nses/http://standards.nctm.org/ How does the Appendix B definition of professional development expand the scope of your partnership design? Consider coherence factors: such as the coherence and alignment of curricular resources, current/future classroom facilities, to School Improvement goals, and needs whether teacher-identified or from student achievement analysis. *Course evaluations will have to be treated significantly more innovatively than the traditional student evaluations of courses.

2 MATRIX, continued Evaluation Outcome AreaQuestionsMeasureAdvice B. Change in Teacher Content Knowledge From PSSM: How can teachers learn what they need to know (p. 370) From IMSP-Benchmarks:  50 percent cumulative increase in specific scientific, mathematical, and engineering (SME) content expertise, including pedagogical content knowledge and leadership qualifications to be demonstrated through at minimum pre- and post-test methodologies, etc. ▪ Did participants acquire the intended knowledge and skills? Innovatively? ▪ To what extent and with what quality are teachers implementing their new content knowledge and skills (e.g., increased use of inquiry-based science instruction, mathematics problem-solving strategies, etc.?) (PNNM) ▪ To what extent have teachers’ attitudes about sc/math and teaching s/m? Misconceptions-Oriented Standards- Based Resources for Teachers (MOSART) 1 – Physical and Earth and Space Science http://www.cfa.harvard.edu/smgphp/ mosart/about_mosart.html Diagnostic Teacher Assessments in Math and Science (DTAMS) 2 – http://louisville.edu/edu/crmstd/diag_ sci_assess_middle_teachers.html Assessing Teacher Learning about Science Teaching (ATLAST) 3 – http://www.horizon- http://www.horizon- research.com/atlast/http://www.horiz on-research.com/atlast/about/ Learning Mathematics for Teaching (LMT) 4 – http://sitemaker.umich.edu/lmt/home http://sitemaker.umich.edu/lmt/files/L MT_sample_items.pdf http://sitemaker.umich.edu/lmt/home http://sitemaker.umich.edu/lmt/files/L MT_sample_items.pdf Journal entries, analysis through coding 5 1 Primarily designed for teachers, could be used for students. No tests available for life science. 2 Available for middle and elementary math teachers, and middle school science (4 content domains; life, physical and earth/space) developers suggest use for determining teacher knowledge growth over time, and to identify areas of weakness that can be addressed w/further PD 3 Looks at science and pedagogical content knowledge, but in very limited content areas: Force and motion, Plate tectonics, flow of matter and energy in living systems. For middle school teachers. 4 Mathematics content only, elementary number and operations, algebra, geometry 5 Examples from Rachel Shefner Develop strong Action Research scaffolding

3 MATRIX, continued Evaluation Outcome AreaQuestionsMeasureAdvice C. Change in Instructional Practices or Curriculum i. classroom level of the teacher as a participant and in the classrooms they teach ii. school building level, collectively iii. district level iv. at the IHE  50 percent cumulative increase in confidence, effectiveness, and utility of pertinent science, mathematics and engineering teaching skills, technologies, and applications to be demonstrated through self-reflective journals, interviews, and action research, and data trends, etc.  50 percent cumulative increase in application of findings, evaluation of personal utility, and development of critical questioning skills for analysis of pertinent scientifically based educational research for teaching and learning mathematics and science to be demonstrated through scaffolded reflections, journals, interviews, etc.  50 percent cumulative increase using measures identified for sustained excellence and support of content expertise in high-need LEAs, including:  50 percent increase in administrative programmatic support measures for scheduling, preparation, and evaluating the quality mathematics and science teaching and learning, to be demonstrated through pre- post surveys and interviews, etc.  50 percent increase in collaborative decision- making measures between participating teachers and administrators to improve the quality of mathematics and science teaching and learning, with specific focus on deliberate collegial connections to professional development among professional staff and resolution of issues associated with teaching and learning settings, to be demonstrated through pre-post surveys and interviews, etc.  Did participants effectively apply the new knowledge and skills?  How did the partnership activities impact participants in their roles as teacher leaders and/or as classroom teachers? (CPS)  How did the partnership activities contribute to teacher leaders’ ability to support math and science inquiry in schools and classrooms? (CPS)  How did participation contribute to building a professional learning community? If so, how did specific activities contribute to building this community?  How do participants plan to tap into this community in their roles as teacher leaders?  PSSM: p. 370-372: Are good instructional materials chosen, used, accepted? Are assessments aligned to clear instructional goals? Are STEM- and STEM-ed technology supporting learning? i. SLPAI: Science Lesson Plan Analysis Instrument 1 http://hub.mspnet.org/index.cf m/14244 i. video coding tool (Heather Hill) http://sitemaker.umich.edu/lmt/ faq_about_video_codes i, ii, iii SEC http://www.ccsso.org/Projects/ Surveys_of_Enacted_Curriculu m/ http://seconline.wceruw.org/se cWebHome.htm i., ii RTOP 2 http://hub.mspnet.org/index.cf m/8678 http://hub.mspnet.org/index.cf m/8678 i. Journal entries, analysis through coding ii. Principal Survey 3, Local Systemic Change through Teacher Enhancement: Principal Questionnaire From Horizon Research i. Inside the Classroom Observation and Analytic Protocol 4, also from Horizon http://www.horizon- research.com/instruments/clas/ cop.php i,ii,ii 2005-06 Local Systemic Change Classroom Observation Protocol 5 http://www.horizon- research.com/instruments/lsc/c op.php 1 Not to be used in isolation but can capture info on CK, PCK and impact on teaching practice. Used in pilot study on measuring teacher change thru program instruction 2 Observation protocol specifically designed to detect “reform”; but developers do not necessarily equate reform w/quality 3 SEC has a principals survey, but there are difficulties? Does it get at what we want? 4 Developed to measure the quality of an observed K-12 science or mathematics classroom lesson by looking at design, implementation, content and culture. Very long! 5 A shorter version of the above, which makes it more usable. Could do pre/post visits? Advice/guidelines for classroom observations?

4 MATRIX, continued Evaluation Outcome AreaQuestionsMeasureAdvice D. Student Achievement  30 percent increase in student performance in specific SME content areas across multiple effective metrics of formative and summative assessment analysis in measures through cumulative action research analysis, including:  30 percent increase in students’ demonstration of achievement in the “meets expectations” category of Illinois’ large-scale assessment measures;  30 percent increase in students’ demonstration of achievement in the “exceeds expectations” category of Illinois’ large-scale assessment measures; and  At least 30 percent increase in students’ demonstration of achievement through trend analysis of classroom-based, high-quality formative assessments and action research.  What was the impact on the students holistically?  Did it affect student performance or achievement?  Are students more confident as learners?  Is student attendance improving?  From the legislation: SHALL include measurable objectives fro improved student academic achievement on State math/sc assessments or, where applicable, an international math/science study assessment; and may include objectives/ measures for:  Increased participation by students in advanced courses in m/s;  Increased percentages of elementary school teachers with academic majors or minors or group majors or minors, in math/eng/or the sciences:  Increased percentages of secondary school classes in m/s taught by teachers with academic majors in math/ engineering/science MOSART–Physical and Earth and Space Science http://mosart.mspnet.org/ http://mosart.mspnet.org/ OMLI Classroom Observation Protocol – http://hub.mspnet.org/index.cf m/11980 http://hub.mspnet.org/index.cf m/11980 ISAT – http://www.isbe.state.il.us/asse ssment/isat.htm Balanced Assessment – http://balancedassessment.conc ord.org/ MCAS – http://www.doe.mass.edu/mcas /testitems.html NAEP – http://nces.ed.gov/nationsrepor tcard/about/naeptools.asp PISA Items – http://nces.ed.gov/surveys/pisa/ educators.asp Tennessee Value-Added Assessment System (TVAAS) – http://www.sas.com/govedu/ed u/ed_eval.pdf http://epa.sagepub.com/cgi/rep rint/25/3/287 Designed for students, can be used for teachers? Observations are centered around student thinking—not for teachers As we know and love it Open-ended items, higher on Bloom’s taxonomy Items are released every year, could be used as benchmarks Used for national assessment, some released items European assessment items, some released items This is a system of quantitative evaluation of teachers through student test scores and gains using complex statistical analysis attributing the gains (or losses) to teacher effects. The second link is a validation study of the instrument.

5 MATRIX, continued Evaluation Outcome AreaQuestionsMeasureAdvice E. Other Outcomes: Organizational Support and Change? From RFP-benchmarks:  100 percent of participating IHEs committing to sustained program support beyond the term of IMSP funding:  Inter- and intra-college/department administrative support is documented by meeting minutes, letters of continuing support and interviews, etc.;  Deliberate interconnections to teacher preparatory programs and general education programs is documented by meeting minutes, program impact statements and design elements, and interviews, etc.  50 percent cumulative increase using measures identified for sustained excellence and support of content expertise in high-need LEAs, including:  50 percent increase in administrative programmatic support measures for scheduling, preparation, and evaluating the quality mathematics and science teaching and learning, to be demonstrated through pre-post surveys and interviews, etc.  50 percent increase in collaborative decision-making measures between participating teachers and administrators to improve the quality of mathematics and science teaching and learning, with specific focus on deliberate collegial connections to professional development among professional staff and resolution of issues associated with teaching and learning settings, to be demonstrated through pre-post surveys and interviews, etc.  From PSSM: pp 375-378, for teacher leaders, school/district/state/province administrators, higher ed faculty, professional organizations and policymakers  From NSES: program and system standards  Were sufficient resources made available?  Were problems addressed quickly and efficiently?  Was implementation advocated, facilitated, and supported?  Were successes recognized and shared?  Was the support public and overt? At the LEA and its classroom levels, at the IHE and its classroom/course/program levels:  WHAT WAS THE IMSP IMPACT ON THE ORGANIZATION?  HOW DID IMSP AFFECT ORGANIZATIONAL CLIMATE AND PROCEDURES?  How did IMSP affect program, course designing processes?  How did IMSP affect instructional strategies?  How did IMSP affect possible future partnerships?


Download ppt "Categorization of MSP-RETA tools via Guide to PD Program Quality Rubric, Section II: Evaluation from CCSSO Draft modifications for IMSP Evaluation Outcome."

Similar presentations


Ads by Google