Ning Fang Don Millard Division of Undergraduate Education National Science Foundation November 10, 2010.

Slides:



Advertisements
Similar presentations
Critical Reading Strategies: Overview of Research Process
Advertisements

Funding for Education Scholarship Russ Pimmel NSF ASEE Annual Conference June 20, 2006.
Session 5 Intellectual Merit and Broader Significance FISH 521.
NSF Research Proposal Review Guidelines. Criterion 1: What is the intellectual merit of the proposed activity? How important is the proposed activity.
Merit Review and Proposal Preparation Mark Courtney Division of Environmental Biology
NSF Merit Review and Proposal Preparation Mark Courtney, Ph.D Adjunct, Department of Biology New Mexico State University 24 September 2008.
An Excellent Proposal is a Good Idea, Well Expressed, With A Clear Indication of Methods for Pursuing the Idea, Evaluating the Findings, and Making Them.
NSF Merit Review Criteria Revision Background. Established Spring 2010 Rationale: – More than 13 years since the last in-depth review and revision of.
The Proposal Review Process Matt Germonprez Mutual of Omaha Associate Professor ISQA College of IS&T.
Preparation/Content of an NSF proposal NSF proposals are uploaded to the Fastlane website prior to submission (NIH uses Grants.gov): 1.Cover sheet (basic.
NSF Doctoral Dissertation Improvement Grants Improve dissertation research – Provide funds not normally available to graduate students significant data-gathering.
How to Write Grants Version 2009.
Proposal Review Process Mock Review Webinar
Proposal Writing Workshop Features of Effective Proposals: Fellowship Track Washington, DC January 9, 2014.
The IGERT Program Preliminary Proposals June 2008 Carol Van Hartesveldt IGERT Program Director IGERT Program Director.
1 Jill Singer Division of Undergraduate Education Directorate for Education & Human Resources National Science Foundation Sustainability.
1 CCLI Proposal Writing Strategies Tim Fossum Program Director Division of Undergraduate Education National Science Foundation Vermont.
National Science Foundation: Transforming Undergraduate Education in Science, Technology, Engineering, and Mathematics (TUES)
EAS 299 Writing research papers
Two Year College Bert E. Holmes Carson Distinguished Chair of Science at UNC-Asheville and formerly Program Officer in Division of Undergraduate Education.
NSF Office of Integrative Activities Major Research Instrumentation Program November 2007 Major Research Instrumentation EPSCoR PI Meeting November 6-9,
CAREER WORKSHOP APRIL 9, 2014 Required Elements of the Proposal Beth Hodges Director, Office of Proposal Development FLORIDA STATE UNIVERSITY.
Proposal Strengths and Weakness as Identified by Reviewers Russ Pimmel & Sheryl Sorby FIE Conference Oct 13, 2007.
Top Ten Ways To Write a Good Proposal… That Won’t Get Funded.
Effective proposal writing Session I. Potential funding sources Government agencies (e.g. European Union Framework Program, U.S. National Science Foundation,
March 20, 2012 Susan Finger & Sue Fitzgerald Division of Undergraduate Education National Science Foundation March 21, 2012 Sue Fitzgerald & Maura Borrego.
Tips for Writing a Successful Grant Proposal Diana Lipscomb Associate Dean for Faculty and Research CCAS.
WE ARE A COMPLEX LAND. MASLOW’S HIERARCHY OF NEEDS DESIRE TO HELP OTHERS MEANING TO LIFE ESTEEM NEEDS RECOGNITION & APPRECIATION BELONGINGNESS AND LOVE.
Overview of NSF Education R & D Programs with an Emphasis on the TUES Program Louis Everett Susan Finger Sue Fitzgerald.
Funding Opportunities for Chemists at the National Science Foundation Division of Undergraduate Education Pamela Brown, NSF Program Director Division of.
Submitting a Proposal: Best Practices By: Anu Singh Science Assistant
Writing More Effective Proposals Russ Pimmel Abe Nisanci U of Alabama NSF. Share The Future IV March 17, 2003.
 NSF Merit Review Criteria Intellectual Merit Broader Impacts  Additional Considerations Integration of Research & Education Integrating Diversity into.
Proposal Writing Workshop Features of Effective Proposals.
A Roadmap to Success Writing an Effective Research Grant Proposal Bob Miller, PhD Regents Professor Oklahoma State University 2011 Bob Miller, PhD Regents.
Partnerships and Broadening Participation Dr. Nathaniel G. Pitts Director, Office of Integrative Activities May 18, 2004 Center.
Grant Research Basics. Asking the Question  Before you start, you must have both clearly stated research question and primary outcome measure.  What.
Proposal Development Sample Proposal Format Mahmoud K. El -Jafari College of Business and Economics Al-Quds University – Jerusalem April 11,2007.
Writing More Effective NSF Proposals Jeanne R. Small Oklahoma City, Oklahoma March 2, 2006 Division of Undergraduate Education (DUE) National Science Foundation.
Academic Research Enhancement Award (AREA) Program Erica Brown, PhD Director, NIH AREA Program National Institutes of Health 1.
NSF GRFP Workshop Sept 16, 2016 Dr. Julia Fulghum
 How the knowledge created advances our theoretical understanding of the study topic, so that others interested in similar situations but in a different.
Funding your Dreams Cathy Manduca Director, Science Education Resource Center Iowa State University, 2005.
An Excellent Proposal is a Good Idea, Well Expressed, With A Clear Indication of Methods for Pursuing the Idea, Evaluating the Findings, and Making Them.
1. Most of the information presented in this workshop represents the presenter’s opinion and not an official NSF position 2.
Writing a More Effective Proposal Susan Burkett and Stephanie Adams February 9, 2006.
 NSF Merit Review Criteria Intellectual Merit Broader Impacts  Additional Considerations Integration of Research & Education Broadening Participation.
Proposal Writing Workshop Features of Effective Proposals.
CAREER WORKSHOP APRIL 6, 2015 Required Elements of the NSF Proposal Beth Hodges Director, Office of Proposal Development FLORIDA STATE UNIVERSITY.
Proposal Writing Workshop Features of Effective Proposals.
1 Developing a Competitive Proposal ( An Interactive, Web-Based Workshop) Russell Pimmel Division of Undergraduate Education National Science Foundation.
NSF Peer Review: Panelist Perspective QEM Biology Workshop; 10/21/05 Dr. Mildred Huff Ofosu Asst. Vice President; Sponsored Programs & Research; Morgan.
1Mobile Computing Systems © 2001 Carnegie Mellon University Writing a Successful NSF Proposal November 4, 2003 Website: nsf.gov.
 Ensure the title is in line with the requirements of the proposed funding agency if they have any specification for the titled page (some do have.
Course, Curriculum, and Laboratory Improvement (CCLI) Transforming Undergraduate Education in Science, Technology, Engineering and Mathematics PROGRAM.
1. October 25, 2011 Louis Everett & John Yu Division of Undergraduate Education National Science Foundation October 26, 2011 Don Millard & John Yu Division.
How to Obtain NSF Grants Review of Proposal Pieces A workshop providing information on the process of applying for external research awards. Sponsored.
Improving Research Proposals: Writing Proposals and the Proposal Review Process Heather Macdonald (based on material from Richelle Allen-King, Cathy Manduca,
CHAPTER 16 Preparing Effective Proposals. PRELIMINARY CONSIDERATIONS  Conducting a Preliminary Assessment  Prior to Writing the Proposal  How Fundable.
Integrative Graduate Education and Research Traineeship The NSF Reviewers’ Perspective NSF Training Grants Workshop.
NSF INCLUDES Inclusion Across the Nation of Learners of Underrepresented Discoverers in Engineering and Science AISL PI Meeting, March 1, 2016 Sylvia M.
Data Infrastructure Building Blocks (DIBBS) NSF Solicitation Webinar -- March 3, 2016 Amy Walton, Program Director Advanced Cyberinfrastructure.
Intellectual Merit & Broader Impact Statements August 2016
Developing a Competitive Proposal (An Interactive, Web-Based Workshop) Russell Pimmel Division of Undergraduate Education National Science Foundation.
FISH 521 Further proceedings Peer review
Intellectual Merit & Broader Impact Statements August 2018
Intellectual Merit & Broader Impact Statements August 2017
Writing More Effective NSF Proposals
S-STEM (NSF ) NSF Scholarships for Science, Technology, Engineering, & Mathematics Information Materials 6 Welcome! This is the seventh in a series.
Intellectual Merit & Broader Impact Statements August 2019
Presentation transcript:

Ning Fang Don Millard Division of Undergraduate Education National Science Foundation November 10, 2010

Mock Review Webinar - 2  Introduction  Overview of the review process  Instructions on producing a panel review  Organization of participants into teams and designation of panel chairs  Local teams discuss strengths and weaknesses  Teams locally report/discuss results  Reporting to virtual group with ratings – note: facilitators will be asked to select an individual to report BREAK (15min)  Individuals consider ways to improve the proposal  Local teams discuss suggestions for improvement  Teams locally report/discuss results (5min)  Report back to virtual group - note: facilitator picks an individual to report  PD commentary on responses  Think – What have I learned today that I will use in preparing my next proposal?  Share your thoughts with local participants  Facilitator reports results back to virtual group Wrap-up Q&A

Mock Review Webinar - 3  Help participants to:  Become more familiar with the proposal review process  Better understand the TUES-specific criteria  Better understand the use of intellectual merit/broader impact criteria in reviewing proposals  Develop more competitive proposals that effectively meet the expectations of the TUES program

Mock Review Webinar - 4

Mock Review Webinar - 5  Title changed to emphasize the special interest in projects that have the potential to transform undergraduate STEM education  Title changed to emphasize the special interest in projects that have the potential to transform undergraduate STEM education  Review criteria was modified to emphasize the desire for projects that:  Propose materials, processes, or models that have the potential to  Enhance student learning  Be adapted easily by other sites  Involve a significant effort to facilitate adaptation at other sites  Institutionalize the approach at the investigator's college or university as appropriate (e.g., for the Type)  Have the potential to contribute to a paradigm shift in undergraduate STEM education

Mock Review Webinar - 6  Vision: Excellent STEM education for all undergraduate students   Reflects national concerns about producing:   Skilled STEM professionals (including K-12 teachers)   Citizens knowledgeable about STEM and how it relates to their lives   Seeks to build a community of faculty committed to improving undergraduate STEM education   Encourages projects with potential to transform undergraduate STEM education

Mock Review Webinar - 7  Creating Learning Materials and Strategies  Guided by research on teaching and learning  Incorporate and be inspired by advances within the discipline  Implementing New Instructional Strategies  Contribute to understanding on how existing strategies:  Can be widely adopted  Are transferred to diverse settings  Impact student learning in diverse settings  Developing Faculty Expertise  Enable faculty to acquire new knowledge and skills in order to revise their curricula and teaching practices  Involve a diverse group of faculty

Mock Review Webinar - 8  Assessing and Evaluating Student Achievement:  Develop and disseminate valid and reliable tests of STEM knowledge  Collect, synthesize, and interpret information about student understanding, reasoning, practical skills, interests, attitudes or other valued outcomes  Conducting Research on Undergraduate STEM Education:  Explore how:  Effective teaching strategies and curricula enhance learning and attitudes  Widespread practices have diffused through the community  Faculty and programs implement changes in their curriculum

Mock Review Webinar - 9  Projects developing instructional materials and methods should:  Be based on how students learn  Consider transferability and dissemination throughout the project's lifetime  Involve efforts to facilitate adaptation at other sites in more advanced projects

Mock Review Webinar - 10  Expect to award approximately 10%  Total budget: up to $200,000 for 2 to 3 years  $250,000 when 4-year and 2-year schools collaborate  Typically involve a single institution & one program component – but there are exceptions  Contribute to the understanding of undergraduate STEM education  Informative evaluation effort based on the project's specific expected outcomes  Institutionalized at the participating colleges and universities  Deadlines:  May 26, 2011 (A-M)  May 27, 2011 (N-W)

Mock Review Webinar - 11  Type 2 Projects  20 to 25 awards expected  Total budget: up to $600K for 2 to 4 years  Type 3 Projects  3 to 5 awards expected  Budget negotiable, but not to exceed $5M over 5 years  Tues Central Resource Projects  1 to 3 awards expected  Budget negotiable, depending on the scope and scale of the activity  Small focused workshop projects -- 1 to 2 years & up to $100K  Large scale projects -- 3 to 5 years & $300K to $3M Deadline: January 14, 2011

Mock Review Webinar - 13  Reviewers are solicited by program directors   2010 TUES Type 1 Proposals Review:   4 panels/program director (60-70 proposals/PD)   6-8 reviewers/panel   17 proposals/panel – not all read by every panelist   Approximately 130 engineering reviewers  Reviewers assign individual ratings and develop a review 2-3 weeks prior to coming to the panel meeting  Panel meeting is held in DC area – over a 2 day period

Mock Review Webinar - 14  Panel Chair (picked by program director ahead of time) establishes order of proposal review process  Proposals are discussed individually  A “scribe” is designated to capture all of the points brought up in discussion and produce a summary review – called the “Panel Summary”  Panel returns on day 2 to review all the proposals’ panel summaries

Mock Review Webinar - 15  NSF program directors  Informs recommendations relative to funding  Guides pre-award negotiations  Applicants  If proposal is funded:  Provides suggestions for improving project  If proposal is not funded:  Provides information to guide a revision of the proposal

Mock Review Webinar - 16  All proposals are evaluated using the NSB-approved review criterion  Intellectual merit  Broader impacts  The TUES Solicitation provides two sets of suggested questions to help define these criteria  Standard NSF set  TUES specific set

Mock Review Webinar - 17  Suggested questions are only a guide for considering intellectual merit and broader impacts  Suggested questions are NOT:  A complete list of “requirements”  Applicable to every proposal  An official checklist

Mock Review Webinar - 18  Will the project:  Include activities important in advancing knowledge?  Involve qualified proposer(s)?  Contain creative and original concepts?  Have a well conceived and organized plan?  Include sufficient access to resources?

Mock Review Webinar - 19  Will the project:  Advance discovery - promote teaching & learning?  Broaden participation of underrepresented groups?  Enhance the infrastructure?  Include broad dissemination?  Benefit society?

Mock Review Webinar - 20 Will the project:  Produce one or more of the following:  Exemplary materials, processes, or models that enhance student learning and can be adopted by other sites  Important findings related to student learning?  Build on existing knowledge about STEM education?  Have explicit and appropriate expected measurable outcomes integrated into an evaluation plan?  Include an evaluation effort that is likely to produce useful information?  Institutionalize the approach at the investigator's college or university (as appropriate for the Type)

Mock Review Webinar - 21 Will the project:  Involve a significant effort to facilitate adaptation at other sites ?  Contribute to the understanding of STEM education?  Help build and diversify the STEM education community?  Have a broad impact on STEM education in an area of recognized need or opportunity?  Have the potential to contribute to a paradigm shift in undergraduate STEM education ?

Mock Review Webinar - 22

Mock Review Webinar - 23  The Entire Proposal is Used to Inform Reviewers  Project Summary  Project Description  Biographical Sketches  Budget  Supplementary Documentation 23

Mock Review Webinar - 24  A review should indicate an opinion on the merit of the project  The rating should indicate an overall evaluation of the proposal’s merit  Excellent, Very Good, Good, Fair, and Poor  Text boxes are provided to:  Describe positive aspects  Not just list them -- Provide details  Identify concerns (or weaknesses)  Not just list them -- Provide details  Offer suggestions for improvement Rating and text should be consistent

Mock Review Webinar - 25  Describe the strengths and weaknesses (or concerns) relative to the review criteria  Include all of the positive aspects, concerns, and issues  Use a comfortable style and format  Use good grammar, style and complete sentences  A very brief review is not very helpful

Mock Review Webinar - 26  Provide any additional information that you want to transmit  One-sentence summary of the proposal  Summary of the critique  Statement indicating why the proposed project is important or not  Suggestions for improvements  Other comments

Mock Review Webinar - 27   Uses appropriate style   Contains adequate details   Contains understandable, specific, and complete statements   Relates strengths and weakness to review criteria   Indicates why an item is a strength or weakness   Justifies the proposal rating in the written critique A reader should be able to guess the rating from the written text

Mock Review Webinar - 28  Identify a strength or weaknesses  “The evaluation plan is a strength.”  Identify a strength or weaknesses and indicate why it is one  “The evaluation plan includes a competent, independent evaluator...”  “The background discussion is well referenced, shows a good understanding of the prior work, supports the proposed work...”

Mock Review Webinar - 29  Identify a strength or weaknesses and indicate why it is one and why it is important  “The evaluation plan includes a competent, independent evaluator, … and it will guide the investigators as the project evolves and provide a measure of its effectiveness at the end.”  Identify a strength, indicate why it is one, why it is important, and how it could be improved  “The evaluation plan is a strength because it includes a competent, independent evaluator, … and this will guide the investigators as the project evolves and provide a measure of its effectiveness at the end. It could be improved by adding …”

Mock Review Webinar - 30  Be constructive in your written comments  Provide suggestions to help applicants improve their proposals  Do not be overly critical in your ratings  Most meritorious proposals (fundable proposals) have some weaknesses  Some are correctable through negotiations

Mock Review Webinar - 31  How many strengths did you note?  How many weaknesses did you note?  Are these numbers consistent with your rating?  Do your statements indicate why an item is a strength or weakness?  Do they indicate why a strength or weakness is important?

Mock Review Webinar - 32

Mock Review Webinar - 33  Local teams discuss the strengths and weaknesses identified in the individual reviews (20min)  Team “scribe” takes notes during discussion  Teams locally report/discuss results with facilitator’s guidance (10min)  Facilitators guide discussion and selection of an individual to report to the full virtual group  Participating organizations will be called upon by webinar moderator – PLEASE BE READY TO REPORT BACK  Program Directors will offer comments on reports (15min)

Mock Review Webinar - 34  Redundancy is OK – indicative of common perception  Responses will be solicited as follows:  Intellectual Merit  Strengths  Weaknesses  Broader Impacts  Strengths  Weaknesses  Overall Perception  Is this a proposal worthy of funding?

Mock Review Webinar - 35

Mock Review Webinar - 36  Redundancy is OK – indicative of common perception  Responses will be solicited as follows:  Intellectual Merit  Strengths  Weaknesses  Broader Impacts  Strengths  Weaknesses  Overall Perception  Is this a proposal worthy of funding?

Mock Review Webinar - 37

Mock Review Webinar - 38

Mock Review Webinar - 39

Mock Review Webinar - 40

Mock Review Webinar - 41  Individuals consider ways that the proposal could be improved – create a list (5min)  Local teams discuss the suggestions for improvement (10min)  Team “scribe” takes notes during discussion  Teams locally report/discuss results with facilitator’s guidance (5min)  Facilitators guide discussion and selection of an individual to report to the full virtual group  Participating organizations will be called upon by webinar moderator – PLEASE BE READY TO REPORT BACK  Program Directors will offer comments on reports (15min)

Mock Review Webinar - 42  Try to minimize redundant responses  Identify approaches for:  Building on strengths  Overcoming weaknesses  Responses should include potential improvements to:  Idea  Project infrastructure  Project implementation plan  Evaluation plan  Dissemination plan  Proposal participation/involvement

Mock Review Webinar - 43

Mock Review Webinar - 44  Try to minimize redundant responses  Identify approaches for:  Building on strengths  Overcoming weaknesses  Responses should include potential improvements to:  Idea  Project infrastructure  Project implementation plan  Evaluation plan  Dissemination plan  Proposal participation/involvement

Panel Summary and Reviewer Comments

Mock Review Webinar - 46  The proposal is the evolution of the successful studio pedagogy developed at RPI to a mobile studio pedagogy  In this way the proposal builds on a instructional methodology accepted in the STEM literature.  The primary innovation is the development of a low cost virtual instrument board that, when coupled with a computer/laptop, provides students with a portable experimental platform.  The proposal is student focused and grounded in the STEM knowledge base.

Mock Review Webinar - 47  This experimental platform serves as a low cost replacement for the laboratory equipment found in the original RPI studio classrooms and provides the students with an "anywhere, anytime" experimental platform. The PIs are well qualified to carry out the work.  In addition, the proposal was responsive to the cyclic model of knowledge creation contained in the CCLI request for proposals.  The panel felt the proposal would be strengthened by a more detailed evaluation plan with clearly stated, measurable outcomes.

Mock Review Webinar - 48  The broader impacts of the proposed work include the potential to significantly impact a large number of under served students by providing a low cost, portable experimental platform.  In addition, the inclusion of an historically black college or university and a community college should magnify the impact of the proposed pedagogy.  This partnership was made concrete through an instructor exchange program between the institutions.

Mock Review Webinar - 49  While the evaluation team appears to have worked together for several years, some basic information about the qualifications of those individuals should have been included in the proposal.  It is not clear that it will positively impact student learning -- just because a student has the means to perform an experiment 24/7 does not mean that they will.  Proposal could benefit from clarifying or identifying the role of laboratory experiences where actual industrial-quality laboratory equipment is utilized, as too much reliance on the proposed pedagogy might also leave the student with limited experience in the use of the real industrial grade and scale laboratory equipment.  The proposal could be strengthened with more definition of the role of each partner.

Mock Review Webinar - 50  Moving from the original studio to the mobile studio essentially means that the new work is now based on an untested model. The proposal could be strengthened by addressing this issue.  It has not been demonstrated that the mobile studio is as effective as the original studio with its significant faculty involvement and structure.  This proposal would have been much stronger as a Phase I or if some preliminary data on efficacy were given.  The proposal mentions a test that was done at Howard University with this work but no details are given.  It fails to explicitly address how it will focus on components of the cyclic model for innovation in STEM education.

Mock Review Webinar - 51

Mock Review Webinar - 52 Reviewers have:  Many proposals  Ten or more from several areas  Limited time for your proposal  20 minutes for first read  Different experiences in review process  Veterans to novices  Different levels of knowledge in proposal area  Experts to outsiders  Discussions of proposals’ merits at panel meeting  Share expertise and experience 52

Mock Review Webinar - 53  Write down a list of suggestions (guidelines) that you would suggest that a colleague should follow - to deal with practical aspects of the Review Process  2 minutes 53

Mock Review Webinar - 54  Use good style (clarity, organization, etc.)  Be concise, but complete  Write simply but professionally  Avoid jargon and acronyms  Check grammar and spelling  Use sections, headings, short paragraphs & bullets (Avoid dense, compact text)  Reinforce your ideas  Summarize; Highlight (bolding, italics)  Give examples 54

Mock Review Webinar - 55  Use tables, figures – where it makes sense  Reinforce your ideas  Summarize; Highlight (bolding, italics)  Give examples 55

Mock Review Webinar - 56  Provide appropriate level of detail  Pay special attention to Project Summary  Summarize goals, rationale, methods, and evaluation and dissemination plans  Address intellectual merit and broader impacts  Explicitly and independently  Three paragraphs with headings:  “Summary”  “Intellectual Merit”  “Broader Impacts” 56

Mock Review Webinar - 57  Follow the solicitation and GPG  Adhere to page, font size, and margin limitations  Use allotted space but don’t pad the proposal  Follow suggested (or implied) organization  Use appendices sparingly (check solicitation to see if allowed)  Include letters showing commitments from others  “Support letters” are not allowed  Avoid form letters 57

Mock Review Webinar - 58  Prepare credible budget  Consistent with the scope of project  Clearly explain and justify each item  Address prior funding when appropriate  Emphasize results  Sell your ideas but don’t over promote  Proofread the proposal  “Tell a story” and turn a good idea into a competitive proposal 58

Mock Review Webinar - 59  Good idea + need  Right people + infrastructure  Assessment of outcomes that measure effect on student learning (with goals/objectives linked to evaluation)  Active dissemination plan  Efforts to broaden participation of underrepresented groups

Mock Review Webinar - 60

Mock Review Webinar - 61  Individuals consider: What have I learned today that I will be able to used in preparing my next proposal?  Create a list (5min)  Share your list with local participants (5min)  Facilitators report back to virtual group (5min)  Participating organizations will be called upon by webinar moderators – PLEASE BE READY TO REPORT BACK

Mock Review Webinar - 62

Mock Review Webinar - 63  For an impact on undergraduate education to be transformative, it needs to provoke a shift in learning

Mock Review Webinar - 64  Model good practices that increase/improve learning  Facilitate direct experience with the methods and processes of inquiry/experimentation  Empower the student  Foster a sense of wonder, the excitement of discovery and “out-of-the-box” thinking  Enhance ability to produce innovative results  Products  Solutions to problems

Mock Review Webinar - 65  Good idea + need  Right people + infrastructure  Can be readily adopted at other sites  Assessment of outcomes that measure effect on student learning (with goals/objectives linked to evaluation)  Active dissemination plan  Shows promise for institutionalization  Efforts to broaden participation of underrepresented groups

Mock Review Webinar - 66

Ning Fang – Don Millard –