Presentation is loading. Please wait.

Presentation is loading. Please wait.

ARC Discovery Projects Workshop Faculty of Science Professor Helena Nevalainen, ARC College of Experts, BSB Panel Professor Bill Griffin, ARC College of.

Similar presentations


Presentation on theme: "ARC Discovery Projects Workshop Faculty of Science Professor Helena Nevalainen, ARC College of Experts, BSB Panel Professor Bill Griffin, ARC College of."— Presentation transcript:

1 ARC Discovery Projects Workshop Faculty of Science Professor Helena Nevalainen, ARC College of Experts, BSB Panel Professor Bill Griffin, ARC College of Experts, PCE Panel 10 November 2009

2 ARC Discovery Projects and Fellowships Overview: some statistics to begin Process: the black box revealed Track record: we believe you (mostly) Early Career Researchers/Fellowships: where are you in your career path? Body of proposal: logical, exciting, a good story Budget tips: thrifty is trendy in the ARC (covered in January Workshop) Rejoinders: you must write one; short, not personal, opportunity to update

3 Discovery Projects: No. Applications & Results (From ARC Website) ARC as a whole 200520062007200820092010 No.3,4413,7664,0474,1214,1644,085 No. Funded 1,053917822878845925 Total Funds 297.5 Million 273.6 Million 274.4 Million 300.8 Million 288.4 Million 325.6 Million ARC Success Rate 30.9%24.5%20.4%21.4%20.4%22.7%

4 ARC Panels BSB -Biological Sciences & Biotechnology EE -Engineering & Environmental Sciences HCA -Humanities & Creative Arts MIC -Mathematics, Information & Communication Sciences PCE -Physical Chemical & Earth Sciences SBE -Social, Behavioural & Sciences

5 Discovery Projects: Application Results by Panel Panel200820092010 BSB21.4%18.4%22.0% EE19.6%20.5%23.2% HCA21.0%19.5%24.9% MIC21.2%23.2%21.6% PCE22.6%19.5%22.8% SBE21.7% 22.5% Overall21.4%20.4%22.7%

6 Discovery Project Proposal Selection Criteria CriteriaWeighting _______________________________________________________________ - Researcher Track Record & capacity to undertake40% the research - Proposal Project Content - Project Significance & Innovation30% - Project Approach20% - National Benefit10% __________________ 100%

7 What happens to your Proposal? - Each College of Experts member reviews >100 DP applications. Each grant is read by 2 College of Experts members. - Ozreader = discipline expert drawn from pool, reviews ≤20 applications, Assigned by ARC Executive Director - Intreaders = real experts (can be for specific aspects of applications), reviews ≤6 applications. - ARC Executive Director takes care of particular panel, plus other schemes. Assigns Ozreaders to particular grants taking advice from College of Experts (EAC) members. But this is done mainly on the codes and keywords you provide!

8 Key words are VIP

9 What happens to your proposal? You get comments from readers, write a rebuttal Panel members see each other’s scores, and the rebuttals, adjust to minimise differences ARC computes WAPR (rankings) August meeting -- start with highest WAPR and work down until $$$ are gone -- best funding for highest-ranked proposals Winners notified -- October/November

10 Ranking 1.Each reviewer’s weighted score is tallied DP=(TR*0.4) + (S/I*0.3) + (Appr*0.2) + (NB*0.1) LP=(TR*0.2) + (ISI*0.25) + (Appr*0.2) + (NB*0.1) + (Commit*0.25) 2.Applications ranked 1 to N based on weighted scores (N = total number of grants reviewed by reviewer) 3.Application rank is converted to percent rank 1.Application percent rank is weighted according to the number of applications reviewed by the respective reviewers and a weighted average is calculated (WAPR).

11 Recommendation to the Minister research branchARC panel ED EAC 2 EAC 1 applicant Comments only rejoinder scores and ranking discussion ± reranking selection meeting $? Budget considered in detail at this point

12 A Discovery Project Application -Proposal is written like a top journal article from Introduction to end of the Method section (include method analysis) -Is an academic argument on how to advance the academic field/knowledge to prove this is a significant idea(s) -Has specific, consistent, meaningful research objectives; Research Question/hypotheses derived from a critical literature review; RQ’s matched to studies proposed in the Approach -Proposal should be consistent with scoring of selection criteria: - Track Record (eg carefully compose team), S&I (eg do a crit. lit review), Approach (eg specific & matched to research questions or hypotheses), & National Benefit (application of the results) -Do multiple drafts, get feedback, do a pilot study – needs to be written very well so it is good relative to the best -Based on Funding Rules & Instructions to Applicants http://www.arc.gov.au/ncgp/dp/dp_fundingrules.htm

13 Discovery Project Application Form PART A – Administrative Summary A2 – Proposal Title A5 – Summary of Proposal (100 words) Summary of National/Community Benefit (100 words) A6 – Keywords, Research Classifications PART B - Personnel, including Track Record and Research Record Relative to Opportunity PART C – Budget PART D – Report on any current ARC Projects PART E – Project Description

14 Proposal Title & 100 Word Summary Think of a catchy title, eg. “Body fluids: sweet protection against infection?” Two 100-word summaries: 1.“Abstract” that needs to give a clear idea of what is proposed and why it is important (outcomes) -- written for a scientist. 2.Summary for propaganda purposes (for use by media when successful ones are announced -- simple language, emphasis on importance and potential outcomes (national benefit)

15 TRACK RECORD Based on past achievement – high level of input from broad base of outside sources (eg journals, societies…) Correlates with other criteria, as well – feasibility (approach) – significance and innovation TR scores tend to add up, ie weighted to higher track records of CIs and PIs (rather than averaged) plenty of opportunity to make clear in application

16 Papers/yr 1st authorJournalPhD stud.ECRAbst. Major impact Other sign SCO RE quality- adusted (Teaching)(Research) impactcurrent on field track record <1=70, 1=75, 2=80, <2=70, 2=75, 4=80, >0.5=2, <0.5=-1,0 to 3≥2=22 >2/yr= 1 (reality check) 4=85, ≥5=90 6=85, ≥8 = 90none=-2(% in HI)>5=4 >4=2, <1=-10 to 20 to 3 AB 9024 212100 CD 7721 1 182 E 024 23 90 F 8123 118886 G75 23 21183 H 8021 1185 I 2 11189 J80 2 182 K 8311 21 189 L85 11 11290 M85 1 1187 N85 21 2190 O82 2 83 P70 11 72 Q83 12 18790 R 822 11 85 S89 11 1192 T90 2 22297 U82 2 184 V86 12 189 W78 21 2 184 X83 1 185 Y82 21 2 18883

17 B10 research record relative to opportunity 1 read and follow the instructions most significant contributions to this research field (B10.1) – don’t hold back (we will believe you, generally) (but over-the-top is picked up and discounted) – focus on your impact directly (narrowly) on research outcomes in this half page, ie how you have changed/moved this area of research

18 Do not do this: “I have carried out research in area x for 20 years and have published significant papers, and have obtained 20 mill in research funds blah blah.” Do this: – I discovered x (see papers 1,2,3) which resulted in an international effort to find y (citations n). – I discovered that a results in b such that the previously accepted paradigm was incorrect (papers 4,5,6). This has led to numerous other groups…. – The outcomes of my research have resulted in z being used by …… in a commercial……. (evidence, see publications 5,6,7+) B10 research record relative to opportunity 2

19 B10 research record relative to opportunity 3 Significant publications in last 5 yrs (B10.2) Ensure that authorship role is clear on all publications (account for differences in conventions of discipline, journal, team) – do not assume all reviewers will know conventions – place explanation in an obvious place Enhance basic information with evidence of impact (think RQF, ERA whatever), succinctly include information on: – reviews, sales, other impact of books – impact factors, citations (H-index), other impact of articles – acceptance rate (if appropriate), other impact of conference papers – relevance/impact indices of other publications

20 B10 research record relative to opportunity 4 Ten best career publications (B10.3) Unlimited space: Provide clear evidence of impact (think again of RQF): – number of times publication has been cited, referred to, etc… – any type of (positive) editorial reaction to publication – practical outcomes – at some point it would be worth giving average citation rates compared to average in field etc.

21 B10 research record relative to opportunity 5 10.4 other evidence of impact and contributions half page continue theme begun in 10.1, broadening emphasis to wider recognition of your research record prizes, awards, patents, experience in industry, editorial boards… – reviewer is good, associate editor (expert panel) is better – member of society is good, officer is better – participant in conference is good, organiser is better – presentations are good, invitations are better broader recognition of your research (eg consultancies) place all achievements in context, eg – award given every 5 years – first non-American to receive…

22 B10 relative to opportunity 6 10.5 other aspects of career…relevant to assessment half page use as required be clear, succinct and reasonable – major illnesses or injuries – relocations – time off for maternity/paternity leave – changes of career, research directions – Other – Can take the opportunity to present reviews of papers recently submitted.

23 B10 research record relative to opportunity 7 10.6 fellowship supplementary information one page fellowships are good additions to applications follow instructions, addressing all points Can be at same institution but give strong reasons

24 No. & success Rate of ECR-Only Proposals (≤ 5 years since PhD) for 2010 DisciplineECR-Only Proposals Received ECR-Only Proposals Approved for Funding Success rate ECR-Only Proposals: % of Total Approved funding BSB971919.6%10.3% EE891516.9%10.4% HCA1141311.4%7.1% MIC1302116.2%13.5% PCE931516.1%8.5% SBE1111513.5%6.9% All6349815.5%9.4%

25 2010 Discovery Project Fellowships Need excellent scores on both (1) Project and (2) Fellow - Australian Postdoc Fellowships (≤3 years since PhD): - 2010 Success Rate = 17.1% - Can be 3 years (100% ARC Salary Support) or 4 years (75%) - Australian Research Fellows/Queen Elizabeth II Fellows - ≤ 8 years since PhD, also ≤ 13 years if previously had ARF/QEII: - 2010 Success Rate = 17.8% - Can be 50% ARC Salary Support. Has a much better success rate - Australian Professional Fellowships - ≤ 13 years since PhD - 2010 Success Rate = 16.3% - Can be 50% ARC Salary Support. Has a much better success rate

26 Body of proposal Construct for the right audience (ie College of Experts, Oz- readers and Int-readers) Consider that CoE member might not know the field Make it exciting -- but watch out for obvious hype…. Have clear aims and hypotheses linked to approach Use preliminary data (VIP), but make sure that it reproduces well in copies, do not use small fonts Show how your previous research is relevant and how you are leading the field Keep reminding yourself of the assessment weightings:

27 Rejoinders Always provide one Usually used to discard or reduce weighting of an assessor that may have been too harsh. It does make an impact so construct it carefully. Do not get personal. Can provide additional findings or publications. Update your progress on the topic since submission.

28 Budget tips No point going for teaching relief (BSB) No point indexing salaries Provide good justification (will reduce degree of cut depending on ranking) -- including the roles of requested Research Assistants, etc MU very good (uniquely?) at providing HDR scholarships -- don’t request them, but put them in as University’s contribution Do not make project absolutely dependent on a large budget. Remember average cut is about 40% (BSB), average budget is about 300- 350K over 3 years.


Download ppt "ARC Discovery Projects Workshop Faculty of Science Professor Helena Nevalainen, ARC College of Experts, BSB Panel Professor Bill Griffin, ARC College of."

Similar presentations


Ads by Google