Towards indicators for ‘opening up’ science and technology policy Ismael Rafols, Tommaso Ciarli, Patrick van Zwanenberg and Andy Stirling Ingenio (CSIC-UPV),

Slides:



Advertisements
Similar presentations
Overlay Maps of Science (2010 update)
Advertisements

Overlay Maps of Science Ismael Rafols 1,2, Alan Porter 2 and Loet Leydesdorff 3 1 SPRU, University of Sussex, Brighton 2 School of Public Policy, Georgia.
Discourses and Framings of Climate Change: What Literatures Do We Need to Review? To realize synergies there is a need to indentify common objectives for.
Tailoring Foresight to Field Specificities Antoine Schoen, Université Paris Est (LATTS, ESIEE Management) and JRC-IPTS Totti Könnölä, JRC-IPTS Philine.
J. David Tàbara Institute of Environmental Science and Technology Autonomous University of Barcelona Integrated Climate Governance.
ESRC/DfID Poverty Alleviation Conference 9/9/14
Measuring science with meaningful metrics Interdisciplinarity and transdisciplinarity Edwin Horlings 23 januari 2014.
See ( OECD-JRC handbook on CI The ‘pros’: Can summarise complex or multi-dimensional issues in view of supporting decision-makers.
Good Evaluation Planning – and why this matters Presentation by Elliot Stern to Evaluation Network Meeting January 16 th 2015.
Horizon 2020 and its impact on university research strategies Univ.-Prof. Dr. Susanne Weigelin-Schwiedrzik Vice-Rector for Research and Career Development.
Linking the Fairs to the 2013 Ontario Curriculum Social Studies 1 to 6 and History and Geography 7 and 8.
1 EU-SPRI Conference Manchester, 19 June 2014 Avian influenza narratives and research landscapes Matthew L. Wallace · INGENIO (CSIC-UPV) Ismael Rafols.
Research evaluation at CWTS Meaningful metrics, evaluation in context
Asha Balakrishnan Vanessa Peña Bhavya Lal Task Leader November 5, 2011
Working together across disciplines Challenges for the natural and social sciences.
Disciplinary Coherence of Units These slides complement the article How journal rankings can suppress interdisciplinary research: A comparison between.
Disciplinary diversity of the Association of Business Schools’ rankings These slides complement the article How journal rankings can suppress interdisciplinary.
Intermediation of Units These slides expand the article How journal rankings can suppress interdisciplinary research. A comparison between innovation studies.
Disciplinary Diversity of Units These slides expand the article How journal rankings can suppress interdisciplinary research: A comparison between innovation.
3. Challenges of bibliometrics DATA: many problems linked to the collection of data FIELDS – DISCIPLINES : various classifications, not satisfactory INDICATORS.
1 Academic Rankings of Universities in the OIC Countries April 2007 April 2007.
PPA 502 – Program Evaluation
IFSA 2004 Workshop 5 Combined micro-economic and ecological assessment tools for sustainable rural development in the context of Farming Systems Analysis.
T H O M S O N S C I E N T I F I C Editorial Development James Testa, Director.
Quality teaching in higher education, an OECD approach Fabrice Hénard, December 2010.
Week 3 1 S514: Social Aspects of IT. 2 Disciplines related to SI Social ScienceManagementComputer Sci. Science & Technology Studies MIS Information Science.
Strengthening the Science-Policy Interface Towards Sustainable Development Dr. Gisbert Glaser International Council for Science (ICSU) Side Event at 2.
Guillaume Rivalle APRIL 2014 MEASURE YOUR RESEARCH PERFORMANCE WITH INCITES.
M254 Arts & Engineering Research Fall 2013, Studio 2611, Elings Hall Tues-Thurs 12:00 to 1:50pm Experimental.
OST Workshop 12 May 2014, Paris The debate on uses and consequences of STI indicators Paul Wouters, Sarah de Rijcke and Ludo Waltman, Centre for Science.
Section 2: Science as a Process
EVALUATING NSF EPSCOR INITIATIVES: TRACKING AND ASSESSING THE DEVELOPMENT OF EPSCOR-SUPPORTED RESEARCH CAPACITY Julia Melkers, Associate Professor School.
International perspectives on e- learning: mapping strategy to practice Gráinne Conole Towards a pan-Canada e-learning research agenda.
Understanding Data Analytics and Data Mining Introduction.
WP 5: Assessment of Transition Pathways to Regional Sustainability of Agriculture “to build on findings from WP3 to inform participatory identification.
Innovation for Growth – i4g Universities are portfolios of (largely heterogeneous) disciplines. Further problems in university rankings Warsaw, 16 May.
Bibliometrics: coming ready or not CAUL, September 2005 Cathrine Harboe-Ree.
1 Issues in Assessment in Higher Education: Science Higher Education Forum on Scientific Competencies Medellin-Colombia Nov 2-4, 2005 Dr Hans Wagemaker.
Well-being and multidimensional deprivation: some results from the OECD Better Life Initiative Nicolas Ruiz.
Advancing foresight methodology through networked conversations Ted Fuller Peter De Smedt Dale Rothman European Science Foundation COllaboration in Science.
9 December 2005 Toward Robust European Air Pollution Policies Workshop, Göteborg, October 5-7, 2005.
Integrated Risk Management Charles Yoe, PhD Institute for Water Resources 2009.
STRATEGIC ENVIRONMENTAL ASSESSMENT METHODOLOGY AND TECHNIQUES.
Reconstructing unruly ecological complexities. ecological complexity poses many challenges to conventional scientific ways of knowing.
Technology Enhanced Learning at University - How can learning enhancement be demonstrated? Adrian Kirkwood & Linda Price IET, The Open University.
Overlay Maps of Science Ismael Rafols 1,2, Alan Porter 2 and Loet Leydesdorff 3 2 SPRU, University of Sussex, Brighton 2 School of Public Policy, Georgia.
Advanced Decision Architectures Collaborative Technology Alliance An Interactive Decision Support Architecture for Visualizing Robust Solutions in High-Risk.
Introduction to Earth Science Section 2 Section 2: Science as a Process Preview Key Ideas Behavior of Natural Systems Scientific Methods Scientific Measurements.
“Social” Multicriteria Evaluation: Methodological Foundations and Operational Consequences Giuseppe Munda Universitat Autonoma de Barcelona Dept. of Economics.
1 Making a Grope for an Understanding of Taiwan’s Scientific Performance through the Use of Quantified Indicators Prof. Dr. Hsien-Chun Meng Science and.
Universiteit Antwerpen Conference "New Frontiers in Evaluation", Vienna, April 24th-25th Reliability and Comparability of Peer Review Results Nadine.
Three Critical Matters in Big Data Projects for e- Science Kerk F. Kee, Ph.D. Assistant Professor, Chapman University Orange, California
DARM 2013: Assessment and decision making Mikko V. Pohjola, Nordem Oy, (THL)
Toward a vulnerability/adaptation methodology Thomas E. Downing Stuart Franklin Sukaina Bharwani Cindy Warwick Gina Ziervogel Stockholm Environment Institute.
How to measure the impact of R&D on SD ? Laurence Esterle, MD, PhD Cermes and Ifris France Cyprus, 16 – 17 October L. ESTERLE Linking science and.
S ystems Analysis Laboratory Helsinki University of Technology 1 Decision Analysis Raimo P. Hämäläinen Systems Analysis Laboratory Helsinki University.
A New Measure of Knowledge Diffusion Stephen Carley; Alan Porter Georgia Tech.
Methods for assessing current and future coastal vulnerability to climate change Dr. Francesca Santoro Università Ca’ Foscari and Euro-Mediterranean Centre.
DELIBERATION JACQUIE BURGESS DEPARTMENT OF GEOGRAPHY, UCL EUROPEAN RESEARCH 2002 CONFERENCE.
Working with your archive organization: Broadening your user community Robert R. Downs, PhD Socioeconomic Data and Applications Center (SEDAC) Center for.
Middle Years Programme The unique benefits of the MYP.
Working with Your Archive : Broadening Your User Community Robert R. Downs, PhD NASA Socioeconomic Data and Applications Center (SEDAC) Center for International.
Bringing Diversity into Impact Evaluation: Towards a Broadened View of Design and Methods for Impact Evaluation Sanjeev Sridharan.
Towards indicators for ‘opening up’ science and technology policy Ismael Rafols Ingenio (CSIC-UPV), Universitat Politècnica de València SPRU (Science Policy.
European Research 2002, Workshop on Uncertainty, Brussels, November 2002 Extended Peer Reviews in Science for Sustainability 1 Extended Peer Reviews Joachim.
Disciplinary structure and topical complexity in SSH—the IMPACT EV mission Sándor Soós, András Schubert, Zsófia Vida.
Giuseppe Munda Universitat Autonoma de Barcelona
Interpreting Foresight Process Impacts:
Opening the black box of national research performance
Dr Hayley MacGregor, 13 Dec 2011
Presentation transcript:

Towards indicators for ‘opening up’ science and technology policy Ismael Rafols, Tommaso Ciarli, Patrick van Zwanenberg and Andy Stirling Ingenio (CSIC-UPV), Universitat Politècnica de València & SPRU —Science and Technology Policy Research, University of Sussex Stockholm, October 2013 Building on work with Loet Leydesdorff and Alan Porter

Paper born out of the reflection on the contrast between interdisciplinary research and journal rankings Interdisciplinary maps versus rankings

On the role of scientific advice in policy (scientometric is the science of science –hence scientific advice) The linearity-autonomy model of scientific advice (Jasanoff, 2011)  Scientific knowledge is the best possible foundation for public decisions  Scientists should establish the facts that matter independently. –S&T indicators produce evidence of these facts. However, this (enlightenment) model has been challenged  The mechanisms to establish facts and make decisions is a social process –“knowledge enables power, but power structures knowledge” (Stirling, 2012)  Modes of advice: The pure scientist vs. honest broker (Pielke, 2007) What is (should be) the role of STI indicators in policy advice? Closing down vs. Opening up

The challenge Problems with current use of S&T indicators Use of conventional S&T indicators is en *problematic* (as many technologies, in particular those closely associated with power, e.g. nuclear)  Narrow inputs (only pubs!)  Scalar outputs (rankings!)  Aggregated solutions --missing variation  Opaque selections and classifications (privately owned databases)  Large, leading scientometric groups embedded in government / consultancy, with limited possibility of public scrutiny  Sometimes even mathematically debatable  Impact Factor of journals (only 2 years, ambiguity in document types)  Average number of citations (pubs) in skewed distributions

From S&T indicators for justification and disciplining… Justification in decision-making Weak justification, “Give me a number, any number!” Strong justification, “Show in numberrs that X is the best choice!” S&T Indicators have a performative role:  They don’t just measure. Not ‘just happen to be used’ in science policy (neutral)  Constitutive part incentive structure for “disciplining” (loaded)  They signal to stakeholders what is important. Institutions use these techniques to discipline subjects  Articulate framings, goals and narratives on performance, collaboration, interdisciplinarity…

… towards S&T indicators as tools for deliberation Yet is possible to design indicators that foster plural reflection rather than justifying or reinforcing dominant perspectives This shift is facilitated by trends pushed by ICT and visualisation tools  More inputs (pubs, pats, but also news, webs, etc.)  Multidimensional outputs (interactive maps)  Multiple solutions -- highlighting variation, confidence intervals  More inclusive and contrasting classifications (by-passing private data ownership? Pubmed, Arxiv)  More possibilities for open scrutiny (new research groups)

1. Conceptual framework: “broadening out” vs. “opening up” policy appraisal

Policy use of S&T indicators: Appraisal Appraisal: ‘the ensemble of processes through which knowledges are gathered and produced in order to inform decision-making and wider institutional commitments’ Leach et al. (2008) Breadth: extent to which appraisal covers diverse dimensions of knowledge Openness: degree to which outputs provide an array of options for policies.

Policy use of S&T indicators: Appraisal Appraisal: ‘the ensemble of processes through which knowledges are gathered and produced in order to inform decision-making and wider institutional commitments’ Leach et al. (2010) Example: Allocation of resources based on research “excellence” Breadth: extent to which appraisal covers diverse dimensions of knowledge Narrow: citations/paper Broad: citations, peer interview, stakeholder view, media coverage, altmetrics Openness: degree to which outputs provide an array of options for policies. Closed: fixed composite measure of variables  unitary and prescriptive Open: consideration of various dimensions  plural and conditional

narrow broad closing-downopening-up range of appraisals inputs (issues, perspectives, scenarios, methods) effect of appraisal ‘outputs’ on decision-making Leach et al Appraisal methods: broad vs. narrow & closing vs. opening

narrow broad closing-downopening-up range of appraisals inputs (issues, perspectives, scenarios, methods) effect of appraisal ‘outputs’ on decision-making Appraisal methods: broad vs. narrow & close vs. open cost-benefit analysis open hearings consensus conference scenario workshops citizens’ juries multi-criteria mapping q-method sensitivity analysis narrative-based participant observation decision analysis risk assessmentstructured interviews Stirling et al. (2007)

narrow broad closing-downopening-up range of appraisals inputs (issues, perspectives, scenarios, methods) effect of appraisal ‘outputs’ on decision-making Appraisal methods: broad vs. narrow & closing vs. opening Most conventional S&T indicators??

narrow broad closing-downopening-up range of appraisals inputs (issues, perspectives, scenarios, methods) effect of appraisal ‘outputs’ on decision-making Broadening out S&T Indicators Conventional S&T indicators?? Broadening out Incorporation plural analytical dimensions: global & local networks hybrid lexical-actor nets etc. New analytical inputs: media, blogsphere.

narrow broad closing-downopening-up range of appraisals inputs (issues, perspectives, scenarios, methods) effect of appraisal ‘outputs’ on decision-making Appraisal methods: broad vs. narrow & closing vs. opening Journal rankings University rankings Unitary measures that are opaque, tendency to favour the established perspectives … and easily translated into prescription European Innovation Scoreboard

narrow broad closing-downopening-up range of appraisals inputs (issues, perspectives, scenarios, methods) effect of appraisal ‘outputs’ on decision-making Opening up S&T Indicators Conventional S&T Indicators?? opening-up Making explicit underlying conceptualisations and creating heuristic tools to facilitate exploration NOT about the uniquely best method Or about the unitary best explanation Or the single best prediction

2. Examples of Opening Up a.Broadening out AND Opening up b.Opening up WITH NARROW inputs

narrow broad closing-downopening-up range of appraisals inputs (issues, perspectives, scenarios, methods) effect of appraisal ‘outputs’ on decision-making 1. Preserving multiple dimensions in broad appraisals Conventional S&T indicators?? Leach et al Broadening out opening-up

Composite Innovation Indicators (25-30 indicators) European (Union) Innovation Scoreboard Grupp and Schubert (2010) show that order is highly dependent on indicators weightings. Sensitivity analysis

Solution: representing multiple dimensions (critique by Grupp and Schubert, 2010) Use of spider diagrams allows comparing like with like U-rank, University performance Comparison tools (Univ. Twente) 5.4 Community trademarks indicator

U-Map: Comparison of Universities in Multiple Dimensions

2. Examples of Opening Up b. Opening up WITH NARROW inputs

narrow broad closing-downopening-up range of appraisals inputs (issues, perspectives, scenarios, methods) effect of appraisal ‘outputs’ on decision-making Opening up S&T Indicators Conventional S&T Indicators?? Leach et al opening-up Making explicit underlying conceptualisations and creating heuristic tools to facilitate exploration NOT about the uniquely best method Or about the unitary best explanation Or the single best prediction

Manchester Inn Inst Warwick Business School Disciplinary Diversity of Publications Variety1920 Shannon Entropy Disciplinary Diversity of References Variety1720 Shannon Entropy Disciplinary diversity of Citations Variety2224 Shannon Entropy Interdisciplinarity as diversity

Rafols, Porter and Leydesdorff (2010) Cognitive Sci. Agri Sci Biomed Sci Chemistry Physics Engineering Env Sci & Tech Matls Sci Infectious Diseases Psychology Social Studies Clinical Med Computer Sci Business & MGT Geosciences Ecol Sci Econ Polit. & Geography Health & Social Issues A Global Map of Science 222 SCI-SSCI Subject Categories

Warwick Business School Subject Categories of publications Nodes labelled if >0.5% publications

Manchester MIoIR Subject Categories of publications Nodes labelled if >0.5% publications

Heuristics of diversity (Stirling, 1998; 2007) Diversity: ‘attribute of a system whose elements may be apportioned into categories’ Characteristics: Variety: Number of distinctive categories Balance: Evenness of the distribution Disparity: Degree to which the categories are different. Variety BalanceDisparity Herfindahl (concentration):  i p i 2 Shannon (Entropy):  i p i ln p i Dissimilarity:  i d i Generalised Diversity (Stirling)  ij(i  j) (p i p j )  (d ij ) 

Manchester Innov Inst Warwick Business S Diversity of Publications Variety 1920 Balance Disparity Shannon Entropy Rao-Stirling Diversity Diversity of References Variety 1720 Balance Disparity Shannon Entropy Rao-Stirling Diversity Diversity of Citations Variety 2224 Balance Disparity Shannon Entropy Rao-Stirling Diversity Comparing degree of interdisciplinarity of two university units: Manchester is more??

Multiple concepts of interdisciplinarity: Conspicuous lack of consensus but most indicators aim to capture the following concepts Integration (diversity & coherence) Research that draws on diverse bodies of knowledge Research that links different disciplines Intermediation Research that lies between or outside the dominant disciplines

Diversity ISSTI Edinburgh WoS Cats of references Assessing interdisciplinarity

ISSTI Edinburgh Observed/Expected Cross-citations Coherence Assessing interdisciplinarity

ISSTI Edinburgh References Intermediation Assessing interdisciplinarity

Summary: IS (blue) units are more interdisciplinary than BMS (orange) More Diverse Rao-Stirling Diversity More Coherent Observed/Expected Cross-Citation Distance More Interstitial Average Similarity

2. Excellence: Opening Up Perspectives Provide different perspectives of performance (alternative measures of the same type of indicator)

Are measures of “excellence” consistent and robust? Good Average Bad Van Eck, Waltman et al. (2013) More basic More applied Clinical neurology Is basic always better than applied ? Citations: not stable to changes in classification and granularity (Zitt et al., 2005; Adams et al., 2008).

Measures of “excellence” Which one is more meaningful??

The new Leiden ranking ( ) Different measures of performance MNC, MNCS, MNCJ, Top 10%, Under different conditions (fractional, language) Include confidence interval (bootstrapping)

3. Summary and conclusions

S&T indicator as a tools to open up the debate ‘Conventional’ use of indicators (‘Pure scientist ‘--Pielke)  Purely analytical character (i.e. free of normative assumptions)  Instruments of objectification of dominant perspectives  Aimed at legitimising /justifying decisions (e.g. excellence)  Unitary and prescriptive advice Opening up scientometrics (‘Honest broker’ --Pielke)  Aimed at locating the actors in their context and dynamics  Not predictive, or explanatory, but exploratory  Construction of indicators is based on choice of perspectives  Make explicit the possible choices on what matters  Supporting debate  Making science policy more ‘socially robust’  Plural and conditional advice Barré (2001, 2004, 2010), Stirling (2008)

Strategies for opening up or how to “keep it complex” yet “manageable” Presenting contrasting perspectives  At least TWO, in order to give a taste of choice Simultaneous visualisation of multiple properties / dimensions  Allowing the user take its own perspective Interactivity  Allowing the user give its own weigh to criteria / factors  Allowing the user manipulate visuals.

Is ‘opening up’ worth the effort? (1) Sustaining diversity in S&T system Decrease in diversity. Potential unintended consequence of the evaluation machine: Why diversity matters Systemic (‘ecological’) understanding of the S&T  S&T outcomes depend on synergistic interactions between disparate elements. Dynamic understanding of excellence and relevance  New social needs, challenges, expectations from S&T Manage diverse portfolios to hedge against uncertainty in research  Office of Portfolio Analysis (National Institutes of Health) Open possibility for S&T to work for the disenfranchised  Topics outside dominant science (e.g. neglected diseases)

Is ‘opening up’ worth the effort? (2) Building robustness against bias Do conventional indicators tend to favour incumbents? Hypothesis: Elites and incumbents (directly or not) influence choice of indicators, which tend to benefit them… “knowledge enables power, but power structures knowledge” (Stirling, 2012)  Crown indicator –Standard measure of performance (~ ) –‘systematic underrating of low-ranked scientists’ (Opthof and Leydesdorff, 2010) (Not spotted for 15 years!)  Journal rankings in Business and Management. –systematic underrating of interdisciplinary (heterodox) depts. (Rafols et al., 2012).  Others?? H-index?? –favours established academics over younger.

‘lock-in’ to policy favoured by incumbent power structures multiple practices, and processes, for informing social agency (emergent and unstructured as well as deliberately designed ) complex, dynamic, inter- coupled and mutually- reinforcing socio- technical configurations in science narrow scope of attention Conventional Policy Dynamics SOCIAL APPRAISAL GOVERNANCE COMMITMENTS simple ‘unitary’ prescriptions POSSIBLE FUTURES expert judgements / ‘evidence base’ “best / optimal /legitimate” S&T indicators risk assessment cost-benefit analysis also: restricted options, knowledges, uncertainties in participation incomplete knowledges Res. Excellence $ IIIIII GUIDANCE / NARRATIVE Stirling (2010)

POSSIBLE PATHWAYS MULTIPLE TRAJECTORIES SOCIAL APPRAISAL GOVERNANCE COMMITMENTS broad-based processes of ‘precautionary appraisal’ ‘opening up’ with ‘plural conditional’ outputs to policymaking dynamic portfolios pursuing diverse trajectories viable options under: conditions, dissonant views, sensitivities, scenarios, maps, equilibria, pathways, discourses multiple: methods, criteria, options, frames, uncertainties, contexts, properties, perspectives Breadth, Plurality and Diversity Sustainability $                 Stirling (2010)

S&T indicator as a tools to open up the debate ‘conventional’ use of indicators  Instruments of objectification  Analytical character (i.e. free of normative assumptions)  Aimed at making decisions (e.g. excellence)  Unitary and prescritive advice Opening up scientometrics  Construction of indicators is based on choice of perspectives  implicit normative choice on what matters  Aimed at locating the actors in their context and dynamics  Not predictive, or explanatory, but exploratory  Supporting debate  making science policy more ‘socially robust’  Plural and conditional advice Barré (2001, 2004, 2010), Stirling (2008)

Heuristics of diversity (Stirling, 1998; 2007) Diversity: ‘attribute of a system whose elements may be apportioned into categories’ Characteristics: Variety: Number of distinctive categories Balance: Evenness of the distribution Disparity: Degree to which the categories are different. Variety BalanceDisparity Herfindahl (concentration):  i p i 2 Shannon (Entropy):  i p i ln p i Dissimilarity:  i d i Generalised Diversity (Stirling)  ij(i  j) (p i p j )  (d ij ) 

Rafols, Porter and Leydesdorff (2010) Cognitive Sci. Agri Sci Biomed Sci Chemistry Physics Engineering Env Sci & Tech Matls Sci Infectious Diseases Psychology Social Studies Clinical Med Computer Sci Business & MGT Geosciences Ecol Sci Econ Polit. & Geography Health & Social Issues A Global Map of Science 222 SCI-SSCI Subject Categories CD-ROM version of the JCR of SCI and SSCI of Matrix of cross-citations between journals (9,000 x 9,000) Collapse to ISI Subject Category matrix (222 x 222) Create similarity matrix using Salton ’ s cosine

Diversity indexes Stirling Generalised Diversity

Diversity indexes Stirling Generalised Diversity  =0,  =0 Number of disciplines

Diversity indexes : Stirling Generalised Diversity  =0,  =1 Simpson (Herfindahl) Index

Diversity indexes Generalised Stirling Diversity  =1,  =1 quadratic entropy

Different aspects of diversity are uncorrelated r var,bal = 0.18, p <.001 r var,dis = 0.32, p <.001 r bal,dis = -0.20, p <.001 Yegros et al. (2010) Which diversity measure should we choose?

Leiden Ranking: high correlation but important individual differences Spearman‘s rank correlation coefficient matrix. (Thanks to Daniel Sirtes and Ludo Waltman for sharing this data)