Reward and punishment mechanism for research data sharing

Slides:



Advertisements
Similar presentations
Authorship APS Professional Skills Course:
Advertisements

4. Public Goods.
Thinking Like an Economist
Peter Griffith and Megan McGroddy 4 th NACP All Investigators Meeting February 3, 2013 Expectations and Opportunities for NACP Investigators to Share and.
OPEN ACCESS PUBLICATION ISSUES FOR NSF OPP Advisory Committee May 30, /24/111 |
Producing Archive Ready Data Sets IASSIST 2006 Margaret Hedstrom Jinfang Niu Kaye Marz.
© 2006 Prentice Hall Business Publishing The Economic Way of Thinking, 11/e Heyne/Boettke/Prychitko “The Economic Way of Thinking” 11 th Edition Chapter.
New DFG Information Infrastructure Projects Dr. Stefan Winkler-Nees; Birmingham, 28. March 2011 New DFG Information Infrastructure Projects.
1 Sharing Research Data in Hong Kong (position paper) Professor John Bacon-Shone Associate Director, Knowledge Exchange The University of Hong Kong Forum.
Chapter 1 The Economic Way of Thinking
Pay-for-Performance: A Decision Guide for Purchasers Guide Prepared for: Agency for HealthCare Research and Quality U.S. Department of Health and Human.
CIFREM Doctoral Program in Economics and Management University of Trento, Italy Thesis Proposal Development (July 2006) Francesca Bortolami
1 Ch. 10: The Federal Budget and Fiscal Policy James R. Russell, Ph.D., Professor of Economics & Management, Oral Roberts University ©2005 Thomson Business.
Copyright © 2006 Elsevier, Inc. All rights reserved Chapter 24 Using Nursing Research in Practice.
Economic evaluation of drugs for rare diseases CENTRE FOR HEALTH ECONOMICS K Claxton, C McCabe, A Tsuchiya Centre for Health Economics and Department of.
Pay-for-Performance: A Decision Guide for Purchasers Guide Prepared for: Agency for HealthCare Research and Quality U.S. Department of Health and Human.
Peter Granda Archival Assistant Director / Data Archives and Data Producers: A Cooperative Partnership.
Vision for academic geographic data access Dr David Medyckyj-Scott GRADE Project Director EDINA.
April 29th, Chapter 6 Decision Making 6.1 The Nature of Management decisions 6.1 The Nature of Management decisions 6.2The decision making process.
United Nations Statistics Division Dissemination of IIP data.
Journal of Economic Behavior and Organization Presented by: Kuan Chen.
Institutional Analysis Lecture 6: Collective Action.
Financial Management Chapter 1- Introduction to Accounting & Finance Session Number N1.
FUNDAMENTALS OF CONTROLLING DHANYA ASOKAN M120005MS.
NRF Open Access Statement
Game theory Chapter 28 and 29
Satisfaction Games in Graphical Multi-resource Allocation
Saving Saving money can be difficult. Many people do not know where to start. There are many ways to save and places to cut cost. You will find suggestions.
Essay theme 2: Environmental Politics
A checking account is an account held at a bank or credit union into which account owners deposit funds.
EMBA Presentation October 15, 2016
An Economic Cause of Data Strategy Failure
Chapter 1: The Economic Way of Thinking Section 2: Economic Choice Today: Opportunity Cost (pg.12-17)
Optimal Stopping.
Investment Analysis and Portfolio management
Provide advice in Life Insurance
Commission structutre in an AFP System *
Enrolling in Clinical Trials
Survey Lead – Pier Valitutti Presented by – Nader Afshar
Research with human participants at Carnegie Mellon University
Chapter 17 Suggested Questions: 2, 3, 5, 7
Hillsborough Public School
Types of Financial Institutions, Interest Spread, Risk/Return Relationship, and Savings options SSEPF2:a-d.
CHAPTER 2: THE SUPPLY OF LABOR
Public Finance, 10th Edition
It’s just as exciting as you think!
Agency Requirements: NOAA Administrative Order Management of environmental and geospatial data and information This training module is part of.
Cost of Credit 2015 Educurious Partners--All rights reserved UNIT 3 LESSON 8.
Credit Cards: More Than Plastic
Game theory Chapter 28 and 29
Dissemination Workshop for African countries on the Implementation of International Recommendations for Distributive Trade Statistics May 2008,
NIH Public Access Policy: RPPR and MY NCBI
Computer-Mediated Communication
Major Expenditures: Housing, Transportation and Food
Ian Ramsey C of E School GCSE ICT Smart working Software choices.
Continuous improvement is the process by which organizations frequently review their procedures, aiming to correct errors or problems. The most effective.
Presented to IEEE Standards Education Committee 11 April 2014
University of South Florida
Funding body requirements
Taxes and Taxation.
Determine Different Accounting Methods Used by Different Fund Types
Introduction to Economics
Cost Sharing May 15, 2013.
Another Example Consider a very popular computer game played by millions of people all over the world. The average score of the game is known to be
Provider Peer Grouping: Project Overview
High School Financial Planning Program
Financial Institutions
Data Warehousing Concepts
ePerformance: A Process Crosswalk May 2010
“Take Charge of Your Finances” Advanced Level
Presentation transcript:

Reward and punishment mechanism for research data sharing Jinfang Niu University of Michigan

Concepts and Definitions Research data Data created by scientific researchers primarily for research purposes. Data created primarily for commercial, clinical or administrative purposes, such as user purchase history, patient medical records, or government records are excluded.

Data sharing for collaboration Very large data sets A researcher has collected data, but he/she doesn’t have enough time or capability to analyze the data. One researcher’s data can be used to address some research questions that he is not interested in.

Data sharing for reuse Data have value beyond the purpose for which they were originally collected. Reused for Verification Replication Different research questions Meta-analysis Decision making & information Teaching

Sharing through archives Producers (prepare & deposit) Archive (process & disseminate) Users Funding agency

Publicly funded research data are public goods Non-rivalry: one individual’s use of the data does not reduce the amount available for other people. Not-excluded: no one is excluded from the online data archive whether or not they deposited data.

Theories about public goods Free-riding in the voluntary contribution to public goods tends to be the dominant strategy in a non-cooperative game. The achievement of a pareto - optimal via decentralized methods in the presence of public goods is fundamentally incompatible with individual incentives.

Beyond public goods For normal public goods In data sharing contributors benefit the same as free-riders. Contributors are not harmed from their contributions In data sharing The benefit of data preparation largely goes to the users. Assumption: It is not very likely that a data depositor uses his own data deposited into a data archive Data producers might be harmed by sharing data.

Incentive mechanisms implemented NIH In the case of noncompliance, NIH can take various actions to protect the federal government's interests. In some instances, NIH may make data sharing an explicit term and condition of subsequent awards. It does not offer rewards for doing well in data-sharing plan. ESRC The final payment of an award will be withheld until data has been deposited in accordance with the requirements. The requirements of the data sharing policy are now a condition of funding. Both: citation of data is not mandatory, co-authorship is not allowed as a condition for sharing data.

Producers’ expectations Some sort of acknowledgments for data deposit, such as a certificate If data deposit is mandatory to receive new funding, or a prerequisite for publishing a paper derived from the data … Some grantees are strongly against punishment ---- From survey data

A simple math model Purpose: to show the effects of punishment and reward. Punishment: the data producer will be punished if the effort he spent on data preparation is lower than a threshold. Reward: when the deposited data is used, the producer of the data gets a reward.

Model set-up A data producer: Assumptions has total fund P chooses  for research Chooses e for data preparation (P =  + e) benefits () (’() >0) from research Assumptions () is concave, and ’() >0 The data producer always tries to maximize his benefit.

No reward & no punishment Data producer: Choice: e = 0 Maximized benefit: (P) User: no data to use, benefit: 0 Social benefit: (P)

Punishment only Producer: If e >= e’, benefit: () = (P-e), max: (P - e’) If e < e’, benefit: () – f, (f>0), max: (P) – f

When Ω(P - e’) > (P) – f The data producer would meet the threshold to avoid the punishment. Two explanations: The punishment is severe enough. The threshold is easy to meet. When Ω(P - e’) = (P) – f The data depositor is indifferent between getting punished and meet the threshold.

When Ω(P - e’) < (P) – f The data producer would choose not to prepare data and get punished. Two explanations: the punishment is not severe enough the threshold is too costly to meet. Conclusion: In the punishment only scenario, data producer’s benefit is max [Ω(P - e’), (P) – f].

User when the data producer would rather be punished, benefit: 0. Social benefit: (P) Max [Ω(P - e’), (P) – f] = (P) – f. The data producer loses f, but the funding agency gets f. : (P) – f +f +0 = (P). Too weak punishment or too high a threshold for punishment is not effective.

Social benefit: Ω(P - e’) + v * (e’) When the producer chooses to meet the threshold, The user would use the data with probability (e’) (’( e) > 0, ( 0) =0). Expected benefit: v * (e’) Social benefit: Ω(P - e’) + v * (e’) Max [Ω(P - e’), (P) – f] = (P - e’) when (P) < Ω(P - e’) + v * (e’), the punishment and threshold are effective. When (P) > Ω(P - e’) + v * (e’), the punishment and threshold are not effective.

Reward only When the deposited data is used, the producer gets a reward r. The deposited data has a probability (e) (’( e) > 0, ( 0) =0) of being used. Producer benefit: () + r * (e) <=> (P-e) + r * (e).

If there is a value of e which makes [(P-e) + r If there is a value of e which makes [(P-e) + r * (e)]’ = 0 <=> r * ’(e) = ’(P-e), the producer’s benefit is maximized when the marginal benefit of spending an additional amount of funding on research is equal to the product of reward and the marginal probability of being used.

If the reward(r) is so big that r If the reward(r) is so big that r * ’(e) > ’(P-e) is always true, the data depositor would spend all funding available on data preparation. The utility is maximized when e = P If the reward is so small that r * ’(e) < ’(P-e) is always true, the data producer would not deposit data and there is no data to use. The producer’s benefit is maximized when e=0 The social benefit is (P ). The small reward is not effective.

Suppose the producer’s benefit is maximized at e = e*, User’s benefit: v* (e*). Producer’s benefit: (P-e*) + r * (e*), The funding agency: - r * (e*). Social benefit: (P-e*) + v*( e*). (P) < = (P-e*) + v*( e*), reward causes at least as much social benefit than no reward & no punishment. Warning: do not make the reward too big or too small.

Characterize punishment and reward Coercive and uniform Pros and cons: Makes all data accessible to the public. All data producers have to prepare and deposit data to avoid punishment even if their data sets are not likely to be used.

Cumulative and Average Monthly Access Rates for the 10 Most Frequently (Top 10) and 10 Least Frequently (Bottom 10) Accessed Data Sets. Cumulative Average (per month) Top 10 Bottom 10 5185 27 1037 3 2345 25 260 2.8 2234 24 319 2.7 1651 183 1623 23 232 2.6 1328 21 148 2.3 1267 20 271 2.2 1229 16 137 1.8 932 14 104 1.6 851 11 95 1.2

Rewards Inducive and selective Pros and cons: Related to the actual use of data Difficult to anticipate actual use. Not all data are accessible to the public

Conclusions Publicly funded research data are public goods. They have some features beyond public goods that make free - riding even more attractive under voluntary contribution. Incentive mechanisms should be available to encourage data sharing Punishments or rewards need to be carefully decided to avoid un - intended consequences

Thanks! This research was supported by the National Science Foundation (NSF Award Number IIS-0456022) as part of a larger project entitled “Incentives for data producers to created archive-ready data sets.”