1 Field Experiment on Incentive of Communication, Compilation and Evaluation 1. Motivation for the study 2. What is the Field Experiment 3. Framework of.

Slides:



Advertisements
Similar presentations
Costs and Benefits.
Advertisements

Comparing One Sample to its Population
Michael R. Baye, Managerial Economics and Business Strategy, 3e. ©The McGraw-Hill Companies, Inc., 1999 Managerial Economics & Business Strategy Chapter.
When sanctions cause non-cooperative behavior in a social dilemma situation: a study using the “Industrial Waste Illegal Dumping Game” 13th International.
Hawawini & VialletChapter 7© 2007 Thomson South-Western Chapter 7 ALTERNATIVES TO THE NET PRESENT VALUE RULE.
Review of the Basic Logic of NHST Significance tests are used to accept or reject the null hypothesis. This is done by studying the sampling distribution.
Chapter 14 Comparing two groups Dr Richard Bußmann.
Chapter 9 THE ECONOMICS OF INFORMATION Copyright ©2002 by South-Western, a division of Thomson Learning. All rights reserved. MICROECONOMIC THEORY BASIC.
Copyright © 2004 South-Western 27 The Basic Tools of Finance.
Evaluating Hypotheses
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 5-1 Chapter 5 Some Important Discrete Probability Distributions Statistics.
Lecture 9: One Way ANOVA Between Subjects
Inferences About Means of Two Independent Samples Chapter 11 Homework: 1, 2, 4, 6, 7.
© 2005 The McGraw-Hill Companies, Inc., All Rights Reserved. Chapter 5 Making Systematic Observations.
7-1 Copyright ©2011 Pearson Education, Inc. publishing as Prentice Hall Chapter 7 Sampling and Sampling Distributions Statistics for Managers using Microsoft.
Basic Tools of Finance Finance is the field that studies how people make decisions regarding the allocation of resources over time and the handling of.
Hypothesis Testing Is It Significant?. Questions What is a statistical hypothesis? What is the null hypothesis? Why is it important for statistical tests?
Chapter One of Your Thesis
Non-parametric statistics
Home Production Defined Home production - purposeful activities performed in individual households that result in goods and services that enable a family.
© 2014 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
Chapter 10 Hypothesis Testing
LEARNING PROGRAMME Hypothesis testing Intermediate Training in Quantitative Analysis Bangkok November 2007.
© 2008 McGraw-Hill Higher Education The Statistical Imagination Chapter 9. Hypothesis Testing I: The Six Steps of Statistical Inference.
Testing Hypotheses about a Population Proportion Lecture 29 Sections 9.1 – 9.3 Tue, Oct 23, 2007.
Hypothesis Testing II The Two-Sample Case.
Tests of significance & hypothesis testing Dr. Omar Al Jadaan Assistant Professor – Computer Science & Mathematics.
Copyright © 2005 Brooks/Cole, a division of Thomson Learning, Inc Chapter 24 Statistical Inference: Conclusion.
Psy B07 Chapter 8Slide 1 POWER. Psy B07 Chapter 8Slide 2 Chapter 4 flashback  Type I error is the probability of rejecting the null hypothesis when it.
Hawawini & VialletChapter 71 ALTERNATIVES TO THE NPV RULE.
QNT 531 Advanced Problems in Statistics and Research Methods
Experiments & Statistics. Experiment Design Playtesting Experiments don’t have to be “big”--many game design experiments take only 30 minutes to design.
8-1 COMPLETE BUSINESS STATISTICS by AMIR D. ACZEL & JAYAVEL SOUNDERPANDIAN 6 th edition (SIE)
RMTD 404 Lecture 8. 2 Power Recall what you learned about statistical errors in Chapter 4: Type I Error: Finding a difference when there is no true difference.
Chap 20-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 20 Sampling: Additional Topics in Sampling Statistics for Business.
© Copyright McGraw-Hill CHAPTER 3 Data Description.
User Study Evaluation Human-Computer Interaction.
TOPIC THREE Chapter 4: Understanding Risk and Return By Diana Beal and Michelle Goyen.
1 Psych 5500/6500 t Test for Two Independent Means Fall, 2008.
Copyright ©2011 Pearson Education 7-1 Chapter 7 Sampling and Sampling Distributions Statistics for Managers using Microsoft Excel 6 th Global Edition.
A Course In Business Statistics 4th © 2006 Prentice-Hall, Inc. Chap 9-1 A Course In Business Statistics 4 th Edition Chapter 9 Estimation and Hypothesis.
Section 10.3: Large-Sample Hypothesis Tests for a Population Proportion.
Lecture 16 Section 8.1 Objectives: Testing Statistical Hypotheses − Stating hypotheses statements − Type I and II errors − Conducting a hypothesis test.
1 Nonparametric Statistical Techniques Chapter 17.
1 Psych 5500/6500 Introduction to the F Statistic (Segue to ANOVA) Fall, 2008.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 10 Comparing Two Populations or Groups 10.1.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 10 Comparing Two Populations or Groups 10.1.
Statistics for Managers Using Microsoft Excel, 5e © 2008 Pearson Prentice-Hall, Inc.Chap 7-1 Statistics for Managers Using Microsoft® Excel 5th Edition.
Chapter 10: Hypothesis Testing Using a Single Sample.
Business Statistics, A First Course (4e) © 2006 Prentice-Hall, Inc. Chap 9-1 Chapter 9 Fundamentals of Hypothesis Testing: One-Sample Tests Business Statistics,
AP Statistics Section 11.1 B More on Significance Tests.
Testing Hypotheses about a Population Proportion Lecture 29 Sections 9.1 – 9.3 Wed, Nov 1, 2006.
The t-distribution William Gosset lived from 1876 to 1937 Gosset invented the t -test to handle small samples for quality control in brewing. He wrote.
Chapter 13: Inferences about Comparing Two Populations Lecture 8b Date: 15 th November 2015 Instructor: Naveen Abedin.
Outline of Today’s Discussion 1.The Chi-Square Test of Independence 2.The Chi-Square Test of Goodness of Fit.
 Assumptions are an essential part of statistics and the process of building and testing models.  There are many different assumptions across the range.
DEVELOPED BY MARY BETH FURST ASSOCIATE PROFESSOR, BUCO DIVISION AMY CHASE MARTIN DIRECTOR OF FACULTY DEVELOPMENT AND INSTRUCTIONAL MEDIA UNDERSTANDING.
Yu-Hsuan Lin Catholic University of Korea, Korea University of York, U.K. 5 th Congress of EAAERE, Taipei, 06 th – 07 th August 2015.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Chapter 9 Introduction to the t Statistic
Behavioral economics Chapter 30
3. The X and Y samples are independent of one another.
Risk and Return.
Hypothesis Testing Is It Significant?.
T-test for 2 Independent Sample Means
Comparing Populations
marketing research with Spss
What are their purposes? What kinds?
Behavioral economics Chapter 30
Chapter 9 Test for Independent Means Between-Subjects Design
Presentation transcript:

1 Field Experiment on Incentive of Communication, Compilation and Evaluation 1. Motivation for the study 2. What is the Field Experiment 3. Framework of the experiment 4. The flow of the experiment 5. Analysis of Results Takuya Nakaizumi

2 Motivation for the study 1. Does monetary Incentive affect the conversation or editing? 1. Editors or writers have their own opinion.They will reject the opinion of others on their writing or compilation especially when they do not agree with or do not like it. 2. Even in such circumstances, incentives may make them edit in a way different from what they would have done otherwise. 2. In this experiment, we tested whether a monetary incentive affected the editing of a conversation in the BBS.

3 Evaluation of the editing 1. To give an editor adequate incentive, we need an evaluation of the performance on which the reward depends 2. The editing must have a true value. However it is quite difficult to estimate. 3. Thus in this study we employ a peer review system. The editing was evaluated by all the participants of the conversation in the BBS. And the rewards of the editor depended on the evaluation of the participants of the conversation of the BBS.

4 Evaluation Behavior 1.Participants of the conversation of the BBS, might appreciate the editing more if the editing attracts the participants. 2.The reward for the editor might affect just the effort of the editor. The reward itself for the editor might not affect the evaluation from the other participants. 3.Thus higher rewards lead to more attractive editing. → We check it by experiment

5 What is a Field Experiment? Harrison and List [2004] : Our primary point is that dissecting the characteristics of field experiments helps define what might be better called an ideal experiment, in the sense that one is able to observe a subject in a controlled setting but where the subject does not perceive any of the controls as being unnatural and there is no deception being practiced. The distinction between the laboratory and the field 1. In a field experiment, sample populations are drawn from many different domains, while sample is student of ordinal laboratory experiment. 2. The laboratory environment might not be fully representative of the field environment. c.f. winner’s curse

6 Determine the Field context of an Experiment Harrison and List [2004] : the subject (participant) pool, the information that the subjects bring to the task, the commodity, the task or trading rules applied, the stakes, the environment in which the subjects operate.

7 Taxonomy of Field Experiment Harrison and List [2004] : 1. ‘artefactual’ field experiment: ‘ non-standard’ subjects, or experimental participants from the market of interest. 2. ‘framed field experiment’: as the same as an artefactual field experiment but with field context in the commodity, task, stakes, or information set of the subjects, the commodity, 3. ‘natural field experiment’: ‘natural field experiment’ is the same as a framed field experiment but where the environment is one where the subjects naturally undertake these tasks and where the subjects do not know that they are participants in an experiment. Our field experiment

8 The Flow of the Experiment 1 1. Pre-survey of the topic and attitude of the participants 2. Discussion on BBS on specific topic Religion, Politics, Marketing, Finance (Japan only) Discussion is rewarded if they contribute more than four times. 3. Choice of the editor of the discussion by Calculating Social Influence Score 4. Editing of the editor 5. Evaluation of the editing by the participants and the editor is then rewarded according to the evaluation.

9 The Flow of the Experiment 2 1. Pre-survey 1. Opinions of the participants on which topic they prefer to discuss. 2. Variance of the opinions in each domain 2. Discussion on BBS 1. Religion, Politics (Marketing, Finance in Japan only) 2. The most and least conflicting topic based on the variance of opinions in each domain.

10 The Flow of the Experiment 3 1. Calculate (SIS) Social Influence Score l Based on mutual evaluation l Like in Google, the SIS evaluation carries more weight if the evaluator has a higher SIS. l Long tail distribution 2. Editing/Summarizing Assigned to the discussant with the highest SIS Two types of reward for the editing, $20 or $80. If others’ assessment is low, then the rewards is reduced. “How satisfied are you with the summary by the selected editor? In order to guarantee the quality of the discussion summary, if the editor is rated less than 'not satisfactory' (A 2 out of a scale of 7) on average, then the bonus for editing will be reduced by half.”

11 Experimental conditions Domains Economic: Marketing, Investment (in Japan only) Non-Economic: Politics, Religions Difficulties of topics: variance in survey High: The most conflicting topic Low: The least conflicting topic Amount of reward to a editor High:$80 (¥8,000 in Japan) Low:$20 (¥2,000 in Japan)

12 Incentive to discuss and edit Discussion on BBS If they post more than four comments, they get the participant fee. Higher rewards for a editors becomes an incentive for participants to discuss the topic. Editing of BBS High:$80 (¥8,000 in Japan) or Low:$20 (¥2,000 in Japan) Higher rewards let the editor flatter the participants more and that may raise the actual score. Evaluation of Editing There is no monetary incentive to evaluate the editing!!

13 Summary of Experiment Data In Japan reward to each editor Doma- ins Difficulty of topics Total BBS Partici- pants In each group Partici- pants at the beginning Partici- pants who continue to the end ¥8,00042 types8×3= (24 editor) (editor) ¥2,00042 types8×3= (24 editor) (editor) total42 types (48 editor) (editor)

14 Summary of Experiment Data In the U.S.(and combined total) reward to each editor Domain religion, politics Difficulty of topics Total BBS Partici- pants In each group Partici- pants at the beginning Partici-pants who continue to the end $8022 types4×3= (12 editor) (editor) $2022 types4×3= (12 editor) (editor) total22 types (24 editor) (editor) U.S., and Japan 2,42 types24 or or (72 editor) (editor)

15 Basic Model (1) Evaluation depends on the quality of editing by editor 0, that is X 0 Assumption 1: i participant’s utility is Thus the evaluation score of i by j, When editor edit the conversation, the editor does not know the evaluation of the other participants. Thus she/he face uncertainty and we describe it by ε, ε 〜

16 Basic Model (2) Quality of editing is assumed to be the function of the effort of editor: X 0 =X 0 (x)=x and reward of the editor, W(s 0 ) is And Cost Function is αc (x), (α>0, c'(x)>0, c" (x)>0, ) Thus editor’s expected benefit of editing, E[B (x)] is

17 Basic Model(3) Then (1) From →Proposition 1There is interior solution of (1)

18 Basic Model(4) That means x 0 is non decreasing function of w And non-decreasing function of α and Thus and Hypothesis : 1) Higher rewards with easy topic lead to highest score. 2)Lower rewards with difficult topic lead to lowest score. 3)The score of Higher reward with difficult topic and Lower reward with easy topic are between them

19 Results of Experiment (1): Evaluation Score (1) Difficulty Reward for editing Difference of Topic: (p-value of GWT) Easy (0.1635) Difficult * (0.0017) Total * (0.03)

20 Results of Experiment (2): length of editing (effort)

21 Results of Experiment (3): Analysis of Evaluation Score(1) Higher reward with easy topic Ordinal incentive theory lower reward with difficult topic lower reward with easy topic higher reward with difficult topic Experimental results lower with difficult lower with easy higher with easy Higher with difficult

22 Results of Experiment (4): Analysis of Evaluation Score(2) Higher reward with easy topic incentive theory with spite lower reward with difficult topic lower reward with easy topic higher reward with difficult topic Experimental results lower with difficult lower with easy higher with easy Higher with difficult

23 Results of Experiment (5): Analysis of Evaluation Score(3) Higher reward with easy topic incentive theory with fairness evaluation lower reward with difficult topic lower reward with easy topic higher reward with difficult topic Experimental results lower with difficult lower with easy higher with easy Higher with difficult

24 Results of Experiment (6): Evaluation Score (2) Low reward ($20,or ¥2,000) High reward ($80,or ¥8,000) Difference (p-value of GWT) Japan (average, variance sample) * The U.S Total * (0.03)

25 Results of Experiment (7): Evaluation Score of The U.S. Difficulty Reward for editing Difference Easy Difficult Total

26 Results of Experiment (8): Evaluation Score of Japan Difficulty Reward for editing Difference Easy Difficult * Total *

27 Results of Experiment (9): Evaluation Score (3)

28 Results of Experiment (10): Evaluation Score of The U.S. Domain Reward for editing Difference Religion Politics * Total

29 Analysis of the results Hypothesis is rejected. Then How the participants evaluate the editing? →Spite or altruistic ? Valuation of the editing by outsiders or Can the market value of the editing be estimated? Evaluation of Evaluators

30 Possibility of the behavioral evaluation Suppose the evaluation function depends on not only the quality but also the rewards the editor gets. The evaluation score is based on both the effort function of the editor and the evaluation function of the other participants as an evaluator.

31 Future exercise How the participants evaluate the editing? Valuation of the editing by outsiders or Can the market value of the editing be estimated? Evaluation of Evaluators Spite or altruistic ?