Is Ignoring Public Information Best Policy? Reinforcement Learning In Information Cascade Toshiji Kawagoe Future University - Hakodate and Shinichi Sasaki.

Slides:



Advertisements
Similar presentations
Motivating Markov Chain Monte Carlo for Multiple Target Tracking
Advertisements

Eliciting Honest Feedback 1.Eliciting Honest Feedback: The Peer-Prediction Model (Miller, Resnick, Zeckhauser) 2.Minimum Payments that Reward Honest Feedback.
Classification. Introduction A discriminant is a function that separates the examples of different classes. For example – IF (income > Q1 and saving >Q2)
Markov Decision Process
This paper reports an experimental study based on the popular Chinos game, in which three players, arranged in sequence, have to guess the total number.
The Structure of Networks with emphasis on information and social networks T-214-SINE Summer 2011 Chapter 16 Ýmir Vigfússon.
DECISION MARKETS WITH GOOD INCENTIVES Yiling Chen (Harvard), Ian Kash (Harvard), Internet and Network Economics, 2011.
Markov Decision Processes
Minority Games A Complex Systems Project. Going to a concert… But which night to pick? Friday or Saturday? You want to go on the night with the least.
Introduction  Bayesian methods are becoming very important in the cognitive sciences  Bayesian statistics is a framework for doing inference, in a principled.
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
12 June, STD( ): learning state temporal differences with TD( ) Lex Weaver Department of Computer Science Australian National University Jonathan.
Reinforcement Learning Rafy Michaeli Assaf Naor Supervisor: Yaakov Engel Visit project’s home page at: FOR.
Chapter 2: Bayesian Decision Theory (Part 1) Introduction Bayesian Decision Theory–Continuous Features All materials used in this course were taken from.
1 Information Markets & Decision Makers Yiling Chen Anthony Kwasnica Tracy Mullen Penn State University This research was supported by the Defense Advanced.
Social Learning. A Guessing Game Why are Wolfgang Puck restaurants so crowded? Why do employers turn down promising job candidates on the basis of rejections.
1 Hybrid Agent-Based Modeling: Architectures,Analyses and Applications (Stage One) Li, Hailin.
True/False. False True Subject May Go Here True / False ? Type correct answer here. Type incorrect answer here.
Bayesian Models Honors 207, Intro to Cognitive Science David Allbritton An introduction to Bayes' Theorem and Bayesian models of human cognition.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
AP STATISTICS.   Theoretical: true mathematical probability  Empirical: the relative frequency with which an event occurs in a given experiment  Subjective:
Chapter 4 Probability Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
Bottom-Up Coordination in the El Farol Game: an agent-based model Shu-Heng Chen, Umberto Gostoli.
A multiple-choice test consists of 8 questions
CSCI 121 Special Topics: Bayesian Network Lecture #1: Reasoning Under Uncertainty.
Utility Theory & MDPs Tamara Berg CS Artificial Intelligence Many slides throughout the course adapted from Svetlana Lazebnik, Dan Klein, Stuart.
Bayes for Beginners Presenters: Shuman ji & Nick Todd.
Theoretical and Experimental Probability Today you will learn to: calculate the theoretical and experimental probabilities of an event. M07.D-S.3.1.1:
Bayesian vs. frequentist inference frequentist: 1) Deductive hypothesis testing of Popper--ruling out alternative explanations Falsification: can prove.
Methodological Problems in Cognitive Psychology David Danks Institute for Human & Machine Cognition January 10, 2003.
1 Demand for Repeated Insurance Contracts with Unknown Loss Probability Emilio Venezian Venezian Associates Chwen-Chi Liu Feng Chia University Chu-Shiu.
1 Lesson 4 Attitudes. 2 Lesson Outline   Last class, the self and its presentation  What are attitudes?  Where do attitudes come from  How are they.
DCM – the theory. Bayseian inference DCM examples Choosing the best model Group analysis.
1 Chapter 12 Probabilistic Reasoning and Bayesian Belief Networks.
PROBABILITY IN OUR DAILY LIVES
Copyright © 2014 by McGraw-Hill Higher Education. All rights reserved. Essentials of Business Statistics: Communicating with Numbers By Sanjiv Jaggia and.
On rare events and the economics of small decisions Ido Erev, Technion Examples: Using safety devices, cheating in exams, selecting among websites, stopping.
SCIENTIFIC METHOD SOLVING A PROBLEM – WITH SCIENCE!! MR. ROCKENSIES.
McGraw-Hill/Irwin Copyright © 2008 by The McGraw-Hill Companies, Inc. All rights reserved. CHAPTER 2 Tools of Positive Analysis.
HAYEK’S POLITICAL ECONOMY: ITS RELEVANCE FOR DEVELOPING ECONOMIES Shankaran Nambiar, Senior Research Fellow and Head, Policy Studies Division Malaysian.
1 1 Slide Simulation Professor Ahmadi. 2 2 Slide Simulation Chapter Outline n Computer Simulation n Simulation Modeling n Random Variables and Pseudo-Random.
Coordination and Learning in Dynamic Global Games: Experimental Evidence Olga Shurchkov MIT The Economic Science Association World Meeting 2007.
Transfer Learning in Sequential Decision Problems: A Hierarchical Bayesian Approach Aaron Wilson, Alan Fern, Prasad Tadepalli School of EECS Oregon State.
Paging Area Optimization Based on Interval Estimation in Wireless Personal Communication Networks By Z. Lei, C. U. Saraydar and N. B. Mandayam.
© 2015 McGraw-Hill Education. All rights reserved. Chapter 16 Decision Analysis.
Chapter 1 Section 2 Review
Discussions on “ Demand for Repeated Insurance Contracts with Unknown Loss Probability ” Jason Yeh Chinese University of Hong Kong 2007 ARIA - Quebec City.
Simulate a Problem and Make an Organized List COURSE 3 LESSON 11-6 On a multiple-choice test, each question has 4 possible answers. You know the answers.
Reinforcement Learning for Mapping Instructions to Actions S.R.K. Branavan, Harr Chen, Luke S. Zettlemoyer, Regina Barzilay Computer Science and Artificial.
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Chapter 4 Probability.
Cascade Principles, Bayes Rule and Wisdom of the Crowds Lecture 6 (Largely drawn from Kleinberg book)
11.2 Probability SLG: To find the probability of an event using theoretical, experimental, and simulation methods.
Yu-Hsuan Lin Catholic University of Korea, Korea University of York, U.K. 5 th Congress of EAAERE, Taipei, 06 th – 07 th August 2015.
KATHOLIEKE UNIVERSITEIT LEUVEN Learning about Compliance under Asymmetric Information Carmen Arguedas Universidad Autónoma de Madrid and Sandra Rousseau.
Bayes’ Theorem Suppose we have estimated prior probabilities for events we are concerned with, and then obtain new information. We would like to a sound.
11-2 Basic Probability.
Lecture 1.31 Criteria for optimal reception of radio signals.
Markov Decision Processes
Incorporating New Information to Decision Trees (posterior probabilities) MGS Chapter 6 Part 3.
Bubbles , Info Cascades & Behavioral Economics.
Chapter 8: Hypothesis Testing and Inferential Statistics
Reinforcement Learning
Course: Autonomous Machine Learning
Experimental Inquiry Template.
Toshiji Kawagoe Future University – Hakodate and Hirokazu Takizawa
11-2 Basic Probability.
LECTURE 23: INFORMATION THEORY REVIEW
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Complete the family of four
Presentation transcript:

Is Ignoring Public Information Best Policy? Reinforcement Learning In Information Cascade Toshiji Kawagoe Future University - Hakodate and Shinichi Sasaki Future University - Hakodate Artificial Economics 2006, Aalborg, Denmark, Sep , 2006

Motivation Dynamic extension of information cascade model –develop a new model featured by a reinforcement learning to extend the model of information cascade. Examine the role of public information in correct guessing the true state of the world –conduct a series of simulations in order check which factor, public information or agent’s learning, is crucial to induce information cascade in a dynamic environment.

Model 1 Information cascade –when it is optimal for each agent to ignore its private information and follow the decisions made by the other agents who have decided before. The theory of information cascade was initiated by several researchers in economics and it has been tested in the laboratory with human subjects by a number of experimental economists

Model 2 Each agent –makes a decision sequentially one by one. –tries to predict the true state of the world. –knows prior probability distribution of the state of the world. –receives a noisy signal of the state of the world privately. –also knows likelihood of receiving that noisy signal conditional on the state. –observes publicly announced past decisions made by the other agents (public information).

Theoretical Prediction Each agent can calculate posterior probability of the realized state of world by utilizing its private information, likelihood and public information to predict true state of the world. But if a small number of agents accidentally made same decisions, then it is easily shown by Bayes rule that an agent who observes such public information should ignore its private information and follows its predecessors’ decisions.

Good and Bad Cascade Good cascade –If the prediction made by agents who are involved in an information cascade is true state of the world. Bad cascade –The prediction made by agents who are involved in an information cascade is not correct state of the world.

Symmetric condition Prior probability of the state A and B –p(A) = p(B) = ½ Likelihood of receiving a signal a (b) in the state A (B) –p(a|A) = p(b|B) = 2/3 At least two agents make same prediction, an agent who observes such public information follows the predecessors’ prediction.

Asymmetric condition Prior probability of the state A and B –p(A) = p(B) = ½ Likelihood of receiving a signal a (b) in the state A (B) –p(a|A) = 6/7 p(b|B) = 2/7 If at least four agents predict A, then predicting A is best response for an agent who observes such public information, and if at least an agent predict B, then predicting B is best response for subsequent agents whatever signal they receive.

Three types of agent Type B –Agents who follows Bayes rule. Type RNP –Agents who follow reinforcement learning without public information Type RP –Agents who follow reinforcement learning with public information

Frequency of information cascade in symmetric case A-cascadeB-cascade LengthTypeBRNPB 3Good Bad Good Bad Good Bad

Frequency of information cascade in asymmetric case A-cascadeB-cascade LengthTypeBRNPB 3Good Bad Good Bad Good Bad

RP model Updating rule if agent i predicts A and it is correct (incorrect) at period t. otherwise Probability of predicting A

Conclusion 1 Previous studies reported that agents’ behaviors were still affected by their private signals even after a longer cascade has already occurred. We tried to answer that question by developing a new model featured by a reinforcement learning to extend the model of information cascade.

Conclusion 2 Our results showed that –incorporating agent’s past experience of prediction had great impact on the emergence of information cascade –providing public information was rather detrimental for agents who would like to predict correct state of the world –So, ignoring public information, and believing one’s own private information and experiences is much better for making correct prediction.