Neza Vodopivec Applied Math and Scientific Computation Program Advisor: Dr. Jeffrey Herrmann Department of Mechanical Engineering.

Slides:



Advertisements
Similar presentations
Estimation of Means and Proportions
Advertisements

Previous Lecture: Distributions. Introduction to Biostatistics and Bioinformatics Estimation I This Lecture By Judy Zhong Assistant Professor Division.
STATISTICAL INFERENCE PART V CONFIDENCE INTERVALS 1.
Uncertainty and confidence intervals Statistical estimation methods, Finse Friday , 12.45–14.05 Andreas Lindén.
Ch11 Curve Fitting Dr. Deshi Ye
Chapter 8 Estimating Single Population Parameters
Lecture 3 Miscellaneous details about hypothesis testing Type II error
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Lecture 4 Chapter 11 wrap-up
Evaluating Hypotheses
4-1 Statistical Inference The field of statistical inference consists of those methods used to make decisions or draw conclusions about a population.
Sampling Distributions
Inference about a Mean Part II
Estimation 8.
Experimental Evaluation
Chapter 11: Inference for Distributions
Inferences About Process Quality
6.1 Confidence Intervals for the Mean (Large Samples)
Inferential Statistics
1 Simple Linear Regression 1. review of least squares procedure 2. inference for least squares lines.
Chapter 15 Nonparametric Statistics
Estimation Goal: Use sample data to make predictions regarding unknown population parameters Point Estimate - Single value that is best guess of true parameter.
Analysis of Simulation Results Andy Wang CIS Computer Systems Performance Analysis.
Neza Vodopivec Applied Math and Scientific Computation Program Advisor: Dr. Jeffrey Herrmann Department of Mechanical Engineering.
Review of Statistical Inference Prepared by Vera Tabakova, East Carolina University ECON 4550 Econometrics Memorial University of Newfoundland.
Copyright © Cengage Learning. All rights reserved. 12 Simple Linear Regression and Correlation.
Simulation Output Analysis
1 Statistical Analysis - Graphical Techniques Dr. Jerrell T. Stracener, SAE Fellow Leadership in Engineering EMIS 7370/5370 STAT 5340 : PROBABILITY AND.
1-1 Copyright © 2015, 2010, 2007 Pearson Education, Inc. Chapter 23, Slide 1 Chapter 23 Comparing Means.
Chapter 6 Confidence Intervals 1 Larson/Farber 4th ed.
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
1 CSI5388: Functional Elements of Statistics for Machine Learning Part I.
Lecture 14 Sections 7.1 – 7.2 Objectives:
Confidence Intervals 1 Chapter 6. Chapter Outline Confidence Intervals for the Mean (Large Samples) 6.2 Confidence Intervals for the Mean (Small.
Population All members of a set which have a given characteristic. Population Data Data associated with a certain population. Population Parameter A measure.
Statistical Review We will be working with two types of probability distributions: Discrete distributions –If the random variable of interest can take.
Confidence Intervals for the Mean (Large Samples) Larson/Farber 4th ed 1 Section 6.1.
Section 8.1 Estimating  When  is Known In this section, we develop techniques for estimating the population mean μ using sample data. We assume that.
Chapter 6 Lecture 3 Sections: 6.4 – 6.5.
The Examination of Residuals. Examination of Residuals The fitting of models to data is done using an iterative approach. The first step is to fit a simple.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
Unit 6 Confidence Intervals If you arrive late (or leave early) please do not announce it to everyone as we get side tracked, instead send me an .
Limits to Statistical Theory Bootstrap analysis ESM April 2006.
Confidence Interval Estimation For statistical inference in decision making:
A New Approach for Comparing Means of Two Populations By Brad Moss And Sponsored by Dr. Chand Chauhan.
Statistics 300: Elementary Statistics Sections 7-2, 7-3, 7-4, 7-5.
Sampling and estimation Petter Mostad
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 25 Paired Samples and Blocks.
Chapter 6 Lecture 3 Sections: 6.4 – 6.5. Sampling Distributions and Estimators What we want to do is find out the sampling distribution of a statistic.
8.1 Estimating µ with large samples Large sample: n > 30 Error of estimate – the magnitude of the difference between the point estimate and the true parameter.
Chapter 8 Estimation ©. Estimator and Estimate estimator estimate An estimator of a population parameter is a random variable that depends on the sample.
1 Statistical Analysis - Graphical Techniques Dr. Jerrell T. Stracener, SAE Fellow Leadership in Engineering EMIS 7370/5370 STAT 5340 : PROBABILITY AND.
Statistics for Business and Economics 8 th Edition Chapter 7 Estimation: Single Population Copyright © 2013 Pearson Education, Inc. Publishing as Prentice.
Chapter Confidence Intervals 1 of 31 6  2012 Pearson Education, Inc. All rights reserved.
STATISTICS People sometimes use statistics to describe the results of an experiment or an investigation. This process is referred to as data analysis or.
Chapter 6 Confidence Intervals 1 Larson/Farber 4th ed.
1 Simple Linear Regression Chapter Introduction In Chapters 17 to 19 we examine the relationship between interval variables via a mathematical.
The normal distribution
Chapter 9 Roadmap Where are we going?.
Estimating the Value of a Parameter Using Confidence Intervals
Hypotheses and test procedures
Sampling Distributions and Estimation
Inference about Comparing Two Populations
Copyright © Cengage Learning. All rights reserved.
CONCEPTS OF ESTIMATION
Discrete Event Simulation - 4
The normal distribution
Ch13 Empirical Methods.
Estimation Goal: Use sample data to make predictions regarding unknown population parameters Point Estimate - Single value that is best guess of true parameter.
Presentation transcript:

Neza Vodopivec Applied Math and Scientific Computation Program Advisor: Dr. Jeffrey Herrmann Department of Mechanical Engineering

The Bass model (Bass, 1969), which was originally developed to model the diffusion of new products in marketing, can be applied to the diffusion of information. The model is based on the assumption that people get their information from two sources: word of mouthadvertising

The Bass model describes the change in the fraction of a population that has become aware of a piece of information: where F(t) is the aware fraction of the population, p is the advertising coefficient, and q is the word-of-mouth coefficient.

We can formulate an agent-based model inspired by the classical Bass model. We discretize the problem and make the following modifications: 1. Instead of taking a deterministic time aggregate, we update probabilistically. 2. Instead of allowing each agent to be influenced by the entire population, it is influenced only by its neighbors.

 The agent-based Bass model is a discrete-time model in which each agent has one of two states at each time step t: (1) unaware or (2) aware.  At time t=0, all agents are unaware.  At each time step, an unaware agent has an opportunity to become aware. Its state changes with p, the probability that it becomes aware due to advertising or due to word of mouth.  The probability of that an agent becomes aware due to word of mouth increases as a function of the fraction of its neighbors who became aware in previous time steps.  Once an agent becomes aware, it remains aware for the rest of the simulation.

In the summer of this year, Dr. Herrmann directed Ardechir Auzolle in implementing this model and explaining its properties. The codebase, written in NetLogo, turned out not to be fast enough to handle large networks. I have coded this model in MATLAB with the hope of producing a faster, more memory efficient implementation. First, I coded a basic implementation as a reference. Then, I implemented the model using a more efficient updating rule and taking advantage of sparse data structures.

Arbitrarily identify the N agents with the set 1, …, N. Let D denote the |D|×2 matrix listing all (directed) edges of the graph as ordered pairs of nodes. INPUT: matrix D, parameters p and q. 1. Keep track of the state of the agents in a length-N bit vector initialized to all zeros. 2. Create an adjacency matrix A such that A(i,j) = 1 if (i,j) is a directed edge in D and 0 otherwise. 3. Create another adjacency matrix B to track aware neighbors. B(i,j) = 1 if (i,j) is a directed edge in D and agent i is aware. 4. At each time step, for each agent: a. Check the bit vector to determine if the agent is already aware. If so, skip it. b. Make the agent newly aware with probability p. c. Look up the agent’s upstream neighbors in A. Look up the agent’s aware upstream neighbors in B. d. Determine what fraction of the upstream neighbors are aware. Make the agent newly aware with probability q times that fraction. e. Once all agents have been processed, record the newly aware ones as aware in the bit vector. Stop once all agents have become aware, or after a maximum number of iterations. OUTPUT: complete history of the bit vector.

 In order to decide whether to change the status of an unaware node, the node’s number of unaware upstream nodes (its “awareness number”) must be computed. The basic implementation effectively recomputes each node’s awareness number from scratch at every time step.  But changes in the awareness number are entirely due to nodes which have just become aware. Basic Neighbor-Set  A possible improvement: a preliminary pass through just the newly-aware nodes which updates just their downstream nodes. After this preliminary step, we can proceed as in the basic implementation, but without needing to recompute awareness numbers.

 Our new updating procedure suggests a further possible improvement: replacing the network’s adjacency matrix with a sparse data structure which reflects the structure of the updating rule.  Information about adjacency can be stored by rewriting the adjacency relation as a function f: V ⟶ 2 V which returns a node’s downstream nodes.  Concretely, this function is most naturally implemented as a vector of length |D| concatenating the output sets of together with a list of pointers marking the start of each set.

Bin-Laden DataMemorySimulation TimeTotal Time NetLogo--~ 3 min. -- Basic MB12.92 sec.21.0 min. Neighbor Set18.55 MB0.74 sec.1.2 min. Irene DataMemorySimulation TimeTotal Time NetLogo--~ 10 sec-- Basic30.31 MB0.73 sec.1.2 min. Neighbor Set1.85 MB0.06 sec.6.5 sec.

 At each time step, our algorithm produces not a number but rather a random variable. We would like to understand its distribution.  We would like to answer two questions: 1. Where is the distribution centered? We can answer this by computing the sample mean, then surrounding it with a confidence interval within which we expect the true mean to lie. 2. Where is the bulk of the distribution’s mass? We can answer this by constructing prediction intervals within which we think our next sample is likely to fall.

 We can create a new distribution from an old one by sampling from the original distribution in batches of size n and considering the distribution of the batch means.  The distribution of sample means will have a mean which equals the original distribution’s mean and a variance which is 1/n times the variance of the original distribution.  The Central Limit Theorem states that the distribution of the sample means, when scaled to have unit variance, will converge to a normal distribution as n⟶.  In other words, the distribution of a mean tends to be normal, even when the distribution from which the mean is computed is decidedly non-normal.

Because the CLT says that the sample mean is approximately normally distributed, for an unknown mean μ. (1) will be approximately standard normal. So we have (2) P (-1.96 ≤ Z ≤ 1.96) = Substituting (1) into (2) and solving for μ we have: The rearranged bounds are known as a confidence interval for the mean.

 We can use our past observations to predict the outcome if we were to run the simulation one more time.  Let x 1,…, x n+1 be a random sample from our distribution of simulation results. A 95% prediction interval for x n+1 is an interval with random endpoints and such that

 If we are sampling from an unknown distribution, we might try using quantiles to form our prediction intervals.  Let be the cumulative distribution function for a random variable. The quantile function is the inverse distribution function,.  Simply put, the quantile function gives the value of a random variable given the probability of obtaining at most that value.

 Parametric prediction intervals require that the distribution from which we are sampling is normal.  A 95% parametric prediction interval for x n+1 is where is the sample mean, S is the sample standard deviation, and t 2.5 the 2.5 th percentile of the t distribution with n-1 degrees of freedom. Let’s assume for a moment that the simulation results are normally distributed at each time step. This assumption allows us to analyze them using parametric prediction intervals.

The nonparametric and the parametric prediction intervals nearly overlap – in this case, the normality assumption didn’t produce significantly different results.

 At each time step, our simulation produces a random variable. In order to compute parametric prediction intervals, we assumed the random variable came from a normal distribution.  So how ‘close’ is this random variable to being normally distributed?  We can examine the distribution graphically with quantile- quantile plots. We can also perform tests for normality, such as the Kolmogorov-Smirnov goodness-of-fit test.

 A q-q plot is often used to determine if the underlying distribution of a data sample matches a particular theoretical distribution.  Let Q 1 (p) and Q 2 (p) be quantile functions for two distributions respectively. A q-q plot is a parametric curve with coordinates (Q 1 (p), Q 2 (p)) and quantile intervals as the parameter.  If the points lie roughly on the line x = y, then the distributions are the same.

Validation

 Let’s focus on the case of a fully-connected network.  Let x(t) denote the fraction of agents who are aware at time t. If the probability that advertising makes an agent aware is p and the independent probability that word of mouth makes the agent aware is qx, then, by the Inclusive-Exclusive Principle, the probability that an agent becomes aware is:  Estimating the fraction of agents who become aware over a single time step as the probability a(t) of becoming aware times the fraction (1- x(t)) of unaware agents, we arrive at the recurrence relation: with the initial condition..

The solution to the recurrence indeed tracks the output of the simulation, suggesting that the implementation is correct.

 Our simulation assumes an advertising probability of p and a word-of-mouth probability of qx over a time step of 1 hour.  If we wish to rework the simulation to run over some smaller time step ∆t, we must reimagine p and q as probabilities per hour and instead write out advertising probability as p ∆t and word-of-mouth probability as q ∆t x.  Inserting ∆t this way into the previous recurrence, we obtain: again starting with.

The solutions to the recurrence continue to track our simulation.

It turns out that the recurrence schemes are (global) first-order approximations to the solution of the original logistic ODE (the Bass model).

 OctoberDevelop basic simulation code. Develop code for statistical analysis of results.  November Validate simulation code by checking corner cases, sampled cases, and by relative testing. Validate code against analytic model.  DecemberValidate simulation against existing NetLogo implementation. Prepare mid-year presentation and report.  JanuaryInvestigate efficiency improvements to code. Incorporate sparse data structures.

Bass, Frank (1969). “A new product growth model for consumer durables”. Management Science 15 (5): p. 215–227. Chandrasekaran, Deepa and Tellis, Gerard J. (2007). “A Critical Review of Marketing Research on Diffusion of New Products”. Review of Marketing Research, p ; Marshall School of Business Working Paper No. MKT Dodds, P.S. and Watts, D.J. (2004). “Universal behavior in a generalized model of contagion”. Phys. Rev. Lett. 92, Mahajan, Vijay; Muller, Eitan and Bass, Frank (1995). “Diffusion of new products: Empirical generalizations and managerial uses”. Marketing Science 14 (3): G79–G88. Rand, William M. and Rust, Roland T. (2011). “Agent-Based Modeling in Marketing: Guidelines for Rigor (June 10, 2011)”. International Journal of Research in Marketing; Robert H. Smith School Research Paper No. RHS