+ Chapter 16 – Markov Chain Random Walks McCTRWs – the fast food of random walks.

Slides:



Advertisements
Similar presentations
Pretty-Good Tomography Scott Aaronson MIT. Theres a problem… To do tomography on an entangled state of n qubits, we need exp(n) measurements Does this.
Advertisements

Hypothesis testing 5th - 9th December 2011, Rome.
Probabilistic Reasoning over Time
Lecture 19: Parallel Algorithms
Review of Identifying Causal Effects Methods of Economic Investigation Lecture 13.
Simple Linear Regression 1. 2 I want to start this section with a story. Imagine we take everyone in the class and line them up from shortest to tallest.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 18 Sampling Distribution Models.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 18 Sampling Distribution Models.
Measuring MOTION.
Statistics review of basic probability and statistics.
Fate & Transport of Contaminants in Environmental Flows 2015.
Chapter 13 MIMs - Mobile Immobile Models. Consider the Following Case You have two connected domains that can exchange mass
The Rate of Concentration of the stationary distribution of a Markov Chain on the Homogenous Populations. Boris Mitavskiy and Jonathan Rowe School of Computer.
25 May Quick Sort (11.2) CSE 2011 Winter 2011.
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Computational statistics 2009 Random walk. Computational statistics 2009 Random walk with absorbing barrier.
Copyright (c) Bani K. Mallick1 STAT 651 Lecture #18.
1 Lecture 25: Parallel Algorithms II Topics: matrix, graph, and sort algorithms Tuesday presentations:  Each group: 10 minutes  Describe the problem,
GG313 Lecture 8 9/15/05 Parametric Tests. Cruise Meeting 1:30 PM tomorrow, POST 703 Surf’s Up “Peak Oil and the Future of Civilization” 12:30 PM tomorrow.
Quicksort
Chapter 4 Numerical Solutions to the Diffusion Equation.
Linear Regression 2 Sociology 5811 Lecture 21 Copyright © 2005 by Evan Schofer Do not copy or distribute without permission.
Physics of fusion power
Relationships Among Variables
David Luebke 1 8/17/2015 CS 332: Algorithms Linear-Time Sorting Continued Medians and Order Statistics.
Psy B07 Chapter 1Slide 1 ANALYSIS OF VARIANCE. Psy B07 Chapter 1Slide 2 t-test refresher  In chapter 7 we talked about analyses that could be conducted.
Reaction-Diffusion Systems - Continued Reactive Random Walks.
Linear Regression Inference
Simulation II IE 2030 Lecture 18. Outline: Simulation II Advanced simulation demo Review of concepts from Simulation I How to perform a simulation –concepts:
CORRELATION & REGRESSION
Brian Renzenbrink Jeff Robble Object Tracking Using the Extended Kalman Particle Filter.
RMTD 404 Lecture 8. 2 Power Recall what you learned about statistical errors in Chapter 4: Type I Error: Finding a difference when there is no true difference.
ENSEMBLE LEARNING David Kauchak CS451 – Fall 2013.
1 LES of Turbulent Flows: Lecture 1 Supplement (ME EN ) Prof. Rob Stoll Department of Mechanical Engineering University of Utah Fall 2014.
Finding dense components in weighted graphs Paul Horn
Liang Ge.  Introduction  Important Concepts in MCL Algorithm  MCL Algorithm  The Features of MCL Algorithm  Summary.
CS433: Modeling and Simulation Dr. Anis Koubâa Al-Imam Mohammad bin Saud University 15 October 2010 Lecture 05: Statistical Analysis Tools.
Nonparametric Statistics aka, distribution-free statistics makes no assumption about the underlying distribution, other than that it is continuous the.
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
Copyright © 2009 Pearson Education, Inc. Chapter 18 Sampling Distribution Models.
Lesson 10 The Fields of Accelerating Charges. Class 29 Today we will: learn about threads and stubs of accelerating point charges. learn that accelerating.
Hypothesis testing Intermediate Food Security Analysis Training Rome, July 2010.
A Comparison of Two MCMC Algorithms for Hierarchical Mixture Models Russell Almond Florida State University College of Education Educational Psychology.
 Let’s take everything we have learned so far and now add in the two other processes discussed in the introduction chapter – advection and retardation.
Optimising Cuts for HLT George Talbot Supervisor: Stewart Martin-Haugh.
Predicting Output from Computer Experiments Design and Analysis of Computer Experiments Chapter 3 Kevin Leyton-Brown.
Application 3: Estimating the Effect of Education on Earnings Methods of Economic Investigation Lecture 9 1.
Statistical-Time Access Fairness Index of One-Bit Feedback Fair Scheduler Fumio Ishizaki Dept. of Systems Design and Engineering Nanzan University, Japan.
Data Analysis Econ 176, Fall Populations When we run an experiment, we are always measuring an outcome, x. We say that an outcome belongs to some.
Chapter 15 – CTRW Continuous Time Random Walks. Random Walks So far we have been looking at random walks with the following Langevin equation  is a.
Chapter 4 The Classical Model Copyright © 2011 Pearson Addison-Wesley. All rights reserved. Slides by Niels-Hugo Blunch Washington and Lee University.
1 OUTPUT ANALYSIS FOR SIMULATIONS. 2 Introduction Analysis of One System Terminating vs. Steady-State Simulations Analysis of Terminating Simulations.
1 Psych 5510/6510 Chapter 13 ANCOVA: Models with Continuous and Categorical Predictors Part 2: Controlling for Confounding Variables Spring, 2009.
UNIT 5.  The related activities of sorting, searching and merging are central to many computer applications.  Sorting and merging provide us with a.
Reaction-Diffusion Systems Reactive Random Walks.
Week 6. Statistics etc. GRS LX 865 Topics in Linguistics.
Example x y We wish to check for a non zero correlation.
Detector Effects and Backgrounds In the following, I examine the following possible sources of spurious signals which could simulate the parity signal:
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Chapter 18 Sampling Distribution Models.
Statistics for Business and Economics Module 1:Probability Theory and Statistical Inference Spring 2010 Lecture 4: Estimating parameters with confidence.
This was written with the assumption that workbooks would be added. Even if these are not introduced until later, the same basic ideas apply Hopefully.
Your friend has a hobby of generating random bit strings, and finding patterns in them. One day she come to you, excited and says: I found the strangest.
Generating Random Variates
Ch 4. Language Acquisition: Memoryless Learning 4.1 ~ 4.3 The Computational Nature of Language Learning and Evolution Partha Niyogi 2004 Summarized by.
Random signals Honza Černocký, ÚPGM.
IL Marking Get out your CPU / Memory answers Swap with someone else
Linear Algebra Review.
Simulation-Based Approach for Comparing Two Means
How to handle missing data values
Regression Forecasting and Model Building
Presentation transcript:

+ Chapter 16 – Markov Chain Random Walks McCTRWs – the fast food of random walks

+ Overview We are in the home stretch now and this last topic is one that is very dear to my heart as it is something that I worked very actively on myself. One of my close friends who was a postdoc in Spain at the same time as me actually ‘invented’ this method and he and I have continued to work on it a lot (including that this is a core part of Nikki’s PhD thesis). We have run this model in many many different problems and so far it has proven itself robust to almost any challenge we have thrown its way…. What we still do not know how to do is how to parameterize it for real data from real systems.. But we’re getting there.

+ Basic Premise of Most Random walk Models Our particle position and time equations are given by  and  are random numbers. We have considered them with finite mean and variance, infinite variance and infinite mean and shown all sorts of behaviors, but we have still always assumed one thing I.I.D (Independent and Identically Distributed)

+ IID Identically Distributed – this means that the random number is always drawn from the same distribution. How might you relax this? Do you think it’s important? In what context? Independent What do you mean by independent? Does this make sense to you How might we relax this assumption and when is it important?

+ Consider the following classical problem Flow between two flat plates We already know from our work on Taylor dispersion how this system will behave at late times (t>L^2/D) But what about at earlier times than this? Can we use some of the CTRW ideas we have discussed to date? y x

+ How about this? Define some distance  x I can measure how much time it takes particles to move this distance, right? Therefore I can say  x is fixed,  t=0 and  is random y x xx And now I just run my CTRW, right? And because p(t) has finite mean and variance it will converge to Fickian Dispersion (which I expect from Taylor’s theory) – Foresee any issues?

+ Let’s test it Consider the following case U=3/2 (1-y^2) -1<y<1 (parabolic flow between two plates) Let’s consider  x=10 (you can pick whatever you want – within limits as Nikki has discovered, but lets not get sidetracked by that) Test our idea for D=0.1,0.01 and Basically, I run a particle tracking simulation with a pulse initial condition at x=0 (represented by N particles) and then measure how long it takes each particle to travel a distance of  x=10. This then gives me a vector of N travel times that I can sample from to make the next jump over  x=10 (i.e. to 20) Let’s test this model’s ability to predict breakthrough at x=20, 40 and 100

+ Step 1 – Measure travel time distribution Run the code rw_ptau.m Note here that we use what is called a flux weighted initial condition (just for ease) This code runs a random walk that resolves the velocity field, i.e. In this code we store the vector, called timer, that measures the amount of time it takes each particle to cross a distance  x=10. Save timer

+ Step 2 – Generate data for comparison Run the same random walk code that models everything too Measure BTCs at further distances – rw_btcs.m Measure the amount of time it takes particles to travel a distance of 20, 40 and 100, which allows us to build brreakthrough curves btc20, btc40 and btc100.

+ Step 3 - Run our “Upscaled” Model We now run a code that implements a CTRW and no longer resolves the velocity field, but aims to capture the different speeds of each particle by saying it takes a random time to cross a distance  x. That random time is chosen from the vector timer. Run this code to generate arrival times/BTCs for x=20,40, 100 Sampled randomly from timer

+ Step 4 – Compare the model outputs Run the code compare.m The code plot the BTCs at x=20,40 and 100 for both the fully resolved and upscaled CTRW model.

+ For D=0.1 Step 1 – Generate Timer

+ For D=0.1 Step 2– Generate BTCs

+ For D=0.1 Step 3 – Run CTRW

+ For D=0.1 Step 4 – Compares Virtually Indistinguishable

+ For D=0.01 Step 1 – Generate Timer

+ For D=0.01 Step 2– Generate BTCs

+ For D=0.01 Step 3 – Run CTRW

+ For D=0.01 Step 4 – Compares Not terrible But certainly Not Indistinguishable

+ For D=0.001 Step 1 – Generate Timer

+ For D=0.001 Step 2– Generate BTCs

+ For D=0.001 Step 3 – Run CTRW

+ For D=0.001 Step 4 – Compares Seems to be getting even worse

+ So what is going on? For large D the proposed CTRW model looks pretty good… As D gets smaller the match between model and measured gets worse and worse – WHY? What is different from the perspective of the CTRW model as D get smaller and smaller? How might you address this? And thereby improve our model so that we can match predicted and observed consistently across a range of values of D?

+ What is missing? The CTRW as we have defined it assumes that successive jumps in time are completely uncorrelated and independent of one another. Our simulations suggest that this may only be true for large diffusion coefficients – does this align with your intuition? As D gets smaller particles are likely to stay on similar streamlines and so it is likely that successive  s will not be independent. How can mea measure and thus impose such correlation

+ Travel Time Distribution x11x11 x22x22 If uncorrelated I just sample arbitrarily from p(  )

+ Travel Time Distribution x11x11 x22x22 Imagine I pick a specific value of  1 as shown. Now because of correlation  2 is restricted where it can come from

+ Let’s break p(  ) into bins – for sake of depiction say 4 x11x11 x22x2 What correlation is saying is that  2 can only come from bin 2 or 3 now and we can measure the probability that it comes from bin 2 or bin 3.

+ What we have to measure Run a simulation with N particles (say N=one million) and for each particle measure how long it takes to cross  x 1 and how long it takes to cross  x 2 The distribution of the two should be the same Now break p(t) into M bins and attribute each particles  1 to one bin and each particles  2 to a bin. Measure how likely it is that a particle with  1 from one bin has a  1 from another one and represent it as a matrix x11x11 x22x22

+ Let’s break p(  ) into bins – for sake of depiction say 4 x11x11 x22x2 What correlation is saying is that  2 can only come from bin 2 or 3 now and we can measure the probability that it comes from bin 2 or bin

+ Transition Matrix (Next Bin) (Initial Bin) How to read this If  1 comes from bin 1 then it has an 80% chance that  2 will be from bin 1 and 20% that it will be from bin 2 and 0% chance from bins 3 or 4 If  1 comes from bin 3 then it has an 0% chance that  2 will be from bin 1 and 15% that it will be from bin 2, 75% from bin 3 and and 10% chance from bin 4 Thus, I randomly generate  1 from p(  ). Then I generate another random number to assess where  2 will come from and then only sample from that bin. I generate a new number to assess where  3 will come from based on  2 and so on until I have jumped my desired number of steps

+ Example – same case as before with D=0.001 (where CTRW did not work) Recall this is what p(  ) looked like If will break this into 25 bins The higher the number of bins the better you resolve correlation – typically is good.

+ Check that indeed p(  ) is the same for  1 and  2 Run code rw_ptau_2dx.m

+ Measure the transition matrix Run code transition_matrix.m 11 22 Bin 1 is the smallest travel times (i.e. fastest particles). Bin 25 is the largest travel times (slowest particles)

+ Cumulative Transition Matrix This Matrix adds columns together… last column has to be 1 Why would I do this? Easier for coding… You know  1, which means you know which row you are in. To determine next bin generate a random number U=[0,1] This determines your next bin by looking along the columns..

+ Now use this transition matrix to enforce correlations in a CTRW code See code mcctrw.m (sample below)

+ How does this compare… Recall this is what we want to reproduce…

+ This is what the CTRW without correlation gave us Not a good match

+ How about our new algorithm… Almost perfect.. Including correlation the way we have seems to fix the issues What about larger D?

+ For D=0.01 What uncorrelated model gave us Better than smaller D But not Perfect

+ What does transition matrix look like for our new model? Compare to smaller D

+ What does transition matrix look like for our new model? D=0.01 D=0.001 What’s different – Does it make sense to you?

+ How about our new algorithm… Almost perfect.. Including correlation the way we have seems to fix the issues What about larger D?

+ Finally for largest D=0.1 What uncorrelated model gave us Virtually Indistinguishable

+ Transition Matrix for this Case Again, compare to before and does it make sense?

+ Comparison Again, virtually perfect (but unnecessary to include correlations)

+ Summary of Algorithm for McCTRW Step 1 – Generate a random number   from p(  ) Step 2 – Determine what bin is   in Step 3 – Generate a random number and use transition matrix to assess what the next bin will be Step 4 – Generate   from the new bin Step 5 - Generate a random number and use transition matrix to assess what the next bin will be Step 6 – Generate   from the new bin And so on until the desired number of jumps have passed.

+ Ok – it works for flow between two plates – does it work in other flows? For this let’s dig in the litterature… Thursday