Hiroki Sayama sayama@binghamton.edu NECSI Summer School 2008 Week 3: Methods for the Study of Complex Systems Stochastic Systems Hiroki Sayama sayama@binghamton.edu.

Slides:



Advertisements
Similar presentations
3D Geometry for Computer Graphics
Advertisements

Probability Distributions CSLU 2850.Lo1 Spring 2008 Cameron McInally Fordham University May contain work from the Creative Commons.
1 The PageRank Citation Ranking: Bring Order to the web Lawrence Page, Sergey Brin, Rajeev Motwani and Terry Winograd Presented by Fei Li.
Ch 9.1: The Phase Plane: Linear Systems
Google’s PageRank By Zack Kenz. Outline Intro to web searching Review of Linear Algebra Weather example Basics of PageRank Solving the Google Matrix Calculating.
Hierarchy in networks Peter Náther, Mária Markošová, Boris Rudolf Vyjde : Physica A, dec
Lecture 3: Markov processes, master equation
Entropy Rates of a Stochastic Process
Mobile Intelligent Systems 2004 Course Responsibility: Ola Bengtsson.
Visual Recognition Tutorial1 Random variables, distributions, and probability density functions Discrete Random Variables Continuous Random Variables.
CHAPTER 6 Statistical Analysis of Experimental Data
Boyce/DiPrima 9th ed, Ch 11.2: Sturm-Liouville Boundary Value Problems Elementary Differential Equations and Boundary Value Problems, 9th edition, by.
Lecture II-2: Probability Review
Modern Navigation Thomas Herring
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
Elec471 Embedded Computer Systems Chapter 4, Probability and Statistics By Prof. Tim Johnson, PE Wentworth Institute of Technology Boston, MA Theory and.
Presented By: - Chandrika B N
: Appendix A: Mathematical Foundations 1 Montri Karnjanadecha ac.th/~montri Principles of.
Piyush Kumar (Lecture 2: PageRank) Welcome to COT5405.
Chapter Twelve Census: Population canvass - not really a “sample” Asking the entire population Budget Available: A valid factor – how much can we.
Sullivan – Fundamentals of Statistics – 2 nd Edition – Chapter 11 Section 1 – Slide 1 of 34 Chapter 11 Section 1 Random Variables.
Random Walks and Semi-Supervised Learning Longin Jan Latecki Based on : Xiaojin Zhu. Semi-Supervised Learning with Graphs. PhD thesis. CMU-LTI ,
8 th Grade Math Common Core Standards. The Number System 8.NS Know that there are numbers that are not rational, and approximate them by rational numbers.
The PageRank Citation Ranking: Bringing Order to the Web Lawrence Page, Sergey Brin, Rajeev Motwani, Terry Winograd Presented by Anca Leuca, Antonis Makropoulos.
PageRank. s1s1 p 12 p 21 s2s2 s3s3 p 31 s4s4 p 41 p 34 p 42 p 13 x 1 = p 21 p 34 p 41 + p 34 p 42 p 21 + p 21 p 31 p 41 + p 31 p 42 p 21 / Σ x 2 = p 31.
Low-Dimensional Chaotic Signal Characterization Using Approximate Entropy Soundararajan Ezekiel Matthew Lang Computer Science Department Indiana University.
Introduction to Chaos by: Saeed Heidary 29 Feb 2013.
1 1 COMP5331: Knowledge Discovery and Data Mining Acknowledgement: Slides modified based on the slides provided by Lawrence Page, Sergey Brin, Rajeev Motwani.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
L 7: Linear Systems and Metabolic Networks. Linear Equations Form System.
1/18/2016Atomic Scale Simulation1 Definition of Simulation What is a simulation? –It has an internal state “S” In classical mechanics, the state = positions.
1 EE571 PART 3 Random Processes Huseyin Bilgekul Eeng571 Probability and astochastic Processes Department of Electrical and Electronic Engineering Eastern.
CHAPTER – 1 UNCERTAINTIES IN MEASUREMENTS. 1.3 PARENT AND SAMPLE DISTRIBUTIONS  If we make a measurement x i in of a quantity x, we expect our observation.
6/11/2016Atomic Scale Simulation1 Definition of Simulation What is a simulation? –It has an internal state “S” In classical mechanics, the state = positions.
Mathematical Formulation of the Superposition Principle
Shan Lu, Jieqi Kang, Weibo Gong, Don Towsley UMASS Amherst
Hiroki Sayama NECSI Summer School 2008 Week 3: Methods for the Study of Complex Systems Dynamical Systems and Phase Space Hiroki.
Hiroki Sayama NECSI Summer School 2008 Week 2: Complex Systems Modeling and Networks Network Models Hiroki Sayama
Modeling and Simulation Dr. Mohammad Kilani
Availability Availability - A(t)
Applied Discrete Mathematics Week 11: Relations
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
Link-Based Ranking Seminar Social Media Mining University UC3M
PageRank and Markov Chains
Hiroki Sayama NECSI Summer School 2008 Week 3: Methods for the Study of Complex Systems Information Theory p I(p)
Understanding Standards Event Higher Statistics Award
CHE 391 T. F. Edgar Spring 2012.
3.1 Expectation Expectation Example
Degree and Eigenvector Centrality
Centrality in Social Networks
Autonomous Cyber-Physical Systems: Dynamical Systems
Quantum One.
Quantum One.
BUS7010 Quant Prep Statistics in Business and Economics
Piyush Kumar (Lecture 2: PageRank)
Boyce/DiPrima 10th ed, Ch 7.5: Homogeneous Linear Systems with Constant Coefficients Elementary Differential Equations and Boundary Value Problems, 10th.
Lecture 2 – Monte Carlo method in finance
STOCHASTIC HYDROLOGY Random Processes
EE513 Audio Signals and Systems
8. Stability, controllability and observability
MAE 82 – Engineering Mathematics
Virtual University of Pakistan
Stability Analysis of Linear Systems
Maths for Signals and Systems Linear Algebra in Engineering Lectures 13 – 14, Tuesday 8th November 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR)
Geology Geomath Chapter 7 - Statistics tom.h.wilson
Shan Lu, Jieqi Kang, Weibo Gong, Don Towsley UMASS Amherst
Eigenvalues and Eigenvectors
Linear Time Invariant systems
Ch 7.5: Homogeneous Linear Systems with Constant Coefficients
Presentation transcript:

Hiroki Sayama sayama@binghamton.edu NECSI Summer School 2008 Week 3: Methods for the Study of Complex Systems Stochastic Systems Hiroki Sayama sayama@binghamton.edu

Four approaches to complexity Nonlinear Dynamics Complexity = No closed-form solution, Chaos Information Complexity = Length of description, Entropy Computation Complexity = Computational time/space, Algorithmic complexity Collective Behavior Complexity = Multi-scale patterns, Emergence

Stochastic system Deterministic system xt = F(xt-1, t) Stochastic system (w/ discrete states) Px(s; t) = Ss’ Px(s|s’) Px(s’; t-1) Px(s; t): Probability for x to be s at time t Px(s|s’): Transition probability for x in state s’ to change to s in one time step Assumed time-independent

Exercise Write the model of the following system in mathematical form Numbers on arrows are transition probabilities 1 1/4 3/4

Random Walk

Random walk Time series in which random noise is added to the system’s previous state xt = xt-1 + d d : random variable (e.g., +1 or -1) Many stochastic events that accumulate and form the system’s history

Exercise Derive a mathematical representation of random walk using probability distribution, assuming: Possible states: integers (unbounded) Probability of moving upward and downward: 50%-50% Px(s|s’) = ½ d(s, s’+1) + ½ d(s, s’-1) d(a,b): Kronecker’s delta function

Where will you be after t steps? Exact position of individual random walk is hard to predict, but you can still statistically describe where it is likely to go

Exercise Show analytically that the expected position of a particle in random walk will not change over time <s> = Ss s Px(s; t)

Exercise Show analytically that the root square means (RMS) distance a particle travels in random walk will grow over time with power=1/2 √<(s-s0)2> = { Ss (s-s0)2 Px(s; t) }1/2

Dependence on time scale Distribution of positions flattens over time (diffusion of probability)

After sufficiently long time Positions of many random-walking particles follow a “normal distribution” A.k.a. Gaussian distribution, bell curve ~ e-(x-m)2/(2s2)/(2ps2)1/2 m: mean s: standard deviation s m

Relation to the central limit theorem Sum (and hence average) of many independent random events will approximately follow a normal distribution (regardless of probability distribution of those random events) Each step of random walk corresponds to an independent event Position of a particle corresponds to the sum of those events

Self-Similar Properties of Individual Random Walk

Structure of a single random walk Random walk produces self-similar trajectory

Fractal time series Random walk is a very simple example of “fractal” time series Time series whose portion has statistical properties similar to those of the whole series itself Shows stochastic self-similarity The word “fractal” came from the fact that these structures have fractional (non-integral) dimensions

Power law (scaling law) P(w) ~ w-k Random walk’s power spectrum shows an inverse power law distribution For such systems, averaging system states over time fails w Power of small fluctuation Power of large fluctuation The more data you collect, the larger fluctuation you find (scaling)

Exercise: Failure of averaging Produce time series data using a random walk model Average the system state for the first t steps Plot this average as a function of t Does the average become more accurate as you increase t?

Dynamics of fractal time series They are always fluctuating, but not completely random There are gentle correlations between past, present, and future (regularity) The similar dynamics appear over a wide range of time scales, without any characteristic frequency

Fractal time series in society Many other real-world time series generated by complex social/ economical systems (i.e., many interacting parts) have fractal characteristics Stock prices Currency exchange rates

Fractal time series in biology Many biological/physiological time series data are found to be fractal Heartbeat rates, breathing intervals, human gaits, body temperature, etc. They are usually fractal if a subject is normal and healthy If they look very regular, the subject is probably not healthy For details see, e.g., http://www.physionet.org/tutorials/fmnc/

Transition Probability Matrix and Asymptotic Probability Distribution

Time evolution of probabilities If we consider probability distribution Px(si; t) as a meta-state of the system, its time evolution is linear Px(si; t) = Sj Px(si|sj) Px(sj; t-1) pt = A pt-1 pt: Probability vector A: Transition probability matrix If original state space is finite and discrete

Convenient properties of transition probability matrix The product of two TPMs is also a TPM All TPMs have eigenvalue 1 |l|  1 for all eigenvalues of any TPM If the transition network is strongly connected, the TPM has one and only one eigenvalue 1 (no degeneration)

Exercise: Prove the following The product of two TPMs is also a TPM All TPMs have eigenvalue 1 |l|  1 for all eigenvalues of any TPM If the transition network is strongly connected, the TPM has one and only one eigenvalue 1 (no degeneration)

Answer (1) All TPMs have eigenvalue 1 You can show that there exists a non- zero vector q that satisfies A q = q, i.e. (A-I) q = 0 → |A-I| = 0 This holds when column vectors of A-I are linearly dependent with each other (i.e., A-I maps vectors to a subspace of fewer dimensions)

Answer (1) A-I actually looks like this: Note that each column vector is in a subspace s1+s2+...+sn = 0 → |A-I| = 0 a11-1 a12 … a1n a21 a22-1 a2n : an1 an2 ann-1

Answer (2) |l|  1 for all eigenvalues of any TPM → |l|  1 For any l, Amq = lmq (q: eigenvector) Am is a product of TPM, therefore it must be a TPM as well whose elements are all <= 1 Amq = lmq can’t diverge → |l|  1

TPM and asymptotic probability distribution |l|  1 for all eigenvalues of any TPM If the transition network is strongly connected, the TPM has one and only one eigenvalue 1 (no degeneration) → This eigenvalue is a unique dominant eigenvalue and the probability vector will eventually converge to its corresponding eigenvector

An Application: Google’s “PageRank”

’s “PageRank” PageRank Explained (from http://www.google.com/technology/) “PageRank relies on the uniquely democratic nature of the web by using its vast link structure as an indicator of an individual page's value. In essence, Google interprets a link from page A to page B as a vote, by page A, for page B. But, Google looks at more than the sheer volume of votes, or links a page receives; it also analyzes the page that casts the vote. Votes cast by pages that are themselves "important" weigh more heavily and help to make other pages “important.” ”

PageRank explained mathematically Lawrence Page, Sergey Brin, Rajeev Motwani, Terry Winograd, 'The PageRank Citation Ranking: Bringing Order to the Web' (1998): http://www-db.stanford.edu/~backrub/pageranksub.ps Node: Web pages link: WWW hyperlinks State: Temporary “importance” of that node Its coefficient matrix is a transition probability matrix that can be obtained by dividing each column of the adjacency matrix by the number of 1’s in that column.

Example s1 0 0.5 0 0 1 0 0 0 0.5 0 1 0.5 0 0 0 0 0 0.5 0 0 0 0 0.5 0.5 0 s2 s5 s3 s4 (In the actual implementation of PageRank, positive non-zero weights are forcedly assigned to all pairs of nodes in order to make the entire network strongly connected)

Interpreting the PageRank network as a stochastic system (1) State of each node can be viewed as a relative population that are visiting the webpage at t At next timestep, the population will distribute to other webpages linked from that page evenly s1 s2 s5 s3 s4

Interpreting the PageRank network as a stochastic system (2) Importance of each webpage can be measured by calculating the final equilibrium state of such population distribution s1 s2 s5 s3 s4

Therefore... Just one dominant eigenvector of the TPM of a strongly connected network always exists, with l = 1 This shows the equilibrium distribution of the population over WWW So, just solve x = Ax and you will get the PageRank for all the web pages on the World Wide Web!!

Exercise s1 0 0.5 0 0 1 0 0 0 0.5 0 1 0.5 0 0 0 0 0 0.5 0 0 0 0 0.5 0.5 0 s2 s5 s3 s4 Calculate the PageRank of each node in the above network (the network is already strongly connected so you can directly calculate its dominant eigenvector)

Summary Stochastic systems may be modeled using probability distributions Random walk shows ~t1/2 RMS growth and fractal time series patterns Time evolution of probability distributions may be modeled and analyzed as a linear system Asymptotic distribution can be uniquely obtained if the transition network is strongly connected

FYI: Master equation A continuous-time analog of the same stochastic processes dP(s)/dt = Ss’≠s ( R(s|s’) P(s’) – R(s’|s) P (s) ) R(a|b): Rate of transition from b to a