Mathematical Analysis of Complex Networks and Databases Philippe Blanchard Dima Volchenkov.

Slides:



Advertisements
Similar presentations
Chapter 0 Review of Algebra.
Advertisements

Common Variable Types in Elasticity
10.4 Complex Vector Spaces.
Signals & Systems Predicting System Performance February 27, 2013.
5.1 Real Vector Spaces.
Chapter 4 Euclidean Vector Spaces
Common Variable Types in Elasticity
Link Analysis: PageRank
CHAPTER ONE Matrices and System Equations
10/11/2001Random walks and spectral segmentation1 CSE 291 Fall 2001 Marina Meila and Jianbo Shi: Learning Segmentation by Random Walks/A Random Walks View.
Entropy Rates of a Stochastic Process
DARPA Oasis PI Meeting Hilton Head, SC March 2002.
Chapter 5 Orthogonality
6. One-Dimensional Continuous Groups 6.1 The Rotation Group SO(2) 6.2 The Generator of SO(2) 6.3 Irreducible Representations of SO(2) 6.4 Invariant Integration.
Ch 7.3: Systems of Linear Equations, Linear Independence, Eigenvalues
Chapter 3 Determinants and Matrices
II–2 DC Circuits I Theory & Examples.
PHY 042: Electricity and Magnetism
資訊科學數學11 : Linear Equation and Matrices
Quantum One: Lecture 8. Continuously Indexed Basis Sets.
Applied Discrete Mathematics Week 10: Equivalence Relations
 Limit Cycle Isolated closed trajectories Analysis is, in general, quite difficult. For nonlinear systems, two simple results exist. Limit Cycle
2009 Mathematics Standards of Learning Training Institutes Algebra II Virginia Department of Education.
Absolute error. absolute function absolute value.
GROUPS & THEIR REPRESENTATIONS: a card shuffling approach Wayne Lawton Department of Mathematics National University of Singapore S ,
1 February 24 Matrices 3.2 Matrices; Row reduction Standard form of a set of linear equations: Chapter 3 Linear Algebra Matrix of coefficients: Augmented.
Linear Algebra/Eigenvalues and eigenvectors. One mathematical tool, which has applications not only for Linear Algebra but for differential equations,
A Study of The Applications of Matrices and R^(n) Projections By Corey Messonnier.
CHAPTER FIVE Orthogonality Why orthogonal? Least square problem Accuracy of Numerical computation.
+ Review of Linear Algebra Optimization 1/14/10 Recitation Sivaraman Balakrishnan.
Elementary Linear Algebra Anton & Rorres, 9th Edition
tensor calculus 03 - tensor calculus - tensor analysis.
1 PATTERN COMPARISON TECHNIQUES Test Pattern:Reference Pattern:
Chapter 3 Vectors in n-space Norm, Dot Product, and Distance in n-space Orthogonality.
PageRank. s1s1 p 12 p 21 s2s2 s3s3 p 31 s4s4 p 41 p 34 p 42 p 13 x 1 = p 21 p 34 p 41 + p 34 p 42 p 21 + p 21 p 31 p 41 + p 31 p 42 p 21 / Σ x 2 = p 31.
David Renardy.  Simple Group- A nontrivial group whose only normal subgroups are itself and the trivial subgroup.  Simple groups are thought to be classified.
Non-Euclidean Example: The Unit Sphere. Differential Geometry Formal mathematical theory Work with small ‘patches’ –the ‘patches’ look Euclidean Do calculus.
AGC DSP AGC DSP Professor A G Constantinides©1 Signal Spaces The purpose of this part of the course is to introduce the basic concepts behind generalised.
Relevant Subgraph Extraction Longin Jan Latecki Based on : P. Dupont, J. Callut, G. Dooms, J.-N. Monette and Y. Deville. Relevant subgraph extraction from.
Review of Matrix Operations Vector: a sequence of elements (the order is important) e.g., x = (2, 1) denotes a vector length = sqrt(2*2+1*1) orientation.
Mathematical Tools of Quantum Mechanics
COMS Network Theory Week 5: October 6, 2010 Dragomir R. Radev Wednesdays, 6:10-8 PM 325 Pupin Terrace Fall 2010.
1 Objective To provide background material in support of topics in Digital Image Processing that are based on matrices and/or vectors. Review Matrices.
Matrices, Vectors, Determinants.
Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley.
Lecture XXVI.  The material for this lecture is found in James R. Schott Matrix Analysis for Statistics (New York: John Wiley & Sons, Inc. 1997).  A.
Graphics Graphics Korea University kucg.korea.ac.kr Mathematics for Computer Graphics 고려대학교 컴퓨터 그래픽스 연구실.
CHAPTER ONE Matrices and System Equations Objective:To provide solvability conditions of a linear equation Ax=b and introduce the Gaussian elimination.
Lecture 11 Inner Product Spaces Last Time Change of Basis (Cont.) Length and Dot Product in R n Inner Product Spaces Elementary Linear Algebra R. Larsen.
Tutorial 6. Eigenvalues & Eigenvectors Reminder: Eigenvectors A vector x invariant up to a scaling by λ to a multiplication by matrix A is called.
Random Walks and Diffusions on Networks and Databases Dimitri Volchenkov (Bielefeld University)
Introduction to Random Walks and Diffusions to Network and Databases: from Electric Networks to Urban Spatial Networks Dimitri Volchenkov (Bielefeld University.
Is it possible to geometrize infinite graphs?
Mathematical Analysis of Complex Networks and Databases
Geometrize everything with Monge-Kantorovich?
Path-integral distance for the data analysis
Ca’ Foscari University of Venice;
Data Analysis of Multi-level systems
Review of Matrix Operations
Random remarks about random walks
Markov Chains Mixing Times Lecture 5
Mid-Term Exam preparation
Deterministic Gossiping
GROUPS & THEIR REPRESENTATIONS: a card shuffling approach
Chapter 3 Linear Algebra
Maths for Signals and Systems Linear Algebra in Engineering Lectures 10-12, Tuesday 1st and Friday 4th November2016 DR TANIA STATHAKI READER (ASSOCIATE.
Mathematical Models of Control Systems
16. Mean Square Estimation
Linear Vector Space and Matrix Mechanics
1st semester a.y. 2018/2019 – November 22, 2018
Presentation transcript:

Mathematical Analysis of Complex Networks and Databases Philippe Blanchard Dima Volchenkov

A network is any method of sharing information between systems consisting of many individual units, a measurable pattern of relationships among entities in a social, ecological, linguistic, musical, financial, etc. space What is a network/database? We suggest that these relationships can be expressed by large but finite matrices (often: with positive entries, symmetric)

Discovering the important nodes and quantifying differences between them in a graph is not easy, since the graph does not possess a metric space structure.

Συμμετρεῖν - to measure together Symmetry w.r.t. permutations (rearrangments) of objects G  A (adjacency matrix of the graph)

Συμμετρεῖν - to measure together Symmetry w.r.t. permutations (rearrangments) of objects   : [ ,A]=0, Automorphisms  G  A (adjacency matrix of the graph) A permutation matrix

Συμμετρεῖν - to measure together   : [ ,A]=0,  only trivial automorphisms  G  A (adjacency matrix of the graph)

Συμμετρεῖν - to measure together   : [ ,A]=0,  only trivial automorphisms  G  A (adjacency matrix of the graph) A permutation matrix is a stochastic matrix. We can extend the notion of automorphisms on the class of stochastic matrices.  T: [T, A]=0, Fractional automorphisms, or stochastic automorphisms

Συμμετρεῖν - to measure together   : [ ,A]=0,  only trivial automorphisms  G  A (adjacency matrix of the graph) A permutation matrix is a stochastic matrix. We can extend the notion of automorphisms on the class of stochastic matrices.  T: [T, A]=0, Fractional automorphisms, or stochastic automorphisms

Συμμετρεῖν - to measure together   : [ ,A]=0,  only trivial automorphisms  G  A (adjacency matrix of the graph) A permutation matrix is a stochastic matrix. We can extend the notion of automorphisms on the class of stochastic matrices.  T: [T, A]=0, Fractional automorphisms, or stochastic automorphisms We may remember the Birkhoff-von Neumann theorem asserting that every doubly stochastic matrix can be written as a convex combination of permutation matrices:        Compact graphs (trees, cycles)

Συμμετρεῖν - to measure together  T: [T, A]=0, Fractional automorphisms G  A (adjacency matrix of the graph) Infinitely many fractional automorphisms: Each T can be considered as a transition matrix of a Markov chain, a random walk defined on the graph/database.

Plan of the talk 1.Data/Graph probabilistic geometric manifolds; 2.Riemannian probabilistic geometry. The relations between the curvature of probabilistic geometric manifold and an intelligibility of the network/database; 3.The data dynamical model; data stability;

Fractional automorphisms establish an equivalence relation between the states (nodes) i ∼ j if an only if (T n ) ij > 0 for some n ≥ 0 and (T m ) ij > 0 for some m ≥ 0, and have all their states in one (communicating) equivalence class. The shortest-path distance, insensitive to the structure of the graph: The length of a walk The distance = “a Feynman path integral” sensitive to the global structure of the graph. Random Walks/ fractional automorphisms assign some probability to every possible path: Distance related to fractional automorphisms

Random walks (fractional automorphisms) on the graph/database  is the “laziness parameter”. ~ processes invariant w.r.t time-dilations “Nearest neighbor random walks”

Random walks (fractional automorphisms) on the graph/database  is the “laziness parameter”. ~ processes invariant w.r.t time-dilations, time units “Nearest neighbor random walks” “Scale- dependent random walks”

Random walks (fractional automorphisms) on the graph/database  is the “laziness parameter”. ~ processes invariant w.r.t time-dilations, time units “Nearest neighbor random walks” “Scale- dependent random walks” “Scale- invariant random walks (of maximal path-entropy)” All paths are equi-probable.

G  A   : [ ,A]=0, Automorphisms  T: [T, A]=0 , Green function  Green functions serve roughly an analogous role in partial differential equations as do Fourier series in the solution of ordinary differential equations. Green functions in general are distributions, not necessarily proper functions. x x' We can define a scalar product: Geometry From symmetry to geometry (a generalized inverse)

The problem is that From symmetry to geometry As being a member of a multiplicative group under the ordinary matrix multiplication, the Laplace operator possesses a group inverse (a special case of Drazin inverse) with respect to this group, L ♯, which satisfies the conditions: [L, L ♯ ] = [L ♯, A] =0 Green functions:

The problem is that From symmetry to geometry Green functions: The most elegant way is by considering the eigenprojection of the matrix L corresponding to the eigenvalue λ 1 = 1−μ 1 = 0 where the product in the idempotent matrix Z is taken over all nonzero eigenvalues of L.

The inner product between any two vectors The dot product is a symmetric real valued scalar function that allows us to define the (squared) norm of a vector Probabilistic Euclidean metric structure

Spectral representations of the probabilistic Euclidean metric structure The spectral representation of the (mean) first passage time to the node i ∈ V, the expected number of steps required to reach the node i ∈ V for the first time starting from a node randomly chosen among all nodes of the graph accordingly to the stationary distribution π. The kernel of the generalized inverse operator

Spectral representations of the probabilistic Euclidean metric structure The commute time, the expected number of steps required for a random walker starting at i ∈ V to visit j ∈ V and then to return back to i, The first-hitting time is the expected number of steps a random walker starting from the node i needs to reach j for the first time The matrix of first-hitting times is not symmetric, H ij ≠ H ji, even for a regular graph.

Electric resistance / Power grid networks a b An electrical network is considered as an interconnection of resistors. Kirchhoff circuit law can be described by the Kirchhoff circuit law,

Electric resistance / Power grid networks a b An electrical network is considered as an interconnection of resistors. Kirchhoff circuit law can be described by the Kirchhoff circuit law, Given an electric current from a to b of amount 1 A, the effective resistance of a network is the potential difference between a and b,

The effective resistance allows for the spectral representation: a b Electric resistance / Power grid networks The relation between the commute time of RW and the effective resistance: The (mean) first passage time to a node is nothing else but its electric potential in the resistance network.

Cities are the biggest editors of our life: built environments constrain our visual space and determine our ability to move thorough by structuring movement space. Some places in urban environments are easily accessible, others are not; well accessible places are more favorable to public, while isolated places are either abandoned, or misused. In a long time perspective, inequality in accessibility results in disparity of land prices: the more isolated a place is, the less its price would be. In a lapse of time, structural isolation would cause social isolation, as a host society occupies the structural focus of urban environments, while the guest society would typically reside in outskirts, where the land price is relatively cheap. The (mean) first-passage time in cities

SoHo East Harlem Federal Hall Bowery East Village Times Square

mean household income The data on the mean household income per year provided by

The data taken from the

The determinants of minors of the k th order of Ψ define an orthonormal basis in the

The squares of these determinants define the probability distributions over the ordered sets of k indexes: satisfying the natural normalization condition,

The squares of these determinants define the probability distributions over the ordered sets of k indexes: satisfying the natural normalization condition, The simplest example of such a probability distribution is the stationary distribution of random walks over the graph nodes.

The recurrence probabilities as principal invariants The Cayley – Hamilton theorem in linear algebra asserts that any N × N matrix is a solution of its associated characteristic polynomial. where the roots  are the eigenvalues of T, and {I k } N k=1 are its principal invariants, with I 0 = 1. As the powers of T determines the probabilities of transitions, we obtain the following expression for the probability of transition from i to j in t = N + 1 steps as the sign alternating sum of the conditional probabilities: |I 1 | = Tr T is the probability that a random walker stays at a node in one time step, |I N | = |det T| expresses the probability that the random walks revisit an initial node in N steps. where p ij (N+1-k) are the probabilities to reach j from i faster than in N + 1 steps, and |I k | are the k -steps recurrence probabilities quantifying the chance to return in k steps.

Probabilistic Riemannian geometry T x M  R N-1 x uiui ujuj  We can determine a node/entry dependent basis of vector fields on the probabilistic manifold: … and then define the metric tensor at each node/entry (of the database) by Small changes to data in a database/weights of nodes would rise small changes to the probabilistic geometric representation of database/graph. We can think of them as of the smooth manifolds with a Riemannian metric. Standard calculus of differential geometry…

Traps: (Mean) First Passage Time > Recurrence Time Mazes and labyrinths Probabilistic hypersurfaces of negative curvature “Confusing environments” It might be difficult to reach a place, but we return to the place quite often provided we reached that.

Z/ 12 Z Music = the cyclic group over the discrete space of notes: Motivated by the logarithmic pitch perception in humans, music theorists represent pitches using a numerical scale based on the logarithm of fundamental frequency. Probabilistic hypersurfaces of positive curvature Landmarks: (Mean) First Passage Time < Recurrence Time “Intelligible environments” Landmarks establishes a wayguiding structure that facilitates understanding of the environment. An example: The resulting linear pitch space in which octaves have size 12, semitones have size 1, and the number 69 is assigned to the note "A4".

A discrete model of music (MIDI) as a simple Markov chain In a musical dice game, a piece is generated by patching notes X t taking values from the set of pitches that sound good together into a temporal sequence.

First passage times to notes resolve tonality In music theory, the hierarchical pitch relationships are introduced based on a tonic key, a pitch which is the lowest degree of a scale and that all other notes in a musical composition gravitate toward. A successful tonal piece of music gives a listener a feeling that a particular (tonic) chord is the most stable and final. Namely, every pitch in a musical piece is characterized with respect to the entire structure of the Markov chain by its level of accessibility estimated by the first passage time to it that is the expected length of the shortest path of a random walk toward the pitch from any other pitch randomly chosen over the musical score. The values of first passage times to notes are strictly ordered in accordance to their role in the tone scale of the musical composition. The basic pitches for the E minor scale are "E", "F", "G", "A", "B", "C", and "D". The E major scale is based on "E", "F", "G", "A", "B", "C", and "D". The A major scale consists of "A", "B", "C", "D", "E", "F", and "G". Tonality structure of music The recurrence time vs. the first passage time over 804 compositions of 29 Western composers.