NETWORK ANALYSIS BASED UPON THE RENYI' ENTROPIES OF THE ASSOCIATED MARKOV MONOID TRANSFORMATION. USC CAS IMI Summer School on Network Science May 2013.

Slides:



Advertisements
Similar presentations
Chapter 6 Matrix Algebra.
Advertisements

Matrix Representation
3D Geometry for Computer Graphics
Lecture 3: A brief background to multivariate statistics
November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
BITTOR A New Foundation for Logical & Numeric Uncertainty Technical Overview Markov Monoids are used as a Mathematical Foundation for a New Theory of Logical.
Matrix Algebra Matrix algebra is a means of expressing large numbers of calculations made upon ordered sets of numbers. Often referred to as Linear Algebra.
Matrix Algebra Matrix algebra is a means of expressing large numbers of calculations made upon ordered sets of numbers. Often referred to as Linear Algebra.
Signal Spaces.
What is symmetry? Immunity (of aspects of a system) to a possible change.
Mathematics. Matrices and Determinants-1 Session.
DARPA Oasis PI Meeting Hilton Head, SC March 2002.
3_3 An Useful Overview of Matrix Algebra
X y z z′ R (  1,  2,  3 ) = 11 y′ 11 =x′ 22 22 22 x′′ z′′ =y′′ 33 y′′′ z′′′ = x′′′ 33 33.
Chapter 4.1 Mathematical Concepts
Computer Graphics Recitation 5.
Chapter 4.1 Mathematical Concepts. 2 Applied Trigonometry Trigonometric functions Defined using right triangle  x y h.
Rotation Group.  A metric is used to measure the distance in a space. Euclidean space is delta  An orthogonal transformation preserves the metric. Inverse.
Networks, Lie Monoids, & Generalized Entropy Metrics Networks, Lie Monoids, & Generalized Entropy Metrics St. Petersburg Russia September 25, 2005 Joseph.
Alon Arad Alon Arad Hurst Exponent of Complex Networks.
P460 - angular momentum1 Orbital Angular Momentum In classical mechanics, conservation of angular momentum L is sometimes treated by an effective (repulsive)
ExaSphere Network Analysis Engine © 2006 Joseph E. Johnson, PhD
CSCE 590E Spring 2007 Basic Math By Jijun Tang. Applied Trigonometry Trigonometric functions  Defined using right triangle  x y h.
P460 - angular momentum1 Orbital Angular Momentum In classical mechanics, conservation of angular momentum L is sometimes treated by an effective (repulsive)
Chapter 3 Determinants and Matrices
1 Network Metrics: A Network Analysis Engine ICCS 2006 Conference June 2006 Boston MA Joseph E. Johnson, PhD Department of Physics University of South.
MOHAMMAD IMRAN DEPARTMENT OF APPLIED SCIENCES JAHANGIRABAD EDUCATIONAL GROUP OF INSTITUTES.
New Approach to Quantum Calculation of Spectral Coefficients Marek Perkowski Department of Electrical Engineering, 2005.
Orbital Angular Momentum
Matrices and Determinants
Quantum One: Lecture 7. The General Formalism of Quantum Mechanics.
Stats & Linear Models.
Separate multivariate observations
Basics of Linear Algebra A review?. Matrix  Mathematical term essentially corresponding to an array  An arrangement of numbers into rows and columns.
Differential Equations
Compiled By Raj G. Tiwari
Piyush Kumar (Lecture 2: PageRank) Welcome to COT5405.
Graphics CSE 581 – Interactive Computer Graphics Mathematics for Computer Graphics CSE 581 – Roger Crawfis (slides developed from Korea University slides)
1 February 24 Matrices 3.2 Matrices; Row reduction Standard form of a set of linear equations: Chapter 3 Linear Algebra Matrix of coefficients: Augmented.
Eigenvalue Problems Solving linear systems Ax = b is one part of numerical linear algebra, and involves manipulating the rows of a matrix. The second main.
Recap of linear algebra: vectors, matrices, transformations, … Background knowledge for 3DM Marc van Kreveld.
Digital Image Processing, 3rd ed. © 1992–2008 R. C. Gonzalez & R. E. Woods Gonzalez & Woods Matrices and Vectors Objective.
generates 3-dimensional rotations
Matrices CHAPTER 8.1 ~ 8.8. Ch _2 Contents  8.1 Matrix Algebra 8.1 Matrix Algebra  8.2 Systems of Linear Algebra Equations 8.2 Systems of Linear.
Multivariate Statistics Matrix Algebra I W. M. van der Veld University of Amsterdam.
17. Group Theory 1.Introduction to Group Theory 2.Representation of Groups 3.Symmetry & Physics 4.Discrete Groups 5.Direct Products 6.Symmetric Groups.
Chapter 5 MATRIX ALGEBRA: DETEMINANT, REVERSE, EIGENVALUES.
PHARMACOECONOMIC EVALUATIONS & METHODS MARKOV MODELING IN DECISION ANALYSIS FROM THE PHARMACOECONOMICS ON THE INTERNET ®SERIES ©Paul C Langley 2004 Maimon.
Eigenvalues The eigenvalue problem is to determine the nontrivial solutions of the equation Ax= x where A is an n-by-n matrix, x is a length n column.
Introduction to Linear Algebra Mark Goldman Emily Mackevicius.
Comparing Two Motions Jehee Lee Seoul National University.
Matrices and Determinants
Basic Theory (for curve 01). 1.1 Points and Vectors  Real life methods for constructing curves and surfaces often start with points and vectors, which.
Computer Graphics Mathematical Fundamentals Lecture 10 Taqdees A. Siddiqi
Unsupervised Learning II Feature Extraction
Chapter 4.1 Mathematical Concepts. 2 Applied Trigonometry "Old Henry And His Old Aunt" Defined using right triangle  x y h.
1 Objective To provide background material in support of topics in Digital Image Processing that are based on matrices and/or vectors. Review Matrices.
Matrices, Vectors, Determinants.
Lecture from Quantum Mechanics. Marek Zrałek Field Theory and Particle Physics Department. Silesian University Lecture 6.
Postulates of Quantum Mechanics
Quantum One.
Numerical Analysis Lecture 16.
Orbital Angular Momentum
Dr Huw Owens Room B44 Sackville Street Building Telephone Number 65891
Chapter 3 Linear Algebra
Numerical Analysis Lecture 17.
Math review - scalars, vectors, and matrices
Presentation transcript:

NETWORK ANALYSIS BASED UPON THE RENYI' ENTROPIES OF THE ASSOCIATED MARKOV MONOID TRANSFORMATION. USC CAS IMI Summer School on Network Science May 2013 Joseph E. Johnson, PhD Distinguished Professor Emeritus Physics Department USC May 23, 2013 © 2013

Networks Defined A network is a set of points 1, 2, …(nodes) with connections among pairs of nodes. C ij is called the connection, connectivity, or adjacency matrix & defines the network. C ij gives the strength of connection between nodes i and j (a real non-negative #) C i≠j ≥ 0 One does not normally consider a negative connectivity value C ii is not defined – The connection of a thing to itself does not have meaning. Without a defined diagonal, C is relatively incomplete mathematically There is no meaning to an eigenvalue analysis without a diagonal. C can be extremely large yet many of the C values are zero (a ‘sparse matrix). C ij may be or not be symmetric.  (C ij –  )

We define a “graph” to be a network where the C values are 1 or 0 Thus things are either connected or not and the information is not very “rich” This is a special, somewhat degenerate case of a network. A network of real numbers contains a trillion times the information of a graph All of our network results are automatically applicable to graphs Analysis, visualization, and comparisons of topologies is very difficult The same topology and structure can be realized in a vast number of indistinguishable ways One of the fundamental problems is that there is no natural order to the nodes thus giving rise to the N! different C matrices that describe the exact same network Thus N! different arrangements of the nodes give different C matrices Consequently it is extremely difficult to compare.

Problems & Objectives How do we normally understand very complex physical systems? Consider the sound made by an instrument The wave f(t) is a function we seek to understand. We can expand f(t) usefully using Fourier analysis That is because columns of air and strings have standing waves that are multiples of the basic frequency Each of these frequencies has a different amplitude but the series consists of smaller and smaller terms This allows one to see the dominant system features first then increasingly see features of lesser importance In complex physical phenomena, we seek expansions in some useful series of approximations Often these are orthogonal functions, with terms that provide increasing accuracy of representation

We would like to have something similar for networks: A more solid mathematical foundation for networks generally. Metrics for expansion of topologies similar to Fourier analysis for sounds. Means to compare two topologies and thus also a change in a topology over time.. Means for classifying topologies Perhaps even a metric of “distance” between two topologies Perhaps even a structure that would support a dynamical theory of network evolution To describe what I have found, I need to give some brief background material

Lie Groups and Lie Algebras in Physics A group is a set of elements A, B, and a product with (a) closure AB=C, (b) associativity A(BC)=(AB)C, (c) an identity I with IA=AI=A, and (d) an inverse AA -1 =A -1 A=I But groups like rotations R= (cos  sin  ; sin , cos  ) have an ∞ of elements. Sophius Lie (1890s) used an exponentiation of an infinitesimal transformation R(  )=(cos  sin  ; sin , cos  )=(1, -   L  or R=e  L

Let a LVS be A+B = C and aA=D. A LVS with a metric A*B =  A i B i = AB cos is a Metric Space A LVS L =  a i L i with an antisymmetric product [ L i, L j ] = c ij k L k and Jacobi identity is a Lie Algebra We normally seek a subgroup of the GL(n, R) by requiring something is invariant: x 2 +y 2 = r 2, c 2 t 2 - r 2,  =1, translations, affine, conformal or other requirements All of relativity and quantum theory can be founded on Lie algebras & groups This proceedure also makes the concept of system invariants more obvious.

If our reality is based upon the foundations of physics, and those laws state that the state of everything is described by a vector, then how does the concept of a network, whose state is described by a connection matrix, arise?

The General Linear Lie Group Consider the group of all continuous linear transformations GL(n,R) Seek continuous transformations M that preserve ∑ x i (Markov Type) These are motions on a hyperplane perpendicular to (1, 1, 1, ….1), is invariant One can show: GL(n,R) = Scaling Algebra (S) + Markov Type Algebra (M) But M can take one from positive to negative values (not true Markov) Choose a Lie basis with a 1 at the ij position and a -1 at the jj position: In 2 dimensions: L 12 = (0, 1; 0, -1) and L 21 = (-1, 0; 1, 0) The Lie Algebra is then all elements of that form with L = ij L ij The commutator closes & the Jacobi Identity is satisfied (note that each column of ij L ij sums to 0)

The Lie Group is then M( ) = exp ( ij L ij ) which leaves ∑ x i invariant in x’ = M x With ij non-negative, then one stays in the positive hyperquadrant That is because all probabilities must be non-negative. But the inverse is lost so this is a Markov Monoid (a group without an inverse) Those transformations would take some states to negative probabilities.

The Lie Markov Monoid (MM) M( ) = exp ( ij L ij ) gives all continuous Markov Transformation that are continuously connected to the identity. These transformations work by giving a non-negative fraction of one x i to another x i. This is a rob Peter to pay Paul transformation Valid Markov transformations, represent irreversible diffusion, and have no inverse. There are (n 2 - n) L ij, one for each off diagonal element (i,j) Note that the diagonal elements are the negative sums of that column The rest of GL(n,R) is spanned by the Scaling Group S(n) = exp ( ii L ii ) Where L ii = 1 or 0, an n parameter Abelian Lie group that scales the axes.

Networks are 1 to 1 with MM Consider that any network C ij defines a  C  with L = ij L ij where M = exp( ij L ij ) Simply set ij = C ij for off diagonal terms and automatically the L ij will define the diagonals. For example in three dimensions if C = (0, 1, 2; 3, 0, 4; 5, 8, 0) then L = Then M = e sL is a true Markov transformation where s  is a continuous value Thus every network C defines a Lie algebra element that generates a one parameter Markov monoid transformation. One can study networks by studying the associated MM. The MM models any network as a transformation on a vector that gives a set of “entity flows” that conserve the imaginary entity ($, energy, water, probability, charge...)

An eigenvalue/eigenvector analysis can be done on the MM The eigenvalues are the exponential decreases in the associated eigenvalues which are linear combinations of nodes. These are similar to the normal nodes of oscilation for coupled harmonic oscillators All are less than 1 except for a single eigenvalue of 1 that gives equilibrium It is possible to have complex eigenvalues that correspond to network ‘cycles’.

M Supports Entropy Metrics The expansion of the M=exp ( ij L ij ) gives the degrees of separation Each term in the expansion maintains the Markov property M=1+  L + (1/2) ( L) 2 +… But it would seem we have not made real progress: The model of conserved flows is intuitive but does not solve things C and the associated L and M are often too large to practically execute eigenvalue analysis The MM Lie algebra has no Casimir operators and does not offer any deeper insight However, the columns of M are non-negative and sum to unity Thus the M ij can be thought of as probabilities and thus support a definition of entropy

We define the entropy of each column as the Renyi’ entropy R  R j c  = (1/(1-  ))* log 2 ((∑ i M  ij )  ) is defined for each column and for  degrees of separation. The same can be defined for rows as for columns. These entropies measure the order / disorder structure of the network “flows” The column and row entropies measure the incoming and outgoing flows for each node. They “distill” the information about the underlying topology.

Sorted Entropies Distill The Topology There is no natural order for the nodes of network making it combinatorially impossible to compare network topologies or to monitor a topology over time. But if we sort the Renyi’ entropy values by order, we get an entropy spectra: AND we get a unique node ordering. The entropy spectral curves distill the topology of the network By choosing both row and column Renyi entropies of multiple orders with multiple degrees of separation, the network can be expanded in curves of ‘diminishing’ importance. Thus we propose that these entropy spectral curves provide a powerful representation of the underlying topology and provide practical tools for comparing networks as well as monitoring a network over time.

The Topology is Expanded as a Series of Entropies and Degrees of Separation The collection of these multiple Renyi’ entropies for rows and columns of the Markov matrix, using Markov matrices that are expanded with different degrees of separation, can provide a complete description of any network (except for a possible degeneracy). We now have the expansion in a meaningful set of metrics for the topology Two networks that have different entropy spectra must be themselves different (when the s parameter is standardized with other factors).

Scalar product of  AB for two networks (or for two times) -> metric By forming R ,i c/r for each of two networks, one can take the difference of the column (or row) values (for multiple  and I) squared:  AB = √ (∑ ,i (R A,c, , i – R B,c, , i ) 2 ) as a measure of the “distance between the topologies A and B”. Here the ,i summation covers the Renyi & degrees of separation orders while i ranges over the nodes for each. Of course one could get a distance between the two topologies A and B by just taking the column entropies for the first degree of separation and using only the differences of the second order Renyi entropies.

By standardizing the parameter  and the normalizations of the C and thus the L matrix, then this parameter provides a distance metric between topologies. Then one could describe a given topology in terms of its distance from a set of “reference topologies”.

A Network of Networks Let us next imagine a network C AB defined in terms of nodes which are themselves all possible networks using a connection matrix C AB = exp-  AB ) 2 Note that this function (properly normalized) maximizes when the networks are close This is exactly what we want for the C AB connection matrix. This could provide a framework for partially classifying networks. A given network would be defined (partially) by its distance from a set of reference networks much like we use the positions of masses in our three dimensional space for regular objects. Having the “position” of a network that is changing now allows us to define its “velocity”.

Past Results We have been able to identify system attacks and abuse using the Renyi entropy for network data from a SC university network and using our software, by clicking the aberrant parts of the entropy spectra, to identify the specific offensive nodes. We have studied applications of this to networks of the U.S. economy (utilizing the Leontief Input-Output matrices from the BEA). We have also studied social networks among Physics students who work with each other – a heterogeneous network where some nodes are students, some are their grades, and some are parameters of their efforts. We are now investigating what we believe to be a new type of information network that has not previously been investigated.

The Algorithm – How it all works: Consider that any network C ij defines a  C  with L = ij L ij where M = exp(s ij L ij ) Simply set ij = C ij for off diagonal terms and automatically the L ij will define the diagonals. For example in three dimensions if C = (0, 1, 2; 3, 0, 4; 5, 8, 0) then L = But the C matrix is only defined to within an overall factor and we need to make this matrix smaller so that when the first term is added to the unit matrix I in the expansion, then the diagonal terms are not negative. Lets take the parameter s = “0.01” to get M = *L = Notice that each column sums to unity and that all elements are non-negative. It is straight forward to add any number of more terms OR NOT to include other degrees of separation in the M matrix.

Then the R 2 values for each node (column) are = - log 2 (M M M 31 2 ) etc for the next two columns or These are then sorted in numerical order to form a non- decreasing (or increasing if you prefer) curve. That is the entropy spectral curve for the three node network. When s=0.02 (twice as great then the entropies are: ) At higher values of s, one can include the s 2 term in the expansion (second order degree of separation).

Here is a Larger C Matrix with a Full Calculation and Entropy Plot Using Excel: Given a C Matrix, determine the L matrix, then computer the M matrix using different s values The s value is here set and used only with the first degree of separation (first power of sL). L MatrixThis is the C network matrix with the diagonal set to the negative of the sum of each column s=0.02 This is one example. S must be set so that 1+(the largest diagonal) > 0 M=I+sL or This matrix is a Markov matrix as one can easily verify Renyi 2 = - log2 (sum of squares of elements in that column) Renyi 2 sorted in numerical order: s = 0.001s = 0.005s = 0.01s =.02s =

Entropy curves for that example

Higher Order Degrees of Separation This example only included the first degree of separation By that we mean the first power of L in the expansion M =e sL Each higher order will alter the entropy curve slightly The amount of change will depend upon the s value. The M matrix will be Markov no matter how many or few degrees of separation one includes.

Networks of Different Numbers of Nodes One of the fundamental aspects of networks in the real world is that the number of nodes is constantly changing. In social networks one is constantly adding new members and loosing other members. The same phenomena is true in all networks thus presenting a difficult problem, namely: How can we compare network topologies as they acquire new nodes and lose some current nodes? We suggest the answer is to still use the comparison of the Renyi entropies as above but to smooth the values and scale the spectral curve between two fixed points – say 0 and 1. Thus a million node network would have the sorted entropies plotted every 1E-6 of the distance between 0 and 1. Then one is still comparing the entropy spectral shape as nodes come and go.

Open Questions of Interest Can one construct a dynamical theory based upon the lowest order changes in the components of expansion of low orders Renyi entropy and low degrees of separation using entropy spectral differences? Identify invariants as real networks evolve over time. What criteria will allow one to prove that the lower order expansions contain the dominant components of the network and allow us to ignore higher order terms? How are the expansion parameter (s value) and trace(L) values best standardized across different networks for these calculations? How are the Renyi row and Renyi column entropy spectra different for practical networks? How can fundamental topologies (random, scale free, rings, cliques, clusters, trees, etc.) be best classified using these entropy spectra?

Thank You for your interest. The author encourages partnering on research with him using these concepts. This research was supported by grants from DARPA with the author as PI The IP of this research is protected by U.S. Patent B2 owned by USC Web Site: Office: Physics Department PSC Room 405 University of South Carolina Columbia SC 29208

Some Plots From Previous Research Plots show (a) the entropy (by column or row) sorted by magnitude against (b) time and (c) sorted node number. We have performed the computations in real time here for internet traffic to identify anomalies. The software allows one to click on an entropy curve anomaly and identify the associated node in spite of the constantly changing sort order.

Column Entropy - Order 1

Column Entropy - Order 2

Column Entropy - Order 3

Order 1 – Order 2 Difference Plot

Order 2 – Order 3 Difference Plot

Column/Row Ratio Plot (Symmetry Plot) – Order 2