Relevant Subgraph Extraction Longin Jan Latecki Based on : P. Dupont, J. Callut, G. Dooms, J.-N. Monette and Y. Deville. Relevant subgraph extraction from.

Slides:



Advertisements
Similar presentations
Based on slides by Y. Peng University of Maryland
Advertisements

. Markov Chains. 2 Dependencies along the genome In previous classes we assumed every letter in a sequence is sampled randomly from some distribution.
Lecture 5 Graph Theory. Graphs Graphs are the most useful model with computer science such as logical design, formal languages, communication network,
8.3 Representing Relations Connection Matrices Let R be a relation from A = {a 1, a 2,..., a m } to B = {b 1, b 2,..., b n }. Definition: A n m  n connection.
Graph-02.
Introduction to Graph “theory”
1 Slides based on those of Kenneth H. Rosen Slides by Sylvia Sorkin, Community College of Baltimore County - Essex Campus Graphs.
Link Analysis: PageRank
 Graph Graph  Types of Graphs Types of Graphs  Data Structures to Store Graphs Data Structures to Store Graphs  Graph Definitions Graph Definitions.
Topics Review of DTMC Classification of states Economic analysis
Random Walks Ben Hescott CS591a1 November 18, 2002.
Андрей Андреевич Марков. Markov Chains Graduate Seminar in Applied Statistics Presented by Matthias Theubert Never look behind you…
Entropy Rates of a Stochastic Process
More on Rankings. Query-independent LAR Have an a-priori ordering of the web pages Q: Set of pages that contain the keywords in the query q Present the.
Applied Discrete Mathematics Week 12: Trees
048866: Packet Switch Architectures Dr. Isaac Keslassy Electrical Engineering, Technion Review.
Chapter 9 Graph algorithms Lec 21 Dec 1, Sample Graph Problems Path problems. Connectedness problems. Spanning tree problems.
Link Analysis, PageRank and Search Engines on the Web
1 Markov Chains Algorithms in Computational Biology Spring 2006 Slides were edited by Itai Sharon from Dan Geiger and Ydo Wexler.
CSE 321 Discrete Structures Winter 2008 Lecture 25 Graph Theory.
Markov Chains Chapter 16.
CS6800 Advanced Theory of Computation Fall 2012 Vinay B Gavirangaswamy
The effect of New Links on Google Pagerank By Hui Xie Apr, 07.
6. Markov Chain. State Space The state space is the set of values a random variable X can take. E.g.: integer 1 to 6 in a dice experiment, or the locations.
The Erdös-Rényi models
Entropy Rate of a Markov Chain
Feature Matching Longin Jan Latecki. Matching example Observe that some features may be missing due to instability of feature detector, view change, or.
Graph Theoretic Concepts. What is a graph? A set of vertices (or nodes) linked by edges Mathematically, we often write G = (V,E)  V: set of vertices,
Random Walks and Semi-Supervised Learning Longin Jan Latecki Based on : Xiaojin Zhu. Semi-Supervised Learning with Graphs. PhD thesis. CMU-LTI ,
1 CS104 : Discrete Structures Chapter V Graph Theory.
Based on slides by Y. Peng University of Maryland
Graphs.  Definition A simple graph G= (V, E) consists of vertices, V, a nonempty set of vertices, and E, a set of unordered pairs of distinct elements.
Markov Chains and Random Walks. Def: A stochastic process X={X(t),t ∈ T} is a collection of random variables. If T is a countable set, say T={0,1,2, …
Data Structures & Algorithms Graphs
Chapter 10 Graph Theory Eulerian Cycle and the property of graph theory 10.3 The important property of graph theory and its representation 10.4.
An Introduction to Graph Theory
Matrices Section 2.6. Section Summary Definition of a Matrix Matrix Arithmetic Transposes and Powers of Arithmetic Zero-One matrices.
CSCI 115 Chapter 8 Topics in Graph Theory. CSCI 115 §8.1 Graphs.
8.4 Closures of Relations Definition: The closure of a relation R with respect to property P is the relation obtained by adding the minimum number of.
Stochastic Processes and Transition Probabilities D Nagesh Kumar, IISc Water Resources Planning and Management: M6L5 Stochastic Optimization.
PERTEMUAN 26. Markov Chains and Random Walks Fundamental Theorem of Markov Chains If M g is an irreducible, aperiodic Markov Chain: 1. All states are.
11. Markov Chains (MCs) 2 Courtesy of J. Bard, L. Page, and J. Heyl.
Fault Tree Analysis Part 11 – Markov Model. State Space Method Example: parallel structure of two components Possible System States: 0 (both components.
1 Closures of Relations Based on Aaron Bloomfield Modified by Longin Jan Latecki Rosen, Section 8.4.
Theory of Computational Complexity Probability and Computing Lee Minseon Iwama and Ito lab M1 1.
Week 11 - Wednesday.  What did we talk about last time?  Graphs  Paths and circuits.
1 GRAPH Learning Outcomes Students should be able to: Explain basic terminology of a graph Identify Euler and Hamiltonian cycle Represent graphs using.
رياضيات متقطعة لعلوم الحاسب MATH 226. Chapter 10.
Let E denote some event. Define a random variable X by Computing probabilities by conditioning.
Applied Discrete Mathematics Week 14: Trees
Markov Chains and Random Walks
Availability Availability - A(t)
Applied Discrete Mathematics Week 13: Graphs
Special Graphs By: Sandeep Tuli Astt. Prof. CSE.
Search Engines and Link Analysis on the Web
Representing Graphs and
Chapter 5 Fundamental Concept
Graph theory Definitions Trees, cycles, directed graphs.
Network analysis.
Deterministic Gossiping
Prof. Paolo Ferragina, Algoritmi per "Information Retrieval"
Discrete-time markov chain (continuation)
Markov Chains Part 5.
Prof. Paolo Ferragina, Algoritmi per "Information Retrieval"
Graph Operations And Representation
Connectivity Section 10.4.
Walks, Paths, and Circuits
Graphs G = (V, E) V are the vertices; E are the edges.
Applied Discrete Mathematics Week 13: Graphs
Based on slides by Y. Peng University of Maryland
Presentation transcript:

Relevant Subgraph Extraction Longin Jan Latecki Based on : P. Dupont, J. Callut, G. Dooms, J.-N. Monette and Y. Deville. Relevant subgraph extraction from random walks in a graph. RR , Catholic University of Louvain, November 2006.

This part is based on P. Dupont, J. Callut, G. Dooms, J.-N. Monette and Y. Deville. Relevant subgraph extraction from random walks in a graph. RR , Catholic University of Louvain, November Relevant Subgraph Extraction Let G = (V, E, A) denote a graph formed by a set V of vertices and a set E of edges, with n = |V | is the number of nodes. A is a weighted n x n adjacency matrix of G. Weights are assumed to be zero if and only if their corresponding edge does not belong to the graph. Otherwise, the weight between any pair of connected nodes is assumed strictly positive. An undirected graph is represented by a symmetric matrix A. We have a special set of nodes of interest We define a node relevance function and will use it to find the relevant subgraph containing K

Random Walk A random walk in a graph can be modeled by a Markov chain describing the sequence of nodes visited during the walk. A state of the Markov chain is associated with each node of the graph. Hence the terms nodes and states are used interchangeably. A random variable X(t) represents the current state of the Markov chain at time t. The probability of transiting to state j at time t+1, given that the current state is i at time t, is given by Thus, from any state i, the (stationary) probability to jump to state j is proportional to the weight a ij of the edge from i to j. The transition matrix P = [p ij ] of the Markov chain is related to the degree and adjacency matrix as P = D -1 A. P is a row-stochastic matrix. P is generally asymmetric. Node relevance nr(i) is the expected number of times a walk starting in a node in K goes through node i before getting absorbed by another node in K. Our goal is to compute the node relevance.

We first consider random walks starting from a given node of interest x in K and ending in any other node of interest y in K with y ≠ x. We define a modified transition matrix for which all nodes of interest but x have been transformed into absorbing states: k = |K|, x Q denotes the (n x k + 1) (n x k + 1) transition matrix between transient states, I is the identity matrix of order k - 1 and x R is a (n x k+1) x (k - 1) matrix. x R ir denotes the probability of a walk being in a transient state i to be absorbed in state r in one step. In an absorbing Markov chain, the node passage time n(x, i) is defined as the number of times a random walk starting in x goes through state i before getting absorbed.

Let {X l = i | X 0 = x} denote the random event of visiting state i in the lth step of a random walk starting in x. The expectation of n(x, i) is the expected number of times such events occur, in other words their respective probabilities, for any time index l: The infinite sum of the successive powers of the matrix x Q is known as the Neumann series and the last equality follows from the theory of stochastic matrices. The matrix is called the fundamental matrix of an absorbing Markov chain. x N xi is the expected number of times a walk starting in the node of interest x goes through state i before getting absorbed in any node of K – {x}, that is, before reaching a distinct node of interest. The expected length n x of any walk starting in x and ending in another node of interest K – {x}, is also the node passage times summed over all transient states: where V T = V – (K – {x})

Recall that node relevance nr(i) is the expected number of times a walk starting in a node in K goes through node i before getting absorbed by another node in K.

Once we have node relevance nr, we can easily select the K-relevant subgraph as the largest connected subgraph S, i.e., S is a connected subgraph with the largest sum It is easy to get S: we add nodes starting with the largest value of nr. We stop once the subgraph becomes disconnected. K-relevant subgraph