M. Sznaier O. Camps Robust Systems Lab Dept. of Electrical and Computer Eng. Northeastern University Compressive Information Extraction TexPoint fonts.

Slides:



Advertisements
Similar presentations
Unsupervised Learning Clustering K-Means. Recall: Key Components of Intelligent Agents Representation Language: Graph, Bayes Nets, Linear functions Inference.
Advertisements

Polynomial dynamical systems over finite fields, with applications to modeling and simulation of biological networks. IMA Workshop on Applications of.
Various Regularization Methods in Computer Vision Min-Gyu Park Computer Vision Lab. School of Information and Communications GIST.
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Globally Optimal Estimates for Geometric Reconstruction Problems Tom Gilat, Adi Lakritz Advanced Topics in Computer Vision Seminar Faculty of Mathematics.
CSCI 347 / CS 4206: Data Mining Module 07: Implementations Topic 03: Linear Models.
Graph Laplacian Regularization for Large-Scale Semidefinite Programming Kilian Weinberger et al. NIPS 2006 presented by Aggeliki Tsoli.
Optimization in Financial Engineering Yuriy Zinchenko Department of Mathematics and Statistics University of Calgary December 02, 2009.
Basic Feasible Solutions: Recap MS&E 211. WILL FOLLOW A CELEBRATED INTELLECTUAL TEACHING TRADITION.
A Convex Optimization Approach to Model (In)validation of Switched ARX Systems with Unknown Switches Northeastern University Yongfang Cheng 1, Yin Wang.
EE462 MLCV Lecture Introduction of Graphical Models Markov Random Fields Segmentation Tae-Kyun Kim 1.
1 s-t Graph Cuts for Binary Energy Minimization  Now that we have an energy function, the big question is how do we minimize it? n Exhaustive search is.
Object Detection by Matching Longin Jan Latecki. Contour-based object detection Database shapes: …..
10/11/2001Random walks and spectral segmentation1 CSE 291 Fall 2001 Marina Meila and Jianbo Shi: Learning Segmentation by Random Walks/A Random Walks View.
MASKS © 2004 Invitation to 3D vision Lecture 8 Segmentation of Dynamical Scenes.
Jie Gao Joint work with Amitabh Basu*, Joseph Mitchell, Girishkumar Stony Brook Distributed Localization using Noisy Distance and Angle Information.
Dual Problem of Linear Program subject to Primal LP Dual LP subject to ※ All duality theorems hold and work perfectly!
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
Support Vector Machines and Kernel Methods
Approximation Algorithms: Bristol Summer School 2008 Seffi Naor Computer Science Dept. Technion Haifa, Israel TexPoint fonts used in EMF. Read the TexPoint.
Radial Basis Function Networks
More Realistic Power Grid Verification Based on Hierarchical Current and Power constraints 2 Chung-Kuan Cheng, 2 Peng Du, 2 Andrew B. Kahng, 1 Grantham.
Linear Algebra and Image Processing
Jinhui Tang †, Shuicheng Yan †, Richang Hong †, Guo-Jun Qi ‡, Tat-Seng Chua † † National University of Singapore ‡ University of Illinois at Urbana-Champaign.
1 Linear Methods for Classification Lecture Notes for CMPUT 466/551 Nilanjan Ray.
Machine Learning CUNY Graduate Center Lecture 3: Linear Regression.
C&O 355 Lecture 2 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A.
C&O 355 Mathematical Programming Fall 2010 Lecture 2 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A A.
Geometry and Algebra of Multiple Views
Cs: compressed sensing
Chapter 14: SEGMENTATION BY CLUSTERING 1. 2 Outline Introduction Human Vision & Gestalt Properties Applications – Background Subtraction – Shot Boundary.
Exploiting the complementarity structure: stability analysis of contact dynamics via sums-of-squares Michael Posa Joint work with Mark Tobenkin and Russ.
EE369C Final Project: Accelerated Flip Angle Sequences Jan 9, 2012 Jason Su.
Sparse Matrix Factorizations for Hyperspectral Unmixing John Wright Visual Computing Group Microsoft Research Asia Sept. 30, 2010 TexPoint fonts used in.
Learning With Structured Sparsity
TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A A A A A A A A Image:
Module networks Sushmita Roy BMI/CS 576 Nov 18 th & 20th, 2014.
Inference of Poisson Count Processes using Low-rank Tensor Data Juan Andrés Bazerque, Gonzalo Mateos, and Georgios B. Giannakis May 29, 2013 SPiNCOM, University.
1 Research Question  Can a vision-based mobile robot  with limited computation and memory,  and rapidly varying camera positions,  operate autonomously.
1 Markov Decision Processes Infinite Horizon Problems Alan Fern * * Based in part on slides by Craig Boutilier and Daniel Weld.
Approximate Inference: Decomposition Methods with Applications to Computer Vision Kyomin Jung ( KAIST ) Joint work with Pushmeet Kohli (Microsoft Research)
Support vector machine LING 572 Fei Xia Week 8: 2/23/2010 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A 1.
Lecture 2: Statistical learning primer for biologists
Lecture #14 Computational methods to construct multiple Lyapunov functions & Applications João P. Hespanha University of California at Santa Barbara Hybrid.
Rank Minimization for Subspace Tracking from Incomplete Data
CPSC 536N Sparse Approximations Winter 2013 Lecture 1 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAAAAAAA.
Large-Scale Matrix Factorization with Missing Data under Additional Constraints Kaushik Mitra University of Maryland, College Park, MD Sameer Sheoreyy.
Ultra-high dimensional feature selection Yun Li
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
Sequential Off-line Learning with Knowledge Gradients Peter Frazier Warren Powell Savas Dayanik Department of Operations Research and Financial Engineering.
A Binary Linear Programming Formulation of the Graph Edit Distance Presented by Shihao Ji Duke University Machine Learning Group July 17, 2006 Authors:
Jianchao Yang, John Wright, Thomas Huang, Yi Ma CVPR 2008 Image Super-Resolution as Sparse Representation of Raw Image Patches.
Super-resolution MRI Using Finite Rate of Innovation Curves Greg Ongie*, Mathews Jacob Computational Biomedical Imaging Group (CBIG) University of Iowa.
Multiplicative updates for L1-regularized regression
Recursive Identification of Switched ARX Hybrid Models: Exponential Convergence and Persistence of Excitation René Vidal National ICT Australia Brian D.O.Anderson.
Motion Segmentation with Missing Data using PowerFactorization & GPCA
René Vidal and Xiaodong Fan Center for Imaging Science
Segmentation of Dynamic Scenes
René Vidal Time/Place: T-Th 4.30pm-6pm, Hodson 301
Segmentation of Dynamic Scenes
A Unified Algebraic Approach to 2D and 3D Motion Segmentation
Segmentation of Dynamic Scenes from Image Intensities
Observability, Observer Design and Identification of Hybrid Systems
Nonnegative polynomials and applications to learning
Nuclear Norm Heuristic for Rank Minimization
Filtering and State Estimation: Basic Concepts
Segmentation of Dynamical Scenes
Linear Time Invariant systems
Outline Sparse Reconstruction RIP Condition
Presentation transcript:

M. Sznaier O. Camps Robust Systems Lab Dept. of Electrical and Computer Eng. Northeastern University Compressive Information Extraction TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A AA A AA A Robust Identification of Hybrid Systems C. Lagoa Dept. of Electrical Eng. Penn State University

Hybrid systems in control: Controller design reasonably well understood Identification/Validation: work in progress u y GσtGσt η

What do these have in common?: Detecting gene activity in a diauxic shift Human tracking and activity analysis Tumor detection in low contrast images In all cases, relevant events comparatively rare and encoded in 1/100 to less than 1/10 6 of the data

What do these have in common?: Detecting gene activity in a diauxic shift Human tracking and activity analysis Tumor detection in low contrast images Claim: A hidden hybrid systems identification problem

Strong prior: – Signal has a sparse representation only a few c i ≠ 0 Signal Recovery: – “sparsify” the coefficients Relax to LP: Compressive Sensing:

Where should we pay attention?: Features (edges, regions, etc.) are important.

Where should we pay attention?: Dynamics are important too!

Strong prior: – Signal has a sparse representation only a few c i ≠ 0 Signal Recovery: – “sparsify” the coefficients Relax to LP: Compressive Sensing:

Strong prior: – Signal has a sparse representation only a few c i ≠ 0 Signal Recovery: – “sparsify” the coefficients Relax to LP: Compressive Sensing: Strong prior: – Actionable information is generated by low complexity dynamical systems. Information extraction: – “sparsify” the dynamics Relax to SDP: Compressive information Extraction

Information extraction as an Id problem: – Model data streams as outputs of piecewise LTI systems – “Interesting” events  Model invariant(s) changes – “Homogeneous” segments  output of a single LTI sub-system u G(  ) y features, pixel values, …

Piecewise Affine (PWA) Systems Id problem Given : – Bounds on noise (|| η|| * · ² ), sub-system order (n o ) – Input/output data (u,y) Find: – A piecewise affine model such that

Piecewise Affine (PWA) Systems Id problem Ill posed, always has a trivial solution Given : – Bounds on noise (|| η|| * · ² ), sub-system order (n o ) – Input/output data (u,y) Find: – A piecewise affine model such that

Given : – Bounds on noise (|| η|| * · ² ), sub-system order (n o ) – Input/output data (u,y) Find: – A piecewise affine model such that with minimum number of switches systems Piecewise Affine (PWA) Systems Id problem

Non-zero g(t) = SWITCH Main idea : PWAS Id problem with min # switches:

Min # switches min||g|| o Main idea : A sparsification problem PWAS Id problem with min # switches:

Formally : PWAS Id problem with min # switches:

Formally : FACT: “exact” solution tktk t k+1 PWAS Id problem with min # switches:

Example: Video segmentation

PWAS Id problem with fixed # subsystems: Activity Analysis Need to tell when we are back to the original system Medical Image Segmentation

Given : – Bounds on noise (|| η|| * · ² ), sub-system order (n o ) – Input/output data (u,y) – Number of sub-models Find: – A piecewise affine model such that: PWAS Id problem with fixed # subsystems: NP-hard, MILP (Bemporad et. Al.)

Given : – Bounds on noise (|| η|| * · ² ), sub-system order (n o ) – Input/output data (u,y) – Number of sub-models Find: – A piecewise affine model such that: PWAS Id problem with fixed # subsystems: Reduces to a rank minimization problem

Given : – Bounds on noise (|| η|| * · ² ), sub-system order (n o ) – Input/output data (u,y) – Number of sub-models Find: – A piecewise affine model such that: PWAS Id problem with fixed # subsystems: Reduces to a SDP

PWAS Id problem in the noise free case: Neither the mode signal nor the parameters, b, are known! Independent of mode signal, linear in parameters, c! GPCA: an algebraic geometric method due to Vidal et al. Main Idea: * = 0

Toy example: 2 first order systems:

Function of the data only System parameters Independent of the data One such equation per data point

GPCA: an algebraic geometric method due to Vidal et al. Main Idea: PWAS Id problem in the noise free case: Embed in a higher dimensional space via Veronese map

GPCA: an algebraic geometric method due to Vidal et al. Main Idea: PWAS Id problem in the noise free case: Solve for c s from the null space of the embedded data matrix. Get b i from c s via polynomial differentiation Details in Vidal et al., 2003

GPCA: an algebraic geometric method due to Vidal et al. Main Idea: What happens with noisy measurements? Solve for c s from the null space of the embedded data matrix. Get b i from c s via polynomial differentiation toto η t η T η η t

GPCA: an algebraic geometric method due to Vidal et al. Main Idea: What happens with noisy measurements? Solve for c s from the null space of the embedded data matrix. Get b i from c s via polynomial differentiation Need to find the null space of a matrix that depends polynomially on the noise t η

GPCA: an algebraic geometric method due to Vidal et al. Main Idea: What happens with noisy measurements? Solve for c s from the null space of the embedded data matrix. Get b i from c s via polynomial differentiation Need to find the null space of a matrix that depends polynomially on the noise. Obvious approach: SVD toto η t η T η η t

Academic Example Noise bound: 0.25

GPCA: an algebraic geometric method due to Vidal et al. Main Idea: What happens with noisy measurements? Solve for c s from the null space of the embedded data matrix. Get b i from c s via polynomial differentiation Need to find the null space of a matrix that depends polynomially on the noise. Minimize rank V s w.r.t η t toto η t η T η η t

Detour: Polynomial Optimization Theorem: (P1) and (P2) are equivalent; that is: From Lasserre 01:

Detour: Polynomial Optimization Theorem: (P1) and (P2) are equivalent; that is: From Lasserre 01:

Detour: Polynomial Optimization From Lasserre 01: Affine in m i

Detour: Polynomial Optimization From Lasserre 01: Affine in m i Hausdorff, Hamburger moments problem. Set of LMIs.

Rank is not a polynomial function. Can we use ideas from polynomial optimization? – YES Optimization Problem 1: What happens with noisy measurements?

Rank is not a polynomial function. Can we use ideas from polynomial optimization? – YES Optimization Problem 1: What happens with noisy measurements? Optimization Problem 2:

Rank is not a polynomial function. Can we use ideas from polynomial optimization? – YES Optimization Problem 1: What happens with noisy measurements? Optimization Problem 2: Convex constraint set!!

Rank is not a polynomial function. Can we use ideas from polynomial optimization? – YES Optimization Problem 1: What happens with noisy measurements? Optimization Problem 2: Fact: – There exists a rank deficient solution for Problem 2 if and only if there exists a rank deficient solution for Problem 1. – If c belongs to the nullspace of the solution of Problem 2, there exists a noise value with such that c belongs to the nullspace of

Rank is not a polynomial function. Can we use ideas from polynomial optimization? – YES Optimization Problem 1: What happens with noisy measurements? Optimization Problem 2: Problem 2 – Matrix rank minimization – Subject to LMI constraints Use a convex relaxation (e.g. log-det heuristic of Fazel et al.) to solve Problem 2 Find a vector c in the nullspace Estimate noise by root finding (V s c = 0 polynomials of one variable) Proceed as in noise-free case

Academic Example

Noise bound: 0.25 Academic Example

Parameter estimation trueMoments-basedGPCA p Submodel 1p p p Submodel 2p p p Submodel 3p p Error ||Δp|| 2 Moments-based: GPCA:

Example: Human Activity Analysis WALKBENDWALK

Example: Recovering 3D Geometry Example: denoising and clustering

Example: Image Segmentation Original image GPCA segmentation “dynamics” based segmentation

Given: – A nominal hybrid model of the form: – A bound on the noise (||η|| ∞ ≤ε) – Experimental Input/Output Data Determine: – whether there exist noise and switching sequences – consistent with a priori information and experimental data Reduces to SPD via moments and duality Model (In)validation of SARX Systems Equivalent to checking emptyness of a semialgebraic set

Semi-algebraic Consistency Set

One of the submodels is active at time t (logical OR) Semi-algebraic Consistency Set * = 0

Semi-algebraic Consistency Set The model is invalid if and only if is empty. Possible to use Positivstellensatz to get invalidation certificates. However, easier to utilize problem structure via moment-based polynomial optimization: Model is invalid iff o*>0

Certificates for (In)Validation The model is invalid if and only if there exist an N such that solution of the moments-based relaxation is positive! Numerically, it is easier to examine the dual problem – Model is invalid if where p* is the maximum of the dual SDP of the N th order relaxation.

Problem has a sparse structure (running intersection property holds) Details in Lasserre 06 Polynomial Optimization pnapna +P n a +1 PTPT + … + No need to consider all cross moments!

Problem has a sparse structure (running intersection property holds) A moments-based relaxation (convergent as N ↑ ): Convex SDP! Polynomial Optimization standard relaxation: O((Tn y ) 2N ) variables exploiting structure: O((n a n y ) 2N ) variables

Example: Activity Monitoring Set of “normal” activities: walking and waiting Estimate center of mass with background subtraction Identified model for walk: Model for wait: Training sequence for WALK

Example: Activity Monitoring A priori hybrid model: walking and waiting, 4% noise Test sequences of hybrid behavior: WALK, WAIT RUN WALK, JUMP Not InvalidatedInvalidated

Identifying Sparse Dynamical Networks Who is in the same team? Who reacts to whom?

Given time series data: What causes what? (Granger causality) Are there hidden inputs? Identifying Sparse Dynamical Networks

Formalization as a graph id problem: Given time series data: What causes what? (Granger causality) Are there hidden inputs?

Each time series becomes a node in the graph Formalization as a graph id problem:

Each time series becomes a node in the graph Formalization as a graph id problem: Each link series becomes a system = a1a1 a2a2 a3a3 ??? ?

Problem Formulation In matrix form: where and for a single node

Problem Formulation Complete network structure can be written as where

A Sparsification Problem: Find block sparse solutions to: Heuristic solution: group-lasso:

A Sparsification Problem: Find block sparse solutions to: A better heuristics: re-weighted group-lasso: link strength 1 st order difference of u u ∂u

Distributed solution: Solve for one node at a time: Complexity of the global solution O(P 8 (PN+T) 8 ) vs Complexity of the distributed solution O(P (PN+T) 8 )

Examples Australian Open Doubles Tennis Final game Network identified using Distributed Method Network identified using Competing Method

Examples Interesting structure is unveiled: A strong relationship was identified between the Brazilian Real and Canadian Dollar, the 8th and 9th largest world economies, respectively. The strongest connection of the Australian Dollar is with the New Zealand Dollar, as it would be expected from geographical proximity. One of the strongest connection of Chinese Yuan is to the United States Dollar. This is expected since the Yuan is indeed pegged to the US Dollar, rather than floating freely.

Examples Our method is capable giving more insight through identified external inputs The two identified jumps at Sep/10/2007(A) Oct/3/2007(B) Google Insight Results with headlines for ‘Brazil Real’ Keyword A)’Brazil real gains beyond 1.9 per dollar on inflows’ by Reuters on Sep/13/2007 B)’Brazil Real Strengthens Beyond 1.80 Per Dollar, a 7-Year High’ by Bloomberg Oct/11/2007

Finding “coordinated”activites: A “maximal clique” type problem

Finding “coordinated”activites: Rank minimization with sparsity constraints

Strong prior: – Signal has a sparse representation only a few c i ≠ 0 Signal Recovery: – “sparsify” the coefficients Relax to LP: Compressive Sensing: Strong prior: – Actionable information is generated by low complexity dynamical systems. Information extraction: – “sparsify” the dynamics Relax to SDP: Compressive information Extraction

– Data as manifestation of hidden, “sparse” dynamic structures – Extracting information from high volume data streams: finding changes in dynamic invariants (often no need to find the models) – Dynamic models as very compact, robust data surrogates – An Interesting connection between several communities: Control, computer vision, systems biology, compressive sensing, machine learning,…. Dynamic models as the key to encapsulate and analyze (extremely) high dimensional data Compressive Information Extraction:

Acknowledgements: Many thanks to: – Workshop organizers – Students Dr. N. Ozay, M. Ayazoglu – Funding agencies (NSF, AFOSR, DHS) More information as