Chebyshev Estimator Presented by: Orr Srour. References Yonina Eldar, Amir Beck and Marc Teboulle, "A Minimax Chebyshev Estimator for Bounded Error Estimation"

Slides:



Advertisements
Similar presentations
Ordinary Least-Squares
Advertisements

Duality for linear programming. Illustration of the notion Consider an enterprise producing r items: f k = demand for the item k =1,…, r using s components:
General Linear Model With correlated error terms  =  2 V ≠  2 I.
MS&E 211 Quadratic Programming Ashish Goel. A simple quadratic program Minimize (x 1 ) 2 Subject to: -x 1 + x 2 ≥ 3 -x 1 – x 2 ≥ -2.
Globally Optimal Estimates for Geometric Reconstruction Problems Tom Gilat, Adi Lakritz Advanced Topics in Computer Vision Seminar Faculty of Mathematics.
1 Modeling and Optimization of VLSI Interconnect Lecture 9: Multi-net optimization Avinoam Kolodny Konstantin Moiseev.
Dragan Jovicic Harvinder Singh
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
C&O 355 Mathematical Programming Fall 2010 Lecture 15 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AA A.
Optimization of thermal processes2007/2008 Optimization of thermal processes Maciej Marek Czestochowa University of Technology Institute of Thermal Machinery.
Lecture 8 – Nonlinear Programming Models Topics General formulations Local vs. global solutions Solution characteristics Convexity and convex programming.
Graph Laplacian Regularization for Large-Scale Semidefinite Programming Kilian Weinberger et al. NIPS 2006 presented by Aggeliki Tsoli.
The General Linear Model. The Simple Linear Model Linear Regression.
A Randomized Polynomial-Time Simplex Algorithm for Linear Programming Daniel A. Spielman, Yale Joint work with Jonathan Kelner, M.I.T.
Visual Recognition Tutorial
Approximation Algoirthms: Semidefinite Programming Lecture 19: Mar 22.
Minimax Estimators Dominating the Least-Squares Estimator Zvika Ben-Haim and Yonina C. Eldar Technion - Israel Institute of Technology.
Feasibility, uncertainty and interpolation J. A. Rossiter (Sheffield, UK)
Totally Unimodular Matrices Lecture 11: Feb 23 Simplex Algorithm Elliposid Algorithm.
Minimaxity & Admissibility Presenting: Slava Chernoi Lehman and Casella, chapter 5 sections 1-2,7.
Unconstrained Optimization Problem
A Constraint Generation Approach to Learning Stable Linear Dynamical Systems Sajid M. Siddiqi Byron Boots Geoffrey J. Gordon Carnegie Mellon University.
Tutorial 10 Iterative Methods and Matrix Norms. 2 In an iterative process, the k+1 step is defined via: Iterative processes Eigenvector decomposition.
6 6.3 © 2012 Pearson Education, Inc. Orthogonality and Least Squares ORTHOGONAL PROJECTIONS.
AMSC 6631 Sparse Solutions of Linear Systems of Equations and Sparse Modeling of Signals and Images: Midyear Report Alfredo Nava-Tudela John J. Benedetto,
Machine Learning Week 4 Lecture 1. Hand In Data Is coming online later today. I keep test set with approx test images That will be your real test.
1. The Simplex Method.
© by Yu Hen Hu 1 ECE533 Digital Image Processing Image Restoration.
Linear Programming System of Linear Inequalities  The solution set of LP is described by Ax  b. Gauss showed how to solve a system of linear.
AN ORTHOGONAL PROJECTION
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
Orthogonality and Least Squares
Vector Norms and the related Matrix Norms. Properties of a Vector Norm: Euclidean Vector Norm: Riemannian metric:
Section 2.3 Properties of Solution Sets
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
I.4 Polyhedral Theory 1. Integer Programming  Objective of Study: want to know how to describe the convex hull of the solution set to the IP problem.
Robust Optimization and Applications Laurent El Ghaoui IMA Tutorial, March 11, 2003.
SYSTEMS Identification Ali Karimpour Assistant Professor Ferdowsi University of Mashhad Reference: “System Identification Theory For The User” Lennart.
CPSC 536N Sparse Approximations Winter 2013 Lecture 1 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAAAAAAA.
Large-Scale Matrix Factorization with Missing Data under Additional Constraints Kaushik Mitra University of Maryland, College Park, MD Sameer Sheoreyy.
Mathematical Analysis of MaxEnt for Mixed Pixel Decomposition
Inequalities for Stochastic Linear Programming Problems By Albert Madansky Presented by Kevin Byrnes.
Linear & Nonlinear Programming -- Basic Properties of Solutions and Algorithms.
Linear Programming Back to Cone  Motivation: From the proof of Affine Minkowski, we can see that if we know generators of a polyhedral cone, they.
C&O 355 Lecture 19 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A A A A A A A A A A.
Generalization Error of pac Model  Let be a set of training examples chosen i.i.d. according to  Treat the generalization error as a r.v. depending on.
Linear Programming Chap 2. The Geometry of LP  In the text, polyhedron is defined as P = { x  R n : Ax  b }. So some of our earlier results should.
Approximation Algorithms based on linear programming.
Regularized Least-Squares and Convex Optimization.
6 6.5 © 2016 Pearson Education, Ltd. Orthogonality and Least Squares LEAST-SQUARES PROBLEMS.
1 Chapter 4 Geometry of Linear Programming  There are strong relationships between the geometrical and algebraic features of LP problems  Convenient.
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
Georgina Hall Princeton, ORFE Joint work with Amir Ali Ahmadi
Proving that a Valid Inequality is Facet-defining
Nonnegative polynomials and applications to learning
Computability and Complexity
Chap 9. General LP problems: Duality and Infeasibility
Polyhedron Here, we derive a representation of polyhedron and see the properties of the generators. We also see how to identify the generators. The results.
Polyhedron Here, we derive a representation of polyhedron and see the properties of the generators. We also see how to identify the generators. The results.
Chapter 8. General LP Problems
Rank-Sparsity Incoherence for Matrix Decomposition
I.4 Polyhedral Theory (NW)
Maths for Signals and Systems Linear Algebra in Engineering Lecture 6, Friday 21st October 2016 DR TANIA STATHAKI READER (ASSOCIATE PROFFESOR) IN SIGNAL.
Back to Cone Motivation: From the proof of Affine Minkowski, we can see that if we know generators of a polyhedral cone, they can be used to describe.
I.4 Polyhedral Theory.
Proving that a Valid Inequality is Facet-defining
Chapter 8. General LP Problems
Simplex method (algebraic interpretation)
Chapter 8. General LP Problems
Chapter 2. Simplex method
Presentation transcript:

Chebyshev Estimator Presented by: Orr Srour

References Yonina Eldar, Amir Beck and Marc Teboulle, "A Minimax Chebyshev Estimator for Bounded Error Estimation" (2007), to appear in IEEE Trans. Signal Proc. Amir Beck and Yonina C. Eldar, Regularization in Regression with Bounded Noise: A Chebyshev Center Approach, SIAM J. Matrix Anal. Appl. 29 (2), (2007). Jacob (Slava) Chernoi and Yonina C. Eldar, Extending the Chebyshev Center estimation Technique, TBA

Chebyshev Center - Agenda Introduction CC - Basic Formulation CC - Geometric Interpretation CC - So why not..? Relaxed Chebyshev Center (RCC) Formulation of the problem Relation with the original CC Feasibility of the original CC Feasibility of the RCC CLS as a CC relaxation CLS vs. RCC Constraints formulation Extended Chebyshev Center

Notations y – boldface lowercase = vector y i - i ’ th component of the vector y A - boldface uppercase = matrix - hat = the estimated vector of x = A – B is PD, PSD

The Problem Estimate the deterministic parameter vector from observations with: A – n x m model matrix w – perturbation vector.

LS Solution When nothing else is known, a common approach is to find the vector that minimizes the data error: Known as “ least squares ”, this solution can be written explicitly: (Assuming A has a full column rank)

But … In practical situations A is often ill- conditioned -> poor LS results

Regularized LS Assume we have some simple prior information regarding the parameter vector x. Then we can use the regularized least squares (RLS):

But … But what if we have some prior information regarding the noise vector as well … ? What if we have some more complicated information regarding the parameter vector x?

Assumptions From now on we assume that the noise is norm-bounded : And that x lies in a set defined by: (hence C is the intersection of k ellipsoids)

Assumptions The feasible parameter set of x is then given by: (hence Q is compact) Q is assumed to have non-empty interior

Constrained Least Squares (CLS) Given the prior knowledge, a popular estimation strategy is: - Minimization of the data error over C - But: the noise constraint is unused … More importantly, it doesn ’ t necessarily lead to small estimation error:

Chebyshev Center The goal: estimator with small estimation error Suggested method: minimize the worst-case error over all feasible vectors

Chebyshev Center – Geometric Interpretation Alternative representation: -> find the smallest ball (hence its center and its radius r ) which encloses the set Q.

Chebyshev Center – Geometric Interpretation

Chebyshev Center This problem is more commonly known as finding “ Chebyshev ’ s Center ”. Pafnuty Lvovich Chebyshev –

Chebyshev Center – The problem The inner maximization is non-convex Computing CC is a hard optimization problem Can be solved efficiently over the complex domain for intersection of 2 ellipsoids

Relaxed Chebyshev Center (RCC) Let us consider the inner maximization first: and:

Relaxed Chebyshev Center (RCC) Denoting, we can write the optimization problem as: with: Concave Not Convex

Relaxed Chebyshev Center (RCC) Let us replace G with: And write the RCC as the solution of: Convex

Relaxed Chebyshev Center (RCC) T is bounded The objective is concave (linear) in The objective is convex in We can replace the order: min-max to max-min

Relaxed Chebyshev Center (RCC) The inner minimization is a simple quadratic problem resulting with Thus the RCC problem can be written as: Note: this is a convex optimization problem.

RCC as an upper bound for CC RCC is not generally equal to the CC (except for k = 1 over the the complex domain) Since we have: Hence the RCC provides an upper bound on the optimal minimax value.

CC – The problem

RCC Solution Theorem: The RCC estimator is given by:

RCC Solution Where are the optimal solution of: subject to:

RCC Solution – as SDP Or as a semidefinite program (SDP): s.t.:

Feasibility of the CC Proposition: is feasible. Proof: Let us write the opt. problem as: with: 1. Convex in 2. strictly convex: 3. has a UNIQUE solution

Feasibility of the CC Let us assume that is infeasible, and denote by y its projection onto Q. By the projection theorem: and therefore:

Feasibility of the CC So: Which using the compactness of Q implies: But this contradicts the optimality of. Hence: is unique and feasible.

Feasibility of the RCC Proposition: is feasible. Proof: Uniqueness follows from the approach used earlier. Let us prove feasibility by showing that any solution of the RCC is also a solution of the CC.

Feasibility of the RCC Let be a solution for the RCC problem. Then: Since: We get:

CLS as CC relaxation We now show that CLS is also a (looser) relaxation of the Chebyshev center. Reminder:

CLS as CC relaxation Note that is equivalent to Define the following CC relaxation: unharmed relaxed

CLS as CC relaxation Theorem: The CLS estimate is the same as the relaxed CC over V (here CCV). Proof: Les us assume is the CCV solution, and the RCC solution,. The RCC is a strictly convex problem, so its solution is unique:

CLS as CC relaxation Define It is easy to show that (hence it is a valid solution for the CCV)

CLS as CC relaxation Denote by the objective of the CCV. By definition: contradicting the optimality of. > 0

CLS vs. RCC Now, as in the proof of the feasibility of the RCC, we know that: And so: Which means that the CLS estimate is the solution of a looser relaxation than that of the RCC.

Modeling Constraints The RCC optimization method is based upon a relaxation of the set Q Different characterizations of Q may lead to different relaxed sets. Indeed, the RCC depends on the specific chosen form of Q. (unlike CC and CLS)

Linear Box Constraints Suppose we want to append box constraints upon x: These can also be written as: Which of the two is preferable … ?

Linear Box Constraints Define:

Linear Box Constraints Suppose, then: Since, it follows that: Which can be written as:

Linear Box Constraints Hence: T1 is a looser relaxation -> T2 is preferable.

Linear Box Constraints An example in R2: The constraints have been chosen as the intersection of: A randomly generated ellipsoid [-1, 1] x [-1, 1]

Linear Box Constraints

An example – Image Deblurring x is a raw vector of a 16 x 16 image. A is a 256 x 256 matrix representing atmospheric turbulence blur (4 HBW, 0.8 STD). w is a WGN vector with std The observations are Ax+w We want x back …

An example – Image Deblurring LS: RLS: with CLS: RCC:

Chebyshev Center - Agenda Introduction CC - Basic Formulation CC - Geometric Interpretation CC - So why not..? Relaxed Chebyshev Center (RCC) Formulation of the problem Relation with the original CC Feasibility of the original CC Feasibility of the RCC CLS as a CC relaxation CLS vs. RCC Constraints formulation Extended Chebyshev Center

Questions..? Now is the time …