A Newton Method for Linear Programming Olvi L. Mangasarian University of California at San Diego.

Slides:



Advertisements
Similar presentations
Engineering Optimization
Advertisements

Dragan Jovicic Harvinder Singh
Lecture 8 – Nonlinear Programming Models Topics General formulations Local vs. global solutions Solution characteristics Convexity and convex programming.
Wangmeng Zuo, Deyu Meng, Lei Zhang, Xiangchu Feng, David Zhang
Separating Hyperplanes
Inexact SQP Methods for Equality Constrained Optimization Frank Edward Curtis Department of IE/MS, Northwestern University with Richard Byrd and Jorge.
The Disputed Federalist Papers : SVM Feature Selection via Concave Minimization Glenn Fung and Olvi L. Mangasarian CSNA 2002 June 13-16, 2002 Madison,
Decomposable Optimisation Methods LCA Reading Group, 12/04/2011 Dan-Cristian Tomozei.
The Most Important Concept in Optimization (minimization)  A point is said to be an optimal solution of a unconstrained minimization if there exists no.
CS38 Introduction to Algorithms Lecture 15 May 20, CS38 Lecture 15.
Support Vector Machines
CSCI 3160 Design and Analysis of Algorithms Tutorial 6 Fei Chen.
Numerical Optimization
1-norm Support Vector Machines Good for Feature Selection  Solve the quadratic program for some : min s. t.,, denotes where or membership. Equivalent.
Dual Problem of Linear Program subject to Primal LP Dual LP subject to ※ All duality theorems hold and work perfectly!
Martin Burger Institut für Numerische und Angewandte Mathematik European Institute for Molecular Imaging CeNoS Total Variation and related Methods Numerical.
Proximal Support Vector Machine Classifiers KDD 2001 San Francisco August 26-29, 2001 Glenn Fung & Olvi Mangasarian Data Mining Institute University of.
Support Vector Machines Formulation  Solve the quadratic program for some : min s. t.,, denotes where or membership.  Different error functions and measures.
Constrained Optimization Rong Jin. Outline  Equality constraints  Inequality constraints  Linear Programming  Quadratic Programming.
Binary Classification Problem Learn a Classifier from the Training Set
Unconstrained Optimization Problem
1 Computational Learning Theory and Kernel Methods Tianyi Jiang March 8, 2004.
Active Set Support Vector Regression
Support Vector Regression David R. Musicant and O.L. Mangasarian International Symposium on Mathematical Programming Thursday, August 10, 2000
1 Multiple Kernel Learning Naouel Baili MRL Seminar, Fall 2009.
Optimality Conditions for Nonlinear Optimization Ashish Goel Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.
An Introduction to Optimization Theory. Outline Introduction Unconstrained optimization problem Constrained optimization problem.
Constrained Optimization Rong Jin. Outline  Equality constraints  Inequality constraints  Linear Programming  Quadratic Programming.
Approximation Algorithms: Bristol Summer School 2008 Seffi Naor Computer Science Dept. Technion Haifa, Israel TexPoint fonts used in EMF. Read the TexPoint.
Mathematical Programming in Support Vector Machines
Incremental Support Vector Machine Classification Second SIAM International Conference on Data Mining Arlington, Virginia, April 11-13, 2002 Glenn Fung.
Frank Edward Curtis Northwestern University Joint work with Richard Byrd and Jorge Nocedal February 12, 2007 Inexact Methods for PDE-Constrained Optimization.
Duality Theory 對偶理論.
Feature Selection in Nonlinear Kernel Classification Olvi Mangasarian & Edward Wild University of Wisconsin Madison Workshop on Optimization-Based Data.
Feature Selection in Nonlinear Kernel Classification Olvi Mangasarian Edward Wild University of Wisconsin Madison.
Linear Programming Boosting by Column and Row Generation Kohei Hatano and Eiji Takimoto Kyushu University, Japan DS 2009.
Support Vector Machines for Data Fitting and Classification David R. Musicant with Olvi L. Mangasarian UW-Madison Data Mining Institute Annual Review June.
Machine Learning Seminar: Support Vector Regression Presented by: Heng Ji 10/08/03.
Support Vector Machines in Data Mining AFOSR Software & Systems Annual Meeting Syracuse, NY June 3-7, 2002 Olvi L. Mangasarian Data Mining Institute University.
Proximal Support Vector Machine Classifiers KDD 2001 San Francisco August 26-29, 2001 Glenn Fung & Olvi Mangasarian Data Mining Institute University of.
Frank Edward Curtis Northwestern University Joint work with Richard Byrd and Jorge Nocedal January 31, 2007 Inexact Methods for PDE-Constrained Optimization.
Nonlinear Data Discrimination via Generalized Support Vector Machines David R. Musicant and Olvi L. Mangasarian University of Wisconsin - Madison
RSVM: Reduced Support Vector Machines Y.-J. Lee & O. L. Mangasarian First SIAM International Conference on Data Mining Chicago, April 6, 2001 University.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Exact Differentiable Exterior Penalty for Linear Programming Olvi Mangasarian UW Madison & UCSD La Jolla Edward Wild UW Madison December 20, 2015 TexPoint.
Data Mining via Support Vector Machines Olvi L. Mangasarian University of Wisconsin - Madison IFIP TC7 Conference on System Modeling and Optimization Trier.
1  Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
CPSC 536N Sparse Approximations Winter 2013 Lecture 1 N. Harvey TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: AAAAAAAAAA.
Inexact SQP methods for equality constrained optimization Frank Edward Curtis Department of IE/MS, Northwestern University with Richard Byrd and Jorge.
Part 3. Linear Programming 3.2 Algorithm. General Formulation Convex function Convex region.
Nonlinear Programming In this handout Gradient Search for Multivariable Unconstrained Optimization KKT Conditions for Optimality of Constrained Optimization.
Linear Programming Chapter 9. Interior Point Methods  Three major variants  Affine scaling algorithm - easy concept, good performance  Potential.
Proximal Plane Classification KDD 2001 San Francisco August 26-29, 2001 Glenn Fung & Olvi Mangasarian Second Annual Review June 1, 2001 Data Mining Institute.
Generalization Error of pac Model  Let be a set of training examples chosen i.i.d. according to  Treat the generalization error as a r.v. depending on.
OR II GSLM
Massive Support Vector Regression (via Row and Column Chunking) David R. Musicant and O.L. Mangasarian NIPS 99 Workshop on Learning With Support Vectors.
Linear Programming Piyush Kumar Welcome to CIS5930.
Minimal Kernel Classifiers Glenn Fung Olvi Mangasarian Alexander Smola Data Mining Institute University of Wisconsin - Madison Informs 2002 San Jose, California,
1 Support Vector Machines: Maximum Margin Classifiers Machine Learning and Pattern Recognition: September 23, 2010 Piotr Mirowski Based on slides by Sumit.
Exact Differentiable Exterior Penalty for Linear Programming
Multiplicative updates for L1-regularized regression
Computational Optimization
Chapter 5. Sensitivity Analysis
CS5321 Numerical Optimization
Part 3. Linear Programming
CS5321 Numerical Optimization
Part 3. Linear Programming
University of Wisconsin - Madison
University of Wisconsin - Madison
Minimal Kernel Classifiers
Presentation transcript:

A Newton Method for Linear Programming Olvi L. Mangasarian University of California at San Diego

Outline  Fast Newton method for class of linear programs  Large number of constraints (millions)  Moderate number of variables (hundreds)  Method is based on an overlooked fact  Dual of asymptotic exterior penalty problem of primal LP  Provides an exact least 2-norm solution of the dual LP  Use a finite value of the penalty parameter  Exact least 2-norm dual solution generates:  Highly accurate primal solution when it is unique  Globally convergent Newton method established  Eleven lines of MATLAB code  Method solves LP with 2-million constraints & 100 variables  14-place accuracy in 17.4 minutes on 400Mhz machine  CPLEX 6.5 ran out of memory on same problem  Dual method for million-variable LP

Primal-Dual LPs Primal Exterior Penalty & Dual Least 2-Norm Problem Dual LP: Exact Dual Least 2-Norm: Primal LP: Asymptotic Primal Exterior Penalty:

The Plus Function

Dual Least 2-Norm: Karush-Kuhn-Tucker Optimality Conditions for Least 2-Norm Dual Lagrangian: KKT Conditions: Equivalently:

Karush-Kuhn-Tucker Optimality Conditions for Least 2-Norm Dual (Continued) Thus : Hence: Thus: solves the exterior penalty problem for the primal LP We have:

Equivalence of Exact Least 2-Norm Dual Solution to Finite-Parameter Primal Penalty Minimization

Primal Exterior Penalty Primal Exterior Penalty(  exact least 2-norm dual solution) : Gradient: Generalized Hessian: where:

LP Newton Algorithm (LPN)

LPN Convergence

Remarks on LPN

lpgen: Random Solvable LP Generator %lpgen: Generate random solvable lp: min c'x s.t. Ax =< b; A:m-by-n %Input: m,n,d(ensity); Output: A,b,c; (x,u): primal-dual solution pl=inline('(abs(x)+x)/2');%pl(us) function tic;A=sprand(m,n,d);A=100*(A-0.5*spones(A)); u=sparse(10*pl(rand(m,1)-(m-3*n)/m)); x=10*spdiags((sign(pl(rand(n,1)-rand(n,1)))),0,n,n)*(rand(n,1)-rand(n,1)); c=-A'*u;b=A*x+spdiags((ones(m,1)-sign(pl(u))),0,m,m)*10*ones(m,1);toc0=toc; format short e;[m n d toc0]  Elements of A uniformly distributed between –50 and +50  Primal random solution x with elements in [-10,10], approximately half of which are zero  Dual random solution u in [0,10], approximately 3n of which are positive

lpnewt1: MATLAB LPN Algorithm without Armijo %lpnewt1: Solve primal LP: min c'x s.t. Ax= =0 %Input: c,A,b,epsi,delta,tol,itmax;Output:v(l2norm dual sol),z primal sol epsi=1e-3;tol=1e-12;delta=1e-4;itmax=100;%default inputs pl=inline('(abs(x)+x)/2');%pl(us) function tic;i=0;z=0;y=((A(1:n,:))'*A(1:n,:)+epsi*eye(n))\(A(1:n,:))'*b(1:n);%y0 while (i tol & toc<1800) df=A'*pl((A*y-b))+epsi*c; d2f=A'*spdiags(sign(pl(A*y-b)),0,m,m)*A+delta*speye(n); z=y;y=y-d2f\df; i=i+1; end toc1=toc;v=pl(A*y-b)/epsi;t=find(v);z=A(t,:)\b(t);toc2=toc; format short e;[epsi delta tol i-1 toc1 toc2 norm(x-y,inf) norm(x-z,inf)]

LPN & CPLEX 6.5 Comparison (LPN without Armijo vs. Dual Simplex) (400 Mhz Pentium II 2Gig) (oom=out of memory)

What About the Case of: n>> m ?  LPN is not appropriate  LPN needs to invert a very large n-by-n matrix  Use instead DLPN (Dual LPN):  Find least 2-norm primal solution  By solving dual exterior penalty problem with finite penalty parameter

Equivalence of Exact Least 2-Norm Primal Probelm and Asymptotic Dual Penalty Problem

Difficulty: Nonnegativity Constrained Dual Penalty  Use an exterior penalty to handle nonnegativity constraint  By:  Replace:

D LPN & CPLEX 7.5 Comparison (DLPN without Armijo vs. Primal Simplex ) (1.8 MHz Pentium 4 1Gig)(oom=out of memory;error 1001) (infs/unbd=presolve determines infeasibility/unboundedness;error 1101 )

Conclusion  LPN & DLPN: Fast new methods for solving linear programs  LPN capable of solving problems with millions of constraints and hundreds of variables  DLPN capable of solving problems with hundreds of constraints and millions of variables  Eleven lines of MATLAB code for each of LPN & DLPN  Competitive with state-of-the-art CPLEX  Very suitable for classification problems  Armijo stepsize not needed in many applications  Typical termination in 3 to 30 steps

Future Work  Investigate & establish finite termination for LPN  Extend to convex quadratic programming  Establish convergence/termination under no assumptions for LPN  Generate self-contained codes for classification and data- mining problems

Paper & MATLAB Code Available