Constrained Optimization Rong Jin. Outline  Equality constraints  Inequality constraints  Linear Programming  Quadratic Programming.

Slides:



Advertisements
Similar presentations
3.6 Support Vector Machines
Advertisements

Support Vector Machine
Lecture 9 Support Vector Machines
EMGT 501 HW #1 Solutions Chapter 2 - SELF TEST 18
Classification / Regression Support Vector Machines
Copyright (c) 2003 Brooks/Cole, a division of Thomson Learning, Inc
Easy Optimization Problems, Relaxation, Local Processing for a small subset of variables.
Separating Hyperplanes
The Simplex Method: Standard Maximization Problems
Support Vector Machines
Optimization in Engineering Design 1 Lagrange Multipliers.
Dual Problem of Linear Program subject to Primal LP Dual LP subject to ※ All duality theorems hold and work perfectly!
Support Vector Classification (Linearly Separable Case, Primal) The hyperplanethat solves the minimization problem: realizes the maximal margin hyperplane.
SVM QP & Midterm Review Rob Hall 10/14/ This Recitation Review of Lagrange multipliers (basic undergrad calculus) Getting to the dual for a QP.
Economics 214 Lecture 37 Constrained Optimization.
EMGT 501 HW # (b) (c) 6.1-4, Due Day: Sep. 21.
Project  Now it is time to think about the project  It is a team work Each team will consist of 2 people  It is better to consider a project of your.
Finite Mathematics & Its Applications, 10/e by Goldstein/Schneider/SiegelCopyright © 2010 Pearson Education, Inc. 1 of 99 Chapter 4 The Simplex Method.
Constrained Optimization Rong Jin. Outline  Equality constraints  Inequality constraints  Linear Programming  Quadratic Programming.
Lecture 35 Constrained Optimization
Support Vector Machine (SVM) Classification
Support Vector Machines and Kernel Methods
Lecture outline Support vector machines. Support Vector Machines Find a linear hyperplane (decision boundary) that will separate the data.
Lecture 10: Support Vector Machines
D Nagesh Kumar, IIScOptimization Methods: M2L5 1 Optimization using Calculus Kuhn-Tucker Conditions.
5.6 Maximization and Minimization with Mixed Problem Constraints
Optimality Conditions for Nonlinear Optimization Ashish Goel Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A.
Greg GrudicIntro AI1 Support Vector Machine (SVM) Classification Greg Grudic.
Optimization Theory Primal Optimization Problem subject to: Primal Optimal Value:
Chapter 4 The Simplex Method
A brief maximum entropy tutorial. Overview Statistical modeling addresses the problem of modeling the behavior of a random process In constructing this.
Machine Learning Week 4 Lecture 1. Hand In Data Is coming online later today. I keep test set with approx test images That will be your real test.
Duality Theory 對偶理論.
The Simplex Method. Standard Linear Programming Problem Standard Maximization Problem 1. All variables are nonnegative. 2. All the constraints (the conditions)
Duality Theory  Every LP problem (called the ‘Primal’) has associated with another problem called the ‘Dual’.  The ‘Dual’ problem is an LP defined directly.
Department Of Industrial Engineering Duality And Sensitivity Analysis presented by: Taha Ben Omar Supervisor: Prof. Dr. Sahand Daneshvar.
CS Statistical Machine learning Lecture 18 Yuan (Alan) Qi Purdue CS Oct
Maximum Entropy (ME) Maximum Entropy Markov Model (MEMM) Conditional Random Field (CRF)
Computational Intelligence: Methods and Applications Lecture 23 Logistic discrimination and support vectors Włodzisław Duch Dept. of Informatics, UMK Google:
4  The Simplex Method: Standard Maximization Problems  The Simplex Method: Standard Minimization Problems  The Simplex Method: Nonstandard Problems.
Machine Learning Weak 4 Lecture 2. Hand in Data It is online Only around 6000 images!!! Deadline is one week. Next Thursday lecture will be only one hour.
D Nagesh Kumar, IIScOptimization Methods: M2L4 1 Optimization using Calculus Optimization of Functions of Multiple Variables subject to Equality Constraints.
CS 478 – Tools for Machine Learning and Data Mining SVM.
Kernel Methods: Support Vector Machines Maximum Margin Classifiers and Support Vector Machines.
Optimization unconstrained and constrained Calculus part II.
Survey of Kernel Methods by Jinsan Yang. (c) 2003 SNU Biointelligence Lab. Introduction Support Vector Machines Formulation of SVM Optimization Theorem.
Mathe III Lecture 8 Mathe III Lecture 8. 2 Constrained Maximization Lagrange Multipliers At a maximum point of the original problem the derivatives of.
Chapter 4 Sensitivity Analysis, Duality and Interior Point Methods.
(iii) Lagrange Multipliers and Kuhn-tucker Conditions D Nagesh Kumar, IISc Introduction to Optimization Water Resources Systems Planning and Management:
Calculus-Based Optimization AGEC 317 Economic Analysis for Agribusiness and Management.
Nonlinear Programming In this handout Gradient Search for Multivariable Unconstrained Optimization KKT Conditions for Optimality of Constrained Optimization.
A Brief Maximum Entropy Tutorial Presenter: Davidson Date: 2009/02/04 Original Author: Adam Berger, 1996/07/05
Economics 2301 Lecture 37 Constrained Optimization.
Greg GrudicIntro AI1 Support Vector Machine (SVM) Classification Greg Grudic.
Kernel Methods: Support Vector Machines Maximum Margin Classifiers and Support Vector Machines.
Inequality Constraints Lecture 7. Inequality Contraints (I) n A Review of Lagrange Multipliers –As we discussed last time, the first order necessary conditions.
Support Vector Machine: An Introduction. (C) by Yu Hen Hu 2 Linear Hyper-plane Classifier For x in the side of o : w T x + b  0; d = +1; For.
Mathe III Lecture 8 Mathe III Lecture 8. 2 Constrained Maximization Lagrange Multipliers At a maximum point of the original problem the derivatives of.
Copyright © 2006 Brooks/Cole, a division of Thomson Learning, Inc. Linear Programming: An Algebraic Approach 4 The Simplex Method with Standard Maximization.
1 Support Vector Machines: Maximum Margin Classifiers Machine Learning and Pattern Recognition: September 23, 2010 Piotr Mirowski Based on slides by Sumit.
Chapter 11 Optimization with Equality Constraints
Support Vector Machines Introduction to Data Mining, 2nd Edition by
The Simplex Method: Standard Minimization Problems
Recap: Conditional Exponential Model
Machine Learning Week 3.
Chapter 5. The Duality Theorem
Lecture 18. SVM (II): Non-separable Cases
The Improved Iterative Scaling Algorithm: A gentle Introduction
Calculus-Based Optimization AGEC 317
Constraints.
Presentation transcript:

Constrained Optimization Rong Jin

Outline  Equality constraints  Inequality constraints  Linear Programming  Quadratic Programming

Optimization Under Equality Constraints  Maximum Entropy Model  English ‘in’  French {dans (1), en (2), à (3), au cours de (4), pendant (5)}

Reducing variables  Representing variables using only p 1 and p 4  Objective function is changed  Solution: p 1 = 0.2, p 2 = 0.3, p 3 =0.1, p 4 = 0.2, p 5 = 0.2

Maximum Entropy Model for Classification  It is unlikely that we can use the previous simple approach to solve such a general  Solution: Lagrangian

Equality Constraints: Lagrangian  Introduce a Lagrange multiplier for the equality constraint  Construct the Lagrangian  Necessary condition A optimal solution for the original optimization problem has to be one of the stationary point of the Lagrangian

Example:  Introduce a Lagrange multiplier for constraint  Construct the Lagrangian  Stationary points

Lagrange Multipliers  Introducing a Lagrange multiplier for each constraint  Construct the Lagrangian for the original optimization problem

Lagrange Multiplier  We have more variables p 1, p 2, p 3, p 4, p 5 and, 1, 2, 3  Necessary condition (first order condition) A local/global optimum point for the original constrained optimization problem  a stationary point of the corresponding Lagrangian Original Entropy Function Constraints

Stationary Points for Lagrangian All probabilities p 1, p 2, p 3, p 4, p 5 are expressed as functions of Lagrange multipliers s

Dual Problem  p 1, p 2, p 3, p 4, p 5 are expressed as functions of s  We can even remove the variable 3  However, it is still difficult to obtain the solution s such that the constraints are satisfied  Dual problem Substitute the expression for ps into the Lagrangian Find the s that MINIMIZE the substituted Lagrangian

Dual Problem Finding s such that the above objective function is minimized Original Lagrangian Substituted Lagrangian Expression for ps

Dual Problem  Using dual problem Constrained optimization  unconstrained optimization  Need to change maximization to minimization  Only valid when the original optimization problem is convex/concave (strong duality) Dual Problem Primal Problem x*= * When convex/concave

Maximum Entropy Model for Classification  Introduce a Lagrange multiplier for each linear constraint

Maximum Entropy Model for Classification  Construct the Lagrangian for the original optimization problem Original Entropy Function Consistency Constraint Normalization Constraint

Stationary Points Stationary points: first derivatives are zero Sum of conditional probabilities must be one Conditional Exponential Model !

Support Vector Machine  Having many inequality constraints  Solving the above problem directly could be difficult  Many variables: w, b,   Unable to use nonlinear kernel function

Inequality Constraints: Modified Lagrangian  Introduce a Lagrange multiplier for the inequality constraint  Construct the Lagrangian  Kuhn-Tucker condition A optimal solution for the original optimization problem will satisfy the following conditions Non-negative Lagrange Multiplier Two cases: 1.g(x) = c, 2.g(x) > c  =0

Example:  Introduce a Lagrange multiplier for constraint  Construct the Lagrangian  KT conditions  Expressing objective function using  Solution is =3

SVM Model  Lagrange multipliers for inequality constraints Min  Max +  

SVM Model  Lagrangian for SVM model  Kuhn-Tucker condition

Dual Problem for SVM  Express w, b,  using  and   Finding solution satisfying KT conditions is difficult

Dual Problem for SVM  Rewrite the Lagrangian function using only  and   Simplify using KT conditions

Dual Problem for SVM  Final dual problem Maximize  Minimize

Quadratic Programming Find Subject to

Linear Programming Find Subject to Very very useful algorithm papers 100+ books 10+ courses 100s of companies Main methods Simplex method Interior point method Most important: how to convert a general problem into the above standard form

Example  Need to change max to min Find Subject to

Example  Need change  to  Find Subject to

Example  Need to convert the inequality Find Subject to

Example  Need change |x 3 | Find Subject to