1 Efficient Parallel Software for Large-Scale Semidefinite Programs Makoto Tokyo-Tech Katsuki Chuo University MSC Yokohama.

Slides:



Advertisements
Similar presentations
Cognitive Radio Communications and Networks: Principles and Practice By A. M. Wyglinski, M. Nekovee, Y. T. Hou (Elsevier, December 2009) 1 Chapter 12 Cross-Layer.
Advertisements

Hybrid BDD and All-SAT Method for Model Checking Orna Grumberg Joint work with Assaf Schuster and Avi Yadgar Technion – Israel Institute of Technology.
A NOVEL APPROACH TO SOLVING LARGE-SCALE LINEAR SYSTEMS Ken Habgood, Itamar Arel Department of Electrical Engineering & Computer Science GABRIEL CRAMER.
Globally Optimal Estimates for Geometric Reconstruction Problems Tom Gilat, Adi Lakritz Advanced Topics in Computer Vision Seminar Faculty of Mathematics.
Automatic Control Laboratory, ETH Zürich Automatic dualization Johan Löfberg.
Sum of Squares and SemiDefinite Programmming Relaxations of Polynomial Optimization Problems The 2006 IEICE Society Conference Kanazawa, September 21,
Multidisciplinary Free Material Optimization of 2D and Laminate Structures Alemseged G Weldeyesus, PhD student Mathias Stolpe, Senior Scientist Stefanie.
Convex Position Estimation in Wireless Sensor Networks
Venkataramanan Balakrishnan Purdue University Applications of Convex Optimization in Systems and Control.
Graph Laplacian Regularization for Large-Scale Semidefinite Programming Kilian Weinberger et al. NIPS 2006 presented by Aggeliki Tsoli.
Javad Lavaei Department of Electrical Engineering Columbia University Graph-Theoretic Algorithm for Nonlinear Power Optimization Problems.
Scalable Stochastic Programming Cosmin Petra and Mihai Anitescu Mathematics and Computer Science Division Argonne National Laboratory Informs Computing.
OpenFOAM on a GPU-based Heterogeneous Cluster
Routing in WSNs through analogies with electrostatics December 2005 L. Tzevelekas I. Stavrakakis.
SDPA: Leading-edge Software for SDP Informs ’ 08 Tokyo Institute of Technology Makoto Yamashita Mituhiro Fukuda Masakazu Kojima Kazuhide Nakata.
Scalable Multi-Stage Stochastic Programming Cosmin Petra and Mihai Anitescu Mathematics and Computer Science Division Argonne National Laboratory DOE Applied.
Amir-Hamed Mohsenian-Rad, Jan Mietzner, Robert Schober, and Vincent W.S. Wong University of British Columbia Vancouver, BC, Canada {hamed, rschober,
Semidefinite Programming
Direct Convex Relaxations of Sparse SVM Antoni B. Chan, Nuno Vasconcelos, and Gert R. G. Lanckriet The 24th International Conference on Machine Learning.
EE 685 presentation Optimization Flow Control, I: Basic Algorithm and Convergence By Steven Low and David Lapsley Asynchronous Distributed Algorithm Proof.
Support Vector Regression David R. Musicant and O.L. Mangasarian International Symposium on Mathematical Programming Thursday, August 10, 2000
Amir-Hamed Mohsenian-Rad, Jan Mietzner, Robert Schober, and Vincent W.S. Wong University of British Columbia Vancouver, BC, Canada {hamed, rschober,
Pablo A. Parrilo ETH Zürich Semialgebraic Relaxations and Semidefinite Programs Pablo A. Parrilo ETH Zürich control.ee.ethz.ch/~parrilo.
1 Parallel Simulations of Underground Flow in Porous and Fractured Media H. Mustapha 1,2, A. Beaudoin 1, J. Erhel 1 and J.R. De Dreuzy IRISA – INRIA.
ME964 High Performance Computing for Engineering Applications “The real problem is not whether machines think but whether men do.” B. F. Skinner © Dan.
Scalable Stochastic Programming Cosmin G. Petra Mathematics and Computer Science Division Argonne National Laboratory Joint work with.
MUMPS A Multifrontal Massively Parallel Solver IMPLEMENTATION Distributed multifrontal.
An approach for solving the Helmholtz Equation on heterogeneous platforms An approach for solving the Helmholtz Equation on heterogeneous platforms G.
High Performance Solvers for Semidefinite Programs
Optimization for Operation of Power Systems with Performance Guarantee
Scalable Multi-Stage Stochastic Programming
Extensions to the OSiL schema: Matrix and cone programming Horand I. Gassmann, Dalhousie University Jun Ma, Kipp Martin, Imre Polik.
1 High-Performance Implementation of Positive Matrix Completion for SDPs Makoto Yamashita (Tokyo Institute of Technology) Kazuhide Nakata (Tokyo Institute.
Sparsity in Polynomial Optimization IMA Annual Program Year Workshop "Optimization and Control" Minneapolis, January 16-20, 2007 Masakazu Kojima Tokyo.
June 21, 2007 Minimum Interference Channel Assignment in Multi-Radio Wireless Mesh Networks Anand Prabhu Subramanian, Himanshu Gupta.
RESOURCES, TRADE-OFFS, AND LIMITATIONS Group 5 8/27/2014.
1 ECE 517: Reinforcement Learning in Artificial Intelligence Lecture 17: TRTRL, Implementation Considerations, Apprenticeship Learning Dr. Itamar Arel.
Statistical Sampling-Based Parametric Analysis of Power Grids Dr. Peng Li Presented by Xueqian Zhao EE5970 Seminar.
Towards a Billion Routing Lookups per Second in Software  Author: Marko Zec, Luigi, Rizzo Miljenko Mikuc  Publisher: SIGCOMM Computer Communication Review,
Stochastic optimization of energy systems Cosmin Petra Argonne National Laboratory.
Shriram Sarvotham Dror Baron Richard Baraniuk ECE Department Rice University dsp.rice.edu/cs Sudocodes Fast measurement and reconstruction of sparse signals.
On the Use of Sparse Direct Solver in a Projection Method for Generalized Eigenvalue Problems Using Numerical Integration Takamitsu Watanabe and Yusaku.
Javad Lavaei Department of Electrical Engineering Columbia University Convex Relaxation for Polynomial Optimization: Application to Power Systems and Decentralized.
INFOMRS Charlotte1 Parallel Computation for SDPs Focusing on the Sparsity of Schur Complements Matrices Makoto Tokyo Tech Katsuki Fujisawa.
EE 685 presentation Optimization Flow Control, I: Basic Algorithm and Convergence By Steven Low and David Lapsley.
Chapter 3 Algorithms Complexity Analysis Search and Flow Decomposition Algorithms.
Introduction to Semidefinite Programs Masakazu Kojima Semidefinite Programming and Its Applications Institute for Mathematical Sciences National University.
Multi-area Nonlinear State Estimation using Distributed Semidefinite Programming Hao Zhu October 15, 2012 Acknowledgements: Prof. G.
1 Enclosing Ellipsoids of Semi-algebraic Sets and Error Bounds in Polynomial Optimization Makoto Yamashita Masakazu Kojima Tokyo Institute of Technology.
Linear Algebra Libraries: BLAS, LAPACK, ScaLAPACK, PLASMA, MAGMA
Nonsmooth Optimization for Optimal Power Flow over Transmission Networks GlobalSIP 2015 Authors: Y. Shi, H. D. Tuan, S. W. Su and H. H. M. Tam.
Data Structures and Algorithms in Parallel Computing Lecture 7.
Practical Message-passing Framework for Large-scale Combinatorial Optimization Inho Cho, Soya Park, Sejun Park, Dongsu Han, and Jinwoo Shin KAIST 2015.
1 Ellipsoid-type Confidential Bounds on Semi-algebraic Sets via SDP Relaxation Makoto Yamashita Masakazu Kojima Tokyo Institute of Technology.
On Optimizing Collective Communication UT/Texas Advanced Computing Center UT/Computer Science Avi Purkayastha Ernie Chan, Marcel Heinrich Robert van de.
1 Parallel Software for SemiDefinite Programming with Sparse Schur Complement Matrix Makoto Tokyo-Tech Katsuki Chuo University Mituhiro.
Improving Performance of The Interior Point Method by Preconditioning Project Proposal by: Ken Ryals For: AMSC Fall 2007-Spring 2008.
Linear Programming Chapter 9. Interior Point Methods  Three major variants  Affine scaling algorithm - easy concept, good performance  Potential.
Parametric Quadratic Optimization Oleksandr Romanko Joint work with Alireza Ghaffari Hadigheh and Tamás Terlaky McMaster University January 19, 2004.
Constraint Programming for the Diameter Constrained Minimum Spanning Tree Problem Thiago F. Noronha Celso C. Ribeiro Andréa C. Santos.
Massive Support Vector Regression (via Row and Column Chunking) David R. Musicant and O.L. Mangasarian NIPS 99 Workshop on Learning With Support Vectors.
National Taiwan University Department of Computer Science and Information Engineering An Approximation Algorithm for Haplotype Inference by Maximum Parsimony.
Hybrid Parallel Implementation of The DG Method Advanced Computing Department/ CAAM 03/03/2016 N. Chaabane, B. Riviere, H. Calandra, M. Sekachev, S. Hamlaoui.
1 An approach based on shortest path and connectivity consistency for sensor network localization problems Makoto Yamashita (Tokyo Institute of Technology)
Hybrid BDD and All-SAT Method for Model Checking
A computational loop k k Integration Newton Iteration
Georgina Hall Princeton, ORFE Joint work with Amir Ali Ahmadi
Nonnegative polynomials and applications to learning
Sudocodes Fast measurement and reconstruction of sparse signals
A computational loop k k Integration Newton Iteration
Presentation transcript:

1 Efficient Parallel Software for Large-Scale Semidefinite Programs Makoto Tokyo-Tech Katsuki Chuo University MSC Yokohama [2010/09/08]

2 Outline 1.SemiDefinite Programming 2.Conversion of stability condition for differential inclusions to an SDP 3.Primal-Dual Interior-Point Methods and its parallel implementation 4.Numerical Results

3 Many Applications of SDP  Control Theory Stability Condition for Differential Inclusions Discrete-Time Optimal Control Problem  Via SDP relaxation Polynomial Optimization Problem Sensor Network Problem Quadratic Assignment Problem  Quantum Chemistry/Information  Large SDP ⇒ Parallel Solver

4 Standard form of SDP

5 Stability condition for differential inclusions to standard SDP .  Does the solution remain in a bounded region?  i.e.,  Yes, if Boyd et al

6 .  To hold this inequality,  Bounding the condition number ⇒ SDP. Conversion to SDP

7 SDP from SCDI .  Feasible solution ⇒ Boundness of the solution  Some translation for standard SDP by e.g. YALMIP [J. L ö fberg].

8 Discrete-Time Optimal Control Problems  This Problem [Coleman et al] can be formulated as SDP via SparsePOP [Kim et al].

9 Primal-Dual Interior-Point Methods  Both Primal and Dual simultaneously in Polynomial-time  Many software are developed SDPA [Yamashita et al] SDPT3 [Toh et al] SeDuMi [Sturm et al] CSDP [Borcher et al]

10 Algorithmic Framework of Primal-Dual Interior-Point Methods Feasible Region of Optimal Solution Initial Point Target Point Central Path Search Direction Step Length to keep interior property The most computational time is consumed by the Search Direction

11 Bottlenecks in PDIPM and SDPARA  To obtain the direction, we solve 1.ELEMENTS 2.CHOLESKY  In SDPARA, parallel computation is applied to these two bottlenecks ProblemELEMENTSCHOLESKYTotal SCDI DTOC Xeon 5460,3.16GHz

12 Nonzero pattern of Schur complement matrix (B) Fully dense Schur complement matrix Fully dense Schur complement matrix Sparse Schur complement matrix Sparse Schur complement matrix SCDI DTOC

13 Exploitation of Sparsity in SDPA  We change the formula by row-wise  We keep this scheme on parallel computation F1F1 F2F2 F3F3

14 Row-wise distribution for dense Schur complement matrix  4 CPU is available  Each CPU computes only their assigned rows .  No communication between CPUs  Efficient memory management

15 Fomula-Cost Based distribution for sparse Schur complement Load on each CPU CPU1:195 CPU2:187 CPU3:189 CPU4:192 Average:190.75

16 Parallel Computation for CHOLESKY  We employ ScaLAPACK [Blackford et.al] ⇒ Dense MUMPS [Amestoy et.al] ⇒ Sparse  Different data storage enhance the parallel Cholesky factorization

17 Problems for Numerical Results  16 nodes Xeon X5460 (3.16GHz) 48GB memory

18 Computation time on SDP [SCDI1] Xeon X5460(3.16GHz) 48GB memory/node Total times ELEMENTS times CHOLESKY times ELEMENTS attains high scalability

19 Computation time on SDP [DTOC1] Xeon X5460(3.16GHz) 48GB memory/node Total 4.85 times ELEMENTS times CHOLESKY 4.34 times Parallel Sparse Cholesky is difficult ELEMENTS is still enhanced

20 Comparison with PCSDP [Ivanov et al] 1.SDPARA is faster than PCSDP 2.The scalability of SDPARA is higher 3.Only SDPARA can solve DTOC Time is second, O.M.:out of memory

21 Concluding Remarks & Future works 1.SDP has many applications including control theory 2.SDPARA solves Larse-scale SDPs effectively by parallel computation 3.Appropriate parallel computations are the key of SDPARA implementation  Improvement on Multi-Threading for sparse Schur complement matrix