A CONDENSATION-BASED LOW COMMUNICATION LINEAR SYSTEMS SOLVER UTILIZING CRAMER'S RULE Ken Habgood, Itamar Arel Department of Electrical Engineering & Computer.

Slides:



Advertisements
Similar presentations
Parallel Jacobi Algorithm Steven Dong Applied Mathematics.
Advertisements

Lecture 19: Parallel Algorithms
Chapter 2 Solutions of Systems of Linear Equations / Matrix Inversion
A NOVEL APPROACH TO SOLVING LARGE-SCALE LINEAR SYSTEMS Ken Habgood, Itamar Arel Department of Electrical Engineering & Computer Science GABRIEL CRAMER.
Dense Matrix Algorithms. Topic Overview Matrix-Vector Multiplication Matrix-Matrix Multiplication Solving a System of Linear Equations.
Algebraic, transcendental (i.e., involving trigonometric and exponential functions), ordinary differential equations, or partial differential equations...
MATH 685/ CSI 700/ OR 682 Lecture Notes
1 Parallel Algorithms II Topics: matrix and graph algorithms.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. Parallel Programming in C with MPI and OpenMP Michael J. Quinn.
Numerical Analysis – Linear Equations(I) Hanyang University Jong-Il Park.
SOLVING SYSTEMS OF LINEAR EQUATIONS. Overview A matrix consists of a rectangular array of elements represented by a single symbol (example: [A]). An individual.
CISE301_Topic3KFUPM1 SE301: Numerical Methods Topic 3: Solution of Systems of Linear Equations Lectures 12-17: KFUPM Read Chapter 9 of the textbook.
Numerical Algorithms Matrix multiplication
Part 3 Chapter 9 Gauss Elimination
Chapter 9 Gauss Elimination The Islamic University of Gaza
Linear Algebraic Equations
Computes the partial dot products for only the diagonal and upper triangle of the input matrix. The vector computed by this architecture is added to the.
1 Lecture 25: Parallel Algorithms II Topics: matrix, graph, and sort algorithms Tuesday presentations:  Each group: 10 minutes  Describe the problem,
ECIV 301 Programming & Graphics Numerical Methods for Engineers Lecture 17 Solution of Systems of Equations.
ECIV 520 Structural Analysis II Review of Matrix Algebra.
Math for CSLecture 31 LU Decomposition Pivoting Diagonalization Gram-Shcmidt Orthogonalization Lecture 3.
Ordinary least squares regression (OLS)
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. Parallel Programming in C with MPI and OpenMP Michael J. Quinn.
1 Systems of Linear Equations (Optional) Special Matrices.
1 EE 616 Computer Aided Analysis of Electronic Networks Lecture 4 Instructor: Dr. J. A. Starzyk, Professor School of EECS Ohio University Athens, OH,
ECIV 301 Programming & Graphics Numerical Methods for Engineers REVIEW II.
Monica Garika Chandana Guduru. METHODS TO SOLVE LINEAR SYSTEMS Direct methods Gaussian elimination method LU method for factorization Simplex method of.
Mujahed AlDhaifallah (Term 342) Read Chapter 9 of the textbook
Math for CSLecture 31 LU Decomposition Pivoting Diagonalization Gram-Shcmidt Orthogonalization Lecture 3.
Chapter 12 Gaussian Elimination (II) Speaker: Lung-Sheng Chien Reference book: David Kincaid, Numerical Analysis Reference lecture note: Wen-wei Lin, chapter.
CISE-301: Numerical Methods Topic 1: Introduction to Numerical Methods and Taylor Series Lectures 1-4: KFUPM.
Assignment Solving System of Linear Equations Using MPI Phạm Trần Vũ.
Chapter 8 Objectives Understanding matrix notation.
Scientific Computing Linear Systems – LU Factorization.
Advanced Computer Graphics Spring 2014 K. H. Ko School of Mechatronics Gwangju Institute of Science and Technology.
1 Secure Cooperative MIMO Communications Under Active Compromised Nodes Liang Hong, McKenzie McNeal III, Wei Chen College of Engineering, Technology, and.
An Effective Dynamic Scheduling Runtime and Tuning System for Heterogeneous Multi and Many-Core Desktop Platforms Authous: Al’ecio P. D. Binotto, Carlos.
CISE-301: Numerical Methods Topic 1: Introduction to Numerical Methods and Taylor Series Lectures 1-4: KFUPM CISE301_Topic1.
Introduction to Numerical Analysis I MATH/CMPSC 455 PA=LU.
Engineering Analysis ENG 3420 Fall 2009 Dan C. Marinescu Office: HEC 439 B Office hours: Tu-Th 11:00-12:00.
ΑΡΙΘΜΗΤΙΚΕΣ ΜΕΘΟΔΟΙ ΜΟΝΤΕΛΟΠΟΙΗΣΗΣ 4. Αριθμητική Επίλυση Συστημάτων Γραμμικών Εξισώσεων Gaussian elimination Gauss - Jordan 1.
CMPS 1371 Introduction to Computing for Engineers MATRICES.
Chemical Engineering Majors Author(s): Autar Kaw
Background Gaussian Elimination Fault Tolerance Single or multiple core failures: Single or multiple core additions: Simultaneous core failures and additions:
EE616 Dr. Janusz Starzyk Computer Aided Analysis of Electronic Circuits Innovations in numerical techniques had profound import on CAD: –Sparse matrix.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 1 Part 3 - Chapter 9 Linear Systems of Equations: Gauss Elimination.
On the Use of Sparse Direct Solver in a Projection Method for Generalized Eigenvalue Problems Using Numerical Integration Takamitsu Watanabe and Yusaku.
Direct Methods for Sparse Linear Systems Lecture 4 Alessandra Nardi Thanks to Prof. Jacob White, Suvranu De, Deepak Ramaswamy, Michal Rewienski, and Karen.
Chapter 9 Gauss Elimination The Islamic University of Gaza
Jungpyo Lee Plasma Science & Fusion Center(PSFC), MIT Parallelization for a Block-Tridiagonal System with MPI 2009 Spring Term Project.
Dense Linear Algebra Sathish Vadhiyar. Gaussian Elimination - Review Version 1 for each column i zero it out below the diagonal by adding multiples of.
ECE 530 – Analysis Techniques for Large-Scale Electrical Systems Prof. Hao Zhu Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign.
1 Chapter 7 Numerical Methods for the Solution of Systems of Equations.
1. Systems of Linear Equations and Matrices (8 Lectures) 1.1 Introduction to Systems of Linear Equations 1.2 Gaussian Elimination 1.3 Matrices and Matrix.
Linear Systems Dinesh A.
PARALLEL COMPUTATION FOR MATRIX MULTIPLICATION Presented By:Dima Ayash Kelwin Payares Tala Najem.
1 Numerical Methods Solution of Systems of Linear Equations.
Part 3 Chapter 9 Gauss Elimination
PIVOTING The pivot or pivot element is the element of a matrix, or an array, which is selected first by an algorithm (e.g. Gaussian elimination, simplex.
Linear Equations.
Lecture 22: Parallel Algorithms
Chapter 10: Solving Linear Systems of Equations
CHAPTER OBJECTIVES The primary objective of this chapter is to show how to compute the matrix inverse and to illustrate how it can be.
Engineering Analysis ENG 3420 Fall 2009
CSCE569 Parallel Computing
Major: All Engineering Majors Author(s): Autar Kaw
Numerical Algorithms Quiz questions
Dense Linear Algebra (Data Distributions)
Simultaneous Linear Equations Gaussian Elimination with Partial Pivoting
August 8, 2006 Danny Budik, Itamar Elhanany Machine Intelligence Lab
Presentation transcript:

A CONDENSATION-BASED LOW COMMUNICATION LINEAR SYSTEMS SOLVER UTILIZING CRAMER'S RULE Ken Habgood, Itamar Arel Department of Electrical Engineering & Computer Science The University of Tennessee GABRIEL CRAMER ( )

EECS Department / University of Tennessee EECS Department / University of Tennessee Outline 2  Motivation & problem statement  Algorithm review  Numerical accuracy & stability  Parallel Implementation  Communication Results Source:

EECS Department / University of Tennessee EECS Department / University of Tennessee Introduction  Mainstream approach: Gaussian Elimination  e.g. LU decomposition  Looking for a lower communication overhead, efficient parallel solver  Targeting an unpopular approach: Cramer’s Rule 3

EECS Department / University of Tennessee EECS Department / University of Tennessee LU Communication Pattern Source:  Communication for distributed LU decomposition L00 U00 U01U02 L10A11A12 L20A21A22  Three sequential steps 1. Top left computes and sends 2. Row and column leads compute and send 3. Remaining processors factorize their blocks  One-to-one communication  Idle time while leads processing 4

EECS Department / University of Tennessee EECS Department / University of Tennessee Outline 5  Motivation & problem statement  Algorithm review  Numerical accuracy & stability  Parallel Implementation  Communication Results Source:

EECS Department / University of Tennessee EECS Department / University of Tennessee Proposed Algorithm Flow 6

EECS Department / University of Tennessee EECS Department / University of Tennessee Matrix “Mirroring”  Mirroring example  Applying Chio’s condensation yields: 7

EECS Department / University of Tennessee EECS Department / University of Tennessee Outline 8  Motivation & problem statement  Algorithm review  Numerical accuracy & stability  Parallel Implementation  Communication Results Source:

EECS Department / University of Tennessee EECS Department / University of Tennessee Accuracy and Numerical Stability  Backward error estimation  Theoretical estimate of rounding error  E matrix depends on two items The largest element in A or b The growth factor of the algorithm  Same growth factor as LU-decomposition with partial pivoting 9

EECS Department / University of Tennessee EECS Department / University of Tennessee Forward Error Comparisons Matrix Size κ(A) Max MatlabMax GSL Avg Matlab Avg GSL 1000 x E E E E x E E E E x E E E E x E E E E x E E E E x E E E E x E E E E x E E E E-11 10

EECS Department / University of Tennessee EECS Department / University of Tennessee Forward Error - Residual Matrix Sizeκ(A) Max Residual Avg Residual 1000 x E E x E E x E E x E E x E E x E E x E E x E E-09 11

EECS Department / University of Tennessee EECS Department / University of Tennessee MATLAB Matrix Gallery Special Matrix Avg MatlabResidual Matlab Residual clement — Tridiagonal matrix with zero diagonal entries1.40E E E+144 lehmer — Symmetric positive definite matrix2.49E E E-06 circul — Circulant matrix3.23E E E-09 chebspec — Chebyshev spectral differentiation matrix9.12E E+042.0E-01 lesp — Tridiagonal matrix with real, sensitive eigenvalues9.56E E E-10 minij — Symmetric positive definite matrix5.14E E E-06 orthog — Orthogonal and nearly orthogonal matrices1.03E E E-08 randjorth — Random J-orthogonal matrix1.55E E E-04 12

EECS Department / University of Tennessee EECS Department / University of Tennessee Outline 13  Motivation & problem statement  Algorithm review  Numerical accuracy & stability  Parallel Implementation  Communication Results Source:

EECS Department / University of Tennessee EECS Department / University of Tennessee Serial Performance Results support the theoretical ~2.5x complexity ratio 14

EECS Department / University of Tennessee EECS Department / University of Tennessee Algorithm Processing Flow 15

EECS Department / University of Tennessee EECS Department / University of Tennessee Overview of Parallel Implementation 16

EECS Department / University of Tennessee EECS Department / University of Tennessee Parallel Implementation (cont’) 17

EECS Department / University of Tennessee EECS Department / University of Tennessee  Two phases of parallel communication  Parallel Chio’s  Gather Columns  Overall Bandwidth Communication Complexity N: Original matrix size, P: number of processors, F: gather columns size 18

EECS Department / University of Tennessee EECS Department / University of Tennessee Communication Overhead 19

EECS Department / University of Tennessee EECS Department / University of Tennessee  Point at which Communication “dead time” matches computational workload Where’s the Breakeven Point?  Assuming d C =.05 and N = 1000, the breakeven processors point would be P ~142 20

EECS Department / University of Tennessee EECS Department / University of Tennessee Closing Thoughts …  Proposed O(N 3 ) Cramer’s Rule method  Significantly lower communications overhead  Many more “broadcasts” than “unicasts”  Comm. function of problem size not processors  Next steps …  Optimize parallel implementation  Spare matrix version 21

EECS Department / University of Tennessee EECS Department / University of Tennessee Thank you 22