CADD: Component-Averaged Domain Decomposition Dan Gordon Computer Science University of Haifa Rachel Gordon Aerospace Engg. Technion January 13, 20081.

Slides:



Advertisements
Similar presentations
Parallel Jacobi Algorithm Steven Dong Applied Mathematics.
Advertisements

Lecture 19: Parallel Algorithms
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
03/29/2006, City Univ1 Iterative Methods with Inexact Preconditioners and Applications to Saddle-point Systems & Electromagnetic Maxwell Systems Jun Zou.
CS 290H 7 November Introduction to multigrid methods
Applied Linear Algebra - in honor of Hans SchneiderMay 25, 2010 A Look-Back Technique of Restart for the GMRES(m) Method Akira IMAKURA † Tomohiro SOGABE.
July 2011 High-order schemes for high-frequency Helmholtz equation 1 Parallel solution of high-frequency Helmholtz equation using high- order finite difference.
Segmentation and Fitting Using Probabilistic Methods
Linear Systems of Equations
Rayan Alsemmeri Amseena Mansoor. LINEAR SYSTEMS Jacobi method is used to solve linear systems of the form Ax=b, where A is the square and invertible.
OpenFOAM on a GPU-based Heterogeneous Cluster
Modern iterative methods For basic iterative methods, converge linearly Modern iterative methods, converge faster –Krylov subspace method Steepest descent.
Prénom Nom Document Analysis: Linear Discrimination Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Convergent and Correct Message Passing Algorithms Nicholas Ruozzi and Sekhar Tatikonda Yale University TexPoint fonts used in EMF. Read the TexPoint manual.
Sparse Matrix Algorithms CS 524 – High-Performance Computing.
1 Systems of Linear Equations Iterative Methods. 2 B. Iterative Methods 1.Jacobi method and Gauss Seidel 2.Relaxation method for iterative methods.
Avoiding Communication in Sparse Iterative Solvers Erin Carson Nick Knight CS294, Fall 2011.
Message Passing Algorithms for Optimization
Sparse Matrix Methods Day 1: Overview Day 2: Direct methods
The Landscape of Ax=b Solvers Direct A = LU Iterative y’ = Ay Non- symmetric Symmetric positive definite More RobustLess Storage (if sparse) More Robust.
High Performance Computing 1 Parallelization Strategies and Load Balancing Some material borrowed from lectures of J. Demmel, UC Berkeley.
CS240A: Conjugate Gradients and the Model Problem.
Monica Garika Chandana Guduru. METHODS TO SOLVE LINEAR SYSTEMS Direct methods Gaussian elimination method LU method for factorization Simplex method of.
Chapter 13 Finite Difference Methods: Outline Solving ordinary and partial differential equations Finite difference methods (FDM) vs Finite Element Methods.
Chapter 10 Real Inner Products and Least-Square (cont.)
Photo-realistic Rendering and Global Illumination in Computer Graphics Spring 2012 Stochastic Radiosity K. H. Ko School of Mechatronics Gwangju Institute.
July 1, 2010Parallel solution of the Helmholtz equation1 Parallel solution of the Helmholtz equation with large wave numbers Dan Gordon Computer Science.
Jason P. Stockmann 1 and R. Todd Constable 1,2 Yale University, Department of Biomedical Engineering 1, Department of Diagnostic Radiology 2, New Haven,
Kalman filtering techniques for parameter estimation Jared Barber Department of Mathematics, University of Pittsburgh Work with Ivan Yotov and Mark Tronzo.
Deep Learning – Fall 2013 Instructor: Bhiksha Raj Paper: T. D. Sanger, “Optimal Unsupervised Learning in a Single-Layer Linear Feedforward Neural Network”,
1 / 20 Arkadij Zakrevskij United Institute of Informatics Problems of NAS of Belarus A NEW ALGORITHM TO SOLVE OVERDEFINED SYSTEMS OF LINEAR LOGICAL EQUATIONS.
EE369C Final Project: Accelerated Flip Angle Sequences Jan 9, 2012 Jason Su.
Module 1: Statistical Issues in Micro simulation Paul Sousa.
CFD Lab - Department of Engineering - University of Liverpool Ken Badcock & Mark Woodgate Department of Engineering University of Liverpool Liverpool L69.
Sept COMP60611 Fundamentals of Concurrency Lab Exercise 2 Notes Notes on the finite difference performance model example – for the lab… Graham Riley,
The swiss-carpet preconditioner: a simple parallel preconditioner of Dirichlet-Neumann type A. Quarteroni (Lausanne and Milan) M. Sala (Lausanne) A. Valli.
Lecture 7 - Systems of Equations CVEN 302 June 17, 2002.
Systems of Equations Summary. Independent The equations of a linear system are independent if none of the equations can be derived algebraically from.
Parallelizing Gauss-Seidel Solver/Pre-conditioner Aim: To parallelize a Gauss-Seidel Solver, which can be used as a pre-conditioner for the finite element.
Case Study in Computational Science & Engineering - Lecture 5 1 Iterative Solution of Linear Systems Jacobi Method while not converged do { }
Domain Decomposition in High-Level Parallelizaton of PDE codes Xing Cai University of Oslo.
October 2008 Integrated Predictive Simulation System for Earthquake and Tsunami Disaster CREST/Japan Science and Technology Agency (JST)
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Direct Methods for Linear Systems Lecture 3 Alessandra Nardi Thanks to Prof. Jacob White, Suvranu De, Deepak Ramaswamy, Michal Rewienski, and Karen Veroy.
Parallel and Distributed Simulation Time Parallel Simulation.
Introduction to Scientific Computing II Overview Michael Bader.
Lecture 21 MA471 Fall 03. Recall Jacobi Smoothing We recall that the relaxed Jacobi scheme: Smooths out the highest frequency modes fastest.
23/5/20051 ICCS congres, Atlanta, USA May 23, 2005 The Deflation Accelerated Schwarz Method for CFD C. Vuik Delft University of Technology
MULTISCALE COMPUTATIONAL METHODS Achi Brandt The Weizmann Institute of Science UCLA
Lecture 6 - Single Variable Problems & Systems of Equations CVEN 302 June 14, 2002.
An Object-Oriented Software Framework for Building Parallel Navier-Stokes Solvers Xing Cai Hans Petter Langtangen Otto Munthe University of Oslo.
Monte Carlo Linear Algebra Techniques and Their Parallelization Ashok Srinivasan Computer Science Florida State University
Programming Massively Parallel Graphics Multiprocessors using CUDA Final Project Amirhassan Asgari Kamiabad
Adaptive grid refinement. Adaptivity in Diffpack Error estimatorError estimator Adaptive refinementAdaptive refinement A hierarchy of unstructured gridsA.
ICCV 2007 National Laboratory of Pattern Recognition Institute of Automation Chinese Academy of Sciences Half Quadratic Analysis for Mean Shift: with Extension.
Multipole-Based Preconditioners for Sparse Linear Systems. Ananth Grama Purdue University. Supported by the National Science Foundation.
Parallel Computing Activities at the Group of Scientific Software Xing Cai Department of Informatics University of Oslo.
An Introduction to Computational Fluids Dynamics Prapared by: Chudasama Gulambhai H ( ) Azhar Damani ( ) Dave Aman ( )
Conjugate gradient iteration One matrix-vector multiplication per iteration Two vector dot products per iteration Four n-vectors of working storage x 0.
Monte Carlo Linear Algebra Techniques and Their Parallelization Ashok Srinivasan Computer Science Florida State University
Relaxation Methods in the Solution of Partial Differential Equations
Xing Cai University of Oslo
Solving Systems of Linear Equations: Iterative Methods
Lecture 22: Parallel Algorithms
Craig Schroeder October 26, 2004
Supported by the National Science Foundation.
Introduction to Scientific Computing II
A Software Framework for Easy Parallelization of PDE Solvers
Parallelizing Unstructured FEM Computation
Ph.D. Thesis Numerical Solution of PDEs and Their Object-oriented Parallel Implementations Xing Cai October 26, 1998.
Presentation transcript:

CADD: Component-Averaged Domain Decomposition Dan Gordon Computer Science University of Haifa Rachel Gordon Aerospace Engg. Technion January 13, Component-Averaged Domain Decomposition

January 13, 2008 Component-Averaged Domain Decomposition 2 Outline of Talk  CADD – Algebraic explanation  CADD – DD explanation  The Kaczmarz algorithm (KACZ)  KACZ  CARP (a CADD method)  Applications of CARP  CARP-CG: CG acceleration of CARP  Sample results

January 13, 2008 Component-Averaged Domain Decomposition 3 CADD – Algebraic Explanation – 1  The equations are divided into blocks (not necessarily disjoint)  Initial estimate: vector x=(x 1,…,x n )  Suppose x 1 is a variable (component of x) that appears in 3 blocks  x 1 is “cloned” as y 1, z 1, t 1 in the different blocks.  Perform one (or more) iterative operation(s) on each block (independently, in parallel)

January 13, 2008 Component-Averaged Domain Decomposition 4 CADD – Algebraic Explanation – 2  The internal iterations in each block produce 3 new values for the clones of x 1 : y 1 ’, z 1 ’, t 1 ’  The next iterative value of x 1 is x 1 ’ = (y 1 ’ + z 1 ’ + t 1 ’)/3  The next iterate is x’ = (x 1 ’,..., x n ’)  Repeat iterations as needed for convergence

January 13, 2008 Component-Averaged Domain Decomposition 5 CADD – non-overlapping domains xx y domain A domain B external grid point of A clone of x 1

January 13, 2008 Component-Averaged Domain Decomposition 6 CADD – overlapping domains xx y 01 1 domain A domain B external grid point of A clone of x 1 A∩BA∩BA∩BA∩B

January 13, 2008 Component-Averaged Domain Decomposition 7 CADD – Parallel Implementation  Every processor is in charge of one domain  Mode: BLOCK-PARALLEL –All processors operate in parallel (each on its domain) –Processors exchange clone values –All values are updated –New values = average of clones

January 13, 2008 Component-Averaged Domain Decomposition 8 CADD Vs. Standard DD Methods  Difference is in handling external grid points:  CADD: A clone of the external grid point is modified by the domain's equations and contributes to new value of grid point.  Standard DD: Value of external grid point is fixed (contributes to RHS)

January 13, 2008 Component-Averaged Domain Decomposition 9 KACZ: The Kaczmarz algorithm  Iterative method, due to Kaczmarz (1937). Rediscovered for CT as ART  Basic idea: Consider the hyperplane defined by each equation  Start from an arbitrary initial point  Successively project current point onto the next hyperplane, in cyclic order

January 13, 2008 Component-Averaged Domain Decomposition 10 KACZ: Geometric Description eq. 1 eq. 2 eq. 3 initial point

January 13, 2008 Component-Averaged Domain Decomposition 11 KACZ with Relaxation Parameter  KACZ can be used with a relaxation parameter λ  λ=1: project exactly on the hyperplane  λ<1: project in front of hyperplane  λ>1: project beyond the hyperplane  Cyclic relaxation: Eq. i is assigned a relaxation parameter λ i

January 13, 2008 Component-Averaged Domain Decomposition 12 Convergence Properties of KACZ  KACZ with relaxation (0< λ <2) converges for consistent systems: –Herman, Lent & Lutz, 1978 –Trummer, 1981  For inconsistent systems, KACZ converges cyclically: –Tanabe, 1971 –Eggermont, Herman & Lent, 1981 (for cyclic relaxation parameters).

January 13, 2008 Component-Averaged Domain Decomposition 13 Another view of KACZ  Given the systemAx = b  Consider the "normal equations" systemAA T y = b, x=A T y  Well-known: KACZ is simply SOR applied to the normal equations  The relaxation parameter of KACZ is the usual relax. par. of SOR

January 13, 2008 Component-Averaged Domain Decomposition 14 CARP: Component-Averaged Row Projections  CARP is a CADD method with KACZ iterations in each domain:  The system Ax=b is divided into blocks B 1,…,B n (need not be disjoint)  Each processor is assigned one (or more) block  All the blocks are processed in parallel  Results from blocks are “merged” to form the next iterate  Merging is done by CADD averaging

January 13, 2008 Component-Averaged Domain Decomposition 15 Overview of CARP domain A domain B KACZiterationsKACZiterations averaging cloning KACZ in superspace (with cyclic relaxation)

January 13, 2008 Component-Averaged Domain Decomposition 16 Convergence of CARP  Averaging Lemma: the component- averaging and cloning operations of CARP are equivalent to KACZ row- projections in a certain superspace (with λ = 1)   CARP is equivalent to KACZ in the superspace, with cyclic relaxation parameters – known to converge

January 13, 2008 Component-Averaged Domain Decomposition 17 "Historical" Note  The term "component averaging" was first used by Censor, Gordon & Gordon (2001) for CAV and BICAV, used for image reconstruction in CT  These are Cimmino-type algorithms, with weights related to the sparsity of the system matrix  CADD, CARP use this concept in a different sense

January 13, 2008 Component-Averaged Domain Decomposition 18 CARP Application: Solution of stiff linear systems from PDEs  Elliptic PDEs w/large convection term result in stiff linear systems (large off-diagonal elements)  CARP is very robust on these systems, as compared to leading solver/preconditioner combinations  Downside: Not always efficient

January 13, 2008 Component-Averaged Domain Decomposition 19 CARP Application: Electron Tomography (joint work with J.-J. Fernández)  3D reconstructions: Each processor is assigned a block of consecutive slices. Data is in overlapping blobs.  The blocks are processed in parallel.  The values of shared variables are transmitted between the processors which share them, averaged, and redestributed.

January 13, 2008 Component-Averaged Domain Decomposition 20 CARP-CG: CG acceleration of CARP  CARP is KACZ in some superspace (with cyclic relaxation parameter)  Björck & Elfving (BIT 79): developed CGMN, which is a (sequential) CG- acceleration of KACZ (double sweep, fixed relax. parameter)  We extended this result to allow cyclic relaxation parameters  Result: CARP-CG

January 13, 2008 Component-Averaged Domain Decomposition 21 CARP-CG: Properties  Same robustness as CARP  Very significant improvement in performance on stiff linear systems derived from elliptic PDEs  Very competitive runtime compared to leading solver/preconditioner combinations on systems derived from convection-dominated PDEs  Improved performance in ET

CARP-CG on PDEs  CG acceleration of projection methods was done before, but with block projections, requiring multi- coloring. CARP-CG avoids this.  Tests were run on 9 convection- dominated PDEs, comparing CARP- CG with restarted GMRES, CGS and Bi-CGSTAB, with and without various preconditioners. Also tested: CGNR.  Domain: unit cube – 80x80x80. January 13, 2008 Component-Averaged Domain Decomposition 22

January 13, 2008 Component-Averaged Domain Decomposition ,000 equations

January 13, 2008 Component-Averaged Domain Decomposition ,000 equations

January 13, 2008 Component-Averaged Domain Decomposition ,000 equations

January 13, 2008 Component-Averaged Domain Decomposition ,000 equations

January 13, 2008 Component-Averaged Domain Decomposition ,000 equations

January 13, 2008 Component-Averaged Domain Decomposition 28 CARP-CG: ET results

January 13, 2008 Component-Averaged Domain Decomposition 29 Future research on CADD  Different internal solvers: 1. Guaranteed convergence? 2. Symmetrizability?  Combine CADD with multigrid  CFD applications  Additional biomedical applications  CADD with preconditioners  Compare CADD with other DD methods, such as additive Schwarz

January 13, 2008 Component-Averaged Domain Decomposition 30 CARP Publications  CARP: SIAM J. Sci. Comput. 27 (2005)  Electron tomography: J. Par. & Dist. Comput. (2007), in press.  CARP-CG: Submitted for publication.