Barry Smith Algebraic Solvers in FASTMath Part 1 – Introduction to several packages unable to present today Part 2 – Portable Extensible Toolkit for Scientific.

Slides:



Advertisements
Similar presentations
Sandia is a multiprogram laboratory operated by Sandia Corporation, a Lockheed Martin Company, for the United States Department of Energy under contract.
Advertisements

Lecture 8 Hybrid Solvers based on Domain Decomposition Xiaoye Sherry Li Lawrence Berkeley National Laboratory, USA crd-legacy.lbl.gov/~xiaoye/G2S3/
Advanced Computational Software Scientific Libraries: Part 2 Blue Waters Undergraduate Petascale Education Program May 29 – June
MATLAB Presented By: Nathalie Tacconi Presented By: Nathalie Tacconi Originally Prepared By: Sheridan Saint-Michel Originally Prepared By: Sheridan Saint-Michel.
Numerical Parallel Algorithms for Large-Scale Nanoelectronics Simulations using NESSIE Eric Polizzi, Ahmed Sameh Department of Computer Sciences, Purdue.
Multilevel Incomplete Factorizations for Non-Linear FE problems in Geomechanics DMMMSA – University of Padova Department of Mathematical Methods and Models.
MA5233: Computational Mathematics
Parallel Decomposition-based Contact Response Fehmi Cirak California Institute of Technology.
Sparse Matrix Methods Day 1: Overview Day 2: Direct methods
1 A Matrix Free Newton /Krylov Method For Coupling Complex Multi-Physics Subsystems Yunlin Xu School of Nuclear Engineering Purdue University October 23,
The Landscape of Ax=b Solvers Direct A = LU Iterative y’ = Ay Non- symmetric Symmetric positive definite More RobustLess Storage (if sparse) More Robust.
Symmetric Definite Generalized Eigenproblem
PETSc Portable, Extensible Toolkit for Scientific computing.
Direct and iterative sparse linear solvers applied to groundwater flow simulations Matrix Analysis and Applications October 2007.
1 With contributions from: I. Demeshko (SNL), S. Price (LANL) and M. Hoffman (LANL) Saturday, March 14, 2015 SIAM Conference on Computational Science &
Sandia is a multiprogram laboratory operated by Sandia Corporation, a Lockheed Martin Company, for the United States Department of Energy’s National Nuclear.
MUMPS A Multifrontal Massively Parallel Solver IMPLEMENTATION Distributed multifrontal.
Iterative and direct linear solvers in fully implicit magnetic reconnection simulations with inexact Newton methods Xuefei (Rebecca) Yuan 1, Xiaoye S.
Massively Parallel Magnetohydrodynamics on the Cray XT3 Joshua Breslau and Jin Chen Princeton Plasma Physics Laboratory Cray XT3 Technical Workshop Nashville,
Processing of a CAD/CAE Jobs in grid environment using Elmer Electronics Group, Physics Department, Faculty of Science, Ain Shams University, Mohamed Hussein.
SciDAC Software Introduction Osni Marques Lawrence Berkeley National Laboratory DOE Workshop for Industry Software Developers March 31,
S.S. Yang and J.K. Lee FEMLAB and its applications POSTEC H Plasma Application Modeling Lab. Oct. 25, 2005.
Page 1 Trilinos Software Engineering Technologies and Integration Numerical Algorithm Interoperability and Vertical Integration –Abstract Numerical Algorithms.
1 Using the PETSc Parallel Software library in Developing MPP Software for Calculating Exact Cumulative Reaction Probabilities for Large Systems (M. Minkoff.
Large-Scale Stability Analysis Algorithms Andy Salinger, Roger Pawlowski, Ed Wilkes Louis Romero, Rich Lehoucq, John Shadid Sandia National Labs Albuquerque,
Overview of MHD and extended MHD simulations of fusion plasmas Guo-Yong Fu Princeton Plasma Physics Laboratory Princeton, New Jersey, USA Workshop on ITER.
Amesos Sparse Direct Solver Package Ken Stanley, Rob Hoekstra, Marzio Sala, Tim Davis, Mike Heroux Trilinos Users Group Albuquerque 3 Nov 2004.
ANS 1998 Winter Meeting DOE 2000 Numerics Capabilities 1 Barry Smith Argonne National Laboratory DOE 2000 Numerics Capability
PDE2D, A General-Purpose PDE Solver Granville Sewell Mathematics Dept. University of Texas El Paso.
ParCFD Parallel computation of pollutant dispersion in industrial sites Julien Montagnier Marc Buffat David Guibert.
ML: Multilevel Preconditioning Package Trilinos User’s Group Meeting Wednesday, October 15, 2003 Jonathan Hu Sandia is a multiprogram laboratory operated.
Automatic Differentiation: Introduction Automatic differentiation (AD) is a technology for transforming a subprogram that computes some function into a.
Efficient Integration of Large Stiff Systems of ODEs Using Exponential Integrators M. Tokman, M. Tokman, University of California, Merced 2 hrs 1.5 hrs.
1 1  Capabilities: Dynamic load balancing and static data partitioning -Geometric, graph-based, hypergraph-based -Interfaces to ParMETIS, PT-Scotch, PaToH.
1 SciDAC TOPS PETSc Work SciDAC TOPS Developers Satish Balay Chris Buschelman Matt Knepley Barry Smith.
Computational Aspects of Multi-scale Modeling Ahmed Sameh, Ananth Grama Computing Research Institute Purdue University.
Danny Dunlavy, Andy Salinger Sandia National Laboratories Albuquerque, New Mexico, USA SIAM Parallel Processing February 23, 2006 SAND C Sandia.
1 1 What does Performance Across the Software Stack mean?  High level view: Providing performance for physics simulations meaningful to applications 
Amesos Interfaces to sparse direct solvers October 15, :30 – 9:30 a.m. Ken Stanley.
1 1  Capabilities: Scalable algebraic solvers for PDEs Freely available and supported research code Usable from C, C++, Fortran 77/90, Python, MATLAB.
1 1  Capabilities: Building blocks for block-structured AMR codes for solving time-dependent PDE’s Functionality for [1…6]D, mixed-dimension building.
Parallel and Distributed Computing Research at the Computing Research Institute Ananth Grama Computing Research Institute and Department of Computer Sciences.
Sandia is a multiprogram laboratory operated by Sandia Corporation, a Lockheed Martin Company, for the United States Department of Energy’s National Nuclear.
Update on Sandia Albany/FELIX First-Order Stokes (FELIX-FO) Solver Irina K. Tezaur Sandia National Laboratories In collaboration with Mauro Perego, Andy.
Implementing Hypre- AMG in NIMROD via PETSc S. Vadlamani- Tech X S. Kruger- Tech X T. Manteuffel- CU APPM S. McCormick- CU APPM Funding: DE-FG02-07ER84730.
October 2008 Integrated Predictive Simulation System for Earthquake and Tsunami Disaster CREST/Japan Science and Technology Agency (JST)
Cracow Grid Workshop, November 5-6, 2001 Concepts for implementing adaptive finite element codes for grid computing Krzysztof Banaś, Joanna Płażek Cracow.
Connections to Other Packages The Cactus Team Albert Einstein Institute
1 1  Capabilities: Serial (C), shared-memory (OpenMP or Pthreads), distributed-memory (hybrid MPI+ OpenM + CUDA). All have Fortran interface. Sparse LU.
On the Use of Finite Difference Matrix-Vector Products in Newton-Krylov Solvers for Implicit Climate Dynamics with Spectral Elements ImpactObjectives 
1 1  Capabilities: Serial (thread-safe), shared-memory (SuperLU_MT, OpenMP or Pthreads), distributed-memory (SuperLU_DIST, hybrid MPI+ OpenM + CUDA).
Discussion Slides S.E. Kruger, J.D. Callen, J. Carlson, C.C. Hegna, E.D. Held, T. Jenkins, J. Ramos, D.D. Schnack, C.R. Sovinec, D.A. Spong ORNL SWIM Meet.
Report from LBNL TOPS Meeting TOPS/ – 2Investigators  Staff Members:  Parry Husbands  Sherry Li  Osni Marques  Esmond G. Ng 
ML: A Multilevel Preconditioning Package Copper Mountain Conference on Iterative Methods March 29-April 2, 2004 Jonathan Hu Ray Tuminaro Marzio Sala Sandia.
Algebraic Solvers in FASTMath Argonne Training Program on Extreme-Scale Computing August 2015.
Consider Preconditioning – Basic Principles Basic Idea: is to use Krylov subspace method (CG, GMRES, MINRES …) on a modified system such as The matrix.
C OMPUTATIONAL R ESEARCH D IVISION 1 Defining Software Requirements for Scientific Computing Phillip Colella Applied Numerical Algorithms Group Lawrence.
Center for Extended MHD Modeling (PI: S. Jardin, PPPL) –Two extensively developed fully 3-D nonlinear MHD codes, NIMROD and M3D formed the basis for further.
On the Path to Trinity - Experiences Bringing Codes to the Next Generation ASC Platform Courtenay T. Vaughan and Simon D. Hammond Sandia National Laboratories.
Conjugate gradient iteration One matrix-vector multiplication per iteration Two vector dot products per iteration Four n-vectors of working storage x 0.
Unstructured Meshing Tools for Fusion Plasma Simulations
Challenges in Electromagnetic Modeling Scalable Solvers
Shrirang Abhyankar IEEE PES HPC Working Group Meeting
Parallel Unstructured Mesh Infrastructure
Meros: Software for Block Preconditioning the Navier-Stokes Equations
Trilinos Software Engineering Technologies and Integration
Salient application properties Expectations TOPS has of users
Embedded Nonlinear Analysis Tools Capability Area
Ph.D. Thesis Numerical Solution of PDEs and Their Object-oriented Parallel Implementations Xing Cai October 26, 1998.
Presentation transcript:

Barry Smith Algebraic Solvers in FASTMath Part 1 – Introduction to several packages unable to present today Part 2 – Portable Extensible Toolkit for Scientific computing

2 2  Capabilities: Serial (C), shared-memory (Pthreads or OpenMP, distributed-memory in MPI; with Fortran interface. Sparse LU decomposition, triangular solution with multiple right-hand sides Incomplete LU (ILU) preconditioner (serial SuperLU) Sparsity-preserving ordering  Minimum degree ordering applied to A T A or A T +A [MMD, Liu `85]  ‘Nested-dissection’ applied to A T A or A T +A [(Par)Metis, (PT)-Scotch] User-controllable pivoting: partial pivoting, threshold pivoting, static pivoting Condition number estimation Iterative refinement Componentwise error bounds  Download:  Further information: contact: Sherry Li, Developers: Sherrry Li, Jim Demmel, John Gilbert, Laura Grigori, Meiyue Shao, Ichitaro Yamazaki SuperLU – supernodal sparse LU linear solver

3 3  A new fusion reactor ITER is being constructed in France, which is used to study how to harness fusion, creating clean energy using nearly inexhaustible hydrogen as the fuel. If successful, ITER will produce 10 times as much energy than it uses — but that success hinges on accurately designing the device.  One major simulation goal is to predict microscopic MHD instabilities of burning plasma in ITER. This involves solving extended and nonlinear Magnetohydrodynamics equations. Two fluid-based codes have been developed, using different finite element discretization methods: M3D-C1 (Steve Jardin et al. PPPL), and NIMROD (Carl Sovinec, Univ. Wisconsin, and others)  These are truly multiphysics simulations involving several coupled PDEs: continuity, Maxwell, momentum, electron energy, ion energy, etc. They are also of multiscale with extremely wide range of time and spatial scales, and high anisotropy. Therefore, the discretized linear systems are indefinite and very ill-conditioned.  SuperLU_DIST has been used in the production codes. For example, the NIMROD simulation time was improved from 3 weeks to 3 days, compared to using some preconditioned iterative solvers. In many iterations of nonlinear solver, the largest linear systems are of dimension 500K, complex and nonsymmetric. FASTMath SciDAC Institute SuperLU: Application highlight in fusion energy research  R Z

4 4  There were over 24,000 downloads of SuperLU in FY It was the third most-downloaded software at LBNL Over the years, it has consistently been the most- downloaded software in the Computing Sciences organization at LBNL  It has been used in many high-end simulation codes, such as ASCEM (reactive flow and transport), M3D-C 1 /NIMROD (fusion tokamak design), Omega3P (accelerator cavity design), OpenSees (earthquake engineering), PHOENIX (modeling of stellar and planetary atmospheres).  It is adopted in many commercial mathematical libraries and simulation software, including AMD (circuit simulation), Boeing (aircraft design), Chevron (geology), Cray's LibSci, FEMLAB, HP's MathLib, IMSL, NAG, OptimaNumerics, Python (SciPy), Walt Disney Feature Animation. SuperLU: General usage and impact

5 5  Capabilities: Algebraic multigrid (AMG) algorithms  Smoothed aggregation AMG (Poisson, elasticity)  Petrov Galerkin AMG (convection-diffusion)  H(curl) AMG (electro-magnetic) Coarsening algorithms  Uncoupled, MIS, METIS, ParMETIS Smoothers  Jacobi, (symmetric) Gauss-Seidel, polynomial (Chebyshev), ILU, block variants, user-provided  Download: trilinos.sandia.gov/download  Further information: trilinos.sandia.gov/packages/ml Trilinos / ML : Aggregation-based Algebraic Multigrid

6 6  Drekar simulation software low mach incompressible Navier Stokes Magneto-hydrodynamics  Drekar makes extensive use of Trilinos framework: Panzer, Phalanx, Intrepid, ML, Ifpack, Amesos, Zoltan, AztecOO  Simulations have been run on up to 512K Bluegene/Q cores and 128K cores of Titan. ML preconditioning in large-scale MHD simulations

7 7  Capabilities: Newton-Based Nonlinear Solver  Linked to Trilinos linear solvers for scalability  Matrix-Free option Globalizations for improved robustness  Line Searches, Trust Region, Homotopy methods Customizable: C++ abstractions at every level Extended by LOCA package  Parameter continuation, Stability analysis, Bifurcation tracking  Download: Part of Trilinos (trilinos.sandia.gov)  Further information: Andy Salinger Trilinos/NOX Nonlinear Solver

8 8  Ice Sheets modeled by nonlinear Stokes’s equation Initial solve is fragile: Full Newton fails Homotopy continuation on regularization parameter “  ” saves the day Trilinos/NOX: Robustness for Ice Sheet Simulation: PISCEES SciDAC Application project (BER-ASCR) Greenland Ice Sheet Surface Velocities (constant friction model)  =  =  =  =10 -10

9 9 FASTMath SciDAC Institute NOX and ML are part of larger Trilinos solver stack: Linear solvers, Equations solvers, Analysis tools Analysis Tools UQ (sampling) Parameter Studies Optimization Analysis Tools (black-box) LinearSolvers Direct Solvers Linear Algebra Algebraic Preconditioners Iterative Solvers EquationSolvers UQ Solver Nonlinear Solver Time Integration Optimization Continuation Sensitivity Analysis Stability Analysis Analysis Tools (embedded) Linear Solver Interface Nonlinear Model Interface Your Model Here Solved Problem Interface NOX Multilevel Preconditioners ML

10  Capabilities: Compute a few eigenpairs of a Hermitian and non-Hermitian matrix Both standard and generalized eigenvalues Extremal and interior eigenvalues Reverse communication allows easy integration with application MPI/BLACS communication  Download:  Further information: beyond PARPACK EIGPEN (based on penalty trace minimization for computing many eigenpairs) Parallel multiple shift-invert interface for computing many eigenpairs PARPACK

11  MFDn (Many-body Fermion Dynamics for nuclear physics) is the most robust and efficient configuration interaction code for studying nuclear structure  Solve the nuclear quantum many-body problem  PARPACK allows low excitation states to be extracted from a low dimensional Krylov subspace FASTMath SciDAC Institute MFDn: Nuclear Configuration Interaction Calculation

12  A list of drivers provided in $ARPACKTOPDIR/PARPACK/EXAMPLES  Reverse communication interface  Hybrid MPI/OpenMP implementation  MATLAB interface available (eigs) FASTMath SciDAC Institute PARPACK Usage 10continue call pdsaupd(comm, ido,….) if (ido.eq. 1.or. -1) then matvec(…,workd(ipntr(1)), workd(ipntr(2)…. endif goto 10