1 1  Capabilities: Scalable algebraic solvers for PDEs Freely available and supported research code Usable from C, C++, Fortran 77/90, Python, MATLAB.

Slides:



Advertisements
Similar presentations
A Workflow Engine with Multi-Level Parallelism Supports Qifeng Huang and Yan Huang School of Computer Science Cardiff University
Advertisements

A NOVEL APPROACH TO SOLVING LARGE-SCALE LINEAR SYSTEMS Ken Habgood, Itamar Arel Department of Electrical Engineering & Computer Science GABRIEL CRAMER.
Michal Merta Alena Vašatová Václav Hapla David Horák
Advanced Computational Software Scientific Libraries: Part 2 Blue Waters Undergraduate Petascale Education Program May 29 – June
1 1 Capabilities: Suite of time integrators and nonlinear solvers  ODE integrators: (CVODE) variable order and step stiff BDF and non-stiff Adams, (ARKode)
1 A Common Application Platform (CAP) for SURAgrid -Mahantesh Halappanavar, John-Paul Robinson, Enis Afgane, Mary Fran Yafchalk and Purushotham Bangalore.
Extending the capability of TOUGHREACT simulator using parallel computing Application to environmental problems.
Parallel Algorithms in STAPL Implementation and Evaluation Jeremy Vu, Mauro Bianco, Nancy Amato Parasol Lab, Department of Computer.
MA5233: Computational Mathematics
Landscape Erosion Kirsten Meeker
Parallelization of FFT in AFNI Huang, Jingshan Xi, Hong Department of Computer Science and Engineering University of South Carolina.
PETSc Portable, Extensible Toolkit for Scientific computing.
An overview of the DANSE software architecture Michael Aivazis Caltech DANSE Kick-Off Meeting Pasadena Aug 15, 2006.
Anne Mascarin DSP Marketing The MathWorks
Lecture 29 Fall 2006 Lecture 29: Parallel Programming Overview.
An approach for solving the Helmholtz Equation on heterogeneous platforms An approach for solving the Helmholtz Equation on heterogeneous platforms G.
CS 591x – Cluster Computing and Programming Parallel Computers Parallel Libraries.
1 TOPS Solver Components Language-independent software components for the scalable solution of large linear and nonlinear algebraic systems arising from.
ICOM 5995: Performance Instrumentation and Visualization for High Performance Computer Systems Lecture 7 October 16, 2002 Nayda G. Santiago.
1 Using the PETSc Parallel Software library in Developing MPP Software for Calculating Exact Cumulative Reaction Probabilities for Large Systems (M. Minkoff.
ANS 1998 Winter Meeting DOE 2000 Numerics Capabilities 1 Barry Smith Argonne National Laboratory DOE 2000 Numerics Capability
MATLAB
ParCFD Parallel computation of pollutant dispersion in industrial sites Julien Montagnier Marc Buffat David Guibert.
Discontinuous Galerkin Methods and Strand Mesh Generation
Ch 1. A Python Q&A Session Spring Why do people use Python? Software quality Developer productivity Program portability Support libraries Component.
Nick Draper 05/11/2008 Mantid Manipulation and Analysis Toolkit for ISIS data.
Components for Beam Dynamics Douglas R. Dechow, Tech-X Lois Curfman McInnes, ANL Boyana Norris, ANL With thanks to the Common Component Architecture (CCA)
Strategies for Solving Large-Scale Optimization Problems Judith Hill Sandia National Laboratories October 23, 2007 Modeling and High-Performance Computing.
Center for Component Technology for Terascale Simulation Software CCA is about: Enhancing Programmer Productivity without sacrificing performance. Supporting.
1 SciDAC TOPS PETSc Work SciDAC TOPS Developers Satish Balay Chris Buschelman Matt Knepley Barry Smith.
Cellular Automata BIORemediation BIORemediation system system Cellular Automata BIORemediation BIORemediation system system M.C.Baracca P.Ornelli G.Clai.
ACES WorkshopJun-031 ACcESS Software System & High Level Modelling Languages by
Computational Aspects of Multi-scale Modeling Ahmed Sameh, Ananth Grama Computing Research Institute Purdue University.
1 1 What does Performance Across the Software Stack mean?  High level view: Providing performance for physics simulations meaningful to applications 
1 1  Capabilities: Building blocks for block-structured AMR codes for solving time-dependent PDE’s Functionality for [1…6]D, mixed-dimension building.
Implementing Hypre- AMG in NIMROD via PETSc S. Vadlamani- Tech X S. Kruger- Tech X T. Manteuffel- CU APPM S. McCormick- CU APPM Funding: DE-FG02-07ER84730.
October 2008 Integrated Predictive Simulation System for Earthquake and Tsunami Disaster CREST/Japan Science and Technology Agency (JST)
Cracow Grid Workshop, November 5-6, 2001 Concepts for implementing adaptive finite element codes for grid computing Krzysztof Banaś, Joanna Płażek Cracow.
MPI: Portable Parallel Programming for Scientific Computing William Gropp Rusty Lusk Debbie Swider Rajeev Thakur.
Computational Science & Engineering meeting national needs Steven F. Ashby SIAG-CSE Chair March 24, 2003.
Connections to Other Packages The Cactus Team Albert Einstein Institute
Linear Algebra Libraries: BLAS, LAPACK, ScaLAPACK, PLASMA, MAGMA
Mantid Manipulation and Analysis Toolkit for ISIS data.
1 1  Capabilities: Serial (C), shared-memory (OpenMP or Pthreads), distributed-memory (hybrid MPI+ OpenM + CUDA). All have Fortran interface. Sparse LU.
“NanoElectronics Modeling tool – NEMO5” Jean Michel D. Sellier Purdue University.
Photos placed in horizontal position with even amount of white space between photos and header Sandia National Laboratories is a multi-program laboratory.
CCA Common Component Architecture Distributed Array Component based on Global Arrays Manoj Krishnan, Jarek Nieplocha High Performance Computing Group Pacific.
Mantid Manipulation and Analysis Toolkit for Instrument data.
1 1  Capabilities: PCU: Communication, threading, and File IO built on MPI APF: Abstract definition of meshes, fields, and their algorithms GMI: Interface.
Algebraic Solvers in FASTMath Argonne Training Program on Extreme-Scale Computing August 2015.
From the customer’s perspective the SRS is: How smart people are going to solve the problem that was stated in the System Spec. A “contract”, more or less.
C OMPUTATIONAL R ESEARCH D IVISION 1 Defining Software Requirements for Scientific Computing Phillip Colella Applied Numerical Algorithms Group Lawrence.
Today's Software For Tomorrow's Hardware: An Introduction to Parallel Computing Rahul.S. Sampath May 9 th 2007.
Quality of Service for Numerical Components Lori Freitag Diachin, Paul Hovland, Kate Keahey, Lois McInnes, Boyana Norris, Padma Raghavan.
Is MPI still part of the solution ? George Bosilca Innovative Computing Laboratory Electrical Engineering and Computer Science Department University of.
Parallel OpenFOAM CFD Performance Studies Student: Adi Farshteindiker Advisors: Dr. Guy Tel-Zur,Prof. Shlomi Dolev The Department of Computer Science Faculty.
VisIt Project Overview
Hui Liu University of Calgary
A survey of Exascale Linear Algebra Libraries for Data Assimilation
Scalable Interfaces for Geometry and Mesh based Applications (SIGMA)
Shrirang Abhyankar IEEE PES HPC Working Group Meeting
MPI: Portable Parallel Programming for Scientific Computing
Programming Models for SimMillennium
HPC Modeling of the Power Grid
Polly Baker Division Director: Data, Mining, and Visualization
Salient application properties Expectations TOPS has of users
Short Course Siena, 5-6 October 2006
Ph.D. Thesis Numerical Solution of PDEs and Their Object-oriented Parallel Implementations Xing Cai October 26, 1998.
Presentation transcript:

1 1  Capabilities: Scalable algebraic solvers for PDEs Freely available and supported research code Usable from C, C++, Fortran 77/90, Python, MATLAB Uses MPI; encapsulates communication details within higher-level objects Tracks the largest DOE but commonly used on moderately sized systems (i.e., machines 1/10th to 1/100th the size of the largest system) 1,100 unique downloads per 300 petsc-maint/petsc-users s per 80 petsc-dev s per week Developed as a platform for experimentation Polymorphism: Single user interface for given functionality; multiple implementations  IS, Vec, Mat, KSP, PC, SNES, TS, etc. No optimality without interplay among physics, algorithmics, and architectures  Download and further info: Public questions: archived Private questions: not archived PETSc: Portable, Extensible Toolkit for Scientific computing

2 2 PETSc: Application highlights  Applications include: acoustics, aerodynamics, air pollution, arterial flow, bone fractures, brain surgery, cancer surgery, cancer treatment, carbon sequestration, cardiology, cells, CFD, combustion, concrete, corrosion, data mining, dentistry, earthquakes, economics, fission, fusion, glaciers, ground water flow, linguistics, mantel convection, magnetic films, materials science, medical imaging, ocean dynamics, oil recovery, page rank, polymer injection molding, polymeric membranes, quantum computing, seismology, semiconductors, rockets, relativity, surface water flow  UNIC (PROTEUS-SN): Neutron transport Uses preconditioned Krylov methods, finalist in 2009 Gordon Bell competition, runs with over 500 billion unknowns on 222,912 cores of Cray XT5 & 294,912 cores of BG/P  PFLOTRAN: Subsurface flow and reactive transport Uses preconditioned Newton-Krylov methods, runs with up 2 billion unknowns on 224,000 cores of Cray XT6

3 3  Algorithms, (parallel) debugging aids, low-overhead profiling  Composability Try new algorithms by choosing from product space and composing existing algorithms (multilevel, domain decomposition, splitting)  Experimentation It is not possible to pick the solver a priori. What will deliver best/competitive performance for a given physics, discretization, architecture, and problem size? PETSc’s response: Expose an algebra of composition so new solvers can be created at runtime; users can develop custom application-specific approaches Important to keep solvers decoupled from physics and discretization because we also experiment with those  Philosophy: Everything has a plugin architecture Vectors, matrices, coloring/ordering/partitioning algorithms Preconditioners, Krylov accelerators Nonlinear solvers, time integrators Spatial discretizations/topology PETSc: Overview of approach