Albert-Einstein-Institut www.aei-potsdam.mpg.de Exploring Distributed Computing Techniques with Ccactus and Globus Solving Einstein’s Equations, Black.

Slides:



Advertisements
Similar presentations
National Institute of Advanced Industrial Science and Technology Ninf-G - Core GridRPC Infrastructure Software OGF19 Yoshio Tanaka (AIST) On behalf.
Advertisements

Towards a Virtual European Supercomputing Infrastructure Vision & issues Sanzio Bassini
Beowulf Supercomputer System Lee, Jung won CS843.
Gabrielle Allen*, Thomas Dramlitsch*, Ian Foster †, Nicolas Karonis ‡, Matei Ripeanu #, Ed Seidel*, Brian Toonen † * Max-Planck-Institut für Gravitationsphysik.
Presented by Scalable Systems Software Project Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported by.
Cactus in GrADS Dave Angulo, Ian Foster Matei Ripeanu, Michael Russell Distributed Systems Laboratory The University of Chicago With: Gabrielle Allen,
Cactus in GrADS (HFA) Ian Foster Dave Angulo, Matei Ripeanu, Michael Russell.
Problem-Solving Environments: The Next Level in Software Integration David W. Walker Cardiff University.
Supporting Efficient Execution in Heterogeneous Distributed Computing Environments with Cactus and Globus Gabrielle Allen, Thomas Dramlitsch, Ian Foster,
GridLab: Dynamic Grid Applications for Science and Engineering A story from the difficult to the ridiculous… Ed Seidel Max-Planck-Institut für Gravitationsphysik.
Cactus Code and Grid Programming Here at GGF1: Gabrielle Allen, Gerd Lanfermann, Thomas Radke, Ed Seidel Max Planck Institute for Gravitational Physics,
GridLab & Cactus Joni Kivi Maarit Lintunen. GridLab  A project funded by the European Commission  The project was started in January 2002  Software.
1 GRID D. Royo, O. Ardaiz, L. Díaz de Cerio, R. Meseguer, A. Gallardo, K. Sanjeevan Computer Architecture Department Universitat Politècnica de Catalunya.
Cactus-G: Experiments with a Grid-Enabled Computational Framework Dave Angulo, Ian Foster Chuang Liu, Matei Ripeanu, Michael Russell Distributed Systems.
Simo Niskala Teemu Pasanen
Chapter 2 Computer Clusters Lecture 2.1 Overview.
Cactus 4.0. Cactus Computational Toolkit and Distributed Computing Solving Einstein’s Equations –Impact on computation Large collaborations essential.
Cactus Tools for the Grid Gabrielle Allen Max Planck Institute for Gravitational Physics, (Albert Einstein Institute)
Cornell Theory Center Aug CCTK The Cactus Computational Toolkit Werner Benger Max-PIanck-Institut für Gravitationsphysik (Albert-Einstein-Institute.
Cactus Computational Frameowork Freely available, modular, environment for collaboratively developing parallel, high- performance multi-dimensional simulations.
Albert-Einstein-Institut Cactus: Developing Parallel Computational Tools to Study Black Hole, Neutron Star (or Airplane...) Collisions.
The Cactus Code: A Parallel, Collaborative, Framework for Large Scale Computing Gabrielle Allen Max Planck Institute for Gravitational Physics, (Albert.
The MicroGrid: A Scientific Tool for Modeling Grids Andrew A. Chien SAIC Chair Professor Department of Computer Science and Engineering University of California,
The Astrophysics Simulation Collaboratory Portal Case Study of a Grid-Enabled Application Environment HPDC-10 San Francisco Michael Russell, Gabrielle.
CoG Kit Overview Gregor von Laszewski Keith Jackson.
Cactus Project & Collaborative Working Gabrielle Allen Max Planck Institute for Gravitational Physics, (Albert Einstein Institute)
Dynamic Grid Simulations for Science and Engineering Ed Seidel Max-Planck-Institut für Gravitationsphysik (Albert Einstein Institute) NCSA, U of Illinois.
Albert-Einstein-Institut Using Supercomputers to Collide Black Holes Solving Einstein’s Equations on the Grid Solving Einstein’s.
1 Cactus in a nutshell... n Cactus facilitates parallel code design, it enables platform independent computations and encourages collaborative code development.
Applications for the Grid Here at GGF1: Gabrielle Allen, Thomas, Dramlitsch, Gerd Lanfermann, Thomas Radke, Ed Seidel Max Planck Institute for Gravitational.
Evaluation of Agent Teamwork High Performance Distributed Computing Middleware. Solomon Lane Agent Teamwork Research Assistant October 2006 – March 2007.
Nomadic Grid Applications: The Cactus WORM G.Lanfermann Max Planck Institute for Gravitational Physics Albert-Einstein-Institute, Golm Dave Angulo University.
The Globus Project: A Status Report Ian Foster Carl Kesselman
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
ARGONNE NATIONAL LABORATORY Climate Modeling on the Jazz Linux Cluster at ANL John Taylor Mathematics and Computer Science & Environmental Research Divisions.
Perspectives on Grid Technology Ian Foster Argonne National Laboratory The University of Chicago.
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.
Introduction to Grid Computing Ed Seidel Max Planck Institute for Gravitational Physics
Ed Seidel Albert Einstein Institute Sources of Gravitational Radiation A.1 Development of CACTUS n Training in the use of the Cactus.
Developing Applications on Today’s Grids Tom Goodale Max Planck Institute for Gravitational Physics
The Grid the united computing power Jian He Amit Karnik.
1October 9, 2001 Sun in Scientific & Engineering Computing Grid Computing with Sun Wolfgang Gentzsch Director Grid Computing Cracow Grid Workshop, November.
The Cactus Code: A Problem Solving Environment for the Grid Gabrielle Allen, Gerd Lanfermann Max Planck Institute for Gravitational Physics.
F. Douglas Swesty, DOE Office of Science Data Management Workshop, SLAC March Data Management Needs for Nuclear-Astrophysical Simulation at the Ultrascale.
Cactus/TIKSL/KDI/Portal Synch Day. Agenda n Main Goals:  Overview of Cactus, TIKSL, KDI, and Portal efforts  present plans for each project  make sure.
GridLab WP-2 Cactus GAT (CGAT) Ed Seidel, AEI & LSU Co-chair, GGF Apps RG, Gridstart Apps TWG Gabrielle Allen, Robert Engel, Tom Goodale, *Thomas Radke.
New and Cool The Cactus Team Albert Einstein Institute
Connections to Other Packages The Cactus Team Albert Einstein Institute
1 Grid Activity Summary » Grid Testbed » CFD Application » Virtualization » Information Grid » Grid CA.
| nectar.org.au NECTAR TRAINING Module 4 From PC To Cloud or HPC.
Cactus Grid Computing Gabrielle Allen Max Planck Institute for Gravitational Physics, (Albert Einstein Institute)
2/22/2001Greenbook 2001/OASCR1 Greenbook/OASCR Activities Focus on technology to enable SCIENCE to be conducted, i.e. Software tools Software libraries.
Conundrum Talk, LBL May 2000 The Cactus Code: A Framework for Parallel Computing Gabrielle Allen Albert Einstein Institute Max Planck Institute for Gravitational.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
Dynamic Grid Computing: The Cactus Worm The Egrid Collaboration Represented by: Ed Seidel Albert Einstein Institute
Motivation: dynamic apps Rocket center applications: –exhibit irregular structure, dynamic behavior, and need adaptive control strategies. Geometries are.
New and Cool The Cactus Team Albert Einstein Institute
Cactus Workshop - NCSA Sep 27 - Oct Generic Cactus Workshop: Summary and Future Ed Seidel Albert Einstein Institute
Metacomputing Within the Cactus Framework What and why is Cactus? What has Cactus got to do with Globus? Gabrielle Allen, Thomas Radke, Ed Seidel. Albert-Einstein-Institut.
HPC University Requirements Analysis Team Training Analysis Summary Meeting at PSC September Mary Ann Leung, Ph.D.
Page : 1 SC2004 Pittsburgh, November 12, 2004 DEISA : integrating HPC infrastructures in Europe DEISA : integrating HPC infrastructures in Europe Victor.
Developing HPC Scientific and Engineering Applications: From the Laptop to the Grid Gabrielle Allen, Tom Goodale, Thomas.
Cactus Project & Collaborative Working
Clouds , Grids and Clusters
The Cactus Team Albert Einstein Institute
Grid Computing AEI Numerical Relativity Group has access to high-end resources in over ten centers in Europe/USA They want: Bigger simulations, more simulations.
Grid Computing.
Exploring Distributed Computing Techniques with Ccactus and Globus
Dynamic Grid Computing: The Cactus Worm
Presentation transcript:

Albert-Einstein-Institut Exploring Distributed Computing Techniques with Ccactus and Globus Solving Einstein’s Equations, Black Holes, and Gravitational Wave Astronomy Cactus, a new community simulation code framework: Grid enabling capabilities Previous Metacomputing experiments What we learned form those Current work, improvements –The present state Future development, goals Thomas Dramlitsch Albert-Einstein-Institut MPI-Gravitationsphysik (and AEI-ANL-NCSA-LBL team) Thomas Dramlitsch Albert-Einstein-Institut MPI-Gravitationsphysik (and AEI-ANL-NCSA-LBL team)

Albert-Einstein-Institut What is Cactus?: new concept in community developed simulation code infrastructure Numerical/computational infrastructure to solve PDE’s Freely available, open community source code: spirit of gnu/linux Developed as Response to Needs of these projects It’s production-software Cactus Divided in “Flesh” (core) and “Thorns” (modules or collections of subroutines) –User choice between Fortran, C, C ++ ; automated interface between them –Parallelism largely automatic and hidden (if desired) from user –Checkpointing / Restart capabilities Many parallel utilities / features enabled by Cactus –Parallel IO: FlexIO, HDF5; Data streaming, remote visualization/steering –Elliptic solvers: PETSc –And of course Metacomputing A Vision: any application can plug into Cactus to be Grid enabled Demo tomorrow night at HPDC

Albert-Einstein-Institut Modularity of Cactus... Application 1a Cactus Flesh Application 2... Sub-app AMR (Grace, etc) MPI layer 1I/O layer 2 Remote Steer 3 Globus Metcomputing Services User selects desired functionality... Application 1b

Albert-Einstein-Institut Metacomputing: harnessing power when and where it is needed Einstein equations typical of apps that require extreme memory, speed –many Flops per grid zone (~ ) –Finite differences on regular grids –Communications of variables through derivatives: ghost zones Largest supercomputers too small! Networks very fast! –OC-12 and higher very common in US –G-Win: 622 Mbits Potsdam-Berlin-Garching, connect multiple supercomputers –Gigabit networking to US possible “Seamless computing and visualization from anywhere” Many metacomputing experiments in progress

Albert-Einstein-Institut High performance: Full 3D Einstein Equations solved on NCSA NT Supercluster, Origin 2000, T3E Excellent scaling on many architectures –Origin up to 256 processors –T3E up to 1024 –NCSA NT cluster up to 128 processors Achieved 142 Gflops/s on 1024 node T3E-1200 (benchmarked for NASA NS Grand Challenge) But, of course, we want much more… metacomputing, meaning connected computers...

Albert-Einstein-Institut Metacomputing the Einstein Equations: Connecting T3E’s in Berlin, Garching, San Diego Want to migrate this technology to the generic user...

Albert-Einstein-Institut Scaling of Cactus on two T3Es on different continents San Diego & Berlin Berlin & Munich

Albert-Einstein-Institut Scaling of Cactus on Multiple SGIs at Remote Sites Argonne & NCSA

Albert-Einstein-Institut Analysis of previous metacomputing experiments It worked! (That’s the main thing we wanted at SC98…) Cactus was not optimized for metacomputing: messages too small, latency etc.. Mpich-G could perform better, e.g. intra-machine communication one order of magnitude slower than native MPI –Mpich-G2 improves this... Communication is non-trivial (not “embarrassingly parallel”) and very intensive Experiments showed: –For some problems, this is feasible –We to improve performance significantly with work on optimization of Cactus and Mpich-G –That’s what we did!

Albert-Einstein-Institut Optimizing Cactus Communication Layers for Metacomputing Made the communication layer(s) much more flexible: –Can specify size and number of messages, in order to achieve best performance with the underlying network (bandwith, latency) –Reduced communication to a bare minimum –Overlapping of communication with other cpu’s –Overlapping of communication and Computation Made the load balancing of cactus more flexible (Matei Ripeanu): –Cactus now allows to decompose the total problem into pieces of different size, according to cpu-power, number of cpu’s used on one machine etc... Cactus compiles (out of the box) with globus and mpich on most common architectures (T3e, Irix, SP-2,…?)

Albert-Einstein-Institut Optimizing Mpich-G: Used Mpich-G2 MPICH-G2 is a completely rewritten communication layer Can distinguish between inter- and intra-machine communication –It uses the vendor’s supplied mpi for intra-machine communication –Uses TCP/IP between machines This means optimal performance in a metacomputing environment Works with Cactus and Globus on all major unix-systems TCP/IP MPI_COMM_WORLD

Albert-Einstein-Institut Current experiments and future plans Current Experiment –Complete testing and production of tightly coupled simulation between different sites in the USA (NCSA, NERSC, ANL, SDSC and others) –Want to use advanced software (Portal, co-scheduling systems etc..) –Want to run across many sites and nodes as possible More General Grid Computing problems –Distribution of multiple grids –Dynamic resource acquisition Aquiring more memory when needed (AMR) Spawning off connected jobs on remote machines Cactus thorn would have access to MDS …

Albert-Einstein-Institut Cactus Computational Toolkit Science, Autopilot, AMR, Petsc, HDF, MPI, GrACE, Globus, Remote Steering... A Portal to Computational Science: The Cactus Collaboratory 1. User has science idea Selects Appropriate Resources Collaborators log in to monitor Steers simulation, monitors performance Composes/Builds Code Components w/Interface... Want to integrate and migrate this technology to the generic user...

Albert-Einstein-Institut German Gigabit Project supported by DFN-Verein Developing Techniques to Exploit High Speed Networks Focus on Remote Steering and Visualization OC-12 Testbed between AEI, ZIB, RZG with built-in application groups ready to use it! Already closely connected to ANL, NCSA, KDI projects AEI

Albert-Einstein-Institut Metacomputing Experiments, Production SC93: remote CM-5 simulation with live viz in CAVE SC95: Heroic I-Way experiments leads to development of Globus. Cornell SP-2, Power Challenge, with live viz in San Diego CAVE SC97: Garching 512 node T3E, launched, controlled, visualized in San Jose SC98: HPC Challenge. SDSC, ZIB, and Garching T3E compute collision of 2 Neutron Stars, controlled from Orlando SC99: Colliding Black Holes using Garching, ZIB T3E’s, with remote collaborative interaction and viz at ANL and NCSA booths April 2000: Attempting to use LANL, NCSA, NERSC, SDSC, ZIB, Garching, NASA-Ames, Maui?, +…? for single simulation! All this technology is available to in main production code for different applications!