Argonne Leadership Computing Facility ALCF at Argonne  Opened in 2006  Operated by the Department of Energy’s Office of Science  Located at Argonne.

Slides:



Advertisements
Similar presentations
University of Chicago Department of Energy The Parallel and Grid I/O Perspective MPI, MPI-IO, NetCDF, and HDF5 are in common use Multi TB datasets also.
Advertisements

The Development of Mellanox - NVIDIA GPUDirect over InfiniBand A New Model for GPU to GPU Communications Gilad Shainer.
© 2007 IBM Corporation IBM Global Engineering Solutions IBM Blue Gene/P Blue Gene/P System Overview - Hardware.
Rhea Analysis & Post-processing Cluster Robert D. French NCCS User Assistance.
Appro Xtreme-X Supercomputers A P P R O I N T E R N A T I O N A L I N C.
Parallel Research at Illinois Parallel Everywhere
Taxanomy of parallel machines. Taxonomy of parallel machines Memory – Shared mem. – Distributed mem. Control – SIMD – MIMD.
IDC HPC User Forum Conference Appro Product Update Anthony Kenisky, VP of Sales.
Information Technology Center Introduction to High Performance Computing at KFUPM.
1 BGL Photo (system) BlueGene/L IBM Journal of Research and Development, Vol. 49, No. 2-3.
An Introduction to Princeton’s New Computing Resources: IBM Blue Gene, SGI Altix, and Dell Beowulf Cluster PICASso Mini-Course October 18, 2006 Curt Hillegas.
Hitachi SR8000 Supercomputer LAPPEENRANTA UNIVERSITY OF TECHNOLOGY Department of Information Technology Introduction to Parallel Computing Group.
IBM RS/6000 SP POWER3 SMP Jari Jokinen Pekka Laurila.
Lecture 1: Introduction to High Performance Computing.
© 2007 IBM Corporation IBM Global Engineering Solutions IBM Blue Gene/P Software Overview.
Real Parallel Computers. Modular data centers Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra,
Critical Flags, Variables, and Other Important ALCF Minutiae Jini Ramprakash Technical Support Specialist Argonne Leadership Computing Facility.
Slide 1 Auburn University Computer Science and Software Engineering Scientific Computing in Computer Science and Software Engineering Kai H. Chang Professor.
LLNL-PRES-XXXXXX This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
U.S. Department of Energy Office of Science Advanced Scientific Computing Research Program CASC, May 3, ADVANCED SCIENTIFIC COMPUTING RESEARCH An.
Research Support Services Research Support Services.
1 Support and Maintenance Kristi Jacobson 7/15/08.
Principles of Scalable HPC System Design March 6, 2012 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Artdaq Introduction artdaq is a toolkit for creating the event building and filtering portions of a DAQ. A set of ready-to-use components along with hooks.
Data Intensive Computing at Sandia September 15, 2010 Andy Wilson Senior Member of Technical Staff Data Analysis and Visualization Sandia National Laboratories.
Improving Network I/O Virtualization for Cloud Computing.
U.S. Department of Energy Office of Science Advanced Scientific Computing Research Program NERSC Users Group Meeting Department of Energy Update September.
Computer Science Section National Center for Atmospheric Research Department of Computer Science University of Colorado at Boulder Blue Gene Experience.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Rensselaer Why not change the world? Rensselaer Why not change the world? 1.
Jaguar Super Computer Topics Covered Introduction Architecture Location & Cost Bench Mark Results Location & Manufacturer Machines in top 500 Operating.
IBM Systems and Technology Group © 2007 IBM Corporation High Throughput Computing on Blue Gene IBM Rochester: Amanda Peters, Tom Budnik With contributions.
BlueGene/L Facts Platform Characteristics 512-node prototype 64 rack BlueGene/L Machine Peak Performance 1.0 / 2.0 TFlops/s 180 / 360 TFlops/s Total Memory.
BG/Q vs BG/P—Applications Perspective from Early Science Program Timothy J. Williams Argonne Leadership Computing Facility 2013 MiraCon Workshop Monday.
IM&T Vacation Program Benjamin Meyer Virtualisation and Hyper-Threading in Scientific Computing.
Stochastic optimization of energy systems Cosmin Petra Argonne National Laboratory.
- Rohan Dhamnaskar. Overview  What is a Supercomputer  Some Concepts  Couple of examples.
CCS Overview Rene Salmon Center for Computational Science.
A Framework for Visualizing Science at the Petascale and Beyond Kelly Gaither Research Scientist Associate Director, Data and Information Analysis Texas.
Leibniz Supercomputing Centre Garching/Munich Matthias Brehm HPC Group June 16.
Brent Gorda LBNL – SOS7 3/5/03 1 Planned Machines: BluePlanet SOS7 March 5, 2003 Brent Gorda Future Technologies Group Lawrence Berkeley.
© 2009 IBM Corporation Motivation for HPC Innovation in the Coming Decade Dave Turek VP Deep Computing, IBM.
1 Cray Inc. 11/28/2015 Cray Inc Slide 2 Cray Cray Adaptive Supercomputing Vision Cray moves to Linux-base OS Cray Introduces CX1 Cray moves.
© Copyright 2012 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice. HP Update IDC HPC Forum.
Modeling Billion-Node Torus Networks Using Massively Parallel Discrete-Event Simulation Ning Liu, Christopher Carothers 1.
Lawrence Livermore National Laboratory S&T Principal Directorate - Computation Directorate Tools and Scalable Application Preparation Project Computation.
ARCHER Advanced Research Computing High End Resource
Template This is a template to help, not constrain, you. Modify as appropriate. Move bullet points to additional slides as needed. Don’t cram onto a single.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
ISTeC Research Computing Open Forum: Using NSF or National Laboratory Resources for High Performance Computing Bhavesh Khemka.
SDM Center High-Performance Parallel I/O Libraries (PI) Alok Choudhary, (Co-I) Wei-Keng Liao Northwestern University In Collaboration with the SEA Group.
Other Tools HPC Code Development Tools July 29, 2010 Sue Kelly Sandia is a multiprogram laboratory operated by Sandia Corporation, a.
Is MPI still part of the solution ? George Bosilca Innovative Computing Laboratory Electrical Engineering and Computer Science Department University of.
PARALLEL AND DISTRIBUTED PROGRAMMING MODELS U. Jhashuva 1 Asst. Prof Dept. of CSE om.
HPC University Requirements Analysis Team Training Analysis Summary Meeting at PSC September Mary Ann Leung, Ph.D.
Petascale Computing Resource Allocations PRAC – NSF Ed Walker, NSF CISE/ACI March 3,
The Evolution of the Italian HPC Infrastructure Carlo Cavazzoni CINECA – Supercomputing Application & Innovation 31 Marzo 2015.
Fermi National Accelerator Laboratory & Thomas Jefferson National Accelerator Facility SciDAC LQCD Software The Department of Energy (DOE) Office of Science.
BLUE GENE Sunitha M. Jenarius. What is Blue Gene A massively parallel supercomputer using tens of thousands of embedded PowerPC processors supporting.
Page : 1 SC2004 Pittsburgh, November 12, 2004 DEISA : integrating HPC infrastructures in Europe DEISA : integrating HPC infrastructures in Europe Victor.
Benefits. CAAR Project Phases Each of the CAAR projects will consist of a: 1.three-year Application Readiness phase ( ) in which the code refactoring.
Introduction to Data Analysis with R on HPC Texas Advanced Computing Center Feb
PARALLEL MODEL OF EVOLUTIONARY GAME DYNAMICS Amanda Peters MIT /13/2009.
Ben Rogers August 18,  High Performance Computing  Data Storage  Hadoop Pilot  Secure Remote Desktop  Training Opportunities  Grant Collaboration.
Advanced Computing Facility Introduction
A Brief Introduction to NERSC Resources and Allocations
University of Technology
Cray Announces Cray Inc.
SiCortex Update IDC HPC User Forum
Presentation transcript:

Argonne Leadership Computing Facility ALCF at Argonne  Opened in 2006  Operated by the Department of Energy’s Office of Science  Located at Argonne National Laboratory (30 miles southwest of Chicago) 1

Argonne Leadership Computing Facility IBM Blue Gene/P, Intrepid (2007)  163,840 processors  80 terabytes of memory  557 teraflops  Energy-efficient system uses one-third the electricity of machines built with conventional parts  #38 on Top500 (June 2012)  #15 on Graph500 (June 2012) The groundbreaking Blue Gene  General-purpose architecture excels in virtually all areas of computational science  Presents an essentially standard Linux/PowerPC programming environment  Significant impact on HPC – Blue Gene systems are consistently found in the top ten list  Delivers excellent performance per watt  High reliability and availability 2

Argonne Leadership Computing Facility IBM Blue Gene/Q, Mira  IBM Blue Gene/Q, Mira – 768,000 processors – 768 terabytes of memory – 10 petaflops – #3 on Top500 (June 2012) – #1 on Graph500 (June 2012) 3 Blue Gene/Q Prototype 2 ranked #1 June 2011

Argonne Leadership Computing Facility Programs for Obtaining System Allocations 4 For more information, visit: / collaborations/index.ph p)

Argonne Leadership Computing Facility The U.S. Department of Energy’s INCITE Program INCITE seeks out large, computationally intensive research projects and awards more than a billion processing hours to enable high- impact scientific advances.  Open to researchers in academia, industry, and other organizations  Proposed projects undergo scientific and computational readiness reviews  More than a billion total hours are awarded to a small number of projects  Sixty percent of the ALCF’s processing hours go to INCITE projects  Call for proposals issued once per year 5

Argonne Leadership Computing Facility 2012 INCITE Allocations by Discipline 6

Argonne Leadership Computing Facility World-Changing Science Underway at the ALCF  Research that will lead to improved, emissions- reducing catalytic systems for industry (Greeley)  Enhancing pubic safety through more accurate earthquake forecasting (Jordan)  Designing more efficient nuclear reactors that are less susceptible to dangerous, costly failures (Fischer)  Accelerating research that may improve diagnosis and treatment for patients with blood-flow complications (Karniadakis)  Protein studies that will apply to a broad range of problems, such as a finding a cure for Alzheimer’s disease, creating inhibitors of pandemic influenza, or engineering a step in the production of biofuels (Baker)  Furthering research to bring green energy sources, like hydrogen fuel, safely into our everyday lives, reducing our dependence on foreign fuels (Khokhlov) 7

Argonne Leadership Computing Facility ALCF Service Offerings  Scientific liaison (“Catalyst”) for INCITE and ALCC projects, providing collaboration along with assistance with proposals and planning  Startup assistance and technical support  Performance engineering and application tuning  Data analysis and visualization experts  MPI and MPI-I/O experts  Workshops and Seminars 8

Argonne Leadership Computing Facility A single node  Can be carved up into multiple MPI ranks, or as a single MPI rank with threads – Up to 4 MPI ranks/node on intrepid, up to 64 MPI ranks/node on mira  SIMD available on the cores, required to reach peak flop rate – 2-way FPU on intrepid, 4-way FPU on mira  Runs a Compute Node Kernel, requires cross-compiling from the front-end login nodes  Forwards I/O operations to an I/O node, which aggregates requests from multiple compute nodes  No virtual memory – 2 GB/node on intrepid, 16 GB/node on mira  No fork()/system() calls 9

Argonne Leadership Computing Facility A partition  Partitions come in pre-defined sizes that gain you isolation from other users  Additionally, you get the I/O nodes connecting you to the GPFS filesystems – requires a scalable I/O strategy!  Partitions can be as small as 512 nodes (16 on development rack), up to the size of the full machine  At the small scale, this is governed by the ratio of I/O nodes to compute nodes  At the large scale, this is governed by the network links required to make a torus, rather than a mesh 10

Argonne Leadership Computing Facility Blue Gene/P hierarchy: 11 1 chip, 20 DRAMs 13.6 GF/s 2.0 GB DDR Supports 4-way SMP 32 Node Cards 1024 chips, 4096 procs 14 TF/s 2 TB 40 Racks 556 TF/s 82TB Rack Intrepid System Compute Card 435 GF/s 64 GB (32 chips 4x4x2) 32 compute, 0-2 IO cards Node Card Front End Node / Service Node System p Servers Linux SLES10

Argonne Leadership Computing Facility 12

Argonne Leadership Computing Facility 13

Argonne Leadership Computing Facility Visualization and Data Analytics  Both systems come with a linux cluster attached to the same GPFS filesystems and network infrastructure  The GPUs on these machines can be used for straight visualization, or to perform data analysis  Software includes VISIT, ParaView, and other viz toolkits 14

Argonne Leadership Computing Facility Programming Models and Development Environment  Basically, all of the lessons from this week apply: MPI, pthreads, OpenMP, using any of C, C++, Fortran – Also have access to things like Global Arrays and other lower-level communication protocols if that’s your thing – Can use XL or GNU compilers, along with LLVM (beta)  I/O using HDF, NetCDF, MPI-I/O, …  Debugging with TAU, HPCToolkit, DDT, TotalView, …  Many supported libraries, like BLAS, PetSc, Scalapack, … 15

Argonne Leadership Computing Facility How do you get involved?  Send to requesting access to the CScADs  Or, go to and request a project of your ownhttps://accounts.alcf.anl.gov 16