Pioneering Science and Technology Office of Science U.S. Department of Energy 1 MCS Vision Petascale Computing Grid Computing Computational Science and.

Slides:



Advertisements
Similar presentations
University of Illinois at Chicago The Future of STAR TAP: Enabling e-Science Research Thomas A. DeFanti Principal Investigator, STAR TAP Director, Electronic.
Advertisements

1 US activities and strategy :NSF Ron Perrott. 2 TeraGrid An instrument that delivers high-end IT resources/services –a computational facility – over.
O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Center for Computational Sciences Cray X1 and Black Widow at ORNL Center for Computational.
University of Chicago Department of Energy The Parallel and Grid I/O Perspective MPI, MPI-IO, NetCDF, and HDF5 are in common use Multi TB datasets also.
U.S. Department of Energy’s Office of Science Basic Energy Sciences Advisory Committee Dr. Daniel A. Hitchcock October 21, 2003
GENI: Global Environment for Networking Innovations Larry Landweber Senior Advisor NSF:CISE Joint Techs Madison, WI July 17, 2006.
SDSC Computing the 21st Century Talk Given to the NSF Sugar Panel May 27, 1998.
The DOE Science Grid Computing and Data Infrastructure for Large-Scale Science William Johnston, Lawrence Berkeley National Lab Ray Bair, Pacific Northwest.
Presented by Scalable Systems Software Project Al Geist Computer Science Research Group Computer Science and Mathematics Division Research supported by.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
Astrophysics, Biology, Climate, Combustion, Fusion, Nanoscience Working Group on Simulation-Driven Applications 10 CS, 10 Sim, 1 VR.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
NPACI: National Partnership for Advanced Computational Infrastructure August 17-21, 1998 NPACI Parallel Computing Institute 1 Cluster Archtectures and.
Simo Niskala Teemu Pasanen
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
UNIVERSITY of MARYLAND GLOBAL LAND COVER FACILITY High Performance Computing in Support of Geospatial Information Discovery and Mining Joseph JaJa Institute.
1 Building National Cyberinfrastructure Alan Blatecky Office of Cyberinfrastructure EPSCoR Meeting May 21,
Computing in Atmospheric Sciences Workshop: 2003 Challenges of Cyberinfrastructure Alan Blatecky Executive Director San Diego Supercomputer Center.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Scientific Data Infrastructure in CAS Dr. Jianhui Scientific Data Center Computer Network Information Center Chinese Academy of Sciences.
SGI Proprietary SGI Update IDC HPC User Forum September, 2008.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
TeraGrid and I-WIRE: Models for the Future? Rick Stevens and Charlie Catlett Argonne National Laboratory The University of Chicago.
U.S. Department of Energy Office of Science Advanced Scientific Computing Research Program CASC, May 3, ADVANCED SCIENTIFIC COMPUTING RESEARCH An.
Seaborg Cerise Wuthrich CMPS Seaborg  Manufactured by IBM  Distributed Memory Parallel Supercomputer  Based on IBM’s SP RS/6000 Architecture.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Cyberinfrastructure Planning at NSF Deborah L. Crawford Acting Director, Office of Cyberinfrastructure HPC Acquisition Models September 9, 2005.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
SciDAC All Hands Meeting, March 2-3, 2005 Northwestern University PIs:Alok Choudhary, Wei-keng Liao Graduate Students:Avery Ching, Kenin Coloma, Jianwei.
The Future of the iPlant Cyberinfrastructure: Coming Attractions.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
DataTAG Research and Technological Development for a Transatlantic Grid Abstract Several major international Grid development projects are underway at.
ARGONNE NATIONAL LABORATORY Climate Modeling on the Jazz Linux Cluster at ANL John Taylor Mathematics and Computer Science & Environmental Research Divisions.
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.
Commodity Grid Kits Gregor von Laszewski (ANL), Keith Jackson (LBL) Many state-of-the-art scientific applications, such as climate modeling, astrophysics,
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Facilities and How They Are Used ORNL/Probe Randy Burris Dan Million – facility administrator.
National Computational Science University of Illinois at Urbana-Champaign Alliance Vision of the Future (9/2000) Commodity cluster hardware –Intel IA-32/64.
Authors: Ronnie Julio Cole David
Cyberinfrastructure What is it? Russ Hobby Internet2 Joint Techs, 18 July 2007.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Scalable Systems Software for Terascale Computer Centers Coordinator: Al Geist Participating Organizations ORNL ANL LBNL.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Computational Science & Engineering meeting national needs Steven F. Ashby SIAG-CSE Chair March 24, 2003.
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
Data Center & Large-Scale Systems (updated) Luis Ceze, Bill Feiereisen, Krishna Kant, Richard Murphy, Onur Mutlu, Anand Sivasubramanian, Christos Kozyrakis.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
2/22/2001Greenbook 2001/OASCR1 Greenbook/OASCR Activities Focus on technology to enable SCIENCE to be conducted, i.e. Software tools Software libraries.
Securing the Grid & other Middleware Challenges Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer.
The Performance Evaluation Research Center (PERC) Participating Institutions: Argonne Natl. Lab.Univ. of California, San Diego Lawrence Berkeley Natl.
SDM Center High-Performance Parallel I/O Libraries (PI) Alok Choudhary, (Co-I) Wei-Keng Liao Northwestern University In Collaboration with the SEA Group.
Presented by NCCS Hardware Jim Rogers Director of Operations National Center for Computational Sciences.
An Architectural Approach to Managing Data in Transit Micah Beck Director & Associate Professor Logistical Computing and Internetworking Lab Computer Science.
Center for Extended MHD Modeling (PI: S. Jardin, PPPL) –Two extensively developed fully 3-D nonlinear MHD codes, NIMROD and M3D formed the basis for further.
Tackling I/O Issues 1 David Race 16 March 2010.
PARALLEL AND DISTRIBUTED PROGRAMMING MODELS U. Jhashuva 1 Asst. Prof Dept. of CSE om.
HPC University Requirements Analysis Team Training Analysis Summary Meeting at PSC September Mary Ann Leung, Ph.D.
Northwest Indiana Computational Grid Preston Smith Rosen Center for Advanced Computing Purdue University - West Lafayette West Lafayette Calumet.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
CyberGIS Prof. Wenwen Li School of Geographical Sciences and Urban Planning 5644 Coor Hall
Charlie Catlett UIUC/NCSA Starlight International Optical Network Hub (NU-Chicago) Argonne National Laboratory U Chicago IIT UIC.
CITA|ICAT Jonathan Dursi HPCS’06 15 May Towards Understanding some Astrophysical Flows using Multiscale Simulations with the FLASH code Jonathan Dursi,
Clouds , Grids and Clusters
Performance Technology for Scalable Parallel Systems
Joint Techs, Columbus, OH
BlueGene/L Supercomputer
TeraScale Supernova Initiative
Presentation transcript:

Pioneering Science and Technology Office of Science U.S. Department of Energy 1 MCS Vision Petascale Computing Grid Computing Computational Science and Engineering Petascale Computing Grid Computing Computational Science and Engineering Increase by several orders of magnitude the computing power that can be applied to individual scientific problems, thus enabling progress in understanding complex physical and biological systems. Interconnect the world’s most important scientific databases, computing systems, instruments and facilities to improve scientific productivity and remove barriers to collaboration. Make high-end computing a core tool for challenging modeling, simulation and analysis problems.

Pioneering Science and Technology Office of Science U.S. Department of Energy 2 MCS Products/Resources Enabling technologies -“middleware” -“tools” -“support applications” Scientific applications Hardware Other fundamental CS research

Pioneering Science and Technology Office of Science U.S. Department of Energy 3 Enabling Technologies Globus Toolkit -Software infrastructure/standards for Grid computing MPICH -Our free implementation of MPI Jumpshot -Software for analysis of message passing pNetCDF -High performance parallel I/O library PetsC -Toolkit for parallel matrix solves Visualization (“Futures lab”) -Scalable parallel visualization software, large-scale displays Access Grid -Collaboration environment

Pioneering Science and Technology Office of Science U.S. Department of Energy 4 Collaboration Technology – the Access Grid Multi-way meetings and conferences over the Internet Using high-quality video/audio technology Large format display: 200+ installations worldwide Easily replicated configurations, open source software

Pioneering Science and Technology Office of Science U.S. Department of Energy 5 The Grid Links People with Distributed Resources on a National Scale Sensors

Pioneering Science and Technology Office of Science U.S. Department of Energy 6 Some key scientific applications Flash -Community code for general Astrophysical phenomena -ASCI project with UC Nek5 -Biological fluids pNeo -Neo-cortex simulations for study of epileptic seizures QMC -Monte Carlo simulations of atomic nuclei Nuclear Reactor Simulations

Pioneering Science and Technology Office of Science U.S. Department of Energy 7 Hardware Chiba City – Software Scalability R&D Addresses scalability issues in system software, open source software, and applications code. 512 CPUs, 256 nodes, Myrinet, 2TB storage, Linux. DOE OASCR funded. Installed in Jazz – Linux Cluster for ANL Applications Supports and enhances ANL application community. 50+ projects from a spectrum of S&E divisions 350 CPUs, Myrinet, 20TB storage. ANL funded. Installed in Achieved 1.1 TF sustained. Blue Gene prototype – coming soon two-rack system scalable to twenty racks

Pioneering Science and Technology Office of Science U.S. Department of Energy 8 Other HPC areas Architecture and Performance Evaluation Programming Models and Languages Systems Software Numerical Methods and Optimization Software components Software Verification Automatic Differentiation

Pioneering Science and Technology Office of Science U.S. Department of Energy 9 I-WIRE Flash Blue Gene NLCF Teragrid I-Wire Impact Two concrete examples of the impact of I-WIRE

Pioneering Science and Technology Office of Science U.S. Department of Energy 10 UIUC/NCSA Starlight International Optical Network Hub (NU-Chicago) Argonne National Laboratory U Chicago IIT UIC Chicago U Chicago Gleacher Center James R. Thompson Center – Illinois Century Network 40 Gb/s Distributed Terascale Facility Network Commercial Fiber Hub I-WIRE State Funded Dark Fiber Optical Infrastructure to support Networking and Applications Research -$11.5M Total Funding - $6.5M FY $5M in process for FY04-5 -Application Driven - Access Grid: Telepresence & Media - TeraGrid: Computational and Data Grids -New Technologies Proving Ground - Optical Network Technologies - Middleware and Computer Science Research Deliverables -A flexible infrastructure to support advanced applications and networking research (Illinois Wired/Wireless Infrastructure for Research and Education)

Pioneering Science and Technology Office of Science U.S. Department of Energy 11 UI-Chicago Illinois Inst. Tech Northwestern Univ-Chicago “Starlight” U of Chicago I-55 I-90/94 I-290 I-294 UIUC/NCSA Urbana (approx 140 miles South) ANL IITUIC UChicago Main UIUC/NCSA Starlight / NU-C ICN UChicago Gleacher Ctr Commercial Fiber Hub Status: I-WIRE Geography Argonne Nat’l Lab (approx 25 miles SW)

Pioneering Science and Technology Office of Science U.S. Department of Energy 12 TeraGrid Vision: A Unified National HPC Infrastructure that is Persistent and Reliable Largest NSF compute resources Largest DOE instrument (SNS) Fastest network Massive storage Visualization instruments Science Gateways Community databases E.g: Geosciences: 4 data collections including high-res CT scans, global telemetry data, worldwide hydrology data, and regional LIDAR terrain data

Pioneering Science and Technology Office of Science U.S. Department of Energy 13 Resources and Services (33TF, 1.1PB disk, 12 PB tape)

Pioneering Science and Technology Office of Science U.S. Department of Energy 14 Los Angeles Atlanta SDSCTACC NCSAPurdueORNL UC/ANLPSC Caltech Chicago IU Current TeraGrid Network Resources: Compute, Data, Instrument, Science Gateways

Pioneering Science and Technology Office of Science U.S. Department of Energy 15 Flash Flash Project -Community Astrophysics code -DOE funded ASCI program at UC/Argonne -4 million per year over ten years -Currently in 7 th year Flash Code/Framework -Heavy emphasis on software engineering, performance, and usability downloads -Active user community -Runs on all major hpc platforms -Public automated testing facility -Extensive user documentation

Pioneering Science and Technology Office of Science U.S. Department of Energy 16 Flash -- Simulating Astrophysical processes Cellular detonation Compressed turbulence Helium burning on neutron stars Richtmyer-Meshkov instability Laser-driven shock instabilities Nova outbursts on white dwarfs Rayleigh-Taylor instability Flame-vortex interactions Gravitational collapse/Jeans instability Wave breaking on white dwarfs Shortly: Relativistic accretion onto NS Orzag/Tang MHD vortex Type Ia Supernova Intracluster interactions Magnetic Rayleigh-Taylor

Pioneering Science and Technology Office of Science U.S. Department of Energy 17 How has fast network helped Flash? Flash in production for five years Generating terabytes of data Currently done “by hand” -Data transferred locally from supercomputing centers for visualization/analysis -Data remotely visualized at UC using Argonne servers -Can harness data storage across several sites Not just “visionary” grid ideas that are useful. Immediate “mundane” things as well!

Pioneering Science and Technology Office of Science U.S. Department of Energy 18 Buoyed Progress in HPC FLASH flagship application for BG/L -Currently being run on 4K processors at Watson -Will run on 16K procs in several months Argonne partnership with Oak Ridge for National Leadership Class Computing Facility -Non-classified computing -BG at Argonne -X1, Black Widow at ORNL -Application focus groups apply for time

Pioneering Science and Technology Office of Science U.S. Department of Energy 19 Petaflops Hardware is Just Around the Corner TeraflopsPetaflops We are here

Pioneering Science and Technology Office of Science U.S. Department of Energy 20 Diverse Architectures for Petaflop Systems vIBM – Blue Gene -Puts processors + cache + network interfaces on same chip -Achieves high packaging density, low power consumption vCray – RS and X1 -10K processor (40 TF) Red Storm at SNL -1K processor (20 TF) X1 to ORNL Emerging -Field Programmable Gate Arrays -Processor in Memory -Streams … v= systems slated for DOE National Leadership Computing Facility

Pioneering Science and Technology Office of Science U.S. Department of Energy 21 NLCF Target Application Areas SciDAC Fusion SciDAC Astrophysics Genomes to Life Nanophase Materials SciDAC Climate SciDAC Chemistry

Pioneering Science and Technology Office of Science U.S. Department of Energy 22 The Blue Gene Consortium: Goals Provide new capabilities to selected applications partnerships. Provide functional requirements for a petaflop/sec version of BG. Build a community around a new class of architecture. -Thirty university and lab partners -About ten hardware partners and about twenty software collaborators Develop a new, sustainable model of partnership. -“Research product” by passing normal “productization” process/costs -Community-based support model (hub and spoke) Engage (or re-engage) computer science researchers with high-performance computing architecture. -Broad community access to hardware systems -Scalable operating system research and novel software research Partnership of DOE, NSF, NIH, NNSA, and IBM will work on computer science, computational science, and architecture development. Kickoff meeting was April 27, 2004, in Chicago.

Pioneering Science and Technology Office of Science U.S. Department of Energy 23 Determining application fit How will applications map onto different petaflop architectures? applications BG/L BG/PBG/Q Vector/parallel X1 BlackWidow Cluster Red Storm ? Massively parallel Fine grain Our focus

Pioneering Science and Technology Office of Science U.S. Department of Energy 24 Application Analysis Process A priori algorithmic performanceanalysis Scalability Analysis (Jumpshot, FPMPI) Refine model Tune code Validate model (BG/L hardware) Look for inauspicious scalability bottlenecksLook for inauspicious scalability bottlenecks May be based on data fromMay be based on data from conventional systems or BGSim conventional systems or BGSim Extrapolate conclusionsExtrapolate conclusions Model 0 th order behaviorModel 0 th order behavior Detailed message passing statistics on BG hardware Detailed architectural adjustments PerformanceAnalysis (PAPI, hpmlib) Rewrite algorithms? Use of memory hierarchyUse of memory hierarchy Use of pipelineUse of pipeline Instruction mixInstruction mix Bandwidth vs. latencyBandwidth vs. latency etc.etc. 0 th order scaling problems

Pioneering Science and Technology Office of Science U.S. Department of Energy 25 High performance resources operate in complex and highly distributed environments Argonne co-founded the Grid -Establishing a persistent, standards-based infrastructure and applications interfaces that enable high performance assess to computation and data -ANL created the Global Grid Forum, an international standards body -We lead development of the Globus Toolkit ANL staff are PIs and key contributors in many grid technology development and application projects -High performance data transport, Grid security, virtual organization mgt., Open Grid Services Architecture, … -Access Grid – group-to-group collaboration via large multimedia displays