Petascale System Requirements for the Geosciences Richard Loft SCD Deputy Director for R&D.

Slides:



Advertisements
Similar presentations
O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Center for Computational Sciences Cray X1 and Black Widow at ORNL Center for Computational.
Advertisements

Parallel computer architecture classification
Supercomputing Challenges at the National Center for Atmospheric Research Dr. Richard Loft Computational Science Section Scientific Computing Division.
One-day Meeting, INI, September 26th, 2008 Role of spectral turbulence simulations in developing HPC systems YOKOKAWA, Mitsuo Next-Generation Supercomputer.
Parallel Research at Illinois Parallel Everywhere
National Center for Atmospheric Research John Clyne 4/27/11 4/26/20111.
1 Cyberinfrastructure Framework for 21st Century Science & Engineering (CIF21) NSF-wide Cyberinfrastructure Vision People, Sustainability, Innovation,
1 Cyberinfrastructure Framework for 21st Century Science & Engineering (CF21) IRNC Kick-Off Workshop July 13,
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA; SAN DIEGO IEEE Symposium of Massive Storage Systems, May 3-5, 2010 Data-Intensive Solutions.
UCAR CONFIDENTIAL NCAR’s Response to upcoming OCI Solicitations Richard Loft SCD Deputy Director for R&D.
B1 -Biogeochemical ANL - Townhall V. Rao Kotamarthi.
Lecture 1: Introduction to High Performance Computing.
Moving Forward: NOAA & Earth Observation Systems Mr. Timothy R.E. Keeney Deputy Assistant Secretary of Commerce for Oceans and Atmosphere National Oceanographic.
4.x Performance Technology drivers – Exascale systems will consist of complex configurations with a huge number of potentially heterogeneous components.
Virtual Machine Hosting for Networked Clusters: Building the Foundations for “Autonomic” Orchestration Based on paper by Laura Grit, David Irwin, Aydan.
Earth Science Sec. 1.1 Introduction to Earth Science and Systems.
Statistical Performance Analysis for Scientific Applications Presentation at the XSEDE14 Conference Atlanta, GA Fei Xing Haihang You Charng-Da Lu July.
NCAR Supercomputing ‘Data Center’ Project An NCAR-led computing ‘facility’ for the study of Earth System Science.
CESD 1 SAGES Scottish Alliance for Geoscience, Environment & Society Observing and Modelling Climate Change Simon Tett, Chair of Earth System Dynamics.
NCAR Annual Budget Review October 8, 2007 Tim Killeen NCAR Director.
THIS IS With Host... Your A- Water on Earth B- Weather Tools C- Water Cycle D- Air Masses E- Teacher’s Choice.
Future role of DMR in Cyber Infrastructure D. Ceperley NCSA, University of Illinois Urbana-Champaign N.B. All views expressed are my own.
SIMPLE DOES NOT MEAN SLOW: PERFORMANCE BY WHAT MEASURE? 1 Customer experience & profit drive growth First flight: June, minute turn at the gate.
Extreme-scale computing systems – High performance computing systems Current No. 1 supercomputer Tianhe-2 at petaflops Pushing toward exa-scale computing.
Computer Science Section National Center for Atmospheric Research Department of Computer Science University of Colorado at Boulder Blue Gene Experience.
U.S. Department of the Interior U.S. Geological Survey U.S. Department of the Interior U.S. Geological Survey Natural Hazards Science – Reducing the World’s.
Programming Models & Runtime Systems Breakout Report MICS PI Meeting, June 27, 2002.
Models in Geoscience Education Cathy Manduca Director, Science Education Resource Center Carleton College, Minnesota.
Problem is to compute: f(latitude, longitude, elevation, time)  temperature, pressure, humidity, wind velocity Approach: –Discretize the.
NCAR Supercomputing ‘Data Center’ Project An NCAR-led computing ‘facility’ for the study of the Earth system May 30, 2006.
Software Engineering Committee Status Report: Preliminary Findings and Recommendations Richard Loft and Gerry Wiener SE Committee Co-chairs National Center.
Geosciences - Observations (Bob Wilhelmson) The geosciences in NSF’s world consists of atmospheric science, ocean science, and earth science Many of the.
Role of Integrated Assessment Modelling (IAM) in climate change policy analysis The Global Integrated Assessment Model (GIAM) An ABARE-CSIRO joint initiative.
Petascale –LLNL Appro AMD: 9K processors [today] –TJ Watson Blue Gene/L: 40K processors [today] –NY Blue Gene/L: 32K processors –ORNL Cray XT3/4 : 44K.
Co-Design 2013 Summary Exascale needs new architectures due to slowing of Dennard scaling (since 2004), multi/many core limits New programming models,
A lower bound to energy consumption of an exascale computer Luděk Kučera Charles University Prague, Czech Republic.
IDC HPC USER FORUM Weather & Climate PANEL September 2009 Broomfield, CO Panel questions: 1 response per question Limit length to 1 slide.
High performance parallel computing of climate models towards the Earth Simulator --- computing science activities at CRIEPI --- Yoshikatsu Yoshida and.
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
© 2009 IBM Corporation Motivation for HPC Innovation in the Coming Decade Dave Turek VP Deep Computing, IBM.
RECAPITALIZING THE NATION’S WEATHER PREDICTION CAPABILITY National Unified Operational Prediction Capability (NUOPC)
RDA Data Support Section. Topics 1.What is it? 2.Who cares? 3.Why does the RDA need CISL? 4.What is on the horizon?
CLIM Fall 2008 What are the Roles of Satellites & Supercomputers in Studying Weather and Climate? CLIM 101.
ESFRI & e-Infrastructure Collaborations, EGEE’09 Krzysztof Wrona September 21 st, 2009 European XFEL.
VAPoR: A Discovery Environment for Terascale Scientific Data Sets Alan Norton & John Clyne National Center for Atmospheric Research Scientific Computing.
11 January 2005 High Performance Computing at NCAR Tom Bettge Deputy Director Scientific Computing Division National Center for Atmospheric Research Boulder,
CCSM Performance, Successes and Challenges Tony Craig NCAR RIST Meeting March 12-14, 2002 Boulder, Colorado, USA.
NCAR Supercomputing ‘Data Center’ Project An NCAR-led computing ‘facility’ for the study of the Earth system.
Exascale climate modeling 24th International Conference on Parallel Architectures and Compilation Techniques October 18, 2015 Michael F. Wehner Lawrence.
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
February 25, 2008 The Emerging Front Range HPC Collaboratory Dr. Rich Loft: Director, Technology Development Computational.
1 Accomplishments. 2 Overview of Accomplishments  Sustaining the Production Earth System Grid Serving the current needs of the climate modeling community.
Status and plans at KEK Shoji Hashimoto Workshop on LQCD Software for Blue Gene/L, Boston University, Jan. 27, 2006.
Exscale – when will it happen? William Kramer National Center for Supercomputing Applications.
NCAR RP Update Rich Loft NCAR RPPI May 7, NCAR Teragrid RP Developments Current Cyberinfrastructure –5.7 TFlops/2048 core Blue Gene/L system –100.
Power and Cooling at Texas Advanced Computing Center Tommy Minyard, Ph.D. Director of Advanced Computing Systems 42 nd HPC User Forum September 8, 2011.
1 Cyber-Enabled Discovery and Innovation Michael Foster May 11, 2007.
Power and Cooling of HPC Data Centers Requirements Roger A Panton Avetec Executive Director DICE
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Data Requirements for Climate and Carbon Research John Drake, Climate Dynamics Group Computer.
Data Discovery and Access to The International Surface Pressure Databank (ISPD) 1 Thomas Cram Gilbert P. Compo* Doug Schuster Chesley McColl* Steven Worley.
Next Generation Data and Computing Facility NCAR “Data Center” project An NCAR-led computing ‘facility’ for the study of the Earth system.
RDA Data Support Section. Topics 1.What is it? 2.Who cares? 3.Why does the RDA need CISL? 4.What is on the horizon?
Sensors and Instrumentation Computational and Data Challenges in Environmental Modelling Dr Peter M Allan Director, Hartree Centre, STFC.
HPC Trends and Challenges in Climate Research Prof. Dr. Thomas Ludwig German Climate Computing Centre & University of Hamburg Hamburg, Germany
1 Data Warehouse Assessments What, Why, and How Noah Subrin Technical Lead SRA International April 24, 2010.
Extreme Scale Infrastructure
THIS IS Jeopardy. THIS IS Jeopardy With Your Host... Mrs. Walsh.
Parallel computer architecture classification
American Meteorological Society
Presentation transcript:

Petascale System Requirements for the Geosciences Richard Loft SCD Deputy Director for R&D

Petascale Collaboratory Reports l Charges: –Identify science drivers for petascale geoscience computing. (Bryan) –Determine technical feasibility and cost of creating such a system by (Loft) l Outcomes : –Collaboratory is in NSF Facility Plan as a facility “under study”. –Initiative to bring vendors and scientists together to discuss petascale geoscience computing

Petascale Geoscience l World class computational capability should be devoted to understanding the earth. –Earthquakes & Tsunamis –Severe weather - Katrina –Global warming –Land use policies l It doesn’t stop at the 1 PFLOPS. l Easy Part: Achieving ~ 1 PFLOPS peak –Nothing exceptional has to happen technologically to do it –Possible by 2008, first by 2010, multiple by 2012 –Highly parallel system - speed of single thread slowing. l Hard Part: Infrastructure. –Facilities to house them. –Efficient applications. –Systems and tools to help us understand the data they produce.

Geoscience really does need Petascale Computing (and Beyond)

Geocience Space Weather Turbulence Atmospheric Chemistry Climate Weather The Sun from the sun’s surface to the earth’s core…

Directly Resolving Weather Phenomena

Weather Research Forecast Model: Katrina Landfall: Aug 29, 2005 After W. Skamarock

WRF Model Performance 9.8% of Peak

Estimated Global NWP Requirements

Petascale Cat Herding Application Experts System Architects Software Architects Numericists geoscience Vendor Sales $ COTS CMOS Open Source Matlab ?

Data Driver l A Petascale system will produce 100 PB of geoscience data per year. l That’s 3.2 GB/s 7x24x365 l Hard: Mining that data torrent for science nuggets. l How do we drive petascale requirements for those DAV systems?

Taming the Torrent: Visualization Example l Key concepts 1.Integrate visualization into analysis process 2.interactively steer the analysis to subsets 3.Employ multiresolution data representation as a data reduction technique Size: Full 1/8 1/64 1/512

Facility Driver Uncertainty in fuel efficiency and power density leads to uncertainty petascale facility planning. You can’t turn a facility on a dime. Efficiency is good. How do we craft power-efficiency benchmarks?

Petascale TCO: Fuel Efficiency Top 20 systems Based on processor power rating only Blue Gene/L: How do we use this stuff?

Petascale Data Center Design Petascale estimate: 2-5 MW by 2010 Reliable Tier 2+ Redundant power/cooling systems Modular design Expandable. Mitigates risk. Floor Space (per module): 20,000 sq. ft. machine room 20,000 sq. ft. mechanical space Power (per module) 4 MW systems plus 4 MW mechanical Land: 13 Acres. A second order term in cost.

Application Driver: Collecting Geoscience Application Requirements How do we extract system requirements for systems that haven’t been built from applications that haven’t been written? No single set of requirements in fact exists. There are very few good models of application performance. Application experts intuition is often wrong. How should applications benchmarks be selected?

1/10 Degree POP Ocean Model Credit: Frank Bryan, NCAR Agulhas Rings

POP X1 Performance 6x: System Architecture Does Matter! After Worley, et al.

POP Sensitivity to System Parameters

1 Degree Barotropic Ocean Model Sensitivity to Latency and Bandwidth (Performance Model) On 4096 processors…

3D pseudospectral code (2048^3) 1.6 GFLOPS sustained 6 µsec latency

Forget the Top 500 list! Rmax ranking vs application reality Are the HPC Challenge Benchmarks good metrics?

HPC Challenge: Flops Results

HPC Challenge: Bandwidth Results

HPC Challenge: Latency/STREAM Results

HPC Challenge: Local/Remote Balance

Discussion - Open Forum How do we drive petascale requirements for DAV systems? How do we extract system requirements for systems that haven’t been built from applications that haven’t been written? How should applications benchmarks be selected? Are the HPC Challenge Benchmarks good metrics? How do we craft power-efficiency benchmarks?

The End

Ready or not, the petascale is coming Proof by extrapolation…

Growth Rate Results - Top 10 Geoscience Centers on the Top500 List l Individual sites come and go. l Individual sites rise and fall. l Collectively, things are smoother. l Annual rate 1st ranked: l Annual rate 1st ranked: l Annual mean growth rate: l Annual rate 10th ranked: l Currently –5 use IBM POWER processors –2 use vector systems –3 use other l Pack is slowly converging. l Use this rate to chart computing future… Log linear fit (mean)

March Towards the Petascale

Extrapolated Petascale System Power requirements … NCAR Facility

Additional Power and Storage Costs Wildcard factors: Katrina, Enron, etc.

Petascale Budget Breakdown ( )