HPC Trends and Challenges in Climate Research Prof. Dr. Thomas Ludwig German Climate Computing Centre & University of Hamburg Hamburg, Germany

Slides:



Advertisements
Similar presentations
Large Scale Computing Systems
Advertisements

Program Analysis and Tuning The German High Performance Computing Centre for Climate and Earth System Research Panagiotis Adamidis.
Preservation and Long Term Access of Data at the World Data Centre for Climate Frank Toussaint N.P. Drakenberg, H. Höck, M. Lautenschlager, H. Luthardt,
Priority Research Direction (I/O Models, Abstractions and Software) Key challenges What will you do to address the challenges? – Develop newer I/O models.
Supercomputing Challenges at the National Center for Atmospheric Research Dr. Richard Loft Computational Science Section Scientific Computing Division.
One-day Meeting, INI, September 26th, 2008 Role of spectral turbulence simulations in developing HPC systems YOKOKAWA, Mitsuo Next-Generation Supercomputer.
Parallel Research at Illinois Parallel Everywhere
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA; SAN DIEGO SDSC RP Update October 21, 2010.
Petascale System Requirements for the Geosciences Richard Loft SCD Deputy Director for R&D.
Claude TADONKI Mines ParisTech – LAL / CNRS / INP 2 P 3 University of Oujda (Morocco) – October 7, 2011 High Performance Computing Challenges and Trends.
KIT – The cooperation of Forschungszentrum Karlsruhe GmbH und Universität Karlsruhe (TH) SX-9, a good Choice ? Steinbuch Centre for Computing.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA; SAN DIEGO IEEE Symposium of Massive Storage Systems, May 3-5, 2010 Data-Intensive Solutions.
UNCLASSIFIED: LA-UR Data Infrastructure for Massive Scientific Visualization and Analysis James Ahrens & Christopher Mitchell Los Alamos National.
Hitachi SR8000 Supercomputer LAPPEENRANTA UNIVERSITY OF TECHNOLOGY Department of Information Technology Introduction to Parallel Computing Group.
Lecture 1: Introduction to High Performance Computing.
1 The development of modern computer systems Early electronic computers Mainframes Time sharing Microcomputers Networked computing.
1 Building National Cyberinfrastructure Alan Blatecky Office of Cyberinfrastructure EPSCoR Meeting May 21,
Future role of DMR in Cyber Infrastructure D. Ceperley NCSA, University of Illinois Urbana-Champaign N.B. All views expressed are my own.
Preserving the Scientific Record: Preserving a Record of Environmental Change Matthew Mayernik National Center for Atmospheric Research Version 1.0 [Review.
VO Sandpit, November 2009 e-Infrastructure to enable EO and Climate Science Dr Victoria Bennett Centre for Environmental Data Archival (CEDA)
Computer Science Section National Center for Atmospheric Research Department of Computer Science University of Colorado at Boulder Blue Gene Experience.
9-Sept-2003CAS2003, Annecy, France, WFS1 Distributed Data Management at DKRZ Distributed Data Management at DKRZ Wolfgang Sell Hartmut Fichtel Deutsches.
Directed Reading 2 Key issues for the future of Software and Hardware for large scale Parallel Computing and the approaches to address these. Submitted.
1 Distributed Energy-Efficient Scheduling for Data-Intensive Applications with Deadline Constraints on Data Grids Cong Liu and Xiao Qin Auburn University.
1 Application Scalability and High Productivity Computing Nicholas J Wright John Shalf Harvey Wasserman Advanced Technologies Group NERSC/LBNL.
System Level Science and System Level Models Ian Foster Argonne National Laboratory University of Chicago Improving IAM Representations of a Science-Driven.
Headline in Arial Bold 30pt HPC User Forum, April 2008 John Hesterberg HPC OS Directions and Requirements.
October 12, 2004Thomas Sterling - Caltech & JPL 1 Roadmap and Change How Much and How Fast Thomas Sterling California Institute of Technology and NASA.
Experts in numerical algorithms and High Performance Computing services Challenges of the exponential increase in data Andrew Jones March 2010 SOS14.
VO Sandpit, November 2009 e-Infrastructure for Climate and Atmospheric Science Research Dr Matt Pritchard Centre for Environmental Data Archival (CEDA)
© 2009 IBM Corporation Motivation for HPC Innovation in the Coming Decade Dave Turek VP Deep Computing, IBM.
CLIM Fall 2008 What are the Roles of Satellites & Supercomputers in Studying Weather and Climate? CLIM 101.
ESFRI & e-Infrastructure Collaborations, EGEE’09 Krzysztof Wrona September 21 st, 2009 European XFEL.
© Copyright 2012 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice. HP Update IDC HPC Forum.
1 THE EARTH SIMULATOR SYSTEM By: Shinichi HABATA, Mitsuo YOKOKAWA, Shigemune KITAWAKI Presented by: Anisha Thonour.
FP7 Support Action - European Exascale Software Initiative DG Information Society and the unit e-Infrastructures EESI Final Conference Exascale key hardware.
Exascale climate modeling 24th International Conference on Parallel Architectures and Compilation Techniques October 18, 2015 Michael F. Wehner Lawrence.
B5: Exascale Hardware. Capability Requirements Several different requirements –Exaflops/Exascale single application –Ensembles of Petaflop apps requiring.
Smart cloud orchestrator - the first implementation in the world at Wroclaw University of Technology for supporting design processes in education at universities.
Exscale – when will it happen? William Kramer National Center for Supercomputing Applications.
Power and Cooling at Texas Advanced Computing Center Tommy Minyard, Ph.D. Director of Advanced Computing Systems 42 nd HPC User Forum September 8, 2011.
DiRAC-3 – The future Jeremy Yates, STFC DiRAC HPC Facility.
Lenovo - Eficiencia Energética en Sistemas de Supercomputación Miguel Terol Palencia Arquitecto HPC LENOVO.
Data Infrastructure foundation of the European Cloud Initiative
Support to scientific research on seasonal-to-decadal climate and air quality modelling Pierre-Antoine Bretonnière Francesco Benincasa IC3-BSC - Spain.
Tackling I/O Issues 1 David Race 16 March 2010.
The Worldwide LHC Computing Grid Frédéric Hemmer IT Department Head Visit of INTEL ISEF CERN Special Award Winners 2012 Thursday, 21 st June 2012.
NASA Earth Exchange (NEX) A collaborative supercomputing environment for global change science Earth Science Division/NASA Advanced Supercomputing (NAS)
CAS2K11 in Annecy, France September 11 – 14, 2011 Data Infrastructures at DKRZ Michael Lautenschlager.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
NASA Earth Exchange (NEX) Earth Science Division/NASA Advanced Supercomputing (NAS) Ames Research Center.
IT-DSS Alberto Pace2 ? Detecting particles (experiments) Accelerating particle beams Large-scale computing (Analysis) Discovery We are here The mission.
The Evolution of the Italian HPC Infrastructure Carlo Cavazzoni CINECA – Supercomputing Application & Innovation 31 Marzo 2015.
Societal applications of large scalable parallel computing systems ARTEMIS & ITEA Co-summit, Madrid, October 30th 2009.
Design and Planning Tools John Grosh Lawrence Livermore National Laboratory April 2016.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
Computing infrastructures for the LHC: current status and challenges of the High Luminosity LHC future Worldwide LHC Computing Grid (WLCG): Distributed.
Defining the Competencies for Leadership- Class Computing Education and Training Steven I. Gordon and Judith D. Gardiner August 3, 2010.
© Thomas Ludwig Prof. Dr. Thomas Ludwig German Climate Computing Center (DKRZ) University of Hamburg, Department for Computer Science (UHH/FBI) Disks,
Energy efficient SCalable
DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING CLOUD COMPUTING
ALICE Computing Model in Run3
Course Description: Parallel Computer Architecture
SDP Kernels Workshop – The Role of Kernels
Chapter 1 Introduction.
TeraScale Supernova Initiative
Coordinator: DKRZ Weather Climate HPC
Panel on Research Challenges in Big Data
What will engineering design practice be like in 2040
Presentation transcript:

HPC Trends and Challenges in Climate Research Prof. Dr. Thomas Ludwig German Climate Computing Centre & University of Hamburg Hamburg, Germany

2 / 42 © DKRZ | more FLOPS, more bytes more science ?! more knowledge and insight ?! more money !!! HP-CAST 18, Hamburg factor 1000 ! factor 100 ? factor 50 ? factor 10 ? A First Cost-Benefit Analysis

3 / 42 © DKRZ | What´s going on in HPC ? HP-CAST 18, Hamburg

4 / 42 © DKRZ | Evolution Curve HP-CAST 18, Hamburg factor 1000 of performance increase every 11 years

5 / 42 © DKRZ | HP-CAST 18, Hamburg MFlops 1KFlops 1Flops 0,001Flops 1GFlops 1TFlops 1PFlops factor 1000 of performance increase every 11 years

6 / 42 © DKRZ | HP-CAST 18, Hamburg MFlops 1KFlops 1Flops 0,001Flops 1GFlops Zuse´s Z3 – 0.3 Flops (1941) 1TFlops 1PFlops

7 / 42 © DKRZ | HP-CAST 18, Hamburg MFlops 1KFlops 1Flops 0,001Flops 1GFlops Zuse´s Z3 – 0.3 Flops (1941) 1TFlops 1PFlops factor 1000 of performance increase every 11 years

8 / 42 © DKRZ | HP-CAST 18, Hamburg MFlops 1KFlops 1Flops 0,001Flops 1GFlops Zuse´s Z3 – 0.3 Flops (1941) 1TFlops 1PFlops 1EFlops 1ZFlops 2030 (?) Hochleistungsrechnen - © Thomas Ludwig factor 1000 of performance increase every 11 years

9 / 42 © DKRZ | What´s going on in climate research ? HP-CAST 18, Hamburg

10 / 42 © DKRZ | The Big Bang of Climate Science HP-CAST 18, Hamburg Picture stolen from Rory Kelly (NCAR) 1956

11 / 42 © DKRZ | GCRM – Global Cloud Resolving Model Finally Cloud Computing HP-CAST 18, Hamburg

12 / 42 © DKRZ | DKRZ today and tomorrow HP-CAST 18, Hamburg

13 / 42 © DKRZ | DKRZ – to provide high performance computing platforms, sophisticated and high capacity data management, and superior service for premium climate science. Mission HP-CAST 18, Hamburg

14 / 42 © DKRZ | IBM Power6, installed 2009 Peak performance: 158 TFLOPS, Linpack 115 TFLOPS Rank 98 in Nov. 2011´s TOP500 list 264 IBM Power6 compute nodes with processor cores More than 26 TB main memory 6+ PB disk space air cond. compute system disk drives air cond. Compute System HP-CAST 18, Hamburg

15 / 42 © DKRZ | 7 Sun StorageTek SL8500 tape libraries slots for tapes 78 tape drives 100+ Petabyte total capacity Data Archive HP-CAST 18, Hamburg

16 / 42 © DKRZ | HP ProLiant DL370 G6 HP XW 9400, HP ProLiant DL585 HP Remote Graphics 10 GE connection to the GPFS disk storage Visualization System HP-CAST 18, Hamburg

17 / 42 © DKRZ | Present, Future, and Exa HP-CAST 18, Hamburg DKRZ 2012 Linpack110 TFLOPS Main memory26 TB Disk space7 PB Tape library100 PB Memory-to-disk30 GB/s Disk-to-tape3 GB/s Application-to-disk too slow

18 / 42 © DKRZ | Present, Future, and Exa HP-CAST 18, Hamburg DKRZ 2012x Linpack110 TFLOPS1 EFLOPS Main memory26 TB260 PB Disk space7 PB70 EB Tape library100 PB1 ZB Memory-to-disk30 GB/s300 TB/s Disk-to-tape3 GB/s30 TB/s Application-to-disk too slow

19 / 42 © DKRZ | Present, Future, and Exa HP-CAST 18, Hamburg DKRZ 2012x Exa Linpack110 TFLOPS1 EFLOPS Main memory26 TB260 PB32-64 PB Disk space7 PB70 EB0.5-1 EB Tape library100 PB1 ZB? Memory-to-disk30 GB/s300 TB/s50 GB/s Disk-to-tape3 GB/s30 TB/s? Application-to-disk too slow

20 / 42 © DKRZ | A “regular” Exascale system as proposed in the documents will not fit typical memory and I/O intensive climate applications Currently, even Petascale systems are not a success story for climate research Observations HP-CAST 18, Hamburg

21 / 42 © DKRZ | Is there a revolution in HPC ? HP-CAST 18, Hamburg

22 / 42 © DKRZ | HP-CAST 18, Hamburg Exascale

23 / 42 © DKRZ | From GFLOPS to TFLOPS was a revolution – From vector machines to massively parallel systems with message passing – Required new algorithms and software From TFLOPS to PFLOPS was more of an evolution – MPI/OpenMP/C/C++/Fortran: just more of it From PFLOPS to EFLOPS will again be a revolution – Clock rate is constant – number of cores increases – Energy efficiency with special accelerator hardware – Resilience will be a big problem –... Exascale Revolution HP-CAST 18, Hamburg

24 / 42 © DKRZ | Hardware challenge – Billions of components The K computer has already around 800 racks ! – Energy consumption – Size of the system memory – System resilience Co-design challenge – Simple rule-of-thumb design choices must be replaced by concepts of co-design to balance design parameters – Optimize trade-offs among applications, software and hardware Exascale Challenges HP-CAST 18, Hamburg

25 / 42 © DKRZ | The applied mathematics challenge – It is not enough to just make the grid finer – Reproducibility of results is also an issue The algorithmic challenge – New multicore-friendly algorithms – Adaptive load balancing – Multiple precision software – Auto-tuning – Integrated fault-tolerance – Energy efficient algorithms Exascale Challenges HP-CAST 18, Hamburg

26 / 42 © DKRZ | The computer science challenge – Programming models – I/O – Tools for debugging and performance analysis The educational challenge – Who will be able to successfully work on these issues? – Who will teach these people? – How many teachers? Exascale Challenges HP-CAST 18, Hamburg

27 / 42 © DKRZ | Good news: We have time until 2018/19 Bad news: We need many more people ! Exascale Observations HP-CAST 18, Hamburg

28 / 42 © DKRZ | Is there a revolution in climate research ? HP-CAST 18, Hamburg

29 / 42 © DKRZ | HP-CAST 18, Hamburg Big Data Wrong fortress, folks!!

30 / 42 © DKRZ | Refer to Jim Gray et al.: “The Fourth Paradigm – Data-Intensive Scientific Discovery” „Data are not consumed by the ideas and innovations they spark but are an endless fuel for creativity“ „Harnessing the Power of Digital Data for Science and Society“ by the Interagency Working Group on Digital Data January 2009 The Fourth Paradigm HP-CAST 18, Hamburg

31 / 42 © DKRZ | Climate research was and is dependent on data – Keep data for long term in order to validate models in the future – Keep all model variables because we do not know what will be necessary in future analyses – Use a fine global grid in order to select later local regions –... DKRZ will archive in 2012 around 8 PByte on tape – This is all output data of simulation runs Climate Research and Data HP-CAST 18, Hamburg

32 / 42 © DKRZ | The 5th coupled model intercomparison project – Provides key input for the next IPCC report – ~20 modeling centers around the world (DKRZ being one of the biggest) – Produces 10s of PBytes of output data from ~60 experiments (“digital born data”) Data are produced without knowing all applications beforehand and these data are projected for interdisciplinary utilization by yet unknown researchers Example: CMIP HP-CAST 18, Hamburg

33 / 42 © DKRZ | Data creation – Model runs in HPC centers (little I big O) Data evaluation – Processing, visualization, sharing, quality control Data archival – Documentation, long term preservation Data dissemination – GUI based catalogue and data access – API based data access – Data federations Data Life Cycle Management HP-CAST 18, Hamburg

34 / 42 © DKRZ | Data creation – German contribution produced at DKRZ – Used ½ year of its compute capacity Data archival – 600 TB raw + 60 TB with quality control Data dissemination – Will be heavily used during the next 5 years Expected CMIP6 data volume: factor higher Example Workflow CMIP HP-CAST 18, Hamburg

35 / 42 © DKRZ | Bad news: We need more people ! Very bad news: We need solutions NOW ! Big Data Observations HP-CAST 18, Hamburg

36 / 42 © DKRZ | From coevolution to corevolution ? HP-CAST 18, Hamburg

37 / 42 © DKRZ | Wikipedia on technical coevolution: “Computer software and hardware can be considered as two separate components but tied intrinsically by coevolution” E.g. numerical weather prediction (NWP) – Started in 1950 – Hardware: ENIAC (385 FLOPS) – Brainware: John von Neumann (appl. maths) E.g. climate models – Started in 1956 by Norman Philips Coevolution HP-CAST 18, Hamburg

38 / 42 © DKRZ | HP-CAST 18, Hamburg Corevolution Data intensive pre-Exascale HPC Wikipedia on technical corevolution: DKRZ will join the army computer scientistsclimate researchers nothing

39 / 42 © DKRZ | Come to see us at ISC´12 booth #140 and learn more about Data life cycle management Massive data I/O DKRZ´s Role HP-CAST 18, Hamburg

40 / 42 © DKRZ | Observations and Conclusions HP-CAST 18, Hamburg

41 / 42 © DKRZ | We need co-design, coevolution, corevolution Future sustained development will heavily depend on an intensified cooperation between hardware, software and application experts Access to human resources will limit our growth Co-Whatsoever HP-CAST 18, Hamburg

42 / 42 © DKRZ | Richardson´s vision in 1916 Perhaps some day in the dim future it will be possible to advance computations faster than the weather advances and at a cost less than the saving to mankind due to the information gained (Weather Prediction by Numerical Process) A Final Cost-Benefit Analysis HP-CAST 18, Hamburg