Possible foreseeable measures for tera-scale data handling Kazutoshi Horiuchi *1 Keiko Takahashi *1 Hirofumi Sakuma *1 Shigemune Kitawaki *2 *1 Frontier.

Slides:



Advertisements
Similar presentations
WHAT IS AN OPERATING SYSTEM? An interface between users and hardware - an environment "architecture ” Allows convenient usage; hides the tedious stuff.
Advertisements

Deep Tropical Convection contribution to climate change.
OPERATING SYSTEM OVERVIEW
Earth Simulator Jari Halla-aho Pekka Keränen. Architecture MIMD type distributed memory 640 Nodes, 8 vector processors each. 16GB shared memory per node.
Introduction and Overview Questions answered in this lecture: What is an operating system? How have operating systems evolved? Why study operating systems?
Operating System Review September 10, 2012Introduction to Computer Security ©2004 Matt Bishop Slide #1-1.
Virtualization. Virtualization  In computing, virtualization is a broad term that refers to the abstraction of computer resources  It is "a technique.
1 Interconnects Shared address space and message passing computers can be constructed by connecting processors and memory unit using a variety of interconnection.
Seaborg Cerise Wuthrich CMPS Seaborg  Manufactured by IBM  Distributed Memory Parallel Supercomputer  Based on IBM’s SP RS/6000 Architecture.
A Metadata Based Approach For Supporting Subsetting Queries Over Parallel HDF5 Datasets Vignesh Santhanagopalan Graduate Student Department Of CSE.
© Janice Regan, CMPT 300, May CMPT 300 Introduction to Operating Systems Principles of I/0 hardware.
A RISC ARCHITECTURE EXTENDED BY AN EFFICIENT TIGHTLY COUPLED RECONFIGURABLE UNIT Nikolaos Vassiliadis N. Kavvadias, G. Theodoridis, S. Nikolaidis Section.
Active Storage and Its Applications Jarek Nieplocha, Juan Piernas-Canovas Pacific Northwest National Laboratory 2007 Scientific Data Management All Hands.
CMAQ Runtime Performance as Affected by Number of Processors and NFS Writes Patricia A. Bresnahan, a * Ahmed Ibrahim b, Jesse Bash a and David Miller a.
System Architecture Directions for Networked Sensors Jason Hill, Robert Szewczyk, Alec Woo, Seth Hollar, David Culler, Kris Pister Presented by Yang Zhao.
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Facilities and How They Are Used ORNL/Probe Randy Burris Dan Million – facility administrator.
High performance parallel computing of climate models towards the Earth Simulator --- computing science activities at CRIEPI --- Yoshikatsu Yoshida and.
CS 591 x I/O in MPI. MPI exists as many different implementations MPI implementations are based on MPI standards MPI standards are developed and maintained.
Multilevel Parallelism using Processor Groups Bruce Palmer Jarek Nieplocha, Manoj Kumar Krishnan, Vinod Tipparaju Pacific Northwest National Laboratory.
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
CCGrid, 2012 Supporting User Defined Subsetting and Aggregation over Parallel NetCDF Datasets Yu Su and Gagan Agrawal Department of Computer Science and.
VAPoR: A Discovery Environment for Terascale Scientific Data Sets Alan Norton & John Clyne National Center for Atmospheric Research Scientific Computing.
Modeling Big Data Execution speed limited by: –Model complexity –Software Efficiency –Spatial and temporal extent and resolution –Data size & access speed.
ATmospheric, Meteorological, and Environmental Technologies RAMS Parallel Processing Techniques.
The impacts of climate change on global hydrology and water resources Simon Gosling and Nigel Arnell, Walker Institute for Climate System Research, University.
Data Management for Decision Support Session-4 Prof. Bharat Bhasker.
1 THE EARTH SIMULATOR SYSTEM By: Shinichi HABATA, Mitsuo YOKOKAWA, Shigemune KITAWAKI Presented by: Anisha Thonour.
CCSM Performance, Successes and Challenges Tony Craig NCAR RIST Meeting March 12-14, 2002 Boulder, Colorado, USA.
Welcome to the PRECIS training workshop
Presented by LCF Climate Science Computational End Station James B. White III (Trey) Scientific Computing National Center for Computational Sciences Oak.
CS4315A. Berrached:CMS:UHD1 Introduction to Operating Systems Chapter 1.
Parallel IO for Cluster Computing Tran, Van Hoai.
Tackling I/O Issues 1 David Race 16 March 2010.
LIOProf: Exposing Lustre File System Behavior for I/O Middleware
07/07/ KDD Services at the Goddard DAAC Robert Mack NASA/Goddard Space Flight Center Distributed Active Archive Center.
© Thomas Ludwig Prof. Dr. Thomas Ludwig German Climate Computing Center (DKRZ) University of Hamburg, Department for Computer Science (UHH/FBI) Disks,
Wolfgang Runte Slide University of Osnabrueck, Software Engineering Research Group Wolfgang Runte Software Engineering Research Group Institute.
Compute and Storage For the Farm at Jlab
Towards development of a Regional Arctic Climate System Model ---
Operating System Overview
Modeling Big Data Execution speed limited by: Model complexity
Lecture 1: Operating System Services
WP18, High-speed data recording Krzysztof Wrona, European XFEL
Chapter 1: Introduction
Network Operating Systems (NOS)
Basic Computer Hardware and Software.
GdX - Grid eXplorer parXXL: A Fine Grained Development Environment on Coarse Grained Architectures PARA 2006 – UMEǺ Jens Gustedt - Stéphane Vialle - Amelia.
Overview of Earth Simulator.
Lecture 1: Network Operating Systems (NOS)
Grid Computing.
Spatial Analysis With Big Data
Introduction to Operating System (OS)
Chapter 1: Introduction
Quick Introduction to OS
Computer Architecture
Chapter 1: Introduction
Basic Computer Hardware and Software.
Invitation to Computer Science 5th Edition
Chapter 17 Parallel Processing
Five Key Computer Components
Language Processors Application Domain – ideas concerning the behavior of a software. Execution Domain – Ideas implemented in Computer System. Semantic.
TeraScale Supernova Initiative
OVERVIEW OF FINITE ELEMENT METHOD
Chapter 2 Operating System Overview
Chapter 1: Introduction
Emulator of Cosmological Simulation for Initial Parameters Study
Data Curation in Climate and Weather
Introduction and Overview
L. Glimcher, R. Jin, G. Agrawal Presented by: Leo Glimcher
Presentation transcript:

Possible foreseeable measures for tera-scale data handling Kazutoshi Horiuchi *1 Keiko Takahashi *1 Hirofumi Sakuma *1 Shigemune Kitawaki *2 *1 Frontier Research System for Global Change *2 Earth Simulator Research and Development Center

Global Change Prediction by an Integrated Three-in-one Research Observation Numerical Simulation Process Study & Modeling Accurate & spatially representative data Optimal monitoring plan Sophisticated high resolution model High performance computing Assimilation data for validation Accurate & spatially representative data ESRDC FRSGC FORSGC

Frontier Research System for Global Change (Project) Funding Bodies Japan Marine Science and Technology Center (JAMSTEC) National Space Development Agency (NASDA) Activities Process Study Model Development (Common) Goal Global Change Prediction

On-Going Process Studies §Climate Variations Research §Hydrological Cycle Research §Global Warming Research §Atmospheric Composition Research §Ecosystem Change Research §Research of International Pacific Research Center §Research of International Arctic Research Center

Current Target of Model Development Group Coupled Model (now based on CCSR/NIES, MOM3) for Climate Change Experiment Cloud Cluster Resolving Ultra High Resolution Model for Prediction of Typhoon/Baiu Evolution Coupled Chemistry - Global Climate Model for Prediction of Atmospheric Composition Change Next Generation Model (Cubic/Icosahedral Grid, CIP method) 4DVAR Ocean Data Assimilation Model (based on MOM3)

Current Target of Coupled Model Development on ES §High Resolution: l Atmosphere Model part: T213L50 l Ocean Model part: 1/10 deg. 53 layers §High Performance Estimation of Acceleration ratio Atmosphere Model: (under estimation) Ocean Model: 300 ~ 400 times (480PEs;60Ns) 5 days for 100 years integration

Earth Simulator Research Development Center (Project) Funding Bodies Japan Atomic Energy Research Institute (JAERI) National Space Development Agency (NASDA) Japan Marine Science and Technology Center (JAMSTEC) Activities Development of High Speed Parallel Computer Understanding and Prediction of Global Change (Common) Goal Global Change Prediction

Characteristics of Earth Simulator §Peak Performance: 40TFLOPS §Number of Processor Nodes: 640 §Number of PEs: 5120 (8PEs/Node) §Interconnection Network: 16GB/s §Total Memory: 10TB §Total Secondary Storage: TB §Total Mass Storage: 1PB (84Drives)

16 Nodes Architecture of Earth Simulator 16 Nodes Interconnection Network TSS Cluster *1Batch Cluster *39 … Fiber Channel Switch … MM VP 0 VP 1VP 7 16 Nodes … MM VP 0 VP 1VP 7 … MM VP 0 VP 1VP 7 … … Fiber Channel Switch … MM VP 0 VP 1VP 7 16 Nodes … MM VP 0 VP 1VP 7 … MM VP 0 VP 1VP 7 … Mass Storage System WAN WS FS … 84 Drives …

I/O Model in Distributed Memory Parallel Computer PPP … communication PPP … PPP … (Unix) File(Data-Distributed Unix) Files Parallel File

Parallel File System on ES To handle distributed data as a logically single file has advantages to develop application softwares and to process post processings M1M1 M2M2 8 9 File Image P P P Processors Disks Proc no.: Np Dist. Size: Sd Dist. Pattern: P= BLOCK/CYCLIC Disk no.: Nd Striping Size: Ss Distribution Mechanism D1D1 D2D2 D3D3 P1P1 P2P2 P3P

Support for Parallel File on Several Levels …… Unix FileParallel File Operating System (with PFS) FAL MPI-IO F90 HPF-RTP F90-RTP Library Compiler C User Program PFSUFSPFSUFSPFS Hardware HPF

Review of Model Development Flow Improvement of Model Execution with Model Evaluation of Results Input Data Output Data Resources for Process Study Results of Process Study Analysis Visualization

Is it Satisfactory about I/O performance? §The faster super computers are, the larger the amount of the output data generated by large-scale simulations. §The large amount of data is stored to secondary storages and/or mass storages whose devices are slower. §Is it satisfactory about I/O performance ?

Amount of Input/Output Data - Coupled Model §To answer the question, the following cases are investigated. l Case I: 1000 Years Integration for the Prediction of Global Warming, Decadal Variability, etc. l Case II: 50 Years Integration for the Analysis of El nino, Dipole Mode Events, Asian Monsoon, etc.

Amount of Input/Output Data - Coupled Model §Atmosphere Model part NOTE: The amount of output data is estimated as 2 byte integer elements §Ocean Model part

Estimated I/O time - Coupled Model NOTE: Time is estimated only on drive’s I/O rate. Multiple drives are assumed to be independent.

Summary of I/O Performance (from the viewpoint of Model Development) §Disk I/O time might be satisfactory. l 0.2% of the simulation time l Less than 2 hours as a total §Tape I/O time might be conspicuous. l 11-35% of the simulation time for 8 tape drives l 1-6 days for 8 tape drives This inefficiency might be critical for iterative works such as model development

How to Shorten Turn Around Time of Model Development §Give up outputting numerical data. §Output necessary minimum data. §Output full data, with executing tape I/O and simulations in parallel, and with tape I/O library being able to extract necessary minimum data for post processing.

Shortening of TAT by Giving up Outputting Numerical Data - - Elapsed time (h) Time increase (%) Elapsed time (h) Time increase (%) CFD solver only + fixed camera + moving camera 1.6M grid (169x92x101)6.2M grid (337x183x101) Elapsed time for the concurrent visualization with RVSLIB in the batch processing mode on SX-4 *The number of computational time steps was *Contour and tracers were displayed at every 10 time steps and visualized animation was stored in a file. *Time integration for moving the tracers was done at every time step for greater accuracy. * This result was provided by NEC

Shortening of TAT by Outputting Necessary Minimum Data §“Browse sets”, into which the large amount of output data is abstracted (spatially and/or temporally) within simulations, should be stored. §Specific regions of output data should be stored. This may be Know-Hows of using ES

Shortening of TAT by Enhancement of Tape I/O Library for Full Output Data §Tape I/O should be executed with simulations in parallel. §In mass storages, output data should be re- organized, and small subsets which are needed for post processing should be able to be extracted. This may be requirements for the improvement of ES

Conclusion §I/O performance was roughly estimated and I/O problem was apprehended. §I/O problem would be avoided with concurrent visualization and/or know-how of the usage. §However we would like to examine the efficient technique for handling the large amount of data continuously to realize comfortable environment for global change prediction.