9-Sept-2003CAS2003, Annecy, France, WFS1 Distributed Data Management at DKRZ Distributed Data Management at DKRZ Wolfgang Sell Hartmut Fichtel Deutsches.

Slides:



Advertisements
Similar presentations
Data Storage Solutions Module 1.2. Data Storage Solutions Upon completion of this module, you will be able to: List the common storage media and solutions.
Advertisements

Introduction to Storage Area Network (SAN) Jie Feng Winter 2001.
Cloud Computing: Theirs, Mine and Ours Belinda G. Watkins, VP EIS - Network Computing FedEx Services March 11, 2011.
Ver 0.1 Page 1 SGI Proprietary Introducing the CRAY SV1 CRAY SV1-128 SuperCluster.
VERITAS Software Corp. BUSINESS WITHOUT INTERRUPTION Fredy Nick SE Manager.
Scale-out Central Store. Conventional Storage Verses Scale Out Clustered Storage Conventional Storage Scale Out Clustered Storage Faster……………………………………………….
Silicon Graphics, Inc. Cracow ‘03 Grid Workshop SAN over WAN - a new way of solving the GRID data access bottleneck Dr. Wolfgang Mertz Business Development.
Silicon Graphics, Inc. Poster Presented by: SGI Proprietary Technologies for Breakthrough Research Rosario Caltabiano North East Higher Education & Research.
Distributed Database Management Systems
Server Platforms Week 11- Lecture 1. Server Market $ 46,100,000,000 ($ 46.1 Billion) Gartner.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
Hitachi SR8000 Supercomputer LAPPEENRANTA UNIVERSITY OF TECHNOLOGY Department of Information Technology Introduction to Parallel Computing Group.
12 Chapter 12 Client/Server Systems Hachim Haddouti.
12 Chapter 12 Client/Server Systems Database Systems: Design, Implementation, and Management, Fifth Edition, Rob and Coronel.
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
NPACI: National Partnership for Advanced Computational Infrastructure August 17-21, 1998 NPACI Parallel Computing Institute 1 Cluster Archtectures and.
National Energy Research Scientific Computing Center (NERSC) The GUPFS Project at NERSC GUPFS Team NERSC Center Division, LBNL November 2003.
Scientific Data Infrastructure in CAS Dr. Jianhui Scientific Data Center Computer Network Information Center Chinese Academy of Sciences.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
STRATEGIES INVOLVED IN REMOTE COMPUTATION
Enterprise Storage A New Approach to Information Access Darren Thomas Vice President Compaq Computer Corporation.
Principles of Scalable HPC System Design March 6, 2012 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Wayne Hogan National Storage Manager Sun Microsystems of Canada, Inc.
Virtualization in the NCAR Mass Storage System Gene Harano National Center for Atmospheric Research Scientific Computing Division High Performance Systems.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
VO Sandpit, November 2009 e-Infrastructure to enable EO and Climate Science Dr Victoria Bennett Centre for Environmental Data Archival (CEDA)
Nov 1, 2000Site report DESY1 DESY Site Report Wolfgang Friebel DESY Nov 1, 2000 HEPiX Fall
M.Lautenschlager (WDCC, Hamburg) / / 1 Semantic Data Management for Organising Terabyte Data Archives Michael Lautenschlager World Data Center.
The University of Bolton School of Games Computing & Creative Technologies LCT2516 Network Architecture CCNA Exploration LAN Switching and Wireless Chapter.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
M.Lautenschlager (WDCC, Hamburg) / / 1 Semantic Data Management for Organising Terabyte Data Archives Michael Lautenschlager World Data Center.
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
Types of Operating Systems
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Parallel Programming on the SGI Origin2000 With thanks to Igor Zacharov / Benoit Marchand, SGI Taub Computer Center Technion Moshe Goldberg,
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
 CASTORFS web page - CASTOR web site - FUSE web site -
Oracle's Distributed Database Bora Yasa. Definition A Distributed Database is a set of databases stored on multiple computers at different locations and.
CLASS Information Management Presented at NOAATECH Conference 2006 Presented by Pat Schafer (CLASS-WV Development Lead)
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
Michael Lautenschlager, Hannes Thiemann, Frank Toussaint WDC Climate / Max-Planck-Institute for Meteorology, Hamburg Joachim Biercamp, Ulf Garternicht,
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
Cray Environmental Industry Solutions Per Nyberg Earth Sciences Business Manager Annecy CAS2K3 Sept 2003.
[IT INFRASTRUCTURE OVERVIEW] Paul Ashurst Executive Director, EIS Global Network Services IT Corporate.
Lautenschlager + Thiemann (M&D/MPI-M) / / 1 Introduction Course 2006 Services and Facilities of DKRZ and M&D Integrating Model and Data Infrastructure.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
Chapter 9: Networking with Unix and Linux. Objectives: Describe the origins and history of the UNIX operating system Identify similarities and differences.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
Evolving Scientific Data Workflow CAS 2011 Pamela Gillman
SAMPLE IMAGE gLite on the Market – Why and How ir Works 4 th EGEE User Forum Thursday, 05 March 2009 Le Ciminiere, Catania, Sicily, Italy Gaël Youinou.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
STORAGE ARCHITECTURE/ MASTER): Where IP and FC Storage Fit in Your Enterprise Randy Kerns Senior Partner The Evaluator Group.
SYSTEMSDESIGNANALYSIS 1 Chapter 21 Implementation Jerry Post Copyright © 1997.
Tackling I/O Issues 1 David Race 16 March 2010.
Pathway to Petaflops A vendor contribution Philippe Trautmann Business Development Manager HPC & Grid Global Education, Government & Healthcare.
Background Computer System Architectures Computer System Software.
LIOProf: Exposing Lustre File System Behavior for I/O Middleware
Presented by Robust Storage Management On Desktop, in Machine Room, and Beyond Xiaosong Ma Computer Science and Mathematics Oak Ridge National Laboratory.
Page : 1 SC2004 Pittsburgh, November 12, 2004 DEISA : integrating HPC infrastructures in Europe DEISA : integrating HPC infrastructures in Europe Victor.
Research and Service Support Resources for EO data exploitation RSS Team, ESRIN, 23/01/2013 Requirements for a Federated Infrastructure.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
Approaches and Challenges in Managing Persistent Identifiers
Network Attached Storage Overview
The INFN Tier-1 Storage Implementation
Research Data Archive - technology
Storage Virtualization
Cloud Web Filtering Platform
Presentation transcript:

9-Sept-2003CAS2003, Annecy, France, WFS1 Distributed Data Management at DKRZ Distributed Data Management at DKRZ Wolfgang Sell Hartmut Fichtel Deutsches Klimarechenzentrum GmbH Wolfgang Sell Hartmut Fichtel Deutsches Klimarechenzentrum GmbH

9-Sept-2003CAS2003, Annecy, France, WFS2 Table of Contents DKRZ - a German HPC Center HPC Systemarchitecture suited for Earth System Modeling The HLRE Implementation at DKRZ Implementing IA64/Linux based Distributed Data Management Some Results Summary DKRZ - a German HPC Center HPC Systemarchitecture suited for Earth System Modeling The HLRE Implementation at DKRZ Implementing IA64/Linux based Distributed Data Management Some Results Summary

9-Sept-2003CAS2003, Annecy, France, WFS3 DKRZ - a German HPCC Mission of DKRZ DKRZ and its Organization DKRZ Services Model and Data Services Mission of DKRZ DKRZ and its Organization DKRZ Services Model and Data Services

9-Sept-2003 CAS2003, Annecy, France, WFS Page 4 In 1987 DKRZ was founded with the Mission to Provide state-of-the-art supercomputing and data service to the German scientific community to conduct top of the line Earth System and Climate Modelling. Provide associated services including high level visualization. In 1987 DKRZ was founded with the Mission to Provide state-of-the-art supercomputing and data service to the German scientific community to conduct top of the line Earth System and Climate Modelling. Provide associated services including high level visualization. Mission of DKRZ

9-Sept-2003 CAS2003, Annecy, France, WFS Page 5 Deutsches KlimaRechenZentrum = DKRZ German Climate Computer Center organised under private law (GmbH) with 4 shareholders investments funded by federal government, operations funded by shareholders usage 50 % shareholders and 50 % community Deutsches KlimaRechenZentrum = DKRZ German Climate Computer Center organised under private law (GmbH) with 4 shareholders investments funded by federal government, operations funded by shareholders usage 50 % shareholders and 50 % community DKRZ and its Organization (1)

9-Sept-2003 CAS2003, Annecy, France, WFS Page 6 DKRZ internal Structure 3 departments for systems and networks visualisation and consulting administration 20 staff in total until restructuring end of 1999 a fourth department supported climate model applications and climate data management DKRZ internal Structure 3 departments for systems and networks visualisation and consulting administration 20 staff in total until restructuring end of 1999 a fourth department supported climate model applications and climate data management DKRZ and its Organization (2)

9-Sept-2003 CAS2003, Annecy, France, WFS Page 7 operations center:DKRZ technical organization of computational ressources (compute-, data- and network-services, infrastructure) advanced visualisation assistance for parallel architectures (consulting and training) operations center:DKRZ technical organization of computational ressources (compute-, data- and network-services, infrastructure) advanced visualisation assistance for parallel architectures (consulting and training) DKRZ Services

9-Sept-2003 CAS2003, Annecy, France, WFS Page 8 competence center:Model & Data professional handling of community models specific scenario runs scientific data handling Model & Data Group external to DKRZ, administered by MPI for Meteorology, funded by BMBF competence center:Model & Data professional handling of community models specific scenario runs scientific data handling Model & Data Group external to DKRZ, administered by MPI for Meteorology, funded by BMBF Model & Data Services

9-Sept-2003CAS2003, Annecy, France, WFS9 HPC Systemarchitecture suited for Earth System Modeling Principal HPC System Configuration Links between Different Services The Data Problem Principal HPC System Configuration Links between Different Services The Data Problem

9-Sept-2003 CAS2003, Annecy, France, WFS Page 10 Principal HPC System Configuration

9-Sept-2003 CAS2003, Annecy, France, WFS Page 11 Functionality and Performance Requirements for Data Service Transparent Access to Migrated Data High Bandwidth for Data Transfer Shared Filesystem Possibility for Adaptation in Upgrade Steps due to Changes in Usage Profile Functionality and Performance Requirements for Data Service Transparent Access to Migrated Data High Bandwidth for Data Transfer Shared Filesystem Possibility for Adaptation in Upgrade Steps due to Changes in Usage Profile Link between Compute Power and Non-Computing Services

9-Sept-2003 CAS2003, Annecy, France, WFS Page 12 Compute server power

9-Sept-2003 CAS2003, Annecy, France, WFS Page 13 Adaptation Problem for Data Server

9-Sept-2003 CAS2003, Annecy, France, WFS Page 14 High Bandwidth between the Coupled Servers Scalability supported by Operating System No Needs for Multiple Copies Record Level Access to Data with High Performance Minimized Data Transfers High Bandwidth between the Coupled Servers Scalability supported by Operating System No Needs for Multiple Copies Record Level Access to Data with High Performance Minimized Data Transfers Pros of Shared Filesystem Coupling

9-Sept-2003 CAS2003, Annecy, France, WFS Page 15 Proprietary Software needed Standardisation still missing Limited Number of Vendors whose Systems can be connected Proprietary Software needed Standardisation still missing Limited Number of Vendors whose Systems can be connected Cons of Shared Filesystem Coupling

9-Sept-2003CAS2003, Annecy, France, WFS16 HLRE Implementation at DKRZ HöchstLeistungsRechnersystem für die Erdsystem- forschung = HLRE High Performance Computer System for Earth System Research Principal HLRE System Configuration HLRE Installation Phases IA64/Linux based Data Services Final HLRE Configuration HöchstLeistungsRechnersystem für die Erdsystem- forschung = HLRE High Performance Computer System for Earth System Research Principal HLRE System Configuration HLRE Installation Phases IA64/Linux based Data Services Final HLRE Configuration

9-Sept-2003 CAS2003, Annecy, France, WFS Page 17 Principal HLRE System Configuration

9-Sept-2003 CAS2003, Annecy, France, WFS Page 18 HLRE Phases Mass Storage Capacity [Tbytes] >720 >1400 >3400 Date Feb Q Q 2003 Nodes CPUs Expected Sustained Performance [Gflops] ca. 200 ca. 350 ca. 500 Expected Increase in Thruputcompared toCRAY C916 ca. 40 ca. 75 ca. 100 Main Memory[Tbytes] Disk-Capacity [Tbytes] ca. 30 ca. 50 ca. 60

9-Sept-2003 CAS2003, Annecy, France, WFS Page 19 DS phase 1: basic structure CS performance increase f = 37 F = f 3/4 = 15 minimal component performance indicated in diagram explicit user access ftp, scp... CS disks with local copies DS disks for cache physically distributed DS NAS architecture CS performance increase f = 37 F = f 3/4 = 15 minimal component performance indicated in diagram explicit user access ftp, scp... CS disks with local copies DS disks for cache physically distributed DS NAS architecture CS client(s) DS other clients GE 180 MB/s 45 MB/s 150 MB/s 375 MB/s 16.5 TB ~ PB 11 TB

9-Sept-2003 CAS2003, Annecy, France, WFS Page 20 Adaptation Option for Data Server

9-Sept-2003 CAS2003, Annecy, France, WFS Page 21 DS phases 2,3: basic structure CS performance increase f = 63/100 F = f 3/4 = 22.4/31.6 minimal component performance indicated in diagram implicit user access local UFS commands CS disks with local copies shared disks (GFS) DS disks for IO buffercache Intel/Linux platforms homogenous HW technological challenge CS performance increase f = 63/100 F = f 3/4 = 22.4/31.6 minimal component performance indicated in diagram implicit user access local UFS commands CS disks with local copies shared disks (GFS) DS disks for IO buffercache Intel/Linux platforms homogenous HW technological challenge CS client(s) DS other clients GE 270/325 MB/s 70/80 MB/s 225/270 MB/s 560/675 MB/s 16.5 TB ~ PB FC 25/30 TB 11 TB

9-Sept-2003CAS2003, Annecy, France, WFS22 Implementing IA64/Linux based Distributed Data Management Overall Phase 1 Configurations Introducing Linux based Distributed HSM Introducing Linux based Distributed DBMS Final Overall Phase 3 Configuration Overall Phase 1 Configurations Introducing Linux based Distributed HSM Introducing Linux based Distributed DBMS Final Overall Phase 3 Configuration

9-Sept-2003Page 23 Proposed final phase 3 configuration HS/MS LAN GE x 48 x 16 x 25 FE x 2/node For PolestarLite AsAmA 16way GFS/Server UVDM AsAmA 16way GFS/Server UVDM UDSN AsAmA 4way GFS/Client Oracle AsAmA 4way GFS/Client Oracle UDSN/UDNL GFS Disk (Polestar) 0.28 x 53 =14.8TB x 36 GFS Disk (Polestar) 0.28 x 53 =14.8TB x 36 x 8 x 2 x 4 FC x 72 x 8 Disk Cache (Polestar) 0.57TB x 15 = 8.5TB Disk Cache (DDN) 0.69TB x 12 = 8.3TB x 72 Local Disk FC- RAID 0.28TB x20 =5.6TB Local Disk FC- RAID 0.28TB x20 =5.6TB Silkworm x 20 x 120 x 32 SX-6SX-6SX-6SX-6SX-6SX-6SX-6SX-6 IXS 24nodes 9940B x B x C x 5 Local Disk (Polestar) 0.14 x 2 = 0.28TB Local Disk (Polestar) 0.14 x 2 = 0.28TB x 2 for Local disk x 2 for Local disk Fibre channel GigabitEther AsAmA 4way GFS/Client Oracle AsAmA 4way GFS/Client Oracle UDSN/UDNL x 2 x 4 AzusA 16way GFS/Server Post processing system UCFM/UDSN Disk FC x 8 Tape FC x 6 Disk FC x 8 Tape FC x 6 x 4 x 16 x 8 Oracle DB (DDN) 2TB x 4 = 8TB x 8 x 4 Oracle Application Server SQLNET Sun 4CPU The Internet AsamA 4CPU SQLNET Migration upon market availability of components

9-Sept-2003CAS2003, Annecy, France, WFS24 Some Results Growth of the Data Archive Growth of Transferrate Observed Transferrates for HLRE FLOPS-Rates Growth of the Data Archive Growth of Transferrate Observed Transferrates for HLRE FLOPS-Rates

9-Sept-2003 CAS2003, Annecy, France, WFS Page 25 DS archive capacity [TB]

9-Sept-2003 CAS2003, Annecy, France, WFS Page 26 DS archive capacity ( )

9-Sept-2003 CAS2003, Annecy, France, WFS Page 27 DS transfer rates [GB/day]

9-Sept-2003 CAS2003, Annecy, France, WFS Page 28 DS transfer rates ( )

9-Sept-2003 CAS2003, Annecy, France, WFS Page 29 DS transfer rates ( )

9-Sept-2003 CAS2003, Annecy, France, WFS Page 30 Observed Transferrates for HLRE LinkSingleStream Transferrate [MB/s] Aggregate Transferrate [MB/s] CS -> DS viaftp, (12.1 SUPER-UX) CS -> DS viaftp, (12.2 SUPER-UX) CS ->localdisk, (12.1 SUPER-UX) > CS -> GFSdisk, (13.1 SUPER-UX) Up to DS -> GFSdisk, (Linux) Up to per node

9-Sept-2003 CAS2003, Annecy, France, WFS Page 31 Observed FLOPS-rates for HLRE 4 node performance > approx.100 GLFOPS ( about 40 % Efficiency) for ECHAM (70-75) MOM Radar Reflection on Sea Ice 24 node performance for Turbulence Code about 470 GFLOPS (30+ % Efficiency)

9-Sept-2003CAS2003, Annecy, France, WFS32 SummarySummary DKRZ provides Computing Resources for Climate Research in Germany on an competitive international level The HLRE System Architecture is suited to cope with a data-intensive Usage Profile Shared Filesystems today are operational in Heterogenous System Environments Standardisation-Efforts for Shared Filesystems needed DKRZ provides Computing Resources for Climate Research in Germany on an competitive international level The HLRE System Architecture is suited to cope with a data-intensive Usage Profile Shared Filesystems today are operational in Heterogenous System Environments Standardisation-Efforts for Shared Filesystems needed

9-Sept-2003CAS2003, Annecy, France, WFS33 Thank you for your attention !

9-Sept-2003 CAS2003, Annecy, France, WFS Page 34 Tape transfer rates ( )

9-Sept-2003 CAS2003, Annecy, France, WFS Page 35 DS transfer requests ( )

9-Sept-2003 CAS2003, Annecy, France, WFS Page 36 DS archive capacity ( )

9-Sept-2003 CAS2003, Annecy, France, WFS Page 37 DS archive capacity ( )