Download presentation
Presentation is loading. Please wait.
Published byGordon Simpson Modified over 9 years ago
1
LLNL’s Data Center and Interoperable Services 5 th Annual ESGF Face-to-Face Conference ESGF 2015 Monterey, CA, USA Dean N. Williams, Tony Hoang, Cameron Harr, Sasha Ames, Jeff Painter Lawrence Livermore National Laboratory December 8-11, 2015 http://aims.llnl.gov
2
Williams 2 LLNL ESGF Data Network
3
Williams 3 AIMS/ESGF Computing Resources
4
Williams 4 AIMS Experimental Compute Cluster System TypeMemory (GBs)CPUsOS Version Master node12864x 2.6GHzRHEL 6.5 Slave nodes 1-81288x 1.2GHzRHEL 6.5 AIMS compute cluster software UV-CDAT SLURM daemon (with munged) Previous test installations: Ceph OSD / MDS (single node used) Cloudera packages – HDFS, YARN (for hadoop, spark), hive
5
Williams 5 AIMS/ESGF Data Transfer Nodes (DTNs) Data transfer nodes run Globus Connect and GridFTP software to allow files to be transferred between DTNs globally. Globus Connect allows for a user-friendly web-browser based search and file transfer GridFTP is the command-line data transfer agent Files are available to users with ESGF certificate/ID Ingest-only nodes are purposed for bulk transfer of new data to LLNL AIMS repository Delivered mid-November and awaiting integration
6
Williams 6 AIMS cluster network 10GigE Switch (Private Network) AIMS/PCMDI Nodes Internet LLNL Enterprise Network Compute Cluster Nodes Firewalls
7
Williams 7 CMIP-5 Data Management Workflow @ LLNL CSS ESGF Datanode (pcmdi9 and pcmdi7) ESGF Datanode (pcmdi9 and pcmdi7) Data transfer node (now gdo2) About 30 Modeling Centers Supercomputer Storage and data transfer servers Secondary computer (converts to CMIP-5 standards) LLNL Environment scientists
8
Williams 8 ACME Data Management Workflow @ LLNL AIMS “Scratch” Storage AIMS “Scratch” Storage CSS ESGF Datanode ESGF Datanode AIMS Analysis Node AIMS Analysis Node Staging server (GDO2) OLCFALCFNERSC Petaflop Supercomputers ACME Scientists 1)Run models 2)Post-processing 3)Transfer data 1)Analysis (at AIMS) “Scratch” Storage AIMS Data Manager 5) Stage to CSS 6) Publish to ESGF climatology data (reduced) histories Analysis Cluster (for postprocessing) LLNL Environment
9
Williams 9 Climate Storage System Information System Type CPUsMemory (GBs) OSDDN Luns ~LUN Size LUN SFC (active/fail- over) ZFS Raid Type ZFS Raid Set Size ZFS Raid Sets Usable Space Dell91032128Solaris 114824TB4/4RaidZ2124960TB Dell91032128TOSS3624TB4/4RaidZ294670TB Sunx444032128Solaris106016TB4/4RaidZ2106768TB The climate storage system environment consists of two Dell 910 servers each with 4 sockets (x 8 cores per socket) and 128 GB of memory. Each of the systems is connected via fiber channel (FC) to a DDN 9900 storage array. Currently there are eight 8 Gbps FC connections to the DDN 9900 unit of which 4 are active and 4 are passive (or used as fail-over). These systems are used as NFS servers and provide a storage archive for the PCMDI front-end servers. The initial CSS system (cssfen1) was installed using Solaris 11 and ZFS. Its file system is composed of 48 LUNs from one of the DDN 9900 units. The second system (cssfen2) was installed using TOSS and ZFS. Its file system is composed of 36 LUNs from a DDN 9900 unit. Other storage used by the Climate environment is allocated from the Green Data Oasis (GDO) environment. This storage is used via the PCMDI GDO zone called "gdo2". Storage space is allocated to the gdo2 zone from the gdofen1 system. The storage space on the gdofen1 system is around 768 TB, however that space is shared with GDO zones running in the GDO environment.
10
Williams 10 What ESGF request of the WGCM Waiting for CMIP6 requirements from WCRP/WGCM/WIP — Prioritized list composed of: must have capabilities; and what would be nice to have capabilities Waiting for timelines for CMIP6 data production and distribution What is the estimated size of the CMIP6 distributed archive and list of the CMIP6 modeling centers CMIP6 data centers must commit to infrastructure modernization to support CMIP6 data scale — Local compute cluster for local data analysis through ESGF portals or other connecting interfaces — High performance networks and data transfer nodes to support data distribution at scale when local analysis capabilities are not sufficient — High performance storage (important data sets on spinning disk rather than tape)
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.