Centre de Calcul de l’Institut National de Physique Nucléaire et de Physique des Particules Data storage services at CC-IN2P3 Jean-Yves Nief.

Slides:



Advertisements
Similar presentations
Computing Infrastructure
Advertisements

Cloud Storage in Czech Republic Czech national Cloud Storage and Data Repository project.
Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
Jean-Yves Nief, CC-IN2P3 Wilko Kroeger, SCCS/SLAC Adil Hasan, CCLRC/RAL HEPiX, SLAC October 11th – 13th, 2005 BaBar data distribution using the Storage.
What is it? Hierarchical storage software developed in collaboration with five US department of Energy Labs since 1992 Allows storage management of 100s.
Generic policy rules and principles Jean-Yves Nief.
IRODS usage at CC-IN2P3 Jean-Yves Nief. Talk overview What is CC-IN2P3 ? Who is using iRODS ? iRODS administration: –Hardware setup. iRODS interaction.
Mass RHIC Computing Facility Razvan Popescu - Brookhaven National Laboratory.
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
Report : Zhen Ming Wu 2008 IEEE 9th Grid Computing Conference.
16 th May 2006Alessandra Forti Storage Alessandra Forti Group seminar 16th May 2006.
SDSC RP Update TeraGrid Roundtable Reviewing Dash Unique characteristics: –A pre-production/evaluation “data-intensive” supercomputer based.
CC - IN2P3 Site Report Hepix Fall meeting 2009 – Berkeley
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
SLAC Experience on Bestman and Xrootd Storage Wei Yang Alex Sim US ATLAS Tier2/Tier3 meeting at Univ. of Chicago Aug 19-20,
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
Integrating JASMine and Auger Sandy Philpott Thomas Jefferson National Accelerator Facility Jefferson Ave. Newport News, Virginia USA 23606
Support in setting up a non-grid Atlas Tier 3 Doug Benjamin Duke University.
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
Overview of grid activities in France in relation to FKPPL FKPPL Workshop Thursday February 26th, 2009 Dominique Boutigny.
CERN – IT Department CH-1211 Genève 23 Switzerland t Working with Large Data Sets Tim Smith CERN/IT Open Access and Research Data Session.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
Status Report of WLCG Tier-1 candidate for KISTI-GSDC Sang-Un Ahn, for the GSDC Tier-1 Team GSDC Tier-1 Team ATHIC2012, Busan,
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Derek Ross E-Science Department DCache Deployment at Tier1A UK HEP Sysman April 2005.
Scientific Storage at FNAL Gerard Bernabeu Altayo Dmitry Litvintsev Gene Oleynik 14/10/2015.
Cloud Computing is a Nebulous Subject Or how I learned to love VDF on Amazon.
Managing Petabytes of data with iRODS at CC-IN2P3
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
KIT – University of the State of Baden-Württemberg and National Laboratory of the Helmholtz Association STEINBUCH CENTRE FOR COMPUTING - SCC
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
SA1 operational policy training, Athens 20-21/01/05 Presentation of the HG Node “Isabella” and operational experience Antonis Zissimos Member of ICCS administration.
W.A.Wojcik/CCIN2P3, Nov 1, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center URL:
BaBar Cluster Had been unstable mainly because of failing disks Very few (
T3g software services Outline of the T3g Components R. Yoshida (ANL)
01. December 2004Bernd Panzer-Steindel, CERN/IT1 Tape Storage Issues Bernd Panzer-Steindel LCG Fabric Area Manager CERN/IT.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
GDB meeting - Lyon - 16/03/05 An example of data management in a Tier A/1 Jean-Yves Nief.
1 5/4/05 Fermilab Mass Storage Enstore, dCache and SRM Michael Zalokar Fermilab.
DCache/XRootD Dmitry Litvintsev (DMS/DMD) FIFE workshop1Dmitry Litvintsev.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
CC-IN2P3 Pierre-Emmanuel Brinette Benoit Delaunay IN2P3-CC Storage Team 17 may 2011.
Storage & Database Team Activity Report INFN CNAF,
An Analysis of Data Access Methods within WLCG Shaun de Witt, Andrew Lahiff (STFC)
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
Centre de Calcul de l’Institut National de Physique Nucléaire et de Physique des Particules Data storage services at CC-IN2P3 Jean-Yves Nief.
November 28, 2007 Dominique Boutigny – CC-IN2P3 CC-IN2P3 Update Status.
IRODS at CC-IN2P3: overview Jean-Yves Nief. Talk overview iRODS in production: –Hardware setup. –Usage. –Prospects. iRODS developpements in Lyon: –Scripts.
Introduction to Data Analysis with R on HPC Texas Advanced Computing Center Feb
Predrag Buncic CERN Data management in Run3. Roles of Tiers in Run 3 Predrag Buncic 2 ALICEALICE ALICE Offline Week, 01/04/2016 Reconstruction Calibration.
ORNL is managed by UT-Battelle for the US Department of Energy OLCF HPSS Performance Then and Now Jason Hill HPC Operations Storage Team Lead
High Performance Storage System (HPSS) Jason Hick Mass Storage Group HEPiX October 26-30, 2009.
BNL Tier1 Report Worker nodes Tier 1: added 88 Dell R430 nodes
CC - IN2P3 Site Report Hepix Spring meeting 2011 Darmstadt May 3rd
Luca dell’Agnello INFN-CNAF
Ákos Frohner EGEE'08 September 2008
The INFN Tier-1 Storage Implementation
CC-IN2P3 Pierre-Emmanuel Brinette IN2P3-CC Storage Team
CTA: CERN Tape Archive Overview and architecture
Research Data Archive - technology
CC and LQCD dimanche 13 janvier 2019dimanche 13 janvier 2019
Presentation transcript:

Centre de Calcul de l’Institut National de Physique Nucléaire et de Physique des Particules Data storage services at CC-IN2P3 Jean-Yves Nief

 Hardware: ◦ Storage on disk. ◦ Storage on tape.  Software: ◦ Storage services.  Pitfalls and lessons learned.  Storage needs in the near future.  Prospects. Agenda Visit of NCSA

Storage at CC-IN2P3: disk Visit of NCSA Parallel File System: GPFS (1.9 PBs) File servers: xrootd, dCache (14 PBs) Used for High Energy Physics (LHC etc…) Mass Storage System: HPSS (600 TBs) Used as a disk cache in front of the tapes. Middlewares: SRM, iRODS (840 TBs) Databases: mySQL, PostGres, Oracle (57 TBs) Software Direct Attached Storage servers (DAS): Dell servers (R720xd + MD1200) ~ 240 servers Capacity: 15 PBs Disk attached via SAS: Dell servers ( R620 + MD3260) Capacity: 1.9 PBs Storage Area Network disk arrays (SAN): IBM V7000 and DCS3700, Hitachi HUS 130. Capacity: 240 TBs Hardware

Storage at CC-IN2P3: tapes Visit of NCSA Mass Storage System: HPSS 25 PBs Max traffic (from HPSS): 100 TBs / day Interfaced with our disk services Backup service: TSM (1 PB) 4 Oracle/STK SL8500 librairies: 40,000 slots (T10K and LTO4) Max capacity: 320 PBs (with T10KD tapes) 111 tape drives 1 IBM TS3500 library: 3500 slots (LTO6) Hardware Software

 4 Oracle/STK SL8500 libraries (4 x slots) for HPSS and TSM.  1 IBM TS3500 library (3500 slots) for TSM.  Tape media: Storage hardware as of today Visit of NCSA MediaCapacity Max Bandwidth Number of drives Service T10K-B1TB120 MB/s59HPSS T10K-C5 TB240 MB/s21HPSS T10K-D8.5 TB252 MB/s31HPSS LTO4800 GB120 MB/s20TSM LTO62.4 TB160 MB/s6TSM

How experiments are dealing with the storage ? Visit of NCSA TSM HPSS iRODS SRM Client apps (local or distant apps) GPFS dCache xrootd File servers / filesystem (local access) Middlewares (wide area access) Mass Storage System Backup

 Purpose: ◦ For heavy I/O operations by many batch jobs in //. ◦ Not a permanent repository.  60 servers, 1.9 PBs of disk space.  1000 GPFS clients installed on our  Used by: ◦ up to 25% of the batch farm. ◦ 84 groups.  Max activity observed: 500 TBs read in 5 days (Planck)  To do: ◦ Migration of metadata storage to flash technology (NetApp EF560). ◦ Reduction of the number of space (35 right now) down to 4. Local access: distributed filesystem Visit of NCSA

 Purpose: ◦ Data access by batch jobs (local + grid). ◦ Pattern: remote access (random I/O, sequential). ◦ Interface with HPSS:  Using Treqs (Tape request scheduler). ◦ dCache service part of the Atlas and CMS federations:  xrootd proxy servers (gateway to our local dCache service). ◦ dCache head nodes db: use of SSD.  To do: ◦ xrootd: lots of tape staging (eg: 20k requests per day)  « automatic » prestaging for some large « productions » needed. Local access: dCache-xrootd Visit of NCSA service# serverscapacityexperiments # clients in // access protocol dCache PiB LCG + « EGEE » VOs (17) dcap, xrootd, WebDAV, gsiftp xrootd442.5 PiB Alice + HEP + astroparticle (13) 8000 xrootd, http

Local access (xrootd/dCache) Visit of NCSA Client HPSS Redirector server: Xrootd T1.root (1) (4) (2) /hpss/in2p3.fr/T1.root ? (etc…) Data server: Xrootd (5) (3) (1) + (2): load balancing (6): random access Data server: Xrootd (6) redirectors Data servers (4) + (5): dynamic staging (TreqS)

 Purpose: ◦ Interface to our back end storage: tapes (main storage).  29 servers (disk + tape movers).  Massive activity: ◦ 27 PiB on tapes (rank #12 in HPSS world), 597 TiB of disk space (Dell R510, R720 servers). ◦ Up to 100 TiB in 22h between HPSS and other storage services.  New drive technology every 30 months.  Life cycle of a tape generation 8 years: ◦ Ramping up in 4 years. ◦ 4 more years in production.  T10KB decommissionning: ◦ tapes to be repacked until end 2017  5 PiB / year. ◦ Represents:  1000 T10K-T2  T10K-C  Or 750 T10K-T2  T10K-D  Tape lifetime: avg 7 years (max 10 years).  Redefine the class of services.  Refactoring of Treqs (Tape request scheduler) by our dev team. Mass Storage service: HPSS Visit of NCSA

 Dynamic staging may result in: ◦ Massive staging requests. ◦ Long delays to access the files. ◦ Tapes capacity is getting bigger!  Increased job inefficiency.  To leverage these drawbacks: ◦ Big files on tapes (several GBs). ◦ Reordering of file requests per tape id (reduce # of mounts/dismounts): up to per day. ◦ Prestaging. ◦ Avoid file scattering on a large amount of tapes (tape families ?).  Refactoring of Treqs (Tape request scheduler) by our dev team. Local access: pitfalls and lessons learned Visit of NCSA

 Purpose: ◦ Data management (storage virtualization, policies). ◦ Data sharing, access, archival etc… in a distributed and heterogeneous environment.  SRM: ◦ Used by Atlas, CMS, LHCb + EGEE/EGI virtual organizations.  iRODS: ◦ Managing > 9 PBs of data (HEP, astro, biology, Arts & Humanities): includes data stored on tapes.  Web interfaces, HTTP/REST, webdav. Middleware: SRM, iRODS Visit of NCSA

 GPFS: ◦ 130 TB (50% being used at the moment)  HPSS: ◦ 78 TB  iRODS: ◦ 90 TB Focus: LSST storage usage Visit of NCSA

Future storage needs (Hard Sci Fi) Visit of NCSA Cumulated Disk (TB) Cumulated Tape (TB) Credit: Rachid Lemrani Hypothesis: LCG: +10% / year HEP like experiment: ratio disk/tape = 30% CTA not included

 Used to debug Qserv.  1 interactive node: ◦ 1 virtual machine ccqservbuild.in2p3.fr. ◦ Mainly used for:  the compilation of the Qserv software.  the deployement of Qserv on the cluster. ◦ Accessible from SLAC.  50 Qserv nodes: ◦ Dell / CC-IN2P3 partnership. ◦ ccqserv{ }.in2p3.fr. ◦ Dell servers R620 and R630  Intel Xeon, 16GB RAM  7 TB of usable storage ◦ run the Qserv software (1 master, 49 slaves). ◦ private subnetwork (nodes accessible from ccqservbuild only). (Credit: Yvan Calas) Qserv Visit of NCSA

 Tape will still have a major role: ◦ Balance between nearline and pure archive storage ?  Hard drives still preferred to SSD: ◦ prices will converge in the next couple of years. ◦ SSD usage limited at the moment for key apps (GPFS metadata, HPSS and dCache databases etc…).  Heavy I/O still and even heavier: ◦ Lots of « random » I/O: still strong needs for seek, read, write. ◦ Simple put/get interaction not sufficient.  Big metadata challenge (storage system metadata).  Data life cycle improvements: ◦ > 100 groups, 5000 users (~ 2000 active users). ◦ Data Management Plan. ◦ archival service beyond simple bit preservation ? ◦ OAIS, dublin core metadata.  Provide interfaces for Open data ? Storage prospects Visit of NCSA

 Assessment of object storage: ◦ Swift: testing it at the moment. ◦ Swift extra value wrt what we already have: to be proven. ◦ Ceph (but it can be more than an object storage): evaluation starting in the next couple of months.  Hadoop, MapReduce type of systems: no plans yet as not much needs expressed by our user community.  Metadata: will increase massively with increased storage. ◦ Limit the amount of files => increased file size. Storage prospects Visit of NCSA