IRODS at CC-IN2P3: overview Jean-Yves Nief. Talk overview iRODS in production: –Hardware setup. –Usage. –Prospects. iRODS developpements in Lyon: –Scripts.

Slides:



Advertisements
Similar presentations
Peter Berrisford RAL – Data Management Group SRB Services.
Advertisements

EHarmony in Cloud Subtitle Brian Ko. eHarmony Online subscription-based matchmaking service Available in United States, Canada, Australia and United Kingdom.
Jean-Yves Nief, CC-IN2P3 Wilko Kroeger, SCCS/SLAC Adil Hasan, CCLRC/RAL HEPiX, SLAC October 11th – 13th, 2005 BaBar data distribution using the Storage.
Generic policy rules and principles Jean-Yves Nief.
Amazon EC2 Quick Start adapted from EC2_GetStarted.html.
IRODS usage at CC-IN2P3 Jean-Yves Nief. Talk overview What is CC-IN2P3 ? Who is using iRODS ? iRODS administration: –Hardware setup. iRODS interaction.
Simplify your Job – Automatic Storage Management Angelo Session id:
Windows Server MIS 424 Professor Sandvig. Overview Role of servers Performance Requirements Server Hardware Software Windows Server IIS.
Experience of a low-maintenance distributed data management system W.Takase 1, Y.Matsumoto 1, A.Hasan 2, F.Di Lodovico 3, Y.Watase 1, T.Sasaki 1 1. High.
Chapter 9: Novell NetWare
IRODS performance test and SRB system at KEK Yoshimi KEK Building data grids with iRODS 27 May 2008.
Hosted by Case Study - Storage Consolidation Steve Curry Yahoo Inc.
OSG Public Storage and iRODS
Introduction to iRODS Jean-Yves Nief. Talk overview Data management context. Some data management goals: –Storage virtualization. –Virtualization of the.
CC - IN2P3 Site Report Hepix Fall meeting 2009 – Berkeley
Grids and Portals for VLAB Marlon Pierce Community Grids Lab Indiana University.
DDN & iRODS at ICBR By Alex Oumantsev History of ICBR  Campus wide Interdisciplinary Center for Biotechnology Research  Core Facility  Funded by the.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
W.A.Wojcik/CCIN2P3, May Running the multi-platform, multi-experiment cluster at CCIN2P3 Wojciech A. Wojcik IN2P3 Computing Center
Overview of day-to-day operations Suzanne Poulat.
Production Data Grids SRB - iRODS Storage Resource Broker Reagan W. Moore
Installation and Development Tools National Center for Supercomputing Applications University of Illinois at Urbana-Champaign The SEASR project and its.
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
Module 1: Configuring Windows Server Module Overview Describe Windows Server 2008 roles Describe Windows Server 2008 features Describe Windows Server.
Microsoft Azure SoftUni Team Technical Trainers Software University
MAGDA Roger Jones UCL 16 th December RWL Jones, Lancaster University MAGDA  Main authors: Wensheng Deng, Torre Wenaus Wensheng DengTorre WenausWensheng.
Centre de Calcul de l’Institut National de Physique Nucléaire et de Physique des Particules Data storage services at CC-IN2P3 Jean-Yves Nief.
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
Tools for collaboration How to share your duck tales…
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Overview of the iPlant Data Store.
8 October 1999 BaBar Storage at CCIN2P3 p. 1 Rolf Rumler BaBar Storage at Lyon HEPIX and Mass Storage SLAC, California, U.S.A. 8 October 1999 Rolf Rumler,
The Global Land Cover Facility is sponsored by NASA and the University of Maryland.The GLCF is a founding member of the Federation of Earth Science Information.
Hepix LAL April 2001 An alternative to ftp : bbftp Gilles Farrache In2p3 Computing Center
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
Introduction to The Storage Resource.
Managing Petabytes of data with iRODS at CC-IN2P3
W.A.Wojcik/CCIN2P3, Nov 1, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center URL:
CC - IN2P3 Site Report Hepix Fall meeting 2010 – Ithaca (NY) November 1st 2010
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
GDB meeting - Lyon - 16/03/05 An example of data management in a Tier A/1 Jean-Yves Nief.
Mobile Analyzer A Distributed Computing Platform Juho Karppinen Helsinki Institute of Physics Technology Program May 23th, 2002 Mobile.
COMP_3:Grid Interoperability and Data Management CC-IN2P3 and KEK Computing Research Center FJPPL Annecy June 15, 2010.
Lynda : Lyon Neuroimaging Database and Applications (1) Institut des Sciences Cognitives UMR 5015 CNRS ; (2) parallel computing ENS-Lyon ; (3)Centre de.
CC-IN2P3 Pierre-Emmanuel Brinette Benoit Delaunay IN2P3-CC Storage Team 17 may 2011.
© 2014 VMware Inc. All rights reserved. Cloud Archive for vCloud ® Air™ High-level Overview August, 2015 Date.
Building Preservation Environments with Data Grid Technology Reagan W. Moore Presenter: Praveen Namburi.
Jean-Yves Nief, CC-IN2P3 CC-IN2P3 KEK-CCIN2P3 meeting on Grids. September 11th – 12th, 2006.
A possible SRB to iRODS Migration (a path from here to there)‏ Adil Hasan RAL.
OSG Storage Forum: iRODS September 21-22, 2010 DICE integrated Rule- Oriented Data System: iRODS Leesa Brieger.
FIFE Architecture Figures for V1.2 of document. Servers Desktops and Laptops Desktops and Laptops Off-Site Computing Off-Site Computing Interactive ComputingSoftware.
Centre de Calcul de l’Institut National de Physique Nucléaire et de Physique des Particules Data storage services at CC-IN2P3 Jean-Yves Nief.
Cofax Scalability Document Version Scaling Cofax in General The scalability of Cofax is directly related to the system software, hardware and network.
Amazon Web Services. Amazon Web Services (AWS) - robust, scalable and affordable infrastructure for cloud computing. This session is about:
CC-IN2P3 data repositories Jean-Yves Nief. What is CC-IN2P3 ? 04/12/2009CC-IN2P3 data repositories2 Federate computing needs of the french community:
CC-IN2P3: A High Performance Data Center for Research Dominique Boutigny February 2011 Toward a future cooperation with Israel.
CCIN2P3 Site Report - BNL, Oct 18, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center.
Consulting Services JobScheduler Architecture Decision Template
An Overview of iRODS Integrated Rule-Oriented Data System
Introduction to PHP FdSc Module 109 Server side scripting and
Introduction to iRODS Jean-Yves Nief.
CC - IN2P3 Site Report Hepix Spring meeting 2011 Darmstadt May 3rd
Consulting Services JobScheduler Architecture Decision Template
The transfer performance of iRODS between CC-IN2P3 and KEK
Cloud based Open Source Backup/Restore Tool
Getting Started.
Getting Started.
CC-IN2P3 Jean-Yves Nief, CC-IN2P3 HEPiX, SLAC
CC and LQCD dimanche 13 janvier 2019dimanche 13 janvier 2019
STATEL an easy way to transfer data
Presentation transcript:

iRODS at CC-IN2P3: overview Jean-Yves Nief

Talk overview iRODS in production: –Hardware setup. –Usage. –Prospects. iRODS developpements in Lyon: –Scripts. –Micro-services. –Drivers. –Resource Monitoring System. –iCommand. SRB to iRODS migration. 27/01/10iRODS at CC-IN2P3: overview2

iRODS CC-IN2P3 9 servers: –3 iCAT servers (metacatalog): Linux SL4, Linux SL5 –6 data servers (200 TB): Thor x4540, Solaris 10. Metacatalog on a dedicated Oracle 11g cluster. HPSS interface: rfio server (using universal MSS driver). Use of fuse-iRODS: –For Fedora-Commons. –For legacy web applications. TSM: backup of some stored data. Monitoring and restart of the services fully automated (crontab + Nagios + SMURF). Accounting: daily report on our web site. 27/01/10iRODS at CC-IN2P3: overview3

iRODS CC-IN2P3 27/01/10iRODS at CC-IN2P3: overview4 iCAT server iCAT server Data server Data server Data server Data server … … DNS alias: ccirods DNS alias: load balanced. redundancy improved. scalability improved.

iRODS usage at CC-IN2P3: TIDRA TIDRA: Rhône-Alpes area data grid. Used by: –Biology. –Biomedical applications: Animal imagery, Human data.  Automatic bulk metadata registration in iRODS based on DICOM files content (Yonny). –Coming soon: synchrotron data (ESRF – Grenoble). Already 3 millions of files registered. Up to connections per day on iRODS. Authentication: using password or grid certificate. Expecting growth: + 20 TBs or more. 27/01/10iRODS at CC-IN2P3: overview5

iRODS usage at CC-IN2P3: Adonis Adonis: french academic platform for Arts and Humanities. Various projects: –Data archives. –Data online access through: Fedora-commons. Web site. –Data access and processing by batch jobs. iRODS at the heart of Adonis. Already 20 TB of data (2 M files), more than 100 TB at the end of this year. 27/01/10iRODS at CC-IN2P3: overview6

iRODS usage at CC-IN2P3: Adonis Ex: archival and data publication of audio files (CRDO). 27/01/10iRODS at CC-IN2P3: overview7 CRDO CINES CC-IN2P3 1.Data transfer: CRDO  CINES (Montpellier). 2.Archived at CINES. 3.iRODS transfer to CC- IN2P3: iput file.tar 4.Automatic untar at Lyon. 5.Automatic registration in Fedora commons. Fedora Archive

iRODS usage: prospects Starting: –Neuroscience: ~60 TB. –IMXGAM: ~ 15 TB ( X and gamma ray imagery). –dChooz (neutrino experiment): ~ 15 TB. Coming soon: LSST (astro): –For the IN2P3 electronic test-bed: ~ 10 TB. –For the DC3b data challenge: ??? TB (DC3b will produce up to 800 TBs). Thinking about a replacement of light weight transfer tool (bbftp). 27/01/10iRODS at CC-IN2P3: overview8

iRODS developements Scripts: –Test of icommands functionnalities. icommand: –iscan (next release 2.3): admin command. Micro-services: –Access control: flexible firewall. –Msi to tar/untar files and register them in iRODS. –Msi to set ACLs on objects/collections. Universal Mass Storage driver. Miscealeneous (related to the Resource Monitoring System): –Choose best resource based on the load. –Automatic setup of status for a server (up or down). 27/01/10iRODS at CC-IN2P3: overview9

iRODS developements: Resource Monitoring System 27/01/10iRODS at CC-IN2P3: overview10 iRODS iCAT server iRODS data server 1.Ask each server for its metrics: rule engine cron task (msi). 2. Performance script launched on each server. Perf script 3. Results sent back to the iCAT. 4. Store metrics into iCAT. DB 5. Compute a «quality factor» for each server stored in an other table: r.e. cron task (msi).

SRB migration to iRODS SRB as of today: –More than 10 experiments using it. –Key component for all of them. –More than 2 PBs handled by the SRB. –Max network traffic 15 TB / day. –500,000 connections per day, even more. –OS: Windows, MAC, Linux (SL, Debian, Suse), Solaris, AIX. –Clients: from laptop to supercomputers. –Connections as far as Australia and Hawaii. Still growing. 27/01/10iRODS at CC-IN2P3: overview11

SRB migration to iRODS Tricky to do it in production: –Java part ok: almost transparent. –Shell commands: need to adapt client side scripts. Will start this year (2 years process) with BioEmergence (extra 60 TBs in iRODS). Won’t do migration for BaBar and SNFactory. 27/01/10iRODS at CC-IN2P3: overview12

iRODS contacts France: –Observatoire de Strasbourg: IVOA. –National French Library. Elsewhere: –USA: SLAC, NASA, LSU (PetaShare), UNC…. –Canada. –Australia: australian data grid. 27/01/10iRODS at CC-IN2P3: overview13

iRODS overall assessement iRODS is very popular. Very flexible, large amount of functionnalities. As SRB, it allows to go way above simple data transfer. Can be interfaced with many different technologies (no limit): –Cloud, Mass Storage, web services, databases, ….  Able to answer a vast amount of needs. Lot of projects = lot of work for us ! Goal for this year: ~ x00 TB (guess: > 300 TBs). Should reach PB scale very quickly. 27/01/10iRODS at CC-IN2P3: overview14

Acknowledgement Thanks to: –Pascal Calvat. –Yonny Cardenas. –Rachid Lemrani. –Thomas Kachelhoffer. –Pierre-Yves Jallud. 27/01/10iRODS at CC-IN2P3: overview15