Jean-Yves Nief, CC-IN2P3, Lyon First Latin American EELA Workshop April 24th – 26th, 2006 Data distribution and aggregation over geographically distant.

Slides:



Advertisements
Similar presentations
National Partnership for Advanced Computational Infrastructure San Diego Supercomputer Center Data Grids for Collection Federation Reagan W. Moore University.
Advertisements

The Storage Resource Broker and.
The Storage Resource Broker and.
Peter Berrisford RAL – Data Management Group SRB Services.
Database Architectures and the Web
Data Grid: Storage Resource Broker Mike Smorul. SRB Overview Developed at San Diego Supercomputing Center. Provides the abstraction mechanisms needed.
NATIONAL PARTNERSHIP FOR ADVANCED COMPUTATIONAL INFRASTRUCTURE SAN DIEGO SUPERCOMPUTER CENTER Particle Physics Data Grid PPDG Data Handling System Reagan.
San Diego Supercomputer CenterNational Partnership for Advanced Computational Infrastructure1 Grid Based Solutions for Distributed Data Management Reagan.
High Performance Computing Course Notes Grid Computing.
A Very Brief Introduction to iRODS
Security Requirements for Shared Collections Storage Resource Broker Reagan W. Moore
Building on the BIRN Workshop BIRN Systems Architecture Overview Philip Papadopoulos – BIRN CC, Systems Architect.
Chronopolis: Preserving Our Digital Heritage David Minor UC San Diego San Diego Supercomputer Center.
Applying Data Grids to Support Distributed Data Management Storage Resource Broker Reagan W. Moore Ian Fisk Bing Zhu University of California, San Diego.
Jean-Yves Nief, CC-IN2P3 Wilko Kroeger, SCCS/SLAC Adil Hasan, CCLRC/RAL HEPiX, SLAC October 11th – 13th, 2005 BaBar data distribution using the Storage.
Grids and Grid Technologies for Wide-Area Distributed Computing Mark Baker, Rajkumar Buyya and Domenico Laforenza.
Generic policy rules and principles Jean-Yves Nief.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
IRODS usage at CC-IN2P3 Jean-Yves Nief. Talk overview What is CC-IN2P3 ? Who is using iRODS ? iRODS administration: –Hardware setup. iRODS interaction.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
CSC 456 Operating Systems Seminar Presentation (11/13/2012) Leon Weingard, Liang Xin The Google File System.
Users’ Authentication in the VRVS System David Collados California Institute of Technology November 20th, 2003TERENA - Authentication & Authorization.
IRODS performance test and SRB system at KEK Yoshimi KEK Building data grids with iRODS 27 May 2008.
Core SRB Technology for 2005 NCOIC Workshop By Michael Wan And Wayne Schroeder SDSC SDSC/UCSD/NPACI.
Introduction to iRODS Jean-Yves Nief. Talk overview Data management context. Some data management goals: –Storage virtualization. –Virtualization of the.
Presenter: Dipesh Gautam.  Introduction  Why Data Grid?  High Level View  Design Considerations  Data Grid Services  Topology  Grids and Cloud.
Rule-Based Data Management Systems Reagan W. Moore Wayne Schroeder Mike Wan Arcot Rajasekar {moore, schroede, mwan, {moore, schroede, mwan,
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
File and Object Replication in Data Grids Chin-Yi Tsai.
Production Data Grids SRB - iRODS Storage Resource Broker Reagan W. Moore
BaBar Data Distribution using the Storage Resource Broker Adil Hasan, Wilko Kroeger (SLAC Computing Services), Dominique Boutigny (LAPP), Cristina Bulfon.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
4/5/2007Data handling and transfer in the LHCb experiment1 Data handling and transfer in the LHCb experiment RT NPSS Real Time 2007 FNAL - 4 th May 2007.
National Partnership for Advanced Computational Infrastructure San Diego Supercomputer Center Persistent Management of Distributed Data Reagan W. Moore.
National Partnership for Advanced Computational Infrastructure San Diego Supercomputer Center Persistent Archive for the NSDL Reagan W. Moore Charlie Cowart.
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
Test Results of the EuroStore Mass Storage System Ingo Augustin CERNIT-PDP/DM Padova.
The Global Land Cover Facility is sponsored by NASA and the University of Maryland.The GLCF is a founding member of the Federation of Earth Science Information.
GCRC Meeting 2004 BIRN Coordinating Center Software Development Vicky Rowley.
Michael Doherty RAL UK e-Science AHM 2-4 September 2003 SRB in Action.
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
Introduction to The Storage Resource.
Managing Petabytes of data with iRODS at CC-IN2P3
Biomedical Informatics Research Network The Storage Resource Broker & Integration with NMI Middleware Arcot Rajasekar, BIRN-CC SDSC October 9th 2002 BIRN.
National Archives and Records Administration1 Integrated Rules Ordered Data System (“IRODS”) Technology Research: Digital Preservation Technology in a.
Rights Management for Shared Collections Storage Resource Broker Reagan W. Moore
The Storage Resource Broker and.
1 GridFTP and SRB Guy Warner Training, Outreach and Education Team, Edinburgh e-Science.
GDB meeting - Lyon - 16/03/05 An example of data management in a Tier A/1 Jean-Yves Nief.
Lynda : Lyon Neuroimaging Database and Applications (1) Institut des Sciences Cognitives UMR 5015 CNRS ; (2) parallel computing ENS-Lyon ; (3)Centre de.
Building Preservation Environments with Data Grid Technology Reagan W. Moore Presenter: Praveen Namburi.
Jean-Yves Nief, CC-IN2P3 CC-IN2P3 KEK-CCIN2P3 meeting on Grids. September 11th – 12th, 2006.
Collection-Based Persistent Archives Arcot Rajasekar, Richard Marciano, Reagan Moore San Diego Supercomputer Center Presented by: Preetham A Gowda.
A System for Monitoring and Management of Computational Grids Warren Smith Computer Sciences Corporation NASA Ames Research Center.
Preservation Data Services Persistent Archive Research Group Reagan W. Moore October 1, 2003.
Data Infrastructure in the TeraGrid Chris Jordan Campus Champions Presentation May 6, 2009.
1 eScience Grid Environments th May 2004 NESC - Edinburgh Deployment of Storage Resource Broker at CCLRC for E-science Projects Ananta Manandhar.
IRODS at CC-IN2P3: overview Jean-Yves Nief. Talk overview iRODS in production: –Hardware setup. –Usage. –Prospects. iRODS developpements in Lyon: –Scripts.
Grid Services for Digital Archive Tao-Sheng Chen Academia Sinica Computing Centre
Building Preservation Environments from Federated Data Grids Reagan W. Moore San Diego Supercomputer Center Storage.
An Overview of iRODS Integrated Rule-Oriented Data System
Introduction to iRODS Jean-Yves Nief.
Arcot Rajasekar Michael Wan Reagan Moore (sekar, mwan,
CC-IN2P3 Jean-Yves Nief, CC-IN2P3 HEPiX, SLAC
VORB Virtual Object Ring Buffers
STATEL an easy way to transfer data
Presentation transcript:

Jean-Yves Nief, CC-IN2P3, Lyon First Latin American EELA Workshop April 24th – 26th, 2006 Data distribution and aggregation over geographically distant sites.

First Latin American EELA Workshop, April 24th-26th Talk overview. Introduction: big science, big data, big problem. SRB: an example of a mature data management tool. Data management and CC-IN2P3: a few examples in various fields (HEP, astrophysics, biomedical applications) using SRB. Data management elsewhere: some interesting data management applications in various area. Pitfalls and challenges: having chosen the right architecture for your project is not the end of the game. Prospects.

Introduction.

First Latin American EELA Workshop, April 24th-26th The present situation. Large amounts of data produced by scientific projects. Order of magnitude right now: 100 TB, ~ PB, millions of records. In many fields: –High Energy Physics (SLAC, Fermilab, CERN etc..). –Astrophysics (simulation projects: Enzo, Terascale Supernova Initiative…, observational data: Eros, MACHO, 2MASS, USNO-B, SDSS, IVOA …). –Earth sciences (Terashake, Terra …). –Biology / Biomedical research (BIRN …).

First Latin American EELA Workshop, April 24th-26th Prospects for the future. Hard to tell but already some indications. Some examples (next decade): –DOE Genomics, GTL program. –Digital libraries for the US administration (NARA). Order of magnitude: ~ EB, trillions of records!  Amount of data and information exploding.  Wider variety of actors: not only big science! Also true for the networking (next slide: source ESnet).

Science AreasToday End2End Throughput 5 years End2End Throughput 5-10 Years End2End Throughput Remarks High Energy Physics 0.5 Gb/s100 Gb/s1000 Gb/shigh bulk throughput Climate (Data & Computation) 0.5 Gb/s Gb/sN x 1000 Gb/shigh bulk throughput SNS NanoScience Not yet started 1 Gb/s1000 Gb/s + QoS for control channel remote control and time critical throughput Fusion Energy0.066 Gb/s (500 MB/s burst) Gb/s (500MB/ 20 sec. burst) N x 1000 Gb/stime critical throughput Astrophysics0.013 Gb/s (1 TBy/week) N*N multicast 1000 Gb/scomputational steering and collaborations Genomics Data & Computation Gb/s (1 TBy/day) 100s of users1000 Gb/s + QoS for control channel high throughput and steering

First Latin American EELA Workshop, April 24th-26th Living in a digital world. Lots of science or digital library projects involving collaborators / users geographically spread. Large computing needs (both CPU and storage). Need for data backup. And / or need for data closed to the users (replica over different sites). Need for collaborative tools to exchange data.  Federate geographically distributed computing facilities.

First Latin American EELA Workshop, April 24th-26th The dawn of cyberinfrastructure (I). What is that ? « An infrastructure based on grids and on application-specific software, tools, and data repositories that support research in a particular discipline. » Why is it needed ? –Need to handle heterogeneous hardware. –Need to handle heterogeneous OS. –Need to handle heterogeneous storage devices. –Need to handle various preservation policies across the distributed environment.

First Latin American EELA Workshop, April 24th-26th The dawn of cyberinfrastructure (II). Virtualization of the storage. Necessary in order to develop client applications transparent to the technology evolution of the underlying storage systems. Virtual organization: –Access rights. –Groups, domains handling: policies for data sharing. –Preservation policies.

First Latin American EELA Workshop, April 24th-26th Requirements (I). Infrastructure independence: Data virtualization: –Management of name spaces independently of the storage repositories. –Support for access operations independently of the storage repositories. Authentication: –Certificate: GSI etc… –Challenge-response mechanism: no pwd sent over the network. –Encrypted password. –Ticket: valid for a given amount of time to access the virtual organization.

First Latin American EELA Workshop, April 24th-26th Requirements (II). Data ownership / Authorization: –Management of the files’ownership across multiple sites (partial or total decoupling between each sites organization and virtual organization). –Access Control Lists valid for the entire virtual organization across the physical domain (group, user etc… levels).

First Latin American EELA Workshop, April 24th-26th Requirements (III). Data operations: –File access: Open, close, read, write, stat… Audit, versions, pinning, checksums, synchronize etc… Parallel I/O, firewall interactions. –Latency management: Bulk operations: register, load, unload, delete etc… Remote procedures: replicate, aggregate, file parsing, I/O requests (FITS, DICOM, HDF5 …). –Metadata management: Annotations, metadata/auditing queries, interface with various information systems (schema extension of the core system).

SRB: Storage Resource Broker.

First Latin American EELA Workshop, April 24th-26th What’s SRB ? Storage Resource Broker: developed by SDSC (San Diego). Provides an uniform interface to heterogeneous storage system (disk, tape, databases) for data distributed in multiple sites. Collaborative tool to share files. Who is using SRB ? –Biology, biomedical applications (e.g: BIRN). –Astrophysics, Earth Sciences (e.g: NASA). –Digital libraries (e.g: NARA). Used world wide: USA, Europe, Asia, Australia.

First Latin American EELA Workshop, April 24th-26th SRB architecture. 1 zone: –1 SRB/MetaCatalog server: contains list of files, physical resources, users registered, etc… –several SRB servers to access the data at their physical location. Site 1 SRB Site 2 SRB Site 3 Application (asking for test1.txt, connecting to site 2) SRB MCAT (1) (4) (2) test1.txt (3)

First Latin American EELA Workshop, April 24th-26th Some SRB features. Files organized in a logical name space with directories, subdirectories. > /home/nief.ccin2p3 # dir /home/nief.ccin2p3 evs_g_isPhysicsEvents_aod # on CC-IN2P3 test1.txt # on Merida Handling replica. Search for the files based on their attributes instead of their physical name and location (site, storage type: disk, tape, databases).  Search by metadata « attached » to the files.

First Latin American EELA Workshop, April 24th-26th Users and ACLs management. Users belong to: –1 zone (ex: IN2P3, Venezuela …). –1 domain (ex: ccin2p3, Merida, Caracas). –1 or several groups. ACL on files and directories. Tickets: –Rights given to temporary users for a limited amount of time.

First Latin American EELA Workshop, April 24th-26th Storage. Mass Storage System (MSS): interface provided for HPSS, Castor and many other MSS. small files management (containers). MSS usage (tapes etc…) transparent for the end user. Logical resources: set of physical resources. resource1: file system IAP resource2: hpss file CC-IN2P3 resource3: file system Merida Able to put a file in the 3 resources in one shot: > Sput –S logical-res test1.txt logical-res

First Latin American EELA Workshop, April 24th-26th Databases. Access to databases through SRB: –Security: SRB server = proxy server  Database can be shielded from the outer world, control on the requests submitted to the database server. –Duplication: very simple copy from a database at one site to an other one. (e.g.: copy of tables from a Oracle db in Lyon to a mySQL db at site X in one shot). Schema extension: –Possibility to link the SRB-MCAT with some other databases (search on SRB objects based on attributes stored in an other db).

First Latin American EELA Workshop, April 24th-26th Interfaces, portability. Interfaces: –Binary commands (Scommands). –APIs: C, Java, Perl, Python. –Web interface (mySRB). –GUI client for Windows (inQ). Portability: –Linux, Windows, Mac OS, Solaris and many more… Databases: –Oracle, DB2, Sybase, PostgreSQL, Informix, mySQL…

Data management and CC-IN2P3: examples using SRB.

First Latin American EELA Workshop, April 24th-26th Who is using CC-IN2P3 ? In green = pre-production. High Energy Physics: –BaBar (SLAC, Stanford). –CMOS (International Linear Collider R&D). –Calice (International Linear Collider R&D). Astroparticle: –Edelweiss (Modane, France). –Pierre Auger Observatory (Argentina). Astrophysics: –SuperNovae Factory (Hawaii). Biomedical applications: –Neuroscience research.

First Latin American EELA Workshop, April 24th-26th Babar, SLAC & CC-IN2P3. BaBar: High Energy Physics experiment closed to Stanford (California). SLAC and CC-IN2P3 first opened to the BaBar collaborators data analysis. Both held complete copies of data (Objectivity). Now only SLAC hold a complete copy of the data. Natural candidates for testing and deployment of grid middleware. Data should be available in a delay of 24/48 hours. SRB: chosen for data distribution of hundreds of TBs of data.

First Latin American EELA Workshop, April 24th-26th SRB BaBar architecture. CC-IN2P3 (Lyon) HPSS/Lyon SRB SLAC (Stanford, CA) SRB MCAT (1) (3) (2) HPSS/SLAC SRB MCAT 2 Zones (SLAC + Lyon)

First Latin American EELA Workshop, April 24th-26th Extra details (BaBar). Hardware: –SUN servers (Solaris 5.8, 5.9): NetraT 1405, V440. Software: –Oracle 10g for the SLAC MCAT. –Oracle 9i for the Lyon MCAT (migration to 10g foreseen). MCATs synchronization: only users and physical resources. Comparison of the MCATs contents to transfer the data. Step (1), (2), (3) multithreaded under client control: very little latency. Advantage: –External client can pick up data from SLAC or Lyon without interacting with the other site.

First Latin American EELA Workshop, April 24th-26th Overall assessment for BaBar. A lot of time saved for developping applications thanks to the SRB. Transparent access to data: –Very useful in an hybrid environment (disk, tape). –Easy to scale the service (adding new servers on the fly). –Not dependent of physical locations changes in the client application. Fully automated procedure. Easy for SLAC to recover corrupted data. 270 TB (460,000 files) shipped to Lyon. Up to 3 TB /day from tape to tape (minimum latency). Going to 5 TB / day soon/

First Latin American EELA Workshop, April 24th-26th Fermilab (US)  CERN SLAC (US)  IN2P3 (FR) 1 Terabyte/day SLAC (US)  INFN Padva (IT) Fermilab (US)  U. Chicago (US) CEBAF (US)  IN2P3 (FR) INFN Padva (IT)  SLAC (US) U. Toronto (CA)  Fermilab (US) Helmholtz-Karlsruhe (DE)  SLAC (US) DOE Lab  DOE Lab SLAC (US)  JANET (UK) Fermilab (US)  JANET (UK) Argonne (US)  Level3 (US) Argonne  SURFnet (NL) IN2P3 (FR)  SLAC (US) Fermilab (US)  INFN Padva (IT) ESNET Traffic with one server on both sides (April 2004).

Neuroscience research (P. Calvat). DICOM IRM Siemens MAGNETOM Sonata Maestro Class 1.5 T (Lyon hospital) Consol Siemens Celsius Xeon (Window NT) Acquisition DICOM Export PC Dell PowerEdge 800  FTP,  File sharing,  … DICOM

First Latin American EELA Workshop, April 24th-26th Neuroscience research (II). Goal: make SRB invisible to the end user. More than 500,000 files registered. Now interfaced within the MATLAB environment: –Data pushed where the CPUs are (CC-IN2P3, ENS Lyon). ~ 1.5 FTE for 3 months… Next step: –Ever growing community (a few TBs / year): Strasbourg hospital to join the project (maybe Marseille, St Etienne…). –Goal: Join the BIRN network (US biomedical network).

SuperNovae Factory. Telescope data stored into the SRB, processed in Lyon (almost online). Collaborative tool + backup (files exchanged between French and US users). Hawaii telescope HPSS/Lyon SRB CC-IN2P3 a few GBs / day SRB HPSS/NERSC Berkeley (project)

Data management elsewhere: a few examples.

Neuroscience: BIRN (I). BIRN = BioInformatics Research Network Brain imagery (human, animals: mice, apes): - fMRI etc… Data sharing and exchange of experimental data for each lab and project.

First Latin American EELA Workshop, April 24th-26th Neuroscience: BIRN (II). BIRN Coordination Center in San Diego: –1 rack (SRB server, database etc…) on each site. –Administration centralized from the BIRN-CC: 24/7. –Sharing software, APIs… –15 millions of files registered (16 TB), 360 users: file search on metadata over the entire sample (impressive!). John Hopkins Hospital: « done more in 6 months than in 18 years ». BIRN: 30 people at the first meeting (2001), 115 in Feb. 2005, more than 200 now  success. Some sites already starting in Europe: Edinburgh, Manchester. Hoping for a french site in the near future.

First Latin American EELA Workshop, April 24th-26th ROADNet (UCSD). Real-time Observatories, Applications, and Data-Management Network. The Problem: –Integrated real-time management of large, distributed, heterogeneous data streams from sensor networks. –Sensors: Seismometers, Accelerometers, Displacement, Barometric pressure, Temperature, Wind Speed, Wind Direction, Infrasound, Hydroacoustic, Differential Pressure Gauges, Strain, Solar Insolation, pH, Electric Current, Electric Potential, Dilution of oxygen, Still Camera Images, Codar. –Multidisciplinary project: Sismology. Oceanography. Hydrology. Meteorology. Etc…

First Latin American EELA Workshop, April 24th-26th ROADNet (UCSD). DATASCOPE Archives/Processing/Review It’s a grid for online studies, handling data streams. (ORB= Online Ring Buffer)

Pitfalls and challenges.

First Latin American EELA Workshop, April 24th-26th Potential pitfalls. To build a successfull environment for data management and distribution over many sites: –Good coordination and communication between the sites administrators: « social » factor. –Manpower: expertise needed in several area (network, sys. admin. and database administration). –Working in different time zones does not make things easy. –Development of monitoring tools. –Automatic recovery of the services in case of services problem: decrease downtime of the services.

First Latin American EELA Workshop, April 24th-26th Hardware requirements. Network: –% packet loss must be low. –High latency network (Round Trip Time > 100 ms): potential show stopper. –Duplication of information services (databases) should be considered (e.g: Belle grid extending in Australia, Japan, South Korea). Servers hardware: –Disk arrays quality: data corruption etc… –Data duplication can be a show stopper in terms of budget. –Database servers scaled correctly.

First Latin American EELA Workshop, April 24th-26th Other requirements. Data integrity (checksum). Backup policy in order to prevent data loss. Scalability of the middleware. Middleware must be multi OS. Fault tolerance of the system. Compatibility of the client application version as a function of the midleware evolution: prevent tough and painfull migration to newer version. Middleware must be as transparent as possible to hardware, databases etc…evolution.

First Latin American EELA Workshop, April 24th-26th Challenges. Is a grid environment always the solution? Not sure !!! Cost in terms of: –Hardware. –Networking. –Manpower (more duplicated sites, more data, more admins). can be prohibitive.

Prospects.

First Latin American EELA Workshop, April 24th-26th Summary and outlook (I). Middleware needed for an efficient data management over multiple site. Scalability might be an issue in the future for the information systems (databases) linked to these middleware: –Inflation of metadata. –Inflation of files.  Web services: not sure that it should not be at the centre of data distribution. Economic and manpower costs often neglected.

First Latin American EELA Workshop, April 24th-26th Summary and outlook (II). SRB: a very good candidate. Is there a real competitor at the moment ? RODS (Rule Oriented Data management System): –Replacement of SRB (open source). –Compatible with SRB (SRB client application could connect to a RODS server). –SDSC leading the project. –CC-IN2P3, one of a few partners going to be involved in the first step.

First Latin American EELA Workshop, April 24th-26th Acknowledgement. Many thanks to: –Reagan Moore and his team (SDSC, USA). –Adil Hasan (CCLRC-RAL, UK). –Wilko Kroeger (SLAC, USA). –Pascal Calvat (CC-IN2P3, France). BaBar: Belle: BIRN: CC-IN2P3: ESnet: ROADNet: SRB: