1Antonella Bozzi – LSC/Virgo meeting Amsterdam 24-9-08 Ligo/Virgo Data Transfer Bulk data replication tools and plans for S6 data replication Antonella.

Slides:



Advertisements
Similar presentations
Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
Advertisements

NATIONAL PARTNERSHIP FOR ADVANCED COMPUTATIONAL INFRASTRUCTURE SAN DIEGO SUPERCOMPUTER CENTER Particle Physics Data Grid PPDG Data Handling System Reagan.
Grid simulation (AliEn) Eugen Mudnić Technical university Split -FESB.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Robust Tools for Archiving and Preserving Digital Data Joseph JaJa, Mike Smorul, and Mike McGann Institute for Advanced Computer Studies Department of.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Workload Management Massimo Sgaravatto INFN Padova.
Magda – Manager for grid-based data Wensheng Deng Physics Applications Software group Brookhaven National Laboratory.
Data Grid Web Services Chip Watson Jie Chen, Ying Chen, Bryan Hess, Walt Akers.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Data Management Kelly Clynes Caitlin Minteer. Agenda Globus Toolkit Basic Data Management Systems Overview of Data Management Data Movement Grid FTP Reliable.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
3 Sept 2001F HARRIS CHEP, Beijing 1 Moving the LHCb Monte Carlo production system to the GRID D.Galli,U.Marconi,V.Vagnoni INFN Bologna N Brook Bristol.
LHCb Applications and GRID Integration Domenico Galli Catania, April 9, st INFN-GRID Workshop.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
The Data Grid: Towards an Architecture for the Distributed Management and Analysis of Large Scientific Dataset Caitlin Minteer & Kelly Clynes.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
EGO Computing Center site report EGO - Via E. Amaldi S. Stefano a Macerata - Cascina (PI) | Stefano Cortese INFN Computing Workshop –
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
Cosener’s House – 30 th Jan’031 LHCb Progress & Plans Nick Brook University of Bristol News & User Plans Technical Progress Review of deliverables.
1 Use of SRMs in Earth System Grid Arie Shoshani Alex Sim Lawrence Berkeley National Laboratory.
8th November 2002Tim Adye1 BaBar Grid Tim Adye Particle Physics Department Rutherford Appleton Laboratory PP Grid Team Coseners House 8 th November 2002.
Problemi e strategie relativi al calcolo off-line di VIRGO Laura Brocco Universita’ di Roma “La Sapienza” & INFN Roma1 for the VIRGO collaboration.
Production Data Grids SRB - iRODS Storage Resource Broker Reagan W. Moore
SAM and D0 Grid Computing Igor Terekhov, FNAL/CD.
Describe workflows used to maintain and provide the RDA to users – Both are 24x7 operations Transition to the NWSC with zero downtime NWSC is new environment.
Grid Workload Management Massimo Sgaravatto INFN Padova.
CYBERINFRASTRUCTURE FOR THE GEOSCIENCES Data Replication Service Sandeep Chandra GEON Systems Group San Diego Supercomputer Center.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
6-10 Oct 2002GREX 2002, Pisa D. Verkindt, LAPP 1 Virgo Data Acquisition D. Verkindt, LAPP DAQ Purpose DAQ Architecture Data Acquisition examples Connection.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
MAGDA Roger Jones UCL 16 th December RWL Jones, Lancaster University MAGDA  Main authors: Wensheng Deng, Torre Wenaus Wensheng DengTorre WenausWensheng.
NOVA Networked Object-based EnVironment for Analysis P. Nevski, A. Vaniachine, T. Wenaus NOVA is a project to develop distributed object oriented physics.
Rule-Based Programming for VORBs Bertram Ludaescher Arcot Rajasekar Data and Knowledge Systems San Diego Supercomputer Center U.C. San Diego.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
Part Four: The LSC DataGrid Part Four: LSC DataGrid A: Data Replication B: What is the LSC DataGrid? C: The LSCDataFind tool.
PHENIX and the data grid >400 collaborators Active on 3 continents + Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
Transaction-based Grid Data Replication Using OGSA-DAI Presented by Yin Chen February 2007.
The DataTransfer status Experience on VSR2 A. Bozzi, L. Salconi – 27 Oct 2009.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
LIGO-G E LIGO Scientific Collaboration Data Grid Status Albert Lazzarini Caltech LIGO Laboratory Trillium Steering Committee Meeting 20 May 2004.
Hepix LAL April 2001 An alternative to ftp : bbftp Gilles Farrache In2p3 Computing Center
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Implementation and performance analysis of.
PHENIX and the data grid >400 collaborators 3 continents + Israel +Brazil 100’s of TB of data per year Complex data with multiple disparate physics goals.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Scott Koranda, UWM & NCSA 14 January 2016www.griphyn.org Lightweight Data Replicator Scott Koranda University of Wisconsin-Milwaukee & National Center.
AHM04: Sep 2004 Nottingham CCLRC e-Science Centre eMinerals: Environment from the Molecular Level Managing simulation data Lisa Blanshard e- Science Data.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
D0 File Replication PPDG SLAC File replication workshop 9/20/00 Vicky White.
10 March Andrey Grid Tools Working Prototype of Distributed Computing Infrastructure for Physics Analysis SUNY.
Homework 1 solutions. Question 1 Solution Q1 Question 2.
Simulation Production System Science Advisory Committee Meeting UW-Madison March 1 st -2 nd 2007 Juan Carlos Díaz Vélez.
Get Data to Computation eudat.eu/b2stage B2STAGE How to shift large amounts of data Version 4 February 2016 This work is licensed under the.
What to expect in TB 2.0 J. Templon, NIKHEF/WP8. Jeff Templon – AWG Meeting, NIKHEF, WP1 u New resource broker architecture Big potato.
The bulk dataTransfer procedures (CNAF, Nov 2015) L. Salconi, A. Bozzi.
Review of PARK Reflectometry Group 10/31/2007. Outline Goal Hardware target Software infrastructure PARK organization Use cases Park Components. GUI /
Simulation Production System
Replica/File Catalog based on RAM SUNY
The Data Grid: Towards an architecture for Distributed Management
U.S. ATLAS Grid Production Experience
An Overview of iRODS Integrated Rule-Oriented Data System
Moving the LHCb Monte Carlo production system to the GRID
The VIRGO DATA ANALYSIS Fulvio Ricci
GGF OGSA-WG, Data Use Cases Peter Kunszt Middleware Activity, Data Management Cluster EGEE is a project funded by the European.
Distributed Data Access and Resource Management in the D0 SAM System
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
VORB Virtual Object Ring Buffers
Technical Capabilities
Presentation transcript:

1Antonella Bozzi – LSC/Virgo meeting Amsterdam Ligo/Virgo Data Transfer Bulk data replication tools and plans for S6 data replication Antonella Bozzi, Stefano Cortese, Livio Salconi European Gravitational Observatory LSC/Virgo meeting Amsterdam

2Antonella Bozzi – LSC/Virgo meeting Amsterdam Useful for data produced continuously that must be replicated permanently at fixed locations 2.More complex than batch transfers of fixed datasets 3.Not suited for user driven transfers of distributed data listed in catalogs Whatever the transfer tool (bbftp, srb, etc.) the real engine is the process that manages the whole automatic procedure: Discovers newly produced files and builds the queues Builds and schedules the transfer jobs (preserving time ordering) Manages the errors (storage at the endpoints and link errors) and resends the files in error Places the files at the target location (creating a directory structure or registering in a database) Bulk in-time replication

3Antonella Bozzi – LSC/Virgo meeting Amsterdam In-time data fluxes Still to refine, based on VSR1-S5 The rates are DAQ rates, the real rate may occur in burst. Post-processed data not considered (Hrec reconstruction) Lyon HPSS Bologna disks/GPFS Cascina disks/NFS LSC raw+trend+hrec 6-7 MB/s Virgo hrec 150KB/s Ligo hrec 130KB/s raw+trend+hrec 6-7 MB/s

4Antonella Bozzi – LSC/Virgo meeting Amsterdam Layout at Cascina frame builders (DAQ) 4TB 10krpm SCSI STOL2 (redundant stream) 4TB 15krpm Fiber STOL1 (main stream)‏ ITF network Storage network dataFinder.pl fflGen.pl dataManager.pl dataSender.pl 140TB Storage farm ST1 - ST10 users remote repos. rawdata flow control msgs flow

5Antonella Bozzi – LSC/Virgo meeting Amsterdam Current architecture: single flux Storage (local)‏ Storage (remote)‏ dataFinder.pl dataSender.pl bbftp (client)‏ bbftpd (daemon) Control flow Data flow

6Antonella Bozzi – LSC/Virgo meeting Amsterdam Various transfer workhorses: bbftp: the simplest, optimized for the transfer only SRB : adapted to work with Fr library, Xrootd+HPSS Is more than a transfer tool, file must be placed in the SRB/Xrootd environment LDR: is more than a transfer tool, handles file metadata and replicas GRID-FTS: is much more than a transfer tool, actually part of the whole GRID architecture, handles natively the file catalog Transfer tools

7Antonella Bozzi – LSC/Virgo meeting Amsterdam Upgrade of Internet link at Cascina: 1Gbps (now 100Mbps) Available throughput from 100Mbps to 500Mbps More simultaneous transfers possible Pass from looping scripts to event driven processes Handle more robustly the exceptions Current study focuses on: Passing from perl to phython Handle queues with MySQL Making the process modular respect to the transfer tool and the placement at the CCs. Improvements under study

8Antonella Bozzi – LSC/Virgo meeting Amsterdam On-line data They go on the same links but should be given priority Post processed data (i.e. different hrec versions) eligible for batch transfers, but maybe suitable to be accessible on- demand in a GRID-like structure Placement at the CCs very tightened with the access method chosen and the file catalog Better decouple it ? Other related issues