Federating Data in the ALICE Experiment

Slides:



Advertisements
Similar presentations
Windows Server ® 2008 File Services Infrastructure Planning and Design Published: June 2010 Updated: November 2011.
Advertisements

ALICE G RID SERVICES IP V 6 READINESS
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
14.1 © 2004 Pearson Education, Inc. Exam Planning, Implementing, and Maintaining a Microsoft Windows Server 2003 Active Directory Infrastructure.
Hands-On Microsoft Windows Server 2003 Administration Chapter 5 Administering File Resources.
MCTS Guide to Microsoft Windows Server 2008 Network Infrastructure Configuration Chapter 8 Introduction to Printers in a Windows Server 2008 Network.
MCTS Guide to Microsoft Windows Server 2008 Network Infrastructure Configuration Chapter 11 Managing and Monitoring a Windows Server 2008 Network.
Staging to CAF + User groups + fairshare Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE Offline week,
ALICE DATA ACCESS MODEL Outline ALICE data access model - PtP Network Workshop 2  ALICE data model  Some figures.
Experiences Deploying Xrootd at RAL Chris Brew (RAL)
G RID SERVICES IP V 6 READINESS
ALICE data access WLCG data WG revival 4 October 2013.
October 8, 2015 University of Tulsa - Center for Information Security Microsoft Windows 2000 DNS October 8, 2015.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES P. Saiz (IT-ES) AliEn job agents.
CYBERINFRASTRUCTURE FOR THE GEOSCIENCES Data Replication Service Sandeep Chandra GEON Systems Group San Diego Supercomputer Center.
Costin Grigoras ALICE Offline. In the period of steady LHC operation, The Grid usage is constant and high and, as foreseen, is used for massive RAW and.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
1 Chapter Overview Preparing to Upgrade Performing a Version Upgrade from Microsoft SQL Server 7.0 Performing an Online Database Upgrade from SQL Server.
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
N EWS OF M ON ALISA SITE MONITORING
CN2140 Server II Kemtis Kunanuraksapong MSIS with Distinction MCT, MCITP, MCTS, MCDST, MCP, A+
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
Site operations Outline Central services VoBox services Monitoring Storage and networking 4/8/20142ALICE-USA Review - Site Operations.
ALICE – networking LHCONE workshop 10/02/ Quick plans: Run 2 data taking Both for Pb+Pb and p+p – Reach 1 nb -1 integrated luminosity for rare.
Update on replica management
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES A. Abramyan, S. Bagansco, S. Banerjee, L. Betev, F. Carminati,
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES A. Abramyan, S. Bagansco, S. Banerjee, L. Betev, F. Carminati,
SLACFederated Storage Workshop Summary For pre-GDB (Data Access) Meeting 5/13/14 Andrew Hanushevsky SLAC National Accelerator Laboratory.
Xrootd Monitoring and Control Harsh Arora CERN. Setting Up Service  Monalisa Service  Monalisa Repository  Test Xrootd Server  ApMon Module.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
Linux Operations and Administration
FTP COMMANDS OBJECTIVES. General overview. Introduction to FTP server. Types of FTP users. FTP commands examples. FTP commands in action (example of use).
JAliEn Java AliEn middleware A. Grigoras, C. Grigoras, M. Pedreira P Saiz, S. Schreiner ALICE Offline Week – June 2013.
AliEn central services Costin Grigoras. Hardware overview  27 machines  Mix of SLC4, SLC5, Ubuntu 8.04, 8.10, 9.04  100 cores  20 KVA UPSs  2 * 1Gbps.
ALICE DATA ACCESS MODEL Outline 05/13/2014 ALICE Data Access Model 2  ALICE data access model  Infrastructure and SE monitoring.
EGI-Engage Data Services and Solutions Part 1: Data in the Grid Vincenzo Spinoso EGI.eu/INFN Data Services.
Rights Management for Shared Collections Storage Resource Broker Reagan W. Moore
03/09/2007http://pcalimonitor.cern.ch/1 Monitoring in ALICE Costin Grigoras 03/09/2007 WLCG Meeting, CHEP.
Dynamic staging to a CAF cluster Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE CAF / PROOF Workshop,
CMS: T1 Disk/Tape separation Nicolò Magini, CERN IT/SDC Oliver Gutsche, FNAL November 11 th 2013.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES The AliEn File Catalogue Jamboree on Evolution of WLCG Data &
1 DIRAC Data Management Components A.Tsaregorodtsev, CPPM, Marseille DIRAC review panel meeting, 15 November 2005, CERN.
Good user practices + Dynamic staging to a CAF cluster Jan Fiete Grosse-Oetringhaus, CERN PH/ALICE CUF,
Building Preservation Environments with Data Grid Technology Reagan W. Moore Presenter: Praveen Namburi.
New Features of Xrootd SE Wei Yang US ATLAS Tier 2/Tier 3 meeting, University of Texas, Arlington,
11 DEPLOYING AN UPDATE MANAGEMENT INFRASTRUCTURE Chapter 6.
Storage discovery in AliEn
1 Remote Installation Service Windows 2003 Server Prof. Abdul Hameed.
The ALICE Analysis -- News from the battlefield Federico Carminati for the ALICE Computing Project CHEP 2010 – Taiwan.
WLCG IPv6 deployment strategy
ALICE internal and external network
Oxana Smirnova, Jakob Nielsen (Lund University/CERN)
Vincenzo Spinoso EGI.eu/INFN
ALICE Monitoring
Data Federation with Xrootd Wei Yang US ATLAS Computing Facility meeting Southern Methodist University, Oct 11-12, 2011.
Torrent-based software distribution
Storage elements discovery
Offline shifter training tutorial
Ákos Frohner EGEE'08 September 2008
Publishing ALICE data & CVMFS infrastructure monitoring
Stephen Burke, PPARC/RAL Jeff Templon, NIKHEF
CTA: CERN Tape Archive Overview and architecture
Computing Infrastructure for DAQ, DM and SC
An Introduction to Computer Networking
20409A 7: Installing and Configuring System Center 2012 R2 Virtual Machine Manager Module 7 Installing and Configuring System Center 2012 R2 Virtual.
Presentation transcript:

Federating Data in the ALICE Experiment Costin.Grigoras@cern.ch

Federating Data in the ALICE Experiment Outline Data access methods in ALICE Storage AAA Storage monitoring SE discovery 2012.08.07 Federating Data in the ALICE Experiment

Data access methods in ALICE Central catalogue of logical file names (LFN) With owner:group and unix-style permissions Size, MD5 of files Metadata on subtrees Each LFN is associated a GUID that can have any number of replicas (PFNs) root://<redirector>//<HH>/<hhhhh>/<GUID> HH and hhhhh are hashes of the GUID Same namespace on all storage elements Files are immutable on the SEs 2012.08.07 Federating Data in the ALICE Experiment

Data access methods in ALICE (2) Exclusive use of xrootd protocol Jobs are (usually) only downloading configuration files with xrdcp Data files are accessed remotely The closest working replica to the job Jobs go to where a copy of the data is At the end of the job N (2..4 typically) replicas are uploaded from the job itself (xrdcp again) Scheduled data transfers only for raw data Andreas' xrd3cp tool (disk server-to-disk server) 2012.08.07 Federating Data in the ALICE Experiment

Federating Data in the ALICE Experiment Some figures 59 disk SEs, 9 tape SEs (T0 and T1s) 57x xrootd, 1x EOS, 2x DPM, 4x CASTOR, 4x dCache 25PB in 220M PFNs Average replication factor is 3 2 copies of the raw data on MSS: Full copy at CERN T0 One distributed copy at T1s (full runs) 2012.08.07 Federating Data in the ALICE Experiment

Federating Data in the ALICE Experiment Storage AAA Storage-independent Handled centrally by the Authen AliEn service Checks client credentials and catalogue permissions and issues access tickets XML block signed and encrypted by Authen The client hands these tickets to the respective storage and (for writes) notifies the catalogue of the successful operation Implemented as xrootd plugin 2012.08.07 Federating Data in the ALICE Experiment

Federating Data in the ALICE Experiment Storage AAA (2) 2012.08.07 Federating Data in the ALICE Experiment

Storage AAA – in deployment Similar to what is in production now Simplified tickets Less text, just signed (no encryption any more) Introducing storage reply envelopes Size and checksum of what the server got Signed by the storage and returned by xrdcp, xrdstat When committing a write the above must match what was booked Can later recheck the files for consistency 2012.08.07 Federating Data in the ALICE Experiment

Storage AAA – in deployment (2) 2012.08.07 Federating Data in the ALICE Experiment

Monitoring – host parameters Integrated in the overall monitoring of ALICE xrootd plugin package also brings a host and service monitoring daemon Monitoring data from xrootd and the daemon is sent to the site MonALISA instance Collected by the central repository and aggregated per cluster http://alimonitor.cern.ch?571 Under deployment: xrootd 3.2.2 with extended monitoring information 2012.08.07 Federating Data in the ALICE Experiment

Storage monitoring – functional tests add / get / delete performed every 2h From a central location Using the full AliEn suite (like any user or job) Results archived for a “reliability” metric Last week * 25% + last day * 75% Separate metrics for read and write 2012.08.07 Federating Data in the ALICE Experiment

Network topology discovery Site MonALISA instances perform between every pair of them Traceroute / tracepath Bandwidth estimation Recording all details we get a good and complete picture of the network topology 2012.08.07 Federating Data in the ALICE Experiment

Federating Data in the ALICE Experiment SE discovery Based on a dynamic “distance” metric from an IP address to a SE Starting from the network topology Same site, same AS, same country, continent... Last functional test excludes non-working SEs Altered by Reliability Remaining free space A small random factor to assure 'democratic' data distribution 2012.08.07 Federating Data in the ALICE Experiment

Federating Data in the ALICE Experiment SE discovery (2) Reading from the closest working replica Simply sorting by the metric for reading, including the non-working SEs, as last resort Writing to the closest working SEs Each SE is associated a tag (“disk”, “tape”, “paper”) Users indicate the number of replicas of each type Default is “disk=2” Not excluding the option of specific target SEs Keep asking until the requirements are met or no more SEs left to try 2012.08.07 Federating Data in the ALICE Experiment

Federating Data in the ALICE Experiment Summary ALICE has the entire storage space federated Via a central catalogue Specific AAA plugin in each SE The storage layer doesn't have any idea about the users and their rights No entity is to be trusted ROOT support as TAlienFile (working with LFNs) The distributed storage infrastructure is transparent to the users Automatically managed 2012.08.07 Federating Data in the ALICE Experiment

Other answers to Fabrizio's questions From a site we only need the entry point and not to delete the automatic alerts automatically The sites are oblivious to our activity, they just provide the boxes, no* customization or reporting* needed Uniform configuration and instrumentation (at least for the xrootd ones) Clearly separated components, recyclable Sites seem quite happy to run a script and forget about the ALICE storage afterwards 2012.08.07 Federating Data in the ALICE Experiment