ATLAS Distributed Analysis tests in the Spanish Cloud

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
A tool to enable CMS Distributed Analysis
December 17th 2008RAL PPD Computing Christmas Lectures 11 ATLAS Distributed Computing Stephen Burke RAL.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
Your university or experiment logo here Caitriana Nicholson University of Glasgow Dynamic Data Replication in LCG 2008.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
ATLAS Distributed Analysis Experiences in STEP'09 Dan van der Ster for the DA stress testing team and ATLAS Distributed Computing WLCG STEP'09 Post-Mortem.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
Experiment Support ANALYSIS FUNCTIONAL AND STRESS TESTING Dan van der Ster, CERN IT-ES-DAS for the HC team: Johannes Elmsheuser, Federica Legger, Mario.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
PanDA Status Report Kaushik De Univ. of Texas at Arlington ANSE Meeting, Nashville May 13, 2014.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks CRAB: the CMS tool to allow data analysis.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
LHCb report to LHCC and C-RSG Philippe Charpentier CERN on behalf of LHCb.
1 Andrea Sciabà CERN The commissioning of CMS computing centres in the WLCG Grid ACAT November 2008 Erice, Italy Andrea Sciabà S. Belforte, A.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Max Baak 1 Efficient access to files on Castor / Grid Cern Tutorial Max Baak, CERN 30 October 2008.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Response of the ATLAS Spanish Tier2 for.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Analysis vs Storage 4 the LHC Experiments (… ATLAS biased view ) Alessandro Di Girolamo CERN / INFN.
1 ATLAS-VALENCIA COORDINATION MEETING 9th May 2008 José SALT TIER2 Report 1.- Present Status of the ES-ATLAS-T2 2.- Computing-Analysis issues Overview.
WLCG November Plan for shutdown and 2009 data-taking Kors Bos.
ATLAS TIER3 in Valencia Santiago González de la Hoz IFIC – Instituto de Física Corpuscular (Valencia)
Presented by: Álvaro Fernandez Casani IFIC – Valencia (Spain) eScience Intrastructure T2-T3 for High Energy Physics Santiago.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
EXPERIENCE WITH ATLAS DISTRIBUTED ANALYSIS TOOLS S. González de la Hoz L. March IFIC, Instituto.
Analysis Facility Infrastructure: ATLAS in Valencia S. González de la Hoz (On behalf of Tier3 Team) IFIC – Institut de Física Corpuscular de València First.
ATLAS Computing Model Ghita Rahal CC-IN2P3 Tutorial Atlas CC, Lyon
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
1-2 March 2006 P. Capiluppi INFN Tier1 for the LHC Experiments: ALICE, ATLAS, CMS, LHCb.
1 Presented by : José F. Salt Cairols UAM (Madrid), 5 th and 6th June 2012 XIII Presential Meeting of the Spanish ATLAS TIER-2.
Presented by: Santiago González de la Hoz IFIC – Valencia (Spain) Experience running a distributed Tier-2 and an Analysis.
S. González de la Hoz, Step09 at IFIC, Spanish ATLAS Tier2 meeting Step09 at IFIC Javier Sánchez Alejandro Lamas Santiago González IFIC – Institut.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
ATLAS Distributed Analysis S. González de la Hoz 1, D. Liko 2, L. March 1 1 IFIC – Valencia 2 CERN.
ATLAS TIER3 in Valencia Santiago González de la Hoz IFIC – Instituto de Física Corpuscular (Valencia)
Daniele Bonacorsi Andrea Sciabà
Status: ATLAS Grid Computing
Real Time Fake Analysis at PIC
LCG Service Challenge: Planning and Milestones
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
Grid site as a tool for data processing and data analysis
Overview of the Belle II computing
Added value of new features of the ATLAS computing model and a shared Tier-2 and Tier-3 facilities from the community point of view Gabriel Amorós on behalf.
Jan 12, 2005 Improving CMS data transfers among its distributed Computing Facilities N. Magini CERN IT-ES-VOS, Geneva, Switzerland J. Flix Port d'Informació.
Data Challenge with the Grid in ATLAS
Bernd Panzer-Steindel, CERN/IT
Status and Prospects of The LHC Experiments Computing
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
PanDA in a Federated Environment
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Readiness of ATLAS Computing - A personal view
The LHC Computing Grid Visit of Her Royal Highness
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Simulation use cases for T2 in ALICE
Ákos Frohner EGEE'08 September 2008
D. van der Ster, CERN IT-ES J. Elmsheuser, LMU Munich
R. Graciani for LHCb Mumbay, Feb 2006
LHC Data Analysis using a worldwide computing grid
Status and Plans of the Spanish ATLAS
ATLAS Distributed Analysis tests in the Spanish Cloud
Using an Object Oriented Database to Store BaBar's Terabytes
The ATLAS Computing Model
Development of LHCb Computing Model F Harris
The LHC Computing Grid Visit of Professor Andreas Demetriou
The LHCb Computing Data Challenge DC06
Presentation transcript:

ATLAS Distributed Analysis tests in the Spanish Cloud Presenter: Santiago González de la Hoz, IFIC (Instituto de Física Corpuscular), Centro mixto CSIC-Universitat de València 1GONZALEZ, S.,1SALT, J., 1AMOROS, G., 1FASSI, F., 1FERNANDEZ, A., 1KACI, M., 1LAMAS, A., 3MARCH, L., 1OLIVER, E., 1SANCHEZ, J., 1VILLAPLANA, M., 1VIVES, R., 2BORREGO, C., 2CAMPOS, M., 2NADAL, J., 2PACHECO, A., 3DEL PESO, J., 3PARDO, J., 3MUñOZ, L.,3FERNANDEZ, P., 3DEL CANO, L., 4ESPINAL, X. 1IFIC-Valencia, 2IFAE-Barcelona, 3UAM-Madrid, 4PIC-Barcelona

Outline The ATLAS experiment Spanish Cloud Distributed Analysis Test Spanish Resources Distributed Analysis Test What tests have been performed in the Spanish Cloud? Lessons learned Conclusions S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)

The ATLAS Experiment The offline computing: - Output event rate: 200 Hz ~ 109 events/year - Average event size (raw data): 1.6 MB/event Processing: - 40,000 of today’s fastest PCs Storage: - Raw data recording rate 320 MB/sec - Accumulating at 5-8 PB/year A solution: Grid technologies GRID computing GRID is used to solve problems of data simulation, storage and analysis. Data per year: ≈ Petabytes event generation simulation of what happens in the detector reconstruction of an event from what happened in the detector S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)

Spanish-Iberian Cloud for ATLAS IFIC IFAE UAM Tier1 at PIC Barcelona Offers storage and processing resources for three LHC experiments: ATLAS, CMS and LHCb. LHC experiments will store a copy of the collected data from the accelerator at CERN and dispatch a secondary copy to the Tier-1s centres in order to guarantee the conservation and integrity of the data. ~5% of the raw data from the LHC accelerator will be stored at PIC. Optical Private Network (OPN) Tier0 (CERN)  Tier1’s. More than 9 PetaBytes in/out PIC in 2008. SWE Cloud: Spain-Portugal Tier1: PIC-Barcelona Tier2’s: UAM, IFAE & IFIC LIP & Coimbra S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)

Spanish Resources Tier1 at PIC It will provide the infrastructure for data re-processing, as the raw data stored will be reprocessed several times per year with new parameters, as calibration and alignment constants improve. Data Storage: Experiments do need large, reliable and scalable storage services. To server the data at the required speed in order to maximize the efficiency of the cluster. Multi-Gigabit Ethernet network architecture, specially designed to enhance high speed data movement between WAN (Tier0, Tier1s, Tier2s) and LAN (CPU farm) . dCache storage system. S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)

Spanish Resources ATLAS Spanish Federated Tier2 IFIC: Valencia (coordinator) IFAE: Barcelona UAM: Madrid ATLAS Spanish Federated Tier2 Ramp-up of Tier-2 Resources (after LHC rescheduling) numbers are cumulative Evolution of ALL ATLAS T-2 resources according to the estimations made by ATLAS CB (Oct.06) Year Spanish ATLAS T-2 assuming a contribution of a 5% to the whole effort Strong increase of resources IFAE UAM IFIC TOTAL CPU (ksi2k) 201 276 438 915 Disk (TB) 104 147 198 449 Present resources of the Spanish ATLAS T-2 (February’09) New acquisitions in progress to get the pledged resources Accounting values are normalized according to WLCG recommendations S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)

Spanish Resources Storage Element System SE (Disk Storage) IFIC StoRM: Posix SRM v2 Lustre: High performance standard file system Shares: 50% IFIC, 25% IFAE and 25% UAM Data (AOD) distribution and DDM FT continuously running from Tier1 to Tier2 Storage Element System SE (Disk Storage) IFIC Lustre+StoRM IFAE dCache/disk+SRM posix UAM dCache A Tier needs a reliable and scalable storage system that can hold the users data, and serve it in an efficient way to users. A first sketch of a Storage system matrix (evaluation of different systems on going at CERN): S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)

Distributed Analysis Tests Goals Distributed Analysis Challenges need to be performed in order to validate site and cloud readiness for the full-scale user load The needs of analysis jobs differ from those of production, so though a site may function for Monte Carlo simulation, may perform poorly for analysis We are trying to identify breaking points and bottlenecks which result from the site/cloud design or configuration ATLAS testing framework An automated DA Challenge framework based on: Ganga submits a bulk of jobs –user like– to the target sites, keep tracking of the status and report once finished. Applications is a real analysis code from physicists, not “hello word” (presently) the Ganga EGEE/LCG and NDGF backend, using its data-based brokering and splitting. Bulk submit jobs to the WMS (direct submission to the CE) Both Posix I/O and “Copy mode” to read/access the input data The jobs are running with “user role” (a real case) S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)

Distributed Analysis Tests Presently the tests submit to the Ganga EGEE/LCG backend A single master “job” on many datasets is submitted Ganga splits this into one subjob per dataset Ganga EGEE/LCG jobs are instrumented to collect statistics Ganga uses by default Posix I/O to access/read the input files on the WN from the SE: dcap://.. is used for dCache rfio://… is used for Castor or DPM file://.. is used for Storm SEs (Lustre) “coy mode” uses dq2-get from the local SE to the WN (has been tested after the posix I/O tests). ATLAS distributed analysis team is working with the Panda team to make these test run on Panda Need to understand how to submit/split similar to above Need to instrument the Panda analysis jobs Metric of each individual tests are collected automatically (http://gangarobot.cern.ch/st/): Performance metrics: Success/failure rate, CPU/Walltime, Events per second, etc.. Error clasiffication: Differente I/O erros Used in Spanish Cloud S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)

What tests have we already performed in the Spanish Cloud In these challenges, we were coordinated with ATLAS Distributed Analysis team (Dan van der Ster and Johannes Elmsheuser) Presently we are interested in submitting large numbers of jobs intensive on data input Until a maximum of 200 jobs has been submitted to each site, thus up to ~1000 total per site. We ran different tests with different number of jobs. The jobs read directly from the SE using posix I/O Focusing on finding site limitations, e.g. heavy loads on storage and network resources, wms, client, etc. Specifically, the tests runs an AOD muon analysis everywhere right now (from M. Biglietti): UserAnalysis pkg, Athena 14.2.20 (ATLAS software version) mc08*AOD*e*s*r5* + some muon datasets replicated to the sites ~40M of events, ~140K files Each job processes an entire dataset, and the system submits one job per dataset S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)

What tests have we already performed in the Spanish Cloud Example Plot: Overall Eficiency, CPU/Walltime (%) CPU/Walltime is the CPU Percent Utilization (%) (%) (%) (%) (%) S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)

What tests have we already performed in the Spanish Cloud Example Plot: Event/second S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)

What tests have we already performed in the Spanish Cloud Example Plot: All sites Under these load (500 jobs and ~40M of events), most jobs were fast (average 54% of CPU and 13 Hz) Specific goal for analysis is to work toward 85% of CPU and 15 Hz per job We are going to try new methods like FileStager (%) S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)

What tests have we already performed in the Spanish Cloud Local posix I/O vs. FileStager: In 2008, we studied local posix I/O: Why? This is how LCG users access the data now. Why? Because of the ACM, and, in theory, posix I/O should minimize the I/O In practice, rfio/dcap/etc… is rarely tuned to Athena’s access pattern. Often, much more data is transferred than needed. Generates high load on network, SE, and disk pools. Tuning the readahead buffers is difficult But there are exceptional sites. Now we are looking at FileStager as an alternative: Pre-copies the next input file with lcg-cp (or anything else) in a background thread. It seems to improve many, but not all, sites. Probably not ideal for TAG (or other random access) analyses. S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)

What tests have we already performed in the Spanish Cloud Simultaneous tests of posix I/O and FileStager in ES cloud: 100 (posix)+50(FS) jobs per site. Results of posix vs. F.S.: Success rates are similar. % CPU Utilisation F.S. improved From 54% to 65% Events/s similarly increased. But, looking at sites individually, we see a different story. Posix FS S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)

What tests have we already performed in the Spanish Cloud Left: Posix I/O, Right: FileStager Posix Using dCache Using Lustre S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)

Lessons learned The results and tests led to the discovery of: Target dataset are not present there (all Tier2s must have all AOD data). (LIP case) Athena releases wrong installed or missing (UAM case). Inconsistency between mappings and permissions between SRM and WNs, so file can’t be accessed locally (IFIC case) Problem with the creation of directories due to a STORM/GFAL (IFIC-LIP case) Wrong access to the files: used rfio:/ instead of file:/ suitable for Lustre (LIP case) Wrong site identification by Ganga (IFAE-PIC case) Bad information published by the site in the information system (LIP Case) The File stager shouldn't be used for sites like LIP and IFIC which are using Lustre as File System In this case it is faster read the files from the SE than copies them to the WN. Maybe better results with faster CPUs, local disks, etc.. S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)

Conclusions Distributed analysis tests are necessary to stress the facilities at a simulated full user load. Very useful first test that allows to identify “big” problems: missing software, missing/bad information published and already to point out some limitation (e.g.: network connection in some sites). Test new parameters (FileStager) to improve access to the data using dCache and Lustre Breaking points and bottlenecks which result from the site/cloud design or configuration has been identified. S. González de la Hoz; 4th EGEE User Forum, March 2009, Catania (Italy)