PRIN STOA-LHC: STATUS BARI BOLOGNA-18 GIUGNO 2014 Giorgia MINIELLO G. MAGGI, G. DONVITO, D. Elia INFN Sezione di Bari e Dipartimento Interateneo.

Slides:



Advertisements
Similar presentations
Xrootd and clouds Doug Benjamin Duke University. Introduction Cloud computing is here to stay – likely more than just Hype (Gartner Research Hype Cycle.
Advertisements

1 CHEP 2000, Roberto Barbera Roberto Barbera (*) Grid monitoring with NAGIOS WP3-INFN Meeting, Naples, (*) Work in collaboration with.
The DataGrid Project NIKHEF, Wetenschappelijke Jaarvergadering, 19 December 2002
A Grid Resource Broker Supporting Advance Reservations and Benchmark- Based Resource Selection Erik Elmroth and Johan Tordsson Reporter : S.Y.Chen.
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
DPM Italian sites and EPEL testbed in Italy Alessandro De Salvo (INFN, Roma1), Alessandra Doria (INFN, Napoli), Elisabetta Vilucchi (INFN, Laboratori Nazionali.
1 Bridging Clouds with CernVM: ATLAS/PanDA example Wenjing Wu
1 Status of the ALICE CERN Analysis Facility Marco MEONI – CERN/ALICE Jan Fiete GROSSE-OETRINGHAUS - CERN /ALICE CHEP Prague.
Distributed Computing for CEPC YAN Tian On Behalf of Distributed Computing Group, CC, IHEP for 4 th CEPC Collaboration Meeting, Sep ,
PROOF: the Parallel ROOT Facility Scheduling and Load-balancing ACAT 2007 Jan Iwaszkiewicz ¹ ² Gerardo Ganis ¹ Fons Rademakers ¹ ¹ CERN PH/SFT ² University.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
1 port BOSS on Wenjing Wu (IHEP-CC)
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Computational grids and grids projects DSS,
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
WNoDeS – Worker Nodes on Demand Service on EMI2 WNoDeS – Worker Nodes on Demand Service on EMI2 Local batch jobs can be run on both real and virtual execution.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Grid Computing at Yahoo! Sameer Paranjpye Mahadev Konar Yahoo!
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
T3 analysis Facility V. Bucard, F.Furano, A.Maier, R.Santana, R. Santinelli T3 Analysis Facility The LHCb Computing Model divides collaboration affiliated.
Changes to CernVM-FS repository are staged on an “installation box" using a read/write file system interface. There is a dedicated installation box for.
11 CLUSTERING AND AVAILABILITY Chapter 11. Chapter 11: CLUSTERING AND AVAILABILITY2 OVERVIEW  Describe the clustering capabilities of Microsoft Windows.
NA61/NA49 virtualisation: status and plans Dag Toppe Larsen CERN
Roberto Barbera Prague, ALICE Multi-site Data Transfer Tests on a Wide Area Network Giuseppe Lo Re Roberto Barbera Work in collaboration with:
Claudio Grandi INFN Bologna CMS Computing Model Evolution Claudio Grandi INFN Bologna On behalf of the CMS Collaboration.
PROOF and ALICE Analysis Facilities Arsen Hayrapetyan Yerevan Physics Institute, CERN.
A Silvio Pardi on behalf of the SuperB Collaboration a INFN-Napoli -Campus di M.S.Angelo Via Cinthia– 80126, Napoli, Italy CHEP12 – New York – USA – May.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
+ AliEn site services and monitoring Miguel Martinez Pedreira.
ELASTIC LSF EXTENSION AT CNAF. credits Vincenzo Ciaschini Stefano Dal Pra Andrea Chierici Vladimir Sapunenko Tommaso Boccali.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
Roberto Barbera Prague, Dipartimento di Fisica dell’Università di Catania and INFN Catania - Italy ALICE Collaboration Update on the network.
CERN IT Department CH-1211 Geneva 23 Switzerland t ES 1 how to profit of the ATLAS HLT farm during the LS1 & after Sergio Ballestrero.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI CERN and HelixNebula, the Science Cloud Fernando Barreiro Megino (CERN IT)
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Considerations on Using CernVM-FS for Datasets Sharing Within Various Research Communities Catalin Condurache STFC RAL UK ISGC, Taipei, 18 March 2016.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
INFN/IGI contributions Federated Clouds Task Force F2F meeting November 24, 2011, Amsterdam.
DGAS Distributed Grid Accounting System INFN Workshop /05/1009, Palau Giuseppe Patania Andrea Guarise 6/18/20161.
Elastic CNAF Datacenter extension via opportunistic resources INFN-CNAF.
Activities and Perspectives at Armenian Grid site The 6th International Conference "Distributed Computing and Grid- technologies in Science and Education"
Bologna, March 30, 2006 Riccardo Zappi / Luca Magnoni INFN-CNAF, Bologna.
Using the CMS Higher Level Trigger Farm as a Cloud Resource David Colling Imperial College London.
St. Petersburg, 2016 Openstack Disk Storage vs Amazon Disk Storage Computing Clusters, Grids and Cloud Erasmus Mundus Master Program in PERCCOM Author:
A Web Based Job Submission System for a Physics Computing Cluster David Jones IOP Particle Physics 2004 Birmingham 1.
S. Pardi Computing R&D Workshop Ferrara 2011 – 4 – 7 July SuperB R&D on going on storage and data access R&D Storage Silvio Pardi
Grid Operations in Germany T1-T2 workshop 2015 Torino, Italy Kilian Schwarz WooJin Park Christopher Jung.
Course: Cluster, grid and cloud computing systems Course author: Prof
Dynamic Extension of the INFN Tier-1 on external resources
Extending the farm to external sites: the INFN Tier-1 experience
Workload Management Workpackage
DPM at ATLAS sites and testbeds in Italy
Distributed storage, work status
ALICE & Clouds GDB Meeting 15/01/2013
SuperB – INFN-Bari Giacinto DONVITO.
Blueprint of Persistent Infrastructure as a Service
Belle II Physics Analysis Center at TIFR
Future of WAN Access in ATLAS
Silvio Pardi R&D Storage Silvio Pardi
ATLAS Cloud Operations
Brief overview on GridICE and Ticketing System
StoRM Architecture and Daemons
Simulation use cases for T2 in ALICE
a VO-oriented perspective
Presentation transcript:

PRIN STOA-LHC: STATUS BARI BOLOGNA-18 GIUGNO 2014 Giorgia MINIELLO G. MAGGI, G. DONVITO, D. Elia INFN Sezione di Bari e Dipartimento Interateneo di Fisica dell’Università e del Politecnico di Bari

BARI PRIN Activities - OUTLINE TEST AND COMPARISON of FILE ACCESS PERFORMANCES VIA XROOTD versus LOCAL ACCESS SET UP AND DEPLOYMENT OF AN ELASTIC CLUSTER BASED ON OPENSTACK IaaS AND CONFIGURED FOR ALICE VIRTUAL ANALYSIS FACILITY A BRIEF OVERVIEW ON BARI TESTBED SET-UP FUTURE GOALS 18/06/14PRIN STOA-LHC: STATUS BOLOGNA

TEST AND COMPARISON of FILE ACCESS PERFORMANCES VIA XROOTD versus LOCAL ACCESS H.E.P. Analysis used: “Observation of a Higgs Boson and measurements of its properties through the decay channel H->ZZ->4l with the CMS experiment”. Dataset: DataMuParked/Run2012A-22Jan2013-v1/AOD. Study of the wall clock time distributions of jobs running two steps of this analysis (one at a time!): STEP 1 more I/O intensive (~ 75% of the whole job duration spent in I/O activitIes, and the remaining ~25% spent in computing activities) and STEP 2 more CPU intensive (~83% computing, ~17% I/O). A Set of files taken from the dataset each job accessed just one file, once reading it on local file system lustre and once via remote file access protocol XROOTD. For every file 100 jobs were submitted to ensure statistics. Only XROOTD redirector can choose the server on which the file requested will be open: A study of wall clock time distributions server by server has also been performed splitting the histogram related to EU XROOTD servers site by site (e.g.: ES, CH, DE, KR…). 18/06/14PRIN STOA-LHC: STATUS BOLOGNA

TEST AND COMPARISON of FILE ACCESS PERFORMANCES VIA XROOTD versus LOCAL ACCESS- RESULTS (1/2) STEP 1 of the analysis: a shift factor between the position of the peaks belonging to the wall clock time distributions of jobs accessing files via xrootd (INCLUDING EU servers) and locally has been calculated. It has been evaluated to be only ~1.1 when considering ONLY IT XROOTD servers and ~1.8 when including all EU servers. A shift factor between EU xrootd histogram's peak and IT xrootd histogram's peak has been evaluated to be ~1.43. Studying the performances about accessing files separating xrootd servers: the wall clock values of jobs opening files on ITALIAN XROOTD servers (BARI + CNAF) are only 15% greater than those of jobs accessing files locally; the peaks belonging to Bari and CNAF histograms are quite compatible (within few units in percentage). The gap between the EU peak and the peaks belonging to IT histograms highlights that the redirection performed by XROOTD on EU servers (OUTSIDE ITALY) clearly constitues a bottleneck in remote data access STEP 2 of the analysis: the new input files (the output files of STEP 1) have been previously allocated on Bari lustre storage and accessed using both methods: Shift factor reduced to ~1.05 in this step. 18/06/14PRIN STOA-LHC: STATUS BOLOGNA

TEST AND COMPARISON of FILE ACCESS PERFORMANCES VIA XROOTD versus LOCAL ACCESS- RESULTS (2/2) 18/06/14PRIN STOA-LHC: STATUS BOLOGNA STEP 1STEP 2

CREATING AN ELASTIC CLUSTER FOR V.A.F. ON PRISMA TESTBED Using the Bari PRISMA TESTBED based on OpenStack, a cluster has been set up and configured for the ALICE VAF; EC2 interface was already available on BARI PRISMA TESTBED; Problems of V.M. contextualization solved in collaboration with Torino and other sites: on Bari PRISMA testbed a “configuration drive” is directly mounted on VMs instead of METADATA SERVER At present, the new version of the CernVM ( ), is able to support the contextualization through “ConfigDrive” (on OpenStack) as an alternative to the “metadata server”. Actually, a direct connection has been used for CernVMFS configuration, but a configuration with proxy is on schedule. The performances of the V.A.F. are under test. We started changing gradually the number of nodes to be requested 18/06/14PRIN STOA-LHC: STATUS BOLOGNA

BARI PRISMA Testbed set-up Rabbit cluster/AMQP Galera Cluster replication HAProxy, keepalived ha1.ba.infn.it ha2.ba.infn.it MySQL queries ha.ba.infn.it MyS QL AM QP Pacemaker/Corosync Active/Passive Pacemaker/Corosync Active/Passive OpenStac k Controller API servers OpenStac k Controller API servers OpenStac k Network Controller OpenStac k Network Controller OpenStac k Network Controller OpenStac k Network Controller OpenStac k Controller API servers OpenStac k Controller API servers OpenStack Public APIs compute node GlusterFS Swift Ceph Object Storage Cluster Mon Cinder Architectural schema by Marica Antonacci

DIMENSIONS AND FEATURES OF INFN BARI/UNIBA TESTBED Implemented within PON-PRISMA project The final configuration foreseen for the testbed is: 500 CPU/core 3TB RAM 110 TB disk (replica 3) 10 Gbit/s internal network 10 Gbit/s geographical network 256 Public IP DIRECT ACCESS At the moment we have 300 CPU/core already available 18/06/14PRIN STOA-LHC: STATUS BOLOGNA

Future Goals Configuration in production of an italian XROOTD redirector in collaboration with PISA This will solve problems of performance and scalability Study of the VAF performances: changing parameters of the cluster configuration (e.g. max number of nodes) testing pilot analysis tasks (VAF vs standard batch analysis on the Grid) Configuring the V.A.F. for CVMFS access via squid proxy; Contribution to the development and testing of the Virtual Storage Element; Configuration of an Analysis Facility based on PROOF, also for CMS 18/06/14PRIN STOA-LHC: STATUS BOLOGNA