Feedback from CMS Andrew Lahiff STFC Rutherford Appleton Laboratory Contributions from Christoph Wissing, Bockjoo Kim, Alessandro Degano CernVM Users Workshop.

Slides:



Advertisements
Similar presentations
SOFTWARE PRESENTATION ODMS (OPEN SOURCE DOCUMENT MANAGEMENT SYSTEM)
Advertisements

Mike Azocar Sr. Developer Technical Specialist Microsoft Corporation
Automated Tests in NICOS Nightly Control System Alexander Undrus Brookhaven National Laboratory, Upton, NY Software testing is a difficult, time-consuming.
1 Bridging Clouds with CernVM: ATLAS/PanDA example Wenjing Wu
CVMFS: Software Access Anywhere Dan Bradley Any data, Any time, Anywhere Project.
Tier-1 experience with provisioning virtualised worker nodes on demand Andrew Lahiff, Ian Collier STFC Rutherford Appleton Laboratory, Harwell Oxford,
OM. Brad Gall Senior Consultant
Abstract The automated multi-platform software nightly build system is a major component in the ATLAS collaborative software organization, validation and.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
1 port BOSS on Wenjing Wu (IHEP-CC)
Status of WLCG Tier-0 Maite Barroso, CERN-IT With input from T0 service managers Grid Deployment Board 9 April Apr-2014 Maite Barroso Lopez (at)
1 Dynamic Application Installation (Case of CMS on OSG) Introduction CMS Software Installation Overview Software Installation Issues Validation Considerations.
J.T Moscicki CERN LCG - Software Process & Infrastructure1 SPI Software Process & Infrastructure for LCG Software Packaging and Distribution LCG Application.
EGEE is a project funded by the European Union under contract IST Testing processes Leanne Guy Testing activity manager JRA1 All hands meeting,
Introduction to CVMFS A way to distribute HEP software on cloud Tian Yan (IHEP Computing Center, BESIIICGEM Cloud Computing Summer School.
CERN Manual Installation of a UI – Oxford July - 1 LCG2 Administrator’s Course Oxford University, 19 th – 21 st July Developed.
Nightly System Growth Graphs Abstract For over 10 years of development the ATLAS Nightly Build System has evolved into a factory for automatic release.
1Offline Weekly Meeting May May 2006 AliRoot Build Integration and (Testing) System Peter Hristov Vagner Morais.
The huge amount of resources available in the Grids, and the necessity to have the most up-to-date experimental software deployed in all the sites within.
1 / 22 AliRoot and AliEn Build Integration and Testing System.
Enabling Grids for E-sciencE System Analysis Working Group and Experiment Dashboard Julia Andreeva CERN Grid Operations Workshop – June, Stockholm.
UMD TIER-3 EXPERIENCES Malina Kirn October 23, 2008 UMD T3 experiences 1.
Organization and Management of ATLAS Nightly Builds F. Luehring a, E. Obreshkov b, D.Quarrie c, G. Rybkine d, A. Undrus e University of Indiana, USA a,
GAAIN Virtual Appliances: Virtual Machine Technology for Scientific Data Analysis Arihant Patawari USC Stevens Neuroimaging and Informatics Institute July.
LCG Generator Meeting, December 11 th 2003 Introduction to the LCG Generator Monthly Meeting.
Infrastructure for QA and automatic trending F. Bellini, M. Germain ALICE Offline Week, 19 th November 2014.
David Adams ATLAS DIAL/ADA JDL and catalogs David Adams BNL December 4, 2003 ATLAS software workshop Production session CERN.
Changes to CernVM-FS repository are staged on an “installation box" using a read/write file system interface. There is a dedicated installation box for.
NA61/NA49 virtualisation: status and plans Dag Toppe Larsen CERN
WLCG Overview Board, September 3 rd 2010 P. Mato, P.Buncic Use of multi-core and virtualization technologies.
Network awareness and network as a resource (and its integration with WMS) Artem Petrosyan (University of Texas at Arlington) BigPanDA Workshop, CERN,
Intro to Datazen.
Servicing HEP experiments with a complete set of ready integrated and configured common software components Stefan Roiser 1, Ana Gaspar 1, Yves Perrin.
Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Tools and techniques for managing virtual machine images Andreas.
2012 Objectives for CernVM. PH/SFT Technical Group Meeting CernVM/Subprojects The R&D phase of the project has finished and we continue to work as part.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Catalin Condurache STFC RAL Tier-1 GridPP OPS meeting, 10 March 2015.
Yannick Patois - Datagrid Software Repository Presentation - March, n° 1 Datagrid Software Repository Presentation CVS, packages and automatic.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
1 CMS Software Installation, Bockjoo Kim, 23 Oct. 2008, T3 Workshop, Fermilab CMS Commissioning and First Data Stan Durkin The Ohio State University for.
Daniele Spiga PerugiaCMS Italia 14 Feb ’07 Napoli1 CRAB status and next evolution Daniele Spiga University & INFN Perugia On behalf of CRAB Team.
Pavel Nevski DDM Workshop BNL, September 27, 2006 JOB DEFINITION as a part of Production.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES CVMFS deployment status Ian Collier – STFC Stefan Roiser – CERN.
Predrag Buncic (CERN/PH-SFT) Software Packaging: Can Virtualization help?
CernVM-FS Infrastructure for EGI VOs Catalin Condurache - STFC RAL Tier1 EGI Webinar, 5 September 2013.
CVMFS: Software Access Anywhere Dan Bradley Any data, Any time, Anywhere Project.
Grid Deployment Board 5 December 2007 GSSD Status Report Flavia Donno CERN/IT-GD.
36 th LHCb Software Week Pere Mato/CERN.  Provide a complete, portable and easy to configure user environment for developing and running LHC data analysis.
Predrag Buncic (CERN/PH-SFT) CernVM Status. CERN, 24/10/ Virtualization R&D (WP9)  The aim of WP9 is to provide a complete, portable and easy.
EMI is partially funded by the European Commission under Grant Agreement RI EMI SA2 Report Andres ABAD RODRIGUEZ, CERN SA2.4, Task Leader EMI AHM,
JRA1 Meeting – 09/02/ Software Configuration Management and Integration EGEE is proposed as a project funded by the European Union under contract.
CernVM-FS – Best Practice to Consolidate Global Software Distribution Catalin CONDURACHE, Ian COLLIER STFC RAL Tier-1 ISGC15, Taipei, March 2015.
Platform & Engineering Services CERN IT Department CH-1211 Geneva 23 Switzerland t PES Agile Infrastructure Project Overview : Status and.
Considerations on Using CernVM-FS for Datasets Sharing Within Various Research Communities Catalin Condurache STFC RAL UK ISGC, Taipei, 18 March 2016.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
CernVM-FS Operations in the CERN IT Storage Group Dan van der Ster (CERN IT-ST) CernVM Users Workshop 6-8 June D. van der Ster2.
CVMFS Alessandro De Salvo Outline  CVMFS architecture  CVMFS usage in the.
The CMS Beijing Tier 2: Status and Application Xiaomei Zhang CMS IHEP Group Meeting December 28, 2007.
Lessons learned administering a larger setup for LHCb
SharePoint 101 – An Overview of SharePoint 2010, 2013 and Office 365
CernVM-FS vs Dataset Sharing
Installation of the ALICE Software
Progress on NA61/NA49 software virtualisation Dag Toppe Larsen Wrocław
Introduction to CVMFS A way to distribute HEP software on cloud
Simulation use cases for T2 in ALICE
CernVM Status Report Predrag Buncic (CERN/PH-SFT).
X in [Integration, Delivery, Deployment]
Module 01 ETICS Overview ETICS Online Tutorials
Presentation transcript:

Feedback from CMS Andrew Lahiff STFC Rutherford Appleton Laboratory Contributions from Christoph Wissing, Bockjoo Kim, Alessandro Degano CernVM Users Workshop 5 th March, 2015

Page 2 Outline > CVMFS > Software deployment in CMS > Content > Installs > CVMFS in user space > CMSSW integration builds

Page 3 Software deployment in CMS > Basically fully based on CVMFS  Recommended strategy by WLCG TEG (Technical Evolution Group) in 2011/12 > Established via WLCG CVMFS task force among all LHC experiments  All sites used by CMS provide CVMFS > Includes the Tier-0 infrastructure > Additional deployment in AFS space at CERN  Backup  Not recommended to be used

Page 4 Content (1): CMSSW > Main CMS software package > Includes “all” from generators over simulation to reconstruction > Packaged as RPMs  Externals shared among several CMSSW releases > ROOT, GEANT, GCC and various libraries > Also packaged as RPMs by CMS > Typical numbers for a CMSSW release  150k+ files  10k+ directories  ~6GB disk space required > Releases typically built for different architectures or GCC versions  Per release on “production architecture”  Others are “development” version  All get installed in CVMFS

Page 5 Content (2): Gridpacks & PDFs > Additional input required for some workflows > Gridpacks  Generator files > Particle Density Functions  PDF sets from LHAPDF > Both include a few large files (100MB+)  Demanding for Squid infrastructure

Page 6 Content (3): CRAB Clients > Tool for submission of Analysis jobs > CRAB2  Frozen, no further updates planned > CRAB3  Recent product > Some additional Grid client tools are needed  Tools to manage VOMS proxies  Can be used from /cvmfs/grid.cern.ch

Page 7 Automatic Deployment > Executed via cron a few times per day > CMSSW and CRAB client installs are fully automatic  Check for new CMSSW releases in “CMS tag collector” and/or “release.map”  Check for new CRAB versions in certain repository directories  Install new versions > Gridpacks are installed via rsync from an agreed EOS directory  New Gridpacks are pushed to CVMFS automatically > LHAPDFs are downloaded from the web and installed in CVMFS  Some steering files to be manually edited before installation > For directories known to become large nested catalogs are created > Error notifications sent via > Feedback  Deployment machine sometimes very slow  Just bad interference of the VM with other VMs on same Hypervisor?

Page 8 CVMFS in User Space > Use case  Opportunistic resource without root access  CVMFS cannot/will not be installed by administrators  Cannot deploy Virtual Machines > Approach  Parrot (from  Allows to provide CVMFS in user space > Experiences > Has worked for some preliminary tests (e.g. CMS jobs at ATLAS sites) > Working on making it production ready

Page 9 CVMFS Integration Builds > Periodic automatic builds of CMSSW are called Integration Builds > Several combinations of platform / architecture / gcc version > 7 base versions of CMSSW plus several flavor variations > 19 releases produced twice a day > The latest two weeks of IBs are always available for developers > The CVMFS service installing & providing the IBs is based on a single Virtual Machine on CERN’s cloud infrastructure and is configured through Puppet

Page 10 Difficulties we encountered > CVMFS is resource hungry when deployed on the cloud infrastructure > Need for 500 IOP/s ceph volumes to publish ~7M files (2 weeks of IBs) > Requires VM with at least 8 GB RAM > Huge number of file descriptors open > CVMFS puppet module > Despite being very efficient, recreating a machine from scratch is not trivial or fast > Because of the above and the data we distribute, to recreate a CMSSW IB repository takes ~ 50 hours to publish these IBs > However we produce more IBs during these 50 hours, so another iteration of installation > Luckily after a few iterations the IBs installation catches up with the building

Page 11 Features we loved > CVMFS developers added several features we appreciate > We always keep the 2 most recent weeks of IBs > Garbage collection eventually becomes a must > FNAL replicates our repository as a stratum 1 > They can garbage collect even if we don’t > Automation is more reliable with CVMFS recovering from strange repo status and fixing itself