Distributed Computing for Small Experiments

Slides:



Advertisements
Similar presentations
Andrew McNab - Manchester HEP - 22 April 2002 EU DataGrid Testbed EU DataGrid Software releases Testbed 1 Job Lifecycle Authorisation at your site More.
Advertisements

Andrew McNab - Manchester HEP - 2 May 2002 Testbed and Authorisation EU DataGrid Testbed 1 Job Lifecycle Software releases Authorisation at your site Grid/Web.
Joining the Grid Andrew McNab. 28 March 2006Andrew McNab – Joining the Grid Outline ● LCG – the grid you're joining ● Related projects ● Getting a certificate.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
Pilots 2.0: DIRAC pilots for all the skies Federico Stagni, A.McNab, C.Luzzi, A.Tsaregorodtsev On behalf of the DIRAC consortium and the LHCb collaboration.
GridPP Steve Lloyd, Chair of the GridPP Collaboration Board.
December 17th 2008RAL PPD Computing Christmas Lectures 11 ATLAS Distributed Computing Stephen Burke RAL.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
Computing and LHCb Raja Nandakumar. The LHCb experiment  Universe is made of matter  Still not clear why  Andrei Sakharov’s theory of cp-violation.
Nick Brook Current status Future Collaboration Plans Future UK plans.
Wenjing Wu Andrej Filipčič David Cameron Eric Lancon Claire Adam Bourdarios & others.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
BESIII Production with Distributed Computing Xiaomei Zhang, Tian Yan, Xianghu Zhao Institute of High Energy Physics, Chinese Academy of Sciences, Beijing.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
The ILC And the Grid Andreas Gellrich DESY LCWS2007 DESY, Hamburg, Germany
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Code Applications Tamas Kiss Centre for Parallel.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow GridPP Computing for Particle.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
Enabling Grids for E-sciencE System Analysis Working Group and Experiment Dashboard Julia Andreeva CERN Grid Operations Workshop – June, Stockholm.
…building the next IT revolution From Web to Grid…
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
Eine Einführung ins Grid Andreas Gellrich IT Training DESY Hamburg
ATLAS is a general-purpose particle physics experiment which will study topics including the origin of mass, the processes that allowed an excess of matter.
Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre 14 February 2006.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
High Energy FermiLab Two physics detectors (5 stories tall each) to understand smallest scale of matter Each experiment has ~500 people doing.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow UK-T0 Meeting 21 st Oct 2015 GridPP.
The GridPP DIRAC project DIRAC for non-LHC communities.
The GridPP DIRAC project DIRAC for non-LHC communities.
Grid Execution Management for Legacy Code Architecture Exposing legacy applications as Grid services: the GEMLCA approach Centre.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES The Common Solutions Strategy of the Experiment Support group.
A Statistical Analysis of Job Performance on LCG Grid David Colling, Olivier van der Aa, Mona Aggarwal, Gidon Moont (Imperial College, London)
LHCb Computing 2015 Q3 Report Stefan Roiser LHCC Referees Meeting 1 December 2015.
Accounting Review Summary and action list from the (pre)GDB Julia Andreeva CERN-IT WLCG MB 19th April
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI solution for high throughput data analysis Peter Solagna EGI.eu Operations.
Bob Jones EGEE Technical Director
Accessing the VI-SEEM infrastructure
WLCG IPv6 deployment strategy
Review of the WLCG experiments compute plans
Status of WLCG FCPPL project
Status: ATLAS Grid Computing
(Prague, March 2009) Andrey Y Shevel
GridPP DIRAC Daniela Bauer & Simon Fayer.
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
Virtualisation for NA49/NA61
Distributed Computing for Small Experiments
Overview of the Belle II computing
Ian Bird GDB Meeting CERN 9 September 2003
How to enable computing
Virtualisation for NA49/NA61
Dagmar Adamova, NPI AS CR Prague/Rez
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Readiness of ATLAS Computing - A personal view
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
UK Status and Plans Scientific Computing Forum 27th Oct 2017
Computing at CEPC Xiaomei Zhang Xianghu Zhao
Simulation use cases for T2 in ALICE
WLCG Collaboration Workshop;
Monitoring of the infrastructure from the VO perspective
Xiaomei Zhang On behalf of CEPC software & computing group Nov 6, 2017
LHCb Grid Computing LHCb is a particle physics experiment which will study the subtle differences between matter and antimatter. The international collaboration.
Ivan Reid (Brunel University London/CMS)
Brian Matthews STFC EOSCpilot Brian Matthews STFC
Gridifying the LHCb Monte Carlo production system
GRIF : an EGEE site in Paris Region
The LHCb Computing Data Challenge DC06
Presentation transcript:

Distributed Computing for Small Experiments Daniela Bauer Imperial College London ICNFP 2017

Overview What is distributed computing ? The Grid – and what it can do for you. How does it work ? (Worked) Examples.

Distributed Computing Local Machine

Distributed Computing Local Batch System Local Machine

Distributed Computing Remote Batch System Local Batch System Local Machine

Distributed Computing Remote Batch System Resource Broker Local Batch System Local Machine

The Grid Originally developed as a tool to analyse data from the LHC: WLCG Webpage 170 computing centres ('sites') in 42 countries Standardized computing environment (+ clouds) Centralised job submission facilities

GridPP A collaboration of 19 UK institutes providing Grid and Cloud based computing services to particle physics and other experiments. Hosts 69k job slots and 51 PB of storage.

How is this relevant to you ? The Grid is not just for the big LHC experiments. E.g. GridPP commits 10% of its resources specifically to non-LHC experiments. Similar initiatives exist in other countries. There is ample opportunity for opportunistic usage. (Almost) Free CPU ! (Almost) Free Storage ! But: No HPC, Licensing, not much GPU (yet). Where's the catch ? → Complex system, effort required.

Grid Terminology: What is a VO ? Wikipedia: In grid computing, a virtual organization (VO) refers to a dynamic set of individuals or institutions defined around a set of resource-sharing rules and conditions. A collaboration corresponding to an experiment: CMS, Atlas, t2k.org, snoplus.snolab.ca, lz, vo.moedal.org, pheno Umbrella VOs: For individual researchers whose projects are too small individually to justify their own VO: WeNMR: Computational biology, gridpp: for researchers at UK universities, vo.france- grilles.fr: for researchers at French universities

What is a 'small' VO ? Small VOs/experiments: < 100 users These VOs have no dedicated computing support Users are typically familiar with the concept of a batch system Most (but not all) users are familiar with the concept of the grid GridPP has so far helped nine small VOs onto the Grid Three of these VOs were new to distributed computing

How do I access The Grid ? You need a certificate to prove who you are: lx01:> openssl x509 -in ~/.globus/usercert.pem -noout -text Certificate:    Data:        Version: 3 (0x2)        Serial Number: 46395 (0xb53b)    Signature Algorithm: sha256WithRSAEncryption        Issuer: C=UK, O=eScienceCA, OU=Authority, CN=UK e-Science CA 2B        Validity            Not Before: Nov  3 17:02:30 2016 GMT            Not After : Dec  3 17:02:30 2017 GMT        Subject: C=UK, O=eScience, OU=Imperial, L=Physics, CN=daniela bauer        Subject Public Key Info:            Public Key Algorithm: rsaEncryption                Public-Key: (2048 bit)                Modulus:                    00:b0:06:60:a8:3e:de:5a:0b:7e:1b:22:24:8b:cc:

Certificates Usually issued by a national certificate authority e.g. France, Germany, UK CERN With your certificate you need to join a VO (or make your own VO): e.g. Joining the Moedal VO (needs certificate loaded in your browser): https://lcg-voms2.cern.ch:8443/voms/vo.moedal.org/register/start.action

Joining a VO

Where to send your jobs: DIRAC The DIRAC framework for distributed computing was originally developed for LHCb. GridPP uses the DIRAC middleware to make their resources available to user communities. This includes Moedal :-) It provides a pilot based job submission framework and a data management system. There are currently at least 11 DIRAC instances in use worldwide. Homepage: diracgrid.org

DIRAC Storage

Accessing a DIRAC instance There are multiple ways to access a DIRAC instance: via scripts (“DIRAC UI”) from any SL6 system. via a python API. using Ganga: https://ganga.web.cern.ch/ganga/ (Usually) accessible to all members of a supported VOs If you are a member of the moedal VO and you would like to try out the system: https://www.gridpp.ac.uk/wiki/Quick_Guide_to_Dirac

Where to get help ? (moedal) access via DIRAC: gridpp-dirac- users@imperial.ac.uk your friendly sysadmin umbrella VO helplists GGUS

Job dashboard

Software Software is usually distributed via cvmfs: uploaded to central cvmfs server ('stratum 0') cvmfs servers are typically hosted by WLCG T0/T1 software is downloaded to the sites as needed caching at sites keeps popular software available for a certain amount of time users access software in /cvmfs/moedal.cern.ch Software must be transferable Anything not available in a standard SL6/SL7 environment must be packaged with your software We can help, but we can't fix your code.

Data Small amounts of storage (~ 10 TB) are usually provided through an affiliated university. Larger amounts are a question between you and your funding agency. Data stored on grid enabled storage ('SE') is accessible from all grid sites and your desktop. A file catalogue keeps track of where your data is stored. The users only sees the Logical File Name (LFN): e.g. /gridpp/user/dirac01.test/dirac01.testfile.txt

Real jobs from real users pheno dune na62 snoplus lz

LSST Currently being build in Chile: http://www.lsst.org/ Current UK contribution: Shape classification of galaxies on data taken by a predecessor (Dark Energy Survey) 100 million galaxies, data divided into 30000 files. Experiment specific software. Classification of a galaxy is an independent measurement, uses 10-20 s of CPU All data for a specific galaxy is contained in the same file This looks a lot like particle physics. No previous experience in grid computing. Settled on Ganga and DIRAC file catalogue for job submission and data access. 40 days FTE work required to setup and successfully complete large scale production workflow.

LuxZeplin Dark matter experiment based in South Dakota: GEANT4 based detector simulations Imperial College acts as their UK Data Centre But data is processed exclusively on the grid Started with experiment specific python scripts using the DIRAC API (physicists) Now: Custom job submission system (software engineer) Recently completed Mock Data Challenge: 133 TB of MC Data comprising 732288 files

The power of opportunistic computing Pledged CPU Cores

pheno/gridpp The pheno VO is used by the phenomenology group based at Durham/UK It uses Ganga/DIRAC to generate home grown MC The NNLO QCD corrections to Z boson (...) The gridpp VO hosts very different projects: GHOST: Geant Human Oncology Simulation Tool Using DIRAC directly, Geant4 based simluations Systematic infrastructure testing (similar to the UK nagios tests, including network tests)

Useful Links/Email vo.france-grilles.fr: http://www.france-grilles.fr/presentation-en/users/ http://www.france-grilles.fr/faq-en/ GridPP (and the 'gridpp' VO): https://www.gridpp.ac.uk/ technical questions: daniela.bauer@imperial.ac.uk, jeremy.coles@cern.ch formal request of collaboration: David Colling (d.colling@imperial.ac.uk) and David Britton (david.britton@glasgow.ac.uk) Moedal: https://lcg-voms2.cern.ch:8443/voms/vo.moedal.org/register/start.action