The CTA Computing Grid Project Cecile Barbier, Nukri Komin, Sabine Elles, Giovanni Lamanna, LAPP, CNRS/IN2P3 Annecy-le-Vieux Cecile Barbier - Nukri Komin1EGI.

Slides:



Advertisements
Similar presentations
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
Advertisements

Batch Production and Monte Carlo + CDB work status Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
K.Harrison CERN, 23rd October 2002 HOW TO COMMISSION A NEW CENTRE FOR LHCb PRODUCTION - Overview of LHCb distributed production system - Configuration.
Stuart K. PatersonCHEP 2006 (13 th –17 th February 2006) Mumbai, India 1 from DIRAC.Client.Dirac import * dirac = Dirac() job = Job() job.setApplication('DaVinci',
DIRAC API DIRAC Project. Overview  DIRAC API  Why APIs are important?  Why advanced users prefer APIs?  How it is done?  What is local mode what.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
Hunt for Molecules, Paris, 2005-Sep-20 Software Development for ALMA Robert LUCAS IRAM Grenoble France.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
3 rd DPHEP Workshop CERN, 7-8 December 2009 G. LAMANNA CTA C herenkov Telescope Array Giovanni Lamanna LAPP - Laboratoire d'Annecy-le-Vieux de Physique.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
CHEP'07 September D0 data reprocessing on OSG Authors Andrew Baranovski (Fermilab) for B. Abbot, M. Diesburg, G. Garzoglio, T. Kurca, P. Mhashilkar.
Nick Brook Current status Future Collaboration Plans Future UK plans.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
Bookkeeping Tutorial. Bookkeeping & Monitoring Tutorial2 Bookkeeping content  Contains records of all “jobs” and all “files” that are created by production.
BESIII Production with Distributed Computing Xiaomei Zhang, Tian Yan, Xianghu Zhao Institute of High Energy Physics, Chinese Academy of Sciences, Beijing.
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Cracow Grid Workshop October 2009 Dipl.-Ing. (M.Sc.) Marcus Hilbrich Center for Information Services and High Performance.
Enabling Grids for E-sciencE EGEE-III INFSO-RI Using DIANE for astrophysics applications Ladislav Hluchy, Viet Tran Institute of Informatics Slovak.
1 LCG-France sites contribution to the LHC activities in 2007 A.Tsaregorodtsev, CPPM, Marseille 14 January 2008, LCG-France Direction.
Bookkeeping Tutorial. 2 Bookkeeping content  Contains records of all “jobs” and all “files” that are produced by production jobs  Job:  In fact technically.
Transformation System report Luisa Arrabito 1, Federico Stagni 2 1) LUPM CNRS/IN2P3, France 2) CERN 5 th DIRAC User Workshop 27 th – 29 th May 2015, Ferrara.
C herenkov Telescope Array Dr. Giovanni Lamanna CNRS tenured research scientist (CTA-LAPP team leader and CTA Computing Grid project coordinator) LAPP.
1 Cherenkov Telescope Array: a production system prototype L. Arrabito 1 C. Barbier 2, J. Bregeon 1, A. Haupt 3, N. Neyroud 2 for the CTA Consortium 1.
1 LHCb computing for the analysis : a naive user point of view Workshop analyse cc-in2p3 17 avril 2008 Marie-Hélène Schune, LAL-Orsay for LHCb-France Framework,
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI VO auger experience with large scale simulations on the grid Jiří Chudoba.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
InSilicoLab – Grid Environment for Supporting Numerical Experiments in Chemistry Joanna Kocot, Daniel Harężlak, Klemens Noga, Mariusz Sterzel, Tomasz Szepieniec.
Enabling Grids for E-sciencE LRMN ThIS on the Grid Sorina CAMARASU.
Science Gateway- 13 th May Science Gateway Use Cases/Interfaces D. Sanchez, N. Neyroud.
EGI-Engage is co-funded by the Horizon 2020 Framework Programme of the European Union under grant number Federated Cloud Update.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
Management of the Data in Auger Jean-Noël Albert LAL – Orsay IN2P3 - CNRS ASPERA – Oct Lyon.
ScotGRID is the Scottish prototype Tier 2 Centre for LHCb and ATLAS computing resources. It uses a novel distributed architecture and cutting-edge technology,
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI solution for high throughput data analysis Peter Solagna EGI.eu Operations.
CTA Cerenkov Telescope Array G. Lamanna (1), C. Vuerli (2) (1) CNRS/IN2P3/LAPP, Annecy (FR) (2) INAF – Istituto Nazionale di Astrofisica, Trieste (IT)
Simulation Production System
Xiaomei Zhang CMS IHEP Group Meeting December
Virtualization and Clouds ATLAS position
Overview of the Belle II computing
Progress on NA61/NA49 software virtualisation Dag Toppe Larsen Wrocław
Belle II Physics Analysis Center at TIFR
Moving the LHCb Monte Carlo production system to the GRID
Pasquale Migliozzi INFN Napoli
Data Challenge with the Grid in ATLAS
Existing Perl/Oracle Pipeline
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Readiness of ATLAS Computing - A personal view
Job workflow Pre production operations:
Artem Trunov and EKP team EPK – Uni Karlsruhe
Simulation use cases for T2 in ALICE
Interoperability & Standards
US CMS Testbed.
WLCG Collaboration Workshop;
Haiyan Meng and Douglas Thain
R. Graciani for LHCb Mumbay, Feb 2006
Overview of Workflows: Why Use Them?
GRIF : an EGEE site in Paris Region
Status and plans for bookkeeping system and production tools
Introduction to the SHIWA Simulation Platform EGI User Forum,
The LHCb Computing Data Challenge DC06
Presentation transcript:

The CTA Computing Grid Project Cecile Barbier, Nukri Komin, Sabine Elles, Giovanni Lamanna, LAPP, CNRS/IN2P3 Annecy-le-Vieux Cecile Barbier - Nukri Komin1EGI User Forum Vilnius,

Cecile Barbier - Nukri KominEGI User Forum Vilnius, CTA-CG CTA Computing Grid LAPP, Annecy Giovanni Lamanna, Nukri Komin Cecile Barbier, Sabine Elles LUPM, Montpellier Georges Vasileiadis Claudia Lavallay, Luisa Arrabito Goal: Bring CTA on the Grid

Cecile Barbier - Nukri KominEGI User Forum Vilnius, CTA-CG Aim provide working environment, tools and services for all tasks assigned to Data Management and Processing Center simulation data processing storage offline analysis user's interface Test Grid computing software around Grid computing estimate computing needs and requirements requests at Lyon, close contact with DESY Zeuthen

Cecile Barbier - Nukri KominEGI User Forum Vilnius, Outline CTA and its Data Management and Processing Centre current activities: massive Monte Carlo Simulations preparation of Meta Data Base short-term plan bring the user on the Grid and to the data ideas for future data management and analysis pipe-line Note: CTA is in preparatory phase here mostly work in progress and ideas

Cecile Barbier - Nukri KominEGI User Forum Vilnius, Current Cherenkov Telescopes

Cecile Barbier - Nukri KominEGI User Forum Vilnius, Cherenkov Telescope Array large array, 30 – 100 telescopes in 3 sizes Preparatory Phase institutes, 22 countries

The CTA Observatory main logical units : Science Operation Centre: organisation of observations Array Operation Centre: the on-site service Science Data Centre: Software development Data analysis Data reduction Data archiving Data dissemination to observers Total expected data volume from CTA: 1 to 10 (?) PB per year (main data stream for permanent storage is of the order of 1 (10 ?) GB/s) MC requirements: tens of CPU years, hundreds of TB Existing ICT-based infrastructures, such as EGEE/EGI and GEANT, are potential solutions to provide the CTA observatory with best use of e-infrastructures. 7 CTA Operational Data Flow

Cecile Barbier - Nukri KominEGI User Forum Vilnius, CTA Virtual Grid Organisation Benefits of the EGEE/EGI Grid institutes can provide easily computing power minimal man power needed, usually sites already supporting LHC can be managed centrally (e.g. for massive simulations) distributed but transparent for all users (compare HESS) CTA Virtual Organisation: vo.cta.in2p3.fr French name, but open to everyone (renaming almost impossible) VO manager: G. LAPP

Cecile Barbier - Nukri KominEGI User Forum Vilnius, CTA VO – computing 14 sites in 5 countries providing access to their computing resources 3 big sites: CC Lyon, DESY Zeuthen, Cyfronet Poland GRIF: several sites of various sizes in/around Paris many small sites (~100 CPUs) 30k logical CPUs, shared with other VOs ~1000 – 2000 CPUs for CTA at any time (based on experience)

Cecile Barbier - Nukri KominEGI User Forum Vilnius, CTA VO – computing load not smooth, only one simulation manager (Nukri Komin) LAPP, CCIN2P3 and DESY Zeuthen among the biggest contributors

Cecile Barbier - Nukri KominEGI User Forum Vilnius, CTA VO – storage each site providing several 100GB up to 10 TB local disk space massive storage (several 100 TB): CC Lyon (including tapes), DESY Zeuthen, Cyfronet massive storage for large temporary files simulations: corsika, will be kept for reprocessing corsika file size GB for proton showers

Cecile Barbier - Nukri KominEGI User Forum Vilnius, Grid Monte Carlo Production first massive use of Grid: MC simulations About good quality runs high requirements per run that only few grid sites can handle up to 4 GB RAM 10 GB local scratch disk space many problems solved, next round much more efficient with automated MC simulation production using the EasiJob tool developed at LAPP

Cecile Barbier - Nukri KominEGI User Forum Vilnius, web interface monitoring Create the job: config files and scripts GANGA job and grid control out: browse results in: configure and start a task CTA VO Grid Operation Centre Grid tools and software developed for CTACG Interface with the community Configuration Monitoring Access to data files central data base    files on grid SEs Automated Simulation Production

Cecile Barbier - Nukri KominEGI User Forum Vilnius, Automated Simulation Production EasiJob – Easy Integrated Job Submission developed by S. Elles within the MUST frame work MUST = Mid-Range Data Storage and Computing Centre widely open to Grid Infrastructure, at LAPP Annecy and Savoie University more general than CTA can be used for any software and every experiment based on GANGA (Gaudi/Athena aNd Grid Alliance) Grid front-end in python developed for Atlas and LHCb, used by many other experiments task configuration, job submission and monitoring, file bookkeeping

Cecile Barbier - Nukri KominEGI User Forum Vilnius, EasiJob – Task Configuration description of a task set of parameters, with default values, define if browsable representation in data base parameter keyword (#key1) job template set of files (input sandbox) keywords will be replaced by data base values web interface example: corsika

Cecile Barbier - Nukri KominEGI User Forum Vilnius, EasiJob – Job Classes configure site classes requirements are based on published parameters e.g. GLueHostMainMemoryRAMSize > 2000 these requirements are interpreted differently at each site sites have different storage capacities creation of job classes large jobs only on a subset of sites job/site matching currently semi-manual close interaction with local admins (in particular Lyon and Zeuthen) web interface

Cecile Barbier - Nukri KominEGI User Forum Vilnius, EasiJob – Job Submission define number of jobs for a task automated job submission to the site with the minimum of waiting jobs submission is paused when too many jobs are pending status monitoring and re-submission of failed jobs keeps track of produced files logical file name (LFN) on the Grid echo statement in execution script monitoring on web page

Cecile Barbier - Nukri KominEGI User Forum Vilnius, EasiJob – Status deployed in Annecy, will be used for next simulations configuration and job submission not open to public want to avoid massive non-sense productions user certificates need to be installed manually idea: provide “software as a service”

Cecile Barbier - Nukri KominEGI User Forum Vilnius, web interface monitoring Create the job: config files and scripts GANGA job and grid control out: browse results in: configure and start a task CTA VO Grid Operation Centre Grid tools and software developed for CTACG Interface with the community Configuration Monitoring Access to data files central data base    files on grid SEs Bookkeeping

Cecile Barbier - Nukri KominEGI User Forum Vilnius, Bookkeeping current simulation: Not all output files kept Production parameters set in EasiJob data base automatically generated web interface [C. Barbier] shows only parameters defined as browsable proposes only values which were produced returns list of lfn (logical file names) starting point for more powerful meta data base

future: complicated data structure we want to keep track of files produced and their relations search for files using the production parameters find information on files, even if the files have been removed Cecile Barbier - Nukri KominEGI User Forum Vilnius, Bookkeeping production file 1... production file 2... DST file... raw data file calibrated file DST real data structure

data: simulations, real data,... meta data: information describing the data logical and physical file name, production parameters, etc. meta information can be in several data bases Cecile Barbier - Nukri KominEGI User Forum Vilnius, Meta Data Management [C. Lavalley, LUP Montpellier]

Cecile Barbier - Nukri KominEGI User Forum Vilnius, Meta Data Management AMI – Atlas Meta Data Interface developed at LPSC Grenoble can interrogate other data bases information can be pushed with AMI clients web, python, C++, Java clients manages access rights: username/password, certificate,... we will deploy AMI for CTA (with LUPM and LPSC) for simulations bookkeeping and file search to be tested for future use in CTA

Cecile Barbier - Nukri KominEGI User Forum Vilnius, Bring the User to the Data You have a certificate? You can submit jobs to the EGI grid : glite-wms-job-submit or use Ganga ( easy to use python front end Grid User Interface needed certificate infrastructure software to download files and submit jobs we are evaluating a way to make Grid UI available Dirac :Distributed Infrastructure with Remote Agent Control

initially developed for LHCb, now generic version very easy to install Grid (and beyond) front-end work load management with pilot jobs : pull mode, no jobs lost due to Grid problems, shorter waiting time before execution integrated Data Management System integrated software management python and web interfaces for job submission LAPP, LUPM and Pic-IFAE Barcelona for setup/testing, will be open to collaboration soon we don't plan to use it for simulations Cecile Barbier - Nukri KominEGI User Forum Vilnius, DIRAC

Cecile Barbier - Nukri KominEGI User Forum Vilnius, Analysis Chain telescopes raw data (level 0) raw data (level 0) calibrated camera images (level 1) calibrated camera images (level 1) photon list (level 2) photon list (level 2) sky maps, lightcurves, spectra (levels 3 and 4) sky maps, lightcurves, spectra (levels 3 and 4) e.g. Fermi Science Tools software available for Linux, Mac, Windows internal data published data

Cecile Barbier - Nukri KominEGI User Forum Vilnius, Data Rates Raw Data: some GB/s, PB/year (level 0) production during night time, max h per day 29 day cycle with peak at new moon Reconstructed Data (level 2, available to public) about 10% of raw data computing requirements: 1h of raw data needs ~200 CPUs*days (today) based on HESS Model++, 28min of 4 telescopes needs 3x 1Ms [M. de Naurois] Results (level 3 and 4, available to public) requirements: to be evaluated

Cecile Barbier - Nukri KominEGI User Forum Vilnius, Data Flow (a possible “Tier” view …) telescopes on-site or nearby computing centre computer cluster at participating institutes local machines, scientist's desktop two or three powerful computing centres Tier remote site: local computing centre or fast internet link

Data Source and Reconstruction Tier 0: local data source (at or near observatory) – make data available on local storage element – possible site for archiving Tier 1: calibration and reconstruction sites – at least 2 sites for redundancy – guaranteed CPU time for calibration and reconstruction – can handle peaks when other site down or re-calibration – requirements: disk space and computing power strong network between Tier 0 and 1: – most computation on Tier 1 weak network connection: – data reduction at Tier 0 ( = Tier 1 is at site)

Data Analysis Tier 2: Science Data Centre(s) – (small) computing clusters at participating institutes – data quality check – 1 st analysis – provide preprocessed data and results Tier 3: scientist's computer – provide individual computing and software – also for non-CTA scientists – data access (data download from nearest Tier 2) – simple installation on all systems → possibly virtual machine

Cecile Barbier - Nukri KominEGI User Forum Vilnius, Summary CTA Computing Grid, several 1000 CPUs at 14 sites currently used for massive simulations simulations tools and services for easy submission and monitoring (LAPP) will set up a meta data base for easy search and use (with LUPM) soon: tests of DIRAC for user analysis (with PIC) future Data Management and Processing Centre on distributed sites (Tier 0,1, (2,3)) Disclaimer: CTA data management system still under study (nothing yet decided !) CTA Computing Grid is one approach under study