LCG experience in Integrating Grid Toolkits

Slides:



Advertisements
Similar presentations
LNL M.Biasotto, Roma, 22 novembre Some notes on WP6 INFN Testbed.
Advertisements

Tony Doyle GridPP2 Proposal, BT Meeting, Imperial, 23 July 2003.
 Contributing >30% of throughput to ATLAS and CMS in Worldwide LHC Computing Grid  Reliant on production and advanced networking from ESNET, LHCNET and.
DataGrid is a project funded by the European Union 22 September 2003 – n° 1 EDG WP4 Fabric Management: Fabric Monitoring and Fault Tolerance
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
27-29 September 2002CrossGrid Workshop LINZ1 USE CASES (Task 3.5 Test and Integration) Santiago González de la Hoz CrossGrid Workshop at Linz,
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
Workload Management Massimo Sgaravatto INFN Padova.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
EGEE is a project funded by the European Union under contract IST JRA1 Testing Activity: Status and Plans Leanne Guy EGEE Middleware Testing.
1 DataGRID Application Status and plans
EMI INFSO-RI SA2 - Quality Assurance Alberto Aimar (CERN) SA2 Leader EMI First EC Review 22 June 2011, Brussels.
CERN LCG-1 Status and Issues Ian Neilson for LCG Deployment Group CERN Hepix 2003, Vancouver.
INFSO-RI Enabling Grids for E-sciencE SA1: Cookbook (DSA1.7) Ian Bird CERN 18 January 2006.
LCG and HEPiX Ian Bird LCG Project - CERN HEPiX - FNAL 25-Oct-2002.
EGEE is a project funded by the European Union under contract IST Testing processes Leanne Guy Testing activity manager JRA1 All hands meeting,
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
Bob Jones – June n° 1 EDG release Schedule Bob Jones.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
05/29/2002Flavia Donno, INFN-Pisa1 Packaging and distribution issues Flavia Donno, INFN-Pisa EDG/WP8 EDT/WP4 joint meeting, 29 May 2002.
D0RACE: Testbed Session Lee Lueking D0 Remote Analysis Workshop February 12, 2002.
First attempt for validating/testing Testbed 1 Globus and middleware services WP6 Meeting, December 2001 Flavia Donno, Marco Serra for IT and WPs.
CERN LCG Deployment Overview Ian Bird CERN IT/GD LHCC Comprehensive Review November 2003.
EGEE is a project funded by the European Union under contract IST Middleware Planning for LCG/EGEE Bob Jones EGEE Technical Director e-Science.
DataGRID WPMM, Geneve, 17th June 2002 Testbed Software Test Group work status for 1.2 release Andrea Formica on behalf of Test Group.
JRA Execution Plan 13 January JRA1 Execution Plan Frédéric Hemmer EGEE Middleware Manager EGEE is proposed as a project funded by the European.
EGEE is a project funded by the European Union under contract IST EGEE Services Ian Bird SA1 Manager Cork Meeting, April
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
29/1/2002A.Ghiselli, INFN-CNAF1 DataTAG / WP4 meeting Cern, 29 January 2002 Agenda  start at  Project introduction, Olivier Martin  WP4 introduction,
Status Organization Overview of Program of Work Education, Training It’s the People who make it happen & make it Work.
LCG CERN David Foster LCG WP4 Meeting 20 th June 2002 LCG Project Status WP4 Meeting Presentation David Foster IT/LCG 20 June 2002.
Nanbor Wang, Balamurali Ananthan Tech-X Corporation Gerald Gieraltowski, Edward May, Alexandre Vaniachine Argonne National Laboratory 2. ARCHITECTURE GSIMF:
David Foster LCG Project 12-March-02 Fabric Automation The Challenge of LHC Scale Fabrics LHC Computing Grid Workshop David Foster 12 th March 2002.
CERN LCG Deployment Overview Ian Bird CERN IT/GD LCG Internal Review November 2003.
12 March, 2002 LCG Applications Area - Introduction slide 1 LCG Applications Session LCG Launch Workshop March 12, 2002 John Harvey, CERN LHCb Computing.
DataGrid is a project funded by the European Commission under contract IST rd EU Review – 19-20/02/2004 The EU DataGrid Project Three years.
CERN Certification & Testing LCG Certification & Testing Team (C&T Team) Marco Serra - CERN / INFN Zdenek Sekera - CERN.
JRA1 Meeting – 09/02/ Software Configuration Management and Integration EGEE is proposed as a project funded by the European Union under contract.
EGEE is a project funded by the European Union under contract IST Issues from current Experience SA1 Feedback to JRA1 A. Pacheco PIC Barcelona.
12 March, 2002 LCG Applications Area - Introduction slide 1 LCG Applications Session LCG Launch Workshop March 12, 2002 John Harvey, CERN LHCb Computing.
Grid Deployment Technical Working Groups: Middleware selection AAA,security Resource scheduling Operations User Support GDB Grid Deployment Resource planning,
J Jensen / WP5 /RAL UCL 4/5 March 2004 GridPP / DataGrid wrap-up Mass Storage Management J Jensen
JRA1 Middleware re-engineering
Bob Jones EGEE Technical Director
Workload Management Workpackage
Status of Task Forces Ian Bird GDB 8 May 2003.
Regional Operations Centres Core infrastructure Centres
BaBar-Grid Status and Prospects
EGEE Middleware Activities Overview
DataGrid Quality Assurance
U.S. ATLAS Grid Production Experience
Grid Deployment Area Status Report
Database Readiness Workshop Intro & Goals
Grid related projects CERN openlab LCG EDG F.Fluckiger
Readiness of ATLAS Computing - A personal view
Testing for patch certification
Testbed Software Test Plan Status
CMS report from FNAL demo week Marco Verlato (INFN-Padova)
LCG middleware and LHC experiments ARDA project
A conceptual model of grid resources and services
Operating the LCG and EGEE Production Grid for HEP
Leigh Grundhoefer Indiana University
Ian Bird LCG Project - CERN HEPiX - FNAL 25-Oct-2002
Report on GLUE activities 5th EU-DataGRID Conference
Porting LCG to IA64 Andreas Unterkircher CERN openlab May 2004
gLite The EGEE Middleware Distribution
Presentation transcript:

LCG experience in Integrating Grid Toolkits Zdenek Sekera IT Division , CERN 16/11/2018 zdenek.sekera@cern.ch CHEP2003

zdenek.sekera@cern.ch CHEP2003 Outline What is LCG? LCG project goals With whom we work ? Basic LCG milestones What is LCG-0 ? What are the plans for LCG-1 ? What is the process to get to LCG-1 ? Conclusions Authors: Piera Bettini, Ian Bird, Flavia Donno, Maarten Litmaath, Di Qing, Louis Poncet, Andrea Sciaba, Zdenek Sekera, Marco Serra, David Smith 16/11/2018 zdenek.sekera@cern.ch CHEP2003

zdenek.sekera@cern.ch CHEP2003 What is LCG? LCG = LHC Computing Grid Creating the infrastructure for computing needs of LHC HEP experiments (Atlas, CMS, Alice, LHCb) All are worldwide collaborations, consequently the software has to work worldwide 16/11/2018 zdenek.sekera@cern.ch CHEP2003

zdenek.sekera@cern.ch CHEP2003 Project Goals Goal – Prepare and deploy the LHC computing environment to help the experiments to analyze the data coming from the detectors applications - tools, frameworks, environment, persistency computing system  global grid service cluster  automated fabric collaborating computer centres  grid CERN-centric analysis  global analysis environment central role of data challenges This is not another grid technology project – it is a grid deployment project 16/11/2018 zdenek.sekera@cern.ch CHEP2003

zdenek.sekera@cern.ch CHEP2003 LCG goals LCG goal is to bring physics world together by creating a user friendly, production quality environment for data processing and physics analysis. How? By integrating different grid toolkits or grid middleware into a homogenous package to guarantee the interoperability among different “ways of doing things”. LCG does NOT write grid middleware. We want physicist to use the same commands no matter where they work, in the USA, Europe, Asia. This makes them feel “at home” and surely their efficiency is much higher. 16/11/2018 zdenek.sekera@cern.ch CHEP2003

What is “production quality”? It is all of the following in no particular order: availability 24 x 7 performance stability, robustness user friendliness maintainability user support 16/11/2018 zdenek.sekera@cern.ch CHEP2003

zdenek.sekera@cern.ch CHEP2003 Who are our partners? LCG is currently a “customer” of: iVDGL - VDT toolkit (including Globus & Condor toolkits) EDG - European DataGrid project EDT - European DataTag project (monitoring) Globus - underlying software GLUE schema – HICB (DataTag & iVDGL) product HICB = HEP Intergrid Coordination Board LCG can be considered a joint effort of all. It pulls together needed components from existing projects. We have to give credit to all developers in all these organizations for their effort in helping us to achieve the final LCG goal: to provide a well-defined user friendly environment. 16/11/2018 zdenek.sekera@cern.ch CHEP2003

zdenek.sekera@cern.ch CHEP2003 Basic LCG milestones LCG is focused on batch & experiments data challenges (phase 1) LCG release milestones: February - LCG-0 (deployment test, not publicly available) July - LCG-1 (production pilot) First publicly available LCG service November - LCG-1 (production system) Performance release needed for data challenges in 2004. We have a first set of integrated software and currently use it for deployment tests (LCG-0) 16/11/2018 zdenek.sekera@cern.ch CHEP2003

zdenek.sekera@cern.ch CHEP2003 LCG-0 WorldGrid demo last year (e.g Supercomputing 2002) demonstrated interoperability between VDT-EDG toolkits We went a step further by integrating VDT and EDG into a single package consisting of EDG 1.4.3 + VDT 1.1.6 + “Glue” schema It’s purpose is to setup infrastructure and release process rather than Grid functionality itself The important task is to determine the necessary work to be done by laboratories to get integrated 16/11/2018 zdenek.sekera@cern.ch CHEP2003

zdenek.sekera@cern.ch CHEP2003 LCG-0 deployment We are deploying it on tier 1 centers, goal is about 10 sites in the initial deployment Currently deployed by: CERN, RAL, CNAF, Taiwan being installed by: FNAL, U of Tokyo we expect a few more sites later 16/11/2018 zdenek.sekera@cern.ch CHEP2003

What did we learn with LCG-0 Release process is difficult: Many specific issues for different sites (how many service machines, what services on service machines, security issues) Packaging is an issue (USA x Europe, rpm x tar, pacman x LCFG) Installation is complicated, we cannot force ourselves on sites with our installation tools when they have already their own Configuration is major problem, it is far too complex Testing is a huge issue, we need to test different architectures, features, networking, interoperability 16/11/2018 zdenek.sekera@cern.ch CHEP2003

zdenek.sekera@cern.ch CHEP2003 Research? Deployment? If we want to/need to reconfigure service nodes, all neccessary changes are now propagated to configuration files manually. That cannot stay this way, we have to find much more automatic ways. Compare to Linux installation 3-4 years ago and today. We must probably go the same way of understanding what the real configuration issues are, only we would like to be much faster. The integration of worker nodes must be resolved by sites themselves, we will help. These issues (and more) make the difference between the research project ending with a demo and the product to be used for a production. 16/11/2018 zdenek.sekera@cern.ch CHEP2003

zdenek.sekera@cern.ch CHEP2003 What will be LCG-1? Expected to be: EDG 2.0 + VDT 1.1.7 (Globus 2.2.4) We are just beginning LCG-1. We have a Certification & Testing testbed at CERN consisting of 40 machines we can configure as desired. Certification & Testing is a joint project (e.g. VDT is testing Globus and Condor). 16/11/2018 zdenek.sekera@cern.ch CHEP2003

LCG-1 Certification Testbed U of Wisconsin 40 machines at CERN 16/11/2018 zdenek.sekera@cern.ch CHEP2003

zdenek.sekera@cern.ch CHEP2003 LCG-1 certification basic grid functionality connectivity grid services security resource brokering data management (replication, catalog) configurability error recovery real world applications site verification suite 16/11/2018 zdenek.sekera@cern.ch CHEP2003

Test and Validation process Developers machines Build system Development Testbed ~15cpu Certification Testbed ~40cpu Production Unit Test Build Integration Certification Production WPs add unit tested code to CVS repository Run nightly build & auto. tests Individual WP tests Grid certification Certified public release for use by apps. Build system Integration Team Test Group Users Tagged package WPs Tagged release selected for certification Overall release tests Certified release selected for deployment Application Certification Fix problems Appl. Representatives Releases candidate Releases candidate Tagged Releases Certified Releases Office hours 24x7 Bugzilla anomalies reports 16/11/2018 zdenek.sekera@cern.ch CHEP2003

zdenek.sekera@cern.ch CHEP2003 Conclusions With LCG-0 we have proved it is possible to integrate different toolkits (VDT & EDG) into a single package that can be repeatable deployed. We have learned the difficult issues (installation, configuration, testing) that will require a special attention for LCG-1. The process how to get the first (July) LCG-1 release out has been setup and is operational. Incremental releases will be required to correct problems according to their priorities up to November LCG-1 release and possibly beyond. There is a difficult and long road ahead towards LCG-1 due to the complex software and aggressive schedule but we believe it can be managed with a systematic determined approach. It is a highly collaborative project so everybody has to contribute. 16/11/2018 zdenek.sekera@cern.ch CHEP2003