Grid Projects: EU DataGrid and LHC Computing Grid Oxana Smirnova Lund University October 29, 2003, Košice.

Slides:



Advertisements
Similar presentations
CERN STAR TAP June 2001 Status of the EU DataGrid Project Fabrizio Gagliardi CERN EU-DataGrid Project Leader June 2001
Advertisements

Stephen Burke - WP8 Status - 9/5/2002 Partner Logo WP8 Status Stephen Burke, PPARC/RAL.
Andrew McNab - Manchester HEP - 22 April 2002 EU DataGrid Testbed EU DataGrid Software releases Testbed 1 Job Lifecycle Authorisation at your site More.
EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
DataGrid is a project funded by the European Commission under contract IST WP2 – R2.1 Overview of WP2 middleware as present in EDG 2.1 release.
Plateforme de Calcul pour les Sciences du Vivant SRB & gLite V. Breton.
NIKHEF Testbed 1 Plans for the coming three months.
CERN The European DataGrid Project Technical status Bob Jones (CERN) Deputy Project Leader.
The DataGrid Project NIKHEF, Wetenschappelijke Jaarvergadering, 19 December 2002
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
The EU DataGrid - Introduction The European DataGrid Project Team
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
The EDG Testbed Deployment Details The European DataGrid Project
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
CERN TERENA Lisbon The Grid Project Fabrizio Gagliardi CERN Information Technology Division May, 2000
Andrew McNab - Manchester HEP - 5 July 2001 WP6/Testbed Status Status by partner –CNRS, Czech R., INFN, NIKHEF, NorduGrid, LIP, Russia, UK Security Integration.
13-May-03D.P.Kelsey, WP8 CA and VO organistion1 CA’s and Experiment (VO) Organisation WP8 Meeting EDG Barcelona, 13 May 2003 David Kelsey CCLRC/RAL, UK.
Computational grids and grids projects DSS,
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
DataGrid WP1 Massimo Sgaravatto INFN Padova. WP1 (Grid Workload Management) Objective of the first DataGrid workpackage is (according to the project "Technical.
IST E-infrastructure shared between Europe and Latin America High Energy Physics Applications in EELA Raquel Pezoa Universidad.
WP8 Status – Stephen Burke – 30th January 2003 WP8 Status Stephen Burke (RAL) (with thanks to Frank Harris)
Training and the NGS Mike Mineter
The European DataGrid Project Team The EU DataGrid.
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
GridPP Presentation to AstroGrid 13 December 2001 Steve Lloyd Queen Mary University of London.
The Grid approach for the HEP computing problem Massimo Sgaravatto INFN Padova
CEOS WGISS-21 CNES GRID related R&D activities Anne JEAN-ANTOINE PICCOLO CEOS WGISS-21 – Budapest – 2006, 8-12 May.
CLRC and the European DataGrid Middleware Information and Monitoring Services The current information service is built on the hierarchical database OpenLDAP.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
J.J.Blaising April 02AMS DataGrid-status1 DataGrid Status J.J Blaising IN2P3 Grid Status Demo introduction Demo.
The European DataGrid Project Fabrizio Gagliardi EU DataGrid Project Leader CERN
…building the next IT revolution From Web to Grid…
Introduction to GRID computing and overview of the European Data Grid The European DataGrid Project
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
INFSO-RI Enabling Grids for E-sciencE Experience of using gLite for analysis of ATLAS combined test beam data A. Zalite / PNPI.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
The EU DataGrid The European DataGrid Project Team
29/1/2002A.Ghiselli, INFN-CNAF1 DataTAG / WP4 meeting Cern, 29 January 2002 Agenda  start at  Project introduction, Olivier Martin  WP4 introduction,
1 P.Kunszt Openlab Lessons learned from Data Management in the EU DataGrid Peter Kunszt CERN IT/DB EU DataGrid Data Management
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
Performance of The NorduGrid ARC And The Dulcinea Executor in ATLAS Data Challenge 2 Oxana Smirnova (Lund University/CERN) for the NorduGrid collaboration.
Partner Logo A Tier1 Centre at RAL and more John Gordon eScience Centre CLRC-RAL HEPiX/HEPNT - Catania 19th April 2002.
EGEE-0 / LCG-2 middleware Practical.
LCG CERN David Foster LCG WP4 Meeting 20 th June 2002 LCG Project Status WP4 Meeting Presentation David Foster IT/LCG 20 June 2002.
15-Feb-02Steve Traylen, RAL WP6 Test Bed Report1 RAL/UK WP6 Test Bed Report Steve Traylen, WP6 PPGRID/RAL, UK
DataGrid is a project funded by the European Commission under contract IST rd EU Review – 19-20/02/2004 The EU DataGrid Project Three years.
LHC Computing, SPC-FC-CC-C; H F Hoffmann1 CERN/2379/Rev: Proposal for building the LHC computing environment at CERN (Phase 1) Goals of Phase.
The DataGrid Project NIKHEF, Wetenschappelijke Jaarvergadering, 19 December 2002
All-sky search for continuous gravitational waves: tests in a grid environment Cristiano Palomba INFN Roma1 Plan of the talk: Computational issues Computing.
Stephen Burke – Sysman meeting - 22/4/2002 Partner Logo The Testbed – A User View Stephen Burke, PPARC/RAL.
Status of gLite-3.0 deployment and uptake Ian Bird CERN IT LCG-LHCC Referees Meeting 29 th January 2007.
Bob Jones – Project Architecture - 1 March n° 1 Project Architecture, Middleware and Delivery Schedule Bob Jones Technical Coordinator, WP12, CERN.
9-Jul-02D.P.Kelsey, DataGrid Security1 EU DataGrid Security 9 July 2002 UK Security Task Force Meeting #2 David Kelsey CLRC/RAL, UK
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
10-Feb-00 CERN HepCCC Grid Initiative ATLAS meeting – 16 February 2000 Les Robertson CERN/IT.
The European DataGrid Project
The EDG Testbed Deployment Details
DataGrid WP6/CA CA Trust Matrices
General Project Manager
Short update on the latest gLite status
LCG experience in Integrating Grid Toolkits
LHC Data Analysis using a worldwide computing grid
Gridifying the LHCb Monte Carlo production system
gLite The EGEE Middleware Distribution
Presentation transcript:

Grid Projects: EU DataGrid and LHC Computing Grid Oxana Smirnova Lund University October 29, 2003, Košice

Outlook Precursors: attempts to meet tasks of HEP computing EDG: the first global Grid development project LCG: deploy computing environment for LHC experiments

Characteristics of HEP computing Eventindependence Event independence  Data from each collision is processed independently: trivial parallelism  Mass of independent problems with no information exchange Massivedatastorage Massive data storage  Modest event size: 1 – 10 MB (not ALICE though)  Total is very large – Petabytes for each experiment. Mostlyreadonly Mostly read only  Data never changed after recording to tertiary storage  But is read often! A tape is mounted at CERN every second! Resilience rather than ultimate reliability  Individual components should not bring down the whole system  Reschedule jobs on failed equipment Modestfloatingpointneeds Modest floating point needs  HEP computations involve decision making rather than calculation

Department    Desktop CERN – Tier 0 MONARC report: Tier 1 FNAL RAL IN2P3 622 Mbps 2.5 Gbps 622 Mbps 155 mbps Tier2 Lab a Uni. b Lab c Uni. n MONARC: hierarchical regional centres model

EU Datagrid project In certain aspects was initiated as a MONARC follow- up, introducing the Grid technologies Started on January 1, 2001, to deliver by end 2003  Aim: to develop a Grid middleware suitable for High Energy physics, Earth Observation, biomedical applications and live demonstrations  9.8 MEuros EU funding over 3 years  Development based on existing tools, e.g., Globus, LCFG, GDMP etc Maintains development and applications testbeds, which include several sites across the Europe

EDG overview : Main partners CERN – International (Switzerland/France) CNRS – France ESA/ESRIN – International (Italy) INFN – Italy NIKHEF – The Netherlands PPARC – UK Slide by EU DatGrid

Research and Academic Institutes CESNET (Czech Republic) Commissariat à l'énergie atomique (CEA) – France Computer and Automation Research Institute, Hungarian Academy of Sciences (MTA SZTAKI) Consiglio Nazionale delle Ricerche (Italy) Helsinki Institute of Physics – Finland Institut de Fisica d'Altes Energies (IFAE) - Spain Istituto Trentino di Cultura (IRST) – Italy Konrad-Zuse-Zentrum für Informationstechnik Berlin - Germany Royal Netherlands Meteorological Institute (KNMI) Ruprecht-Karls-Universität Heidelberg - Germany Stichting Academisch Rekencentrum Amsterdam (SARA) – Netherlands Swedish Research Council - Sweden EDG overview : Assistant Partners Industrial Partners Datamat (Italy) IBM-UK (UK) CS-SI (France) Slide by EU DatGrid

EDG work-packages  WP1: Work Load Management System  WP2: Data Management  WP3: Grid Monitoring / Grid Information Systems  WP4: Fabric Management  WP5: Storage Element, MSS support  WP6: Testbed and demonstrators  WP7: Network Monitoring  WP8: High Energy Physics Applications  WP9: Earth Observation  WP10: Biology  WP11: Dissemination  WP12: Management

Simplified Grid deployment approach Homogeneous structure  All the sites must run with the same OS and kernel (Linux, RedHat7.3)  Recommended central installation via LCFG service (installs entire machine from scratch on each reboot)  Exceptions are possible, but not supported Invasive installation  Requires massive existing cluster re-configuration  Needs to be installed on every compute node

Basic EDG services Workload management  Resource Broker (RB) and Job Submission Service (JSS)  Logging and Bookkeeping Service (L&B)  Information Index (II)  User Interface (UI) Data management  Replica Location Service (RLS)  Replica Metadata Catalog (RMC)  Replica Optimization Service (ROS) Information and monitoring service  Relational Grid Monitoring Architecture (R-GMA) Fabric management Mass storage management Virtual Organization management

Typical EDG site composition Site-specific: User Interface (UI) Computing Element or Service (CE)  Gatekeeper (GK)  Worker Nodes (WN) do have client APIs for accessing EDG services and information Storage Element (SE) Monitoring Node (MON)  R-GMA servlets for the site  ROS Common: Resource Broker (RB) RLS  Local Replica Catalog (LRC)  RMC Information Catalog (IC)

Organization of user access Users must have valid personal Globus-style certificates  Group or anonymous certificates are not allowed  Certificate Issuing Authority (CA) must be endorsed by the EDG Security Group  If there is no approved CA in your country/region, France catches all Users must belong to one of the accepted Virtual Organizations (VO)  LHC experiments, biomedical and Earth Observation applications, and some EDG teams  VO lists are managed by experiments/teams representatives  Users can belong to several VOs  Users with identical names or a user with several certificates can not belong to a same VO  Local system administrators still have a full control To “log into the Grid”, users make use of the private certificate to issue a public proxy Grid sites accept requests only from users whose certificates are signed by CAs that a site accepts

EDG applications testbed EDG is committed to create a stable testbed to be used by applications for real tasks  This started to materialize in August 2002…  …and coincided with the ATLAS DC1  CMS joined in December  ALICE, LHCb – smaller scale tests At the moment (October 2003) consists of ca. 15 sites in 8 countries Most sites are installed from scratch using the EDG tools (require/install RedHat 7.3)  Some have installations on the top of existing resources  A lightweight EDG installation is available Central element: the Resource Broker (RB), distributes jobs between the resources  Most often, a single RB is used  Some tests used RBs “attached” to User Interfaces  In future, may be an RB per Virtual Organization (VO) or/and per user ?

EDG Applications Testbed snapshot

Basic EDG functionality as of today UI CASTOR RLS CE RB do rfcp rfcp replicate RM jdl +R-GMA NFS RSL Output RM Input Output

EDG status The EDG1 was not a very satisfactory prototype  Highly unstable behavior  Somewhat late deployment  Many missing features and functionalities The EDG2 is released and deployed for applications on October 20, 2003  Many services have been re-written since EDG1  Some functionality have been added, but some have been lost  Stability is still the issue, esp. the Information System performance  Little has been done to streamline applications environment deployment  No production-scale tasks have been shown to perform reliably yet No development will be done beyond this point  Bug fixing will continue for a while  Some “re-engineering” is expected to be done by the next EU-sponsored project – EGEE

The future: LCG LCG  LHC Computing Grid Goal: to deploy an adequate information and computational infrastructure for the LHC experiments Means of achieving: using the modern distributed computing and data analysis tools and utilities – The Grid Resources: large computing centers around the World as the basic elements  Research institutes, laboratories and universities are also members of the data analysis chain  No need to concentrate the computing power at CERN

LCG Timeline September 2001: the project is approved by the CERN Council Duration: 2002 to 2008  Phase 1: prototyping, testing  Phase 2: deployment of the LHC computing infrastructure2 November 2003: a functioning LCG-1 prototype (a criterion: 30 consecutive days of non-stop operation); includes 10 regional centers May 2004: research lab and institutes are joining with their resources December 2004: LCG-3, 50% of expected by 2007 performance IX/ Phase 1 Phase 2 XI/03V/04XII/04

LCG organization Financing:  CERN and other states participating in LHC projects  Business partners  LHC experiments  National research foundations and computing centers  Projects financed my EU and other international funds Structure:  Applications  CERN fabric  Grid technology  Grid deployment FOR MORE INFO:

First priority: LCG-1 Computing clusterNetwork resourcesData storage Operating systemLocal schedulerFile system User accessSecurity Data transfer Information schema Global schedulerData managementInformation system User interfaces Applications Major components and levels: Hardware System software Passive services Active services High level services Closed system (?) HPSS, CASTOR… RedHat Linux NFS, … PBS, Condor, LSF,… VDT (Globus, GLUE) EU DataGrid LCG, experiments

grid for a physics study group grid for a regional group Tier2 Lab a Uni a Lab c Uni n Lab m Lab b Uni b Uni y Uni x Tier3 physics department    Desktop Germany Tier 1 USA UK France Italy Taiwan CERN Tier 1 Japan The LHC Computing Centre CERN Tier 0 LHC Grid: what became of the MONARC hierarchy

LCG status Grid component: almost entirely the EDG solution  Major difference: LCG-1 still has the “old” MDS for the information system  Deployed at the LCG testbed, non-overlapping with the EDG in general, includes non-EU countries like US, Russia or Taiwan  More stable so far than EDG (for MDS?..)  Little or no Grid development  In future, may consider alternative Grid solutions, e.g., the AliEn (though unlikely)  Grid Technology area is on the verge of being dismissed, as LCG will not be doing Grid development LHC Applications component:  A lot of very serious development  Many areas are covered, from generators to Geant4 to data management etc  Unfortunately, has little interaction and co-operation with Grid developers

LCG-1 Testbed

Summary Initiated by CERN, EDG came as the first global Grid R&D project aiming at deploying working services Sailing in uncharted waters, EDG ultimately provided a set of services, allowing to construct a Grid infrastructure Perhaps the most notable EDG achievement is introduction of authentication and authorization standards, now recognized worldwide LCG took a bold decision to deploy EDG as their Grid component for the LCG-1 release The Grid development does not stop with EDG: LCG is open for new solutions, with a strong preference towards OGSA