NorduGrid and LCG middleware comparison

Slides:



Advertisements
Similar presentations
Andrew McNab - Manchester HEP - 22 April 2002 EU DataGrid Testbed EU DataGrid Software releases Testbed 1 Job Lifecycle Authorisation at your site More.
Advertisements

Andrew McNab - Manchester HEP - 2 May 2002 Testbed and Authorisation EU DataGrid Testbed 1 Job Lifecycle Software releases Authorisation at your site Grid/Web.
EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
Plateforme de Calcul pour les Sciences du Vivant SRB & gLite V. Breton.
INFSO-RI Enabling Grids for E-sciencE EGEE Middleware The Resource Broker EGEE project members.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
Makrand Siddhabhatti Tata Institute of Fundamental Research Mumbai 17 Aug
L ABORATÓRIO DE INSTRUMENTAÇÃO EM FÍSICA EXPERIMENTAL DE PARTÍCULAS Enabling Grids for E-sciencE Grid Computing: Running your Jobs around the World.
DataGrid WP1 Massimo Sgaravatto INFN Padova. WP1 (Grid Workload Management) Objective of the first DataGrid workpackage is (according to the project "Technical.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Information System on gLite middleware Vincent.
- Distributed Analysis (07may02 - USA Grid SW BNL) Distributed Processing Craig E. Tull HCG/NERSC/LBNL (US) ATLAS Grid Software.
LCG / ARC Interoperability Status Michael Grønager, PhD (UNI-C / NBI) January 19, 2006, Uppsala.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
June 24-25, 2008 Regional Grid Training, University of Belgrade, Serbia Introduction to gLite gLite Basic Services Antun Balaž SCL, Institute of Physics.
MTA SZTAKI Hungarian Academy of Sciences Introduction to Grid portals Gergely Sipos
INFSO-RI Enabling Grids for E-sciencE OSG-LCG Interoperability Activity Author: Laurence Field (CERN)
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Glite. Architecture Applications have access both to Higher-level Grid Services and to Foundation Grid Middleware Higher-Level Grid Services are supposed.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
INFSO-RI Enabling Grids for E-sciencE Αthanasia Asiki Computing Systems Laboratory, National Technical.
LCG LCG-1 Deployment and usage experience Lev Shamardin SINP MSU, Moscow
David Adams ATLAS ATLAS-ARDA strategy and priorities David Adams BNL October 21, 2004 ARDA Workshop.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
CNAF - 24 September 2004 EGEE SA-1 SPACI Activity Italo Epicoco.
NorduGrid's ARC: A Grid Solution for Decentralized Resources Oxana Smirnova (Lund University/CERN) for the NorduGrid collaboration ISGC 2005, Taiwan.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) Overveiw of the gLite middleware Yaodong Cheng
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI solution for high throughput data analysis Peter Solagna EGI.eu Operations.
EGEE Data Management Services
Gri2Win: Porting gLite to run under Windows XP Platform
Grid2Win: Porting of gLite middleware to Windows platform
Grid2Win Porting of gLite middleware to Windows XP platform
Grid Computing: Running your Jobs around the World
The Information System in gLite middleware
NorduGrid's ARC: A Grid Solution for Decentralized Resources
Regional Operations Centres Core infrastructure Centres
The EDG Testbed Deployment Details
Oxana Smirnova, Jakob Nielsen (Lund University/CERN)
Classic Storage Element
Use of Nagios in Central European ROC
Peter Kacsuk – Sipos Gergely MTA SZTAKI
gLite Information System(s)
Practical: The Information Systems
gLite Grid Services Salma Saber
GDB 8th March 2006 Flavia Donno IT/GD, CERN
Comparison of LCG-2 and gLite v1.0
Accounting at the T1/T2 Sites of the Italian Grid
The Information System in gLite
Grid2Win: Porting of gLite middleware to Windows XP platform
Sergio Fantinel, INFN LNL/PD
Introduction to Grid Technology
The GENIUS portal Roberto Barbera University of Catania and INFN
Grid2Win: Porting of gLite middleware to Windows XP platform
Grid Services Ouafa Bentaleb CERIST, Algeria
LCG-1 Status & Task Forces
Short update on the latest gLite status
Gri2Win: Porting gLite to run under Windows XP Platform
LCG middleware and LHC experiments ARDA project
High Energy Physics Computing Coordination in Pakistan
gLite Information System(s)
Data services in gLite “s” gLite and LCG.
EGEE Middleware: gLite Information Systems (IS)
The GENIUS portal and the GILDA t-Infrastructure
Overview of gLite Middleware
gLite Information System
gLite The EGEE Middleware Distribution
Information Services Claudio Cherubino INFN Catania Bologna
Presentation transcript:

NorduGrid and LCG middleware comparison ARC vs LCG2 NorduGrid and LCG middleware comparison Oxana Smirnova Lund/ATLAS August 2, 2004, CERN

Setting the field NorduGrid LCG Name at birth Nordic Testbed for Wide Area Computing and Data Handling Large Hadron Collider Computing Grid When Started in spring 2001 Started in spring 2002 Who Nordic universities: ATLAS teams, others CERN: IT division and LHC experiments Members and associates Dozens Hundreds Why Provide researchers with adequate computing power by harnessing distributed resources using Grid technologies 2004-08-02

Main activity areas NorduGrid LCG Nordic countries Certificate Authority Grid middleware (Advanced Resource Connector, ARC) development Applications support (mainly ATLAS) Support of resources (those in the universities) User support Grid deployment (over the available resources) includes user and applications support Grid Middleware (took over from EDG, collaboration with EGEE ) Applications (GEANT4, POOL, generators,…) Distributed Analysis Fabric and networking (mainly CERN, regional centers) 2004-08-02

Basic systems and services (figures from the NG Monitor and GridIce) NorduGrid LCG Computing resources Standalone stations, Linux clusters, pools, other UN*X systems PBS, SGE, Condor,… All shared ~40 sites Linux (RH7.3, soon: Scientific Linux) clusters PBS, LSF Some shared ~70 sites Storage resources Disk storage ~30 TB Disk and tape storage ~40 TB Certification Resources and users must be certified by a CA authorized by the common Security Team Resource management Up to resource owners Centrally coordinated User management ~500 users Centralized, VO-based ? Users ? SWEGRID computers count towards both; deploy only ARC so far 2004-08-02

Relation to Middleware NorduGrid LCG Started deploying EDG middleware in 2001 Started deploying EDG middleware in 2002 Started developing own middleware (ARC) in February 2002 Reason: EDG m/w could not support ATLAS DC1 production Tailored EDG middleware (LCG1 and LCG2) Reason: EDG m/w could not support Data Challenges by LHC experiments Uses ARC for production Satisfactory performance for ATLAS DC2, CMS (in Estonia), all kinds of Nordic researchers Uses LCG2 for Data Challenges Not quite satisfactory performance so far Will continue developing ARC, no need to switch to anything else willing to offer ARC to LCG Will develop and deploy gLite, together with EGEE EGEE requires Nordic sites to run LCG2/gLite 2004-08-02

Middleware: general ARC LCG2 License GPL EDG Platforms GNU Linux (runs on RH*, SuSE*, Mandrake*, Debian3.0 etc), tests on Mac OS X (GT2 issues) RedHat 7.3, in future: Scientific Linux, Mac OS X Grid middleware Globus 2.4.3, own patches Globus 2.2.4, patches by VDT GT3 RLS EDG RLS GACL - GSOAP - (?) Condor-G EDG VOMS Core developers 6 ± 2 Some dozens ? 2004-08-02

ARChitecture Goal: no single point of failure 2004-08-02

ARC-Geography 2004-08-02

Components overview (1/2) ARC LCG2 Computing element Grid Manager: resides on a master node Interface to LRMS Input data stage-in (LFN resolution etc) Output data stage-out and eventual registration Job management (kill, restart, logging, clean up etc) Pluggable (e.g. accounting plug-in, benchmarking...) Provides info for infosystem (stateful) CE: resides on a master node Includes some data management tools (not integrated) Storage element « Simple »: stripped down computing element (gridftp server + GACL) « Smart »: Reliable file transfer Communication with various data indexing services Asynchronous data manipulation « Classic »: gridftp server « SE »: SRM-enabled 2004-08-02

Components overview (2/2) ARC LCG2 Worker node Not needed Resides at every WN; mostly data management functionality User Interface A lightweigt client Security: certificate request, proxy... Job submission: matchmaking, scheduling, upload Job management: monitor, kill, proxy renewal etc Grid file movement: copy/erase/register etc Central service Job submission: RB contact Job management: monitor, kill etc GIIS Part of the Computing Element service (« server » package) Separate service (BDII) Resource Broker Not needed (UI’s functionality) Job submission: matchmaking, scheduling, sandbox transfer, logging 2004-08-02

Information System: MDS ARC LCG2 LDAP tree Own Cluster-oriented Glue LDAP schema GRIS At each cluster and SE GIIS No data caching Strict hierarchical system (Grid - Country or an international body - National body - Cluster) Multi-rooted at top and for big countries Caching on Some hierarchy Top-level GIISes replaced by BDII Clients Broker/User Interface Monitor Broker 2004-08-02

No caching, only registrants lists ARC MDS details No caching, only registrants lists Cluster sub-tree 2004-08-02

Grid workload management ARC LCG2 Job description Globus RSL + extensions JDL: Condor classAds User Interface (UI) Local to user or shared, unlimited amount (unknown, often several per user) Typically shared, several per testbed Matchmaking, scheduling Decentralized (local to every client), integrated into UI, relies on Infosys Dedicated Resource Broker (RB), approx. one per VO, relies on infosys Job description transaction UI uploads files directly to chosen clusters Sandbox is transferred via RB Job monitoring, management Direct queries to clusters Queries via the submission RB 2004-08-02

Grid data management ARC LCG2 Replica services Replica Catalog (Globus + patches), RLS (Globus) RLS (EDG) Input data LFN or URL defined in job description, staged by the server, cached If LFN or GUID is defined in job description, the job is steered to the location; otherwise files must be moved by the job (hence necessary to install LCG2 on every worker node) Output data LFN or URL defined in job description, uploaded and registered by the server FN (+LFN +SE) specified in job description, uploaded by the server Automatic, bulk replication Not supported (yet?) 2004-08-02

User management ARC LCG2 VO membership Several options: LDAP servers VOMS Plain text lists (served by HTTP) VO-based functionality - Local user mapping Resource availability Tags publishing Databases Monitoring 2004-08-02

Other services ARC LCG2 Monitoring Based entirely on Information System, no special sensors Uses additional sensors apart of GRISes Bookkeeping Under development (user-friendly interface is missing), Web-service Part of the WMS Distribution Available via FTP and Web: nightly snapshots, weekly tags, stable releases (RPM, tgz) Available via anonymous CVS checkout, LCFG (RPM) Support Web site, support ticket system, Bugzilla, discussion mailing list Web site (!), support ticket system (?), Savannah (?), plenty of mailing lists 2004-08-02

Summary A major difference: ARC concentrates most job and data management functionality in the cluster front-end (the Grid Manager) LCG2 requires installation on each WN; ARC does not Another major difference: Information System Wrong/outdated/insufficient information is the most destructive thing on Earth Another difference: ARC is oriented towards particular persons and small teams, while LCG2 is for large organizations ARC has a lightweight client, LCG2 has none ARC user can monitor each job on the Web, LCG can not ARC is portable to many systems, LCG is not LCG has tape storage facilities, NorduGrid has none LCG has common policies, NorduGrid has none Plenty of other differences, but many commonalities as well Certification Absence of a reliable accounting system No fine-grained information access Ultimately, the common goal – to help them physicists to find the Higgs  2004-08-02