NorduGrid's ARC: A Grid Solution for Decentralized Resources

Slides:



Advertisements
Similar presentations
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks MyProxy and EGEE Ludek Matyska and Daniel.
Advertisements

Grid Standardization from the NorduGrid/ARC perspective Balázs Kónya, Lund University, Sweden NorduGrid Technical Coordinator ETSI Grid Workshop on Standardization,
Current status of grids: the need for standards Mike Mineter TOE-NeSC, Edinburgh.
EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
Plateforme de Calcul pour les Sciences du Vivant SRB & gLite V. Breton.
A conceptual model of grid resources and services Authors: Sergio Andreozzi Massimo Sgaravatto Cristina Vistoli Presenter: Sergio Andreozzi INFN-CNAF Bologna.
Swedish participation in DataGrid and NorduGrid Paula Eerola SWEGRID meeting,
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Current Monte Carlo calculation activities in ATLAS (ATLAS Data Challenges) Oxana Smirnova LCG/ATLAS, Lund University SWEGRID Seminar (April 9, 2003, Uppsala)
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Status of Interoperability Markus Schulz.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
Introduction to NorduGrid ARC / Arto Teräs Slide 1(16) Introduction to NorduGrid ARC Arto Teräs Free and Open Source Software Developers' Meeting.
OSG Middleware Roadmap Rob Gardner University of Chicago OSG / EGEE Operations Workshop CERN June 19-20, 2006.
LCG and HEPiX Ian Bird LCG Project - CERN HEPiX - FNAL 25-Oct-2002.
NorduGrid: a collaboration, the middleware and related infrastructure projects Balázs Kónya Technical Coordinator Nordugrid Collaboration/Lund.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
EGEE / ARC Interoperability Status Michael Grønager, PhD (UNI-C / NBI / EGEE) for the NorduGrid collaboration September 28, 2005, Culham.
LCG / ARC Interoperability Status Michael Grønager, PhD (UNI-C / NBI) January 19, 2006, Uppsala.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks State of Interoperability Laurence Field.
© 2006 Open Grid Forum Enabling Pervasive Grids The OGF GIN Effort Erwin Laure GIN-CG co-chair, EGEE Technical Director
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks, An Overview of the GridWay Metascheduler.
Grid Middleware Tutorial / Grid Technologies IntroSlide 1 /14 Grid Technologies Intro Ivan Degtyarenko ivan.degtyarenko dog csc dot fi CSC – The Finnish.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks State of Interoperability Laurence Field.
OSG Tier 3 support Marco Mambelli - OSG Tier 3 Dan Fraser - OSG Tier 3 liaison Tanya Levshina - OSG.
CEOS WGISS-21 CNES GRID related R&D activities Anne JEAN-ANTOINE PICCOLO CEOS WGISS-21 – Budapest – 2006, 8-12 May.
INFSO-RI Enabling Grids for E-sciencE OSG-LCG Interoperability Activity Author: Laurence Field (CERN)
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Alex Read, Dept. of Physics Grid Activities in Norway R-ECFA, Oslo, 15 May, 2009.
Performance of The NorduGrid ARC And The Dulcinea Executor in ATLAS Data Challenge 2 Oxana Smirnova (Lund University/CERN) for the NorduGrid collaboration.
Globus and PlanetLab Resource Management Solutions Compared M. Ripeanu, M. Bowman, J. Chase, I. Foster, M. Milenkovic Presented by Dionysis Logothetis.
FP6−2004−Infrastructures−6-SSA E-infrastructure shared between Europe and Latin America gLite Information System Claudio Cherubino.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI Services for Distributed e-Infrastructure Access Tiziana Ferrari on behalf.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
J Jensen / WP5 /RAL UCL 4/5 March 2004 GridPP / DataGrid wrap-up Mass Storage Management J Jensen
NorduGrid's ARC: A Grid Solution for Decentralized Resources Oxana Smirnova (Lund University/CERN) for the NorduGrid collaboration ISGC 2005, Taiwan.
Grid Computing with Debian, Globus and ARC Mattias Ellert, Uppsala Universitet (.se) Steffen Möller, Universität zu Lübeck (.de) Anders Wäänänen, Niels.
Massimo Sgaravatto INFN Padova
Jean-Philippe Baud, IT-GD, CERN November 2007
Bob Jones EGEE Technical Director
Aleksandr Konstantinov, Weizhong Qiang NorduGrid collaboration
Workload Management Workpackage
Grid Computing: Running your Jobs around the World
The EDG Testbed Deployment Details
Oxana Smirnova, Jakob Nielsen (Lund University/CERN)
EGEE Middleware Activities Overview
Peter Kacsuk – Sipos Gergely MTA SZTAKI
Stephen Pickles Technical Director, GOSC
Globus —— Toolkits for Grid Computing
NorduGrid and LCG middleware comparison
Interoperability & Standards
LCG middleware and LHC experiments ARDA project
A conceptual model of grid resources and services
Ian Bird LCG Project - CERN HEPiX - FNAL 25-Oct-2002
Report on GLUE activities 5th EU-DataGRID Conference
From Prototype to Production Grid
EGEE Middleware: gLite Information Systems (IS)
The GENIUS portal and the GILDA t-Infrastructure
gLite The EGEE Middleware Distribution
Grid Computing Software Interface
Sergio Andreozzi Laurence Field Balazs Konya
Information Services Claudio Cherubino INFN Catania Bologna
Presentation transcript:

NorduGrid's ARC: A Grid Solution for Decentralized Resources Oxana Smirnova (Lund University/CERN) for the NorduGrid collaboration GDB, June 22, 2005, CERN

NorduGrid history 2001-2002: a part of the NORDUNet2 program, aimed to enable Grid middleware and applications in the Nordic countries Middleware: EDG Applications: HEP (ATLAS), theoretical physics Participants: academic groups from 4 Nordic countries Denmark: Research Center COM, DIKU, NBI Finland: HIP Norway: U. of Bergen, U. of Oslo Sweden: KTH, Stockholm U., Lund U., Uppsala U. (ATLAS groups) Since end-2002 is a research collaboration between Nordic academic institutes Open to anybody, non-binding Hardware: mostly rental resources and those belonging to users Since end-2003 focuses only on middleware Develops own Grid middleware: the Advanced Resource Connector (ARC) 6 core developers, many contributing student projects Provides middleware to research groups and national Grid projects ARC is now installed on ~50 sites (~5000 CPUs) in 13 countries all over the World www.nordugrid.org

ARC history NorduGrid had strong links with EDG WP6: active work with the ITeam; Nordic CA WP8: active work with ATLAS DC1 WP2: contribution to GDMP Attempts to contribute to RC, Infosystem Had to diverge from EDG in 2002 January 2002: became increasingly aware that EDG Is not suitable for non-dedicated resources with a non-CERN OS Won’t deliver a production-level middleware in time February 2002: developed own lightweight Grid architecture March 2002: prototypes of the core services in place April 2002: first live demos ran May 2002: entered a continuous production mode Since 2004, used by more and more national Grid projects, not necessarily related to NorduGrid or HEP/CERN www.nordugrid.org

ARC Grid A Grid based on ARC middleware Driven (so far) mostly by the needs of the LHC experiments One of the world’s largest production-level Grids Close cooperation with other Grid projects: EU DataGrid (2001-2003) SWEGRID, DCGC … NDGF LCG EGEE Assistance in Grid deployment outside the Nordic area Recently introduced: the ARC Community VO to join those who share their resources www.nordugrid.org

Goals The system must be: Light-weight Portable Non-intrusive: Resource owners retain full control; Grid is effectively a yet another user (with many faces though) No requirements w.r.t. OS, resource configuration, etc. Clusters need not be dedicated Runs independently of other existing Grid installation Client part must be easily installable by a novice user Trivial tasks must be trivial to perform Strategy: start with something simple that works for users and add functionality gradually www.nordugrid.org

Architecture Each resource has a front-end Authenticates users, interprets tasks, interacts with LRMS, publishes information, moves data Each user can have an independent lightweight brokering client (or many) Resource discovery, matchmaking, job submission and manipulation, monitoring Grid topology is achieved by an hierarchical, multi-rooted set of indexing services Monitoring relies entirely on the information system Ad-hoc data management, for the beginning www.nordugrid.org

Components Goal: no single point of failure www.nordugrid.org

Components Computing resources: Linux clusters/pools or workstations Addition of non-Linux resources is possible via Linux front-ends Front-end: Runs custom pluggable GridFTP server for all the communications Accepts job requests and formulates jobs for LRMS/fork Performs most data movement (stage in and out), cache management, interacts with data indexing systems Manages user work area Performs all kinds of job management upon client request Publishes system and job information www.nordugrid.org

Components Client: a lightweight User Interface with the built-in Resource Broker A set of command line utilities Minimal and simple Under the hood: resource discovery, matchmaking, optimization, job submission Complete support for single job management Basic functionality for multiple job management Support for single file manipulations Portals and GUI clients are being developed www.nordugrid.org

Components Information System: based on Globus-patched OpenLDAP: it uses GRIS and GIIS back-ends Keeps strict registration hierarchy Multi-rooted Effectively provides a pseudo-mesh architecture, similar to file sharing networks Information is only kept on the resource; never older than 30 seconds Own schema and providers www.nordugrid.org

Components Storage: any kind of storage system with a disk front-end Own GridFTP server implementation with pluggable back-ends Ordinary file system access Grid Access Control Lists (GACL) based access “Smart" Storage Element - WS based data service with direct support for Indexing Services (Globus’ RC, RLS) tape storage systems are being acquired www.nordugrid.org

Functionality Single-login access to a heterogeneous, distributed set of computing and storage resources Based on pre-WS Globus libraries and API Complete up-to-date information on the available resources Serial batch job submission to best resources available Matchmaking, brokering Job monitoring and management Seamless input/output data movement Basic data management Indexing, movement www.nordugrid.org

Features Light-weight Portable Dynamic, heterogeneous Non-intrusive: Resource owners retain full control No requirements w.r.t. OS, resource configuration, etc. Clusters need not be dedicated Runs independently of other existing Grid installation Client part is easily installable by a novice user Optimized for large data processing, bulk job submission Simple monitoring of the entire system and each job Specialization: oriented towards serial batch jobs Parallel jobs are perfectly possible, but only within a cluster; no optimization Interactive tasks are not supported No optimization for “Hello World” short jobs www.nordugrid.org

Distribution, availability At ftp.nordugrid.org: Stable releases, including: Binary RPMs and tar-balls are available for most Linux platforms Source RPMs Standalone client tar-ball for installation by a non-privileged user Only 13 MB when unpacked Contains all the EU Grid PMA approved CA keys Includes all the basic Globus client tools Weekly development builds Nightly builds CVS at cvs.nordugrid.org License: GPL More info, complete documentation, contacts at www.nordugrid.org www.nordugrid.org

On interoperability Some historical perspective 1988: Condor. No data management, no security infrastructure 1993: Legion/Avaki. Also calculation-oriented First object-oriented model; commercialized 1997: UNICORE. No data management still 1997: SRB. Only data management Source unavailable 1998: Globus Toolkit 1. GSI, GridFTP, but still no data management, very mixed success, MDS – a bottleneck Was about to die, but received some appreciation in Europe and became a de-facto standard: 2001: EDG. Makes use of GT2, Condor. Some data management (RC, GDMP) 2002: VDT. A “delivery channel” offering GT2, Globus etc – no data management 2002: NorduGrid/ARC. Makes use of GT2 libraries and data management (RC). Implements Web Services for some components 2002: OGSA. Object-oriented approach, modularity, STANDARDS UNICORE quickly moves towards OGSA standards 2003: LCG-1. Largely inherits EDG 2003: Grid3. Based on VDT. No data management 2005: GT4. Implements Web Services for some components 2005: gLite. EDG-line, but implements Web Services for some components, comprehensive basis for data management ... What more?... Note: most of the mentioned solutions are incompatible with each other www.nordugrid.org

On interoperability-2 Some quotes: Rob Baxter (EPCC, NeSC), October 2002: “do we have a real chance of a single, uniform architecture for distributed systems integration?” (speaking of OGSA) Mark Linesch (GGF Chairman), June 2005: “OGSA is in the early stages of development and standardization“ “GGF distinguishes between the OGSA architectural process, OGSA profiles and specifications, and OGSA software. All of these are important to maintain coherence around OGSA and Grid standards” “At the time of writing, we have OGSA Use Case documents, OGSA Architectural documents and drafts of an OGSA Basic Profile document and OGSA Roadmap document. We do not yet have any OGSA-compliant software implementations or OGSA compliance tests” www.nordugrid.org

On interoperability-3 For simplicity, let’s “forget” about UNICORE, Avaki, SRB, GT4 etc But for how long? UNICORE enjoys EU support, adhere to OGSA – maybe this is the standard to follow? GT4 has improved security infrastructure, Reliable File Transfer etc – perhaps this is the standard? LCG-2 and gLite are much closer to ARC than to e.g. UNICORE Short list of “our” services and solutions: Service/component LCG-2, gLite ARC Basis GT2 from VDT GT2 own patch, GT3 pre-WS Data transfer GridFTP, SRM v? (DPM) GridFTP, SRM v1.1 client Data management EDG RLS, Fireman & Co, LFC RC, RLS, Fireman Information LDAP, GLUE1.1, MDS+BDII, R-GMA LDAP, ARC schema, MDS-GIIS Job description JDL (based on classAds) RSL Job submission Condor-G to GRAM GridFTP VO management VOMS, gLite VOMS, CAS (?) VOMS www.nordugrid.org

On interoperability-4 Notes: Bottom line: Cooperation between ARC and Condor led in October 2004 to Condor-G version that can submit jobs to ARC GridFTP (translation from ARC infosystem schema to GLUE was developed by Rod Walker). Was meant to be used by LCG – but nobody configured an RB this way yet ARC does not use Condor components, can not submit jobs vice versa GLUE2 schema is expected to be developed soon, with participation of NorduGrid, OSG and others. All chances to get a common resource representation. “Rome” Common Resource Management initiative (includes Globus, UNICORE, LCG, EGEE, NorduGrid, NAREGI) converged on usage of GGF JSDL for job description JSDL v1.0 is still rudimentary, but is the least common denominator Bottom line: LCG and NorduGrid are quite close already and have all chances to get closer It might make sense to move together to a worldwide standard in cooperation with other Grid projects, instead of inventing new ways or converging to each other Respecting GGF recommendations (OGSA, SRM, JSDL etc) might be a good starting point Apparent substantial differences between gLite 1.1 and LCG-2, plus non-standard SRM are confusing – in the interoperability perspective. www.nordugrid.org

Conclusion NorduGrid’s ARC is a reliable and robust Grid middleware, supporting distributed production facilities already for almost 3 years, non-stop The middleware is in development, everybody is welcomed to use and contribute ARC is meant to offer Grid solution for decentralized, opportunistic resource usage: Using ARC does not give an automatic access to any resource: please negotiate with the resource owners (create Virtual Organizations) Deploying ARC does not open doors to all the users: only resource owners decide whom to authorize ARC developers are deeply interested and willing to take part in global Grid standardization and interoperability efforts Currently, only site and user certification is standardized, and to some extent – data transfer www.nordugrid.org