Www.nordugrid.org NorduGrid: a collaboration, the middleware and related infrastructure projects Balázs Kónya Technical Coordinator Nordugrid Collaboration/Lund.

Slides:



Advertisements
Similar presentations
Grid Standardization from the NorduGrid/ARC perspective Balázs Kónya, Lund University, Sweden NorduGrid Technical Coordinator ETSI Grid Workshop on Standardization,
Advertisements

30-31 Jan 2003J G Jensen, RAL/WP5 Storage Elephant Grid Access to Mass Storage.
Current status of grids: the need for standards Mike Mineter TOE-NeSC, Edinburgh.
Data Management Expert Panel - WP2. WP2 Overview.
EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
A conceptual model of grid resources and services Authors: Sergio Andreozzi Massimo Sgaravatto Cristina Vistoli Presenter: Sergio Andreozzi INFN-CNAF Bologna.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Massimo Cafaro GridLab Review GridLab WP10 Information Services Massimo Cafaro CACT/ISUFI University of Lecce, Italy.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
CSC Grid Activities Arto Teräs HIP Research Seminar February 18th 2005.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
Kate Keahey Argonne National Laboratory University of Chicago Globus Toolkit® 4: from common Grid protocols to virtualization.
Open Science Grid Software Stack, Virtual Data Toolkit and Interoperability Activities D. Olson, LBNL for the OSG International.
A Lightweight Platform for Integration of Resource Limited Devices into Pervasive Grids Stavros Isaiadis and Vladimir Getov University of Westminster
Grid Computing - AAU 14/ Grid Computing Josva Kleist Danish Center for Grid Computing
GT Components. Globus Toolkit A “toolkit” of services and packages for creating the basic grid computing infrastructure Higher level tools added to this.
HPDC 2007 / Grid Infrastructure Monitoring System Based on Nagios Grid Infrastructure Monitoring System Based on Nagios E. Imamagic, D. Dobrenic SRCE HPDC.
LCG and HEPiX Ian Bird LCG Project - CERN HEPiX - FNAL 25-Oct-2002.
NorduGrid Architecture and tools CHEP2003 – UCSD Anders Wäänänen
2005 © SWITCH Perspectives of Integrating AAI with Grid in EGEE-2 Christoph Witzig Amsterdam, October 17, 2005.
Grid Workload Management Massimo Sgaravatto INFN Padova.
EGEE / ARC Interoperability Status Michael Grønager, PhD (UNI-C / NBI / EGEE) for the NorduGrid collaboration September 28, 2005, Culham.
LCG / ARC Interoperability Status Michael Grønager, PhD (UNI-C / NBI) January 19, 2006, Uppsala.
Alex Read, Dept. of Physics Grid Activity in Oslo CERN-satsingen/miljøet møter MN-fakultetet Oslo, 8 juni 2009 Alex Read.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Grid Middleware Tutorial / Grid Technologies IntroSlide 1 /14 Grid Technologies Intro Ivan Degtyarenko ivan.degtyarenko dog csc dot fi CSC – The Finnish.
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
1October 9, 2001 Sun in Scientific & Engineering Computing Grid Computing with Sun Wolfgang Gentzsch Director Grid Computing Cracow Grid Workshop, November.
OSG Tier 3 support Marco Mambelli - OSG Tier 3 Dan Fraser - OSG Tier 3 liaison Tanya Levshina - OSG.
GRIDS Center Middleware Overview Sandra Redman Information Technology and Systems Center and Information Technology Research Center National Space Science.
INFSO-RI Enabling Grids for E-sciencE OSG-LCG Interoperability Activity Author: Laurence Field (CERN)
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
Conference name Company name INFSOM-RI Speaker name The ETICS Job management architecture EGEE ‘08 Istanbul, September 25 th 2008 Valerio Venturi.
Andrew McNabSecurity Middleware, GridPP8, 23 Sept 2003Slide 1 Security Middleware Andrew McNab High Energy Physics University of Manchester.
Introduction to Grids By: Fetahi Z. Wuhib [CSD2004-Team19]
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Alain Roy Computer Sciences Department University of Wisconsin-Madison Condor & Middleware: NMI & VDT.
KnowARC objectives & challenges Balázs Kónya/Lund University Oslo, 1 st KnowARC Conference.
Status Organization Overview of Program of Work Education, Training It’s the People who make it happen & make it Work.
Alex Read, Dept. of Physics Grid Activities in Norway R-ECFA, Oslo, 15 May, 2009.
7. Grid Computing Systems and Resource Management
Easy Access to Grid infrastructures Dr. Harald Kornmayer (NEC Laboratories Europe) Dr. Mathias Stuempert (KIT-SCC, Karlsruhe) EGEE User Forum 2008 Clermont-Ferrand,
Performance of The NorduGrid ARC And The Dulcinea Executor in ATLAS Data Challenge 2 Oxana Smirnova (Lund University/CERN) for the NorduGrid collaboration.
International Symposium on Grid Computing (ISGC-07), Taipei - March 26-29, 2007 Of 16 1 A Novel Grid Resource Broker Cum Meta Scheduler - Asvija B System.
Globus and PlanetLab Resource Management Solutions Compared M. Ripeanu, M. Bowman, J. Chase, I. Foster, M. Milenkovic Presented by Dionysis Logothetis.
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
FP6−2004−Infrastructures−6-SSA E-infrastructure shared between Europe and Latin America gLite Information System Claudio Cherubino.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
John Kewley e-Science Centre All Hands Meeting st September, Nottingham GROWL: A Lightweight Grid Services Toolkit and Applications John Kewley.
EGI Technical Forum Amsterdam, 16 September 2010 Sylvain Reynaud.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
The European KnowARC Project Péter Stefán, NIIF/HUNGARNET/KnowARC TNC2009, June 2009, Malaga.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
OSG Status and Rob Gardner University of Chicago US ATLAS Tier2 Meeting Harvard University, August 17-18, 2006.
NDGF – a Joint Nordic Production Grid Lars Fischer ICFA Workshop on HEP Networking, Grid, and Digital Divide Issues for Global e-Science Cracow, 2 October.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
NorduGrid's ARC: A Grid Solution for Decentralized Resources Oxana Smirnova (Lund University/CERN) for the NorduGrid collaboration ISGC 2005, Taiwan.
Bob Jones EGEE Technical Director
Aleksandr Konstantinov, Weizhong Qiang NorduGrid collaboration
NorduGrid's ARC: A Grid Solution for Decentralized Resources
Oxana Smirnova, Jakob Nielsen (Lund University/CERN)
Farid Ould-Saada KnowARC’06 Conference Oslo 7-9/11/06
Ian Bird LCG Project - CERN HEPiX - FNAL 25-Oct-2002
Report on GLUE activities 5th EU-DataGRID Conference
Information Services Claudio Cherubino INFN Catania Bologna
Presentation transcript:

NorduGrid: a collaboration, the middleware and related infrastructure projects Balázs Kónya Technical Coordinator Nordugrid Collaboration/Lund University OSG Consortium Meeting, August 22, 2006

outline  NorduGrid Collaboration  The Advanced Resource Connector (ARC) middleware –Design & main components –Status & future development –Interoperability & Standards  Infrastructure projects using ARC

NorduGrid collaboration: some history  : a research project of the NORDUNet2 program aimed to enable Grid in the Nordic countries  Since end-2002 is a research collaboration between Nordic academic institutes –Open to anybody, non-binding  Since end-2003 focuses on middleware –Develops own Grid middleware: the Advanced Resource Connector (ARC) –Provides middleware to research groups and national Grid projects  ARC is now installed on ~50 sites (~5000 CPUs) in 14 countries all over the World

The NorduGrid Collaboration From To –EDG>ARC –Tesbed>50 sites –HEP>multidisciplinary +Bio,Chem.,.. –4 Nordic>13 countries –20 cpu’s>5000 cpu’s –2001> from a research project to a research collaboration...from a Grid testbed to a major middleware provider NOT an infrastructure, does not operate or control resources

How did ARC appear  Back in High Energy Physics Institutes from Scandinavia wanted to share their computing resources and jointly contribute to CERN/LHC computing –They needed a Grid! –The Grid hype just begun –Globus was regarded as the “de facto standard” middleware  NO production ready middleware was available or seen on the horizon as of November 2001: –Very alpha Globus GT-2.0 (GRAM-1.5, MDS-2.0); nevertheless Globus & IBM already started to work on OGSA/I, i.e. GT v.3 (which was announced in February 2002) –EDG middleware was in an extremely embryonic phase  Since May 2002 ARC has been used in production Data Challenges

Design philosophy (1/2) 1.The system must be: a)Light-weight b)Portable & modular c)Non-intrusive on the resource side: Resource owners retain full control No requirements w.r.t. OS, resource configuration, etc. Clusters need not be dedicated Runs independently of other existing Grid installation d)Special attention to functionality & performance “Traditionally, Scandinavian design has been associated with simple, uncomplicated designs, functionality and a democratic approach”

Design philosophy (2/2) a)Flexible & powerful on the client part must be easily installable by a novice user trivial tasks must be trivial to perform no dependency on central services No central client(s), create a real distributed system 1.Strategy: start with something simple that works for users and add functionality gradually Source of design illustrations: “Scandinavian Design beyond the Myth”

ARC components Goal: no single point of failure

Architecture key points  Each resource has a front-end –Authenticates users, interprets tasks, interacts with LRMS, publishes information, moves data –Resources are Grid-enabled by the ARC layer deployed on the front-end, no middleware components behind the front-end!  Each user can have an independent lightweight brokering client (or many) –Resource discovery, matchmaking, job submission and manipulation, monitoring  Grid topology is achieved by an hierarchical, multi- rooted set of indexing services  Monitoring relies entirely on the information system  Ad-hoc data management, for the beginning

 Computing resources: Grid-enabled via ARC layer on head node (front-end): –Custom GridFTP server for all the communications –Grid Manager handles job management upon client request, interfaces to LRMS –Performs most data movement (stage in and out), cache management, manages user work area –Publishes resource and job information via LDAP Computing service: the key component no GRAM!!!

Components: Clients  Client: a lightweight User Interface with the built-in Resource Broker –A set of command line utilities –Minimal and simple –Under the hood: resource discovery, matchmaking, optimization, job submission –Complete support for single job management –Basic functionality for multiple job management –Support for single file manipulations –Built upon ARCLIB  Portals and GUI clients are being developed

Components: Infosystem  Information System: based on Globus-patched OpenLDAP: it uses GRIS and GIIS back-ends –Keeps strict registration hierarchy –Multi-rooted –Effectively provides a pseudo-mesh architecture, similar to file sharing networks –Information is only kept on the resource; never older than 30 seconds –Own schema and providers

Components: Storages  Storage: any kind of storage system with a disk front-end –Conventional Storage: Own GridFTP server implementation with pluggable back-ends Ordinary file system access Grid Access Control Lists (GACL) based access –“Smart" Storage Element: WS based data service with direct support for Indexing Services (Globus’ RC, RLS) –no tape storage systems in use so far

Functionality overview  Provides reliable implementation of fundamental Grid services: –The usual grid security: single sign on, Grid ACLs (GACL), VOs (VOMS) –Job submission: direct or via matchmaking and brokering –Information services: resource aggregation, representation, discovery and monitoring –Implements core data management functionality Automated seamless input/output data movement Data Indexing (RLS, Fireman), client-side data movement –Job monitoring & management –Logging service  Builds upon standard open source solutions and protocols –Globus Toolkit® pre-WS libraries (no services, no GRAM!) –OpenLDAP, OpenSSL, SASL, SOAP, GridFTP, GSI

What is ARC today; status  General purpose Open Source European Grid middleware –Being developed & maintained by the NorduGrid Collaboration –Deployment support –Extensive documentation –Support for most of the popular Linux distributions  Lightweight architecture for a dynamic heterogeneous system –non-intrusive on the server side: coexistence with other middlewares on the same resource –standalone client (13 MB, pre-configured tar-ball)  User- & performance-driven development –Production quality software since May 2002 –First middleware ever to contribute to HEP data challenge  Middleware of choice by many national academic projects due to its technical merits –SWISS Grid(s), Finnish M-Grid, NDGF, etc… –Majority of ARC users now are NOT from the HEP community

ARC development future  Production sites run stable releases 0.4.x –Released in April 2004, took 2 years to develop, Globus 2, pre-WS technology, most basic functionality  Release 0.6 (second production ARC release) should be out real soon –Development branch 0.5.x is already used as a release candidate e.g. in ATLAS’ Dulcinea executor and other clients. –Currently working on documentation and packaging  Ambitious development plan to turn ARC into a next generation Grid middleware –SOA, WS interfaces based on current powerful core components –Extend data management functionalities  ARC's developer force: Open source community governed by Nordugrid Collaboration with dedicated personnel: –KnowARC project (17 FTEs) –NDGF software developer team (6 FTE) –NorduNet3 (5 Phd students) –Academic research groups in Umeå (SE), Aalborg (DK)

The KnowARC project  EU FP6 Specific Targeted Research Project –"Grid-enabled Know-how Sharing Technology Based on ARC Services and Open Standards" –Started: June 1 st, 2006 runs for three years –Partners: NorduGrid members and research teams in medicine, bioinformatics, physics, engineering, automotive industry apps, IT (10 partners from 7 countries)  Objectives: –to create a novel, powerful Next Generation Grid middleware based on ARC, widely respected for its simplicity, non-invasiveness and cost- efficiency; –to promote Grid standardization and interoperability; –to contribute to Grid technologies take-up, bridging the gaps between business and academia in Grid development  Will be a major contributor to the development of the next generation ARC middleware that will be the next step after ARC 0.6, addressing current limitations and shortcomings

KnowARC development areas  Core Services –Next generation Grid middleware architecture survey and design –Web Service interfaces over ARC services –Back-ends –Sandboxing & virtualization –Security framework (delegation)  Higher level services –Self-healing flexible storage and user-friendly storage interface –Self healing grid jobs: job migration & job manager –P2P-like information backbone, novel brokering approaches –Scalable accounting service –Dynamic application framework management –ARC-enabled Taverna and flowGuide (workflow engines)  Standards & Interoperability –OGSA –gLite gateway  Applications –Automotive industry –Medical image processing –Statistical genomics –Engineering portal (construction industry)  Quality assurance, software distribution –Build system, testing, support –Profiling, performance analysis, usability studies –Pilot Grid system –ARC in major Linux distributions –New platforms: Windows, Solaris, Mac OS-X

On interoperability of “pre-standard” middlewares  For simplicity, let’s “forget” most of the major middlewares (UNICORE, Avaki, SRB, GT4, Condor, etc) and take a look at only on the Glite(LCG) vs ARC interoperability issue: –both are production level middlewares –both are pre-standard implementations (non WS-based systems) Security Job submission Job description Information Data management Data transfer Basic software stack Service/component ARC protocol via GridFTPmixture of Condor & GRAM GSI, VOMS, GACLGSI, VOMS, MyProxy, CAS (?) RSL, soon JSDL supportJDL (based on classAds) LDAP, ARC schema, ARC-GIIS, interested in “Glue2” LDAP, GLUE1.1, BDII-GIIS, R- GMA, interested (2) “Glue2” RC, RLS, FiremanEDG RLS, Fireman & Co, LFC GridFTP, SRM v1.1 clientGridFTP, SRM v? (DPM) pre-WS GT libraries, own patches GT2 solutions from VDT ARCGlite (LCG)

Interoperability  Main difficulty –middleware stacks are continuously changing –components have closed, undocumented interfaces  GGF GIN effort aims to collect the low-hanging fruits –so far targets only GRAM-based systems, ARC is the exception –Ninf-G invoke remote process module for ARC has been developed  Glue schema initiative –Glue v 1.2 is rather limited –Glue 1.2 vs. ARC mapping: –NorduGrid has been interested in a major Glue revision since Glue v1.1 Glue schema 1.2 meeting notes: “Earliest possible 2.0 discussions October (2005),might be possible to start discussion with NorduGrid on some common design” Glue 2.0 is being postponed NorduGrid Glue 2.0 thoughts thoughts  Job submission gateways through Condor

gLite to ARC gateway

Standardization: the NorduGrid perspective  Interoperability efforts: good starting point in understanding different existing systems and documenting their interfaces  Fragmentation of the standardization landscape and process –by far too many (non-implemented) proposals of numerous Standard Development Organizations –integration concerns: can all these proposals be implemented and used together?  It is desirable that major middleware providers become more dedicated in influencing and implementing emerging standards –major players: Globus, Condor (OSG), Unicore, Glite (EGEE), OMII, Chinese middleware(s), ObjectWeb (Proactive), ARC (NorduGrid) –there is a threat that Grid community will keep producing non- implemented, conflicting standards  NorduGrid would like to avoid emerging “de facto standards” defined by implementations  Interfaces, Interfaces, Interfaces

Standardization: the NorduGrid perspective  Areas where standardization & implementations are urgently needed: –Job description language (JSDL1.x?) JSDL 1.0 is deliberately narrowly focused –Representation of Grid-related objects (Glue2?, CIM?) Computing & storage resources Grid jobs ( simple things such as “job states” are not defined!!!) –Standard interface to computing resources (execution service) Job submission, Job management, Job monitoring (OGSA-BES?) –Standard interface to Storages (SRM?) –(pre-deployed) Grid Application environments/frameworks –Description of Virtual Organizations –Standards related to Grid economy (e.g. Usage info) GGF-UR group declared that “… will not be a Grid usage record”

ARC-enabled Grid infrastructure(s)  Loosely coupled collection of Grid(s) based on ARC middleware –Multidisciplinary Grids/VOs: Swiss Bio, Finnish Material Science Grids –Very low operational costs!!!  Grid of National Grids –SWEGRID, DCGC, NorGrid, M-Grid, SwissGrid, NDGF, etc.. –National operational procedures  Easy to Join: assistance in Grid deployment outside the Nordic area  No central operational centre,no centrally forced VOs, large autonomy  NorduGrid Collaboration is just a provider of the middleware, it does not control resources or operation

Nordic DataGrid Facility NOS-N DK SFN S Nordic Data Grid Facility  NDGF == “Nordic Data Grid Facility” –Idea conceived in 2002 simultaneously with LCG –Goal: create a Nordic Grid infrastructure, primarily for LHC Grid computing (Tier1) – : pilot project funded by the 4 Nordic countries (Denmark, Finland, Norway, Sweden) –NorduGrid/ARC middleware chosen as the basis  June 1 st 2006: NDGF is launched –Nordic production Grid, leveraging national grid resources –Common framework for Nordic production Grid –Co-ordinates & hosts major Grid projects (e.g. the Nordic LHC Tier-1) –Develops Grid middleware (ARC contributor) –Single Point of Entry for collaboration, middleware development/deployment, e- Science projects –Represents the Nordic Grid community internationally  NDGF –Funded (2 MEUR/year) by National Research Councils of the Nordic countries (NOS-N)  NDGF coordinates activities - does not own resources or middleware

Summary  NorduGrid Collaboration develops, maintains & supports an open source Grid middleware –ARC middleware is approaching the milestone of the second stable release (v0.6) –Ambitious development plan after the 0.6 release to be carried out by newly launched projects (KnowARC, NDGF, NorduNet3)  ARC is a reliable, robust, easy-to-use Grid middleware, enabling distributed production facilities already for almost 3 years, non-stop. – ARC is a choice of middleware for numerous national and cross-national Grid projects.  The NorduGrid Community is involved in and devoted towards Grid interoperability and standardization efforts –Interoperability and standardization needs are among the major driving forces behind further ARC development

references  Advanced Resource Connector middleware for lightweight computational Grids, Future Generation Computer Systems (in press) –  Science Grid This Week article on KnowARC – ore.html  web sites: – – –