Download presentation
Presentation is loading. Please wait.
Published byDaniela Griffin Modified over 8 years ago
1
INFSO-RI-508833 Enabling Grids for E-sciencE www.eu-egee.org Overview Dr. Marcel Kunze Institut für Wissenschaftliches Rechnen Forschungszentrum Karlsruhe Federation Meeting at GSI, 7.3.2005
2
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 2 Build a large-scale production grid service to: Support science and technology worldwide Link with and build on national, regional and international initiatives Foster international cooperation both in the creation and the use of the e- infrastructure Why EGEE? Network Operations, Support and training Collaborations Grid infrastructure
3
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 3 The Federations 70 leading institutions in 27 countries, federated in regional Grids Germany: DESY, DKRZ, FhG, FZK, GSI Switzerland: CSCS 32 M Euros EU funding (2004-5), O(100 M) total budget Aiming for a combined capacity of over 20’000 CPUs (one of the largest international Grid infrastructures ever assembled) ~ 300 dedicated staff
4
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 4 EGEE Activities Emphasis on operating a production grid and supporting the end-users 48 % service activities (Grid Operations, Support and Management, Network Resource Provision) SA1-SA2 24 % middleware re-engineering (Quality Assurance, Security, Network Services Development) JRA1-JRA4 28 % networking (Management, Dissemination and Outreach, User Training and Education, Application Identification and Support, Policy and International Cooperation) NA1-NA5
5
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 5 Country providing resources Country anticipating joining In LCG-2: 113 sites, 30 countries >10,000 cpu ~5 PB storage Includes non-EGEE sites: 9 countries 18 sites Computing Resources: Feb 2005
6
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 6 Grid Operations and Support Core Infrastructure Centres –operate core services (VO specific and general Grid services) –develop new management tools –provide support to the Regional Operations Centres Regional Operations Centres –Testing (certification) of new middleware on a variety of platforms before deployment –Deployment of middleware releases + coordination + distribution inside the region –integration of ‘Local’ VO –Development of procedures and capabilities to operate the resources Global Grid User Support (GGUS) –Integration of regional support centres
7
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 7 GGUS Web Portal FZK (See T. Antoni‘s Talk)
8
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 8 Grid Operations The grid is flat, but Hierarchy of responsibility –Essential to scale the operation CICs act as a single Operations Centre –Operational oversight (grid operator) responsibility –rotates weekly between CICs –Report problems to ROC/RC –ROC is responsible for ensuring problem is resolved –ROC oversees regional RCs ROCs responsible for organising the operations in a region –Coordinate deployment of middleware, etc CERN runs operations monitoring center (OMC) and coordinates sites not associated with a ROC CIC RC ROC RC ROC RC ROC RC ROC OMC RC = Resource Centre DESY, FhG, FZK, GSI (See H.Marten‘s Talk)
9
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 9 Deployment of applications PilotNew Pilot applications –High Energy Physics –Biomed applications Generic applications – Deployment under way –Computational Chemistry –Earth science research –EGEODE: first industrial application –Astrophysics (MAGIC) With interest from –Hydrology –Seismology –Grid search engines –Stock market simulators –Digital video etc. DKRZ, FhG, FZK
10
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 10 Outreach & Training Public and technical websites constantly evolving to expand information available and keep it up to date 2 conferences organised –~ 300 @ Cork, ~ 400 @ Den Haag Athens 3rd project conference 18-22 April ’05 Pisa 4th project conference 24-28 October ’05 More than 40 training events (including the GGF grid school) across many countries –~600 people trained induction; application developer; advanced; retreats –Material archive coming online with ~80 presentations Strong links with GILDA testbed and GENIUS portal developed in EU DataGrid FZK (See R.Berlich‘s Talk)
11
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 11 Future EGEE Middleware - gLite Application requirements http://egee-na4.ct.infn.it/requirements/http://egee-na4.ct.infn.it/requirements/ Intended to replace present middleware with production quality services Developed from existing components Aims to address present shortcomings and advanced needs from applications Prototyping short development cycles for fast user feedback Initial web-services based prototypes being tested Globus 2 basedWeb services based gLite-2gLite-1LCG-2LCG-1
12
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 12 Co-existence with deployed infrastructure –Co-existence with LCG-2 and OSG (US) are essential for the EGEE Grid services Site autonomy –Reduce dependence on ‘global, central’ services Open source license Guiding Principles Service oriented approach –Allow for multiple interoperable implementations Lightweight (existing) services –Easily and quickly deployable –Use existing services where possible Condor, EDG, Globus, LCG, … Portable –Being built on Scientific Linux and Windows Security –Sites and Applications Performance/Scalability & Resilience/Fault Tolerance –Comparable to deployed infrastructure EDGVDT... LCG...AliEn
13
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 13 gLite Services and Responsible Clusters Grid Access Service API Access Services Job Provenance Job Management Services Computing Element Workload Management Package Manager Metadata Catalog Data Services Storage Element Data Management File & Replica Catalog Authorization Security Services Authentication Auditing Information & Monitoring Information & Monitoring Services Application Monitoring Site Proxy Accounting JRA3 UK CERN IT/CZ
14
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 14 gLite Services for Release 1 Software stack and origin (simplified) Computing Element –Gatekeeper (Globus) –Condor-C (Condor) –CE Monitor (EGEE) –Local batch system (PBS, LSF, Condor) Workload Management –WMS (EDG) –Logging and bookkeeping (EDG) –Condor-C (Condor) Storage Element –File Transfer/Placement (EGEE) –glite-I/O (AliEn) –GridFTP (Globus) –SRM: Castor (CERN), dCache (FNAL, DESY), other SRMs Catalog –File and Replica Catalog (EGEE) –Metadata Catalog (EGEE) Information and Monitoring –R-GMA (EDG) Security –VOMS (DataTAG, EDG) –GSI (Globus) –Authentication for C and Java based (web) services (EDG)
15
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 15 Job Management Services Efficient and reliable scheduling of computational tasks on the available infrastructure Started with LCG-2 Workload Management System (WMS) –Inherited from EDG –Support partitioned jobs and jobs with dependencies –Support for different replica catalogs for data based scheduling –Modification of internal structure of WMS Task queue: queue of pending submission requests Information supermarket: repository of information on resources Better reliability, better performance, better interoperability, support push and pull mode –Under development Web Services interface supporting bulk submission (after V1.0) Bulk submission supported now by use of DAGs Distributed superscheduling (interaction among multiple WMSs)
16
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 16 Job Management Services II Computing Element (CE) –Service representing a computing resource CE moving towards a VO based local scheduler Incorporates new technologies provided by Condor and Globus –Web service interfaces job management requests (run, cancel, suspend, resume, …) Still under development Policy based notifications on changes of the CE (pull model) Job Provenance –Keeps track of definition of submitted jobs, execution conditions and job life cycle for a long time under development Grid Accounting (DGAS) –Accumulates Grid accounting information about the usage of Grid resources by users / groups (e.g. VOs) for billing and scheduling policies. Under development VOMS –Virtual Organization Membership Service Advanced Reservation service under development Assured backward compatibility (to facilitate migration from current SA1 infrastructure)
17
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 17 Efficient and reliable data storage, movement, and retrieval on the infrastructure Storage Element –Reliable file storage (SRM based storage systems) –Posix-like file access (gLite I/O) –Transfer (gridFTP) File and Replica Catalog –Resolves logical filenames (LFN) to physical location of files (URL understood by SRM) and storage elements –Hierarchical File system like view in LFN space –Single catalog or distributed catalog (under development) deployment possibilities File Transfer and Placement Service –Reliable file transfer and transactional interactions with catalogs Data Scheduler –Scheduled data transfer in the same spirit as jobs are being scheduled taking into account e.g. network characteristics (collaboration with JRA4) –Under development Metadata Catalog –Limited metadata can be attached to the File and Replica Catalog –Interface to application specific catalogs have been defined Data Management Services SRM I/O GridFTP Storage Element TransferAgent FPS Data Scheduler FPS VOs Site boundary LocalCat Catalog MOM Catalog VOs LocalCat Catalog MOM LocalCat Catalog MOM
18
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 18 Information and Monitoring Services Efficient and reliable provision of Grid information and Grid and Application monitoring data R-GMA (Relational Grid Monitoring Architecture) –Implements GGF GMA standard –Development started in EDG, deployed on the production infrastructure for accounting Producer Service Registry Service Consumer Service API Mediator Schema Service Consumer application Producer application Publish Tuples Send Query Receive Tuples Register Locate Query Tuples SQL “CREATE TABLE” SQL “INSERT” SQL “SELECT”
19
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 19 R-GMA Producer, Consumer, Registry and Schema services with supporting tools –Registry replication –Simpler API – matching the next (WS) release Provides smooth transition between old API and WS –coping with life on the Grid: poorly configured networks, firewalls, MySQL corruptions etc Generic Service Discovery API Under development –Web Service version –File (as well as memory and RDBMS) based Producers –Native python interface –Fine grained authorization –Schema replication
20
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 20 Other Reengineering Activities Prototypes of Grid Access Service and Package Manager implemented in the AliEn framework Grid Access Service –Acts on user’s behalf –Discovers and manages Grid services for the user Package Manager –Provides dynamically distribution of application software needed –Does not install Grid middleware
21
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 21 Release Timeline May 2004 Prototype Available to ARDA users Builds June 2004 July 2004 Aug 2004 Sep 2004 Oct 2004 Nov 2004 Dec 2004 Jan 2005 Feb 2005 March 2005 First public nightly build (B 39) Functionality First automated build (B 1) First Integration build (I20041020 B 80) RC1 (I20041217 B 151) I/O Client I/O Server CE, LB, WMS, WN AliEn, R-GMA Data Local Transfer Service, Single Catalog VOMS, Site Configuration RC1 (I20050204 B 206) Today Release 1.0 April 2005
22
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 22 Weekly Release Process DevelopmentIntegrationTesting Software Code Deployment Packages Integration Tests FailPass Fix Functional Tests Testbed Deployment Fail Pass Installation Guide, Release Notes, etc
23
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 23 Total Physical Source Lines of Code (SLOC) SLOC = 583,620 (as of 7 February 2005) Total SLOC by language (dominant language first) C++180816 (30.98%) Java170749 (29.26%) Ansi C134264 (23.01%) Perl 60972 (10.45%) Python 20039 ( 3.43%) sh 11859 ( 2.03%) Yacc 3635 ( 0.62%) jsp 640 ( 0.11%) Lex 335 ( 0.06%) csh 217 ( 0.04%) SLOC Software provided by different sources Diversity increases the complexity
24
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 24 gLite Interoperability Interoperability with other Grids mainly needed at resource level –Same physical resource should be exploitable in different Grids Approach –Reduce requirements on sites CE: globus gatekeeper SE: SRM –Close connection with other projects OSG Use EGEE architecture and design documents as basis for their blueprint Common members in design teams
25
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 25 Applications (I): Grid monitoring –GIIS Monitor + Monitor Graphs –Sites Functional Tests –GOC Data Base –Scheduled Downtimes –Live Job Monitor –GridIce – VO + fabric view –Certificate Lifetime Monitor Operation of Production Service: real-time display of grid operations Accounting information Selection of Monitoring tools:
26
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 26 Applications (II): gPTM3D Goal: Enable PTM3D for the Grid –PTM3D (Poste de Travail Médical 3D) is an interactive radiological image visualization and processing tool developed at LIMSI (CNRS) with clinical usage –Step1: interactive response time for CPU-intensive volume reconstruction Interaction RenderExploreAnalyseInterpretAcquire
27
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 27 Applications (III): GEMS Motivation: –Modern computer simulations of biomolecular systems produce an abundance of data, which could be reused several times by different researchers. data must be catalogued and searchable. GEMS (Grid enabled Molecular Simulations) database and toolkit: –autonomous storage resources –metadata specification –automatic storage allocation and replication policies –interface for distributed computation EGEE Grid GEMS Application Deployment Server Resource Broker Computing Element MPI GEMS program GEMS program GEMS program GEMS programs Working nodes License Server Outbound connectivity
28
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 28 Applications (IV): EGEODE Seismic processing Generic Platform: –Based on Geocluster, an industrial application –Include several standard tools for signal processing, simulation and inversion (model optimization) Open: any user can write new algorithms in new modules (shared or not) Free for academic research Controlled by license keys (opportunity to explore license issue at a grid level) Initial partners F, CH, NL, Russia
29
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 29 Future EGEE has „hit the ground running“ FZK is leading the Swiss/German Federation –Grid Operations, Training, Dissemination Acquiring and porting/developing Grid Applications in further scientific disciplines to make use of the infrastructure (EGAAP Panel) Currently preparing for follow-on project EGEE II –Call for Research Infrastructures (May-September)
30
Enabling Grids for E-sciencE INFSO-RI-508833 EGEE Federation Meeting, GSI, 7.3.2005 30 Contacts EGEE Website http://www.eu-egee.org How to join http://public.eu-egee.org/join/ EGEE Project Office project-eu-egee-po@cern.ch
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.