INFSO-RI-508833 Enabling Grids for E-sciencE www.eu-egee.org www.glite.org EGEE Middleware Claudio Grandi (INFN – Bologna) EGEE-NAREGI meeting March 20-22,

Slides:



Advertisements
Similar presentations
INFSO-RI Enabling Grids for E-sciencE EGEE and gLite Slides by: Erwin Laure EGEE Deputy Middleware Manager.
Advertisements

Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Middleware Claudio Grandi (INFN – Bologna) Workshop Commissione.
FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
Plateforme de Calcul pour les Sciences du Vivant SRB & gLite V. Breton.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Services Abderrahman El Kharrim
EGEE-II INFSO-RI Enabling Grids for E-sciencE Slides based on material from Sergio Andreozzi INFN-CNAF and from Pedro.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Based on material by Sergio Andreozzi INFN-CNAF OMII-Europe All-Hands.
FP6−2004−Infrastructures−6-SSA E-infrastructure shared between Europe and Latin America Giuseppe Andronico INFN Sezione di Catania.
IST E-infrastructure shared between Europe and Latin America Overview of gLite Middleware Pedro Henrique Rausch Bello Instituto.
Makrand Siddhabhatti Tata Institute of Fundamental Research Mumbai 17 Aug
INFSO-RI Enabling Grids for E-sciencE Comparison of LCG-2 and gLite Author E.Slabospitskaya Location IHEP.
INFSO-RI Enabling Grids for E-sciencE gLite Data Management Services - Overview Mike Mineter National e-Science Centre, Edinburgh.
FESR Consorzio COMETA Grid Introduction and gLite Overview Corso di formazione sul Calcolo Parallelo ad Alte Prestazioni (edizione.
OSG Middleware Roadmap Rob Gardner University of Chicago OSG / EGEE Operations Workshop CERN June 19-20, 2006.
INFSO-RI Enabling Grids for E-sciencE Logging and Bookkeeping and Job Provenance Services Ludek Matyska (CESNET) on behalf of the.
EGEE-III INFSO-RI Enabling Grids for E-sciencE The Medical Data Manager : the components Johan Montagnat, Romain Texier, Tristan.
INFSO-RI Enabling Grids for E-sciencE Workload Management System Mike Mineter
INFSO-RI Enabling Grids for E-sciencE Status and Plans of gLite Middleware Erwin Laure 4 th ARDA Workshop 7-8 March 2005.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Security and Job Management.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Using gLite API Vladimir Dimitrov IPP-BAS “gLite middleware Application Developers.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks JRA1 summary Claudio Grandi EGEE-II JRA1.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE middleware: gLite Data Management EGEE Tutorial 23rd APAN Meeting, Manila Jan.
Enabling Grids for E-sciencE Introduction Data Management Jan Just Keijser Nikhef Grid Tutorial, November 2008.
INFSO-RI Enabling Grids for E-sciencE The gLite Workload Management System Elisabetta Molinari (INFN-Milan) on behalf of the JRA1.
June 24-25, 2008 Regional Grid Training, University of Belgrade, Serbia Introduction to gLite gLite Basic Services Antun Balaž SCL, Institute of Physics.
EGEE-II INFSO-RI Enabling Grids for E-sciencE JRA1 in EGEE II Claudio Grandi (INFN and CERN) EGEE II Transition Meeting.
FP6−2004−Infrastructures−6-SSA E-infrastructure shared between Europe and Latin America gLite Overview Roberto Barbera Univ. of.
1 Andrea Sciabà CERN Critical Services and Monitoring - CMS Andrea Sciabà WLCG Service Reliability Workshop 26 – 30 November, 2007.
INFSO-RI Enabling Grids for E-sciencE EGEE Middleware reengineering Claudio Grandi – JRA1 Activity Manager - INFN EGEE Final EU.
Glite. Architecture Applications have access both to Higher-level Grid Services and to Foundation Grid Middleware Higher-Level Grid Services are supposed.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
INFSO-RI Enabling Grids for E-sciencE The gLite File Transfer Service: Middleware Lessons Learned form Service Challenges Paolo.
Enabling Grids for E-sciencE gLite for ATLAS Production Simone Campana, CERN/INFN ATLAS production meeting May 2, 2005.
INFSO-RI Enabling Grids for E-sciencE Introduction Data Management Ron Trompert SARA Grid Tutorial, September 2007.
FP6−2004−Infrastructures−6-SSA E-infrastructure shared between Europe and Latin America Alexandre Duarte CERN IT-GD-OPS UFCG LSD 1st EELA Grid School.
INFSO-RI Enabling Grids for E-sciencE gLite Middleware Status Frédéric Hemmer, JRA1 Manager, CERN On behalf of JRA1.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Grid2Win : gLite for Microsoft Windows Roberto.
INFSO-RI Enabling Grids for E-sciencE /10/20054th EGEE Conference - Pisa1 gLite Configuration and Deployment Models JRA1 Integration.
EGEE-II INFSO-RI Enabling Grids for E-sciencE gLite and Condor present and future Claudio Grandi (INFN – Bologna)
David Adams ATLAS ATLAS-ARDA strategy and priorities David Adams BNL October 21, 2004 ARDA Workshop.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Middleware reengineering Claudio Grandi (INFN – Bologna) EGEE Final.
INFSO-RI Enabling Grids for E-sciencE Policy management and fair share in gLite Andrea Guarise HPDC 2006 Paris June 19th, 2006.
Segundo Taller Latino Americano de Computación GRID – Primer Taller Latino Americano de EELA – Primer Tutorial Latino Americano de EELA
EGEE-II INFSO-RI Enabling Grids for E-sciencE Practical using WMProxy advanced job submission.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Data Management cluster summary David Smith JRA1 All Hands meeting, Catania, 7 March.
EGEE-II INFSO-RI Enabling Grids for E-sciencE middleware status and plans Claudio Grandi (INFN and CERN) John White.
13th EELA Tutorial, La Antigua, 18-19, October E-infrastructure shared between Europe and Latin America FP6−2004−Infrastructures−6-SSA
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
First South Africa Grid Training June 2008, Catania (Italy) OVERVIEW of the gLite COMPONENTS Marcello Iacono Manno FIRST.
OSG Status and Rob Gardner University of Chicago US ATLAS Tier2 Meeting Harvard University, August 17-18, 2006.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Job Management Claudio Grandi.
INFSO-RI Enabling Grids for E-sciencE gLite Overview Riccardo Bruno, Salvatore Scifo gLite - Tutorial Catania, dd.mm.yyyy.
INFSO-RI Enabling Grids for E-sciencE Padova site report Massimo Sgaravatto On behalf of the JRA1 IT-CZ Padova group.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI Services for Distributed e-Infrastructure Access Tiziana Ferrari on behalf.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Enabling Grids for E-sciencE Work Load Management & Simple Job Submission Practical Shu-Ting Liao APROC, ASGC EGEE Tutorial.
Enabling Grids for E-sciencE Claudio Cherubino INFN DGAS (Distributed Grid Accounting System)
EGEE Data Management Services
StoRM: a SRM solution for disk based storage systems
GDB 8th March 2006 Flavia Donno IT/GD, CERN
Comparison of LCG-2 and gLite v1.0
gLite Middleware Status
Slides contributed by EGEE Team
Current status of gLite
Short update on the latest gLite status
Data Management cluster summary
Overview of gLite Middleware
Based on material by Sergio Andreozzi INFN-CNAF
Presentation transcript:

INFSO-RI Enabling Grids for E-sciencE EGEE Middleware Claudio Grandi (INFN – Bologna) EGEE-NAREGI meeting March 20-22, 2006 CERN, Switzerland

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March Outline Processes and Releases Subsystems –Information System and Monitoring –Security Infrastructure –Workload Management –Data Management Summary

INFSO-RI Enabling Grids for E-sciencE Process and Releases

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March gLite Software Process DevelopmentIntegrationTesting Software Code Deployment Packages Integration Tests FailPass Fix Functional Tests Testbed Deployment Fail Pass Installation Guide, Release Notes, etc

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March gLite Releases gLite Special Release for SC File Transfer Service gLite Service Release ~60 Defect Fixes gLite Special Release for SC File Transfer Service April 2005 May 2005 July 2005 Aug 2005 Sep 2005 Oct 2005 Nov 2005 Dec 2005 Jan 2006 Feb 2006 June 2005 gLite 1.0 Condor-C CE gLite I/O R-GMA WMS L&B VOMS Single Catalog gLite 1.1 File Transfer Service Metadata catalog Functionality gLite 1.2 File Transfer Agents Secure Condor-C gLite 1.3 File Placement Service FTS multi-VO Refactored RGMA & CE gLite 1.4 VOMS for Oracle SRMcp for FTS WMproxy LBproxy DGAS QF1.0.12_01_2005 QF1.0.12_02_2005 QF1.0.12_03_2005 QF1.0.12_04_2005 QF1.1.0_05_2005 QF1.1.0_06_2005 QF1.1.0_07_2005 QF1.1.0_08_2005 QF1.1.0_09_2005 QF1.1.0_10_2005 QF1.1.2_11_2005 QF1.1.2_12_2005 QF1.1.2_13_2005 QF1.1.2_16_2005 QF1.2.0_14_2005 QF1.2.0_15_2005 QF1.3.0_17_2005 QF1.3.0_18_2005 QF1.3.0_19_2005 QF1.3.0_20_2005 QF1.3.0_22_2005 QF1.3.0_21_2005 QF1.3.0_23_2005 QF1.3.0_24_2005 QF1.4.1__25_2005 QF1.4.1__28_2005 QF1.4.1__26_2005 QF1.4.1__29_2005 QF1.4.1__30_2005 QF1.4.1__27_2005 gLite 1.5 Shallow resub. in WMS GPBOX gLite I/O encryption Amga Hydra QF1.4.1__31_2005 QF1.4.1__32_2005 QF1.4.1__33_2005 QF1.4.1__34_2005

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March After gLite Converge from LCG and gLite to a single middleware stack called gLite. The first version will be gLite –Process controlled by the Technical Coordination Group –gLite and LCG have been the last independent releases Components in gLite –Certified:  All components already in LCG plus upgrades this already includes new versions of VOMS, R-GMA and FTS  The Workload Management System (with LB, CE, UI) of gLite –Tested to some degree and with limited deployment support:  The DGAS accounting system  Data management tools as needed by the Biomed community Hydra, AMGA, secure access to data

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March Schedule for gLite After gLite 3.0.0: –March 31 st : code freeze for development release gLite –April 30 th : end of integration –May 31 st : end of certification. Deployment on PPS –July 31 st : release of production version gLite Start deployment at sites –September : gLite installed at sites and usable. PRODUCTION!! Deploy in prod n PPS Certification Tuesday 28/2/06 gLite 3.0.0β exits certification and enters pre-production FebruaryAprilMayJuneMarch Friday 28/4/06 PPS phase ends. gLite passes from PPS to Production. Wednesday 15/3/06 gLite 3.0.0β available to users in the PPS Deployment of gLite 3.0.0β in PPS Continual bug fixing and patches passed to PPS Thursday 1/6/06 LCG Service Challenge 4 (SC4) starts!!

INFSO-RI Enabling Grids for E-sciencE Security Infrastructure

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March Authentication Based on trusted third parties (TTPs) –Regional Certificate Authorities (CA) X.509 PKI infrastructure –Grid Security Infrastructure (from Globus 2.4.3, as in VDT-1.2.2) Federations of CAs exists –Europe EUGridPMA, Asia-Pacific Region APGridPMA, Americas TAGPMA –International Grid Trust Federation (IGTF) established in October 2005 Short-Lived Credential Services (SLCS)  proxies Site-integrated credential services (SICS) –issue short-lived credentials to its local users (e.g. Kerberos at FNAL) Proxy store: MyProxy (Version 1.14) from VDT –Allows clean proxy renewal –working with VDT to provide VOMS-aware access to MyProxy 

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March Authorization VO-specific attributes: VOMS –VOMS issues Attribute Certificates that are attached to proxies and provide users with additional capabilities –are the base for Authorization process Web services for delegation –using portType (Java, for axis) and GridSite (C for Apache) –need to define a standard for interoperation Authorization framework –gLite Authorization Framework (compatible with XACML policies)  mainly used for Java-based applications –LCAS/LCMAPS  mainly used for C-based applications (e.g. GT2-GRAM, GridFTP server) –support for VOMS, blacklists, gridmap files 

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March VO Policy management: GPBOX GPBOX: Distributed VO policy management tool –Interface to define fine-grained VO policies  based on VOMS groups and roles –Store policies and propagate to sites  enforcement of policies done at sites – sites may accept/reject policies –May be interfaced to dynamic sources of information  e.g. an accounting system to provide fair share –Standards Compliant (RBAC, XACML, GSI)PBox Policies Group A : high and low priority CEs Group B : low priority CEs CE HIGH CE LOW RB VOMS server Group A Group B

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March ! Accounting DGAS: accumulates Grid accounting information –User, JobId, user VO, VOMS FQAN(role,capabilities), SI2K, SF2K, system usage (cpuTime, wallTime…),… –allows billing and scheduling policies –levels of granularity: from single jobs to VO or grid aggregations –Privacy: only the user or VO manager can access information  site managers can keep accounting information available just for site internal analysis –Sites can substitute DGAS metering system with their own Limited support in gLite Usage accounting Account balancing, resource pricing, (billing) usage records accounting data Usage Analysis Usage Metering Interface

INFSO-RI Enabling Grids for E-sciencE Information System and Monitoring

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March R-GMA The Relational Grid Monitoring Architecture (R-GMA) provides a uniform method to access and publish both information and monitoring data. From a user's perspective, an R-GMA installation currently appears similar to a single relational database. Relational implementation of the GGF’s Grid Monitoring Architecture (GMA) Producer Service Registry Service Consumer Service API Schema Service Consumer application Producer application Publish Tuples Send Query Receive Tuples Register Locate Query Tuples SQL “CREATE TABLE” SQL “INSERT” SQL “SELECT” 

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March Service discovery the gLite Service Discovery provides a standard set of methods for locating Grid services hides underlying information system plug-ins for R-GMA, BDII and XML files (others could be developed if required) API available for Java and C/C++ command line version also available Used by WMS and Data Management clients Production Services still using BDII as the Information System 

INFSO-RI Enabling Grids for E-sciencE Workload Management

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March Computing Element Service representing a computing resource Condor-C GSI enabled –Based on Condor , will migrate to in next release Uses CEMon to publish information –support for R-GMA and bdII –GLUE 1.1 (migration to GLUE 1.2 for next release) Batch Local ASCII Helper (BLAH) –Interface between the CE and the Local Resource ManagerSsystem –submit, cancel and query –Support for hold and resume  To be used to put a job on hold, waiting for e.g. the staging of the input data 

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March ICE-CREAM CREAM –web service Computing Element –Cream WSDL allows defining custom user interface –C++ CLI interface allows direct submission Lightweight Fast notification of job status changes –via CEMon Improved security –no “fork-scheduler” Will support for bulk jobs on the CE –optimization of staging of input sandboxes for jobs with shared files ICE (Interface to Cream Environment) –being integrated in WMS for submissions to CREAM

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March CEMon Web service to publish status of a computing resource to clients –WMS or individual clients Synchronous queries or asynchronous notification of events Clients subscribe and are notified according to user defined policies  on job status  on CE characteristics and status –May be used to pull jobs Included in VDT and used in OSG for resource selection In gLite will be available but the baseline is that the WMS queries the bdII 

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March Job information Logging and Bookkeeping service –Tracks jobs during their lifetime (in terms of events) –L&B Proxy provides faster, synchronous and more efficient access to L&B services to Workload Management Services –Support for “CE reputability ranking“  Maintains recent statistics of job failures at CE’s  Feeds back to WMS to aid planning –Working on inclusion of L&B in the VDT Job Provenance –Long term job information storage –Useful for debugging, post-mortem analysis, comparison of job executions in different environments –Useful for statistical analysis 

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March Workload Management System Helps the user accessing computing resources –resource brokering –management of input and output –management of complex workflows Job specification in JDL (based on Condor classad 0.9.6) –Building a JDL ↔ JSDL translator Backward compatible with LCG-2 Support for shallow resubmission –Resubmission happens in case of failure only when the job didn't start Support for MPI job even if the file system is not shared between CE and Worker Nodes (WN) Support collection of information from many sources –CEMon, bdII, R-GMA Support for Data management interfaces (DLI and StorageIndex) Support for execution of all DAG nodes within a single CE –chosen by user or by the WMS matchmaker Support for file peeking during job execution (Job File Perusal) Initial support for pilot job –prepare the execution environment, then get and execute the user job 

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March WMProxy WMProxy is a SOAP Web service providing access to the Workload Management System (WMS) Job characteristics specified via JDL –jobRegister  create id  map to local user and create job dir  register to L&B  return id to user –input files transfer  GridFTP 1.12 –jobStart  register sub-jobs to L&B  map to local user and create sub-job dir’s  unpack sub-job files  deliver jobs to WM 

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March Complex Workflows Direct Acyclic Graph (DAG) is a set of jobs where the input, output, or execution of one or more jobs depends on one or more other jobs A Collection is a group of jobs with no dependencies –basically a collection of JDL’s A Parametric job is a job having one or more attributes in the JDL that vary their values according to parameters Using compound jobs it is possible to have one shot submission of a (possibly very large, up to thousands) group of jobs –Submission time reduction  Single call to WMProxy server  Single Authentication and Authorization process  Sharing of files between jobs –Availability of both a single Job Id to manage the group as a whole and an Id for each single job in the group nodeE nodeC nodeA nodeD nodeB 

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March WMProxy Shared Input sandbox –Useful when submitting compound jobs –When sub-jobs input sandboxes contain instances of the same file, these are transferred only once and made available by WMProxy to all involved sub-jobs  lower data size to transfer  minimize number of calls to file transfer service Asynchronous job start will be available in next release –Upon call to the jobStart operation, WMProxy can complete processing of the request in background –Control is returned to the client soon after the request has been accepted by WMProxy and the corresponding event has been logged to LB –All time-consuming actions needed to complete the processing of the request are performed “behind the scene” by the service  From user point of view makes submission time (almost) independent from the number of jobs 

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March gLite WMS tests Submission slower with gLite NS but after that all jobs are scheduled much faster But using the WMProxy prompt back after a few seconds (almost independent of number of job!) Submission test with Network Server –10 threads of 200 jobs –multithread active on WMS LCG-2 gLite w.r.t job submission time All jobs scheduled Prompt back w.r.t. submission of job #1 LCG-2 gLite 3.0.0

INFSO-RI Enabling Grids for E-sciencE Data Management

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March Storage Element A disk or tape storage resource Common interface: SRMv1 –negotiable transfer protocols (Gridftp, gsidcap, RFIO, …) –work in progress to migrate to SRMv2 –Various implementation from LCG and other external projects  disk-based: DPM, dCache  tape-based: Castor, dCache –Support for ACLs in DPM, in future in Castor and dCache Posix-like file access: –Grid File Access Layer (GFAL) by LCG. Provides also:  Abstractions: Storage Element, File Catalog, Information System  Support for ACL in the SRM layer –gLite I/O  Support for file ACL via the Fireman catalog  Interfaced to SRM Storage Elements (Castor, dCache and DPM)  Configuration using the common Service Discovery interfaces  Limited support in gLite (replaced by GFAL)   !

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March The Fireman catalog Resolves logical file names (LFN) to physical location of files (URL understood by SRM) and storage elements Oracle and MySQL versions available Secure services, using VOMS groups, ACL support CLI and simple API for C/C++ wrapping a lot of the complexity for easy usage Attribute support Symbolic link support Exposing interfaces suitable for matchmaking (StorageIndex and DLI) Limited support in gLite (replaced by LFC) !

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March The AMGA Metadata Catalog Joint JRA1-NA4 (ARDA) development Authentication based on Password, X509 Cert, Grid Proxy Posix-ACLs and Unix permissions for entries and collections –not yet on individual attributes Built-in group-management like AFS or via VOMS SOAP and Text front-ends Streamed Bulk Operations –vital for WAN operations Scales well up to 100 concurrent clients –back-end limit Support for replication –also via LCG-3D for Oracle backend only Limited support in gLite !

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March Hydra Hydra keystore –store keys for data encryption –3 instances: at least 2 need to be available for decryption –Limited support in gLite Demonstrated with the SRM- DICOM demo at EGEE Pisa conference (Oct’05) !

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March File Transfer Service Reliable, scalable and customizable file transfer reliable: –transfer requests are jobs  in case of failure a transfer may be retried  subject to generic and VO-specific policies scalable: –more files in a single transfer request –File Transfer queues allow inter-VO sharing customizable: –VO-specific File Transfer Agents 

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March The FTS key concept: Channel Logical unit of management –Represent a mono- directional network pipe between two sites Independently manageable –State –Number of streams –Number of concurrent transfers Production Channels –Dedicated Tier0-Tier1, Tier1-Tier1 or Tier1-Tier2 transfers  efficient bulk distribution Non-production Channels –Open networks shared with other applications. 

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March FTS architecture Client with SOAP API and CLI –Roles  Administrator of the service  Channel Manager  VO Manager  Submitter User  Regular User  Vetoed User Web Service Database –Oracle and MySQL Agents act on database conditions 

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March FTS at work Service Challenge 3 Rerun (January 2006) –All sites achieved target rate –8/11 sites achieved nominal rate Service Challenge 4 (start June 2006) –increase stability, sustainability, robustness –support for SRMv2 –integration with experiment frameworks 

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March Summary Complete middleware stack –security infrastructure –information system and monitoring –workload management –data management Developed according to a well defined process Controlled by the EGEE Technical Coordination Group gLite will be available on the production infrastructure in less than 2 months Development is continuing to provide increased robustness, usability and functionality

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March

Enabling Grids for E-sciencE INFSO-RI Claudio Grandi - EGEE-NAREGI Interoperability Meeting - CERN March Security Architecture