Download presentation
Presentation is loading. Please wait.
Published byJasper Lyons Modified over 9 years ago
1
DataTag is a project funded by the European UnionCHEP 2003 – 24-28 March 2003 - The WorldGRID transatlantic testbed A successful example of Grid interoperability across EU and US domains Flavia Donno (Formerly of DataTAG WP4, LCG) Flavia.Donno@cern.ch http://chep03.ucsd.edu/files/249.ppt CHEP 2003 – 24-28 March – n o. (1)
2
DataTag is a project funded by the European Union CHEP 2003 – 24-28 March 2003 – n o 2 Talk Outline Motivation Participants Interoperability issues Solutions Architecture Monitoring/Support Spin off F. Donno CERN/IT and INFN R. Gardner University of Chicago Applications CMS ATLAS Monitoring with Nagios Monitoring with Ganglia Conclusions Next Steps
3
DataTag is a project funded by the European Union CHEP 2003 – 24-28 March 2003 – n o 3 Participants DataTAG: F. Donno, CERN - INFN S. Andreozzi, R. Barbera, V. Ciaschini, S. Fantinel, A. Ghiselli, M. Mazzucato, D. Rebatto, G. Tortone, L. Vaccarossa, M. Verlato, C. Vistoli – INFN M. Draoli, CNR-Rome CrossGrid: A. Garcia, M. Hardt, FZK - Germany J. Marco, UC - Spain M.David, J. Gomes, LIP - Portugal O. Maroney, U.Bristol, UK Trillium/iVDGL: P. Avery, J. Rodriguez - U. Florida E. Deelman, N. Olomu - USC/ISI J. Gieraltowski, S. Gose, E. May, J. Schopf – Argonne Afaq, J. Annis, R. Glossum, R. Pordes, V. Sekrhi – Fermilab W. Deng, J. Smith, D. Yu - BNL A. DeSmit, A. Roy - Wisconsin C. Dumitrescu, I. Foster, R. Gardner, U. Chicago L. Grundhoefer,J. Hicks, F. Luehring, L. Meehan - U. Indiana S. Youssef, Boston University B. Moe - Milwaukee D. Olson – LBNL S. Singh - Caltech
4
DataTag is a project funded by the European Union CHEP 2003 – 24-28 March 2003 – n o 4 Motivations Basic collaboration between European and US Grid projects Interoperability between Grid domains for applications submitted by users from different virtual organizations Controlled use of shared resources subject to agreed policy Integrated use of heterogeneous resources from iVDGL and DataGrid/CrossGrid testbed domains Goal: Build a “transatlantic grid” based on the existent European and American Grids with the goal of offering transparent access to the distributed computing infrastructure necessary to the “data-intensive” modern applications
5
DataTag is a project funded by the European Union CHEP 2003 – 24-28 March 2003 – n o 5 Interoperability Issues several OS Many grids with several OS (RH 6.2, RH 7.x, Fermi Linux, CERN Linux,…), several compilers and software components. Grid Architectures Different Grid Architectures (VDT server/client vs. Computing Elements, Storage Elements, User Interfaces, …) minimum set of core services optional services Need to identify minimum set of core services and define collective/optional services Common protocols/Same or compatible versions of the software Authentication and Authorization Authentication and Authorization mechanism: authority trusting, user authentication/authorization via LDAP VO Servers. Grid resource description Grid resource description/status: Globus schema vs. EDG schema vs. GLUE schema Grid Data management Several Grid Data management Tools Software distribution and configuration Software distribution and configuration : rpm based vs. PACMAN
6
DataTag is a project funded by the European Union CHEP 2003 – 24-28 March 2003 – n o 6 Solutions several OS Many grids with several OS (RH 6.2, RH 7.x, Fermi Linux, CERN Linux,…), several compilers and software components. Partition WorldGrid in subdomains with uniform or compatible set of basic services. Such resources will advertise themselves with specific targets to the applications (such as RH6.2). Try to keep the subdomains as large as possible.
7
DataTag is a project funded by the European Union CHEP 2003 – 24-28 March 2003 – n o 7 Solutions Grid Architectures Different Grid Architectures (VDT server/client vs. Computing Elements, Storage Elements, User Interfaces, …)CE UI SE RB VDT Client VDT Server RC IS RC IS
8
DataTag is a project funded by the European Union CHEP 2003 – 24-28 March 2003 – n o 8 Solutions minimum set of core services optional services Need to identify minimum set of core services and define collective/optional services Common protocols/Same or compatible versions of the software Globus and Condor core services (GRAM, GSI, MDS, GridFTP, …) Resource Broker, User Interface and JDL, Data Management high level tools (edg-replica- manager, MAGDA, Globus Replica Catalog, …) collective optional services not installed universally User Grid Portals (Genius, GRAPPA, …): a variety available not to change the User Interface to the GRID
9
DataTag is a project funded by the European Union CHEP 2003 – 24-28 March 2003 – n o 9 Solutions Authentication and Authorization Authentication and Authorization mechanism: authority trusting, user authentication/authorization via LDAP VO Servers. DOE and EDG certificates universally accepted DataTAG and iVDGL VO LDAP servers trusted mkgridmap tool universally installed Local security policy sites agreed to allow access to grid demonstration users (kerberos, …)
10
DataTag is a project funded by the European Union CHEP 2003 – 24-28 March 2003 – n o 10 Solutions Grid resource description Grid resource description/status: Globus schema vs. EDG schema vs. GLUE schema three coexistent schemas in place (Globus, EDG, GLUE) installed on all resources Some tool (monitoring) working with all of them EDG middleware using both EDG and GLUE US tools using none or Globus
11
DataTag is a project funded by the European Union CHEP 2003 – 24-28 March 2003 – n o 11 Solutions Software distribution and configuration Software distribution and configuration : rpm based vs. PACMAN Created WorldGrid distribution (rpm/LCFGng and PACMAN) Effort to ensure coherency and automatic configuration
12
DataTag is a project funded by the European Union CHEP 2003 – 24-28 March 2003 – n o 12 Final Architecture CE UI SE RB VDT Client VDT Server RC IS
13
DataTag is a project funded by the European Union CHEP 2003 – 24-28 March 2003 – n o 13 Monitoring and Support Two monitoring tools VO based in place: edt-monitor based on Nagios and iVDGL based on Ganglia (see talk from R. Gardner) Support infrastructure: Support infrastructure: to support site administrators during the installation and configuration procedure. Also for problem fixing during normal operation to support site administrators during the installation and configuration procedure. Also for problem fixing during normal operation
14
DataTag is a project funded by the European Union CHEP 2003 – 24-28 March 2003 – n o 14 Spin-off GLUE schema: WorldGrid has allowed to prove the validity of the GLUE schema and encouraged EDG to deploy it VOMS: The authentication/authorization problems were identified and parallel research activities started, like the one on Virtual Organization Manager Service VOMS: The authentication/authorization problems were identified and parallel research activities started, like the one on Virtual Organization Manager Service GLUE Packaging: A working group is trying to find a solution for a standardization of the packaging, distribution and configuration problem for a software release GLUE Packaging: A working group is trying to find a solution for a standardization of the packaging, distribution and configuration problem for a software release GLUE Testing: The problem of verifying an installation and validate a site for joining the Grid has been addressed and a working group has started GLUE Testing: The problem of verifying an installation and validate a site for joining the Grid has been addressed and a working group has started Support: A first operation/monitoring center has started in US taking advantage of the monitoring tools. Other centers in EU Support: A first operation/monitoring center has started in US taking advantage of the monitoring tools. Other centers in EU LCG-0: After the demonstration at IST2002 and SC2002, LCG has based his first middleware distribution on the WorldGrid experience LCG-0: After the demonstration at IST2002 and SC2002, LCG has based his first middleware distribution on the WorldGrid experience
15
DataTag is a project funded by the European UnionCHEP 2003 – 24-28 March 2003 - Rob Gardner University of Chicago on behalf of the WG group The WorldGRID transatlantic testbed, Part 2 A successful example of Grid interoperability across EU and US domains
16
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 16 16 Talk Outline l Motivation l Participants l Interoperability issues l Solutions l Architecture l Monitoring/Support l Spin off F. Donno CERN/IT and INFN Part 2 R. Gardner University of Chicago Applications CMS ATLAS Monitoring with Nagios Monitoring with Ganglia Conclusions Next Steps
17
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 17 17 Installing Apps on 2 Grids l We needed a way to get applications from three experiments (VO’s) setup on the execution sites l On DataTAG resources, selected CE’s were loaded with CMS or ATLAS rpms l On iVDGL resources, we Pacmanized binaries (rpms and tarballs) of bundled applications –%pacman –get iVDGL:ScienceGrid >Atlas-kit, Atlas-ATLFAST >CMS-MOP, EDG-CMS >SDSS Astrotools –binaries, and run time environments 3 experiments
18
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 18 18 ATLSIM Job Write to Grid Storage Element Read from Grid Storage Element Input Data Output ZEBRA Grid Storage Apache EnginFrame GENIUS https+java/xml+rfb WEB Browser ED G UI Loc al WS the Grid EDG+GSI ATLAS and CMS with GENIUS see R. Barbera’s Genius talk this conference
19
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 19 19 RB/JSS II SE input data location Replica Catalog TOP GIIS... CE Executable = "/usr/bin/env"; Arguments = "zsh prod.dc1_wrc 00001"; VirtualOrganization="datatag"; Requirements=Member(other.GlueHostApplicationSoftware RunTimeEnvironment,"ATLAS-3.2.1" ); Rank = other.GlueCEStateFreeCPUs; InputSandbox={"prod.dc1_wrc",“rc.conf","plot.kumac"}; OutputSandbox={"dc1.002000.test.00001.hlt.pythia_jet_17.log","dc1.002000.test.00001.hlt.pythia_jet_17. his","dc1.002000.test.00001.hlt.pythia_jet_17.err","plot.kumac"}; ReplicaCatalog="ldap://dell04.cnaf.infn.it:9211/lc=ATLAS,rc=GLUE,dc=dell04,dc=cnaf,dc=infn,dc=it"; InputData = {"LF:dc1.002000.evgen.0001.hlt.pythia_jet_17.root"}; StdOutput = " dc1.002000.test.00001.hlt.pythia_jet_17.log"; StdError = "dc1.002000.test.00001.hlt.pythia_jet_17.err"; DataAccessProtocol = "file"; JDL GLUE-aware files WN ATLAS sw data registration GLUE-Schema based Information System WorldGrid Testbed JDL Job GENIUS UI see WorldGrid Poster this conf.
20
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 20 20 CMS Applications l Monte Carlo Production chain on Grid –CMKIN: generation physics events with PYTHIA –CMSIM: simulation of the detector with GEANT3 l CMS production software installed in the WN’s l Job workflow and data management –CMKIN jobs sent by the RB to WN with CMS software, store the output at nearby SE >register LFN to the RC –CMSIM jobs sent by the RB to WN nearby SE >Register LFN to the RC
21
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 21 21 ATLAS Applications l Grappa and Genius submissions l ATLAS Detector Simulations –Simulation of the detector response using ATLSIM (GEANT3) –Based on DC1 Grid script l ATLAS production software installed in the WN’s
22
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 22 22 Grappa Portal Engine Web browser interface h ttps Cactus framework Script interface Resource A Resource Z... MAGDA: replica and metadata Compute Elements Java CoG submission, monitoring Storage Elements: - Disk/HPSS input files Grappa and ATLAS see D. Engh this conf.
23
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 23 23 Job Submission Animation
24
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 24 24 VO Monitoring l Initial Requirements: –Grid-level resource activity, utilization, and performance monitoring; –VO-level resource activity and resource utilization monitoring; –Customized views: >Hardware resources (clusters, sites, grids); >VO usages, jobs, work-types; l Design Goals: –Scalability over large number of resources and networks; –Simplicity and distributed architecture; l Two approaches –iVDGL: built on popular Ganglia resource monitoring package from UC Berkeley –DataTAG: built on popular Nagios package http://www.nagios.org/
25
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 25 25 Web php client Grid Aggregation Round Robin DB Tool Logging & Bookeeping UI RB JSSCE RRDB Tool gmond Site a RRDB Tool gmond Site b DataTAG VO Ganglia iVDGL
26
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 26 26 Site Level VO Usage and Policy
27
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 27 27 VO Nagios Monitoring based on Nagios (a host and service monitoring engine) [detailed information on: http://www.nagios.org] host local plug-ins – collect info from OS - CPU load - RAM - disk - jobs MDS plug-ins - collect aggregate info from GRIS - number of running/waiting jobs - number of total/free CPUs history graphs for all monitoring metrics aggregate info/graphs per Site and Virtual Organization
28
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 28 28 Status and Summary Map grid-aggregate monitors 3-level status map
29
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 29 29 VO Usage Graphs MDS collected site and aggregated montiors see G. Tortone et. al., this conference
30
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 30 30 WorldGrid Next Steps l New developments in DataTAG: SRM solution –Test/experiment with SRM solution for Storage Element access (multiple implementations of the protocol) Data Management tools –Test/experiment with advanced Data Management tools such as Globus-EDG/RLS Grid Resource Discovery –Propose alternative Grid Resource Discovery mechanisms based on WEB services monitoringOGSA –Improve the monitoring tools taking advantage of OGSA WorldGrid GOC –Develop a WorldGrid GOC, coordinated operations centers l Continue themes in iVDGL: –site-friendly installations, untouched by humans –multi-VO (controlled use of shared resources) –pursue concept of ‘projects’
31
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 31 31 A project consists of A (typically small) list of distinguished names or VO(s). Email and phone contact. A software environment expressed as a Pacman package. Local disk space requirements. A url describing the project. Basic site management operations: Join a project Leave a project Pause a project Site manager commands Projects as unit of access
32
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 32 32 % worldgrid –info -join -leave -pause -kill -update -getCA -setForum Example Site Manager Commands
33
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 33 33 iVDGL WorldGrid FAQ Forum Help Joined projects Demo ATLASDC2-higgs ChimeraTest8 Projects Certified Demo CMS-DC2-SUSY ChimeraTest8 ChimeraTest9 ATLASDC2-higgs SDSC-scan45 Installed Software WorldGrid ScienceGrid ProjectAccess CAs Batch jobs Performance 10/150 G used in WorkSpace History
34
R. Gardner CHEP 2003 – 24-28 March 2003 – n o 34 34 Conclusions l Lessons from WorldGrid 2002 –Grid building >Packaging and configuration key >GLUE meta-packaging study launched, report available >Testing and site validation –Interoperability >Configuration of common MDS schema allowed joint use of VDT and EDG middleware installations >good experience for LCG –Integrate two very different grids >“Top down” EDG-style of Grids with high level services >“Bottoms up” VDT-style grids providing core services with –Transatlantic cooperation can be fun!
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.