Enabling Grids for E-sciencE www.eu-egee.org Experience Supporting the Integration of LHC Experiments Computing Systems with the LCG Middleware Simone.

Slides:



Advertisements
Similar presentations
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Torsten Antoni – LCG Operations Workshop, CERN 02-04/11/04 Global Grid User Support - GGUS -
Advertisements

A tool to enable CMS Distributed Analysis
Client/Server Grid applications to manage complex workflows Filippo Spiga* on behalf of CRAB development team * INFN Milano Bicocca (IT)
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
CERN - IT Department CH-1211 Genève 23 Switzerland t Monitoring the ATLAS Distributed Data Management System Ricardo Rocha (CERN) on behalf.
GRACE Project IST EGAAP meeting – Den Haag, 25/11/2004 Giuseppe Sisto – Telecom Italia Lab.
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
CERN IT Department CH-1211 Geneva 23 Switzerland t The Experiment Dashboard ISGC th April 2008 Pablo Saiz, Julia Andreeva, Benjamin.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
INFSO-RI Enabling Grids for E-sciencE SA1: Cookbook (DSA1.7) Ian Bird CERN 18 January 2006.
GGF12 – 20 Sept LCG Incident Response Ian Neilson LCG Security Officer Grid Deployment Group CERN.
Enabling Grids for E-sciencE Overview of System Analysis Working Group Julia Andreeva CERN, WLCG Collaboration Workshop, Monitoring BOF session 23 January.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Julia Andreeva CERN (IT/GS) CHEP 2009, March 2009, Prague New job monitoring strategy.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
INFSO-RI Enabling Grids for E-sciencE Project Gridification: the UNOSAT experience Patricia Méndez Lorenzo CERN (IT-PSS/ED) CERN,
OSG Area Coordinator’s Report: Workload Management April 20 th, 2011 Maxim Potekhin BNL
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.
David Adams ATLAS ADA, ARDA and PPDG David Adams BNL June 28, 2004 PPDG Collaboration Meeting Williams Bay, Wisconsin.
Enabling Grids for E-sciencE System Analysis Working Group and Experiment Dashboard Julia Andreeva CERN Grid Operations Workshop – June, Stockholm.
INFSO-RI Enabling Grids for E-sciencE SA1 and gLite: Test, Certification and Pre-production Nick Thackray SA1, CERN.
CEOS WGISS-21 CNES GRID related R&D activities Anne JEAN-ANTOINE PICCOLO CEOS WGISS-21 – Budapest – 2006, 8-12 May.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Overview of STEP09 monitoring issues Julia Andreeva, IT/GS STEP09 Postmortem.
1 LHCb on the Grid Raja Nandakumar (with contributions from Greig Cowan) ‏ GridPP21 3 rd September 2008.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
LCG ARDA status Massimo Lamanna 1 ARDA in a nutshell ARDA is an LCG project whose main activity is to enable LHC analysis on the grid ARDA is coherently.
Julia Andreeva, CERN IT-ES GDB Every experiment does evaluation of the site status and experiment activities at the site As a rule the state.
INFSO-RI Enabling Grids for E-sciencE Experience of using gLite for analysis of ATLAS combined test beam data A. Zalite / PNPI.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Antonio Retico CERN, Geneva 19 Jan 2009 PPS in EGEEIII: Some Points.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Ricardo Rocha CERN (IT/GS) EGEE’08, September 2008, Istanbul, TURKEY Experiment.
INFSO-RI Enabling Grids for E-sciencE The gLite File Transfer Service: Middleware Lessons Learned form Service Challenges Paolo.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Plans for Service Challenge 3 Ian Bird LHCC Referees Meeting 27 th June 2005.
Julia Andreeva on behalf of the MND section MND review.
The GridPP DIRAC project DIRAC for non-LHC communities.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Monitoring of the LHC Computing Activities Key Results from the Services.
MND review. Main directions of work  Development and support of the Experiment Dashboard Applications - Data management monitoring - Job processing monitoring.
IAG – Israel Academic Grid, EGEE and HEP in Israel Prof. David Horn Tel Aviv University.
INFSO-RI SA2 ETICS2 first Review Valerio Venturi INFN Bruxelles, 3 April 2009 Infrastructure Support.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
INFSO-RI Enabling Grids for E-sciencE UNOSAT and Geant4: Experiences of their merge in the LCG Environment Patricia Méndez Lorenzo.
Enabling Grids for E-sciencE CMS/ARDA activity within the CMS distributed system Julia Andreeva, CERN On behalf of ARDA group CHEP06.
INFSO-RI Enabling Grids for E-sciencE gLite Test and Certification Effort Nick Thackray CERN.
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
The GridPP DIRAC project DIRAC for non-LHC communities.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks EGEE Operations: Evolution of the Role of.
SAM Status Update Piotr Nyczyk LCG Management Board CERN, 5 June 2007.
Status of gLite-3.0 deployment and uptake Ian Bird CERN IT LCG-LHCC Referees Meeting 29 th January 2007.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
LCG Workshop User Support Working Group 2-4 November 2004 – n o 1 Some thoughts on planning and organization of User Support in LCG/EGEE Flavia Donno LCG.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES The Common Solutions Strategy of the Experiment Support group.
ALICE Physics Data Challenge ’05 and LCG Service Challenge 3 Latchezar Betev / ALICE Geneva, 6 April 2005 LCG Storage Management Workshop.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
WLCG Accounting Task Force Update Julia Andreeva CERN GDB, 8 th of June,
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Enabling Grids for E-sciencE Claudio Cherubino INFN DGAS (Distributed Grid Accounting System)
GRID & Parallel Processing Koichi Murakami11 th Geant4 Collaboration Workshop / LIP - Lisboa (10-14/Oct./2006) 1 GRID-related activity in Japan Go Iwai,
INFN-GRID Workshop Bari, October, 26, 2004
INFNGRID Workshop – Bari, Italy, October 2004
Simulation use cases for T2 in ALICE
LCG middleware and LHC experiments ARDA project
LCG Operations Workshop, e-IRG Workshop
Monitoring of the infrastructure from the VO perspective
Presentation transcript:

Enabling Grids for E-sciencE Experience Supporting the Integration of LHC Experiments Computing Systems with the LCG Middleware Simone Campana LCG Experiment Integration and Support CERN-IT / INFN-CNAF

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 2 Mandate of the LCG/EIS Team  EIS Experiment Integration and Support  EIS : Experiment Integration and Support Team integrating Help LHC Experiments integrating their production environment with the Grid Middleware and utilities. support Offer support during all steps of integration process  understanding of the middleware functionality  testing new prototypal components  getting on the LCG Infrastructure. One person dedicated to each LHC Experiment  Production  Production is the main focus. User Support Experiment Support does not mean User Support. GOC Experiment Support does not mean GOC.

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 3 Main Tasks  Integration Middleware functionality and usage Functionality tests Customized distributions and missing tools Discuss requirements  And bring them to the attention of the developers  Experiment and User Support Documentation: Manuals, Guides, FAQ First line User Support Monitoring experiment specific production system  Provide infrastructure expertise Monitoring/Managing Services  GRID and Experiment Specific Solving site-related problems Service Challenge Second Level Support (on shift)

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 4 Tools… Tools… Tools…  Data Management Customized version of LCG Data Management clients  Workload Management Monitoring of the job “standard error” and “standard output”  g-peek Estimate job normalized CPU and Wall Clock time left on CPU  Information System C++ Generic API (with ldap and R-GMA backends) User friendly querying tools  Generic Framework for Job Submission Intensively used by GEANT4  Many others …  Several functionalitiesintegrated in the Middleware  Several functionalities provided by the tools have been integrated in the Middleware See the g-peek functionality

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 5 Monitoring Tools ATLAS SC3 Service Monitor LHCb specific Site Functional Tests

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 6 Experiment Software Installation Lcg-ManageSoftware Lcg-ManageVOTag Tank&Spark gssklog Lcg-asis UI WN CE

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 7 VO-BOX developed and packaged  First prototype developed and packaged by EIS.  Evaluation of the Globus GSI-enabled ssh server and relative configuration  Development of a ad-hoc proxy renewal server with relative user level tool  Overall configuration of the node type Inclusion of UI clients and gssklog installation issues discussions on possible evolution  Following up installation issues and further discussions on possible evolution

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 8 EIS on ALICE Data Challenges 04 and 05  EIS For Data Challenges 04 and 05 Offered support for the integration of ALICE framework with LCG services  Integration with existing LCG services  Development of new tools Follow up of production exercise  Provided solution for site specific problems  Follow up of services deployment at the sites Collected ALICE requirements for middleware developers

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 9 EIS on ALICE ALICE specific user level tools  Development ALICE specific user level tools Integration of Monalisa monitoring system with LCG  Later, the tools have been generalized for other use-cases FTS transfer handling client  Then integrated in the ALICE framework Publication of VO specific services in the Information System  Included as part of the VO-BOX middleware component

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 10 Some Results of the last PDC04  ◘ Statistics after phase 1 (ended April 4, 2004): ➸ ALICE::CERN::LCG is the interface to LCG-2 ➸ ALICE::Torino::LCG is the interface to GRID.IT 4 ~ 1.3 million files, 26 TB data volume S. Bagnasco. SC3 Detailed Planning Workshop, CERN 13.June, 05 )

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 11 EIS in ATLAS  Support in the development of ATLAS framework Data Management Workload management  Operational support Exclusion of problematic sites Follow up of site configuration problems Understanding of failures and suggestion of solutions Number of jobs per day Data Challenge 2 Rome Production production Large event production for Rome workshop Physics Rome workshop EIS support activities

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 12 Rome Production experience on LCG  Jobs distributed to 45 different computing resources  Ratio generally proportional to the size of the cluster indicates an overall good job distribution.  No site in particular ran large majority of jobs. The site with the largest number of CPU resources (CERN), contributed for about 11% of the ATLAS production. Other major sites ran between 5% and 8% of the jobs each. robust and fault-tolerant system  Achievement toward a more robust and fault-tolerant system does not rely on a small number of large computing centers. The percentage of ATLAS jobs run at each LCG site

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 13 EIS in ATLAS  Service Challenge 3 Support to the ATLAS Data Management System  File Transfer Service (FTS) and LCG File Catalog (LFC)  Prototype Data Location Interface (DLI) developed ATLAS WMS and DDM integration. Role in the technical coordination of the ATLAS Service Challenge activities  ensuring the readiness of the sites before and during the exercise  following up issues with the different services.  Testing Several new glite components (WMS, gpbox, FTS …) In the context of the task force and in collaboration with ARDA  User Support Analysis on LCG produced data

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 14 EIS in CMS  LFCPOOL file catalog  LFC evaluation as a POOL file catalog use case: local file catalog performance tests Results: LFC and POOL_LFC interface issues discovered and fixed  LFCData Location System  LFC evaluation as a Data Location System implementation of a Python API performance tests Results: LFC was found to be an valid implementation of a DLS; performance issues discovered and fixed

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 15 EIS in CMS  Service Challenge 3 fake analysis job submission analysis of job failures and related statistics Results: much better understanding of the stability of the LCG infrastructure when intensively used  Support active in the solution of Grid-related problems for the MC production and user analysis (CRAB) activities VO management  CMS VO management

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 16 The CMS Analysis Jobs Taken from the CMS Dashboard (ARDA)

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 17 EIS in LHCb  EIS supported LHCb along many activities: Data Challenge 04 Data Challenge 04 Service Challenge 3 Service Challenge 3 Analysis exercise Analysis exercise  Operation support chasing/tackling sites and middleware related problems developing experiment specific monitoring tools  T1-T1 transfer monitor for SC3  VO oriented plug-ins for SFT

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 18 EIS in LHCb  Integration of LHCb framework and LCG middleware Offering suggestions for an optimized middleware usage Development of user level tools  Query the information system, interactions with SRM, LFC, DLI. Repackaging or customized version of existing tools  lcg_utils and GFAL  User Support Especially for analysis users Using the GGUS portal  Testing of new components CREAM CE, g-pbox, WMS …

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 19 The LHCb Data Challenge DIRAC alone LCG in action /day LCG paused Phase 1 Completed /day LCG restarted 187 M Produced Events 61% efficiency for LCG Number of Jobs run versus time Jobs run in LCG and Dirac-only sites

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 20  WISDOM: research on malaria medical care Major success in EGEE  1 million of potential medicines tested in 1 week  1000 CPUs employed in EGEE/LCG Support to Biomedical community and the WISDOM project no-HEP  First no-HEP VO supported by EIS Different needs, access pattern, user scenarios Scattered and heterogeneous community  Main support activities for Biomed: Improvement of Job submission strategy Adaptation of application to Grid Environment Oparational support User Support Data Challenge  Biomedical Data Challenge in July - August 2005 ~70000 jobs run 1 TB of data produced equivalent of ~70 CPU years computed.

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 21 GEANT4  GEANT4simulation of particle interactions with matter  GEANT4: simulation of particle interactions with matter. HEP and nuclear experiments, medical, accelerator, space physics 3 major productions on LCG  First 2 hosted by dteam and alice, third as a real VO Aimed to test new version of software “Gridification” process  EIS support in GEANT “Gridification” process Development of tools for job submission an handling  Then extended and generalized for other VOs Creation and administration of the GEANT VO  Contact point for the EGEE ROC managers Operational support during production

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 22 Relief Projects of UNOSAT  Case Study  Case Study: Indian Ocean Tsunami Relief and Development 29 th Dec 2004: First Map distributed online to field users January 2005: 200,000 tsunami maps downloaded in total UNOSAT has a huge amount of data to be stored Good amount of storage provided by CERN Running and storing data in LCG/EGEE can certainly assist UNOSAT in their purposes In Summer 2005 the collaboration with LCG started Gridification prcess similar to GEANT4 experience Gridification prcess similar to GEANT4 experience

Enabling Grids for E-sciencE CHEP06 – February 2006 – Mumbay (India) 23 Summary Our mailing list: Our WEB site: integratingVOsoftware environment GRID middleware  EIS provides help integrating VO specific software environment with GRID middleware Direct experiment support via a contact persons Special middleware distributions documentation User support Data Challenges, Service Challenges and Distributed Productions  Data Challenges, Service Challenges and Distributed Productions Follow up of operational issues  maintaing experiment specific services  assisting sites with configuration problems Not anymore “sporadic” exercises. interesting a productive experience  Overall a very interesting a productive experience LHC experiments and other VOs seem to find EIS team very supportive