1 ALICE Grid Status David Evans The University of Birmingham GridPP 16 th Collaboration Meeting QMUL 27-29 June 2006.

Slides:



Advertisements
Similar presentations
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks MyProxy and EGEE Ludek Matyska and Daniel.
Advertisements

S.L.LloydATSE e-Science Visit April 2004Slide 1 GridPP – A UK Computing Grid for Particle Physics GridPP 19 UK Universities, CCLRC (RAL & Daresbury) and.
1 ALICE Grid Status David Evans The University of Birmingham GridPP 14 th Collaboration Meeting Birmingham 6-7 Sept 2005.
GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
Tony Doyle - University of Glasgow GridPP EDG - UK Contributions Architecture Testbed-1 Network Monitoring Certificates & Security Storage Element R-GMA.
Storage Review David Britton,21/Nov/ /03/2014 One Year Ago Time Line Apr-09 Jan-09 Oct-08 Jul-08 Apr-08 Jan-08 Oct-07 OC Data? Oversight.
Southgrid Status Pete Gronbech: 21 st March 2007 GridPP 18 Glasgow.
B A B AR and the GRID Roger Barlow for Fergus Wilson GridPP 13 5 th July 2005, Durham.
Tony Doyle GridPP2 Proposal, BT Meeting, Imperial, 23 July 2003.
The LHC experiments AuthZ Interoperation requirements GGF16, Athens 16 February 2006 David Kelsey CCLRC/RAL, UK
Andrew McNab - Manchester HEP - 22 April 2002 EU DataGrid Testbed EU DataGrid Software releases Testbed 1 Job Lifecycle Authorisation at your site More.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Torsten Antoni – LCG Operations Workshop, CERN 02-04/11/04 Global Grid User Support - GGUS -
Data Management Expert Panel. RLS Globus-EDG Replica Location Service u Joint Design in the form of the Giggle architecture u Reference Implementation.
GridPP From Prototype to Production David Britton 21/Sep/06 1.Context – Introduction to GridPP 2.Performance of the GridPP/EGEE/wLCG Grid 3.Some Successes.
LCG Tiziana Ferrari - SC3: INFN installation status report 1 Service Challenge Phase 3: Status report Tiziana Ferrari on behalf of the INFN SC team INFN.
Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
ATLAS Tier-3 in Geneva Szymon Gadomski, Uni GE at CSCS, November 2009 S. Gadomski, ”ATLAS T3 in Geneva", CSCS meeting, Nov 091 the Geneva ATLAS Tier-3.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
Grid and CDB Janusz Martyniak, Imperial College London MICE CM37 Analysis, Software and Reconstruction.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
1 Status of the ALICE CERN Analysis Facility Marco MEONI – CERN/ALICE Jan Fiete GROSSE-OETRINGHAUS - CERN /ALICE CHEP Prague.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
ATLAS-Specific Activity in GridPP EDG Integration LCG Integration Metadata.
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
Experience with the WLCG Computing Grid 10 June 2010 Ian Fisk.
BaBar Grid Computing Eleonora Luppi INFN and University of Ferrara - Italy.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
EGEE is a project funded by the European Union under contract IST Testing processes Leanne Guy Testing activity manager JRA1 All hands meeting,
The ILC And the Grid Andreas Gellrich DESY LCWS2007 DESY, Hamburg, Germany
GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
GLite – An Outsider’s View Stephen Burke RAL. January 31 st 2005gLite overview Introduction A personal view of the current situation –Asked to be provocative!
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
Status of PDC’07 and user analysis issues (from admin point of view) L. Betev August 28, 2007.
Overview of grid activities in France in relation to FKPPL FKPPL Workshop Thursday February 26th, 2009 Dominique Boutigny.
1 LHCb on the Grid Raja Nandakumar (with contributions from Greig Cowan) ‏ GridPP21 3 rd September 2008.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
LCG ARDA status Massimo Lamanna 1 ARDA in a nutshell ARDA is an LCG project whose main activity is to enable LHC analysis on the grid ARDA is coherently.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow UK-T0 Meeting 21 st Oct 2015 GridPP.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
Site Report: Prague Jiří Chudoba Institute of Physics, Prague WLCG GridKa+T2s Workshop.
The CMS Top 5 Issues/Concerns wrt. WLCG services WLCG-MB April 3, 2007 Matthias Kasemann CERN/DESY.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
The GridPP DIRAC project DIRAC for non-LHC communities.
Materials for Report about Computing Jiří Chudoba x.y.2006 Institute of Physics, Prague.
Production Activities and Results by ALICE Patricia Méndez Lorenzo (on behalf of the ALICE Collaboration) Service Challenge Technical Meeting CERN, 15.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Enabling Grids for E-sciencE INFSO-RI Enabling Grids for E-sciencE Gavin McCance GDB – 6 June 2007 FTS 2.0 deployment and testing.
ALICE experiences with CASTOR2 Latchezar Betev ALICE.
David Adams ATLAS ATLAS Distributed Analysis (ADA) David Adams BNL December 5, 2003 ATLAS software workshop CERN.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
The GridPP DIRAC project DIRAC for non-LHC communities.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks EGEE Operations: Evolution of the Role of.
Status of gLite-3.0 deployment and uptake Ian Bird CERN IT LCG-LHCC Referees Meeting 29 th January 2007.
J Jensen/J Gordon RAL Storage Storage at RAL Service Challenge Meeting 27 Jan 2005.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
ALICE Physics Data Challenge ’05 and LCG Service Challenge 3 Latchezar Betev / ALICE Geneva, 6 April 2005 LCG Storage Management Workshop.
Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.
The ALICE Production Patricia Méndez Lorenzo (CERN, IT/PSS) On behalf of the ALICE Offline Project LCG-France Workshop Clermont, 14th March 2007.
Availability of ALICE Grid resources in Germany Kilian Schwarz GSI Darmstadt ALICE Offline Week.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
Update on Plan for KISTI-GSDC
Simulation use cases for T2 in ALICE
LHC Data Analysis using a worldwide computing grid
Ivan Reid (Brunel University London/CMS)
Presentation transcript:

1 ALICE Grid Status David Evans The University of Birmingham GridPP 16 th Collaboration Meeting QMUL June 2006

2 Outline of Talk The ALICE Experiment The ALICE Experiment ALICE computing requirements ALICE computing requirements ALICE Grid – AliEn ALICE Grid – AliEn Analysis using AliEn Analysis using AliEn Status of ALICE Data Challenge 2006 Status of ALICE Data Challenge 2006 Summary and Outlook Summary and Outlook

3 The ALICE Experiment ALICE is one of the four main LHC experiments at CERN. ALICE is one of the four main LHC experiments at CERN. Only one dedicated to heavy-ion physics. Only one dedicated to heavy-ion physics. –Study of QCD under extreme conditions ~ 1000 collaborators ~ 1000 collaborators ~ 100 institutions ~ 100 institutions Birmingham is only UK institute involved Birmingham is only UK institute involved

4 ALICE Requirements Data taking (each year) Data taking (each year) –1 month of Pb-Pb data ~ 1 PByte –Also p-p for rest of the year ~ 1 PByte Large scale simulation effort Large scale simulation effort –1 Pb-Pb event: ~ 8 hrs (3 GHz) Data Reconstruction Data Reconstruction Data analysis Data analysis Smaller Collaboration than ATLAS or CMS but similar computing requirements. Smaller Collaboration than ATLAS or CMS but similar computing requirements.

5 Profile of CPU requirements Total CERN T0 CERN T1 Ext Tier 1 Ext Tier 2 35 MSK2K Jan 07Sept 08 Nov 09

6 Tier Hierarchy MONARC Model MONARC Model Cloud Model (Tier free) used Cloud Model (Tier free) used in ALICE data challenges (native AliEn sites – for LCG site we comply with Tier model) in ALICE data challenges (native AliEn sites – for LCG site we comply with Tier model) Tier 0 RAW data master copy Data reconstruction (1 st pass) Prompt analysis Tier 1 Copy of RAW reconstruction Scheduled analysis Tier 2 MC production Partial copy of ESD Data analysis

7 ALICE Gridd - AliEn AliEn (ALICE Environment) – Grid framework developed by ALICE – used in production for ~5 years. AliEn (ALICE Environment) – Grid framework developed by ALICE – used in production for ~5 years. Based on WEB services and standard protocols. Based on WEB services and standard protocols. Built around open source code Built around open source code –Less than 5% is native AliEn code (mainly PERL). To date, > 500,000 ALICE jobs have been run under AliEn control worldwide. To date, > 500,000 ALICE jobs have been run under AliEn control worldwide.

8 AliEn Pull Protocol One of the major differences between ALiEn and LCG grids is that AliEn uses the pull rather thanpush protcol. One of the major differences between ALiEn and LCG grids is that AliEn uses the pull rather thanpush protcol. EDG/Globus model: EDG/Globus model: ALiEn model: ALiEn model: userserver Resource Broker userserver Resource Broker job list

9 LCG / gLite ALICE is committed to using as much common grid applications as possible. ALICE is committed to using as much common grid applications as possible. Changes have been made to make AliEn work with LCG Changes have been made to make AliEn work with LCG –E.g. changes to File Catalogue (FC) LFC (Local File Catalogue or LCG File Catalogue) –V0 Box at each Tier 1 and Tier 2 –Globus/GSI compatible authentication Interface AliEn gLite in development Interface AliEn gLite in development

10 Analysis Core of ALICE computing model is AliRoot Core of ALICE computing model is AliRoot –Uses ROOT framework Couple AliEn with ROOT for Grid-based analysis. Couple AliEn with ROOT for Grid-based analysis. –Use PROOF – Parallel ROOT Facility –To the user its like using ROOT 4-tier architecture: 4-tier architecture: –ROOT client session, API server (AliEn + PROOF), Site PROOF master servers, PROOF slave servers. Data from DC2006 only accessible via Grid Data from DC2006 only accessible via Grid

11 PROOF Each node has PROOF slave Each site has PROOF master server Uses pull protocol i.e. the slaves ask the master for work packets. Slower slaves get smaller work packets etc. Client API Server AliEn FC …. List of sites with data

12 ALICE Data Challenge 2006 (PDC06) Last challenge before the start of data taking Last challenge before the start of data taking Test of all Grid components Test of all Grid components –AliEn as a ALICE interface to the Grid and much, much more –LCG/gLite baseline services (WMS, DMS) Test of computing centres infrastructure Test of computing centres infrastructure Major test of stability of all of the above Major test of stability of all of the above

13 Grid software deployment and running LCG sites are operated through the VO-box framework LCG sites are operated through the VO-box framework –All ALICE sites need one –Relatively extended deployment cycle, a lot of configuration and version update issues had to be solved –Situation is quite routine now Data management Data management –This year – xrootd as disk pool manager on all sites –The installation/configuration procedures have just been released –xrootd integrated in other storage management solutions (CASTOR, DPM, dCache) – under development Data replication (FTS) Data replication (FTS) –We use it for scheduled replication of data between the computing centres (RAW from T0->T1, MC production T2->T1, etc…) –Fully incorporated in the AliEn FTD, to be extensively tested in July

14 VO box support and operation In additional to the standard LCG components, the VO-box runs ALICE-specific software components In additional to the standard LCG components, the VO-box runs ALICE-specific software components –V0-boxes now at RAL Tier 1 and Birmingham Tier 2 –Birmingham ALICE students are testing ALiEn for analysis purposes through Birmingham Tier 2. The installation and maintenance of these is entirely our responsibility: The installation and maintenance of these is entirely our responsibility: –Support for UK V0-box supplied by CERN (no UK manpower available) Site related problems are handled by the site admins Site related problems are handled by the site admins LCG services problems are reported to GGUS LCG services problems are reported to GGUS

15 Operation status Running in a continuous mode since 24/05 Running in a continuous mode since 24/05 VO-boxes: VO-boxes: –monthly releases of AliEn (curently v.2-10), LCG and soon tests of gLite 3.0 Central ALICE services: Central ALICE services: –AliEn machinery and API Service is developed/deployed and maintained by the AliEn team Site services: Site services: –Stability testing of both AliEn and LCG components –The interfaces AliEn-LCG/gLite are still in development –A gLite V0-box has already been provided at CERN and first tests performed.

16 Running status – one month

17 Sites contributions in the past 2 months 60%T1, 40%T2 (almost half from 2 T2 sites!) 60%T1, 40%T2 (almost half from 2 T2 sites!) RAL: 0.7%

18 Running status – site averages Pledged resources – 4000 CPUs Pledged resources – 4000 CPUs Our average is on a 12% level Our average is on a 12% level –Due to central and site services malfunctions –Mostly due to sites providing less CPUs than pledged

19 Stability improvements This is a data challenge, so there is always place for improvement: This is a data challenge, so there is always place for improvement: –AliEn is undergoing gradual fixes and new features are added –The LCG software will undergo a quantum leap – move from LCG to gLite –Site infrastructure – VO-box, etc… also needs solidification, especially at the T2s –Monitoring and control – continuously adding new features

20 Outlook PDC06 has started as planned PDC06 has started as planned –This is the last exercise before the beam! –It is a test of all Grid tools/services we will use in 2007 »If not in PDC06, good chance is that they will not be ready –It is also a large-scale test the computing infrastructure – computing, storage and network performance

21 Outlook (2) Outlook (2) We have all pieces needed to run production on the Grid (some untested). We have all pieces needed to run production on the Grid (some untested). The exercise started 2 months ago and will continue until the end of the year The exercise started 2 months ago and will continue until the end of the year At the moment, we are optimising the use of resources – attempting to get from the sites the promised resources At the moment, we are optimising the use of resources – attempting to get from the sites the promised resources Next phase of the plan is a test of the file transfer utilities of LCG (FTS) and integration with AliEn FTD Next phase of the plan is a test of the file transfer utilities of LCG (FTS) and integration with AliEn FTD In parallel to that we will run event production as usual In parallel to that we will run event production as usual

22 Summary AliEn is a Grid framework developed by ALICE using 95% open source code (e.g SOAP) and 5 % AliEn specific (perl) code. AliEn is a Grid framework developed by ALICE using 95% open source code (e.g SOAP) and 5 % AliEn specific (perl) code. AliEn evolving to take into account EGEE/gLite framework and to work with LCG. AliEn evolving to take into account EGEE/gLite framework and to work with LCG. –New user interfaces developed –PROOF for analysis developed –Better authentication/authorisation developed Data Challenge 2006 – since April – going well Data Challenge 2006 – since April – going well V0 boxes at RAL T1 and Bham T2 V0 boxes at RAL T1 and Bham T2 Lack of computing resources a worry. Lack of computing resources a worry.