LCG Service Challenge: Planning and Milestones

Slides:



Advertisements
Similar presentations
LCG Tiziana Ferrari - SC3: INFN installation status report 1 Service Challenge Phase 3: Status report Tiziana Ferrari on behalf of the INFN SC team INFN.
Advertisements

INFN Testbed status report L. Gaido WP6 meeting CERN - October 30th, 2002.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Stefano Belforte INFN Trieste 1 CMS SC4 etc. July 5, 2006 CMS Service Challenge 4 and beyond.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
Status Report on Tier-1 in Korea Gungwon Kang, Sang-Un Ahn and Hangjin Jang (KISTI GSDC) April 28, 2014 at 15th CERN-Korea Committee, Geneva Korea Institute.
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
CHEP – Mumbai, February 2006 The LCG Service Challenges Focus on SC3 Re-run; Outlook for 2006 Jamie Shiers, LCG Service Manager.
BINP/GCF Status Report BINP LCG Site Registration Oct 2009
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
SRM 2.2: tests and site deployment 30 th January 2007 Flavia Donno, Maarten Litmaath IT/GD, CERN.
WLCG Service Schedule June 2007.
LCG Service Challenge Phase 4: Piano di attività e impatto sulla infrastruttura di rete 1 Service Challenge Phase 4: Piano di attività e impatto sulla.
Southgrid Technical Meeting Pete Gronbech: 26 th August 2005 Oxford.
Status Report of WLCG Tier-1 candidate for KISTI-GSDC Sang-Un Ahn, for the GSDC Tier-1 Team GSDC Tier-1 Team 12 th CERN-Korea.
LCG Service Challenges: Planning for Tier2 Sites Update for HEPiX meeting Jamie Shiers IT-GD, CERN.
LCG Service Challenges: Planning for Tier2 Sites Update for HEPiX meeting Jamie Shiers IT-GD, CERN.
Status Report of WLCG Tier-1 candidate for KISTI-GSDC Sang-Un Ahn, for the GSDC Tier-1 Team GSDC Tier-1 Team ATHIC2012, Busan,
WLCG 1 Service Challenge 4: Preparation, Planning and Outstanding Issues at INFN Workshop sul Calcolo e Reti dell'INFN Jun.
Site Report: Prague Jiří Chudoba Institute of Physics, Prague WLCG GridKa+T2s Workshop.
Tier-1 Andrew Sansum Deployment Board 12 July 2007.
Plans for Service Challenge 3 Ian Bird LHCC Referees Meeting 27 th June 2005.
LHCb report to LHCC and C-RSG Philippe Charpentier CERN on behalf of LHCb.
Report from GSSD Storage Workshop Flavia Donno CERN WLCG GDB 4 July 2007.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
The Grid Storage System Deployment Working Group 6 th February 2007 Flavia Donno IT/GD, CERN.
Status of gLite-3.0 deployment and uptake Ian Bird CERN IT LCG-LHCC Referees Meeting 29 th January 2007.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Status of GSDC, KISTI Sang-Un Ahn, for the GSDC Tier-1 Team
SRM 2.2: experiment requirements, status and deployment plans 6 th March 2007 Flavia Donno, INFN and IT/GD, CERN.
Storage & Database Team Activity Report INFN CNAF,
LCG Introduction John Gordon, STFC-RAL GDB June 11 th, 2008.
Dissemination and User Feedback Castor deployment team Castor Readiness Review – June 2006.
The status of IHEP Beijing Site WLCG Asia-Pacific Workshop Yaodong CHENG IHEP, China 01 December 2006.
HEPiX IPv6 Working Group David Kelsey (STFC-RAL) GridPP33 Ambleside 22 Aug 2014.
INFN Site Report R.Gomezel November 5-9,2007 The Genome Sequencing University St. Louis.
1-2 March 2006 P. Capiluppi INFN Tier1 for the LHC Experiments: ALICE, ATLAS, CMS, LHCb.
Operations Workshop Introduction and Goals Markus Schulz, Ian Bird Bologna 24 th May 2005.
Servizi core INFN Grid presso il CNAF: setup attuale
WLCG IPv6 deployment strategy
Davide Salomoni INFN-CNAF Bologna, Jan 12, 2006
DPM at ATLAS sites and testbeds in Italy
London Tier-2 Quarter Owen Maroney
WLCG Tier-2 Asia Workshop TIFR, Mumbai 1-3 December 2006
“A Data Movement Service for the LHC”
Dirk Duellmann CERN IT/PSS and 3D
October 28, 2013 at 14th CERN-Korea Committee, Geneva
Service Challenge 3 CERN
3D Application Tests Application test proposals
Data Challenge with the Grid in ATLAS
INFN-GRID Workshop Bari, October, 26, 2004
Database Readiness Workshop Intro & Goals
CMS — Service Challenge 3 Requirements and Objectives
ATLAS activities in the IT cloud in April 2008
Update on Plan for KISTI-GSDC
Proposal for obtaining installed capacity
Luca dell’Agnello INFN-CNAF
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
The INFN TIER1 Regional Centre
Olof Bärring LCG-LHCC Review, 22nd September 2008
John Gordon, STFC GDB October 12th 2011
The INFN Tier-1 Storage Implementation
Bernd Panzer-Steindel CERN/IT
Data Management cluster summary
Summary of Service Challenges
R. Graciani for LHCb Mumbay, Feb 2006
LHC Data Analysis using a worldwide computing grid
WLCG Collaboration Workshop: Outlook for 2009 – 2010
Overview & Status Al-Ain, UAE November 2007.
The LHCb Computing Data Challenge DC06
Presentation transcript:

LCG Service Challenge: Planning and Milestones Tiziana.Ferrari@cnaf.infn.it INFN CNAF Review Mar 1, 2006

Service Challenge 4 schedule SC3 and pre-SC4 results SC4: Outline Service Challenge 4 schedule SC3 and pre-SC4 results SC4: CNAF hw resources and milestones Service planning Personnel

SC Milestones 2006 January SC3 disk repeat – Nominal rate (200 MB/s) capped at 150MB/s/ February CHEP Workshop; T1-T1 Use Cases, SC3 disk - tape repeat (50MB/s, 5 drives) March Detailed plan for SC4 service agreed (M/W + DM service enhancements). gLite 3.0 release beta testing April SC4 disk – disk (200 MB/s) and disk – tape (75 MB/s) throughput tests. gLite 3.0 release available for distribution May Installation, configuration and testing of gLite 3.0 release at sites June Start of SC4 production tests by experiments of ‘T1 Use Cases’ T2 Workshop: identification of key Use Cases and Milestones for T2s July Tape throughput tests at full nominal rates! August T2 Milestones – debugging of tape results if needed September LHCC review – rerun of tape tests if required October WLCG Service Officially opened. Capacity continues to build up November 1st WLCG ‘conference’. All sites have network / tape h/w in production(?) December Final service / middleware review leading to early 2007 upgrades for LHC data taking??

Disk – disk throughput test rerun (Jan 06)

Alice: Running jobs (Jan-Feb 06) Alice Running Jobs [CS] FarmLast value Min Avg  Max  SUM    0    0    786.2    3651  Bari    0    0    11.28    84  CNAF    6.929    0    209.2    1072  Torino    1.108    0    21.99    47  SUM CNAF

LHCb: FTS channel mesh testing Jan-Feb 2006 CNAF

SC3: CMS, Phase 1 report (Sep-Nov 05) Objective: 10 or 50 TB per Tier-1 and ~5 TB per Tier-2 (source: L.Tuura)

SC3: LHCb report Phase 1 (Oct-Nov 05, Data moving) Less than 1TB stripped DSTs replicated. At INFN most of this data already existed with only a few files missing from the dataset therefore it was only necessary to replicate a small fraction of the files from CERN (source: A. C. Smith) Configuration of an entire CNAF-to-Tier1 channel matrix (for replication of stripped data)

Production phase started on Nov 2 SC3: ATLAS Production phase started on Nov 2 5932 files copied and registered at CNAF: 89 “Failed replication” events 14 “No replicas found”

SC4 Hardware Resources at CNAF Total capacities: 2500 CPU slots (1500 physical CPUs) disk: 400 TB, including the Castor buffer space tape: 200 TB (4 9940B + 6 LTO2 drives) 2 Gb/s available bandwidth for LHC (CERN - CNAF) LHC fraction: up to 2500 CPU slots (1500 physical CPUs) - all worker nodes are shared disk: 112 TB, including the Castor front-end tape: 160 TB (4 9940B + 6 LTO2 drives)   All CPU installed with SLC 3.0.4 and LCG 2.6.0 Additional test farm available Total capacities: 2500 CPU slots (1500 physical CPUs) disk: 400 TB, including the Castor buffer space tape: 200 TB (4 9940B + 6 LTO2 drives) 2 Gb/s available bandwidth for LHC (CERN - CNAF) LHC fraction: up to 2500 CPU slots (1500 physical CPUs) - all worker nodes are shared disk: 112 TB, including the Castor front-end tape: 160 TB (4 9940B + 6 LTO2 drives)   All CPU installed with SLC 3.0.4 and LCG 2.6.0 Additional test farm available Total capacities: 2500 CPU slots (1500 physical CPUs) disk: 400 TB, including the Castor buffer space tape: 200 TB (4 9940B + 6 LTO2 drives) 2 Gb/s available bandwidth for LHC (CERN - CNAF) LHC fraction: up to 2500 CPU slots (1500 physical CPUs) - all worker nodes are shared disk: 112 TB, including the Castor front-end tape: 160 TB (4 9940B + 6 LTO2 drives)   All CPU installed with SLC 3.0.4 and LCG 2.6.0 Additional test farm available Total capacities: 2500 CPU slots (1500 physical CPUs) disk: 400 TB, including the Castor buffer space tape: 200 TB (4 9940B + 6 LTO2 drives) 2 Gb/s available bandwidth for LHC (CERN - CNAF) LHC fraction: up to 2500 CPU slots (1500 physical CPUs) - all worker nodes are shared disk: 112 TB, including the Castor front-end tape: 160 TB (4 9940B + 6 LTO2 drives)   All CPU installed with SLC 3.0.4 and LCG 2.6.0 Additional test farm available SC4 Hardware Resources at CNAF 2500 CPU slots (out of 1000 physical CPUs) LHC: up to 2500 CPU slots (Worker Nodes are shared) SLC 3.0.4 and LCG 2.6.0 Disk: 400 TB (including the Castor buffer space) LHC: 100 TB (including the Castor buffer space) Tape: 260 TB (7 9940B + 6 LTO2 drives) LHC: 160 TB Network available bandwidth (CERN – CNAF): 10 Gb/s CNAF – T2 connectivity already tested (Bari, Catania, Legnaro, Milan, Pisa, Torino) CNAF – Karlsruhe connectivity (1 Gb/s, MPLS): under implementation

LCG CNAF Milestones (1st Q 2006) 16.01.06  150 MB/s (SC3 disk-disk, throughput test rerun) ☺ Feb 2006  upgrade to CASTOR v2 (installation started on Dec 1 2005)  ongoing Feb 2006  purchase of additional 120 TB on tape  ongoing Feb 2006  all required sw baseline services deployed (SRM, LFC, FTS, CE, RB, BDII, RGMA) ☺ Mar 2006  setup of CNAF – Karlsruhe backup connection Mar 2006  evaluation of dCache and StoRM (for disk-only SRM) Tier-2’s: Definition of INFN Tier-2 service plan is ongoing

Candidate Tier-2 sites in SC3 (Oct 05) Torino (ALICE): FTS, LFC, dCache (LCG 2.6.0) Storage Space: 2 TBy Milano (ATLAS): FTS, LFC, DPM 1.3.7 Storage space: 5.29 TBy Pisa (ATLAS/CMS): FTS, PhEDEx, POOL file cat, PubDB, LFC, DPM 1.3.5 Storage space: 5 TBy available, 5 TBy expected Legnaro (CMS): FTS, PhEDEx, Pool file cat., PubDB, DPM 1.3.7 (1 pool, 80 Gby) Storage space: 4 TBy Bari (ALICE/CMS): FTS, PhEDEx, POOL file cat., PubDB, LFC, dCache, DPM Storage space: 1.4 TBy available, 4 TBy expected Catania (ALICE): DPM and Classic SE (Storage space: 1.8 Tby) LHCb CNAF Catania hw configuration

SC4 Service Planning – CNAF (1/2) COMPO- NENT NEEDED BY PILOT USE PRODU-CTION COMMENT STATUS VOMS ALL Mar - Apr 2006 1 June 2006 Installation for production: May 06 Deployed Myproxy Mar- Apr 2006 BDII/GLUE FTS ALICE, ATLAS LHCb LFC CMS, ATLAS ALICE Lcg-utils GFAL RB Glite 3.0

SC4 Service Planning – CNAF (2/2) COMPO- NENT NEEDED BY PILOT USE PRODU-CTION COMMENT STATUS CE – classic gLite 3.0 ALL Mar - Apr 2006 1 June 2006 Installation for production: May 06 Deployed (Classic CE and CREAM) gPBox CMS, ATLAS Mar- Apr 2006 TBD v.1.0.16 on INFN Certification testbed RGMA/ GridICE ALL (GridICE) Deployed FTS ALICE, ATLAS LHCb Deployed (GridICE server 1.8.3-2) LFC ALICE APEL DGAS ALL (DGAS) VOBOX Until May CASTOR DPM dCache Casto2 and dCache under testing, Installation in May StoRM also under testing

Personnel Storage group: Castor, Castor2, dCache, FTS Network group: LHC OPN CNAF – CERN link configuration Farming group: from October 2005 Grid operations group: installation, testing, monitoring of SC services SC coordination at INFN: Tiziana Ferrari (CNAF) Michele Michelotto (INFN Padova, deputy) INFN SC mailing list Tier-2 relying on local man-power, working in close collaboration with CNAF