Collaboration Meeting

Slides:



Advertisements
Similar presentations
EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
Advertisements

 Contributing >30% of throughput to ATLAS and CMS in Worldwide LHC Computing Grid  Reliant on production and advanced networking from ESNET, LHCNET and.
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Les Les Robertson WLCG Project Leader WLCG – Worldwide LHC Computing Grid Where we are now & the Challenges of Real Data CHEP 2007 Victoria BC 3 September.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
Stefano Belforte INFN Trieste 1 CMS SC4 etc. July 5, 2006 CMS Service Challenge 4 and beyond.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
LHCC Comprehensive Review – September WLCG Commissioning Schedule Still an ambitious programme ahead Still an ambitious programme ahead Timely testing.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
LCG and HEPiX Ian Bird LCG Project - CERN HEPiX - FNAL 25-Oct-2002.
DOSAR Workshop, Sao Paulo, Brazil, September 16-17, 2005 LCG Tier 2 and DOSAR Pat Skubic OU.
John Gordon STFC-RAL Tier1 Status 9 th July, 2008 Grid Deployment Board.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.
LCG LHC Computing Grid Project – LCG CERN – European Organisation for Nuclear Research Geneva, Switzerland LCG LHCC Comprehensive.
LCG Service Challenges: Planning for Tier2 Sites Update for HEPiX meeting Jamie Shiers IT-GD, CERN.
Tony Doyle - University of Glasgow 8 July 2005Collaboration Board Meeting GridPP Report Tony Doyle.
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Dr Ian Bird LCG Project Leader 6 th October 2009 Telecom 2009 Youth Forum.
ESFRI & e-Infrastructure Collaborations, EGEE’09 Krzysztof Wrona September 21 st, 2009 European XFEL.
1 LHCb on the Grid Raja Nandakumar (with contributions from Greig Cowan) ‏ GridPP21 3 rd September 2008.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Plans for Service Challenge 3 Ian Bird LHCC Referees Meeting 27 th June 2005.
U.S. ATLAS Computing Facilities Overview Bruce G. Gibbard Brookhaven National Laboratory U.S. LHC Software and Computing Review Brookhaven National Laboratory.
LCG Accounting Update John Gordon, CCLRC-RAL WLCG Workshop, CERN 24/1/2007 LCG.
LHCC Referees Meeting – 28 June LCG-2 Data Management Planning Ian Bird LHCC Referees Meeting 28 th June 2004.
WLCG Status Report Ian Bird Austrian Tier 2 Workshop 22 nd June, 2010.
Status of gLite-3.0 deployment and uptake Ian Bird CERN IT LCG-LHCC Referees Meeting 29 th January 2007.
ARDA Massimo Lamanna / CERN Massimo Lamanna 2 TOC ARDA Workshop Post-workshop activities Milestones (already shown in December)
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Status of GSDC, KISTI Sang-Un Ahn, for the GSDC Tier-1 Team
ALICE Physics Data Challenge ’05 and LCG Service Challenge 3 Latchezar Betev / ALICE Geneva, 6 April 2005 LCG Storage Management Workshop.
HEPiX IPv6 Working Group David Kelsey (STFC-RAL) GridPP33 Ambleside 22 Aug 2014.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
LHC collisions rate: Hz New PHYSICS rate: Hz Event selection: 1 in 10,000,000,000,000 Signal/Noise: Raw Data volumes produced.
1-2 March 2006 P. Capiluppi INFN Tier1 for the LHC Experiments: ALICE, ATLAS, CMS, LHCb.
ScotGRID is the Scottish prototype Tier 2 Centre for LHCb and ATLAS computing resources. It uses a novel distributed architecture and cutting-edge technology,
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
JRA1 Middleware re-engineering
The CMS Experiment at LHC
Bob Jones EGEE Technical Director
Workload Management Workpackage
(Prague, March 2009) Andrey Y Shevel
LCG Service Challenge: Planning and Milestones
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
Grid site as a tool for data processing and data analysis
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
SA1 Execution Plan Status and Issues
Ian Bird GDB Meeting CERN 9 September 2003
POW MND section.
Data Challenge with the Grid in ATLAS
Update on Plan for KISTI-GSDC
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Readiness of ATLAS Computing - A personal view
Olof Bärring LCG-LHCC Review, 22nd September 2008
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
LCG middleware and LHC experiments ARDA project
Connecting the European Grid Infrastructure to Research Communities
LCG experience in Integrating Grid Toolkits
Ian Bird LCG Project - CERN HEPiX - FNAL 25-Oct-2002
LHC Data Analysis using a worldwide computing grid
Collaboration Board Meeting
Future EU Grid Projects
ATLAS DC2 & Continuous production
LHCb thinking on Regional Centres and Related activities (GRIDs)
gLite The EGEE Middleware Distribution
The LHC Computing Grid Visit of Professor Andreas Demetriou
The LHCb Computing Data Challenge DC06
Presentation transcript:

Collaboration Meeting GridPP Overview Tony Doyle 11 January 2006 Collaboration Meeting

OR The Icemen Cometh

Collaboration Meeting Outline High level LHC, Expt. and Grid Plans.. 2005 Outturn and the Goldilocks Problem.. “Some of the challenges for next year” gLite is not too late.. Pulling together on the good ship Grid Coming together to enable a discovery Plans and Resolutions (from the year of the rooster to the year of the dog..) 11 January 2006 Collaboration Meeting

Collaboration Meeting A. When Cometh the LHC? “Main objectives are to terminate installation in February 2007 and enable first collisions in summer 2007” Lyn Evans 11 January 2006 Collaboration Meeting

B. When Cometh the Detectors? e.g. ATLAS “with good will and great efforts from everybody we can be confident that the Technical Coordination Team will manage to have ATLAS installed by June 2007” 11 January 2006 Collaboration Meeting

Collaboration Meeting C. When Cometh the Grid? Service Challenges – UK deployment plans End point April ’07 Context: The virtual LHC Computing Centre 11 January 2006 Collaboration Meeting

Collaboration Meeting Grid Overview Aim: by 2008 (full year’s data taking) CPU ~100MSi2k (100,000 CPUs) Storage ~80PB - Involving >100 institutes worldwide Build on complex middleware being developed in advanced Grid technology projects, both in Europe (Glite) and in the USA (VDT) Prototype went live in September 2003 in 12 countries Extensively tested by the LHC experiments in September 2004 197 sites, 13,797 CPUs, 5PB storage in September 2005 11 January 2006 Collaboration Meeting

2005 Metrics and Quality Assurance: wider concerns Target Current status Q2 2006 Target values Number of Users ~ 1000 ≥ 3000 Number of sites 120 50 Number of CPU ~12000 9500 at month 15 Number of Disciplines 6 ≥ 5 Multinational 24 ≥ 15 countries 11 January 2006 Collaboration Meeting

Collaboration Meeting 2005 Tier-1 GOC Accounting 11 January 2006 Collaboration Meeting

2005 Grid and Non-Grid Tier-1/A CPU Use Fraction ~50% 11 January 2006 Collaboration Meeting

2005 Grid and Non-Grid Tier-1/A CPU Use OC user Users! 70% “Target” OC! 11 January 2006 Collaboration Meeting

Collaboration Meeting 2005 Tier-1 Efficiency overall efficiency = S (CPU times) / S (wall times) Jan ’05 Pre-Grid Efficiency High Apr ’05 Remote data access problems Dec ’05 General Improvements Mixed view from “VOs” Need to test data access I/O-bound jobs possible Good to have this data! 11 January 2006 Collaboration Meeting

Some of the challenges for next year See Jamie’s talk Castor 2 Good progress, rapid reaction to problems But some way still to go with testing - stressing reliability, performance Can only be done with participation of experiments Distribution to other sites being planned Distributed database services Architecture and plan agreed now But still to deploy pilot services - timing is worryingly tight End-to-end testing of the DAQ-T0-T1 chain recording, calibration and alignment, reconstruction, distribution Full Tier-1 work load testing – Recording, reprocessing, ESD distribution, analysis, Tier-2 support Understanding the CERN Analysis Facility batch analysis interactive analysis Startup scenarios Schedule may be better known after next spring’s Chamonix meeting

Collaboration Meeting High Level View from recent LHCC review is ~ Service Challenges - OK (established) Throughput - not OK (not sufficiently tested) Baseline services – OK (defined, not completely established) Practical steps – OK (we need to improve communication) MoUs – OK (we need to sign off, covered last meting) Concerns: Significant delay in gLite, Castor2, distributed data management, database services all are late Middleware and experiment connections are too weak Analysis models are untested 11 January 2006 Collaboration Meeting

Collaboration Meeting gLite Stack Requirement Storage Element Basic File Transfer Reliable File Transfer Catalogue Services Data Management tools Compute Element Workload Management VO Agents VO Membership Services DataBase Services Posix-like I/O Application Software Installation Tools Job Monitoring Reliable Messaging Information System 15 Baseline Services for a functional Grid We rely upon gLite components This middleware builds upon VDT (Globus and Condor) and meets the requirements of all the basic scientific use cases: Green (amber) areas are (almost) agreed as part of the shared generic middleware stack by each of the application areas Red are areas where generic middleware competes with application-specific software. www.glite.org 11 January 2006 Collaboration Meeting

Collaboration Meeting gLite Pack 11 January 2006 Collaboration Meeting

Middleware Re-engineering A series of gLite releases have been produced (1.1, 1.2, 1.3, and 1.4) Driven by application and deployment needs Focus on defect fixing gLite deployed on a Pre-Production Service and made available for application use Independent evaluation by NGS gLite components also available via VDT (US) gLite components deployed on the infrastructure Emphasis is now on release of gLite 1.5 Will continue… see Steve Fisher’s talk EGEE phase 2 starts in April 2006 www.glite.org 11 January 2006 Collaboration Meeting

Some of the challenges for next year File transfers Good initial progress (except dCache->DPM- currently) But some way still to go with testing - stressing reliability, performance Can only be done with participation of experiments Distribution to other sites being planned Distributed VO services Plan agreed – T1 will sign off and then VO boxes may be deployed by T2s But still to deploy pilot services - CMS (OK) LHCb (OK) ATLAS ALICE End-to-end testing of the T0-T1-T2 chain MC production, reconstruction, distribution Full Tier-1 work load testing Recording, reprocessing, ESD distribution, analysis, Tier-2 support Understanding the “Analysis Facility” batch analysis @ T1 and T2 interactive analysis Startup scenarios Schedule is known at high level and defined for Service Challenges – testing time ahead (in many ways) 11 January 2006 Collaboration Meeting

Collaboration Meeting Themes for 2006 Think how you can help by either: measuring throughput for experiments; testing gLite, providing feedback; working practically with experiments; running your analysis on the Grid; helping Grid adoption (in other fields). OR combine elements of 1-5. Emphasis on end-to-end (vertical) integration. 11 January 2006 Collaboration Meeting

Pulling together with the experiments? Hopefully the effort in pulling the Grid boat out is more equal.. However many discoveries made in Grid circles are currently being re-discovered by experiment users Succinct user documentation will help 11 January 2006 Collaboration Meeting

Pulling together with the experiments? There are currently not enough users in the Grid School for the Gifted Having smart users helps (the current ones are) The system may be too complex, requiring too much work by the user? Or the (virtual) help desk may not be enough? Or the documentation may be misleading? 11 January 2006 Collaboration Meeting

2005: Functional Tests 2006: File Transfers number of sites passing the SFT tests Log data lost total grid sites Successful Year of Functional Tests, with bar raised throughout the year. “Functional” ≠ “Performant”. Need to test network-file transfers -file placement-file collection transfers, working with experiments. 11 January 2006 Collaboration Meeting

Collaboration Meeting Come together.. Physics discovery requires many elements to work.. The Icemen Cometh 2009 2008 2007 2008-09 “Physics discovery” 2006-07 “Performance Testing” 2004-05 “Functional Testing” 2006 2005 2004 11 January 2006 Collaboration Meeting

Collaboration Meeting A (Light) Summary Priors indicate a Higgs mass 114.4 < mH [GeV] <219 The Grid Service will be launched in April 2007 The Detectors will be complete in June 2007 The LHC will provide first collisions in Summer 2007 These will enable data analyses such that the Higgs will be discovered on May 29th 2009 probably… large corrections If the Higgs particle is discovered, the Grid will be one of three major components 11 January 2006 Collaboration Meeting

Collaboration Meeting 2005: The Year of the Tier-1 “World's biggest grid seeks secrets of the universe” “It's in Didcot and it's running on open source” Silicon.Com Published: Thursday 24 November 2005 http://www.silicon.com/publicsector/0,3800010403,39154492,00.htm 11 January 2006 Collaboration Meeting

Collaboration Meeting Summary This meeting focuses on the challenges of a New Year The Old Year was the year of the (Tier-1) rooster Tier-1/A utilisation was too low (at the start) and too high (in the end), but just right overall? Tier-2 utilisation was too low (throughout) The New Year is the year of the (Tier-2) dog? The seamless vision of a T0-T1-T2 structure hidden behind a transparent Grid requires (relatively rapid) testing The New Year will be better, if we resolve to make (Many) T1 T2, Grid Expt. Performance Tests we have (but they must all ultimately be successful) 11 January 2006 Collaboration Meeting