Key Project Drivers - FY10 Ruth Pordes, June 15th 2009.

Slides:



Advertisements
Similar presentations
 Contributing >30% of throughput to ATLAS and CMS in Worldwide LHC Computing Grid  Reliant on production and advanced networking from ESNET, LHCNET and.
Advertisements

May 9, 2008 Reorganization of the OSG Project The existing project organization chart was put in place at the beginning of It has worked very well.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Jan 2010 Current OSG Efforts and Status, Grid Deployment Board, Jan 12 th 2010 OSG has weekly Operations and Production Meetings including US ATLAS and.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
Assessment of Core Services provided to USLHC by OSG.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
Key Project Drivers - FY11 Ruth Pordes, June 15th 2010.
OSG Operations and Interoperations Rob Quick Open Science Grid Operations Center - Indiana University EGEE Operations Meeting Stockholm, Sweden - 14 June.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
INFSO-RI Enabling Grids for E-sciencE SA1: Cookbook (DSA1.7) Ian Bird CERN 18 January 2006.
GGF12 – 20 Sept LCG Incident Response Ian Neilson LCG Security Officer Grid Deployment Group CERN.
May 8, 20071/15 VO Services Project – Status Report Gabriele Garzoglio VO Services Project – Status Report Overview and Plans May 8, 2007 Computing Division,
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
Apr 30, 20081/11 VO Services Project – Stakeholders’ Meeting Gabriele Garzoglio VO Services Project Stakeholders’ Meeting Apr 30, 2008 Gabriele Garzoglio.
OSG Area Coordinators Meeting Proposal Chander Sehgal Fermilab
Discussion Topics DOE Program Managers and OSG Executive Team 2 nd June 2011 Associate Executive Director Currently planning for FY12 XD XSEDE Starting.
GridPP3 Project Management GridPP20 Sarah Pearce 11 March 2008.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Responsibilities of ROC and CIC in EGEE infrastructure A.Kryukov, SINP MSU, CIC Manager Yu.Lazin, IHEP, ROC Manager
GridPP Deployment & Operations GridPP has built a Computing Grid of more than 5,000 CPUs, with equipment based at many of the particle physics centres.
Key Project Drivers - an Update Ruth Pordes, June 14th 2008, V2: June 23 rd. These slides are in addition to the information available in
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
BNL Tier 1 Service Planning & Monitoring Bruce G. Gibbard GDB 5-6 August 2006.
Open Science Grid Open Science Grid: Beyond the Honeymoon Dane Skow Fermilab September 1, 2005.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks SA1: Grid Operations Maite Barroso (CERN)
INFSO-RI Enabling Grids for E-sciencE EGEE SA1 in EGEE-II – Overview Ian Bird IT Department CERN, Switzerland EGEE.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Ricardo Rocha CERN (IT/GS) EGEE’08, September 2008, Istanbul, TURKEY Experiment.
Status Organization Overview of Program of Work Education, Training It’s the People who make it happen & make it Work.
Report from the WLCG Operations and Tools TEG Maria Girone / CERN & Jeff Templon / NIKHEF WLCG Workshop, 19 th May 2012.
OSG Project Manager Report for OSG Council Meeting OSG Project Manager Report for OSG Council Meeting May 12, 2009 Chander Sehgal.
The OSG and Grid Operations Center Rob Quick Open Science Grid Operations Center - Indiana University ATLAS Tier 2-Tier 3 Meeting Bloomington, Indiana.
April 26, Executive Director Report Executive Board 4/26/07 Things under control Things out of control.
Jan 2010 OSG Update Grid Deployment Board, Feb 10 th 2010 Now having daily attendance at the WLCG daily operations meeting. Helping in ensuring tickets.
INFSO-RI Enabling Grids for E-sciencE An overview of EGEE operations & support procedures Jules Wolfrat SARA.
Operations Activity Doug Olson, LBNL Co-chair OSG Operations OSG Council Meeting 3 May 2005, Madison, WI.
CMS Usage of the Open Science Grid and the US Tier-2 Centers Ajit Mohapatra, University of Wisconsin, Madison (On Behalf of CMS Offline and Computing Projects)
LHC Computing, CERN, & Federated Identities
Sep 25, 20071/5 Grid Services Activities on Security Gabriele Garzoglio Grid Services Activities on Security Gabriele Garzoglio Computing Division, Fermilab.
Open Science Grid: Beyond the Honeymoon Dane Skow Fermilab October 25, 2005.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Monitoring of the LHC Computing Activities Key Results from the Services.
LCG Accounting Update John Gordon, CCLRC-RAL WLCG Workshop, CERN 24/1/2007 LCG.
OSG Report for DOE/NSF Joint Oversight Group U.S. Large Hadron Collider Program OSG Report for DOE/NSF Joint Oversight Group U.S. Large Hadron Collider.
IAG – Israel Academic Grid, EGEE and HEP in Israel Prof. David Horn Tel Aviv University.
Ian Bird Overview Board; CERN, 8 th March 2013 March 6, 2013
Why a Commercial Provider should Join the Academic Cloud Federation David Blundell Managing Director 100 Percent IT Ltd Simple, Flexible, Reliable.
1 Open Science Grid.. An introduction Ruth Pordes Fermilab.
WLCG Status Report Ian Bird Austrian Tier 2 Workshop 22 nd June, 2010.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
Grid Deployment Technical Working Groups: Middleware selection AAA,security Resource scheduling Operations User Support GDB Grid Deployment Resource planning,
OSG Area Coordinators Meeting Security Team Report Mine Altunay 8/15/2012.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Campus Grids Working Meeting Report Rob Gardner University of Chicago OSG All Hands March 10, 2010.
Grid Colombia Workshop with OSG Week 2 Startup Rob Gardner University of Chicago October 26, 2009.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Role and Challenges of the Resource Centre in the EGI Ecosystem Tiziana Ferrari,
Bob Jones EGEE Technical Director
Gene Oleynik, Head of Data Storage and Caching,
Key Project Drivers - FY10 Ruth Pordes, June 15th 2009
Regional Operations Centres Core infrastructure Centres
Open Science Grid Progress and Status
Ian Bird GDB Meeting CERN 9 September 2003
POW MND section.
Cloud Computing R&D Proposal
Leigh Grundhoefer Indiana University
LHC Data Analysis using a worldwide computing grid
Presentation transcript:

Key Project Drivers - FY10 Ruth Pordes, June 15th 2009

May 9, 2008 The top 10: Key Goals for FY10 Top 5: Support for LHC user running, support for Tier-1, Tier-2s and focus on Tier-3s. Support software, analyses, and data management for LIGO based on requests. Easier and more usable incremental upgrades of software for “any” reason – security, faults, functionality. Support for other OSG stakeholders requests and increased number of non-physics and campus beneficiaries in OSG. Timely and appropriate security and operational response to problems, as well as records of expectations, including SLAs and Policy, across the board. Next 5: Continued improved technical and operational support for data storage use, access and policies. Wider adoption of Pilot based workload management, progress in transparency between campus & wide area resources, policies for improved usability & efficiency. Articulation and implementation of Security authorization, identification & policy. Success of OSG Satellites for moving the infrastructure forward Better understanding of role of (separately) MPI, Virtual Machines and Cloud resources and policies in the OSG environment. 2

May 9, 2008 Internal Items needing to be Addressed Put things in your program of work that you think need doing without regard to the effort available. I would also prefer any holes you are worried about get included in the area plans rather than left to chance later. Meet responsibilities as part of Management of an effective project – update WBS, make reports, attend Area Coordinator meetings etc as expected. Area Metrics. Documentation. Integration of Training. Communication and Publishing. Planning for the Future of OSG. 3

May 9, 2008 Additional Info 4

May 9, LIGO Application needs:  Full support of and Inspiral Analysis across the majority of OSG sites.  Support for data movement and placement on OSG sites for LIGO applications.  Evaluation of another LIGO science application on the OSG. Middleware needs:  Support for native packaging and source distributions.  Strengthen connections to Condor and Pegasus in OSG support.  Integration of and support for LIGO security infrastructure. Service needs: Critical? YSecurity monitoring, incident response, notification and mitigation YAccounting -Integration of accounting with OSG accounting reports YIntegration and system validation of new and updated middleware. YTicket Handling YEffective Grid-wide job execution of and Inspiral Analysis YReporting of trends in usage, reliability, job state, job monitoring

May 9, WLCG US ATLAS and US CMS will be taking data from October  We must be communicative, responsive and flexible to needs on the ground – Tier-2, Tier-3 and Tier-1. WLCG will continue operations  We must continue to work closely with the US ATLAS and US CMS management to understand, react sensibly to and contribute to WLCG requirements and decisions.

May 9, US ATLAS and US CMS resource increase Summary of US ATLAS Tier2s %Increase CPU (kSI2K, HEP-SPEC06)6, ,47630,72420% Disk (Tbytes)2,4673,06725% Summary of US CMS Tier2s CPU(kSI2K, HEP-SPEC06)7, ,00054,00076% Disk (Tbytes)2, % ATLAS BNL Tier-1 CPU(kSI2K, HEP-SPEC06)7, ,34851,06074% Disk (Tbytes)5,82211,637100% Tape (Tbytes)3,2776,28692% CMS FNAL Tier-1 CPU(kSI2K, HEP-SPEC06)5,100 – 20,40044,400117% Disk (Tbytes)2,6004,10057% Tape (Tbytes)7,10011,00055% From WLCG MOU. (says updates to US CMS T2 numbers will be made after initial data taking.)

May 9, 2008 From review.. Continued Challenges:  Scientific productivity argument could be presented in a more compelling manner.  Can’t rely on interaction with VOs to effectively represent what’s happening in end user community.  Didn’t tell OSG’s excellent story in a cohesive way. Accomplishments:  OSG is a very critical piece of the infrastructure for the LHC and potentially important for LIGO.  This initiative has significant potential value for other large-scale science endeavors.  Clear that model for incorporating new HW resources can be successful – i.e. enabled CDF resources for D0 re-processing. 8

May 9, 2008 Futures Planning – proposed & status Phase 1 by June 30, 2009 – delayed by a few weeks Gather and document requirements & expectations from major stakeholders for 2010 to 2015 Solicit guidance from OSG Council on key directions for future Phase 2 by Aug 2009 Face-to-Face Council Mtg – 2-4 page plan overall – not Each 2-page plan for each OSG work area (or functional unit) List/abstract of satellite proposals Phase 3 by Dec 2009 Analysis => outline for proposal Document Architecture Identify particpating senior personnel (and institutions) Phase 4 at March 2010 All Hands Meeting Endorsement of proposal by stakeholders OSG Future proposal to NSF/DOE by March 30,

May 9, WLCG MOU Goals - OSG support for LHC Tier-2s continues provision of managed disk storage providing permanent and/or temporary data storage for files and databases; provision of access to the stored data by other centres of the WLCG operation of an end-user analysis facility provision of other services, e.g. simulation, according to agreed Experiment requirements; ensure network bandwidth and services for data exchange with Tier1 Centres, as part of an overall plan agreed between the Experiments and the Tier1 Centres concerned. All storage and computational services shall be “grid enabled” according to standards agreed between the LHC Experiments and the regional centres. ServiceMaximum delay in responding to operational problems Average availability measured on an annual basis Prime timeOther periods End-user analysis facility2 hours72 hours95% Other services12 hours72 hours95%

May 9, WLCG MOU - OSG provides a Grid Operations Center Annex ハ 3.4.Grid Operations Services This section lists services required for the operation and management of the grid for LHC computing. This section reflects the current (September 2005) state of experience with operating grids for high energy physics. It will be refined as experience is gained. Grid Operations Centres – Responsible for maintaining configuration databases, operating the monitoring infrastructure, pro-active fault and performance monitoring, provision of accounting information, and other services that may be agreed. Each Grid Operations Centre shall be responsible for providing a defined sub-set of services, agreed by the WLCG Collaboration. Some of these services may be limited to a specific region or period (e.g. prime shift support in the country where the centre is located). Centres may share responsibility for operations as agreed from time to time by the WLCG Collaboration. User Support for grid and computing service operations:  First level (end-user) helpdesks are assumed to be provided by LHC Experiments and/or national or regional centres, and are not covered by this MoU.  Grid Call Centres – Provide second level support for grid-related problems, including pro-active problem management. These centres would normally support only service staff from other centres and expert users. Each call centre shall be responsible for the support of a defined set of users and regional centres and shall provide coverage during specific hours.

May 9, WLCG Operations Centers in the US Indiana University iGOC Scope of the serviceOpen Science Grid Operations Centre Period during which the centre operates as the primary monitoring centre 24  7  52 BNL, Fermilab Scope of the serviceUS-ATLAS and US-CMS Virtual Organisation Support Centre respectively Period during which the centre operates as the primary monitoring centre 24  7  52

May 9, US LHC Service Needs CriticalInterface to WLCG? YSecurity monitoring, incident response, notification and mitigationCollaboration with EGEE, WLCG YAccounting - CPU, Storage & EfficienciesYes YGOC BDII Information System with accurate information published by all OSG sites that support US CMS VO. Reliability publish accurate information to WLCG BDII YUS ATLAS specific accounting reportsNo YReliability and Availability monitoringYes YIntegration and system validation of new and updated middleware.Test interoperation of new releases with EGEE NUser/VO monitoring and validation using the RSV infrastructurePerhaps YTicket HandlingBi-directional:US LHC OSG GGUS, including alarms. YSRM V2.2 Storage at Tier-2sTrack WLCG deployments. YCE interface to meet experiment throughput needs YReporting of trends in usage, reliability, job state, job monitoring. YGrid wide Information system accessible to ATLAS applications.No NTroubleshooting and user support, especially support from centralized expert group for use of storage on OSG sites YSRM V2.2 Storage at Tier-2sTrack WLCG deployments. YCE interface to meet CMS throughput needs based on GlideinWMS workload management. YReporting of trends in usage, reliability, job state, job monitoring. Site level dashboard of usage, job state, efficiencies and errors across US CMS OSG sites.

May 9, Run II Needs Continued support for opportunistic use of OSG resources for simulation needs. Maintain deployed middleware and services compatible with existing experiment software. Throughput  DZero : 5 M events/week  CDF: 5 M events/week?

May 9, STAR Support for xrootd on OSG sites. Support for VMs and use of Commercial Clouds.