Status of CMS-HI Compute Proposal for USDOE

Slides:



Advertisements
Similar presentations
Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
Advertisements

Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
DATA PRESERVATION IN ALICE FEDERICO CARMINATI. MOTIVATION ALICE is a 150 M CHF investment by a large scientific community The ALICE data is unique and.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
16 October 2005 Collaboration Meeting1 Computing Issues & Status L. Pinsky Computing Coordinator ALICE-USA.
CMS STEP09 C. Charlot / LLR LCG-DIR 19/06/2009. Réunion LCG-France, 19/06/2009 C.Charlot STEP09: scale tests STEP09 was: A series of tests, not an integrated.
Tier 1 Facility Status and Current Activities Rich Baker Brookhaven National Laboratory NSF/DOE Review of ATLAS Computing June 20, 2002.
PPDG and ATLAS Particle Physics Data Grid Ed May - ANL ATLAS Software Week LBNL May 12, 2000.
Status of 2015 pledges 2016 requests RRB Report Concezio Bozzi INFN Ferrara LHCb NCB, November 3 rd 2014.
LHC Computing Review Recommendations John Harvey CERN/EP March 28 th, th LHCb Software Week.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
1. Maria Girone, CERN  Q WLCG Resource Utilization  Commissioning the HLT for data reprocessing and MC production  Preparing for Run II  Data.
7April 2000F Harris LHCb Software Workshop 1 LHCb planning on EU GRID activities (for discussion) F Harris.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
STAR Off-line Computing Capabilities at LBNL/NERSC Doug Olson, LBNL STAR Collaboration Meeting 2 August 1999, BNL.
BNL Wide Area Data Transfer for RHIC & ATLAS: Experience and Plans Bruce G. Gibbard CHEP 2006 Mumbai, India.
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National Laboratory.
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory Review of U.S. LHC Software and Computing Projects Fermi National Laboratory November.
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
23.March 2004Bernd Panzer-Steindel, CERN/IT1 LCG Workshop Computing Fabric.
Predrag Buncic Future IT challenges for ALICE Technical Workshop November 6, 2015.
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
U.S. ATLAS Computing Facilities Overview Bruce G. Gibbard Brookhaven National Laboratory U.S. LHC Software and Computing Review Brookhaven National Laboratory.
The ATLAS Computing Model and USATLAS Tier-2/Tier-3 Meeting Shawn McKee University of Michigan Joint Techs, FNAL July 16 th, 2007.
US ATLAS Tier 1 Facility Rich Baker Deputy Director US ATLAS Computing Facilities October 26, 2000.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
GDB, 07/06/06 CMS Centre Roles à CMS data hierarchy: n RAW (1.5/2MB) -> RECO (0.2/0.4MB) -> AOD (50kB)-> TAG à Tier-0 role: n First-pass.
LHCb Current Understanding of Italian Tier-n Centres Domenico Galli, Umberto Marconi Roma, January 23, 2001.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
Hall D Computing Facilities Ian Bird 16 March 2001.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
The CMS-HI Computing Plan Vanderbilt University
Ian Bird WLCG Workshop San Francisco, 8th October 2016
The Beijing Tier 2: status and plans
The Vanderbilt Effort in CMS Vanderbilt University
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
U.S. ATLAS Tier 2 Computing Center
CMS-HI Offline Computing
The Vanderbilt Effort in CMS Vanderbilt University
Charles Maguire for VU-RHIC group
Data Challenge with the Grid in ATLAS
The CMS-HI Computing Plan Vanderbilt University
Vanderbilt Tier 2 Project
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
ALICE Computing Model in Run3
Bernd Panzer-Steindel CERN/IT
New strategies of the LHC experiments to meet
Scientific Computing At Jefferson Lab
Preparations for the CMS-HI Computing Workshop in Bologna
Near Real Time Reconstruction of PHENIX Run7 Minimum Bias Data From RHIC Project Goals Reconstruct 10% of PHENIX min bias data from the RHIC Run7 (Spring.
Vanderbilt University
Nuclear Physics Data Management Needs Bruce G. Gibbard
Proposal to BNL/DOE to Use ACCRE Farm for PHENIX Real Data Reconstruction in Charles Maguire for the VU PHENIX group Carie Kennedy, Paul Sheldon,
Preparations for Reconstruction of Run6 Level2 Filtered PRDFs at Vanderbilt’s ACCRE Farm Charles Maguire et al. March 14, 2006 Local Group Meeting.
Collaboration Board Meeting
Status of CMS-HI Compute Proposal for USDOE
Heavy Ion Physics Program of CMS Proposal for Offline Computing
Heavy Ion Physics Program of CMS Proposal for Offline Computing
CMS-HI Offline Computing
Preparations for Reconstruction of Run7 Min Bias PRDFs at Vanderbilt’s ACCRE Farm (more substantial update set for next week) Charles Maguire et al. March.
ATLAS DC2 & Continuous production
The ATLAS Computing Model
LHCb thinking on Regional Centres and Related activities (GRIDs)
Development of LHCb Computing Model F Harris
Expanding the PHENIX Reconstruction Universe
Presentation transcript:

Status of CMS-HI Compute Proposal for USDOE Charles F. Maguire (Vanderbilt University) for the CMS-HI Institutions Version 1, 15:10 CDT on July 1 July 2, 2008 CMS CRB Meeting

HI Institutions in CMS US (all U.S.D.O.E-NP except UC Davis which is NSF): Colorado, Iowa, Kansas, LANL, Maryland, Minnesota, MIT, Vanderbilt, UC Davis, UI Chicago Nuclear Physics division of DOE separate from HEP Non-US: Athens, Auckland, Budapest, CERN, Chongbuk, Cukurova, Korea U, Lisbon, Lyon, Moscow, Mumbai, Seoul, Zagreb, Paris? Overall ~100 people (60 PhDs, 40 students) July 2, 2008 CMS CRB Meeting

CMS-HI Status for the US DOE Certain US groups have been receiving core grant support for CMS-HI simulations, R&D: research scientists, postdocs, students, travel, computing DOE Office of Nuclear Physics has launched its support for the LHC HI program in 2007 LHC physics is now part of the long term RHI/NP plan in US Both the ALICE and the CMS experiments are recognized US ALICE Construction of EM Calorimeter for ALICE, ~13 M$ project, CD-1 signed ~10 institutions interested, including LBL, LLNL,ORNL ~50-70 people US CMS HLT farm, ~2M$ project + ZDC ~0.5 M$ + Computing (to be funded) ~10 institutions interested, including LANL ~50 people July 2, 2008 CMS CRB Meeting

Status of CMS-HI Groups in US CMS proposal to DOE-NP review in October 2006 CPUs for online farm in FY08-FY10 (consistent with DAQ/Cittolin plans) ~2M$: Prototype HLT farm part of CMS DAQ for purchase now Review request: “do more studies”. Launched a ~1 year physics project to simulate jet+g in HI events plus complete revamp of HI software Operating funds Cat-A + travel starting in FY07 and FY08 (16 PhD) Review called for a formal CMS-HI computing proposal Review follow-up activity Response to review recommendations almost concluded Computing proposal is being completed now Note: the “turn-on” RHIC>CMS is managed very carefully by the DOE-NP, and not all groups can start now Managed via renewals and supplements of individual groups e.g. Maryland, Vanderbilt July 2, 2008 CMS CRB Meeting

Internal Decisions about Computing CMS-HI Physics Will Need a Dedicated Compute Center Different production and analysis schedules from HEP program Different physics goals US CMS-HI Institutions Reviewed Compute Bids Reviews occurred during late 2007 and early 2008 Used external (RHIC) consultants to comment on 3 bids Conclusion: Vanderbilt University will be the lead institution on the proposal to DOE Main computer center to be situated on Vanderbilt Campus Charlie Maguire as the Principal Investigator Some fraction of computers and disk space will go to MIT All other institutions will have excellent network connections This includes the overseas CMS-HI institutions July 2, 2008 CMS CRB Meeting

Reasons for Vanderbilt Choice Local group has solid experience in RHI computing Responsible for simulation of PHENIX since 1992 Remote nearly-real-time reconstruction of PHENIX data at Vanderbilt in 2007 (30 TBytes input raw data, 20 TBytes output, see slides 7 and 8) Vanderbilt has strong history in RHI Physics at RHIC Good physics judgment on priorities and to prepare for data Vanderbilt group will work together with the university computing staff (10 persons at ACCRE facility) Large existing computational facility for all of Vanderbilt Strong interaction with local HEP group in CMS (Paul Sheldon, Will Johns, and Dan Engh) July 2, 2008 CMS CRB Meeting

ACCRE Compute Facility at Vanderbilt ACCRE Advanced Computing Center for Research and Education www.accre.vanderbilt.edu $8.5 M start-up grant from VU $1.5 M additional NSF funding Currently has ~1500 CPUs, may grow to ~3000 CPUs in 5 years exclusive of CMS-HI purchases Serves a wide spectrum of university research including especially medical applications July 2, 2008 CMS CRB Meeting

Real RHI Data Reconstruction at ACCRE PHENIX Raw Data (2007) Reconstruction Project RHIC->Vanderbilt->RCF Near-real time reconstruction effort, ~few days latency from calibration processes at local RHIC buffers 30 TBytes of raw data transferred to ACCRE during 6 weeks (no taping) 20 TBytes of reconstruction output returned to RCF for PHENIX users Highly automated assembly line procedures, used ~100 PERL scripts including web-based flow monitoring PHENIX 2007 data volumes are comparable to what is expected for first year of CMS-HI production Near real-time aspect not a factor for CMS-HI, but tape archiving will be July 2, 2008 CMS CRB Meeting

Overview of CMS-HI Compute Proposal Bring RAW data to Vanderbilt, archive to tape Expecting 300 TBytes of data transfers when nominal luminosity is achieved Use of CERN Tier0 for immediate calibrations and limited reco Do real data reconstruction with preparation for two passes per year MC production at MIT and other places Distribute AODs to all members of CMS HI group around the world AODs will be processed at Vanderbilt for most US CMS-HI groups The USDOE-NP is highly influenced by the RCF model at RHIC We will use all the CMS tools: CRAB, PhEdEx, PAT, ... Prepare by using existing centers this summer to exercise CMSSW software Build up to about 3000 CPUs (VU+MIT) over ~5 years, starting next year (US FY’09) The CMS-HI compute center will be part of a larger Vanderbilt University computing center (ACCRE) Many synergies, research projects that can help us at VU (REDDnet) Possibility for “opportunistic computing” beyond the 3000 CPU allocation July 2, 2008 CMS CRB Meeting

Input to the Draft Proposal Proposal is using information about timing and data reconstruction from HLT and g-jet studies (USDOE review “jet challenge”) Some 3000 CPUs running for 12 months (reco, and real+MC analysis) Total CPU power = 4.8MSI2K, about 10% of CMS-HEP Tier1+Tier2 power ~400 TB of disk storage Tape archive (1.6 PBytes over 5 years) 10 Gbps network connection to CERN and other places (e.g. MIT) Good and debugged connections to all CMS HI institutions July 2, 2008 CMS CRB Meeting

Network Issue to Be Resolved Action Item from USDOE-NP/Esnet Worshop USDOE-NP and ESnet Organization Workshop was held in May 2008 Purpose was to review WAN and LAN projected needs for the next 5 years Major US nuclear physics experiments were requested to give case studies RACF (RHIC and ATLAS Computing Facility) for PHENIX and STAR at RHIC LBNL for US-ALICE Vanderbilt University for US-CMS-HI JLAB (Jefferson Laboratory) for CEBAF Action Item Affecting Both US-ALICE and US-CMS-HI Possibility that the LHCnet capabilities would be saturated by HEP needs CMS-HI forecasts 300 TBytes to be transported in one month (100% of CMS-HI raw data) US-ALICE forecasts 100 TBytes to be transported in four months (10% of ALICE raw data) LHCnet is a US-HEP priority Related issue of US Congress mandated cap on HEP expenditures for the LHC What alternative paths exists to LHCnet for trans-Atlantic transport of data? These alternative links are being investigated by the staff of the US Internet2 organization The costs and consequences of the alternative links should be well understood by all sides Both US-ALICE and US-CMS-HI must include such studies in their respective compute proposals July 2, 2008 CMS CRB Meeting

Computing at Present for CMS-HI The CMS-HI compute proposal will take additional time and effort to get approved by the USDOE First draft is almost completed Discussions have taken place with the FNAL Tier1 experts Preliminary version is now available to CMS Tier0 experts Need to get the proposal to the DOE as soon as practical Proposal will be scrutinized by external DOE reviewers Earliest funding would be after FY’09 starts (October 2008) In the interim MIT continues computing for CMS-HI The HI-Tier-3 is living in symbiosis with HEP-Tier-2 There are ~130 CPU, 30 TB of disk dedicated to HI, 1GB/core, to be upgraded to 2GB/core shortly There is the possibility of opportunistic access to >1600 CPUs of the Tier-2 July 2, 2008 CMS CRB Meeting

Future Computing for CMS-HI Discussions at CMS-HI June Meeting CERN Tier0 Moscow Budapest MIT Seoul Vanderbilt Paris… Others? July 2, 2008 CMS CRB Meeting

Summary A Dedicated CMS-HI Compute Facility is to be Proposed in the US Facility at Vanderbilt will function as a combined Tier1/Tier2 Receipt and archiving to tape of the raw data from Tier0, ~300 TBytes/year Reconstruction of raw data into RECO and AOD files Processing of AOD files for analysis for US (and others) users Distribution of some AOD files to overseas CMS-HI facilities Approximate 25% of the CPUs and disk resources to be put at MIT Continues MIT’s role as MC producer for CMS-HI Retains and expands expertise of RHI group at MIT in CMSSW Proposal Time Scale (www.hep.vanderbilt.edu/~maguirc/CMS-HI/cmsHIComputingProposal.pdf) First draft should be completed within 2 weeks Distribution to CMS computing experts for their comments Revised draft should be submitted to the USDOE-NP as soon as practical Integration of CMS-HI Compute Facility with Rest of CMS Computing CMS-HI compute facility should function as much as possible like other CMS Tier1 and Tier2 facilities CMS-HI institutions must take advantage of all of the developments and tools available at the other CMS compute facilities July 2, 2008 CMS CRB Meeting