LQCD Project Overview Don Holmgren LQCD Project Progress Review May 25-26, 2006 Fermilab.

Slides:



Advertisements
Similar presentations
I/O and the SciDAC Software API Robert Edwards U.S. SciDAC Software Coordinating Committee May 2, 2003.
Advertisements

SciDAC Software Infrastructure for Lattice Gauge Theory
Ninth Lecture Hour 8:30 – 9:20 pm, Thursday, September 13
Nuclear Physics in the SciDAC Era Robert Edwards Jefferson Lab SciDAC 2009 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this.
4/2/2002HEP Globus Testing Request - Jae Yu x Participating in Globus Test-bed Activity for DØGrid UTA HEP group is playing a leading role in establishing.
Dynamical Chiral Fermions The `Grail’ – dyn. chiral fermions Generation of dyn. chiral fermions configs –RBC on the RIKEN QCDOC – Jan 05 (some %) –UKQCD.
1 Lattice QCD Now and in > 5 Years Aida X. El-Khadra (UIUC) Beauty 2003, Oct 14-18, 2003.
Effective Methods for Software and Systems Integration
Jefferson Lab Status Hall A collaboration Dec. 16, 2013 R. D. McKeown Deputy Director For Science.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
Dr. Tom WayCSC What is Software Engineering? CSC 4700 Software Engineering Lecture 1.
The Lattice Initiative at Jefferson Lab Robert Edwards Jefferson Lab JLab in a close partnership with MIT has formed the Lattice Hadron Physics Collaboration.
SciDAC Software Infrastructure for Lattice Gauge Theory Richard C. Brower Annual Progress Review JLab, May 14, 2007 Code distribution see
Simulating Quarks and Gluons with Quantum Chromodynamics February 10, CS635 Parallel Computer Architecture. Mahantesh Halappanavar.
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
QCD Project Overview Ying Zhang September 26, 2005.
SC4 Workshop Outline (Strong overlap with POW!) 1.Get data rates at all Tier1s up to MoU Values Recent re-run shows the way! (More on next slides…) 2.Re-deploy.
BROOKHAVEN SCIENCE ASSOCIATES National Synchrotron Light Source II Project Management Jim Yeck Deputy Director (Project Management)
Lattice QCD in Nuclear Physics Robert Edwards Jefferson Lab CCP 2011 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.:
Physics Steven Gottlieb, NCSA/Indiana University Lattice QCD: focus on one area I understand well. A central aim of calculations using lattice QCD is to.
LQCD Clusters at JLab Chip Watson Jie Chen, Robert Edwards Ying Chen, Walt Akers Jefferson Lab.
R. Ryne, NUG mtg: Page 1 High Energy Physics Greenbook Presentation Robert D. Ryne Lawrence Berkeley National Laboratory NERSC User Group Meeting.
May 8, 20071/15 VO Services Project – Status Report Gabriele Garzoglio VO Services Project – Status Report Overview and Plans May 8, 2007 Computing Division,
Scientific Computing Experimental Physics Lattice QCD Sandy Philpott May 20, 2011 IT Internal Review 12GeV Readiness.
HEPiX Karlsruhe May 9-13, 2005 Operated by the Southeastern Universities Research Association for the U.S. Department of Energy Thomas Jefferson National.
Tier 1 Facility Status and Current Activities Rich Baker Brookhaven National Laboratory NSF/DOE Review of ATLAS Computing June 20, 2002.
1 Metrics for the Office of Science HPC Centers Jonathan Carter User Services Group Lead NERSC User Group Meeting June 12, 2006.
Lattice QCD and the SciDAC-2 LQCD Computing Project Lattice QCD Workflow Workshop Fermilab, December 18, 2006 Don Holmgren,
ILDG Middleware Status Chip Watson ILDG-6 Workshop May 12, 2005.
LQCD Project Review Response Germantown, MD Aug 8, 2005.
Proposed 2007 Acquisition Don Holmgren LQCD Project Progress Review May 25-26, 2006 Fermilab.
09/02 ID099-1 September 9, 2002Grid Technology Panel Patrick Dreher Technical Panel Discussion: Progress in Developing a Web Services Data Analysis Grid.
U.S. ATLAS Tier 1 Planning Rich Baker Brookhaven National Laboratory US ATLAS Computing Advisory Panel Meeting Argonne National Laboratory October 30-31,
High Energy Nuclear Physics and the Nature of Matter Outstanding questions about strongly interacting matter: How does matter behave at very high temperature.
May 25-26, 2006 LQCD Computing Review1 Jefferson Lab 2006 LQCD Analysis Cluster Chip Watson Jefferson Lab, High Performance Computing.
Lattice QCD and GPU-s Robert Edwards, Theory Group Chip Watson, HPC & CIO Jie Chen & Balint Joo, HPC Jefferson Lab TexPoint fonts used in EMF. Read the.
SciDAC Software Infrastructure for Lattice Gauge Theory Richard C. Brower QCD Project Review May 24-25, 2005 Code distribution see
LQCD Computing Review Bakul Banerjee Fermi National Accelerator Laboratory May 24, 2005.
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory Review of U.S. LHC Software and Computing Projects Fermi National Laboratory November.
HEPiX FNAL ‘02 25 th Oct 2002 Alan Silverman HEPiX Large Cluster SIG Report Alan Silverman 25 th October 2002 HEPiX 2002, FNAL.
SURA BOT 11/5/02 Lattice QCD Stephen J Wallace. SURA BOT 11/5/02 Lattice.
1 Lattice QCD Clusters Amitoj Singh Fermi National Accelerator Laboratory.
1 Cluster Development at Fermilab Don Holmgren All-Hands Meeting Jefferson Lab June 1-2, 2005.
Network design Topic 6 Testing and documentation.
Site Report on Physics Plans and ILDG Usage for US Balint Joo Jefferson Lab.
USQCD regional grid Report to ILDG /28/09ILDG14, June 5, US Grid Usage  Growing usage of gauge configurations in ILDG file format.  Fermilab.
What is QCD? Quantum ChromoDynamics is the theory of the strong force
HEP and NP SciDAC projects: Key ideas presented in the SciDAC II white papers Robert D. Ryne.
CD FY09 Tactical Plan Status FY09 Tactical Plan Status Report for Neutrino Program (MINOS, MINERvA, General) Margaret Votava April 21, 2009 Tactical plan.
Computing Performance Recommendations #10, #11, #12, #15, #16, #17.
U.S. Department of Energy’s Office of Science Midrange Scientific Computing Requirements Jefferson Lab Robert Edwards October 21, 2008.
Physics Results from CDF and Prospects for a FY 2011 Run Kevin Pitts / University of Illinois DOE S&T Review of Scientific User Facilities June 30 – July.
UKQCD NeSCAC Irving, 24/1/061 January 06 UKQCD meeting Staggered fermion project Alan Irving University of Liverpool.
Projects, Tools and Engineering Patricia McBride Computing Division Fermilab March 17, 2004.
RUP RATIONAL UNIFIED PROCESS Behnam Akbari 06 Oct
PEER 2003 Meeting 03/08/031 Interdisciplinary Framework Major focus areas Structural Representation Fault Systems Earthquake Source Physics Ground Motions.
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
CD FY08 Tactical Plan Status FY08 Tactical Plan Status Report for SCF/QCD James N. Simone and Amitoj Singh With Bakul Banerjee, Bill Boroski, Don Holmgren,
Fermi National Accelerator Laboratory & Thomas Jefferson National Accelerator Facility SciDAC LQCD Software The Department of Energy (DOE) Office of Science.
Grid development at University of Manchester Hardware architecture: - 1 Computer Element and 10 Work nodes Software architecture: - EasyGrid to submit.
LQCD Computing Project Overview
Project Management – Part I
Computational Requirements
Lattice QCD Computing Project Review
Systems Implementation,
Project Management – Part II
LQCD Computing Operations
Scientific Computing At Jefferson Lab
Patrick Dreher Research Scientist & Associate Director
International Lattice Data Grid
Presentation transcript:

LQCD Project Overview Don Holmgren LQCD Project Progress Review May 25-26, 2006 Fermilab

May 25, 2006LQCD Project Overview2 Project Scope The construction/purchase of new clusters/systems The operation of the U.S. QCDOC at Brookhaven, the existing SciDAC clusters at Fermilab and Jefferson Lab, and the systems acquired during the project Labor for system administration, hardware support, and site management Not in scope (provided by SciDAC/Base): Labor for scientific software support Software development

May 25, 2006LQCD Project Overview3 Existing Systems (as of Oct 1) BNL: US QCDOC, nodes, 4.2 Tflops Fermilab: QCD, 127 nodes, 0.15 Tflops Pion, 518 nodes, 0.86 Tflops JLab: 2M, 128 nodes, 0.06 Tflops (retired May 2006) 3G, 256 nodes, 0.19 Tflops 4G, 384 nodes, 0.46 TFlops

May 25, 2006LQCD Project Overview4 LQCD Staffing – FY06 Project + Base/SciDAC FTE sysadmin / technician Scientific software & user support site management BNL FNAL JLab Manpower increases at FNAL, JLab after FY06 as size of operated systems grows.

May 25, 2006LQCD Project Overview5 LQCD Milestones Top level milestones As listed in LQCD OMB FY2007 Exhibit 300 Revisions to be discussed in a later talk Deployment and delivery in each year as shown Also, a progress review by June 30 in FY06-FY08 YearDeliver (Tflops-Yrs) Deploy (Tflops) FY FY FY FY

May 25, 2006LQCD Project Overview6 Definitions TFlops = average of domain wall fermion (DWF) and MILC asqtad performance. asqtad and DWF are the two most common “actions” used in US LQCD calculations Ratio of DWF:asqtad performance is typically 1.2:1, but this varies by machine (as high as 1.4:1) “Top500” TFlops are considerably higher (2:1 – 3:1) TFlops-yr = available time-integrated performance during an 8000-hour year Remaining 800 hours are assumed to be consumed by engineering time and other downtime

May 25, 2006LQCD Project Overview7 Budget Breakdown $9.2M total, $2.5M FY06-FY08, $1.7M FY09 HEP funding: 80% NP Funding: 20% Mixture of operations and equipment: YearOperationsEquipment FY06$650K$1850K FY07$884K$1616K FY08$960K$1540K FY09$1032K$668K Total$3526K$5674K

May 25, 2006LQCD Project Overview8 LQCD Metafacility The LQCD project has no need for a computational grid Scientific allocations are for 12 months and a given allocation runs at one site only Intermediate data products are large so local access more efficient Hardware is heterogeneous (QCDOC, and gigE mesh, Myrinet, and Infiniband clusters)  no binary compatibility possible A data grid will be a convenience for users Minimal impact if delayed 99% of all data access is local Users currently rely on scp and occasionally grid tools to move files ILDG (International Lattice Data Grid) software deployment this summer will provide a metadata catalog, a replica catalog, and data access services

May 25, 2006LQCD Project Overview9 LQCD Metafacility cont’d Common Runtime Environment Goal: Allow site independent Makefiles, batch and run scripts Standards include file system layouts, interactive, batch, and parallel execution environments Part of the SciDAC project (external dependency) Status: draft standard available, implementation by end of July 0.15 FTE assigned (Metafacility Operations Manager) Defines and coordinates ILDG and Common Run Environment deployments Coordinates documentation and testing

May 25, 2006LQCD Project Overview10 Integrated Baseline Review – May 2005 Major recommendations regarding deployments from May ’05 Integrated Baseline Review: Build one system (cluster) per year Alternate sites (Fermilab/Jefferson Lab) Choose site and design of each purchase to maximize science output

May 25, 2006LQCD Project Overview11 Project Response Delivered at August 8, 2005 meeting in Germantown Proposed deployment in FY06: 0.4 Tflops cluster at JLab (0.2 Tflops SciDAC/base, 0.2 Tflops project), based on Infiniband and FY05 Fermilab “Pion” cluster, large enough to do DWF algorithm development and analysis of MILC asqtad configurations with DWF fermions Critical to establish Infiniband expertise at JLab to prepare for a large cluster in FY2007 Release to production by June 30 Status: released May 1

May 25, 2006LQCD Project Overview12 Project Response Deployment in FY06, cont’d : 2.0 Tflops cluster at FNAL (0.2 Tflops SciDAC / supplemental, 1.8 Tflops project), based on Infiniband, sufficient for configuration generation and analysis of fine lattices Wait for FBDIMM memory architecture, which promises the memory bandwidth required for meeting our performance goals Release to production by Sept 30 Status: On schedule, Opteron-based, delivery in July/August FBDIMM architecture has disappointing performance Subsequent deployments would be at most one per year, alternating sites We will discuss in a later talk the advantages of combining FY08/FY09

May 25, 2006LQCD Project Overview13 Significance and Relevance of Research Major research objectives Calculate weak interaction matrix elements to the accuracy needed for precise tests of the Standard Model. Determine the properties of strongly interacting matter at high temperatures and densities. Calculate the masses of strongly interacting particles and obtain a quantitative understanding of their internal structure.

May 25, 2006LQCD Project Overview14 Significance and Relevance of Research Relevance to experimental programs in high energy and nuclear physics: Weak matrix elements: Babar (SLAC), D0 and CDF (FNAL), CLEO-c (Cornell) High temperature/density QCD: RHIC (BNL) Hadron structure: CEBAF (JLab), RHIC (BNL)

May 25, 2006LQCD Project Overview15 Some Recent Achievements Validation through calculations of quantities that are well determined experimentally (validation of calculational approach): QuantityLattice QCDExperiment  s (M Z ) ± ± f K / f  ± ± V us ± ± gAgA ± ±

May 25, 2006LQCD Project Overview16 Some Recent Achievements LQCD predictions that have been verified by experiment – these analyses were enabled by the availability of the Fermilab clusters. QuantityLattice QCDExperiment fDfD 201 ± 3 ± 17 MeV223 ± MeV f Ds 249 ± 3 ± 16 MeV279 ± 17 ± 20 MeV m Bc 6304 ± 20 MeV6287 ± 5 MeV

May 25, 2006LQCD Project Overview17 Form Factor for the Semileptonic Decay of the K Meson The semileptonic form factor f + (q 2 ) for the decay of a D meson into a K meson and leptons, as a function of the momentum transfer to the leptons q 2. The orange curve is the lattice prediction, and the violet points are the experimental results of the Belle Collaboration.

May 25, 2006LQCD Project Overview18 Talks Today Computational Requirements JLab 2006 Acquisition Details Fermilab 2006 Acquisition Details Proposed 2007 Acquisition Plan Project Management

Backup Slides

May 25, 2006LQCD Project Overview20 Manpower - BNL Management Eric Blum 0.10 FTE (LQCD) Tom Schlagel 0.04 FTE (LQCD) Hardware Support Joe DePace 0.35 FTE (LQCD) Don Gates 0.35 FTE (LQCD) System Administration John Reddy 0.05 FTE (LQCD) Software Support Efstratios Efstathiadis 0.50 FTE = 0.25 SciDAC Base Chulwoo Jung 0.37 FTE (SciDAC I and II) Enno Schultz 0.25 FTE (SciDAC I and II)

May 25, 2006LQCD Project Overview21 Manpower - JLab Chip Watson 0.75 FTE = 0.15 LQCD SciDAC Base Balint Joo 1.00 FTE = 0.15 LQCD SciDAC Ying Chen 1.00 FTE = 0.20 LQCD SciDAC Base Jie Chen 0.90 FTE (SciDAC) Sherman White 0.50 FTE = 0.40 LQCD Base Robert Edwards 0.20 FTE (Base) New Hire (October 2006) 1.00 FTE = 0.85 LQCD Base

May 25, 2006LQCD Project Overview22 Manpower - FNAL Don Holmgren 1.00 FTE = 0.25 LQCD-PM LQCD SciDAC Base Bakul Banerjee 0.50 FTE = 0.25 LQCD-PM Base Amitoj Singh 1.00 FTE = 0.50 LQCD SciDAC Kurt Ruthmansdorfer 1.00 FTE = 1.00 LQCD Jim Simone 0.50 FTE (SciDAC) Jim Kowalkowski 0.50 FTE (SciDAC/Base) Bob Forster 0.25 FTE LQCD New Hire (August 2006) 1.00 FTE (SciDAC/Base)