SciDAC Software Infrastructure for Lattice Gauge Theory

Slides:



Advertisements
Similar presentations
UKQCD GridPP NeSCAC Irving, 4/2/041 9 th GridPP Collaboration Meeting QCDgrid: Status and Future Alan Irving University of Liverpool.
Advertisements

I/O and the SciDAC Software API Robert Edwards U.S. SciDAC Software Coordinating Committee May 2, 2003.
Nuclear Physics in the SciDAC Era Robert Edwards Jefferson Lab SciDAC 2009 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this.
Lattice Quantum Chromodynamic for Mathematicians Richard C. Brower Yale University May Day 2007 Tutorial in “ Derivatives, Finite Differences and.
Dynamical Chiral Fermions The `Grail’ – dyn. chiral fermions Generation of dyn. chiral fermions configs –RBC on the RIKEN QCDOC – Jan 05 (some %) –UKQCD.
Lattice QCD Comes of Age y Richard C. Brower XLIst Rencontres de Moriond March QCD and Hadronic interactions at high energy.
P. Vranas, IBM Watson Research Lab 1 Gap domain wall fermions P. Vranas IBM Watson Research Lab.
Adaptive Multigrid for Lattice QCD James Brannick October 22, 2007 Penn State
BU SciDAC Meeting Balint Joo Jefferson Lab. Anisotropic Clover Why do it ?  Anisotropy -> Fine Temporal Lattice Spacing at moderate cost  Combine with.
QDP++ and Chroma Robert Edwards Jefferson Lab
Design Considerations Don Holmgren Lattice QCD Project Review May 24, Design Considerations Don Holmgren Lattice QCD Computing Project Review Cambridge,
Algorithms for Lattice Field Theory at Extreme Scales Rich Brower 1*, Ron Babich 1, James Brannick 2, Mike Clark 3, Saul Cohen 1, Balint Joo 4, Tony Kennedy.
SciDAC Software Infrastructure for Lattice Gauge Theory DOE meeting on Strategic Plan --- April 15, 2002 Software Co-ordinating Committee Rich Brower ---
HackLatt MILC with SciDAC C Carleton DeTar HackLatt 2008.
MILC Code Basics Carleton DeTar KITPC MILC Code Capabilities Molecular dynamics evolution –Staggered fermion actions (Asqtad, Fat7, HISQ,
1 Parallel Computing—Introduction to Message Passing Interface (MPI)
HackLatt MILC Code Basics Carleton DeTar HackLatt 2008.
SciDAC Software Infrastructure for Lattice Gauge Theory DOE Grant ’01 -- ’03 (-- ’05?) All Hands Meeting: FNAL Feb. 21, 2003 Richard C.Brower Quick Overview.
I/O and the SciDAC Software API Robert Edwards U.S. SciDAC Software Coordinating Committee May 2, 2003.
The science of simulation falsification algorithms phenomenology machines better theories computer architectures non-perturbative QFT experimental tests.
1 Down Place Hammersmith London UK 530 Lytton Ave. Palo Alto CA USA.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
The QCDOC Project Overview and Status Norman H. Christ DOE LGT Review May 24-25, 2005.
LQCD Project Overview Don Holmgren LQCD Project Progress Review May 25-26, 2006 Fermilab.
Lattice QCD and GPU-s Robert Edwards, Theory Group Chip Watson, HPC & CIO Jie Chen & Balint Joo, HPC Jefferson Lab TexPoint fonts used in EMF. Read the.
QCDgrid UKQCD Achievements and Future Priorities Who and what Achievements QCDgrid middleware Future priorities Demo of meta-data catalogue browser Alan.
SciDAC Software Infrastructure for Lattice Gauge Theory Richard C. Brower Annual Progress Review JLab, May 14, 2007 Code distribution see
Simulating Quarks and Gluons with Quantum Chromodynamics February 10, CS635 Parallel Computer Architecture. Mahantesh Halappanavar.
QCD Project Overview Ying Zhang September 26, 2005.
9/16/2000Ian Bird/JLAB1 Planning for JLAB Computational Resources Ian Bird.
Lattice QCD in Nuclear Physics Robert Edwards Jefferson Lab CCP 2011 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.:
Chroma I: A High Level View Bálint Joó Jefferson Lab, Newport News, VA given at HackLatt'06 NeSC, Edinburgh March 29, 2006.
SciDAC Software Infrastructure for Lattice Gauge Theory Richard C. Brower & Robert Edwards June 24, 2003.
LQCD Clusters at JLab Chip Watson Jie Chen, Robert Edwards Ying Chen, Walt Akers Jefferson Lab.
UKQCD QCDgrid Richard Kenway. UKQCD Nov 2001QCDgrid2 why build a QCD grid? the computational problem is too big for current computers –configuration generation.
Lattice QCD and the SciDAC-2 LQCD Computing Project Lattice QCD Workflow Workshop Fermilab, December 18, 2006 Don Holmgren,
Crystal Ball Panel ORNL Heterogeneous Distributed Computing Research Al Geist ORNL March 6, 2003 SOS 7.
HackLatt MILC Code Basics Carleton DeTar First presented at Edinburgh EPCC HackLatt 2008 Updated 2013.
Proposed 2007 Acquisition Don Holmgren LQCD Project Progress Review May 25-26, 2006 Fermilab.
09/02 ID099-1 September 9, 2002Grid Technology Panel Patrick Dreher Technical Panel Discussion: Progress in Developing a Web Services Data Analysis Grid.
May 25-26, 2006 LQCD Computing Review1 Jefferson Lab 2006 LQCD Analysis Cluster Chip Watson Jefferson Lab, High Performance Computing.
Lattice QCD and GPU-s Robert Edwards, Theory Group Chip Watson, HPC & CIO Jie Chen & Balint Joo, HPC Jefferson Lab TexPoint fonts used in EMF. Read the.
Chroma: An Application of the SciDAC QCD API(s) Bálint Joó School of Physics University of Edinburgh UKQCD Collaboration Soon to be moving to the JLAB.
1 1 What does Performance Across the Software Stack mean?  High level view: Providing performance for physics simulations meaningful to applications 
SciDAC Software Infrastructure for Lattice Gauge Theory Richard C. Brower QCD Project Review May 24-25, 2005 Code distribution see
1 Cluster Development at Fermilab Don Holmgren All-Hands Meeting Jefferson Lab June 1-2, 2005.
Site Report on Physics Plans and ILDG Usage for US Balint Joo Jefferson Lab.
Linear Algebra Libraries: BLAS, LAPACK, ScaLAPACK, PLASMA, MAGMA
What is QCD? Quantum ChromoDynamics is the theory of the strong force
A QCD Grid: 5 Easy Pieces? Richard Kenway University of Edinburgh.
COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC.
HEP and NP SciDAC projects: Key ideas presented in the SciDAC II white papers Robert D. Ryne.
U.S. Department of Energy’s Office of Science Midrange Scientific Computing Requirements Jefferson Lab Robert Edwards October 21, 2008.
QDP++ and Chroma Robert Edwards Jefferson Lab Collaborators: Balint Joo.
UKQCD NeSCAC Irving, 24/1/061 January 06 UKQCD meeting Staggered fermion project Alan Irving University of Liverpool.
Fermi National Accelerator Laboratory & Thomas Jefferson National Accelerator Facility SciDAC LQCD Software The Department of Energy (DOE) Office of Science.
POE Parallel Operating Environment. Cliff Montgomery.
INTRODUCTION TO HIGH PERFORMANCE COMPUTING AND TERMINOLOGY.
Jun Doi IBM Research – Tokyo Early Performance Evaluation of Lattice QCD on POWER+GPU Cluster 17 July 2015.
LQCD Computing Project Overview
Project Management – Part I
Computational Requirements
Lattice QCD Computing Project Review
Programming Models for SimMillennium
Cyberinfrastructure – the NEOS Project
LQCD Computing Operations
Is System X for Me? Cal Ribbens Computer Science Department
ILDG Implementation Status
Patrick Dreher Research Scientist & Associate Director
Chroma: An Application of the SciDAC QCD API(s)
Presentation transcript:

SciDAC Software Infrastructure for Lattice Gauge Theory Richard C. Brower All Hands Meeting JLab June 1-2, 2005 MAKE SLIDES MIRROR WRITE UP! Code distribution see http://www.usqcd.org/usqcd-software (Links to documentation at bottom)

Major Participants in Software Project UK Peter Boyle Balint Joo (Mike Clark) * Software Coordinating Committee

QCD API Design Completed Components Performance Future OUTLINE Note: Thursday 9:30 Applications of the SciDAC API TBA = Carleton DeTar & Balint Joo

Lattice QCD – extremely uniform (Perfect) Load Balancing: Uniform periodic lattices & identical sublattices per processor. (Complete) Latency hiding: overlap computation /communications Data Parallel: operations on Small 3x3 Complex Matrices per link. Critical kernel : Dirac Solver is ~90% of Flops. Lattice Dirac operator:

C/C++, implemented over MPI, native SciDAC QCD API Optimised for P4 and QCDOC Optimized Dirac Operators, Inverters Level 3 ILDG collab QDP (QCD Data Parallel) QIO Binary Data Files / XML Metadata Level 2 Lattice Wide Operations, Data shifts Exists in C/C++ QLA (QCD Linear Algebra) Level 1 QMP (QCD Message Passing) C/C++, implemented over MPI, native QCDOC, M-via GigE mesh

Fundamental Software Components are in Place: QMP (Message Passing) version 2.0 GigE, in native QCDOC & over MPI. QLA (Linear Algebra) C routines (24,000 generated by Perl with test code). Optimized (subset) for P4 Clusters in SSE assembly. QCDOC optimized using BAGEL -- assembly generator for PowerPC, IBM power, UltraSparc, Alpha… by P. Boyle UK QDP (Data Parallel) API C version for MILC applications code C++ version for new code base Chroma I/0 and ILDG: Read/Write for XML/binary files compliant with ILDG Level 3 (Optimized kernels) Dirac inverters for QCDOC and P4 clusters

Example of QDP++ Expression Typical for Dirac Operator: QDP/C++ code: Portable Expression Template Engine (PETE) Temporaries eliminated, expressions optimized

Gauge Covariant Version of QDP/C++ code ? Lattice gauge algebra has bifundamentals: QDP new type: (Field U, displacement y-x) No shifts! All data movement is implicit in * operator. Gauge non-invariant expressions are impossible, except with special identity field for pure shifts ( Identity[vec])

Performance of Optimized Dirac Inverters QCDOC Level 3 inverters on 1024 processor racks Level 2 on Infiniband P4 clusters at FNAL Level 3 GigE P4 at JLab and Myrinet P4 at FNAL Future Optimization on BlueGene/L et al.

Level 3 on QCDOC

Level 3 Aqstad QCDOC (double precision)

Level 2 Asqtad on single P4 node

Asqtad on Infiniband Cluster

Level 3 Domain Wall CG Inverter y (32 nodes) 42 6 16 (4) 52 20 (6)102 40 6 163 y Ls = 16, 4g is P4 2.8MHz, 800MHz FSB

Alpha Testers on the 1K QCDOC Racks MILC C code: Level 3 Asqtad HMD: 403 * 96 lattice for 2 + 1 flavors (Each trajectory: 12hrs 18 min, Dru Renner) Chroma C++ code: Level 3 domain wall RHMC: 243 * 32 (Ls = 8) for 2+1 flavors (James Osborn & Robert Edwards & UKQCD) I/O: Reading and writing lattices to RAID and host AIX box File transfers: 1 hop file transfers between BNL, FNAL & JLab

BlueGene/L Assessment June ’04: 1024 node BG/L Wilson HMC sustain 1 Teraflop (peak 5.6) Bhanot, Chen, Gara, Sexton & Vranas hep-lat/0409042 Feb ’05: QDP++/Chroma running on UK BG/L Machine June ’05: BU and MIT each acquire 1024 Node BG/L ’05-’06 Software Development: Optimize Linear Algebra (BAGEL assembler from P. Boyle) at UK Optimize Inverter (Level 3) at MIT Optimize QMP (Level 1) at BU

Future Software Directions On going: Optimization, Testing and Hardening of components. New Level 3 and application routines. Executables for BG/L, X1E, XT3, Power 3, PNNL cluster, APEmille… Integration: Uniform Interfaces, Uniform Execution Environment for 3 Labs. Middleware for data sharing with ILDG. Data, Code, Job control as 3 Lab Meta Facility (leverage PPDG) Algorithm Research: Increased performance (leverage TOPS & NSF IRT). New physics (Scattering, resonances, SUSY, Fermion signs, etc?) Generalize API: QCD like extensions of SM, SUSY, New Lattices, Related Apps? (See FermiQCD) Graphics: Debugger, GUI for Job and Data Management, Physics & PR ? Today at 5:30 MG talk by James Brannick

Old MG resultsy to set context for Brannick’s talk: 2x2 Blocks for Dirac Operator  = 1 2-d Lattice, U(x) on links (x) on sites Gauss-Jacobi (Diamond), CG (circle), MG with V cycle (square), W cycle (star) y R. C. Brower, R. Edwards, C.Rebbi,and E. Vicari, "Projective multigrid forWilson fermions", Nucl. Phys.B366 (1991) 689 (aka Spectral AMG, Tim Chatier, 2002)

Limitation of earlier MG methods with gauge invariant but non-adaptive projections for 2-d Schwinger Modely Universality of  vs mq/1/2 Gauss-Jacobi (Diamond), CG (circle), 3 level MG (square & star)  = 3 (cross), 10(plus), 100( square)