Download presentation
Presentation is loading. Please wait.
Published byCristina Holyfield Modified over 9 years ago
1
SUPER Bob Lucas University of Southern California Sept. 23, 2011 SciDAC-3 Institute for Sustained Performance, Energy, and Resilience http://super-scidac.org Bob Lucas, PI University of Southern California Argonne PI: Boyana Norris norris@mcs.anl.gov Support for this work was provided through the Scientific Discovery through Advanced Computing (SciDAC) program funded by the U.S. Department of Energy, Office of Science, Advanced Scientific Computing Research
2
SUPER SciDAC Performance Efforts SciDAC has always had an effort focused on performance Performance Evaluation Research Center (PERC) – Benchmarking, modeling, and understanding Performance Engineering Research Institute (PERI) – Performance engineering, modeling, and engagement – Three SciDAC-e projects Institute for Sustained Performance, Energy, and Resilience (SUPER) – Performance engineering – Energy minimization – Resilient applications – Optimization of the above SUPER http://super-scidac.org210/18/12
3
SUPER Outline SUPER Team Research Directions Broader Engagement SUPER http://super-scidac.org310/18/12
4
SUPER SUPER Team SUPER Gabriel Marin Philip Roth Patrick Worley ORNL David Bailey Lenny Oliker Sam Williams LBNL Bronis de Supinski Daniel Quinlan LLNL UCSD Laura Carrington Ananta Tiwari Jeffrey Hollingsworth UMD Dan Terpstra UTK Allen Malony Sameer Shende Oregon Utah Mary Hall ANL Paul Hovland Boyana Norris Stefan Wild USC Jacque Chame Pedro Diniz Bob Lucas (PI) Rob Fowler Allan Porterfield UNC Shirley Moore UTEP http://super-scidac.org410/18/12
5
SUPER Broadly Based Effort All PIs have independent research projects – SUPER money alone isn’t enough to support any of its investigators – SUPER leverages other work and funding SUPER contribution is integration, results beyond any one group – Follows successful PERI model (tiger teams and autotuning) – Collaboration extends to others having similar research goals John Mellor-Crummey at Rice is an active collaborator Other likely collaborators include LANL, PNNL, Portland State, and UT San Antonio Perhaps Juelich and Barcelona too? SUPER http://super-scidac.org510/18/12
6
SUPER Outline SUPER Team Research Directions Broader Engagement SUPER http://super-scidac.org610/18/12
7
SUPER SUPER Research Objectives Performance Optimization of DOE Applications Automatic performance tuning Energy minimization Resilient computing Optimization of the above Collaboration with SciDAC-3 Institutes Engagement with SciDAC-3 Applications SUPER http://super-scidac.org710/18/12
8
SUPER Optimization of Particle-to-Grid Methods on Multicore Systems (Oliker and Williams, also supported by XStack) GoalsObjectives Develop multicore-specific algorithms for a broad spectrum of particle-mesh applications. Explore optimization, challenged by complex data hazards and irregular data access. Demonstrate a unique synchronization and decomposition approach that improves performance on many of multicore platforms. Our methodology resulted in substantial reduction in memory usage while simultaneously improving performance. Our unique application-specific combination of optimizations achieved 4x and 5.6x improvement for heart and fusion particle-to-grid simulations. This work is a critical step toward designing auto- tuning frameworks, which will extend the result to numerous particle methods and platforms. Published in IEEE Transactions of Parallel and Distributed Processing (TPDS) Progress & Accomplishments Heart Code GTC Speedup of optimized version vs reference parallel code on Nehalem, Opteron, and Niagara platforms Particle-to-grid methods are very common, e.g., fusion, astrophysics, chemistry, and fluid mechanics. Performance is challenged on modern multicores by memory and bandwidth constraints. Our objective is performance enhancement of particle-to-grid applications, working in the context of a heart simulation and a fusion code. http://super-scidac.org810/18/12
9
SUPER Performance Engineering Research Measurement and monitoring – Adopting University of Oregon’s TAU system – Building on UTK’s PAPI measurement library – Also collaborating with Rice and its HPCToolkit Performance Database – Extending TAUdb to enable online collection and analysis Performance modeling – PBound and Roofline models to bound performance expectations – MIAMI to model impact of architectural variation – PSINS to model communication SUPER http://super-scidac.org910/18/12
10
SUPER Generating Performance Models with PBound Use source code analysis to generate performance bounds (best/worst case scenarios) Can be used for – Understanding performance problems on current architectures For example, when presented in the context of roofline models (introduced by Sam Williams, LBNL) – Projecting performance to hypothetical architectures http://super-scidac.org1010/18/12
11
SUPER Example Roofline Model locality walls 11 Remember, memory traffic includes more than just compulsory misses. As such, actual arithmetic intensity may be substantially lower. Walls are unique to the architecture-kernel combination actual FLOP:Byte ratio attainable GFLOP/s 0.5 1.0 1/81/8 2.0 4.0 8.0 16.0 32.0 64.0 128.0 256.0 1/41/4 1/21/2 124816 w/out SIMD mul / add imbalance w/out ILP w/out SW prefetch w/out NUMA Opteron 2356 (Barcelona) peak DP Stream Bandwidth only compulsory miss traffic+write allocation traffic+capacity miss traffic+conflict miss traffic FLOPs Conflict + Capacity + Allocations + Compulsory AI = Model Observed Performance http://super-scidac.org10/18/12
12
SUPER 12 SUPE R 12 Sample PBound Output #include "pbound_list.h" void axpy4(int n, double *y, double a1, double *x1, double a2, double *x2, double a3, double *x3, double a4, double *x4) { #ifdef pbound_log pboundLogInsert("axpy.c@6@5",2,0,5 * ((n - 1) + 1) + 4, 1 * ((n - 1) + 1),3 * ((n - 1) + 1) + 1,4 * ((n - 1) + 1)); #endif } void axpy4(int n, double *y, double a1, double *x1, double a2, double *x2, double a3, double *x3, double a4, double *x4) { register int i; for (i=0; i<=n-1; i++) y[i]=y[i]+a1*x1[i]+a2*x2[i]+a3*x3[i]+a4*x4[i]; } http://super-scidac.org10/18/12
13
SUPER Autotuning Led by Mary Hall, University of Utah Extend PERI autotuning system for future architectures – New TAU front-end for triage – CUDA-CHiLL, Orio to target GPUs – OpenMP-CHiLL for SMP multicores – Active Harmony provides search engine Drive empirical autotuning experimentation Balance threads and MPI ranks in hybrids of OpenMP and MPI Extend to surface/area to volume, or halo size, experiments – Targeted Autotuning Domain-specific languages Users write simple code and leave the tuning to us – Whole program autotuning Parameters, algorithm choice, libraries linked, etc. SUPER http://super-scidac.org1310/18/12
14
SUPER Autotuning PETSc Solvers on GPUs with Orio B. Norris and A. Mametjanov, Argonne ImpactObjectives Generate high-performance implementations sparse matrix algebra kernels defined using a high-level syntax Optimize performance automatically on different architectures, including cache-based multicore CPUs and heterogeneous CPU/GPU systems Integrate autotuning into the Portable Extensible Toolkit for Scientific Computing (PETSc) Efficiently search the space of possible optimizations Optimizations for multicore CPUs and GPUs from the same high-level input. Improved code readability and maintainability. Performance of autotuned kernels typically exceeds that of tuned vendor numerical library implementations. Accomplishments FY2012 Intel Xeon 2.8 GHz E5462 (8 cores), 2.8GHz; GPU: NVIDIA Fermi C2070 Example: Autotuning a structured-grid PDE application using PETSc for GPUs (solid fuel ignition). lower is better Developed new transformations for sparse matrix operations targeting GPUs. Autotuning resulted in improvements ranging from 1.5x to 2x over tuned vendor libraries on different GPUs for a solid fuel ignition PETSc application. “Autotuning stencil-based computations on GPUs.” A. Mametjanov, D. Lowell, C.-C. Ma, and B. Norris. Proceedings of IEEE Cluster 2012. http://tinyurl.com/OrioTool http://super-scidac.org1410/18/12
15
SUPER Pflotran Autotuning Example CompilerOriginalActive HarmonyExhaustive Time (u1,u2)SpeedupTime(u1,u2)Speedup pathscale0.580.32(3,11)1.810.30(3,15)1.93 gnu0.710.47(5,13)1.510.46(5,7)1.54 pgi0.900.53(5,3)1.700.53(5,3)1.70 cray1.130.70(15,5)1.610.69(15,15)1.63 PETSc Triangular Solve Kernel 15 10/18/12
16
SUPER TAU (Oregon) HPC Toolkit (Rice) ROSE (LLNL) MIAMI (ORNL) PAPI (UTK) CHiLL (USC/ISI and Utah) ROSE (LLNL) Orio (Argonne) { OSKI (LBNL) Active Harmony (UMD) GCO (UTK) PerfTrack (LBNL, SDSC, RENCI) Autotuning Framework http://super-scidac.org1610/18/12
17
SUPER Energy Minimization Led by Laura Carrington, University of California at San Diego Develop new energy aware APIs for users – I know the processor on the critical path in my multifrontal code Obtain more precise data regarding energy consumption – Extend PAPI to sample hardware power monitors – Build new generation of PowerMon devices – Extend performance models Transform codes to minimize energy consumption Inform systems to allow them to exploit DVFS SUPER http://super-scidac.org1710/18/12
18
SUPER MILC (6%) GTC (5%) PSCYEE (19%) FT (20.5%) LAMMPS (32%) Green Queue: Customized Large-scale Clock Frequency Scaling (Carrington and Tiwari) Research PlanOverall Objectives The power wall is one of the major factors impeding the development and deployment of exascale systems Solutions are needed that reduce energy from both the hardware and software side Solutions must be easy to use, fully automated, scalable and application-aware Develop static and runtime application analysis tools to quantify application phases that affect both system-wide power draw and performance Use the properties of these application phases to develop models of the power draw Combine power and performance models to create phase- specific Dynamic Voltage-Frequency Scaling (DVFS) settings Developed Green Queue, a framework that can make CPU clock frequency changes based on intra-node (e.g., memory latency bound) and inter-node (e.g., load imbalance) aspects of an application Deployed Green Queue on SDSC’s Gordon supercomputer using one full rack (1024 cores) and a set scientific applications Green Queue shown to reduce the operational energy required for HPC applications by an average of 12% (maximum 32%) Deployed a test system for SUPER team members to facilitate collaboration on energy efficiency research and tool integration Progress & Accomplishments Power monitoring during application runs. Numbers in parenthesis show energy savings. Application Name (% energy saved) http://super-scidac.org1810/18/12
19
SUPER Resilient Computing Led by Bronis de Supinski, Livermore National Laboratory Investigate directive-based API for users – Enable users to express their knowledge w/r resilience – Not all faults are fatal errors – Those that can’t be tolerated can often be ameliorated Automating vulnerability assessment – Modeled on success of PERI autotuning – Conduct fault injection experiments – Determine which code regions or data structures fail catastrophically – Determine what transformations enable them to survive In either event, extend ROSE compiler to implement transformations SUPER http://super-scidac.org1910/18/12
20
SUPER Fault Resilience of an Algebraic Multi-Grid (AMG) Solver (Casas Guix and de Supinski) Goals of the SUPER resilience effortGeneral AMG Features AMG solves linear systems of equations derived from the discretization of partial differential equations. AMG is an iterative method that operates on nested grids of varying refinement. Two operators (restriction and interpolation) propagate the linear system through the different grids. Identify vulnerable data and code regions. Design and implement simple and effective resilience strategies to improve vulnerability of sensitive pieces of code. Long term: develop a general methodology to automatically improve the reliability of generic HPC codes. Developed of a methodology to automatically inject faults to assess the vulnerability of codes to soft errors. Performed a vulnerability study of AMG. Determined that AMG is most vulnerable to soft errors in pointer arithmetic, which lead to fatal segmentation faults. Demonstrated that triple modular redundancy in pointer calculations reduces the vulnerability of AMG to soft errors Presented at ICS 2012 Progress & Accomplishments In three different experiments, increasing pointer replication in AMG reduces the number of fatal segmentation faults. http://super-scidac.org2010/18/12
21
SUPER Optimization Led by Paul Hovland, Argonne National Laboratory Performance, energy, and resilience are implicitly related and require simultaneous optimization E.g., Processor pairing covers soft errors, but halves throughput SUPER Results in a stochastic, mixed integer, nonlinear, multi-objective, optimization problem Only sample small portion of search space: Requires efficient derivative-free numerical optimization algorithms Need to adapt algorithms from continuous to discrete autotuning domain http://super-scidac.org2110/18/12
22
SUPER Optimization: Time vs. Energy Conventional wisdom suggests that best strategy to minimize energy is to minimize execution time However, in practice, real tradeoffs between time and energy are observed – Depends on computation being performed – Depends on problem size SUPER http://super-scidac.org2210/18/12
23
SUPER Outline SUPER Team Research Directions Broader Engagement SUPER http://super-scidac.org2310/18/12
24
SUPER SciDAC-3 Application Partnerships Collaboration with SciDAC Application Partnerships is expected – Yet SUPER funding is spread very thin SUPER investigators included in 12 Application Partnerships – Our time costs money, like everybody else Principles used to arrange teams – Technical needs of the proposal – Familiarity of the people – Proximity SUPER http://super-scidac.org2410/18/12
25
SUPER Application Engagement Led by Pat Worley, Oak Ridge National Laboratory PERI strategy: proactively identify application collaborators – Based on comprehensive application survey at beginning of SciDAC-2 – Exploited proximity and long-term relationships SUPER strategy: broaden our reach – Key is partnering with staff at ALCF, OLCF, and NERSC – Augment TAUdb to capture data from applications from centers – Initial interactions between Oregon and ORNL with and LCFs – Collaborate with other SciDAC-3 investigators Focused engagement as requested by DOE SUPER http://super-scidac.org2510/18/12
26
SUPER SUPER Summary Research components – Automatic performance tuning New focus on portability – Addressing the “known unknowns” Energy minimization Resilient computing – Optimization of the above Near-term impact on DOE computational science applications – Application engagement coordinated with ALCF, NLCF, and NERSC – Tool integration, making research artifacts more approachable – Participation in SciDAC-3 Application Partnerships – Outreach and tutorials SUPER http://super-scidac.org2610/18/12
27
SUPER Bonus Slides SUPER http://super-scidac.org2710/18/12
28
SUPER Outreach and Tutorials Led by David Bailey, Lawrence Berkeley National Laboratory We will offer training to ALCF, OLCF, and NERSC staff – Facilitate limited deployment of our research artifacts We will organize tutorials for end users of our tools – Offer them at widely attended forums such as SC conferences UTK is hosting a Web site, UTEP is maintaining it – www.super-scidac.org UTEP coordinating quarterly news letters SUPER http://super-scidac.org2810/18/12
29
SUPER SciDAC-3 Institute Collaboration PERI was directed not to work with math and CS institutes PERI nevertheless found itself tuning math libraries – E.g., PETSc kernels are computational bottlenecks in PFLOTRAN SciDAC-3 is not SciDAC-2; collaboration is encouraged Initial discussions with FASTMath and QUEST – Optimize math libraries with FASTMath – Optimize end-to-end workflows with QUEST SUPER http://super-scidac.org2910/18/12
30
SUPER Participation in SciDAC-3 Application Partnerships BER Applying Computationally Efficient Schemes for BioGeochemical Cycles (ORNL) BERMultiScale Methods for Accurate, Efficient, and Scale-Aware Models of the Earth Sys. (LBNL) BERPredicting Ice Sheet and Climate Evolution at Extreme Scales (LBNL & ORNL) BESDeveloping Advanced Methods for Excited State Chemistry in the NWChem S/W Suite (LBNL) BESOptimizing Superconductor Transport Properties through Large-scale Simulation (ANL) BESSimulating the Generation, Evolution and Fate of Electronic Excitations in Molecular and Nanoscale Materials with First Principles Methods (LBNL) FESPartnership for Edge Plasma Simulation (ORNL) FESPlasma Surface Interactions (ORNL) HEPCommunity Petascale Project for Accelerator Science and Simulation (ANL) NPA MultiScale Approach to Nuclear Structure and Reactions (LBNL) NPComputing Properties of Hadrons, Nuclei and Nuclear Matter from QCD (UNC) NPNuclear Computational Low Energy Initiative (ANL) SUPER http://super-scidac.org3010/18/12
31
SUPER Outlined code (from ROSE outliner) for (si = 0; si < stencil_size; si++) for (kk = 0; kk < hypre__mz; kk++) for (jj = 0; jj < hypre__my; jj++) for (ii = 0; ii < hypre__mx; ii++) rp[((ri+ii)+(jj*hypre__sy3))+(kk*hypre__sz3)] -= ((Ap_0[((ii+(jj*hypre__sy1))+ (kk*hypre__sz1))+ (((A->data_indices)[i])[si])])* (xp_0[((ii+(jj*hypre__sy2))+(kk*hypre__sz2))+(( *dxp_s)[si])])); CHiLL transformation recipe permute([2,3,1,4]) tile(0,4,TI) tile(0,3,TJ) tile(0,3,TK) unroll(0,6,US) unroll(0,7,UI) Constraints on search 0 ≤ TI, TJ, TK ≤ 122 0 ≤ UI ≤ 16 0 ≤ US ≤ 10 compilers ∈ {gcc, icc} Search space: 122 3 x16x10x2 = 581,071,360 points Autotuning the central SMG2000 kernel http://super-scidac.org3110/18/12
32
SUPER Collaboration with NNSA Evaluating accuracy and performance tradeoffs of the physical model Reducing SMN execution frequency: Improves performance Reduces the quality of the simulation SUPER ParaDiS models dislocation lines in a physical volume The volume is divided into several computational domains The dislocations interact and move in response to physical forces External stress and inter-dislocation interactions determine the forces http://super-scidac.org3210/18/12
33
SUPER SMG2000 Autotuning Result SUPER Selected parameters: TI=122,TJ=106,TK=56,UI=8,US=3,Comp=gcc Performance gain on residual computation: 2.37X Performance gain on full app: 27.23% improvement Parallel search evaluates 490 points and converges in 20 steps http://super-scidac.org3310/18/12
34
SUPER Tool Integration Led by Al Malony, University of Oregon TAU replaces HPCToolkit as primary triage tool TAUdb replaces PERI performance database New tools to enable performance portability CUDA-CHiLL and OpenMP-CHiLL PAPI GPU, Q, RAPL, MIC Integration of autotuning framework and TAUmon Enable online autotuning Already using online binary patch for empirical tuning experiments SUPER http://super-scidac.org3410/18/12
35
SUPER Meeting Schedule OregonSept. 21-22, 2011 UNC RENCIMarch29-30, 2012 ANLSept.25-26, 2012 RiceMarch2013 LBNLSept.2013 UtahFeb.2014 MarylandSept.2014 UCSD SDSCMarch2015 ORNL & UTKSept.2016 USC ISIMarch2017 SUPER http://super-scidac.org3510/18/12
36
SUPER Management Structure Overall management – Bob Lucas and David Bailey Distributed leadership of research – Follows PERI model, adapts as needed Weekly project teleconferences – Tuesdays, 2 PM Eastern – Wednesdays, noon Eastern – First Wednesday of the month is management topics – Technically focused otherwise Regular face-to-face project meetings – Monday mornings at SC – All hands, twice per year, each institution takes a turn hosting Allows students and staff to attend at least once SUPER http://super-scidac.org3610/18/12
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.