Luiz DeRose © Cray Inc. CSCS July 13-15, 2009 Using Hardware Performance Counters on the Cray XT Luiz DeRose Programming Environment Director.

Slides:



Advertisements
Similar presentations
More on Processes Chapter 3. Process image _the physical representation of a process in the OS _an address space consisting of code, data and stack segments.
Advertisements

ARCHER Tips and Tricks A few notes from the CSE team.
CPE 731 Advanced Computer Architecture ILP: Part V – Multiple Issue Dr. Gheith Abandah Adapted from the slides of Prof. David Patterson, University of.
Lecture 6: Multicore Systems
Miss Penalty Reduction Techniques (Sec. 5.4) Multilevel Caches: A second level cache (L2) is added between the original Level-1 cache and main memory.
© Cray Inc. CSC, Finland September 21-24, XT3XT4XT5XT6 Number of cores/socket Number of cores/node Clock Cycle (CC) ??
Profiling your application with Intel VTune at NERSC
Data Marshaling for Multi-Core Architectures M. Aater Suleman Onur Mutlu Jose A. Joao Khubaib Yale N. Patt.
Intel® performance analyze tools Nikita Panov Idrisov Renat.
NewsFlash!! Earth Simulator no longer #1. In slightly less earthshaking news… Homework #1 due date postponed to 10/11.
Automated Instrumentation and Monitoring System (AIMS)
Chapter 12 CPU Structure and Function. CPU Sequence Fetch instructions Interpret instructions Fetch data Process data Write data.
Computer Organization and Architecture
MCTS GUIDE TO MICROSOFT WINDOWS 7 Chapter 10 Performance Tuning.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment Chapter 11: Monitoring Server Performance.
Chapter 14 Chapter 14: Server Monitoring and Optimization.
Chapter 12 CPU Structure and Function. Example Register Organizations.
1 COMP 206: Computer Architecture and Implementation Montek Singh Wed, Nov 9, 2005 Topic: Caches (contd.)
1 Presenter: Chien-Chih Chen Proceedings of the 2002 workshop on Memory system performance.
Performance Engineering and Debugging HPC Applications David Skinner
A Characterization of Processor Performance in the VAX-11/780 From the ISCA Proceedings 1984 Emer & Clark.
Basics of Operating Systems March 4, 2001 Adapted from Operating Systems Lecture Notes, Copyright 1997 Martin C. Rinard.
Multi-core Programming Thread Profiler. 2 Tuning Threaded Code: Intel® Thread Profiler for Explicit Threads Topics Look at Intel® Thread Profiler features.
Performance Measurement on kraken using fpmpi and craypat Kwai Wong NICS at UTK / ORNL March 24, 2010.
MCTS Guide to Microsoft Windows 7
Simultaneous Multithreading: Maximizing On-Chip Parallelism Presented By: Daron Shrode Shey Liggett.
Multi-core Programming VTune Analyzer Basics. 2 Basics of VTune™ Performance Analyzer Topics What is the VTune™ Performance Analyzer? Performance tuning.
Parallel Programming Models Jihad El-Sana These slides are based on the book: Introduction to Parallel Computing, Blaise Barney, Lawrence Livermore National.
Topics covered: Memory subsystem CSE243: Introduction to Computer Architecture and Hardware/Software Interface.
1 Interconnects Shared address space and message passing computers can be constructed by connecting processors and memory unit using a variety of interconnection.
Application performance and communication profiles of M3DC1_3D on NERSC babbage KNC with 16 MPI Ranks Thanh Phung, Intel TCAR Woo-Sun Yang, NERSC.
Overview of CrayPat and Apprentice 2 Adam Leko UPC Group HCS Research Laboratory University of Florida Color encoding key: Blue: Information Red: Negative.
Martin Schulz Center for Applied Scientific Computing Lawrence Livermore National Laboratory Lawrence Livermore National Laboratory, P. O. Box 808, Livermore,
Performance Monitoring Tools on TCS Roberto Gomez and Raghu Reddy Pittsburgh Supercomputing Center David O’Neal National Center for Supercomputing Applications.
Chapter 8 CPU and Memory: Design, Implementation, and Enhancement The Architecture of Computer Hardware and Systems Software: An Information Technology.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment, Enhanced Chapter 11: Monitoring Server Performance.
1 06/09/2011, COSMO GM Xavier Lapillonne Porting the physical parametrizations on GPU using directives X. Lapillonne, O. Fuhrer Eidgenössisches Departement.
Belgrade, 25 September 2014 George S. Markomanolis, Oriol Jorba, Kim Serradell Performance analysis Tools: a case study of NMMB on Marenostrum.
1 How will execution time grow with SIZE? int array[SIZE]; int sum = 0; for (int i = 0 ; i < ; ++ i) { for (int j = 0 ; j < SIZE ; ++ j) { sum +=
NUG Meeting Performance Profiling Using hpmcount, poe+ & libhpm Richard Gerber NERSC User Services
Tool Visualizations, Metrics, and Profiled Entities Overview [Brief Version] Adam Leko HCS Research Laboratory University of Florida.
1 Computer Systems II Introduction to Processes. 2 First Two Major Computer System Evolution Steps Led to the idea of multiprogramming (multiple concurrent.
Operating Systems ECE344 Ashvin Goel ECE University of Toronto Virtual Memory Hardware.
Processes and Virtual Memory
Processor Memory Processor-memory bus I/O Device Bus Adapter I/O Device I/O Device Bus Adapter I/O Device I/O Device Expansion bus I/O Bus.
Single Node Optimization Computational Astrophysics.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 2.
PAPI on Blue Gene L Using network performance counters to layout tasks for improved performance.
Sunpyo Hong, Hyesoon Kim
Projections - A Step by Step Tutorial By Chee Wai Lee For the 2004 Charm++ Workshop.
LECTURE 12 Virtual Memory. VIRTUAL MEMORY Just as a cache can provide fast, easy access to recently-used code and data, main memory acts as a “cache”
Copyright ©: Nahrstedt, Angrave, Abdelzaher1 Memory.
Tuning Threaded Code with Intel® Parallel Amplifier.
Embedded Real-Time Systems Processing interrupts Lecturer Department University.
1© Copyright 2015 EMC Corporation. All rights reserved. NUMA(YEY) BY JACOB KUGLER.
Processes and threads.
CSC 4250 Computer Architectures
How will execution time grow with SIZE?
Copyright © 2011, Elsevier Inc. All rights Reserved.
/ Computer Architecture and Design
Lecture 14 Virtual Memory and the Alpha Memory Hierarchy
Alpha Microarchitecture
Virtual Memory Hardware
* From AMD 1996 Publication #18522 Revision E
Allen D. Malony Computer & Information Science Department
Projections Overview Ronak Buch & Laxmikant (Sanjay) Kale
Memory System Performance Chapter 3
Presentation transcript:

Luiz DeRose © Cray Inc. CSCS July 13-15, 2009 Using Hardware Performance Counters on the Cray XT Luiz DeRose Programming Environment Director Cray Inc.

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 22 Simplified memory hierachy on the Quad Core AMD Opteron …... registers L1 data cache L2 cache 16 SSE2 128-bit registers bit registers 2 x 16 Bytes per clock loads or 1 x 16 Bytes per clock load and store, (76.8 GB/s or 38.4 GB/s on 2.4 Ghz) Main memory  64 Byte cache line, 8 banks  complete data cache lines are loaded from main memory, if not in L2 or L3 cache  if L1 data cache needs to be refilled, then storing back to L2 cache, if L2 needs to be refilled, storing back to L3  Items in L1 and L2 are exclusive, L3 is “sharing aware”  64 Byte cache line  write back cache: data offloaded from L1 data cache are stored here first until they are flushed out to main memory 16 Bytes wide => 10.6 GB/s for DDR2-667, 73ns 16 Bytes per clock, 38.4 GB/s BW …... Shared L3 cache 32 GB/s Remote memory 8GB/s over coherent Hyper Transport, 115ns

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 3 Hardware Performance Counters  AMD Opteron Hardware Performance Counters Four 48-bit performance counters.  Each counter can monitor a single event Count specific processor events »the processor increments the counter when it detects an occurrence of the event » (e.g., cache misses) Duration of events »the processor counts the number of processor clocks it takes to complete an event »(e.g., the number of clocks it takes to return data from memory after a cache miss) Time Stamp Counters (TSC)  Cycles (user time)

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 4 PAPI Predefined Events  Common set of events deemed relevant and useful for application performance tuning Accesses to the memory hierarchy, cycle and instruction counts, functional units, pipeline status, etc. The “papi_avail” utility shows which predefined events are available on the system – execute on compute node  PAPI also provides access to native events The “papi_native_avail” utility lists all AMD native events available on the system – execute on compute node  Information on PAPI and AMD native events pat_help counters man papi_counters For more information on AMD counters: 

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 5 Hardware Counters Selection  PAT_RT_HWPC | Specifies hardware counter events to be monitored  A set number can be used to select a group of predefined hardware counters events (recommended) CrayPat provides 19 groups on the Cray XT systems  Alternatively a list of hardware performance counter event names can be used Maximum of 4 events  Both formats can be specified at the same time, with later definitions overriding previous definitions  Hardware counter events are not collected by default  Hardware counters collection is not supported with sampling on systems running Catamount on the compute nodes

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 6 Accuracy Issues  Pay attention to what is not measured: Out-of-order processors Speculation Lack of standard on what is counted  Microbenchmarks can help determine accuracy of the hardware counters  For more information on AMD counters: architecture manuals:  user interface Kernel Hardware counters  Granularity of the measured code If not sufficiently large enough, overhead of the counter interfaces may dominate

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 7 Hardware Performance Counters Hardware performance counter events: PAPI_TOT_INS Instructions completed PAPI_L1_DCA Level 1 data cache accesses PAPI_FP_OPS Floating point operations DATA_CACHE_MISSES Data Cache Misses CYCLES_USER User Cycles (approx, from clock ticks) Estimated minimum overhead per call of a traced function, which was subtracted from the data shown in this report (for raw data, use the option: -s overhead=include): PAPI_TOT_INS instr PAPI_L1_DCA refs PAPI_FP_OPS ops DATA_CACHE_MISSES misses CYCLES_USER cycles Time microseconds

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 8 PAPI_TLB_DM Data translation lookaside buffer misses PAPI_L1_DCA Level 1 data cache accesses PAPI_FP_OPS Floating point operations DC_MISS Data Cache Miss User_Cycles Virtual Cycles ======================================================================== USER Time% 68.0% Time secs Imb.Time -- secs Imb.Time% -- Calls 0.001M/sec calls PAPI_L1_DCM M/sec misses PAPI_TLB_DM 2.114M/sec misses PAPI_L1_DCA M/sec refs PAPI_FP_OPS M/sec ops User time (approx) secs cycles 96.5%Time Average Time per Call sec CrayPat Overhead : Time 0.2% HW FP Ops / User time M/sec ops 4.3%peak(DP) HW FP Ops / WCT M/sec Computational intensity 0.17 ops/cycle 1.39 ops/ref MFLOPS (aggregate) M/sec TLB utilization refs/miss avg uses D1 cache hit,miss ratios 94.8% hits 5.2% misses D1 cache utilization (M) refs/miss avg uses ======================================================================== PAT_RT_HWPC=1 (Summary with TLB) PAT_RT_HWPC=1 Flat profile data Hard counts Derived metrics

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 9 PAT_RT_HWPC=2 (L1 and L2 Metrics) ======================================================================== USER Time% 57.5% Time secs Imb.Time -- secs Imb.Time% -- Calls 0.001M/sec calls REQUESTS_TO_L2:DATA M/sec req DATA_CACHE_REFILLS: L2_MODIFIED:L2_OWNED: L2_EXCLUSIVE:L2_SHARED 9.691M/sec fills DATA_CACHE_REFILLS_FROM_SYSTEM: ALL M/sec fills PAPI_L1_DCA M/sec refs User time (approx) secs cycles 98.6%Time Average Time per Call sec CrayPat Overhead : Time 0.1% D1 cache hit,miss ratio (R) 88.4% hits 11.6% misses D1 cache utilization 8.65 refs/refill avg uses D2 cache hit,miss ratio 74.8% hits 25.2% misses D1+D2 cache hit,miss ratio 91.8% hits 8.2% misses D1+D2 cache utilization refs/miss avg uses System to D1 refill M/sec lines System to D1 bandwidth MB/sec bytes L2 to Dcache bandwidth MB/sec bytes ========================================================================

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 10 PAT_RT_HWPC=3 (Bandwidth) ======================================================================== USER / mlwxyz_ Time% 44.0% Time Imb.Time Imb.Time% 1.0% Calls 10 QUADWORDS_WRITTEN_TO_SYSTEM: ALL M/sec ops DATA_CACHE_REFILLS: L2_MODIFIED:L2_OWNED: L2_EXCLUSIVE:L2_SHARED M/sec fills DATA_CACHE_REFILLS_FROM_SYSTEM: ALL M/sec fills DATA_CACHE_LINES_EVICTED:ALL M/sec ops User time (approx) secs cycles Average Time per Call sec/call Cycles secs cycles User time (approx) secs cycles Utilization rate 99.4% D2 cache hit ratio 43.3% System to D1 refill M/sec lines System to D1 bandwidth MB/sec bytes L2 to Dcache bandwidth MB/sec bytes L2 to System BW per core MB/sec bytes ========================================================================

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 11 PAT_RT_HWPC=5 (Floating point mix) ======================================================================== USER Time% 58.5% Time secs Imb.Time -- secs Imb.Time% -- Calls 0.001M/sec calls RETIRED_MMX_AND_FP_INSTRUCTIONS: PACKED_SSE_AND_SSE M/sec instr PAPI_FML_INS M/sec ops PAPI_FAD_INS M/sec ops PAPI_FDV_INS 7.258M/sec ops User time (approx) secs cycles 97.0%Time Average Time per Call sec CrayPat Overhead : Time 0.2% HW FP Ops / Cycles 0.19 ops/cycle HW FP Ops / User time M/sec ops 4.7%peak(DP) HW FP Ops / WCT M/sec FP Multiply / FP Ops 35.0% FP Add / FP Ops 65.0% MFLOPS (aggregate) M/sec ========================================================================

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 12 PAT_RT_HWPC=12 (QC Vectorization) ======================================================================== USER Time% 62.6% Time secs Imb.Time -- secs Imb.Time% -- Calls 0.001M/sec calls RETIRED_SSE_OPERATIONS: SINGLE_ADD_SUB_OPS: SINGLE_MUL_OPS 0 ops RETIRED_SSE_OPERATIONS: DOUBLE_ADD_SUB_OPS: DOUBLE_MUL_OPS M/sec ops RETIRED_SSE_OPERATIONS: SINGLE_ADD_SUB_OPS: SINGLE_MUL_OPS:OP_TYPE 0 ops RETIRED_SSE_OPERATIONS: DOUBLE_ADD_SUB_OPS: DOUBLE_MUL_OPS:OP_TYPE M/sec ops User time (approx) secs cycles 99.5%Time Average Time per Call sec CrayPat Overhead : Time 0.2% ========================================================================

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 13 Vectorization Example ======================================================================== USER / calc2_ Time% 28.2% Time secs Imb.Time secs Imb.Time% 11.9% Calls /sec calls RETIRED_SSE_OPERATIONS: SINGLE_ADD_SUB_OPS: SINGLE_MUL_OPS 0 ops RETIRED_SSE_OPERATIONS: DOUBLE_ADD_SUB_OPS: DOUBLE_MUL_OPS M/sec ops RETIRED_SSE_OPERATIONS: SINGLE_ADD_SUB_OPS: SINGLE_MUL_OPS:OP_TYPE 0 ops RETIRED_SSE_OPERATIONS: DOUBLE_ADD_SUB_OPS: DOUBLE_MUL_OPS:OP_TYPE M/sec ops User time (approx) secs cycles 96.2%Time When compiled with fastsse: ======================================================================== USER / calc2_ Time% 24.3% Time secs Imb.Time secs Imb.Time% 26.4% Calls 0.001M/sec calls RETIRED_SSE_OPERATIONS: SINGLE_ADD_SUB_OPS: SINGLE_MUL_OPS 0 ops RETIRED_SSE_OPERATIONS: DOUBLE_ADD_SUB_OPS: DOUBLE_MUL_OPS M/sec ops RETIRED_SSE_OPERATIONS: SINGLE_ADD_SUB_OPS: SINGLE_MUL_OPS:OP_TYPE 0 ops RETIRED_SSE_OPERATIONS: DOUBLE_ADD_SUB_OPS: DOUBLE_MUL_OPS:OP_TYPE M/sec ops User time (approx) secs cycles 100.0%Time

How do I interpret these derived metrics?  The following thresholds are guidelines to identify if optimization is needed: Computational Intensity: < 0.5 ops/ref  This is the ratio of FLOPS by L&S  Measures how well the floating point unit is being used FP Multiply / FP Ops or FP Add / FP Ops: < 25% Vectorization: < 1.5 July 13-15, 2009Luiz DeRose © Cray Inc. Slide 14

Memory Hierarchy Thresholds  TLB utilization: < 90.0% Measures how well the memory hierarchy is being utilized with regards to TLB This metric depends on the computation being single precision or double precision  A page has 4 Kbytes. So, one page fits 512 double precision words or 1024 single precision words TLB utilization < 1 indicates that not all entries on the page are being utilized between two TLB misses  D1 cache utilization: < 1 (D1+D2 cache utilization: < 1) A cache line has 64 bytes (8 double precision words or 16 single precision words) D1 cache utilization < 1 indicates that not all entries on the cache line are being utilized between two cache misses  D1 cache hit (or miss) ratios: 10%)  D2 (L2) cache hit (or miss) ratios: 5%)  D1 + D2 cache hit (or miss) ratios: 8%) D1 and D2 caches on the Opteron are complementary This metric provides a view of the Total Cache hit (miss) ratio July 13-15, 2009Luiz DeRose © Cray Inc. Slide 15

Luiz DeRose © Cray Inc. CSCS July 13-15, 2009 Detecting Load Imbalance on the Cray XT Luiz DeRose Programming Environment Director Cray Inc.

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 17 Motivation for Load Imbalance Analysis  Increasing system software and architecture complexity Current trend in high end computing is to have systems with tens of thousands of processors  This is being accentuated with multi-core processors  Applications have to be very well balanced In order to perform at scale on these MPP systems Efficient application scaling includes a balanced use of requested computing resources  Desire to minimize computing resource “waste” Identify slower paths through code Identify inefficient “stalls” within an application

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 18 Cray Tools Load Imbalance Support  Very few performance tools focus on load imbalance Need standard metrics Need intuitive way of presentation  CrayPat support: Imbalance time and % MPI sync time OpenMP Performance Metrics MPI rank placement suggestions  Cray Apprentice 2 support: Load imbalance visualization

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 19 Imbalance Time  Metric based on execution time  It is dependent on the type of activity: User functions Imbalance time = Maximum time – Average time Synchronization (Collective communication and barriers) Imbalance time = Average time – Minimum time  Identifies computational code regions and synchronization calls that could benefit most from load balance optimization  Estimates how much overall program time could be saved if corresponding section of code had a perfect balance Represents upper bound on “potential savings” Assumes other processes are waiting, not doing useful work while slowest member finishes

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 20 Load balance metric - rationale

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 21 Imbalance %  Represents % of resources available for parallelism that is “wasted”  Corresponds to % of time that rest of team is not engaged in useful work on the given function  Perfectly balanced code segment has imbalance of 0%  Serial code segment has imbalance of 100% Imbalance% = Imbalance time Max Time X N - 1 N 100 X

Call Tree Visualization (Swim3d) July 13-15, 2009Luiz DeRose © Cray Inc. Slide 22

Discrete Unit of Help (DUH Button) July 13-15, 2009Luiz DeRose © Cray Inc. Slide 23

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 24 Load Distribution

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 25 Profile with Load Distribution by Groups Table 1: Profile by Function Group and Function Time % | Time |Imb. Time | Imb. | Calls |Group | | | Time % | | Function | | | | | PE='HIDE' 100.0% | | -- | -- | 2530 |Total | | 83.7% | | -- | -- | 303 |USER || || 32.4% | | | 6.8% | 98 |calc3_ || 27.7% | | | 6.0% | 100 |calc2_ || 21.0% | | | 2.8% | 100 |calc1_ || 2.0% | | | 3.3% | 1 |inital_ ||========================================================= | 16.3% | | -- | -- | 2227 |MPI || || 12.7% | | | 64.1% | 351 |mpi_waitall_ || 2.2% | | | 59.7% | 936 |mpi_isend_ || 1.4% | | | 44.7% | 936 |mpi_irecv_ |==========================================================

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 26 MPI Sync Time  Measure load imbalance in programs instrumented to trace MPI functions to determine if MPI ranks arrive at collectives together  Separates potential load imbalance from data transfer  Sync times reported by default if MPI functions traced  If desired, PAT_RT_MPI_SYNC=0 deactivated this feature

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 27 MPI Sync Time Statistics Time % | Time |Imb. Time | Imb. | Calls |Group | | | Time % | | Function | | | | | PE='HIDE' 100.0% | | -- | -- | |Total | | 76.5% | | -- | -- | 4752 |USER || || 96.0% | | | 3.3% | 12 |sweep_ || 3.2% | | | 3.1% | 12 |source_ || 0.3% | | | 2.9% | 12 |flux_err_ || 0.2% | | | 22.8% | 2280 |snd_real_ || 0.1% | | | 2.9% | 1 |initialize_ || 0.1% | | | 3.2% | 1 |initxs_ || 0.1% | | | 40.3% | 2280 |rcv_real_ ||========================================================= | 16.6% | | -- | -- | 4603 |MPI || || 93.9% | | | 20.7% | 2280 |mpi_recv_ || 5.9% | | | 17.7% | 2280 |mpi_send_ || 0.2% | | | 34.4% | 32 |mpi_allreduce_ ||========================================================= | 6.3% | | -- | -- | 39 |MPI_SYNC || || 61.1% | | | 45.7% | 4 |mpi_bcast_(sync) || 38.7% | | | 63.2% | 32 |mpi_allreduce_(sync) || 0.1% | | | 35.5% | 3 |mpi_barrier_(sync) |==========================================================

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 28 A() { } OMP loop Source code main() { } A() OMP parallel OMP end parallel OpenMP (Ideal) Instrumentation run-time library Compiler generated A() { } // Region main() { } A() master thread Outlined { Function } // Do all threads do I=start,end loop body enddo OpenMP loop { } Parallel_enter Parallel_exit Parallel_begin Parallel_end Loop_enter Loop_exit Function_enter Function_exit

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 29 CrayPat OpenMP Performance Metrics  Per-thread timings  Overhead incurred at enter/exit of Parallel regions Worksharing constructs within parallel regions  Load balance information across threads  Sampling performance data without API  Separate metrics for OpenMP runtime and OpenMP API calls

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 30 OpenMP Data from pat_report  Default view (no options needed to pat_report) Focus on where program is spending its time Shows imbalance across all threads Assumes all requested resources should be used Highlights non-uniform imbalance across threads

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 31 MPI Rank Reorder  MPI rank placement with environment variable  Distributed placement  SMP style placement  Folded rank placement  User provided rank file ????????

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 32 Rank Reorder Example - hycom pat_report -O load_balance Table 2: Load Balance across PE's by FunctionGroup Time % | Cum. | Time | Calls |Group | Time % | | | PE[mmm] 100.0% | 100.0% | | |Total | | 57.7% | 57.7% | | |USER || || 0.5% | 0.5% | | |pe.201 || 0.4% | 58.2% | | |pe.45 || 0.0% | 100.0% | | |pe.184 ||================================================== | 42.3% | 100.0% | | |MPI || || 0.9% | 0.9% | | |pe.184 || 0.3% | 61.9% | | |pe.37 || 0.2% | 100.0% | | |pe.201 |===================================================

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 33 Rank Reorder Example - hycom pat_report -O load_balance -s pe=ALL Table 2: Load Balance across PE's by FunctionGroup Time % | Cum. | Time | Calls |Group | Time % | | | PE 100.0% | 100.0% | | |Total | | 57.7% | 57.7% | | |USER || || 0.5% | 0.5% | | |pe.201 || 0.5% | 1.0% | | |pe.182 || 0.5% | 1.5% | | |pe.200 || 0.5% | 2.0% | | |pe.188 || 0.5% | 2.5% | | |pe || 0.1% | 99.8% | | |pe.248 || 0.1% | 99.9% | | |pe.213 || 0.0% | 99.9% | | |pe.232 || 0.0% | 100.0% | | |pe.240 || 0.0% | 100.0% | | |pe.184 ||===============================================

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 34 Rank Reorder Example - hycom After custom placement (10% performance improvement): Table 2: Load Balance with MPI Sent Message Stats Time % | Time | Sent Msg | Sent Msg | Avg Sent |Group | | Count | Total Bytes | Msg Size | PE[mmm] 100.0% | | | | |Total | | 60.2% | | -- | -- | -- |USER || || 0.5% | | -- | -- | -- |pe.158 || 0.4% | | -- | -- | -- |pe.126 || 0.0% | | -- | -- | -- |pe.184 ||================================================================= | 39.8% | | | | |MPI || || 1.0% | | | | |pe.184 || 0.3% | | | | |pe.126 || 0.3% | | | | |pe.158 |==================================================================

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 35 MPI Rank Placement Suggestions  When to use? Point-to-point communication consumes significant fraction of the program time and have a significant imbalance  pat_report -O mpi_sm_rank_order... When there seems to be a load imbalance of another type  Can get a suggested rank order file based on user time pat_report -O mpi_rank_order...  Can have a different metric for load balance pat_report -O mpi_rank_order -s mro_metric=DATA_CACHE_MISSES...  Information in resulting report Available if MPI functions traced (-g mpi)  Custom placement files automatically generated

MPI Rank Placement Suggestions (cont’d)  See table notes in resulting report from pat_report  Report provides quad core and dual core suggestions  Set MPICH_RANK_REORDER_METHOD environment variable Set to numerical value or MPICH_RANK_ORDER file from pat_report July 13-15, 2009Luiz DeRose © Cray Inc. Slide 36

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 37 Example: -O mpi_rank_order (asura) Notes for table 1: To maximize the locality of point to point communication, choose and specify a Rank Order with small Max and Avg Sent Msg Total Bytes per node for the target number of cores per node. To specify a Rank Order with a numerical value, set the environment variable MPICH_RANK_REORDER_METHOD to the given value. To specify a Rank Order with a letter value 'x', set the environment variable MPICH_RANK_REORDER_METHOD to 3, and copy or link the file MPICH_RANK_ORDER.x to MPICH_RANK_ORDER. Table 1: Sent Message Stats and Suggested MPI Rank Order Sent Msg Total Bytes per MPI rank Max Avg Min Max Min Total Bytes Total Bytes Total Bytes Rank Rank Quad core: Sent Msg Total Bytes per node Rank Max Avg Min Max Node Min Node Order Total Bytes Total Bytes Total Bytes Ranks Ranks d ,124,35,91 86,27,108,63 u ,124,35,91 86,27,108, ,57,58,59 108,109,110, ,57,71,56 74,53,75, ,70,7,71 52,116,53,117

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 38 Example: File MPICH_RANK_ORDER.u (asura) # Suggested custom rank placement: # # pat_report -O mpi_sm_rank_order \ # /home/crayadm/ldr/ASURA/asura10it.x+apa tdt.ap2 # # Targets multi-core processors, based on Sent Msg Total Bytes. # # Program: /work/crayadm/ldr/ASURA/run/asura10it.x # Number PEs: 128 # Cores/Node: 4 # # Heuristic: u # 86,27,108,63,13,67,23,39,70,3,113,17,21,46,40,89 28,36,34,10,7,127,41,105,94,25,12,38,6,75,57,60 56,109,106,68,42,66,43,79,72,45,85,80,33,111,49,107 14,103,114,9,126,52,78,2,55,88,87,118,119,64,15,16 90,102,122,31,37,123,29,59,71,53,98,82,92,124,35,91 5,125,115,11,97,95,30,54,19,4,69,0,62,110,51,112 26,32,121,77,65,100,76,24,58,74,1,18,101,116,84,50 44,96,93,20,83,61,104,47,99,81,120,73,8,117,22,48

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 39 MPI + OpenMP? (some ideas)  When does it pay to add OpenMP to my MPI code? Only add OpenMP when code is network bound Adding OpenMP to memory bound codes will most likely hurt performance rather than help it Look at collective time, excluding sync time: this goes up as network becomes a problem Look at point-to-point wait times: if these go up, network may be a problem

Luiz DeRose © Cray Inc. CSCS July 13-15, 2009 Parallel Performance Analysis and Visualization on the Cray XT Luiz DeRose Programming Environment Director Cray Inc.

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 41 Cray Apprentice 2  Call graph profile  Communication statistics  Time-line view Communication I/O  Activity view  Pair-wise communication statistics  Text reports  Source code mapping  Cray Apprentice 2  is target to help identify and correct: Load imbalance Excessive communication Network contention Excessive serialization I/O Problems

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 42 Statistics Overview Switch Overview display

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 43 Function Profile

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 44 Load Balance View (Aggregated) Min, Avg, and Max Values -1, +1 Std Dev marks

Call Tree View July 13-15, 2009Luiz DeRose © Cray Inc. Slide 45 Function List Load balance overview: Height  Max time Middle bar  Average time Lower bar  Min time Yellow represents imbalance time Zoom Height  exclusive time Width  inclusive time DUH Button: Provides hints for performance tuning Filtered nodes or sub tree

Call Tree View – Function List July 13-15, 2009Luiz DeRose © Cray Inc. Slide 46 Function List off Right mouse click: Node menu e.g., hide/unhide children Sort options % Time, Time, Imbalance % Imbalance time Right mouse click: View menu: e.g., Filter

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 47 Load Balance View (from Call Tree) -1, +1 Std Dev marks Min, Avg, and Max Values

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 48 Source Mapping from Call Tree

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 49 Function Profile

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 50 Distribution by PE, by Call, & by Time

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 51 Environment & Execution Details

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 52 Time Line View (Sweep3D)

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 53 Time Line View (Zoom) User Functions, MPI & SHMEM Line I/O Line

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 54 Time Line View (Fine Grain Zoom)

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 55 Activity View

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 56 Pair-wise Communication

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 57 I/O Overview

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 58 I/O Rates

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 59 Hardware Counters Overview

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 60 Hardware Counters Time Line

Controlling Performance File Size  Performance files can be quite large. There are several run-time environment variables to keep data files down to reasonable sizes  The particular run-time environment variables to use vary depending on the type of experiment being conducted  Sampling: PAT_RT_RECORD_PE  Collect trace for a subset of the PEs PAT_RT_RECORD_THREAD  Collect trace for a subset of the threads PAT_RT_INTERVAL  Specifies the interval, at which the instrumented program is sampled PAT_RT_CALLSTACK  Limit the depth to trace the call stack PAT_RT_HWPC  Avoid collecting hardware counters (unset) PAT_RT_SIZE  The number of contiguous bytes in the text segment available for sampling PAT_RT_WRITE_BUFFER_SIZE  Specifies the size, of a buffer that collects measurement data for a single thread July 13-15, 2009Luiz DeRose © Cray Inc. Slide 61

Controlling Trace File Size  Tracing: PAT_RT_CALLSTACK  Limit the depth to trace the call stack PAT_RT_HWPC  Avoid collecting hardware counters (unset) PAT_RT_RECORD_PE  Collect trace for a subset of the PEs PAT_RT_RECORD_THREAD  Collect trace for a subset of the threads PAT_RT_TRACE_FUNCTION_ARGS  Limit the number of function arguments to be traced PAT_RT_TRACE_FUNCTION_LIMITS  Avoid tracing indicated functions PAT_RT_TRACE_FUNCTION_MAX  Limit the maximum number of traces generated for all functions for a single process PAT_RT_TRACE_THRESHOLD_PCT  Specifies a % of time threshold to enforce when executing in full trace mode PAT_RT_TRACE_THRESHOLD_TIME  Specifies a time threshold to enforce when executing in full trace mode  Use the limit built-in command for ksh(1) or csh(1) to control how much disk space the trace file can consume July 13-15, 2009Luiz DeRose © Cray Inc. Slide 62

July 13-15, 2009Luiz DeRose © Cray Inc. Slide 63 Additional API Functions  int PAT_state (int state) State can have one of the following:  PAT_STATE_ON  PAT_STATE_OFF  PAT_STATE_QUERY  int PAT_record (int state) Controls the state for all threads on the executing PE. As a rule, use PAT_record() unless there is a need for different behaviors for sampling and tracing  int PAT_sampling_state (int state)  int PAT_tracing_state (int state)  int PAT_trace_function (const void *addr, int state) Activates or deactivates the tracing of the instrumented function  int PAT_flush_buffer (void)

Luiz DeRose © Cray Inc. CSCS July 13-15, 2009 Parallel Performance Analysis and Visualization on the Cray XT Questions / Comments Thank You!