SC’13: Hands-on Practical Hybrid Parallel Application Performance Engineering Introduction to VI-HPS Brian Wylie Jülich Supercomputing Centre.

Slides:



Advertisements
Similar presentations
Forschungszentrum Jülich in der Helmholtz-Gesellschaft February 2007 – OGF19 A Collaborative Online Visualization and Steering (COVS) Framework for e-Science.
Advertisements

Barcelona Supercomputing Center. The BSC-CNS objectives: R&D in Computer Sciences, Life Sciences and Earth Sciences. Supercomputing support to external.
Performance Analysis Tools for High-Performance Computing Daniel Becker
K T A U Kernel Tuning and Analysis Utilities Department of Computer and Information Science Performance Research Laboratory University of Oregon.
Supercomputing Institute for Advanced Computational Research © 2009 Regents of the University of Minnesota. All rights reserved. The Minnesota Supercomputing.
Robert Bell, Allen D. Malony, Sameer Shende Department of Computer and Information Science Computational Science.
TAU Parallel Performance System DOD UGC 2004 Tutorial Allen D. Malony, Sameer Shende, Robert Bell Univesity of Oregon.
Nick Trebon, Alan Morris, Jaideep Ray, Sameer Shende, Allen Malony {ntrebon, amorris, Department of.
On the Integration and Use of OpenMP Performance Tools in the SPEC OMP2001 Benchmarks Bernd Mohr 1, Allen D. Malony 2, Rudi Eigenmann 3 1 Forschungszentrum.
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
MDOP 2010: Diagnostic and Recovery Toolset (DaRT) Speaker Fabrizio Grossi
Performance Technology for Complex Parallel Systems REFERENCES.
VMWare Workstation Installation. Starting Vmware Workstation Go to the start menu and start the VMware Workstation program. *Note: The following instructions.
Parallelization with the Matlab® Distributed Computing Server CBI cluster December 3, Matlab Parallelization with the Matlab Distributed.
1 1 Hybrid Cloud Solutions (Private with Public Burst) Accelerate and Orchestrate Enterprise Applications.
1 Score-P – A Joint Performance Measurement Run-Time Infrastructure for Periscope, Scalasca, TAU, and Vampir Markus Geimer 2), Bert Wesarg 1), Brian Wylie.
Go to & click on Install/Download Menu optionhttp://osddlinux.osdd.net/ Go to & click on Install/Download.
DKRZ Tutorial 2013, Hamburg Analysis report examination with CUBE Markus Geimer Jülich Supercomputing Centre.
Analysis report examination with CUBE Alexandru Calotoiu German Research School for Simulation Sciences (with content used with permission from tutorials.
SICSA student induction day, 2009Slide 1 Social Simulation Tutorial Session 6: Introduction to grids and cloud computing International Symposium on Grid.
© Spinnaker Labs, Inc. Google Cluster Computing Faculty Training Workshop Open Source Tools for Teaching.
1 Performance Analysis with Vampir DKRZ Tutorial – 7 August, Hamburg Matthias Weber, Frank Winkler, Andreas Knüpfer ZIH, Technische Universität.
WORK ON CLUSTER HYBRILIT E. Aleksandrov 1, D. Belyakov 1, M. Matveev 1, M. Vala 1,2 1 Joint Institute for nuclear research, LIT, Russia 2 Institute for.
Presented by National Institute for Computational Sciences (NICS): Education, Outreach and Training Julia C. White User Support National Institute for.
Overview of the Computer Resource Team (CRT) Blaise Barney (LLNL) Rob Cunningham (LANL) Barbara Jennings (Sandia) PSAAP Kickoff Meeting July 8, 2008 Albuquerque,
The Pipeline Processing Framework LSST Applications Meeting IPAC Feb. 19, 2008 Raymond Plante National Center for Supercomputing Applications.
TRACEREP: GATEWAY FOR SHARING AND COLLECTING TRACES IN HPC SYSTEMS Iván Pérez Enrique Vallejo José Luis Bosque University of Cantabria TraceRep IWSG'15.
Introduction to Interactive Media Interactive Media Tools: Software.
Score-P – A Joint Performance Measurement Run-Time Infrastructure for Periscope, Scalasca, TAU, and Vampir Alexandru Calotoiu German Research School for.
Presented by KOJAK and SCALASCA Jack Dongarra, Shirley Moore, Farzona Pulatova, and Fengguang Song University of Tennessee and Oak Ridge National Laboratory.
VAMPIR. Visualization and Analysis of MPI Resources Commercial tool from PALLAS GmbH VAMPIRtrace - MPI profiling library VAMPIR - trace visualization.
Technology + Process SDCI HPC Improvement: High-Productivity Performance Engineering (Tools, Methods, Training) for NSF HPC Applications Rick Kufrin *,
Profile Analysis with ParaProf Sameer Shende Performance Reseaerch Lab, University of Oregon
Martin Schulz Center for Applied Scientific Computing Lawrence Livermore National Laboratory Lawrence Livermore National Laboratory, P. O. Box 808, Livermore,
Performance Analysis Tool List Hans Sherburne Adam Leko HCS Research Laboratory University of Florida.
SALSASALSASALSASALSA FutureGrid Venus-C June Geoffrey Fox
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
Software Overview Environment, libraries, debuggers, programming tools and applications Jonathan Carter NUG Training 3 Oct 2005.
Hands-on: NPB-MZ-MPI / BT VI-HPS Team. 10th VI-HPS Tuning Workshop, October 2012, Garching Local Installation VI-HPS tools accessible through the.
A New Parallel Debugger for Franklin: DDT Katie Antypas User Services Group NERSC User Group Meeting September 17, 2007.
LINUX MINT 13 Introduction Linux Mint is a modern, elegant and comfortable operating system which is both powerful and easy to use. The minimum requirements.
Debugging parallel programs. Breakpoint debugging Probably the most widely familiar method of debugging programs is breakpoint debugging. In this method,
1 Typical performance bottlenecks and how they can be found Bert Wesarg ZIH, Technische Universität Dresden.
Lawrence Livermore National Laboratory S&T Principal Directorate - Computation Directorate Tools and Scalable Application Preparation Project Computation.
21 Sep UPC Performance Analysis Tool: Status and Plans Professor Alan D. George, Principal Investigator Mr. Hung-Hsun Su, Sr. Research Assistant.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
Future Grid Future Grid Overview. Future Grid Future GridFutureGridFutureGrid The goal of FutureGrid is to support the research that will invent the future.
The Performance Evaluation Research Center (PERC) Participating Institutions: Argonne Natl. Lab.Univ. of California, San Diego Lawrence Berkeley Natl.
Hands-On Virtual Computing
Parallel Performance Measurement of Heterogeneous Parallel Systems with GPUs Allen D. Malony, Scott Biersdorff, Sameer Shende, Heike Jagode†, Stanimire.
Extreme Scalability Working Group (XS-WG): Status Update Nick Nystrom Director, Strategic Applications Pittsburgh Supercomputing Center May 20, 2010.
Other Tools HPC Code Development Tools July 29, 2010 Sue Kelly Sandia is a multiprogram laboratory operated by Sandia Corporation, a.
Derek Weitzel Grid Computing. Background B.S. Computer Engineering from University of Nebraska – Lincoln (UNL) 3 years administering supercomputers at.
Presented by Jack Dongarra University of Tennessee and Oak Ridge National Laboratory KOJAK and SCALASCA.
Course 03 Basic Concepts assist. eng. Jánó Rajmond, PhD
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Overview of Atmosphere.
Online Performance Analysis and Visualization of Large-Scale Parallel Applications Kai Li, Allen D. Malony, Sameer Shende, Robert Bell Performance Research.
© 2007 IBM Corporation IBM Software Strategy Group IBM Google Announcement on Internet-Scale Computing (“Cloud Computing Model”) Oct 8, 2007 IBM Confidential.
Introduction to Performance Tuning Chia-heng Tu PAS Lab Summer Workshop 2009 June 30,
ENEA GRID & JPNM WEB PORTAL to create a collaborative development environment Dr. Simonetta Pagnutti JPNM – SP4 Meeting Edinburgh – June 3rd, 2013 Italian.
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
Kai Li, Allen D. Malony, Sameer Shende, Robert Bell
DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING CLOUD COMPUTING
Introduction to Parallel Computing: MPI, OpenMP and Hybrid Programming
Performance Technology for Scalable Parallel Systems
Desktop Virtualization
LinkSCEEM-2: A computational resource for the development of Computational Sciences in the Eastern Mediterranean Mostafa Zoubi SESAME Outreach SESAME,
TAU integration with Score-P
Introduction to XSEDE Resources HPC Workshop 08/21/2017
Presentation transcript:

SC’13: Hands-on Practical Hybrid Parallel Application Performance Engineering Introduction to VI-HPS Brian Wylie Jülich Supercomputing Centre

SC’13: Hands-on Practical Hybrid Parallel Application Performance Engineering Virtual Institute – High Productivity Supercomputing Goal: Improve the quality and accelerate the development process of complex simulation codes running on highly-parallel computer systems Start-up funding (2006–2011) by Helmholtz Association of German Research Centres Activities –Development and integration of HPC programming tools Correctness checking & performance analysis –Training workshops –Service Support lists Application engagement –Academic workshops 2

SC’13: Hands-on Practical Hybrid Parallel Application Performance Engineering VI-HPS partners (founders) 3 Forschungszentrum Jülich ■ Jülich Supercomputing Centre RWTH Aachen University ■ Centre for Computing & Communication Technical University of Dresden ■ Centre for Information Services & HPC University of Tennessee (Knoxville) ■ Innovative Computing Laboratory

SC’13: Hands-on Practical Hybrid Parallel Application Performance Engineering VI-HPS partners (cont.) 4 Barcelona Supercomputing Center ■ Centro Nacional de Supercomputación German Research School ■ Laboratory of Parallel Programming Lawrence Livermore National Lab. ■ Centre for Applied Scientific Computing Technical University of Munich ■ Chair for Computer Architecture University of Oregon ■ Performance Research Laboratory University of Stuttgart ■ HPC Centre University of Versailles St-Quentin ■ LRC ITACA

SC’13: Hands-on Practical Hybrid Parallel Application Performance Engineering Productivity tools MUST ■ MPI usage correctness checking PAPI ■ Interfacing to hardware performance counters Periscope ■ Automatic analysis via an on-line distributed search Scalasca ■ Large-scale parallel performance analysis TAU ■ Integrated parallel performance system Vampir ■ Interactive graphical trace visualization & analysis Score-P ■ Community instrumentation & measurement infrastructure 5

SC’13: Hands-on Practical Hybrid Parallel Application Performance Engineering Productivity tools (cont.) KCachegrind ■ Callgraph-based cache analysis [x86 only] MAQAO ■ Assembly instrumentation & optimization [x86 only] mpiP/mpiPview ■ MPI profiling tool and analysis viewer Open MPI ■ Integrated memory checking Open|Speedshop ■ Integrated parallel performance analysis environment Paraver/Extrae ■ Event tracing and graphical trace visualization & analysis Rubik ■ Process mapping generation & optimization [BG only] SIONlib ■ Optimized native parallel file I/O STAT ■ Stack trace analysis tools 6

SC’13: Hands-on Practical Hybrid Parallel Application Performance Engineering Technologies and their integration 7 Optimization Visual trace analysis Automatic profile & trace analysis Error & anomaly detection Hardware monitoring Execution SCALASCA VAMPIR / PARAVER PAPI MUST PERISCOPE KCACHEGRIND TAU RUBIK / MAQAO SYSMON / SIONLIB / OPENMPI STAT SCORE-P MPIP / O|SS / LWM2

SC’13: Hands-on Practical Hybrid Parallel Application Performance Engineering Disclaimer Tools will not automatically make you, your applications or computer systems more productive. However, they can help you understand how your parallel code executes and when / where it's necessary to work on correctness and performance issues. 8

SC’13: Hands-on Practical Hybrid Parallel Application Performance Engineering VI-HPS training & Tuning Workshops Goals –Give an overview of the programming tools suite –Explain the functionality of individual tools –Teach how to use the tools effectively –Offer hands-on experience and expert assistance using tools –Receive feedback from users to guide future development For best results, bring & analyze/tune your own code(s)! VI-HPS Hands-on Tutorial series –SC’08, ICCS’09, SC’09, Cluster’10, SC’10, SC’11, EuroMPI’12, XSEDE’13 (San Diego), SC’13 (Denver) VI-HPS Tuning Workshop series –2008 (Aachen & Dresden), 2009 (Jülich & Bremen), 2010 (Garching & Amsterdam/NL), 2011 (Stuttgart & Aachen), 2012 (St-Quentin/F & Garching), 2013 (Saclay/F & Jülich) 9

SC’13: Hands-on Practical Hybrid Parallel Application Performance Engineering Recent & upcoming events ISC Extreme Scale Tools Tutorial (16 Jun 2013, Leipzig) EuroPar VI-HPS Tools Tutorial (26 Sep 2013, Aachen) 12th VI-HPS Tuning Workshop (7-11 Oct 2013, Jülich) –Hosted by Jülich Supercomputing Centre, FZJ, Germany –Using PRACE Tier-0 Juqueen BlueGene/Q system –Score-P, Scalasca, Vampir, TAU, Periscope, Paraver, MUST,... Further events to be determined –(one-day) tutorials With guided exercises usually using a Live-DVD –(multi-day) training workshops With your own applications on actual HPC systems Check for announced eventswww.vi-hps.org/training Contact us if you might be interested in hosting an event 10

SC’13: Hands-on Practical Hybrid Parallel Application Performance Engineering VI-HPS Linux Live DVD/ISO Bootable Linux installation on DVD (or USB memory stick) Includes everything needed to try out our parallel tools on an 64-bit x86-architecture notebook computer VI-HPS tools: MUST, PAPI, Score-P, Periscope, Scalasca, TAU, Vampir* Also: Eclipse/PTP, TotalView*, etc. * time/capability-limited evaluation licences provided for commercial products GCC (w/ OpenMP), OpenMPI Manuals/User Guides Tutorial exercises & examples Produced by U. Oregon PRL Sameer Shende 11

SC’13: Hands-on Practical Hybrid Parallel Application Performance Engineering VI-HPS Linux Live ISO ISO image approximately 4GB –download latest version from website – –optionally create bootable DVD or USB drive Boot directly from disk –enables hardware counter access and offers best performance, but no save/resume Boot within virtual machine –faster boot time and can save/resume state, but may not allow hardware counter access Boots into Linux environment for HPC –supports building and running provided MPI and/or OpenMP parallel application codes –and experimentation with VI-HPS (and third-party) tools 12