The Performance Evaluation Research Center (PERC) Participating Institutions: Argonne Natl. Lab.Univ. of California, San Diego Lawrence Berkeley Natl.

Slides:



Advertisements
Similar presentations
Performance Testing - Kanwalpreet Singh.
Advertisements

Technology Drivers Traditional HPC application drivers – OS noise, resource monitoring and management, memory footprint – Complexity of resources to be.
What do we currently mean by Computational Science? Traditionally focuses on the “hard sciences” and engineering –Physics, Chemistry, Mechanics, Aerospace,
O AK R IDGE N ATIONAL L ABORATORY U. S. D EPARTMENT OF E NERGY Center for Computational Sciences Cray X1 and Black Widow at ORNL Center for Computational.
ASCR Data Science Centers Infrastructure Demonstration S. Canon, N. Desai, M. Ernst, K. Kleese-Van Dam, G. Shipman, B. Tierney.
U.S. Department of Energy’s Office of Science Basic Energy Sciences Advisory Committee Dr. Daniel A. Hitchcock October 21, 2003
Presentation at WebEx Meeting June 15,  Context  Challenge  Anticipated Outcomes  Framework  Timeline & Guidance  Comment and Questions.
The System and Software Development Process Instructor: Dr. Hany H. Ammar Dept. of Computer Science and Electrical Engineering, WVU.
4.1.5 System Management Background What is in System Management Resource control and scheduling Booting, reconfiguration, defining limits for resource.
Implementation methodology for Emerging Reconfigurable Systems With minimum optimization an appreciable speedup of 3x is achievable for this program with.
© Copyright Richard W. Selby and Northrop Grumman Corporation. All rights reserved. 0 Process Synchronization and Stabilization February 2007 Rick.
Automated Analysis and Code Generation for Domain-Specific Models George Edwards Center for Systems and Software Engineering University of Southern California.
R R R CSE870: Advanced Software Engineering (Cheng): Intro to Software Engineering1 Advanced Software Engineering Dr. Cheng Overview of Software Engineering.
Marakas: Decision Support Systems, 2nd Edition © 2003, Prentice-Hall Chapter Chapter 7: Expert Systems and Artificial Intelligence Decision Support.
Chapter 12: Simulation and Modeling Invitation to Computer Science, Java Version, Third Edition.
All content in this presentation is protected – © 2008 American Power Conversion Corporation Rael Haiboullin System Engineer Capacity Manager.
CMSC 611: Advanced Computer Architecture Benchmarking Some material adapted from Mohamed Younis, UMBC CMSC 611 Spr 2003 course slides Some material adapted.
Module 3: Business Information Systems
1 Building National Cyberinfrastructure Alan Blatecky Office of Cyberinfrastructure EPSCoR Meeting May 21,
An Automated Component-Based Performance Experiment and Modeling Environment Van Bui, Boyana Norris, Lois Curfman McInnes, and Li Li Argonne National Laboratory,
The Performance Evaluation Research Center (PERC) Patrick H. Worley, ORNL Participating Institutions: Argonne Natl. Lab.Univ. of California, San Diego.
© Fujitsu Laboratories of Europe 2009 HPC and Chaste: Towards Real-Time Simulation 24 March
N By: Md Rezaul Huda Reza n
Parallel Programming Models Jihad El-Sana These slides are based on the book: Introduction to Parallel Computing, Blaise Barney, Lawrence Livermore National.
DCS Overview MCS/DCS Technical Interchange Meeting August, 2000.
1 ISA&D7‏/8‏/ ISA&D7‏/8‏/2013 Systems Development Life Cycle Phases and Activities in the SDLC Variations of the SDLC models.
QCD Project Overview Ying Zhang September 26, 2005.
Evaluation of software engineering. Software engineering research : Research in SE aims to achieve two main goals: 1) To increase the knowledge about.
SC’13: Hands-on Practical Hybrid Parallel Application Performance Engineering Introduction to VI-HPS Brian Wylie Jülich Supercomputing Centre.
Future role of DMR in Cyber Infrastructure D. Ceperley NCSA, University of Illinois Urbana-Champaign N.B. All views expressed are my own.
November 13, 2006 Performance Engineering Research Institute 1 Scientific Discovery through Advanced Computation Performance Engineering.
Uncovering the Multicore Processor Bottlenecks Server Design Summit Shay Gal-On Director of Technology, EEMBC.
R. Ryne, NUG mtg: Page 1 High Energy Physics Greenbook Presentation Robert D. Ryne Lawrence Berkeley National Laboratory NERSC User Group Meeting.
Model-Driven Analysis Frameworks for Embedded Systems George Edwards USC Center for Systems and Software Engineering
University of Southern California Center for Systems and Software Engineering Model-Based Software Engineering Supannika Koolmanojwong Spring 2013.
Martin Schulz Center for Applied Scientific Computing Lawrence Livermore National Laboratory Lawrence Livermore National Laboratory, P. O. Box 808, Livermore,
Building an Electron Cloud Simulation using Bocca, Synergia2, TxPhysics and Tau Performance Tools Phase I Doe SBIR Stefan Muszala, PI DOE Grant No DE-FG02-08ER85152.
ASCAC-BERAC Joint Panel on Accelerating Progress Toward GTL Goals Some concerns that were expressed by ASCAC members.
Issues Autonomic operation (fault tolerance) Minimize interference to applications Hardware support for new operating systems Resource management (global.
The System and Software Development Process Instructor: Dr. Hany H. Ammar Dept. of Computer Science and Electrical Engineering, WVU.
ANKITHA CHOWDARY GARAPATI
Debugging parallel programs. Breakpoint debugging Probably the most widely familiar method of debugging programs is breakpoint debugging. In this method,
Scalable Systems Software for Terascale Computer Centers Coordinator: Al Geist Participating Organizations ORNL ANL LBNL.
Software Architecture Evaluation Methodologies Presented By: Anthony Register.
1 DOE Office of Science October 2003 SciDAC Scientific Discovery through Advanced Computing Alan J. Laub.
Software Development Problem Analysis and Specification Design Implementation (Coding) Testing, Execution and Debugging Maintenance.
1 SciDAC High-End Computer System Performance: Science and Engineering Jack Dongarra Innovative Computing Laboratory University of Tennesseehttp://
Computational Science & Engineering meeting national needs Steven F. Ashby SIAG-CSE Chair March 24, 2003.
The System and Software Development Process Instructor: Dr. Hany H. Ammar Dept. of Computer Science and Electrical Engineering, WVU.
1 OFFICE OF ADVANCED SCIENTIFIC COMPUTING RESEARCH The NERSC Center --From A DOE Program Manager’s Perspective-- A Presentation to the NERSC Users Group.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
HEP and NP SciDAC projects: Key ideas presented in the SciDAC II white papers Robert D. Ryne.
Computing Performance Recommendations #10, #11, #12, #15, #16, #17.
ComPASS Summary, Budgets & Discussion Panagiotis Spentzouris, Fermilab ComPASS PI.
CISC Machine Learning for Solving Systems Problems Presented by: Eunjung Park Dept of Computer & Information Sciences University of Delaware Solutions.
UNIT 5.
ERP and Related Technologies
High Risk 1. Ensure productive use of GRID computing through participation of biologists to shape the development of the GRID. 2. Develop user-friendly.
Learning A Better Compiler Predicting Unroll Factors using Supervised Classification And Integrating CPU and L2 Cache Voltage Scaling using Machine Learning.
Advanced User Support in the Swedish National HPC Infrastructure May 13, 2013NeIC Workshop: Center Operations best practices.
PEER 2003 Meeting 03/08/031 Interdisciplinary Framework Major focus areas Structural Representation Fault Systems Earthquake Source Physics Ground Motions.
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
Building PetaScale Applications and Tools on the TeraGrid Workshop December 11-12, 2007 Scott Lathrop and Sergiu Sanielevici.
CD Strategy Session/Briefing on Collaboration Tools: Sept. 12, Collaboration Tools Erik Gottschalk.
Presented by SciDAC-2 Petascale Data Storage Institute Philip C. Roth Computer Science and Mathematics Future Technologies Group.
Advanced Software Engineering Dr. Cheng
Computational Reasoning in High School Science and Math
Model-Driven Analysis Frameworks for Embedded Systems
DESIGN OF EXPERIMENTS by R. C. Baker
Presentation transcript:

The Performance Evaluation Research Center (PERC) Participating Institutions: Argonne Natl. Lab.Univ. of California, San Diego Lawrence Berkeley Natl. Lab.Univ. of Illinois Lawrence Livermore Natl. Lab.Univ. of Maryland Oak Ridge Natl. Lab.Univ. of Tennessee, Knoxville Website:

PERC Overview  An “Integrated Software Infrastructure Center” (ISIC) sponsored under DoE’s SciDAC program.  Funding: approx. $2.4 million per year.  Mission:  Develop a science of performance.  Engineer tools for performance analysis and optimization.  Focus:  Large, grand-challenge calculations, especially SciDAC application projects.

Specific Objectives  Understand key factors in applications that affect performance.  Understand key factors in computer systems that affect performance.  Develop models that accurately predict performance of applications on systems.  Develop an enabling infrastructure of tools for performance monitoring, modeling and optimization.  Validate these ideas and infrastructure via close collaboration with DOE Office of Science and others.  Transfer the technology to end users.

Anticipated Benefits Consider the economic value of improving the performance of a single high-end scientific application code by 20%. Assume:  $10 million computer system lease cost per year.  $10 million per year in site costs, support staff, etc.  10-year lifetime of code.  Code uses 5% of system cycles each year. Savings: $2,000,000. Scientific benefit (additional computer runs and research) is probably much higher.

Anticipated Benefits, cont.  We rely heavily on commercial vendors for high- performance computer systems.  We are invited by vendors to provide guidance on the design of current and future systems. BUT  At present we can provide only vague information – little if any quantitative data or rigorous analysis. The performance monitoring and modeling capability to be developed in PERC will significantly improve our ability to influence future scientific computer systems.

Current State-of-the-Art  Some tools collect performance data, but  They are not targeted to large parallel systems.  They are not able to collect performance data at individual levels of deep memory hierarchies.  A few performance modeling techniques have been developed, but  They are time-consuming to generate, difficult to use, or have limited accuracy.  Some automatic tuning techniques have been developed, but  They have been applied only in limited algorithm domains.  There is no hardened support for real-time optimization.

New Capabilities  Better Benchmarks:  Polished, concise versions of real user codes, representing strategic application areas.  Kernel benchmarks extracted from real codes reduce complexity of analyzing full-size benchmarks.  Low-level benchmarks measure key rates of data access at various levels of memory hierarchy.  Modern performance monitoring tools:  Flexible instrumentation systems capture hardware and software interactions, instruction execution frequencies, memory reference behavior, and execution overheads.  An advanced data management infrastructure tracks performance experiments and data across time and space.

New Capabilities, cont.  Performance modeling:  Application signature tools characterize applications independent of the machine where they execute.  Machine signature tools characterize computer systems, independent of the applications.  Convolution tools combine application and machine signatures to provide accurate performance models.  Statistical models find approximate performance models based on easily measured performance data.  Phase models analyze separate sections of an application, providing overall performance predictions.  Performance bound tools determine ultimate potential of an application on a given system.

New Capabilities, cont.  Performance optimization:  Compile-time optimization mechanisms analyze source code to improve performance.  Self-tuning software automatically tunes code based on real-time measurements of hardware environment.  Performance assertions permit user-specified run- time tests to possibly change the course of the computation depending on results.  Performance portability programming techniques to insure that code runs at near-optimal performance across a variety of modern systems.

Measured Memory Access Patterns

SvPablo Graphical Interface

PAPI Perfometer Interface

Performance Study of SciDAC Application

SciDAC Interactions Codes have been acquired from these projects:  Terascale Simulation of Neutrino-Driven Supernovae  Advanced Computing for 21st Century Accelerators  National Computational Infrastructure for Lattice Gauge Theory  Collaborate Design and Development of Community Climate System Model for Terascale Computers  Numerical Computation of Wave-Plasma Interactions  Accurate Properties for Open-Shell States of Large Molecules  Terascale Optimal PDE Solvers  An Algorithmic and Software Framework for PDEs

Working with PERC  Benchmarking  Application group works with PERC to specify relevant benchmark codes and problems.  PERC characterizes performance, generates performance models, and suggests optimizations.  Performance Tools  PERC trains application developers to use tools.  Application group uses tools in development, providing feedback on functionality and future development For further information:

Summary  Achieving optimal performance on HPC systems has compelling economic and scientific rationales.  Performance is poorly understood – in depth-studies do not exist except in a handful of cases.  PERC will pursue “performance science” and “performance engineering”, including improved benchmarks, monitoring tools, modeling techniques, and optimizers.

ENABLING TECHNOLOGIES Developing a science for understanding performance of scientific applications on high-end computer systems, and engineering strategies for improving performance on these systems. Understand the key factors in applications that affect performance. Understand the key factors in computer systems that affect performance. Develop models that accurately predict performance of applications on systems. Develop an enabling infrastructure of tools for performance monitoring, modeling and optimization. Validate these ideas and infrastructure via close collaboration with DOE SC and other application owners. Transfer the technology to end-users. GOALS Optimize and Simplify: Profiling of real applications Measurement of machine capabilities (emphasis on memory hierarchy) Performance prediction Performance monitoring Informed tuning DOE LABS ANLLBNL LLNLORNL UNIVERSITIES UCSDUI-UC UTennUMD Enhanced Scientific Simulations & Experiments Scientific Simulations & Experiments » Climate Modeling » Astrophysics » Genomics & Proteomics » High Energy Physics Application Signatures Machine Signatures Bound Models Phase Models C O N V O L U T I O N S ROSESvPablo PAPI Sigma++ (IBM) DynInst