Technology + Process SDCI HPC Improvement: High-Productivity Performance Engineering (Tools, Methods, Training) for NSF HPC Applications Rick Kufrin *,

Slides:



Advertisements
Similar presentations
Technology Drivers Traditional HPC application drivers – OS noise, resource monitoring and management, memory footprint – Complexity of resources to be.
Advertisements

Extreme Scalability RAT Report title=RATs#Extreme_Scalability Sergiu Sanielevici,
Presentation at WebEx Meeting June 15,  Context  Challenge  Anticipated Outcomes  Framework  Timeline & Guidance  Comment and Questions.
ACCI TASK FORCES Update CASC September 22, Task Force Introduction Timeline months or less from June 2009 Led by NSF Advisory Committee on.
GENI: Global Environment for Networking Innovations Larry Landweber Senior Advisor NSF:CISE Joint Techs Madison, WI July 17, 2006.
Funding Opportunities at NSF Jane Silverthorne International Arabidopsis Consortium Workshop January 15, 2011.
Life and Health Sciences Summary Report. “Bench to Bedside” coverage Participants with very broad spectrum of expertise bridging all scales –From molecule.
Cyberinfrastructure Center ■ (765) ■ Felix Haas Hall Mission: Create an interdisciplinary facility of national prominence.
Robert Bell, Allen D. Malony, Sameer Shende Department of Computer and Information Science Computational Science.
Simulation and Information Technologies Gregory L. Fenves, UC Berkeley PEER Summative Meeting – June 13, 2007.
TAU Parallel Performance System DOD UGC 2004 Tutorial Allen D. Malony, Sameer Shende, Robert Bell Univesity of Oregon.
Nick Trebon, Alan Morris, Jaideep Ray, Sameer Shende, Allen Malony {ntrebon, amorris, Department of.
INTRODUCTION TO PERFORMANCE ENGINEERING Allen D. Malony Performance Research Laboratory University of Oregon SC ‘09: Productive Performance Engineering.
On the Integration and Use of OpenMP Performance Tools in the SPEC OMP2001 Benchmarks Bernd Mohr 1, Allen D. Malony 2, Rudi Eigenmann 3 1 Forschungszentrum.
A Parallel Structured Ecological Model for High End Shared Memory Computers Dali Wang Department of Computer Science, University of Tennessee, Knoxville.
EInfrastructures (Internet and Grids) - 15 April 2004 Sharing ICT Resources – Discussion of Best Practices in the U.S. Mary E. Spada Program Manager, Strategic.
Iterative computation is a kernel function to many data mining and data analysis algorithms. Missing in current MapReduce frameworks is collective communication,
1 Building National Cyberinfrastructure Alan Blatecky Office of Cyberinfrastructure EPSCoR Meeting May 21,
Don Von Dollen Senior Program Manager, Data Integration & Communications Grid Interop December 4, 2012 A Utility Standards and Technology Adoption Framework.
4.x Performance Technology drivers – Exascale systems will consist of complex configurations with a huge number of potentially heterogeneous components.
Role of Deputy Director for Code Architecture and Strategy for Integration of Advanced Computing R&D Andrew Siegel FSP Deputy Director for Code Architecture.
18:15:32Service Oriented Cyberinfrastructure Lab, Grid Deployments Saul Rioja Link to presentation on wiki.
Effective User Services for High Performance Computing A White Paper by the TeraGrid Science Advisory Board May 2009.
The Climate Prediction Project Global Climate Information for Regional Adaptation and Decision-Making in the 21 st Century.
SC’13: Hands-on Practical Hybrid Parallel Application Performance Engineering Introduction to VI-HPS Brian Wylie Jülich Supercomputing Centre.
Expanding the Accessibility and Impact of Language Technologies for Supporting Education (TFlex): Edinburgh Effort Dr. Myroslava Dzikovska, Prof. Johanna.
October 21, 2015 XSEDE Technology Insertion Service Identifying and Evaluating the Next Generation of Cyberinfrastructure Software for Science Tim Cockerill.
Center for Component Technology for Terascale Simulation Software CCA is about: Enhancing Programmer Productivity without sacrificing performance. Supporting.
Pascucci-1 Valerio Pascucci Director, CEDMAV Professor, SCI Institute & School of Computing Laboratory Fellow, PNNL Massive Data Management, Analysis,
Overcoming Scaling Challenges in Bio-molecular Simulations Abhinav Bhatelé Sameer Kumar Chao Mei James C. Phillips Gengbin Zheng Laxmikant V. Kalé.
Renaissance Computing Institute: An Overview Lavanya Ramakrishnan, John McGee, Alan Blatecky, Daniel A. Reed Renaissance Computing Institute.
BalticGrid-II Project BalticGrid-II Kick-off Meeting, , Vilnius1 Joint Research Activity Enhanced Application Services on Sustainable e-Infrastructure.
HPC Centres and Strategies for Advancing Computational Science in Academic Institutions Organisers: Dan Katz – University of Chicago Gabrielle Allen –
S2I2: Enabling grand challenge data intensive problems using future computing platforms Project Manager: Shel Swenson (USC & GATech)
PerfExplorer Component for Performance Data Analysis Kevin Huck – University of Oregon Boyana Norris – Argonne National Lab Li Li – Argonne National Lab.
Allen D. Malony, Professor  University of Illinois, Urbana-Champaign  Fulbright Research Scholar  The Netherlands  Austria  Alexander von Humboldt.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Five Year ITS Strategic Plan (2010 – 2014). 2 U.S. Department of Transportation Research and Innovative Technology Administration Overall ITS Community.
Allen D. Malony Department of Computer and Information Science TAU Performance Research Laboratory University of Oregon Discussion:
Joint Meeting of the AUS, US, XS Working Groups TG10 Tuesday August 3, hrs Elwood II.
Computational Science & Engineering meeting national needs Steven F. Ashby SIAG-CSE Chair March 24, 2003.
Connections to Other Packages The Cactus Team Albert Einstein Institute
Internet2 End-to-End (e2e) Performance Initiative Laurie Burns Steve Corbató Internet2 Fall Member Meeting Atlanta, Georgia October 31, 2000.
Cyberinfrastructure: Many Things to Many People Russ Hobby Program Manager Internet2.
Big Data in the Geosciences, University Corporation for Atmospheric Research (UCAR/NCAR), and the NCAR Wyoming Supercomputing Center (NWSC) Marla Meehl.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
Making the System Operational Implementation & Deployment
Parallel Performance Measurement of Heterogeneous Parallel Systems with GPUs Allen D. Malony, Scott Biersdorff, Sameer Shende, Heike Jagode†, Stanimire.
Data Conservancy and the US NSF DataNet Initiative Fourth Workshop on Data Preservation and Long-Term Analysis in HEP Sayeed Choudhury Johns Hopkins University.
VLDATA Common solution for the (very-)large data challenge EINFRA-1, focus on topics (4) & (5)
Extreme Scalability Working Group (XS-WG): Status Update Nick Nystrom Director, Strategic Applications Pittsburgh Supercomputing Center May 20, 2010.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI strategy and Grand Vision Ludek Matyska EGI Council Chair EGI InSPIRE.
High Risk 1. Ensure productive use of GRID computing through participation of biologists to shape the development of the GRID. 2. Develop user-friendly.
EU-Russia Call Dr. Panagiotis Tsarchopoulos Computing Systems ICT Programme European Commission.
All Hands Meeting 2005 BIRN-CC: Building, Maintaining and Maturing a National Information Infrastructure to Enable and Advance Biomedical Research.
Northwest Indiana Computational Grid Preston Smith Rosen Center for Advanced Computing Purdue University - West Lafayette West Lafayette Calumet.
Resource Optimization for Publisher/Subscriber-based Avionics Systems Institute for Software Integrated Systems Vanderbilt University Nashville, Tennessee.
Building PetaScale Applications and Tools on the TeraGrid Workshop December 11-12, 2007 Scott Lathrop and Sergiu Sanielevici.
CyberGIS Prof. Wenwen Li School of Geographical Sciences and Urban Planning 5644 Coor Hall
NASA Model-Based Systems Engineering Pathfinder 2016 Summary and Path Forward Karen J. Weiland, Ph.D. Jon Holladay, NASA Systems Engineering Technical.
Kai Li, Allen D. Malony, Sameer Shende, Robert Bell
Productive Performance Tools for Heterogeneous Parallel Computing
XSEDE Value Added and Financial Economies
Performance Technology for Scalable Parallel Systems
Tutorial Outline Welcome (Malony)
Similarities between Grid-enabled Medical and Engineering Applications
TAU integration with Score-P
Inculcating “Parallel Programming” in UG curriculum
Allen D. Malony, Sameer Shende
Making the System Operational Implementation & Deployment
Presentation transcript:

Technology + Process SDCI HPC Improvement: High-Productivity Performance Engineering (Tools, Methods, Training) for NSF HPC Applications Rick Kufrin *, Allen D. Malony +, Shirley Moore §, Nick Nystrom !, Sameer Shende + * National Center for Supercomputing Applications + University of Oregon § University of Tennessee, Knoxville ! Pittsburgh Supercomputing Center Introduction Parallel performance tools lie at the heart of HPC cyberinfrastructure providing powerful techniques for performance observation, analysis, and optimization integrated in technology solutions that are portable, interoperable, and scalable. Now is the time to transfer successful, robust parallel performance infrastructure to an integrated, extensible, and sustainable performance framework and suite of tools that will be improved and supported for the long term. However, productive use of high-performance computing (HPC) technology is as much a human-centric endeavor as it is technology driven. Performance tools are necessary, but not sufficient. If HPC investments are to be maximized, application developers must be trained to be good performance problem solvers and performance engineers. Objectives The project will integrate, harden, and deploy an open, portable, robust performance tools environment for the NSF-funded HPC centers. The widely-used TAU, PAPI, KOJAK, and PerfSuite technologies will be improved and supported as core components. Each tool will be enhanced to better address user needs and evolving scalable HPC technology, and to interoperate as part of a performance engineering framework for routine use in performance evaluation and optimization of domain S&E applications. A training program for performance technology and engineering will be piloted and refined at the Pittsburgh Supercomputing Center and integrated with the TeraGrid Education, Outreach, Training (EOT) program over time. Its objectives will be to educate application developers and students in sound performance evaluation methods, to teach them best practices for engineering high-performance code solutions based on expert tuning strategies, and to train them to use the performance tools effectively. Organization Applications ENZO ENZO is an adaptive mesh refinement (AMR), grid-based hybrid code (hydro+N-Body) which is designed to do simulations of cosmological structure formation. NAMD The NAMD parallel molecular dynamics application is a mature community code widely deployed for research in large-scale biomolecular systems. NEMO3D NEMO3D is a quantum mechanical based simulation tool created to provide quantita- tive predictions for nanometer-scaled semiconductor devices. Training User engagement experts work with users to use tools for optimization User support in TeraGrid introduce tools to TeraGrid user support teams Workshops conduct multiple annual workshops to help instill performance engineering expertise among application developers Quantify tool impact track productivity and usability issues