Extreme Scalability Working Group (XS-WG): Status Update Nick Nystrom Director, Strategic Applications Pittsburgh Supercomputing Center October 22, 2009.

Slides:



Advertisements
Similar presentations
Extreme Scalability RAT Report title=RATs#Extreme_Scalability Sergiu Sanielevici,
Advertisements

User Services Transition To XD TG Quarterly Management Meeting, San Juan 12/7/2010 Amit & Sergiu.
User Support Coordination Objectives for Plan Years 4 and 5 (8/1/2008 to 3/1/2010) Sergiu Sanielevici, GIG Area Director for User Support Coordination.
Priority Research Direction (I/O Models, Abstractions and Software) Key challenges What will you do to address the challenges? – Develop newer I/O models.
Adaptive Infrastructures EPRI/DoD Initiative on Complex Interactive Networks/Systems Joint innovative research ·EPRI and ·Office of the Director of Defense.
TAU Parallel Performance System DOD UGC 2004 Tutorial Allen D. Malony, Sameer Shende, Robert Bell Univesity of Oregon.
A Parallel Structured Ecological Model for High End Shared Memory Computers Dali Wang Department of Computer Science, University of Tennessee, Knoxville.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
NGNS Program Managers Richard Carlson Thomas Ndousse ASCAC meeting 11/21/2014 Next Generation Networking for Science Program Update.
Simo Niskala Teemu Pasanen
1 Building National Cyberinfrastructure Alan Blatecky Office of Cyberinfrastructure EPSCoR Meeting May 21,
3DAPAS/ECMLS panel Dynamic Distributed Data Intensive Analysis Environments for Life Sciences: June San Jose Geoffrey Fox, Shantenu Jha, Dan Katz,
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
N Tropy: A Framework for Analyzing Massive Astrophysical Datasets Harnessing the Power of Parallel Grid Resources for Astrophysical Data Analysis Jeffrey.
Computer System Architectures Computer System Software
Brad Armosky, TACC Shawn T. Brown, PSC Scott Lathrop, UC/ANL Laura McGinnis, PSC TeraGrid’08 Conference, Las Vegas, Nevada June 11, 2008.
Slide 1 Auburn University Computer Science and Software Engineering Scientific Computing in Computer Science and Software Engineering Kai H. Chang Professor.
Effective User Services for High Performance Computing A White Paper by the TeraGrid Science Advisory Board May 2009.
Designing and Evaluating Parallel Programs Anda Iamnitchi Federated Distributed Systems Fall 2006 Textbook (on line): Designing and Building Parallel Programs.
DOE BER Climate Modeling PI Meeting, Potomac, Maryland, May 12-14, 2014 Funding for this study was provided by the US Department of Energy, BER Program.
1 Preparing Your Application for TeraGrid Beyond 2010 TG09 Tutorial June 22, 2009.
Technology + Process SDCI HPC Improvement: High-Productivity Performance Engineering (Tools, Methods, Training) for NSF HPC Applications Rick Kufrin *,
Coordinating the TeraGrid’s User Interface Areas Dave Hart, Amit Majumdar, Tony Rimovsky, Sergiu Sanielevici.
4.2.1 Programming Models Technology drivers – Node count, scale of parallelism within the node – Heterogeneity – Complex memory hierarchies – Failure rates.
The Globus Project: A Status Report Ian Foster Carl Kesselman
Kelly Gaither Visualization Area Report. Efforts in 2008 Focused on providing production visualization capabilities (software and hardware) Focused on.
TeraGrid Quarterly Meeting Dec 5 - 7, 2006 Data, Visualization and Scheduling (DVS) Update Kelly Gaither, DVS Area Director.
Software Working Group Chairman’s Note: This document was prepared by the “software and applications” working group and was received by the entire workshop.
A Framework for Visualizing Science at the Petascale and Beyond Kelly Gaither Research Scientist Associate Director, Data and Information Analysis Texas.
National Center for Supercomputing Applications Barbara S. Minsker, Ph.D. Associate Professor National Center for Supercomputing Applications and Department.
Towards Exascale File I/O Yutaka Ishikawa University of Tokyo, Japan 2009/05/21.
Advanced User Support Amit Majumdar 5/7/09. Outline  Three categories of AUS  Update on Operational Activities  AUS.ASTA  AUS.ASP  AUS.ASEOT.
Leibniz Supercomputing Centre Garching/Munich Matthias Brehm HPC Group June 16.
1October 9, 2001 Sun in Scientific & Engineering Computing Grid Computing with Sun Wolfgang Gentzsch Director Grid Computing Cracow Grid Workshop, November.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Joint Meeting of the AUS, US, XS Working Groups TG10 Tuesday August 3, hrs Elwood II.
1 "Workshop 31: Developing a Hands-on Undergraduate Parallel Programming Course with Pattern Programming SIGCSE The 44 th ACM Technical Symposium.
Lawrence Livermore National Laboratory S&T Principal Directorate - Computation Directorate Tools and Scalable Application Preparation Project Computation.
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
Breakout Group: Debugging David E. Skinner and Wolfgang E. Nagel IESP Workshop 3, October, Tsukuba, Japan.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
Education, Outreach and Training (EOT) and External Relations (ER) Scott Lathrop Area Director for EOT Extension Year Plans.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Advanced User Support for MPCUGLES code at University of Minnesota October 09,
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 1.
Extreme Scalability Working Group (XS-WG): Status Update Nick Nystrom Director, Strategic Applications Pittsburgh Supercomputing Center October 21, 2010.
Extreme Scalability Working Group (XS-WG): Status Update Nick Nystrom Director, Strategic Applications Pittsburgh Supercomputing Center May 20, 2010.
Advanced User Support Amit Majumdar 8/13/09. Outline  Three categories of AUS  Operational Activities  AUS.ASTA  AUS.ASP  ASTA example.
Tackling I/O Issues 1 David Race 16 March 2010.
Advanced User Support Amit Majumdar 8/12/10. Outline  Three categories of AUS  Operational Activities  AUS.ASTA  ASTA examples  AUS.ASP  AUS.ASEOT.
Is MPI still part of the solution ? George Bosilca Innovative Computing Laboratory Electrical Engineering and Computer Science Department University of.
HPC University Requirements Analysis Team Training Analysis Summary Meeting at PSC September Mary Ann Leung, Ph.D.
Advanced User Support in the Swedish National HPC Infrastructure May 13, 2013NeIC Workshop: Center Operations best practices.
TeraGrid Advanced User Support (AUS) Amit Majumdar, SDSC Area Director – AUS TeraGrid Annual Review April 6-7,
Building PetaScale Applications and Tools on the TeraGrid Workshop December 11-12, 2007 Scott Lathrop and Sergiu Sanielevici.
Education, Outreach and Training (EOT) Scott Lathrop Area Director for EOT January 2010.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
© 2008 Pittsburgh Supercomputing Center Proposed ideas for consideration under AUS.
TeraGrid’s Process for Meeting User Needs. Jay Boisseau, Texas Advanced Computing Center Dennis Gannon, Indiana University Ralph Roskies, University of.
Defining the Competencies for Leadership- Class Computing Education and Training Steven I. Gordon and Judith D. Gardiner August 3, 2010.
Shaowen Wang 1, 2, Yan Liu 1, 2, Nancy Wilkins-Diehr 3, Stuart Martin 4,5 1. CyberInfrastructure and Geospatial Information Laboratory (CIGI) Department.
TeraGrid Software Integration: Area Overview (detailed in 2007 Annual Report Section 3) Lee Liming, JP Navarro TeraGrid Annual Project Review April, 2008.
INTRODUCTION TO HIGH PERFORMANCE COMPUTING AND TERMINOLOGY.
SCEC Capability Simulations on TeraGrid
Productive Performance Tools for Heterogeneous Parallel Computing
Introduction to Parallel Computing: MPI, OpenMP and Hybrid Programming
Geoffrey Fox, Shantenu Jha, Dan Katz, Judy Qiu, Jon Weissman
Performance Technology for Scalable Parallel Systems
Parallel Programming By J. H. Wang May 2, 2017.
Department of Computer Science, University of Tennessee, Knoxville
Presentation transcript:

Extreme Scalability Working Group (XS-WG): Status Update Nick Nystrom Director, Strategic Applications Pittsburgh Supercomputing Center October 22, 2009

XS-WG Update | Nystrom | October 22, Extreme Scalability Working Group (XS-WG): Purpose Meet the challenges and opportunities of deploying extreme-scale resources into the TeraGrid, maximizing both scientific output and user productivity. –Aggregate, develop, and share wisdom –Identify and address needs that are common to multiple sites and projects –May require assembling teams and obtaining support for sustained effort XS-WG benefits from active involvement of all Track 2 sites, BlueWaters, tool developers, and users. The XS-WG leverages and combines RPs’ interests to deliver greater value to the computational science community.

XS-WG Update | Nystrom | October 22, XS-WG Participants Nick NystromPSC, XS-WG lead Jay AlamedaNCSA Martin BerzinsUniv. of Utah (U) Paul BrownIU Shawn BrownPSC Lonnie CrosbyNICS, IO/Workflows lead Tim DudekGIG EOT Victor EijkhoutTACC Jeff GardnerU. Washington (U) Chris HempelTACC Ken JansenRPI (U) Shantenu JhaLONI Nick KaronisNIU (G) Dan KatzLONI Ricky KendallORNL Byoung-Do KimTACC Scott LathropGIG, EOT AD Vickie LynchORNL Amit MajumdarSDSC, TG AUS AD Mahin MahmoodiPSC, Tools lead Allen MalonyUniv. of Oregon (P) David O’NealPSC Dmitry PekurovskySDSC Wayne PfeifferSDSC Raghu ReddyPSC, Scalability lead Sergiu SanieleviciPSC Sameer ShendeUniv. of Oregon (P) Ray SheppardIU Alan SnavelySDSC Henry TufoNCAR George TurnerIU John UrbanicPSC Joel WellingPSC Nick WrightSDSC (P) S. Levent Yilmaz*CSM, U. Pittsburgh (P) U: user; P: performance tool developer; G: grid infrastructure developer; *: joined XS-WG since last TG-ARCH update

XS-WG Update | Nystrom | October 22, Technical Challenge Area #1: Scalability and Architecture Algorithms, numerics, multicore, etc. –Robust, scalable infrastructure (libraries, frameworks, languages) for supporting applications that scale to O (10 4–6 ) cores –Numerical stability and convergence issues that emerge at scale –Exploiting systems’ architectural strengths –Fault tolerance and resilience Contributors –POC: Raghu Reddy (PSC) –Members: Reddy, Majumdar, Urbanic, Kim, Lynch, Jha, Nystrom Current activities –Understanding performance tradeoffs in hierarchical architectures e.g. partitioning between MPI/OpenMP for different node architectures, interconnects, and software stacks candidate codes for benchmarking: HOMB, WRF, perhaps others –Characterizing bandwidth-intensive communication performance

XS-WG Update | Nystrom | October 22, Investigating the Effectiveness of Hybrid Programming (MPI+OpenMP) Begun in XS-WG, extended through AUS effort in collaboration with Amit Majumdar Examples of applications with hybrid implementations: WRF, POP, ENZO To exploit more memory per task, threading offers clear benefits. But what about performance? –Prior results are mixed; pure MPI often seems at least as good. –Historically, systems had fewer cores/socket and fewer cores/node than we have today, and far fewer than they will have in the future. –Have OpenMP versions been as carefully optimized? Reasons to look into hybrid implementations now –Current T2 systems have 8-16 cores per node. –Are we at the tipping point for threading offering a win? If not, is there one, and at what core count, and for which kinds of algorithms? –What is the potential for performance improvement?

XS-WG Update | Nystrom | October 22, Hybrid OpenMP-MPI Benchmark (HOMB) Developed by Jordan Soyke, while a student intern at PSC, subsequently enhanced by Raghu Reddy Simple benchmark code Permits systematic evaluation by –Varying computation-communication ratio –Varying message sizes –Varying MPI vs. OpenMP balance Allows characterization of performance bounds –Characterizing the potential hybrid performance of an actual application is possible with adequate understanding of its algorithms and their implementations.

XS-WG Update | Nystrom | October 22, Characteristics of the Benchmark Perfectly parallel with both MPI/OpenMP Perfectly load balanced Distinct computation and communication sections Only nearest-neighbor communication –Currently no reduction operations –No overlap of computation and communication Can easily vary computation/communication ratio Current tests are with large messages

XS-WG Update | Nystrom | October 22, Preliminary Results on Kraken: MPI vs. MPI+OpenMP, 12 threads/node Hybrid could be beneficial because of other reasons: –Application has limited scalability because of the decomposition –Application needs more memory –Application has dynamic load imbalance The hybrid approach provides increasing performance advantage as communication fraction increases. –… for the current core count per node. –Non-threaded sections of an actual application would have an Amdahl’s Law effect; these results constitute a best case limit.

XS-WG Update | Nystrom | October 22, Technical Challenge Area #2: Tools Performance tools, debuggers, compilers, etc. –Evaluate strengths and interactions; ensure adequate installations –Analyze/address gaps in programming environment infrastructure –Provide advanced guidance to RP consultants Contributors POC: Mahin Mahmoodi (PSC) Members: Mahmoodi, Wright, Alameda, Shende, Sheppard, Brown, Nystrom Current activities Focus on testing debuggers and performance tools at large core counts Ongoing, excellent collaboration between SDCI tool projects, plus consideration of complementary tools Submission for a joint POINT/IPM tools tutorial to TG09 Installing and evaluating strengths of tools as they apply to complex production applications

XS-WG Update | Nystrom | October 22, Collaborative Performance Engineering Tutorials TG09: Using Tools to Understand Performance Issues on TeraGrid Machines: IPM and the POINT Project (June 22, 2009) –Karl Fuerlinger (UC Berkeley), David Skinner (NERSC/LBNL), Nick Wright (then SDSC), Rui Liu (NCSA), Allen Malony (Univ. of Oregon), Haihang You (UTK), Nick Nystrom (PSC) –Analysis and optimization of applications on the TeraGrid, focusing on Ranger and Kraken. SC09: Productive Performance Engineering of Petascale Applications with POINT and VI-HPS (Nov. 16, 2009) –Allen Malony and Sameer Shende (Univ. of Oregon), Rick Kufrin (NCSA), Brian Wylie and Felix Wolf (JSC), Andreas Knuepfer and Wolfgang Nagel (TU Dresden), Shirley Moore (UTK), Nick Nystrom (PSC). –Addresses performance engineering of petascale, scientific applications with TAU, PerfSuite, Scalasca, and Vampir. –Includes hands-on exercises using a Live-DVD containing all of the tools, helping to prepare participants to apply modern methods for locating and diagnosing typical performance bottlenecks in real-world parallel programs at scale.

XS-WG Update | Nystrom | October 22, Technical Challenge Area #3: Workflow, data transport, analysis, visualization, and storage Coordinating massive simulations, analysis, and visualization –Data movement between RPs involved in complex simulation workflows; staging data from HSM systems across the TeraGrid –Technologies and techniques for in situ visualization and analysis Contributors –POC: Lonnie Crosby (NICS) –Members: Crosby, Welling, Nystrom Current activities –Focus on I/O profiling and determining platform-specific recommendations for obtaining good performance for common parallel I/O scenarios

XS-WG Update | Nystrom | October 22, Co-organized a Workshop on Enabling Data-Intensive Computing: from Systems to Applications July 30-31, 2009, University of Pittsburgh days: presentations, breakout discussions –architectures –software frameworks and middleware –algorithms and applications Speakers –John Abowd - Cornell University –David Andersen - Carnegie Mellon University –Magda Balazinska - The University of Washington –Roger Barga - Microsoft Research –Scott Brandt - The University of California at Santa Cruz –Mootaz Elnozahy - International Business Machines –Ian Foster - Argonne National labs –Geoffrey Fox - Indiana University –Dave O'Hallaron - Intel Research –Michael Wood-Vasey - University of Pittsburgh –Mazin Yousif - The University of Arizona –Taieb Znati - The National Science Foundation From R. Kouzes et al., The Changing Paradigm of Data-Intensive Computing, IEEE Computer, January 2009

XS-WG Update | Nystrom | October 22, Next TeraGrid/Blue Waters Extreme-Scale Computing Workshop To focus on parallel I/O for petascale applications, addressing: –multiple levels of applications, middleware (HDF, MPI-IO, etc.), and systems –requirements for data transfers to/from archives and remote processing and management facilities. Tentatively scheduled for the week of March 22, 2010, in Austin Builds on preceding Petascale Application Workshops –December 2007, Tempe: general issues of petascale applications –June 2008, Las Vegas: more general issues of petascale applications –March 2009, Albuquerque: fault tolerance and resilience; included significant participation from NNSA, DOE, and DoD