1 TeraGrid and the Path to Petascale John Towns Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing Applications.

Slides:



Advertisements
Similar presentations
Xsede eXtreme Science and Engineering Discovery Environment Ron Perrott University of Oxford 1.
Advertisements

1 US activities and strategy :NSF Ron Perrott. 2 TeraGrid An instrument that delivers high-end IT resources/services –a computational facility – over.
Background Chronopolis Goals Data Grid supporting a Long-term Preservation Service Data Migration Data Migration to next generation technologies Trust.
Dan Stanzione, Training lead CASC Meeting
(e)Science-Driven, Production- Quality, Distributed Grid and Cloud Data Infrastructure for the Transformative, Disruptive, Revolutionary, Next-Generation.
1 Ideas About the Future of HPC in Europe “The views expressed in this presentation are those of the author and do not necessarily reflect the views of.
Simo Niskala Teemu Pasanen
April 2009 OSG Grid School - RDU 1 Open Science Grid John McGee – Renaissance Computing Institute University of North Carolina, Chapel.
1 The eXtremeDigital (XD) Program Barry I. Schneider Office of Cyberinfrastructure National Science Foundation
Core Services I & II David Hart Area Director, UFP/CS TeraGrid Quarterly Meeting December 2008.
Network, Operations and Security Area Tony Rimovsky NOS Area Director
Computing in Atmospheric Sciences Workshop: 2003 Challenges of Cyberinfrastructure Alan Blatecky Executive Director San Diego Supercomputer Center.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
GIG Software Integration: Area Overview TeraGrid Annual Project Review April, 2008.
ASQ World Conference on Quality and Improvement May 24-26, 2010, St. Louis, MO Quality in Chaos: a view from the TeraGrid environment John Towns.
SICSA student induction day, 2009Slide 1 Social Simulation Tutorial Session 6: Introduction to grids and cloud computing International Symposium on Grid.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
The TeraGrid David Hart Indiana University AAAS’09, FEBRUARY 13, 2009.
Advancing Scientific Discovery through TeraGrid Scott Lathrop TeraGrid Director of Education, Outreach and Training University of Chicago and Argonne National.
1 TeraGrid ‘10 August 2-5, 2010, Pittsburgh, PA State of TeraGrid in Brief John Towns TeraGrid Forum Chair Director of Persistent Infrastructure National.
August 2007 Advancing Scientific Discovery through TeraGrid Adapted from S. Lathrop’s talk in SC’07
SAN DIEGO SUPERCOMPUTER CENTER NUCRI Advisory Board Meeting November 9, 2006 Science Gateways on the TeraGrid Nancy Wilkins-Diehr TeraGrid Area Director.
TeraGrid Overview Cyberinfrastructure Days Internet2 10/9/07 Mark Sheddon Resource Provider Principal Investigator San Diego Supercomputer Center
1 Preparing Your Application for TeraGrid Beyond 2010 TG09 Tutorial June 22, 2009.
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
SAN DIEGO SUPERCOMPUTER CENTER Impact Requirements Analysis Team Co-Chairs: Mark Sheddon (SDSC) Ann Zimmerman (University of Michigan) Members: John Cobb.
Coordinating the TeraGrid’s User Interface Areas Dave Hart, Amit Majumdar, Tony Rimovsky, Sergiu Sanielevici.
1 PY4 Project Report Summary of incomplete PY4 IPP items.
October 21, 2015 XSEDE Technology Insertion Service Identifying and Evaluating the Next Generation of Cyberinfrastructure Software for Science Tim Cockerill.
Kelly Gaither Visualization Area Report. Efforts in 2008 Focused on providing production visualization capabilities (software and hardware) Focused on.
Geosciences - Observations (Bob Wilhelmson) The geosciences in NSF’s world consists of atmospheric science, ocean science, and earth science Many of the.
Cyberinfrastructure What is it? Russ Hobby Internet2 Joint Techs, 18 July 2007.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
TeraGrid Allocations Discussion John Towns Director, Persistent Infrastructure National Center for Supercomputing Applications University of Illinois.
August 3, March, The AC3 GRID An investment in the future of Atlantic Canadian R&D Infrastructure Dr. Virendra C. Bhavsar UNB, Fredericton.
Joint Meeting of the AUS, US, XS Working Groups TG10 Tuesday August 3, hrs Elwood II.
TeraGrid Quarterly Meeting Arlington, VA Sep 6-7, 2007 NCSA RP Status Report.
TeraGrid Program Year 5 Overview John Towns Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing Applications University.
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Sergiu April 2006June 2006 Overview of TeraGrid Resources and Services Sergiu Sanielevici, TeraGrid Area Director for User.
Cyberinfrastructure: Many Things to Many People Russ Hobby Program Manager Internet2.
National Center for Supercomputing Applications University of Illinois at Urbana–Champaign Visualization Support for XSEDE and Blue Waters DOE Graphics.
Power and Cooling at Texas Advanced Computing Center Tommy Minyard, Ph.D. Director of Advanced Computing Systems 42 nd HPC User Forum September 8, 2011.
Education, Outreach and Training (EOT) and External Relations (ER) Scott Lathrop Area Director for EOT Extension Year Plans.
Data Area Report Chris Jordan, Data Working Group Lead, TACC Kelly Gaither, Data and Visualization Area Director, TACC April 2009.
TeraGrid Program Year 5 Overview John Towns Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing Applications University.
Data, Visualization and Scheduling (DVS) TeraGrid Annual Meeting, April 2008 Kelly Gaither, GIG Area Director DVS.
Network, Operations and Security Area Tony Rimovsky NOS Area Director
NICS Update Bruce Loftis 16 December National Institute for Computational Sciences University of Tennessee and ORNL partnership  NICS is the 2.
TeraGrid-Wide Operations Von Welch Area Director for Networking, Operations and Security NCSA, University of Illinois April, 2009.
TeraGrid-Wide Operations DRAFT #2 Mar 31 Von Welch.
TeraGrid External Relations SC10 update Elizabeth Leake September 28, 2010.
AT LOUISIANA STATE UNIVERSITY CCT: Center for Computation & Technology Introduction to the TeraGrid Daniel S. Katz Lead, LONI as a TeraGrid.
Software Integration Highlights CY2008 Lee Liming, JP Navarro GIG Area Directors for Software Integration University of Chicago, Argonne National Laboratory.
Visualization Update June 18, 2009 Kelly Gaither, GIG Area Director DV.
TeraGrid Program Year 5 Overview John Towns Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing Applications University.
Petascale Computing Resource Allocations PRAC – NSF Ed Walker, NSF CISE/ACI March 3,
Northwest Indiana Computational Grid Preston Smith Rosen Center for Advanced Computing Purdue University - West Lafayette West Lafayette Calumet.
TG ’08, June 9-13, State of TeraGrid John Towns Co-Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing.
Building PetaScale Applications and Tools on the TeraGrid Workshop December 11-12, 2007 Scott Lathrop and Sergiu Sanielevici.
Education, Outreach and Training (EOT) Scott Lathrop Area Director for EOT January 2010.
Data Infrastructure in the TeraGrid Chris Jordan Campus Champions Presentation May 6, 2009.
SAN DIEGO SUPERCOMPUTER CENTER SDSC Resource Partner Summary March, 2009.
TeraGrid’s Process for Meeting User Needs. Jay Boisseau, Texas Advanced Computing Center Dennis Gannon, Indiana University Ralph Roskies, University of.
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
1 TeraGrid Annual Review (2009) and TeraGrid Extension Proposal John Towns Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for.
XSEDE Value Added and Financial Economies
Joslynn Lee – Data Science Educator
Presentation transcript:

1 TeraGrid and the Path to Petascale John Towns Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing Applications University of Illinois

2 DEISA PRACE Symposium, May11-13, 2009 What is the TeraGrid? World’s largest open scientific discovery infrastructure –supported by US National Science Foundation –extremely user-driven MPI jobs, ssh or grid (GRAM) access, etc. An instrument that delivers high-end resources and services –a computational facility: over a Pflop/s in parallel computing capability will grow to > 1.75 Pflop/s in 2009 –high performance networks –a data storage and management facility: over 30 PB of storage, over 100 scientific data collections –visualization systems, Science Gateways, User Portal A service: help desk and consulting, Advanced User Support (AUS), education and training events and resources Something you can use without financial cost –allocated to US researchers and their collaborators through national peer-review process generally, review of computing, not science

3 DEISA PRACE Symposium, May11-13, 2009 Our Vision of TeraGrid Three part mission: –support the most advanced computational science in multiple domains address key challenges prioritized by users –empower new communities of users partner with science community leaders - “Science Gateways” –provide resources and services that can be extended to a broader cyberinfrastructure partner with campuses and facilities TeraGrid is… –an advanced, nationally distributed, open cyberinfrastructure comprised of supercomputing, storage, and visualization systems, data collections, and science gateways, integrated by software services and high bandwidth networks, coordinated through common policies and operations, and supported by computing and technology experts, that enables and supports leading­edge scientific discovery and promotes science and technology education –a complex collaboration of over a dozen organizations and NSF awards working together to provide collective services that go beyond what can be provided by individual institutions

4 DEISA PRACE Symposium, May11-13, 2009 TeraGrid: greater than the sum of its parts… Single unified allocations process Single point of contact for problem reporting and tracking –especially useful for problems between systems Simplified access to high end resources for science and engineering –single sign-on –coordinated software environments –uniform access to heterogeneous resources to solve a single scientific problem –simplified data movement Expertise in building national computing and data resources Leveraging extensive resources, expertise, R&D, and EOT –leveraging other activities at participant sites –learning from each other improves expertise of all TG staff Leadership in cyberinfrastructure development, deployment and support –demonstrating enablement of science not possible without the TeraGrid-coordinated human and technological resources

5 DEISA PRACE Symposium, May11-13, 2009 Diversity of Resources (not exhaustive) Very Powerful Tightly Coupled Distributed Memory –Ranger (TACC): Sun Constellation, 62,976 cores, 579 Tflop/s, 123 TB RAM –Kraken (NICS): Cray XT5, 66,048 cores, 608 Tflop/s, > 1 Pflop/s in 2009 Shared Memory –Cobalt (NCSA): Altix, 8 Tflop/s, 3 TB shared memory –Pople (PSC): Altix, 5 Tflop/s, 1.5 TB shared memory Clusters with Infiniband –Abe (NCSA): 90 Tflop/s –Lonestar (TACC): 61 Tflop/s –QueenBee (LONI): 51 Tflop/s Condor Pool (Loosely Coupled) –Purdue- up to 22,000 CPUs Visualization Resources –TeraDRE (Purdue): 48 node nVIDIA GPUs –Spur (TACC): 32 nVIDIA GPUs Storage Resources –GPFS-WAN (SDSC) –Lustre-WAN (IU) –various archival resources

6 DEISA PRACE Symposium, May11-13, 2009 Resources to come… Track PSC –large shared memory system in 2010 Track 2d being competed –data-intensive HPC system –experimental HPC system –pool of loosely coupled, high throughput resources –experimental, high-performance grid test bed eXtreme Digital (XD) High-Performance Remote Visualization and Data Analysis Services –service and possibly resources; up to 2 awards (?) Blue Waters (Track NCSA: –1 Pflop/s sustained on serious applications in 2011 Unsolicited proposal for archival storage enhancements pending

7 DEISA PRACE Symposium, May11-13, 2009 How is TeraGrid Organized? TG is set up like a large cooperative research group –evolved from many years of collaborative arrangements between the centers –still evolving! Federation of 12 awards –Resource Providers (RPs) –Grid Infrastructure Group (GIG) Strategically lead by the TeraGrid Forum –made up of the PI’s from each RP and the GIG –led by the TG Forum Chair, who is responsible for coordinating the group (elected position) – John Towns – TG Forum Chair –responsible for the strategic decision making that affects the collaboration Centrally coordinated by the GIG

8 DEISA PRACE Symposium, May11-13, 2009 TeraGrid Participants

9 DEISA PRACE Symposium, May11-13, 2009 Who are the Players? GIG Management –GIG Director: Matthew Heinzel –GIG Director of Science: Dan Katz –Area Directors: Software Integration: Lee Liming/J.P. Navarro Gateways: Nancy Wilkins-Diehr User Services: Sergiu Sanielevici Advanced User Support: Amit Majumdar Data and Visualization: Kelly Gaither Network, Ops, and Security: Von Welch EOT: Scott Lathrop Project Management: Tim Cockerill User Facing Projects and Core Services: Dave Hart TeraGrid Forum –TG Forum Chair: John Towns –Membership: PSC: Ralph Roskies NICS: Phil Andrews ORNL: John Cobb Indiana: Craig Stewart Purdue: Carol Song U Chicago/ANL: Mike Papka NCSA: John Towns LONI: Dan Katz TACC: Jay Boisseau NCAR: Rich Loft SDSC: Richard Moore GIG: Matt Heinzel

10 DEISA PRACE Symposium, May11-13, 2009 Allocations Process National peer-review process –allocates computational, data, and visualization resources –makes recommendations on allocation of advanced direct support services Managed by TeraGrid –GIG and RP Participants in reviews –CORE Services award to manage shared responsibilities TACC: Meeting coordination SDSC: TG Central DB NCSA: POPS, TG Allocations group Currently awarding >10B Normalized Units of resources annually

11 DEISA PRACE Symposium, May11-13, 2009 TeraGrid HPC Usage, B NUs in Q Kraken, Aug Ranger, Feb B NUs in 2007 In 2008, Aggregate HPC power increased by 3.5x NUs requested and awarded quadrupled NUs delivered increased by 2.5x In 2008, Aggregate HPC power increased by 3.5x NUs requested and awarded quadrupled NUs delivered increased by 2.5x

12 DEISA PRACE Symposium, May11-13, 2009 TeraGrid Use by Discipline, 2008 ~2,500 users charged jobs in 2008 Representing 332 institutions, 48 states + PR,DC ~2,500 users charged jobs in 2008 Representing 332 institutions, 48 states + PR,DC 2008 Summary

13 DEISA PRACE Symposium, May11-13, 2009 Impacting Many Agencies NSF DOE NIH NASA DOD International University Other Industry NSF 52% DOE 13% NIH 19% NASA 10% DOD 1% International 0% University 2% Other 2% Industry 1% NSF 49% DOE 11% NIH 15% NASA 9% DOD 5% International 3% University 1% Other 6% Industry 1% Supported Research Funding by Agency Resource Usage by Agency $91.5M in Funded Research Supported 10B NUs Delivered

14 DEISA PRACE Symposium, May11-13, 2009 Geosciences (SCEC) Goal is understanding earthquakes and to mitigate risks of loss of life and property damage. Spans the gamut from largest simulations to midsize jobs to huge number of small jobs For largest runs (Cybershake), where they examine high frequency modes (short wave- length, so higher resolution) of particular interest to civil engineers, need large distributed memory runs using the Track 2 machines at TACC, NICS; 2,000-64,000 cores of Ranger, Kraken. To improve the velocity model that goes into the large simulations, need mid-range core counts jobs doing full 3-D tomography (Tera3D); DTF and other clusters (e.g. Abe); Need large data available on disk (100 TB) Output is large data sets stored at NCSA, or SDSC’s GPFS, IRODS. Moving to DOE machine at Argonne. TG provided help with essential data transfer. Excellent example of coordinated ASTA support- CUI (SDSC) and Urbanic (PSC) interface with consultants at NICS, TACC, &NCSA to smooth migration of code. Improved performance 4x.

15 DEISA PRACE Symposium, May11-13, 2009 Aquaporins - Schulten group,UIUC Aquaporins are proteins which conduct large volumes of water through cell walls while filtering out charged particles like hydrogen ions (protons). Start with known crystal structure, simulate over 100,000 atoms, using NAMD Water moves through aquaporin channels in single file. Oxygen leads the way in. At the most constricted point of channel, water molecule flips. Protons can’t do this.

16 DEISA PRACE Symposium, May11-13, 2009 Animation pointed to by 2003 Nobel chemistry prize announcement for structure of aquaporins (Peter Agre) The simulation helped explain how the structure led to the function Aquaporin Mechanism

17 DEISA PRACE Symposium, May11-13, 2009 Where is TeraGrid Going? Current Program –nominal end date for TeraGrid is March 2010 TeraGrid Phase III: eXtreme Digital Resources for Science and Engineering (XD) –follow-on program for TeraGrid –four Integrating Services Coordination and Management Service (CMS) Technology Audit and Insertion Service (TAIS) Advanced User Support Service (AUSS) Training, Education and Outreach Service (TEOS) –original planned start date on April 2010 CMS, AUSS and TEOS deferred one year

18 DEISA PRACE Symposium, May11-13, 2009 TeraGrid → TeraGrid Extension (?) → XD Transition Planning All current TeraGrid activity areas have effort reserved for TeraGrid → XD transition effort as appropriate –transition issues exist for nearly all areas effort ear-marked to support transition issues Start of XD for CMS/AUSS/TEOS deferred for one year (1 April 2011) –induced TeraGrid Extension Proposal 12-month funding to support most GIG functions and some non-Track 2 RP resources –uncertainty in sequence of events –still need to address many changes in TeraGrid going into presumed Extension Period many resources exit TeraGrid Program Year 5 planning process will likely need to address: –TeraGrid Extension following PY5 –necessarily include transition to XD in extension period