Future Requirements for NSF ATM Computing Jim Kinter, co-chair Presentation to CCSM Advisory Board 9 January 2008.

Slides:



Advertisements
Similar presentations
Capacity Building Mandate We, the participants…recognize the need to support: …A coordinated effort to involve and assist developing countries in improving.
Advertisements

UCSC History. UCSC: A brief history 60s University Placement Committee A lot of field trips/interaction with employers.
DOE Global Modeling Strategic Goals Anjuli Bamzai Program Manager Climate Change Prediction Program DOE/OBER/Climate Change Res Div
U.S. Department of Energy Office of Science Advanced Scientific Computing Research Program NERSC Users Group Meeting Department of Energy Update June 12,
SDSC Computing the 21st Century Talk Given to the NSF Sugar Panel May 27, 1998.
Introduction What is Parallel Algorithms? Why Parallel Algorithms? Evolution and Convergence of Parallel Algorithms Fundamental Design Issues.
The Interplay of Funding Policy for infrastructure at NSF Richard S. Hirsh.
What is a climate model?. Substitutes for reality Closely mimics some essential elements Omits or poorly mimics non-essential elements What is a Model?
SALSASALSASALSASALSA Digital Science Center June 25, 2010, IIT Geoffrey Fox Judy Qiu School.
Core Services I & II David Hart Area Director, UFP/CS TeraGrid Quarterly Meeting December 2008.
Network, Operations and Security Area Tony Rimovsky NOS Area Director
December, 2009 Kent Milfeld, TG Allocations Coordinator.
CCSM Software Engineering Coordination Plan Tony Craig SEWG Meeting Feb 14-15, 2002 NCAR.
FY Division of Human Resources Development Combined COV COV PRESENTATION TO ADVISORY COMMITTEE January 7, 2014.
Bill Kuo Summary of DTC EC Meeting 26 th August 2010.
Information Technology at Purdue Presented by: Dr. Gerry McCartney Vice President and CIO, ITaP HPC User Forum September 8-10, 2008 Using SiCortex SC5832.
National Science Foundation GEOGeosciences HIA PER High-Performance Instrumented Airborne Platform for Environmental Research.
The FY 2009 Budget Thomas N. Cooley, NSF Council of Colleges of Arts and Sciences March 13, 2008.
State of the Federation Winter Meeting Washington, D.C. January 9, 2008.
The Utility of National Academy-Sponsored Decadal Surveys Daniel N. Baker Laboratory for Atmospheric and Space Physics CU-Boulder.
1 TeraGrid ‘10 August 2-5, 2010, Pittsburgh, PA State of TeraGrid in Brief John Towns TeraGrid Forum Chair Director of Persistent Infrastructure National.
NCAR Supercomputing Center (NSC) Project Status Update to the CHAP 4 October 2007 Krista Laursen NSC Project Director.
August 2007 Advancing Scientific Discovery through TeraGrid Adapted from S. Lathrop’s talk in SC’07
U.S. Department of Energy Office of Science Advanced Scientific Computing Research Program NERSC Users Group Meeting Department of Energy Update September.
Federal R&D Funding: A SIAM Perspective Mel Ciment Senior Advisor SIAM, Washington Office Tel: May 6, 2001.
Outline Course Administration Parallel Archtectures –Overview –Details Applications Special Approaches Our Class Computer Four Bad Parallel Algorithms.
Legal Services Advisory Committee (LSAC) Grant Applicant Workshop February 11, 2013.
TeraGrid Overview Cyberinfrastructure Days Internet2 10/9/07 Mark Sheddon Resource Provider Principal Investigator San Diego Supercomputer Center
Identifying Grand Challenges in Climate Change Research: Guiding DOE’s Strategic Planning: Report on the DOE/BERAC workshop March Crystal City For.
SAN DIEGO SUPERCOMPUTER CENTER Impact Requirements Analysis Team Co-Chairs: Mark Sheddon (SDSC) Ann Zimmerman (University of Michigan) Members: John Cobb.
Post-Secondary Education Program Joint AFN/INAC PSE Program Review with representation from NAIIHL and the Labrador Inuit Regional Information Process.
Crystal Ball Panel ORNL Heterogeneous Distributed Computing Research Al Geist ORNL March 6, 2003 SOS 7.
ESIP Federation Air Quality Cluster Partner Agencies.
RI User Support in DEISA/PRACE EEF meeting 2 November 2010, Geneva Jules Wolfrat/Axel Berg SARA.
Morris Aizenman Senior Scientist Directorate for Mathematical & Physical Sciences National Science Foundation Physics and Engineering Sciences Committee.
SCD Update Tom Bettge Deputy Director Scientific Computing Division National Center for Atmospheric Research Boulder, CO USA User Forum May 2005.
Petascale –LLNL Appro AMD: 9K processors [today] –TJ Watson Blue Gene/L: 40K processors [today] –NY Blue Gene/L: 32K processors –ORNL Cray XT3/4 : 44K.
INTO THE NEW YEAR January 3, Objectives Reaffirm principles –China’s interest in exploring ESIP structure prompted review of ESIP evolution (more.
Scalable Systems Software for Terascale Computer Centers Coordinator: Al Geist Participating Organizations ORNL ANL LBNL.
CCSM Tutorial CCSM Software Engineering Group June
ESIP Vision: “Achieve a sustainable world” by Serving as facilitator and advisor for the Earth science information community Promoting efficient flow of.
Report of the Subcommittee on Global Change Research Programs to DOE Office of Science Biological and Environmental Research Advisory Committee April 20,
Climate Change Education Interagency Working Group FY 2008 Potential Increase Funding and the Need for a Coordinating Interagency Education Working Group.
11 January 2005 High Performance Computing at NCAR Tom Bettge Deputy Director Scientific Computing Division National Center for Atmospheric Research Boulder,
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
February 25, 2008 The Emerging Front Range HPC Collaboratory Dr. Rich Loft: Director, Technology Development Computational.
Sergiu April 2006June 2006 Overview of TeraGrid Resources and Services Sergiu Sanielevici, TeraGrid Area Director for User.
1 Overall Architectural Design of the Earth System Grid.
Cyberinfrastructure: Many Things to Many People Russ Hobby Program Manager Internet2.
UPDATE on NSF & MPS Presentation to MPSAC 22 April 2004 Michael S Turner Assistant Director for MPS.
EGEE is a project funded by the European Union under contract IST Generic Applications in EGEE-NA4 Roberto Barbera NA4 Generic Applications.
Department of Energy Office of Science  FY 2007 Request for Office of Science is 14% above FY 2006 Appropriation  FY 2007 Request for HEP is 8% above.
Network, Operations and Security Area Tony Rimovsky NOS Area Director
Research Cloud Proposal & Globus Scott Baily, ACNS.
NICS Update Bruce Loftis 16 December National Institute for Computational Sciences University of Tennessee and ORNL partnership  NICS is the 2.
NOAA R&D High Performance Computing Colin Morgan, CISSP High Performance Technologies Inc (HPTI) National Oceanic and Atmospheric Administration Geophysical.
PPAI Decadal Prediction/Predictability/Variability –Reviewed the WGCM/WGSIP Decadal Prediction Experiment –Reviewed/Revised Decadal WG Prospectus Reviewed.
SAN DIEGO SUPERCOMPUTER CENTER Fran Berman Engineering Advisory Committee Cyberinfrastructure Subcommittee -- Prologue Dr. Francine Berman Director, SDSC.
National Science Foundation Blue Ribbon Panel on Cyberinfrastructure Summary for the OSCER Symposium 13 September 2002.
Visualization Update June 18, 2009 Kelly Gaither, GIG Area Director DV.
PEER 2003 Meeting 03/08/031 Interdisciplinary Framework Major focus areas Structural Representation Fault Systems Earthquake Source Physics Ground Motions.
PLANNING FOR A NOAA SCIENCE CENTER June 28, 2000.
TG ’08, June 9-13, State of TeraGrid John Towns Co-Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing.
TeraGrid’s Process for Meeting User Needs. Jay Boisseau, Texas Advanced Computing Center Dennis Gannon, Indiana University Ralph Roskies, University of.
Performance Technology for Scalable Parallel Systems
HPC System Acquisition and Service Provision
Jeffrey P. Gardner Pittsburgh Supercomputing Center
EXCITED Workshop Suvrajeet Sen, DMII, ENG – CI Coordinator Workshop
The two faces of Cyberinfrastructure: Grids (or Web 2
SIT Chair Priorities and SIT-33 Objectives
Presentation transcript:

Future Requirements for NSF ATM Computing Jim Kinter, co-chair Presentation to CCSM Advisory Board 9 January 2008

Charge to Committee Committee: Formed as sub-committee of AC-GEO to advise ATM on the future of computing; members selected by ATM from atmospheric sciences research community and supercomputing community Background: Recognizing that both advances in technology and the recent creation of the NSF Office of Cyberinfrastructure have opened up a wide range of opportunities for providing the computational services needed by atmospheric research, ATM wishes to survey the possible approaches. ATM seeks input from the atmospheric research community on how best to meet future needs, including how to minimize any potential disruption to individual research programs. Charge: The panel is charged with advising ATM on the merits of different possible strategies for ensuring access to adequate high-end computing services for the atmospheric sciences community over the next decade. In particular, the panel is asked to: 1. Review relevant materials describing the anticipated computational requirements of the atmospheric science research community; 2. Develop a list of different possible strategies for meeting the atmospheric sciences’ computing needs over the period ; 3. Provide an analysis of the merits of the various strategies developed under (2), including a discussion of the costs and benefits; 4. Provide a recommendation to ATM about the best strategy to pursue. Report: Preliminary advice early Feb; report to AC-GEO in April 2008

40 Years of Supercomputing THEN –Top speed: 10 4 FLOPS –In 18 months, my computer will have 2X transistors and be 2X faster –Cold War drives computing industry –Needed big building with lots of cheap power, cooling –Apps: NWP, missile trajectories –ATM computing done at NCAR (FFRDC) NOW –Top speed: FLOPS (+10 9 X) –In 18 months my chips will have 2X transistors –Video gaming and finance sector drive computing industry –Need big building with lots of expensive power, cooling –Apps: NWP, bomb design, circuit design, aerospace design, molecular dynamics, solar interior, human blood flow, earthquake modeling, N-body problem, QCD, … –AR4: CCSM development at NCAR, production elsewhere (DOE, Earth Simulator)

40 Years of Supercomputing THEN –Top speed: 10 4 FLOPS –In 18 months, my computer will have 2X transistors (Moore’s Law) and be 2X faster –Cold War drives computing industry –Needed big building with lots of cheap power, cooling –Apps: NWP, missile trajectories –ATM computing done at NCAR (FFRDC) NOW –Top speed: FLOPS (+10 9 X) –In 18 months my chips will have 2X transistors (Moore’s Law) –Video gaming and finance sector drive computing industry –Need big building with lots of expensive power, cooling –Apps: NWP, bomb design, circuit design, aerospace design, molecular dynamics, solar interior, human blood flow, earthquake modeling, N-body problem, QCD, … –AR4: CCSM development at NCAR, production elsewhere (DOE, Earth Simulator)

40 Years of Supercomputing THEN –Top speed: 10 4 FLOPS –In 18 months, my computer will have 2X transistors (Moore’s Law) and be 2X faster –Cold War drives computing industry –Needed big building with lots of cheap power, cooling –Apps: NWP, missile trajectories –ATM computing done at NCAR (FFRDC) NOW –Top speed: FLOPS (+10 9 X) –In 18 months my chips will have 2X transistors (Moore’s Law) –Entertainment and finance sectors drive computing industry –Need big building with lots of expensive power, cooling –Apps: NWP, bomb design, circuit design, aerospace design, molecular dynamics, solar interior, human blood flow, earthquake modeling, N-body problem, QCD, … –AR4: CCSM development at NCAR, production elsewhere (DOE, Earth Simulator)

40 Years of Supercomputing THEN –Top speed: 10 4 FLOPS –In 18 months, my computer will have 2X transistors (Moore’s Law) and be 2X faster –Cold War drives computing industry –Needed big building with lots of cheap power, cooling –Apps: NWP, missile trajectories –ATM computing done at NCAR (FFRDC) NOW –Top speed: FLOPS (+10 9 X) –In 18 months my chips will have 2X transistors (Moore’s Law) –Entertainment and finance sectors drive computing industry –Need big building with lots of expensive power, cooling –Apps: NWP, bomb design, circuit design, aerospace design, molecular dynamics, solar interior, human blood flow, earthquake modeling, N-body problem, QCD, … –AR4: CCSM development at NCAR, production elsewhere (DOE, Earth Simulator) In between … client-server model and commodity clusters significantly reduced power/cooling requirement BUT … Near future: power cost = system cost

40 Years of Supercomputing THEN –Top speed: 10 4 FLOPS –In 18 months, my computer will have 2X transistors (Moore’s Law) and be 2X faster –Cold War drives computing industry –Needed big building with lots of cheap power, cooling –Apps: NWP, missile trajectories –ATM computing done at NCAR (FFRDC) NOW –Top speed: FLOPS (+10 9 X) –In 18 months my chips will have 2X transistors (Moore’s Law) –Entertainment and finance sectors drive computing industry –Need big building with lots of expensive power, cooling –Apps: NWP, climate modeling, bomb design, circuit design, aerospace design, molecular dynamics, solar interior, human blood flow, earthquake modeling, N-body problem, QCD, … –AR4: CCSM development at NCAR, production elsewhere (DOE, Earth Simulator)

40 Years of Supercomputing THEN –Top speed: 10 4 FLOPS –In 18 months, my computer will have 2X transistors (Moore’s Law) and be 2X faster –Cold War drives computing industry –Needed big building with lots of cheap power, cooling –Apps: NWP, missile trajectories –ATM computing done at NCAR (FFRDC) NOW –Top speed: FLOPS (+10 9 X) –In 18 months my chips will have 2X transistors (Moore’s Law) –Entertainment and finance sectors drive computing industry –Need big building with lots of expensive power, cooling –Apps: NWP, climate modeling, bomb design, circuit design, aerospace design, molecular dynamics, solar interior, human blood flow, earthquake modeling, N-body problem, QCD, … –AR4: CCSM development at NCAR-CSL, production elsewhere (DOE, Earth Simulator)

CSL Resources for CCSM Process –CCSM Production gets special status –All other requests reviewed for computational appropriateness, readiness, criticality and relevance to climate simulation and own scientific goals –Overall “correction” if merit-based allocations are above/below available resources CCSM Production –6.5M CPU-hrs over Dec’07 - May’09 (~750 CPU-yrs) –Issues: Flat sub-allocation of resources to CCSM working groups - why no priorities? Insufficient interaction among WGs to coordinate numerical experiments CCSM Development –3.1M CPU-hrs over Dec’07 - May’09 (~350 CPU-yrs) –Issues: Too little effort to move toward petascale computing Worries about sub-critical human resources for algorithms, HEC etc. Same concerns as expressed for Production request

AR5 Production Elsewhere … DOE - NERSC and ORNL (~40K CPU-yrs/yr) NASA - Columbia (~10K CPU-yrs/yr) NOAA - GFDL??? International - Earth Simulator? (ES-2???) Industry - IBM? Cray?? SGI???

AR5 Production Elsewhere … DOE - NERSC and ORNL (~40K CPU-yrs/yr) NASA - Columbia (~10K CPU-yrs/yr) International - Earth Simulator? (ES-2???) Industry - IBM? Cray?? SGI??? NSF - TeraGrid – 12K CPU-yrs/yr in 2007 – 80K CPU-yrs/yr in 2008 –150K CPU-yrs/yr in 2009 –250K CPU-yrs/yr in 2010 –500K CPU-yrs/yr in 2011

TeraGrid FY2007 Usage PHY 13% AST 14% DMR 7% DMS 0% CHE 15% BIO 24% GEO 8% Industry 8% ENG 7% Other 1% CIS 4% PHY AST DMR DMS CHE BIO GEO Industry ENG CIS Other MPS 49% (incl. ATM) 

TeraGrid HPC Usage by Site FY (4/01/06 through 3/31/07) FY Total: ~110M CPU-hrs or ~12.5K CPU-yrs Dell Intel64 linux (9600) Dell PowerEdge linux (5840) IBM e1350 (3072) Cray XT3 (4136) IBM Blue Gene ( ) IBM Power 4+ (2176) … NCSA 37% SDSC 23% PSC 20% TACC 15% Indiana 2% Purdue 2% ANL 1% ORNL 0% NCSA: 24 ATM-related projects 13 universities 42 CPU-years NCAR: 400 ATM-related projects 100 universities 1.4K CPU-years +69K CPU-yrs in 2008 (Sun Opteron 4-core) +80K CPU-yrs in 2009 (Cray Opteron 8-core) +400K CPU-yrs in 2011 (IBM Power7 8-core)