Dr Mark Parsons Commercial Director, EPCC +44 131 650 5022 HECToR The latest UK National High Performance Computing Service.

Slides:



Advertisements
Similar presentations
HPCx Power for the Grid Dr Alan D Simpson HPCx Project Director EPCC Technical Director.
Advertisements

© 2007 IBM Corporation IBM Global Engineering Solutions IBM Blue Gene/P Blue Gene/P System Overview - Hardware.
© Cray Inc. CSC, Finland September 21-24, XT3XT4XT5XT6 Number of cores/socket Number of cores/node Clock Cycle (CC) ??
ARCHER: The next generation of UK National High Performance Computing Facility Dr Lorna Smith, EPCC. ARCHER CSE Deputy Director Thanks to Alan Simpson.
High-Performance Computing
IBM 1350 Cluster Expansion Doug Johnson Senior Systems Developer.
Appro Xtreme-X Supercomputers A P P R O I N T E R N A T I O N A L I N C.
Supercomputing Challenges at the National Center for Atmospheric Research Dr. Richard Loft Computational Science Section Scientific Computing Division.
Ver 0.1 Page 1 SGI Proprietary Introducing the CRAY SV1 CRAY SV1-128 SuperCluster.
Beowulf Supercomputer System Lee, Jung won CS843.
GPU System Architecture Alan Gray EPCC The University of Edinburgh.
GPGPU Introduction Alan Gray EPCC The University of Edinburgh.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA; SAN DIEGO SDSC RP Update October 21, 2010.
IDC HPC User Forum Conference Appro Product Update Anthony Kenisky, VP of Sales.
Nov COMP60621 Concurrent Programming for Numerical Applications Lecture 6 Chronos – a Dell Multicore Computer Len Freeman, Graham Riley Centre for.
OPTERON (Advanced Micro Devices). History of the Opteron AMD's server & workstation processor line 2003: Original Opteron released o 32 & 64 bit processing.
The UK National Grid Service Using the NGS. Outline NGS Background Getting Certificates Acceptable usage policies Joining VO’s What resources will be.
Claude TADONKI Mines ParisTech – LAL / CNRS / INP 2 P 3 University of Oujda (Morocco) – October 7, 2011 High Performance Computing Challenges and Trends.
NPACI: National Partnership for Advanced Computational Infrastructure Supercomputing ‘98 Mannheim CRAY T90 vs. Tera MTA: The Old Champ Faces a New Challenger.
CSC Site Update HP Nordic TIG April 2008 Janne Ignatius Marko Myllynen Dan Still.
1 HPC and the ROMS BENCHMARK Program Kate Hedstrom August 2003.
1 Computer Science, University of Warwick Architecture Classifications A taxonomy of parallel architectures: in 1972, Flynn categorised HPC architectures.
A New Building Data Center Upgrade capacity and technology step 2011 – May the 4 th– Hepix spring meeting Darmstadt (de) Pascal Trouvé (Facility Manager.
Results Matter. Trust NAG. Numerical Algorithms Group Mathematics and technology for optimized performance Andrew Jones IDC HPC User Forum, Imperial College.
1Training & Education at EPCC Training and Education at EPCC Judy Hardy
ORIGINAL AUTHOR JAMES REINDERS, INTEL PRESENTED BY ADITYA AMBARDEKAR Overview for Intel Xeon Processors and Intel Xeon Phi coprocessors.
Programming for High Performance Computers John M. Levesque Director Cray’s Supercomputing Center Of Excellence.
Bob Thome, Senior Director of Product Management, Oracle SIMPLIFYING YOUR HIGH AVAILABILITY DATABASE.
1 Advanced Storage Technologies for High Performance Computing Sorin, Faibish EMC NAS Senior Technologist IDC HPC User Forum, April 14-16, Norfolk, VA.
Synchronization and Communication in the T3E Multiprocessor.
3- System modelling An architectural model presents an abstract view of the sub-systems making up a system May include major information flows between.
SDSC RP Update TeraGrid Roundtable Reviewing Dash Unique characteristics: –A pre-production/evaluation “data-intensive” supercomputer based.
Principles of Scalable HPC System Design March 6, 2012 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
Company LOGO High Performance Processors Miguel J. González Blanco Miguel A. Padilla Puig Felix Rivera Rivas.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Results of the HPC in Europe Taskforce (HET) e-IRG Workshop Kimmo Koski CSC – The Finnish IT Center for Science April 19 th, 2007.
Jaguar Super Computer Topics Covered Introduction Architecture Location & Cost Bench Mark Results Location & Manufacturer Machines in top 500 Operating.
High Performance Computing Processors Felix Noble Mirayma V. Rodriguez Agnes Velez Electric and Computer Engineer Department August 25, 2004.
Taking the Complexity out of Cluster Computing Vendor Update HPC User Forum Arend Dittmer Director Product Management HPC April,
Frank Casilio Computer Engineering May 15, 1997 Multithreaded Processors.
BlueGene/L Facts Platform Characteristics 512-node prototype 64 rack BlueGene/L Machine Peak Performance 1.0 / 2.0 TFlops/s 180 / 360 TFlops/s Total Memory.
Presented by Leadership Computing Facility (LCF) Roadmap Buddy Bland Center for Computational Sciences Leadership Computing Facility Project.
Cray Innovation Barry Bolding, Ph.D. Director of Product Marketing, Cray September 2008.
Infrastructure Improvements 2010 – November 4 th – Hepix – Ithaca (NY)
2009/4/21 Third French-Japanese PAAP Workshop 1 A Volumetric 3-D FFT on Clusters of Multi-Core Processors Daisuke Takahashi University of Tsukuba, Japan.
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
HPCx:an Overview Dr Arthur Trew Director, EPCC. 210 February 2004IBM Team Talent Meeting what is HPCx? HPCx is the latest in a series of HPC services.
PSC’s CRAY-XT3 Preparation and Installation Timeline.
Patryk Lasoń, Marek Magryś
IDC HPC User Forum April 14 th, 2008 A P P R O I N T E R N A T I O N A L I N C Steve Lyness Vice President, HPC Solutions Engineering
Interconnection network network interface and a case study.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
National Center for Supercomputing Applications University of Illinois at Urbana–Champaign Visualization Support for XSEDE and Blue Waters DOE Graphics.
COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC.
Presented by NCCS Hardware Jim Rogers Director of Operations National Center for Computational Sciences.
Professor Arthur Trew Director, EPCC EPCC: KT in novel computing.
1 Next Generation Correlators, June 26 th −29 th, 2006 The LOFAR Blue Gene/L Correlator Stichting ASTRON (Netherlands Foundation for Research in Astronomy)
Cray XD1 Reconfigurable Computing for Application Acceleration.
Science Support for Phase 4 Dr Alan D Simpson HPCx Project Director EPCC Technical Director.
CDA-5155 Computer Architecture Principles Fall 2000 Multiprocessor Architectures.
Pathway to Petaflops A vendor contribution Philippe Trautmann Business Development Manager HPC & Grid Global Education, Government & Healthcare.
Parallel Computers Today Oak Ridge / Cray Jaguar > 1.75 PFLOPS Two Nvidia 8800 GPUs > 1 TFLOPS Intel 80- core chip > 1 TFLOPS  TFLOPS = floating.
Petascale Computing Resource Allocations PRAC – NSF Ed Walker, NSF CISE/ACI March 3,
Societal applications of large scalable parallel computing systems ARTEMIS & ITEA Co-summit, Madrid, October 30th 2009.
Plans for the National NERC HPC services UM vn 6.1 installations and performance UM vn 6.6 and NEMO(?) plans.
DIRECT IMMERSION COOLED IMMERS HPC CLUSTERS BUILDING EXPERIENCE
Appro Xtreme-X Supercomputers
DIRECT IMMERSION COOLED IMMERS HPC CLUSTERS BUILDING EXPERIENCE
Presentation transcript:

Dr Mark Parsons Commercial Director, EPCC HECToR The latest UK National High Performance Computing Service

EPCC The University of Edinburgh founded EPCC in 1990 to act as the focus for its interests in parallel computing Today, EPCC is a key European supercomputing centre –80 permanent staff –Managing all UK national HPC facilities –Work 50:50 academia and industry Goal is to rival the big US centres –eg. NCSA at the University of Illinois In 2007 we won the contract to host the new UK National Service HPC service - HECToR Technology Transfer Training European Coordination HPC Research Facilities Visitor Programme 2HECToR: IDC HPC Forum October 2008

The HECToR Service HECToR: High End Computing Terascale Resource Procured for UK scientists by Engineering and Physical Sciences Research Council – EPSRC Competitive process involving three procurements –Hardware – CRAY –Accommodation and Management – UoE HPCx LTD –Computational Science and Engineering Support – NAG EPCC won the A&M procurement through its company – UoE HPCx Ltd HECToR is located at The University of Edinburgh HECToR: IDC HPC Forum October 20083

Contractual structure and roles UoE HPCx Ltd is a w holly-owned subsidiary of University of Edinburgh UoE HPCx Ltd awarded main contract for HECToR Service Provision –Runs from 2007 to 2013 –Subcontracts: Hardware (Cray), Helpdesk (EPCC), Systems (EPCC+DL) CSE support from NAG is separate Total contract value is around £115 million HECToR: IDC HPC Forum October EPSRC UoE HPCX Ltd NAG CCLRC/DLUoE/EPCCCray Hardware CSE Helpdesk Systems

HECToR Installation Timeline HECToR: IDC HPC Forum October February 2007 Signing of HECToR Contracts March 2007 Chippewa Falls, WI: XT4 Cabinets being assembled Edinburgh: laying foundations for new plant room

HECToR Installation Timeline HECToR: IDC HPC Forum October April 2007 Edinburgh: Test and Development System (one XT4 cabinet) installed Edinburgh: new building in progress August 2007 Edinburgh: Full 60 Cabinet System installed

HECToR at the ACF HECToR: IDC HPC Forum October 20087

Advanced Computing Facility Constructed 1976 for the University of Edinburgh –1 x 600 m² Computer Room –24-stage DX-based cooling servicing the room through 4 vast walk-in air-handling units –"conventional" downflow system Refurbished 2004 as the Advanced Computing Facility –2 x 300 m² Computer Rooms (one active, one empty concrete shell) –all new chilled-water based plant services, with capacity of 1.2MW Major expansion 2007 for HECToR –2nd Computer Room brought into operation –new-build external plant room to support massive uplift in required capacity –new HV electrical provision (up to 7MW) HECToR: IDC HPC Forum October 20088

Power and Cooling 9HECToR: IDC HPC Forum October 2008

Power and cooling New 470m² plant room for HECToR – 1.5x the area of the room it services UPS provides mins autonomy – must keep cooling running when powering HECToR – diesel engines Currently HECToR uses around 1.2MW We have provision at the ACF up to 7MW Rack power continues to increase: –2002 – IBMp690 10kW per rack –2007 – HECToR Phase 1 18kW per rack –2009 – HECToR Phase 2 38kW per rack (estimate) Now at limits of direct air cooling – next generation must use water cooling – much more efficient HECToR: IDC HPC Forum October

Power and cooling (cont) The average off-coil air temperature is maintained with ease in the range: 12.7° ° (in excess of design spec) The average chilled-water flow temperature is maintained in the range: 7.7° - 8.3° (load independent) The average chilled-water return temperature is maintained in the range: 13.7° ° 60 m³ per sec of air at mean 13° is supplied into the sub-floor Chilled-water flow rate is maintained at 40 litres per second –144,000 litres per hour Because we use “free cooling” when possible the cooling overhead can be brought well below 20% over the year HECToR: IDC HPC Forum October

The Cray XT4 Processing Element 12 Direct Attached Memory 8.5 GB/sec Local Memory Bandwidth 50 ns latency HyperTransport 4 GB/sec MPI Bandwidth AMD Opteron 7.6 GB/sec Cray SeaStar2 Interconnect 6.5 GB/sec Torus Link Bandwidth Copyright (c) 2008 Cray Inc. HECToR: IDC HPC Forum October 2008

Why is HECToR not a room full of PCs? HECToR is expensive because of its communications network Designed for –High bandwidth –Low latency Mandatory requirement –to scale to 10,000+ cores –for real world codes HECToR: IDC HPC Forum October

Scalable Software Architecture: UNICOS/lc –Microkernel on Compute PEs, full featured Linux on Service PEs. –Service PEs specialize by function –Software Architecture eliminates OS “Jitter” –Software Architecture enables reproducible run times –Large machines boot in under 30 minutes, including filesystem 14 Service Partition Specialized Linux nodes Compute PE Login PE Network PE System PE I/O PE Copyright (c) 2008 Cray Inc. HECToR: IDC HPC Forum October 2008

Utilisation Past 6 months – utilisation has remained relatively constant at 70-80% –Drop in June due to extended planned downtime for X2 integration EPSRC and NERC scientists dominate user base Continued growth in number of users – > 500 users 15HECToR: IDC HPC Forum October 2008

HECToR Job Sizes Good spread of job sizes 22% of utilisation is for 1024 processors and above Good increase in larger 4096 processor jobs HECToR Utilisation by Job Size 1 Mar 08 – 1 Sep 08 16HECToR: IDC HPC Forum October 2008

Utilisation by subject area Largest usage: –Chemistry –Environment –Physics –Materials Utilisation by Subject Area 1 Mar 08 – 1 Sep 08 17HECToR: IDC HPC Forum October 2008

Technology refreshes Cray have 4-year contract for hardware provision –Plus possible extension for years 5 and 6 Phase 1 (accepted: September 2007): –60TFlop Cray XT4 Vector system (installed June 2008) –2TFlop Cray X2 vector system (a “BlackWidow”) Phase 2 (installation: Summer 2009): –~60Tflop Cray XT4 (quadcore upgrade) –~200TFlop Cray (tba) Phase 3 (installation: Summer 2011): –technology supplier subject to future tender –anticipate infrastructure requirements approx as per Phase 2 18HECToR: IDC HPC Forum October 2008

Cray XT4 Quadcore Node GB/sec 2 – 8 GB 12.8 GB/sec direct connect memory (DDR 800) 6.4 GB/sec direct connect HyperTransport Cray SeaStar2+ Interconnect 4-way SMP >35 Gflops per node Up to 8 GB per node OpenMP Support within socket Copyright (c) 2008 Cray Inc. HECToR: IDC HPC Forum October 2008

Cray XT5 Quadcore Node GB/sec 2 – 32 GB memory 6.4 GB/sec direct connect HyperTransport Cray SeaStar2+ Interconnect 25.6 GB/sec direct connect memory 8-way SMP >70 Gflops per node Up to 32 GB of shared memory per node OpenMP Support Copyright (c) 2008 Cray Inc. HECToR: IDC HPC Forum October 2008

Dual Core v. Quad Core Core –2.6Ghz clock frequency –SSE SIMD FPU (2flops/cycle = 5.2GF peak) Cache Hierarchy –L1 Dcache/Icache: 64k/core –L2 D/I cache: 1M/core –SW Prefetch and loads to L1 –Evictions and HW prefetch to L2 Memory –Dual Channel DDR2 –10GB/s 667MHz –8GB/s nominal STREAMs Power –103W Core –2.1Ghz clock frequency –SSE SIMD FPU (4flops/cycle = 8.4GF peak) Cache Hierarchy –L1 Dcache/Icache: 64k/core –L2 D/I cache: 512 KB/core –L3 Shared cache 2MB/Socket –SW Prefetch and loads to L1,L2,L3 –Evictions and HW prefetch to L1,L2,L3 Memory –Dual Channel DDR2 –12GB/s 800MHz –10GB/s nominal STREAMs Power –75W HECToR: IDC HPC Forum October Dual CoreQuad Core

From Terascale to Petascale There are many challenges facing HPC today As processors have grown faster they’ve got hotter Manufacturers have responded with multicore processors –As a result we’ve entered a second golden age of parallelism But –Multicore processors are generally clocked slower than single core –Memory bandwidth is not increasing commensurately –The Taiwanese are making bigger devices but not faster –It takes considerable effort to parallelise a code –This hasn’t changed in 20 years –Many codes do not scale –Particularly many ISV engineering codes HECToR: IDC HPC Forum October

Conclusion At 60 TFlops, HECToR is one of the most powerful computers in the world today It’s a fantastic asset for the UK It serves the UK scientific community and the industrial and commercial communities We’re at a very interesting moment in supercomputing –The days of easy programmability are over –Software is the challenge not hardware –Industry and academia need to work closely together because we have the same (hard) problems to solve over the next decade HECToR: IDC HPC Forum October

Thanks and questions HECToR: IDC HPC Forum October