Supermicro © 2009Confidential HPC Case Study & References
Supermicro © 2009Confidential Super Micro Computer, Inc. (NASDAQ: SMCI), a leader in application-optimized, high performance server solutions, participated in the inaugural ceremony for the CERNs LHC (Large Hadron Collider) Project in Geneva. Supermicro’s SuperBlade® servers, housed at CERN, (one of the world’s largest research labs) enabled the LHC Project with superior computational performance, scalability, and energy efficiency. “We are honored to have Supermicro’s industry-leading blade server technology installed at the foundation of this monumental scientific research project,” said Charles Liang, CEO and president of Supermicro. “Our SuperBlade® platforms deliver unsurpassed performance, computing density and energy efficiency, making them ideal for HPC clusters and data centers.” The LHC Project deploys, amongst others, Supermicro’s award-winning SuperBlade servers. These optimized solutions empower Supermicro customers with the most advanced green server technology available, including 93%* peak power supply efficiency, innovative and highly efficient thermal and cooling system designs, and industry-leading performance-per-watt (290+ GFLOPS/kW*). Installation Example – CERN 100+ nodes SuperBlades® along with rack mount servers 14-Blade 10-Blade
Supermicro © 2009Confidential CERN LHC (Large Hadron Collider) Project CERN: Largest Hadron Collider Research Center 14-Blade 10-Blade Source: Dr. Helge Meinhard, CERN
Supermicro © 2009Confidential Source: Dr. Helge Meinhard, CERN Tunnel of 27 km circumference, 4 m diameter, 50…150 m below ground; Detectors at four collision points 15 Petabytes per year for four experiments CERN LHC (Large Hadron Collider) Project
Supermicro © 2009Confidential Installation Example – Research LABs Total 4000 nodes / 4-way AMD Barcelona quad-core With Infiniband connection 300 nodes / 2 DP nodes per 1U 1U Twin™ / Intel Hapertown With Infiniband (onboard) connection Shanghai-ready
Supermicro © 2009Confidential LLNL Hyperion Petescale Cluster
Supermicro © 2009Confidential Supermicro A+ Server 2041M-32R+B H8QM3-2 motherboard 2U 4-socket AMD quad-core One with 64 GB RAM & one with 256 GB RAM Two x16, Two x4 PCIe 1.0 slots One Mellanox Connect-X IB DDR 2-port PCIe CX4 HCA One LSI SAS controller 8 external ports LLNL Hyperion Compute Node
Supermicro © 2009Confidential Speeding up science: CSIRO's CPU-GPU CSIRO GPU supercomputer configuration: The new CSIRO high performance computing cluster will deliver up to 200 plus Teraflops of computing performance and will consist of the following components: Supermicro 100 Dual Xeon E5462 Compute Nodes (i.e. a total of GHz compute cores) with 16GB of RAM, 500GB SATA storage and DDR InfiniBand interconnect 50 Tesla S1070 (200 GPUs with a total of 48,000 streaming processor cores) 96 port DDR InfiniBand Switch On TOP500 / Green500 (June, 2009): Delivered by NEC (Hybrid Cluster): Supermicro Twin + NVIDIA GPUs (50TFLOPS) TOP500: ranking 77 Green500: ranking 20: the most efficient cluster system in the X86 space
Supermicro © 2009Confidential Oil & Gas Application Oil & Gas Exploration - Seismic Data Analysis Supermicro 2U Twin² System – 512 nodes
Supermicro © 2009Confidential HPC Implementation with PRACE Challenge Customer: Swedish National Infrastructure for Computing (SNIC), Royal Institute of Technology (KTH), Sweden, jointly with the Partnership for Advanced Computing in Europe (PRACE) Need: A general purpose HPC cluster for assessment of energy efficiency and compute density achievable using standard components Buying Criteria: Superior performance/watt/sq.ft. Ability to control energy consumption based on workload End-to-End Non-blocking, high throughput IO connectivity Latest server chipset supporting PCI-E Gen2 Non-proprietary x86 architecture Solution Collaborative effort between Supermicro and AMD Infrastructure (SuperBlade ® ): 18 7U SuperBlade® enclosures 10 4-way blades with 240 processor cores per 7U QDR InfiniBand switch (40Gb/s quad data rate) High efficiency, N+1 redundant power supplies (93% efficiency) Blade enclosure management solution including KVM/IP on each node Processor: Six-Core AMD Opteron™ HE processor Chipset: AMD SR5670 chipset supporting HyperTransport™ 3 interface, PCI-E Gen2 IO connectivity, APML power management Computing Capacity: way systems with total 720 processors – 4320 cores
Supermicro © 2009Confidential References LLNL: Lawrence Livermore National Laboratory Dr. Mark Seager Director of the Advanced Simulation and Computing Program V P POBOX 808, L-554, East Ave., Livermore, CA CERN: European Organization for Nuclear Research Dr. Helge Meinhard Server & Storage IT in charge 1211 Geneva 23, Switzerland Tel PRACE Prof. Lennart Johnsson PDC Director, Teknikringen 14, Royal Institute of Technology SE Stockholm, Sweden