Personal Chris Ward CS147 Fall 2008.  Recent offerings from NVIDA show that small companies or even individuals can now afford and own Super Computers.

Slides:



Advertisements
Similar presentations
Parallel Computing Glib Dmytriiev
Advertisements

Hardware specification. The “brain” The brain processes instructions, performs calculations, and manages the flow of information through the body The.
Super Computers By Phuong Vo.
Vectors, SIMD Extensions and GPUs COMP 4611 Tutorial 11 Nov. 26,
Multi-core and tera- scale computing A short overview of benefits and challenges CSC 2007 Andrzej Nowak, CERN
Monte-Carlo method and Parallel computing  An introduction to GPU programming Mr. Fang-An Kuo, Dr. Matthew R. Smith NCHC Applied Scientific Computing.
MULTICORE PROCESSOR TECHNOLOGY.  Introduction  history  Why multi-core ?  What do you mean by multicore?  Multi core architecture  Comparison of.
What is GPGPU? Many of these slides are taken from Henry Neeman’s presentation at the University of Oklahoma.
BY MANISHA JOSHI.  Extremely fast data processing-oriented computers.  Speed is measured in “FLOPS”.  For highly calculation-intensive tasks.  For.
GPU System Architecture Alan Gray EPCC The University of Edinburgh.
XEON PHI. TOPICS What are multicore processors? Intel MIC architecture Xeon Phi Programming for Xeon Phi Performance Applications.
Utilization of GPU’s for General Computing Presenter: Charlene DiMeglio Paper: Aspects of GPU for General Purpose High Performance Computing Suda, Reiji,
Appendix A — 1 FIGURE A.2.2 Contemporary PCs with Intel and AMD CPUs. See Chapter 6 for an explanation of the components and interconnects in this figure.
HPCC Mid-Morning Break High Performance Computing on a GPU cluster Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
GRAPHICS AND COMPUTING GPUS Jehan-François Pâris
A many-core GPU architecture.. Price, performance, and evolution.
GPUs. An enlarging peak performance advantage: –Calculation: 1 TFLOPS vs. 100 GFLOPS –Memory Bandwidth: GB/s vs GB/s –GPU in every PC and.
Supercomputers Daniel Shin CS 147, Section 1 April 29, 2010.
GPU PROGRAMMING David Gilbert California State University, Los Angeles.
1 ITCS 6/8010 CUDA Programming, UNC-Charlotte, B. Wilkinson, Jan 19, 2011 Emergence of GPU systems and clusters for general purpose High Performance Computing.
Central Processing Unit IS – 221 PC Admin 2/26/2007 by Len Krygsman IV.
Heterogeneous Computing Dr. Jason D. Bakos. Heterogeneous Computing 2 “Traditional” Parallel/Multi-Processing Large-scale parallel platforms: –Individual.
Buying a Computer By: Jemima, Sakshi, and Amena BTT101.
GPGPU overview. Graphics Processing Unit (GPU) GPU is the chip in computer video cards, PS3, Xbox, etc – Designed to realize the 3D graphics pipeline.
Real Parallel Computers. Modular data centers Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra,
Michael Monroig Michael Fiorelli.  The Processor is also known as the CPU or Central Processing Unit.  Processors carry out the instructions of computer.
HPCC Mid-Morning Break Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery Introduction to the new GPU (GFX) cluster.
Lecture 2 : Introduction to Multicore Computing Bong-Soo Sohn Associate Professor School of Computer Science and Engineering Chung-Ang University.
Exploiting Disruptive Technology: GPUs for Physics Chip Watson Scientific Computing Group Jefferson Lab Presented at GlueX Collaboration Meeting, May 11,
Comp-TIA Standards.  AMD- (Advanced Micro Devices) An American multinational semiconductor company that develops computer processors and related technologies.
1 ITCS 4/5010 CUDA Programming, UNC-Charlotte, B. Wilkinson, Dec 31, 2012 Emergence of GPU systems and clusters for general purpose High Performance Computing.
Training Program on GPU Programming with CUDA 31 st July, 7 th Aug, 14 th Aug 2011 CUDA Teaching UoM.
Different CPUs CLICK THE SPINNING COMPUTER TO MOVE ON.
Shared memory systems. What is a shared memory system Single memory space accessible to the programmer Processor communicate through the network to the.
Implementation of Parallel Processing Techniques on Graphical Processing Units Brad Baker, Wayne Haney, Dr. Charles Choi.
By Arun Bhandari Course: HPC Date: 01/28/12. GPU (Graphics Processing Unit) High performance many core processors Only used to accelerate certain parts.
Company LOGO High Performance Processors Miguel J. González Blanco Miguel A. Padilla Puig Felix Rivera Rivas.
College of Engineering Representing Numbers in a Computer Section B, Spring 2003 COE1361: Computing for Engineers COE1361: Computing for Engineers 1 COE1361:
The CPU (or Central Processing Unit. Statistics Clock speed – number of instructions that can be executed per second Data width – The number of bits held.
Copyright © 2007 Heathkit Company, Inc. All Rights Reserved PC Fundamentals Presentation 27 – A Brief History of the Microprocessor.
Status of the L1 STS Tracking I. Kisel GSI / KIP CBM Collaboration Meeting GSI, March 12, 2009.
The Central Processing Unit
Taking the Complexity out of Cluster Computing Vendor Update HPC User Forum Arend Dittmer Director Product Management HPC April,
Use of GPUs in ALICE (and elsewhere) Thorsten Kollegger TDOC-PG | CERN |
Large-scale Deep Unsupervised Learning using Graphics Processors
S&T IT Research Support 11 March, 2011 ITCC. Fast Facts Team of 4 positions 3 positions filled Focus on technical support of researchers Not “IT” for.
Emergence of GPU systems and clusters for general purpose high performance computing ITCS 4145/5145 April 3, 2012 © Barry Wilkinson.
CPU Inside Maria Gabriela Yobal de Anda L#32 9B. CPU Called also the processor Performs the transformation of input into output Executes the instructions.
Hardware Acceleration Using GPUs M Anirudh Guide: Prof. Sachin Patkar VLSI Consortium April 4, 2008.
Carlo del Mundo Department of Electrical and Computer Engineering Ubiquitous Parallelism Are You Equipped To Code For Multi- and Many- Core Platforms?
ICAL GPU 架構中所提供分散式運算 之功能與限制. 11/17/09ICAL2 Outline Parallel computing with GPU NVIDIA CUDA SVD matrix computation Conclusion.
GPUs: Overview of Architecture and Programming Options Lee Barford firstname dot lastname at gmail dot com.
Today’s Computers By Sharif Safdary. The right computer for you. Advantages to Laptops Advantages to Laptops Size Size Weight Weight Portability Portability.
GRID Applications for the LHC ALICE Experiment Máté Ferenc Nagy Eötvös University Applied-Physics student Supervisor : Gergely Barnaföldi MTA KFKI RMKI.
MULTICORE PROCESSOR TECHNOLOGY.  Introduction  history  Why multi-core ?  What do you mean by multicore?  Multi core architecture  Comparison of.
High Performance Computing
By Mr. Brincks June 15, Essential Components Computer Case Silver Thermaltake Armor LCS All Aluminum Full Tower Computer Chassis VE2000SWA, BTX.
Cross correlators are a highly computationally intensive part of any radio interferometer. Although the number of operations per data sample is small,
GPGPU introduction. Why is GPU in the picture Seeking exa-scale computing platform Minimize power per operation. – Power is directly correlated to the.
Parallel Computers Today Oak Ridge / Cray Jaguar > 1.75 PFLOPS Two Nvidia 8800 GPUs > 1 TFLOPS Intel 80- core chip > 1 TFLOPS  TFLOPS = floating.
Emergence of GPU systems for general purpose high performance computing ITCS 4145/5145 © Barry Wilkinson GPUIntro.ppt Oct 30, 2014.
Sobolev(+Node 6, 7) Showcase +K20m GPU Accelerator.
Jun Doi IBM Research – Tokyo Early Performance Evaluation of Lattice QCD on POWER+GPU Cluster 17 July 2015.
Presented by Leo Pleše ScienceUp.org
Chapter 10: Computer systems (1)
CS111 Computer Programming
Super Computing By RIsaj t r S3 ece, roll 50.
Multicore and GPU Programming
Multicore and GPU Programming
CSE 102 Introduction to Computer Engineering
Presentation transcript:

Personal Chris Ward CS147 Fall 2008

 Recent offerings from NVIDA show that small companies or even individuals can now afford and own Super Computers.  Just what is a “Super Computer”?

 “A supercomputer is a computer that is at the frontline of current processing capacity, particularly speed of calculation. “ –wikipedia  The term supercomputer itself is rather fluid, and today's supercomputer tends to become tomorrow's ordinary computer.

 In computing, FLOPS (or flops or flop/s) is an acronym meaning FLoating point Operations Per Second  Today’s two fastest Super Computers operate at > 1 petaflop/s (represents one quadrillion floating point operations per second. )

Most applications use double precision math for the following reasons: 1. To minimize the accumulation of round-off error, 2. For ill-conditioned problems that require higher precision 3. The 8 bit exponent defined by the IEEE floating point standard for 32-bit arithmetic will not accommodate the calculation, or 4. There are critical sections in the code which require higher precision.

 Some of the latest GPUs and CPUs have LARGE differences in flops for single vs double precision  Researchers are recognizing this trend and re- examining their algorithms to take advantage of single precision whenever possible

 GPU Graphics Processing Unit  CPU Central Processing unit DeviceTransistor CountYear ReleasedMaker RV AMD GT NVIDIA Dual-Core Itanium Intel Quad-Core Itanium Tukwila Intel

 Tesla S1070 Computer is a four-teraflop 1U system powered by the world’s first one-teraflop processor.  Each of the 4 cores contains 240 GPUs (960 total)  Single Precision floating point performance (peak) 3.73 to 4.14 Tflops  Double Precision floating point performance (peak) 311 to 345 Gflops  Starting at about $3,995 to ~ $10,000

1 Board= 240 Processing Cores ~ 1 TFlop performance 4GB GDDR3 RAM PCI-Express 2.0, Write your code in C and run in Windows or Linux $1,699.99

         details.asp?EdpNo= &CatId= details.asp?EdpNo= &CatId=4044