A Data Communication Reliability and Trustability Study for Cluster Computing Speaker: Eduardo Colmenares Midwestern State University Wichita Falls, TX.

Slides:



Advertisements
Similar presentations
Computing Infrastructure
Advertisements

Providing Fault-tolerance for Parallel Programs on Grid (FT-MPICH) Heon Y. Yeom Distributed Computing Systems Lab. Seoul National University.
System Area Network Abhiram Shandilya 12/06/01. Overview Introduction to System Area Networks SAN Design and Examples SAN Applications.
Building a Distributed Full-Text Index for the Web S. Melnik, S. Raghavan, B.Yang, H. Garcia-Molina.
IBM 1350 Cluster Expansion Doug Johnson Senior Systems Developer.
♦ Commodity processor with commodity inter- processor connection Clusters Pentium, Itanium, Opteron, Alpha GigE, Infiniband, Myrinet, Quadrics, SCI NEC.
Information Technology Center Introduction to High Performance Computing at KFUPM.
LinkSCEEM-2: A computational resource for the development of Computational Sciences in the Eastern Mediterranean Mostafa Zoubi SESAME SESAME – LinkSCEEM.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
IBM RS6000/SP Overview Advanced IBM Unix computers series Multiple different configurations Available from entry level to high-end machines. POWER (1,2,3,4)
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
Teraserver Darrel Sharpe Matt Todd Rob Neff Mentor: Dr. Palaniappan.
CS 550 Amoeba-A Distributed Operation System by Saie M Mulay.
Understanding Operating Systems 1 Overview Introduction Operating System Components Machine Hardware Types of Operating Systems Brief History of Operating.
Parallel Rendering Ed Angel
1 SOFTWARE TECHNOLOGIES BUS Abdou Illia, Spring 2007 (Week 2, Thursday 1/18/2007)
Real Parallel Computers. Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra, Meuer, Simon Parallel.
Research Computing with Newton Gerald Ragghianti Nov. 12, 2010.
Router Architectures An overview of router architectures.
Gordon: Using Flash Memory to Build Fast, Power-efficient Clusters for Data-intensive Applications A. Caulfield, L. Grupp, S. Swanson, UCSD, ASPLOS’09.
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
Prepared by Careene McCallum-Rodney Hardware specification of a computer system.
Real Parallel Computers. Modular data centers Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra,
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
5.3 HS23 Blade Server. The HS23 blade server is a dual CPU socket blade running Intel´s new Xeon® processor, the E5-2600, and is the first IBM BladeCenter.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Basic Computer Structure and Knowledge Project Work.
HPC at IISER Pune Neet Deo System Administrator
ATM and Fast Ethernet Network Interfaces for User-level Communication Presented by Sagwon Seo 2000/4/13 Matt Welsh, Anindya Basu, and Thorsten von Eicken.
High Performance Computing G Burton – ICG – Oct12 – v1.1 1.
Computer Maintenance Unit Subtitle: Bus Structures Excerpted from Copyright © Texas Education Agency, All rights reserved.
 What is an operating system? What is an operating system?  Where does the OS fit in? Where does the OS fit in?  Services provided by an OS Services.
林俊宏 Parallel Association Rule Mining based on FI-Growth Algorithm Bundit Manaskasemsak, Nunnapus Benjamas, Arnon Rungsawang.
Translate the following message:
So, Jung-ki Distributed Computing System LAB School of Computer Science and Engineering Seoul National University Implementation of Package Management.
High Performance User-Level Sockets over Gigabit Ethernet Pavan Balaji Ohio State University Piyush Shivam Ohio State University.
1b.1 Types of Parallel Computers Two principal approaches: Shared memory multiprocessor Distributed memory multicomputer ITCS 4/5145 Parallel Programming,
On the Varieties of Clouds for Data Intensive Computing 董耀文 Antslab Robert L. Grossman University of Illinois at Chicago And Open Data.
© 2012 MELLANOX TECHNOLOGIES 1 The Exascale Interconnect Technology Rich Graham – Sr. Solutions Architect.
Workload-driven Analysis of File Systems in Shared Multi-Tier Data-Centers over InfiniBand K. Vaidyanathan P. Balaji H. –W. Jin D.K. Panda Network-Based.
Hardware Trends. Contents Memory Hard Disks Processors Network Accessories Future.
1 CMG, 2006 Reno Yiping Ding and Ethan Bolker How Many Guests Can You Serve? - On the Number of Partitions.
1 Network Monitoring Mi-Jung Choi Dept. of Computer Science KNU
HPCVL High Performance Computing Virtual Laboratory Founded 1998 as a joint HPC lab between –Carleton U. (Comp. Sci.) –Queen’s U. (Engineering) –U. of.
Parallelization of the Classic Gram-Schmidt QR-Factorization
The Guts. CPU CPU Socket The CPU is generally a 2 inch ceramic square with a silicon chip located inside. The chip usually about the size of a thumbnail.
Laboratório de Instrumentação e Física Experimental de Partículas GRID Activities at LIP Jorge Gomes - (LIP Computer Centre)
1 Buses and types of computer. Paul Strickland Liverpool John Moores University.
COMPUTER ARCHITECTURE. Recommended Text 1Computer Organization and Architecture by William Stallings 2Structured Computer Organisation Andrew S. Tanenbaum.
ITEP computing center and plans for supercomputing Plans for Tier 1 for FAIR (GSI) in ITEP  8000 cores in 3 years, in this year  Distributed.
Computer Systems. This Module Components Home PC Inputs Processor Memory Motherboards Auxiliary Storage Outputs.
Rob Allan Daresbury Laboratory NW-GRID Training Event 25 th January 2007 Introduction to NW-GRID R.J. Allan CCLRC Daresbury Laboratory.
Chapter 13 – I/O Systems (Pgs ). Devices  Two conflicting properties A. Growing uniformity in interfaces (both h/w and s/w): e.g., USB, TWAIN.
Efficiency of small size tasks calculation in grid clusters using parallel processing.. Olgerts Belmanis Jānis Kūliņš RTU ETF Riga Technical University.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 1.
INTRODUCTION OF SYSTEM & APPLICATION SOFTWARE. OPERATING SYSTEM (OS) An operating system, or OS, is a software program that enables the computer hardware.
CIP HPC CIP - HPC HPC = High Performance Computer It’s not a regular computer, it’s bigger, faster, more powerful, and more.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
Computer Performance. Hard Drive - HDD Stores your files, programs, and information. If it gets full, you can’t save any more. Measured in bytes (KB,
Creating Grid Resources for Undergraduate Coursework John N. Huffman Brown University Richard Repasky Indiana University Joseph Rinkovsky Indiana University.
Personal Computer (PC)  Computer advertisement specification Intel® Pentium 4 Processor at 3.06GHz with 512K cache 512MB DDR SDRAM 200GB ATA-100 Hard.
Brief introduction about “Grid at LNS”
Unit Subtitle: Bus Structures Excerpted from
HPC Roadshow Overview of HPC systems and software available within the LinkSCEEM project.
Infiniband Architecture
LinkSCEEM-2: A computational resource for the development of Computational Sciences in the Eastern Mediterranean Mostafa Zoubi SESAME Outreach SESAME,
Introduction to client/server architecture
CS 345A Data Mining MapReduce This presentation has been altered.
Cluster Computers.
Presentation transcript:

A Data Communication Reliability and Trustability Study for Cluster Computing Speaker: Eduardo Colmenares Midwestern State University Wichita Falls, TX

Introduction HPC –Relevant to a variety of sciences, not only CS –More and more researchers considering new technologies capable of considerable computational power Distributed Shared Memory Systems (Clusters) –Combine computational capabilities of multiple nodes via a high speed communication network. Clusters, if public –Multiple users »Need policies to schedule jobs and manage resources –We acknowledge to potential characteristics of interest Performance Consistent time accuracy 2

Ping-Pong Test Purpose: –To measure the end-to-end delay time associated with sending a message back and forth between processes in a cluster of workstations (or any other parallel system) –Two variants 3

Both Ping-Pong Tests Message Size = 8 bytes 4 Ping-Pong-APing-Pong-B

Testing Environments Community-Cluster This cluster has 12TB of public shared Lustre storage and three groups of public and private nodes, all connected by SDR Infiniband and Gigabit Ethernet. The quad-core nodes have Infinihost III Lx (PCI-e) cards, and the older nodes have Infinihost (PCI-X) cards. Public quad-core (512 cpu, 4.77 TF). 64 nodes with dual quad-core Intel 5345 processors (2.33 GHz) and 12GB of memory each. Designated compute-1-x, 2-x. Public single-core (128 cpu, 0.82 TF). 64 nodes with dual single-core Intel "Irwindale" processors (3.2 GHz) and 4 GB of memory each. Designated compute-3-x, 4-x, 5-x. Public AMD dual-core (8 cpu,.04 TF). 1 node with quad dual-core AMD 8218 processors (2.60 GHz) and 64GB of memory. Designated compute-8-1. Multi-user and 3 queues {2WKpar, 48Hquadpar, 2WKpar} 5 My-Cluster Each one of the nodes in this cluster has the following characteristics: one Intel(R) Pentium(R) 4 CPU at 1.70GHz, one 3Com PCI 3c905C Tornado network card. All the nodes in this cluster are interconnected via a 3Com® Super Stack® 3 Switch Port. Table 1 summarizes the major hardware differences between nodes. 1 user NODEMEMORY (MB) HARD DISK MB-T340016A, ATA MB-T340016A, ATA MB-MAXTOR 6L020J1, ATA MB-T340016A, ATA MB-MAXTOR 6L020J1, ATA MB-MAXTOR 6L020J1, ATA MB-WDC WD400BB-75DEA0, ATA MB-MAXTOR 6L040J2, ATA MB-MAXTOR 6L040J2, ATA

Experimental results Ping-Pong-A Community-Cluster (queue: 2WKpar) 6 Community-Cluster - 2WKpar - 9P – 1 st Sample Included Proc-Pair Average (usecs) Stdv (usecs) Min (usecs) Max (usecs) P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P Community-Cluster - 2WKpar - 9P – 1 st Sample Removed Proc-Pair Average (usecs) Stdv (usecs) Min (usecs) Max (usecs) P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P

Experimental results Ping-Pong-A Community-Cluster (queue: 2WKpar) 7

Experimental results Ping-Pong-A Community-Cluster (queue: 48Hpar) 8 Community-Cluster – 48Hpar - 9P – 1 st Sample Removed Proc-Pair Average (usecs) Stdv (usecs) Min (usecs) Max (usecs) P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P

Experimental results Ping-Pong-A Community-Cluster (queue: 48Hpar) 9

Experimental results Ping-Pong-A Community-Cluster (queue: 48Hquadpar) 10 Community-Cluster - 48Hquadpar-9P-1 st Sample Removed Proc-Pair Average (usecs) Stdv (usecs) Min (usecs) Max (usecs) P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P

Experimental results Ping-Pong-A Community-Cluster (queue: 48Hquadpar) 11

Experimental results Ping-Pong-A My-Cluster 12 My-Cluster - 9P - 1 st Sample Removed Proc-Pair Average (usecs) Stdv (usecs) Min (usecs) Max (usecs) P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P My-Cluster - 9P - 1 st SAMPLE INCLUDED Proc-Pair Average (usecs) Stdv (usecs) Min (usecs) Max (usecs) P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P

Experimental results Ping-Pong-A My-Cluster 13 Community-Cluster - 2WKpar - 9P – 1 st Sample Removed Proc-Pair Average (usecs) Stdv (usecs) Min (usecs) Max (usecs) P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P

Experimental results Ping-Pong-A Let’s compare 14 Community-Cluster - 2WKpar - 9P – 1 st Sample Removed Proc-Pair Average (usecs) Stdv (usecs) Min (usecs) Max (usecs) P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P

Experimental results Ping-Pong-B Community-Cluster (queue: 2WKpar) 15 Community-Cluster - 2WKpar - 9P – 1 st Sample Removed Proc-Pair Average (usecs) Stdv (usecs) Min (usecs) Max (usecs) P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P Community-Cluster- 2WKpar - 9P – 1 st SAMPLE INCLUDED Proc-Pair Average (usecs) Stdv (usecs) Min (usecs) Max (usecs) P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P

Experimental results Ping-Pong-B Community-Cluster (queue: 2WKpar) 16

Experimental results Ping-Pong-B Community-Cluster (queue: 48Hpar) 17 Community-Cluster – 48Hpar - 9P – 1 st Sample Removed Proc-Pair Average (usecs) Stdv (usecs) Min (usecs) Max (usecs) P0 – P P0 – P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P

Experimental results Ping-Pong-B Community-Cluster (queue: 48Hpar) 18

Experimental results Ping-Pong-B Community-Cluster (queue: 48Hquadpar) 19 Community-Cluster – 48Hquadpar - 9P – 1 st Sample Removed Proc-Pair Average (usecs) Stdv (usecs) Min (usecs) Max (usecs) P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P

Experimental results Ping-Pong-B Community-Cluster (queue: 48Hquadpar) 20

Experimental results Ping-Pong-B My-Cluster 21 My-Cluster - 9P – 1 st Sample Removed Proc-Pair Average (usecs) Stdv (usecs) Min (usecs) Max (usecs) P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P P0 - P

Experimental results Ping-Pong-B My-Cluster 22

Experimental results Ping-Pong-B Let’s compare 23

Was all this worth it? 24

Conclusions All three queues of the Community-Cluster are prone to pulses of some nature that can negatively impact high performance applications where consecutive steady and accurate time readings are needed to statistically validate a phenomena under study. The one-user cluster, referred to as My-Cluster, presents a much better behavior in terms of disturbances during communication between processes; the communication between processes is clearly more stable and steady for a considerable number of consecutive samples; as a consequence, it constitutes a more appropriate choice for time sensitive analysis. The third and most important conclusion is derived from both ping pong tests, and it is associated with the fact that only the first time that a pair of processes establish a communication, a one-time high fee to pay in terms of time will exist. Consequently, it is possible to think that some sort of communication initialization takes place during this first message, and that such situation does not occur for all subsequent communications. This finding is supported by both ping pong tests, ping-pong-A and ping-pong-B, and occurs in both clusters even though they have different operating systems, hardware architecture and communication capabilities. 25

26 Thank You Questions ?