An Introduction and Overview of Grid Computing Presenter: Xiaofei Cao Patrick Berg.

Slides:



Advertisements
Similar presentations
Distributed Data Processing
Advertisements

1 US activities and strategy :NSF Ron Perrott. 2 TeraGrid An instrument that delivers high-end IT resources/services –a computational facility – over.
ASCR Data Science Centers Infrastructure Demonstration S. Canon, N. Desai, M. Ernst, K. Kleese-Van Dam, G. Shipman, B. Tierney.
IBM 1350 Cluster Expansion Doug Johnson Senior Systems Developer.
High Performance Computing Course Notes Grid Computing.
Dinker Batra CLUSTERING Categories of Clusters. Dinker Batra Introduction A computer cluster is a group of linked computers, working together closely.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
CoreGRID Workpackage 5 Virtual Institute on Grid Information and Monitoring Services Authorizing Grid Resource Access and Consumption Erik Elmroth, Michał.
Introduction and Overview “the grid” – a proposed distributed computing infrastructure for advanced science and engineering. Purpose: grid concept is motivated.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
Simo Niskala Teemu Pasanen
SPRING 2011 CLOUD COMPUTING Cloud Computing San José State University Computer Architecture (CS 147) Professor Sin-Min Lee Presentation by Vladimir Serdyukov.
Distributed Computer Architecture Benjamin Jordan, Kevin Cone, Jason Bradley.
SALSASALSASALSASALSA Digital Science Center June 25, 2010, IIT Geoffrey Fox Judy Qiu School.
Network, Operations and Security Area Tony Rimovsky NOS Area Director
Hands-On Microsoft Windows Server 2008 Chapter 1 Introduction to Windows Server 2008.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Scientific Data Infrastructure in CAS Dr. Jianhui Scientific Data Center Computer Network Information Center Chinese Academy of Sciences.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
and beyond Office of Vice President for Information Technology.
DISTRIBUTED COMPUTING
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
1 Introduction to Middleware. 2 Outline What is middleware? Purpose and origin Why use it? What Middleware does? Technical details Middleware services.
The Grid System Design Liu Xiangrui Beijing Institute of Technology.
The Anatomy of the Grid Mahdi Hamzeh Fall 2005 Class Presentation for the Parallel Processing Course. All figures and data are copyrights of their respective.
1 4/23/2007 Introduction to Grid computing Sunil Avutu Graduate Student Dept.of Computer Science.
Service - Oriented Middleware for Distributed Data Mining on the Grid ,劉妘鑏 Antonio C., Domenico T., and Paolo T. Journal of Parallel and Distributed.
SALSASALSASALSASALSA FutureGrid Venus-C June Geoffrey Fox
Introduction to Grid Computing Ed Seidel Max Planck Institute for Gravitational Physics
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
GRID ARCHITECTURE Chintan O.Patel. CS 551 Fall 2002 Workshop 1 Software Architectures 2 What is Grid ? "...a flexible, secure, coordinated resource- sharing.
Authors: Ronnie Julio Cole David
Copyright © 2011 Curt Hill MIMD Multiple Instructions Multiple Data.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Introduction to Grids By: Fetahi Z. Wuhib [CSD2004-Team19]
TeraGrid Quarterly Meeting Arlington, VA Sep 6-7, 2007 NCSA RP Status Report.
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
7. Grid Computing Systems and Resource Management
Cyberinfrastructure: Many Things to Many People Russ Hobby Program Manager Internet2.
Aneka Cloud ApplicationPlatform. Introduction Aneka consists of a scalable cloud middleware that can be deployed on top of heterogeneous computing resources.
Securing the Grid & other Middleware Challenges Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer.
Network, Operations and Security Area Tony Rimovsky NOS Area Director
GRID ANATOMY Advanced Computing Concepts – Dr. Emmanuel Pilli.
3/12/2013Computer Engg, IIT(BHU)1 PARALLEL COMPUTERS- 1.
1 TCS Confidential. 2 Objective : In this session we will be able to learn:  What is Cloud Computing?  Characteristics  Cloud Flavors  Cloud Deployment.
The Globus Toolkit The Globus project was started by Ian Foster and Carl Kesselman from Argonne National Labs and USC respectively. The Globus toolkit.
All Hands Meeting 2005 BIRN-CC: Building, Maintaining and Maturing a National Information Infrastructure to Enable and Advance Biomedical Research.
Northwest Indiana Computational Grid Preston Smith Rosen Center for Advanced Computing Purdue University - West Lafayette West Lafayette Calumet.
TG ’08, June 9-13, State of TeraGrid John Towns Co-Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
INTRODUCTION TO GRID & CLOUD COMPUTING U. Jhashuva 1 Asst. Professor Dept. of CSE.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
Accessing the VI-SEEM infrastructure
DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING CLOUD COMPUTING
Clouds , Grids and Clusters
Volunteer Computing for Science Gateways
Joint Techs, Columbus, OH
Jeffrey P. Gardner Pittsburgh Supercomputing Center
Grid Computing.
University of Technology
GRID COMPUTING PRESENTED BY : Richa Chaudhary.
Grid Computing B.Ramamurthy 9/22/2018 B.Ramamurthy.
Introduction to Grid Technology
The Anatomy and The Physiology of the Grid
The Anatomy and The Physiology of the Grid
Presentation transcript:

An Introduction and Overview of Grid Computing Presenter: Xiaofei Cao Patrick Berg

Introduction At its most basic level, grid computing is a computer network in which each computer's resources are shared with every other computer in the system. Processing power, memory and data storage are all community resources that authorized users can tap into and leverage for specific tasks.

Comparing with other technologies Ian Foster et al., “Cloud Computing and Grid Computing 360-Degree Compared”; The problems are mostly the same in clouds and Grids. There is a common need to be able to manage large facilities; to define methods by which consumers discover, request, and use resources provided by the central facilities; and to implement the often highly parallel computations that execute on those resources. Details differ, but the two communities are struggling with many of the same issues

Indiana University (IU) o “Big Red” - Big Red is a distributed shared-memory cluster, consisting of 768 IBM JS21 Blades, each with two dual- core PowerPC 970 MP processors, 8GB of memory, and a PCI-X Myrinet 2000 adapter for highbandwidth, low-latency Message Passing Interface (MPI) applications. In addition to local scratch disks, the Big Red compute nodes are connected via gigabit Ethernet to a 266TB GPFS file system, hosted on 16 IBM p505 Power5 systems. Joint Institute for Computational Sciences (JICS) o University of Tennessee and ORNL o Future expansions are being planned that would add a 40-teraflops Cray XT3 system to the TeraGrid and they have plans to expand that to a 170 teraflops Cray XT4 system which in turn will be upgraded to a 10,000+ compute socket Cray system of approximately 1 petaflop. ouisiana Optical Network Initiative (LONI) o “Queen Bee”, the core cluster of LONI, is a 50.7 Teraflops Peak Performance 668 node Dell PowerEdge 1950 cluster running the Red Hat Enterprise Linux 4 operating system. Each node contains two Quad Core Oak Ridge National Laboratory (ORNL) o In this case it is more of a user than a provider. Their users of neutron science facilities (the High Flux Isotope Reactor and the Spallation Neutron Source) will be able to access TeraGrid resources and services for their data storage, analysis, and simulation. National Center for Supercomputing Applications (NCSA) o University of Illinois Urbana- Champaign o Provides 10 teraflops of capability computing through its IBM Linux cluster, which consists of 1,776 Itanium2 processors. In addition to the processing power capability, the NCSA also includes 600 terabytes of secondary storage and 2 petabytes of archival storage capacity. Pittsburgh Supercomputing Center (PSC) o Provides computational power via its 3,000-processor HP Alpha Server system, TCS-1, which offers 6 teraflops of capability coupled uniquely to a 21-node visualization system. It also provides a 128-processor, 512- gigabyte shared-memory HP Marvel system, a 150-terabyte disk cache, and a mass storage system with a capacity of 2.4 petabytes. University of Chicago/Argonne National Laboratory (UC/ANL) o Provides users with high-resolution rendering and remote visualization capabilities via a 1-teraflop IBM Linux cluster with parallel visualization hardware. Texas Advanced Computing Center (TACC) o Provides a 1024-processor Cray/Dell Xeon-based Linux cluster, a 128- processor Sun E25K Terascale visualization machine with 512 gigabytes of shared memory, for a total of 6.75 teraflops of computing/visualization capacity, in addition to a 50 terabyte Sun storage area network. Only half of the cycles produced by these resources are available to TeraGrid users. San Diego Supercomputer Center (SDCS) o It leads the TeraGrid data and knowledge management effort. It provides a data-intensive IBM Linux cluster based on Itanium processors, that reaches over 4 teraflops and 540 terabytes of network disk storage. In addition, a portion of SDSC’s IBM 10- teraflops supercomputer is assigned to the TeraGrid. An IBM HPSS archive currently stores a petabyte of data. Purdue University o Provide 6 teraflops of computing capability, 400 terabytes of data storage capacity, visualization resources, access to life science data sets, and a connection to the Purdue Terrestrial Observatory National Center for Atmospheric Research (NCAR) o Boulder, CO. o “Frost” - BlueGene/L computing system. The 2048-processor system brings 250 teraflops of computing capability and more than 30 petabytes of online and archival data storage to the TeraGrid. Hey, guys, I have a project need to use the TeraGrid to work out the result Computing on the Teragrid is free but you still need to submit a proposal outlining your research for an initial allocation. From the time the proposal is submitted to the time you can log-on and use the Teragrid is usually about 6 months so the first step is to plan ahead. The proposal for an initial allocation of 200,000 core hours is short (2-3 paragraphs) and takes 2-3 weeks to approve. Teragrid

Who support TeraGrid The US National Science Foundation (NSF) issued a solicitation asking for a "distributed terascale facility" from program director Richard L. Hilderbrandt. The TeraGrid project was launched in August 2001 with $53 million in funding to four sites. In October 2003, NSF awarded $10 million to add four sites to TeraGrid as well as to establish a third network hub. n August 2005, NSF's newly created office of cyberinfrastructure extended support for another five years with a $150 million set of awards.

Indian gird(GARUDA)

Chinese National grid

UK E-Science grid

Common business model The business model for Grids (at least that found in academia or government labs) is project- oriented in which the users or community represented by that proposal have certain number of service units (i.e. CPU hours) they can spend. For example, the TeraGrid operates in this fashion, and requires increasingly complex proposals be written for increasing number of computational power. The TeraGrid has more than a dozen Grid sites, all hosted at various institutions around the country.

Is mining bitcoin a grid computing project? Share processing power. (Wasting?) Distributed computing. Using network to communicate with each other. So yes.

Architecture of the Grid Grid Architecture is broadly defined around the concept and requirements of Virtual Organizations (VO). Virtual Organizations are a set of individuals, institutions or systems defined by sharing rules. As such, these Virtual Organizations require highly flexible sharing relationships, and price control over the use of shared resources. Grid Architecture then is centrally focused on interoperability, making it a “protocol” based architecture.

The Grid Architecture Model The architecture of the Grid is composed of five levels, each governed by a set of protocols. Application Layer Collective Layer Resource Layer Connectivity Layer Fabric Layer

The Fabric Layer The Fabric Layer provides access to different resource types, and implements the local, resource-specific operations that occur on specific resources. Some of the resources handled by the fabric layer: – Computational Resources – Storage Resources – Network Resources – Code Repositories

The Connectivity Layer The connectivity layer defines core communication and authentication protocols required for specific network transactions. As such, the requirements for this layer include providing easy transactions and secure connections.

Security of the Grid The security of the grid is handled primarily by the connectivity layer, and assumes that resources in the grid are heterogeneous and dynamic. The key focus of security in the grid is: – Single Sign On – Users log in once and have access to multiple Grid resources (as determined by the Fabric Layer). – Delegation – A program may access resources for a user, and can also delegate additional tasks to other programs. – Integrity and Segregation – Resources belonging to one user cannot be accessed by another who has not been authoritized to access those resources, including during resource-transfer. – Coordinated Resource Allocation, Reservation, and Sharing – Global and local resource usage and allocation policies must be taken into consideration.

The Resource Layer The Resource layer, much like the name implies, manages the transaction and delivery of resources. In particular, it defines protocols for the secure negotiation, initiation, control, accounting and payment of operation on individual resources. There are two types of protocols associated with the resource layer: – Information Protocols – These obtain information about a resource. – Management Protocols – These negotiate access to a shared resource.

The Collective Layer The collective layer focuses more on a global scale, capturing interactions across collections of resources. Essentially, the collective layer builds upon the resource and connectivity layers to implement new sharing behaviors without effecting the resources themselves. A number of different protocols and services can be implemented in the collective layer, such as directory services, brokering services, workload management systems, and monitoring and diagnostics services. What components are used in the collective layer depends largely on each specific Virtual Organization.

The Application Layer The final layer in the overall Grid Architecture, the Application Layer, is comprised of user applications that operate within the Virtual Organization environment. An application calls upon the various services defined by the lower layers, and each layer has protocols to deal with each service request made by an application in the application layer.

Application of the Grid Grids support a large range of applications, including tightly coupled parallel tasks and loosely-coupled applications. These can largely be scheduled on many different computing resources across multiple virtual organizations, and they can be small or large, use one processor or several, and be static or dynamic. These applications typically use message passing interface (MPI) to achieve the needed inter-process communication. On the other hand, Grids have also seen great success in the execution of more loosely coupled applications that tend to be managed and executed through workflow systems or other sophisticated and complex applications.

Molecular Biosciences Physics Astronomical Sciences Chemistry Materials Research Chemical, Thermal Systems Atmospheric Sciences Advanced Scientific Computing Earth Sciences Biological and Critical Systems Ocean Sciences Cross-Disciplinary Activities

Uses of the Grid Today Grids are employed in a number of environments today and used for a number of different applications. and – Searching for extra- terrestrial life and folding proteins. European Grid Infrastructure and LHC Computing Grid – Used in experiments with the CERN Large Hadron Collider. United Devices Cancer Research Project The World Grid Community – Works with 400 other companies and is used for various scientific research. Currently, it has more processing power than the fastest supercomputer available today.

How Grid Projects Work Control Node Contact Control Node Data Sent for Analysis Users with grid software application installed Data Analyzed using CPU Analyzed Data sent back to Node

How Grid Projects Work Continued As more users join the project, the process becomes increasingly parallel. Contact Data Sent Analyzed Data Returned

Ian Foster et al., “Cloud Computing and Grid Computing 360-Degree Compared”; Parvin Asadzadeh et al., “Global Grids and Software Toolkits: A Study of Four Middleware Technologies”; Ian Foster et al., “The Anatomy of the Grid: Enabling Scalable Virtual Organizations”;