Oh-kyoung Kwon Grid Computing Research Team KISTI First PRAGMA Institute MPICH-GX: Grid Enabled MPI Implementation to Support the Private IP and the Fault.

Slides:



Advertisements
Similar presentations
1 Inducements–Call Blocking. Aware of the Service?
Advertisements

Using Matrices in Real Life
Process Description and Control
Libra: An Economy driven Job Scheduling System for Clusters Jahanzeb Sherwani 1, Nosheen Ali 1, Nausheen Lotia 1, Zahra Hayat 1, Rajkumar Buyya 2 1. Lahore.
Distributed Systems Architectures
© 2008 Pearson Addison Wesley. All rights reserved Chapter Seven Costs.
Copyright © 2003 Pearson Education, Inc. Slide 1 Computer Systems Organization & Architecture Chapters 8-12 John D. Carpinelli.
Chapter 1 The Study of Body Function Image PowerPoint
Processes and Operating Systems
1 Copyright © 2013 Elsevier Inc. All rights reserved. Chapter 1 Embedded Computing.
Author: Julia Richards and R. Scott Hawley
1 Copyright © 2013 Elsevier Inc. All rights reserved. Appendix 01.
1 Copyright © 2013 Elsevier Inc. All rights reserved. Chapter 3 CPUs.
Properties Use, share, or modify this drill on mathematic properties. There is too much material for a single class, so you’ll have to select for your.
CSF4 Meta-Scheduler Tutorial 1st PRAGMA Institute Zhaohui Ding or
11 Application of CSF4 in Avian Flu Grid: Meta-scheduler CSF4. Lab of Grid Computing and Network Security Jilin University, Changchun, China Hongliang.
Experiences of Grid Enabled MPI Implementation named MPICH-GX with Atmospheric Applications Oh-kyoung Kwon, KISTI Salvador Castañeda, CICESE PRAGMA 11.
UNITED NATIONS Shipment Details Report – January 2006.
By Rick Clements Software Testing 101 By Rick Clements
1 Hyades Command Routing Message flow and data translation.
Grid Checkpoining Architecture Radosław Januszewski CoreGrid Summer School 2007.
1 RA I Sub-Regional Training Seminar on CLIMAT&CLIMAT TEMP Reporting Casablanca, Morocco, 20 – 22 December 2005 Status of observing programmes in RA I.
Exit a Customer Chapter 8. Exit a Customer 8-2 Objectives Perform exit summary process consisting of the following steps: Review service records Close.
Create an Application Title 1A - Adult Chapter 3.
FACTORING ax2 + bx + c Think “unfoil” Work down, Show all steps.
Year 6 mental test 5 second questions
Year 6 mental test 10 second questions
NGS computation services: API's,
Protocol layers and Wireshark Rahul Hiran TDTS11:Computer Networks and Internet Protocols 1 Note: T he slides are adapted and modified based on slides.
Auto-scaling Axis2 Web Services on Amazon EC2 By Afkham Azeez.
REVIEW: Arthropod ID. 1. Name the subphylum. 2. Name the subphylum. 3. Name the order.
Chapter 1: Introduction to Scaling Networks
1 The phone in the cloud Utilizing resources hosted anywhere Claes Nilsson.
INTRODUCTION TO SIMULATION WITH OMNET++ José Daniel García Sánchez ARCOS Group – University Carlos III of Madrid.
2 |SharePoint Saturday New York City
IP Multicast Information management 2 Groep T Leuven – Information department 2/14 Agenda •Why IP Multicast ? •Multicast fundamentals •Intradomain.
VOORBLAD.
1 Network Address Translation (NAT) Relates to Lab 7. Module about private networks and NAT.
Sample Service Screenshots Enterprise Cloud Service 11.3.
Copyright © 2012, Elsevier Inc. All rights Reserved. 1 Chapter 7 Modeling Structure with Blocks.
Factor P 16 8(8-5ab) 4(d² + 4) 3rs(2r – s) 15cd(1 + 2cd) 8(4a² + 3b²)
Basel-ICU-Journal Challenge18/20/ Basel-ICU-Journal Challenge8/20/2014.
1..
© 2007 Cisco Systems, Inc. All rights reserved.Cisco Public 1 EN0129 PC AND NETWORK TECHNOLOGY I IP ADDRESSING AND SUBNETS Derived From CCNA Network Fundamentals.
31242/32549 Advanced Internet Programming Advanced Java Programming
CONTROL VISION Set-up. Step 1 Step 2 Step 3 Step 5 Step 4.
© 2012 National Heart Foundation of Australia. Slide 2.
1 © 2004, Cisco Systems, Inc. All rights reserved. CCNA 1 v3.1 Module 2 Networking Fundamentals.
April 2003 ONLINE SERVICE DELIVERY Presentation. 2 What is Online Service Delivery? Vision The current vision of the Online Service Delivery program is.
Understanding Generalist Practice, 5e, Kirst-Ashman/Hull
Executional Architecture
Global Analysis and Distributed Systems Software Architecture Lecture # 5-6.
DB analyzer utility An overview 1. DB Analyzer An application used to track discrepancies and other reports in Sanchay Post Constantly updated by SDC.
Providing Fault-tolerance for Parallel Programs on Grid (FT-MPICH) Heon Y. Yeom Distributed Computing Systems Lab. Seoul National University.
Chapter 9: Subnetting IP Networks
Model and Relationships 6 M 1 M M M M M M M M M M M M M M M M
25 seconds left…...
Analyzing Genes and Genomes
Systems Analysis and Design in a Changing World, Fifth Edition
© 2007 Cisco Systems, Inc. All rights reserved.Cisco Public 1 Addressing the Network – IPv4 Network Fundamentals – Chapter 6.
©Brooks/Cole, 2001 Chapter 12 Derived Types-- Enumerated, Structure and Union.
Intracellular Compartments and Transport
PSSA Preparation.
VPN AND REMOTE ACCESS Mohammad S. Hasan 1 VPN and Remote Access.
Essential Cell Biology
TCP/IP Protocol Suite 1 Chapter 18 Upon completion you will be able to: Remote Login: Telnet Understand how TELNET works Understand the role of NVT in.
Traversing symmetric NAT with predictable port allocation function SIN 2014 Dušan Klinec, Vashek Matyáš Faculty of Informatics, Masaryk University.
Presentation transcript:

Oh-kyoung Kwon Grid Computing Research Team KISTI First PRAGMA Institute MPICH-GX: Grid Enabled MPI Implementation to Support the Private IP and the Fault Tolerance

2 Contents Motivation What is MPICH-GX ? Private IP Support Fault Tolerance Support Demos Job Submission Demo using Command Line Interface Grid Portal Demo Experiments w/ CICESE using MM5 Conclusion

3 The Problems (1/2) Running on a Grid presents the following problems: Standard MPI implementations require that all compute nodes are visible to each other, making it necessary to have them on public IP addresses –Public IP addresses for assigning to all compute nodes aren't always readily available –There are security issues when exposing all compute nodes with public IP addresses –At developing countries, the majority of government and research institutions only have public IP addresses

4 The Problems (2/2) Running on a Grid presents the following problems: If multiple processes run on distributed resources and a process dies, it is difficult to find one process What if the interruption of electric power happens on the remote site?

5 What is MPICH-GX? MPICH-GX is a patch of MPICH-G2 to extend following functionalities required in the Grid. Private IP Support Fault Tolerant Support MPICH-GX Abstract Device Interface Globus2 device Private IP module Fault tolerant module Topologyaware Globus I/O... local-job-manager hole-punching-server

6 Private IP Support MPICH-G2 does not support private IP clusters Cluster Head (public IP) Compute Nodes (private IP) Cluster Head (public IP) Compute Nodes (private IP) 1. Job submission 2. Job distribution 3. Job allocation ??? 4. Computation & communication How to do it with private IP? Public IP 5. Report result member s

7 How To Support Private IP (1/2) User-level proxy Use separate application It is easy to implement and portable But it causes performance degradation due to additional user level switching Cluster Head (public IP) Compute Nodes (private IP) Cluster Head (public IP) Compute Nodes (private IP) Relay App. Relay App.

8 How To Support Private IP (2/2) Kernel-level proxy Generally, it is neither easy to implement nor portable But it can minimize communication overhead due to firewall NAT (Network Address Translation) Main mechanisms of Linux masquerading Cluster Head (public IP) Compute Nodes (private IP) Cluster Head (public IP) Compute Nodes (public IP) NAT Out-going message Incoming message

9 Hole Punching Easily applicable kernel-level solution It is a way to access unreachable hosts with a minimal additional effort All you need is a server that coordinates the connections When each client registers with server, it records two endpoints for that client Cluster Head (public IP) Compute Nodes (private IP) Cluster Head (public IP) Compute Nodes (private IP) NAT Server

10 Hole Punching Modification of Hole Punching scheme (1/3) At initialization step, every MPI process behind private network clusters obtains masqueraded address through the trusted entity Cluster Head (public IP) Compute Nodes (private IP) Cluster Head (public IP) Compute Nodes (private IP) NAT Server : :

11 Hole Punching Modification of Hole Punching scheme (2/3) MPI processes distribute two endpoints Cluster Head (public IP) Compute Nodes (private IP) Cluster Head (public IP) Compute Nodes (private IP) NAT Server : :3000

12 Hole Punching Modification of Hole Punching scheme (3/3) Each MPI process tries only one kind of connection If destination belongs to the same cluster, it connects using private endpoint Otherwise, it connects using public endpoint Cluster Head (public IP) Compute Nodes (private IP) Cluster Head (public IP) Compute Nodes (private IP) NAT Server : :30000

13 Fault Tolerant Support We provide a checkpointing-recovery system for Grid. Our library requires no modifications of application source codes. affects the MPICH communication characteristics as less as possible. All of the implementation have been done at the low level, that is, the abstract device level

14 Recovery of Process Failure process PBS Modified GRAM Job Submission Service local-job- manager PBS Modified GRAM local-job- manager process notify resubmit

15 Recovery of Host Failure process PBS Modified GRAM Job Submission Service local-job- manager PBS Modified GRAM local-job- manager process resubmit Host Failur e notify

Job Submission Demo using Command Line Interface

17 MoreDream Toolkit

18 Relation with Globus Tookit MPICH-GX is integrated with GT Both GT 3.x and GT 4.x On GT3.x, it works with both GRASP of MoreDream and gxrun & gxlrm. At the moment, on GT4, it only works with gxrun & gxlrm, not with GRAM4 If MPICH-G4 releases, MPICH-GX will be based on MPICH-G4, which supports GRAM4

19 gxrun & gxlrm gxrun: a client which enables users to submit a Grid job JRDL(Job & Resource Description Language): A Job is described as a JRDL gxlrm: a server which accepts the Grid job on the each resource, and then allocates resources

20 JRDL (1/2) Job & Resource Description Language

21 JRDL (2/2)

22 Workflow of MPICH-GX job

Demonstration for launching the MPICH-GX job

Grid Portal for the MPI job

25 Grid Portal for the MPI job Grid Portal based on Gridsphere General Purpose Web Portal MM5 Grid Portal

26 MM5 Grid Portal The MM5 is a limited-area terrain-following sigma coordinate model designed to simulate mesoscale atmospheric circulation The MM5 Grid Portal is for the MM5 modeling system with which people can use it easily The portal is consisted of 4 building blocks as follows: Terrain-, Pre-, Main-, and Post-Processing As a testbed system, presently, the MM5 Grid Portal is targeted at the specific user group of Korea. But the portal will be gradually extended to the general purpose user services environment

27 Architecture of MM5 Grid Portal System MM5 Portal Service User Web Portal Grid Information Service Job Submission Service ComputeResources Seoul Terrain MPICH-GXMPICH-GX Pre-Processing Main Post-Processing Allocate Resources GridMiddleware Submit Get Status DaejeonBusanGwangjuPohang Allocate Resources

28 Demonstration for launching the Grid job via web portal gridsphere gridsphere

29 Experiences of MPICH-GX with Atmospheric Applications Collaboration efforts with PRAGMA people (Daniel Garcia Gradilla (CICESE), Salvador Castañeda(CICESE), and Cindy Zheng(SDSC)) Testing the functionalities of MPICH-GX with atmospheric applications of MM5/WRF models in our K*Grid Testbed MM5/WRF is a medium scale computer model for creating atmospheric simulations, weather forecasts. Applications of MM5/WRF models work well with MPICH- GX Functionality of the private IP could be usable practically, and the performance of the private IP is reasonable

30 Experiments pluto eolo Hurricane Marty Simulation Santana Winds Simulation output

31 Analyze the output 12 nodes on single cluster (orion cluster): 17.55min 12 nodes on cross-site 6 nodes on orion + 6 nodes on eros, where all nodes have public IP: 21min 6 nodes on orion + 6 nodes on eros, where the nodes on orion have private IP and the nodes on eros have public IP: 24min 16 nodes on cross-site 8 nodes on orion + 8 nodes on eros, where all nodes have public IP: 18min 8 nodes on orion + 8 nodes on eros, where the nodes on orion have private IP and the nodes on eros have public IP: 20min

32 Conclusion and Limitation MPICH-GX is a patch of MPICH-G2 to provide useful functionalities for supporting the private IP and fault tolerance The application of WRF model work well with MPICH-GX at Grid environments. but when using MM5 model, we have scaling problems for clusters bigger than 24 processors using MPICH-GX. The functionality of the private IP could be usable practically, and the performance of the private IP is reasonable MPICH-GX is compatible with Globus Toolkit 4.x. But, doesnt work with GRAM4. Manual installation

33 Now and Future Works We are expecting the release of MPICH-G4 Extension to the MPICH-G4 Continuously, we are testing with more larger input size of MM5/WRF

34 Thanks for your attention !