Does the implementation give solutions for the requirements? Flexibility GridRPC enables dynamic join/leave of QM servers. GridRPC enables dynamic expansion.

Slides:



Advertisements
Similar presentations
$100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200.
Advertisements

$100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500.
Demand Response in Ontario Paul Grod, CEO, Rodan Energy July 11, 2013.
$100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300.
Resource WG Breakout. Agenda How we will support/develop data grid testbed and possible applications (1 st day) –Introduction of Gfarm (Osamu) –Introduction.
National Institute of Advanced Industrial Science and Technology Experiences through Grid Challenge Event Yoshio Tanaka.
National Institute of Advanced Industrial Science and Technology Flexible, robust, and efficient multiscale QM/MD simulation using GridRPC and MPI Yoshio.
National Institute of Advanced Industrial Science and Technology Status report on the large-scale long-run simulation on the grid - Hybrid QM/MD simulation.
National Institute of Advanced Industrial Science and Technology Running flexible, robust and scalable grid application: Hybrid QM/MD Simulation Hiroshi.
Resource WG Report. Projects Applications EOL Ninf-G Climate model GridBlast GOC Gangla / SCMSWeb => Uniform Database Goodness Status map (e.g. IVDGL)
A Proposal of Capacity and Performance Assured Storage in The PRAGMA Grid Testbed Yusuke Tanimura 1) Hidetaka Koie 1,2) Tomohiro Kudoh 1) Isao Kojima 1)
National Institute of Advanced Industrial Science and Technology Ninf-G - Core GridRPC Infrastructure Software OGF19 Yoshio Tanaka (AIST) On behalf.
Severs AIST Cluster (50 CPU) Titech Cluster (200 CPU) KISTI Cluster (25 CPU) Climate Simulation on ApGrid/TeraGrid at SC2003 Client (AIST) Ninf-G Severs.
Three types of remote process invocation
MicroKernel Pattern Presented by Sahibzada Sami ud din Kashif Khurshid.
$100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300.
Data Storage Solutions Module 1.2. Data Storage Solutions Upon completion of this module, you will be able to: List the common storage media and solutions.
Describing Complex Products as Configurations using APL Arrays.
Chapter 20 Oracle Secure Backup.
$100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300 $400 $500 $100 $200 $300.
Advanced Industrial Science and Technology 10 Aug / Globus Retreat Ninf-G: Grid RPC system based on the Globus Toolkit Yoshio Tanaka (AIST, Japan)
Distributed Systems Major Design Issues Presented by: Christopher Hector CS8320 – Advanced Operating Systems Spring 2007 – Section 2.6 Presentation Dr.
© 2007 Cisco Systems, Inc. All rights reserved.Cisco Public ITE PC v4.0 Chapter 1 1 Addressing the Network – IPv4 Network Fundamentals – Chapter 6.
Autonomic Systems Justin Moles, Winter 2006 Enabling autonomic behavior in systems software with hot swapping Paper by: J. Appavoo, et al. Presentation.
Developing an Agricultural Monitoring System from Remote Sensing Data Using GridRPC on Ninf-G Shamim Akther, Yann Chemin, Honda Kiyoshi Asian Institute.
Study of Hurricane and Tornado Operating Systems By Shubhanan Bakre.
VxWorks Real-Time Kernel Connectivity
1 Routing and Scheduling in Web Server Clusters. 2 Reference The State of the Art in Locally Distributed Web-server Systems Valeria Cardellini, Emiliano.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Oracle 10g Database Administrator: Implementation and Administration
Scripting Languages For Virtual Worlds. Outline Necessary Features Classes, Prototypes, and Mixins Static vs. Dynamic Typing Concurrency Versioning Distribution.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
Cross Cluster Migration Remote access support Adianto Wibisono supervised by : Dr. Dick van Albada Kamil Iskra, M. Sc.
5205 – IT Service Delivery and Support
PMIT-6102 Advanced Database Systems
STRATEGIES INVOLVED IN REMOTE COMPUTATION
SensIT PI Meeting, January 15-17, Self-Organizing Sensor Networks: Efficient Distributed Mechanisms Alvin S. Lim Computer Science and Software Engineering.
SRUTHI NAGULAVANCHA CIS 764, FALL 2008 Department of Computing and Information Sciences (CIS) Kansas State University -1- Back up & Recovery Strategies.
KISTI’s Activities on the NA4 Biomed Cluster Soonwook Hwang, Sunil Ahn, Jincheol Kim, Namgyu Kim and Sehoon Lee KISTI e-Science Division.
Chapter 8 Implementing Disaster Recovery and High Availability Hands-On Virtual Computing.
WP9 Resource Management Current status and plans for future Juliusz Pukacki Krzysztof Kurowski Poznan Supercomputing.
MCTS Guide to Microsoft Windows Vista Chapter 4 Managing Disks.
Loosely Coupled Parallelism: Clusters. Context We have studied older archictures for loosely coupled parallelism, such as mesh’s, hypercubes etc, which.
High Performance Computing on Virtualized Environments Ganesh Thiagarajan Fall 2014 Instructor: Yuzhe(Richard) Tang Syracuse University.
Component Technology. Challenges Facing the Software Industry Today’s applications are large & complex – time consuming to develop, difficult and costly.
NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.
NIH Resource for Biomolecular Modeling and Bioinformatics Beckman Institute, UIUC NAMD Development Goals L.V. (Sanjay) Kale Professor.
Supporting Molecular Simulation-based Bio/Nano Research on Computational GRIDs Karpjoo Jeong Konkuk Suntae.
Chapter 2 Processes and Threads Introduction 2.2 Processes A Process is the execution of a Program More specifically… – A process is a program.
OPERATING SYSTEMS CS 3530 Summer 2014 Systems with Multi-programming Chapter 4.
GRIDS Center Middleware Overview Sandra Redman Information Technology and Systems Center and Information Technology Research Center National Space Science.
National Institute of Advanced Industrial Science and Technology APGrid PMA: Stauts Yoshio Tanaka Grid Technology Research Center,
Tanenbaum & Van Steen, Distributed Systems: Principles and Paradigms, 2e, (c) 2007 Prentice-Hall, Inc. All rights reserved DISTRIBUTED SYSTEMS.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
7. Grid Computing Systems and Resource Management
National Institute of Advanced Industrial Science and Technology GGF12 Workshop on Operational Security for the Grid Cross-site authentication and access.
Communications & Networks National 4 & 5 Computing Science.
Week 1 Lecture 1 Oracle Architecture Overview. Learning Objectives Learn about Oracle9i architecture and key Oracle9i software components Discover differences.
National Institute of Advanced Industrial Science and Technology Developing Scientific Applications Using Standard Grid Middleware Hiroshi Takemiya Grid.
Università di Perugia Enabling Grids for E-sciencE Status of and requirements for Computational Chemistry NA4 – SA1 Meeting – 6 th April.
DHCP Vrushali sonar. Outline DHCP DHCPv6 Comparison Security issues Summary.
PARALLEL AND DISTRIBUTED PROGRAMMING MODELS U. Jhashuva 1 Asst. Prof Dept. of CSE om.
Seminar On Rain Technology
TeraGrid Capability Discovery John-Paul “JP” Navarro TeraGrid Area Co-Director for Software Integration University of Chicago/Argonne National Laboratory.
Advanced Operating Systems CS6025 Spring 2016 Processes and Threads (Chapter 2)
OPERATING SYSTEMS CS 3502 Fall 2017
Network Load Balancing
Oracle Architecture Overview
Abstractions for Fault Tolerance
Presentation transcript:

Does the implementation give solutions for the requirements? Flexibility GridRPC enables dynamic join/leave of QM servers. GridRPC enables dynamic expansion of a QM server.Robustness GridRPC detects errors and application can implement a recovery code by itself.Efficiency GridRPC can easily handle multiple clusters. Local MPI provides high performance on a cluster by fine grain parallelism.

Strategy for long run QM simulation will migrate to the other cluster either by intentionally or unintentionally. intentional migration Exceeds the maximum runtime for the cluster Reservation period has expired unintentional migration Any error/fault is detected The next cluster will be selected by either reservation or simple selection algorithm. Selection algorithm considers number of available cpus number of requested cpus records of past utilization Simulation reads a host information file in every time step. A cluster can join to/leave from the experiment on-the-fly.

National Institute of Advanced Industrial Science and Technology Experiments - target simulation - - testbed - - results and lessons learned -

Grid-enabled SIMOX Simulation on Japan- US Grid Testbed at SC2005 A technique to fabricate a micro structure consisting of Si surface on the thin SiO 2 insulator Allows to create higher speed with lower power consumption device

SIMOX simulation on the Grid Simulate SIMOX by implanting five oxygen atoms with their initial velocities much smaller than the usual values. The incident positions of the oxygen atoms relative to the surface crystalline structure of Si differ. 5 QM regions are initially defined Size and No. of QM regions are changed during the simulation 0.11million atoms in total Results of the experiments will demonstrate the sensitivity of the process on the incident position of the oxygen atom when its implantation velocity is small.

Testbed for the experiment Phase 1 Phase 2 Phase 3Phase 4 AIST Super Clusters P32 (2144 CPUs), M64 (528 CPUs), F32 (536 CPUs) TeraGrid Clusters PSC clusters (3000 CPUs), NCSA clusters(1774 CPUs) USC Clusters USC (7280 CPUs) Japan Clusters U-Tokyo (386 CPUs), TITECH (512 CPUs) QM1P32 USC ISTB S QM2P32 NCS A USC Prest o QM3M64 QM4P32 TCS USC P32 QM5P32 TCS USC P32 Reserv e F32 P32 F32

Result of the experiment Phase 1Phase 2Phase 3Phase 4 Experiment Time: days Simulation steps: 270 (~ 54 fs) Longest Calculation Time: 4.76 day

Results of the experiment (cont d) Behavior of oxygen atoms strongly depends on the incident position QM 1 QM 2 QM 3 QM 4 QM 5 v/v 0 Time step Expanding/Dividing QM regions at every 5 time steps Expansion: 47 times, Division: 8 times) Time Step No. of CPUs No. of QM Atoms 270 No. of CPUs/Atoms Succeeded in long-run by intentional/unintentional resource migration Intentional migration Migration triggered by faults

Summary of the experimental results We could verify that our strategy for long run is practical approach Continue the simulation by migrating one cluster to the other one based on reservation We could verify the programming using GridRPC and MPI could implement real Grid- enabled application Dynamic resource allocation / migration Recover from faults Manage hundreds of CPUs on distributed sites

Status and Future Plans Ninf-G Version 5 will be coming! What are differences with Ninf-G4? Lower prerequisites for installation Ninf-G4 needs Globus Library since it uses Globus IO for client/server communications. Ninf-G5 can be installed without Globus. i.e., Ninf-G5 can be installed according to the underlying software environments Three major components (remote process invocation, information retrieval, and client/server communication) can be pluggable. e.g. without Globus, without TCP Work efficiently from a single supercomputer to Grid Other new features will be supported Connection less (client server) Client-side check pointing Ninf-G 5.0.0alpha will be available in this March.

For more info, related links Ninf project ML Ninf-G Users ML (subscribed member s only) Ninf project home page Open Grid Forum GGF GridRPC WG