H IGH E NERGY A CCELERATOR R ESEARCH O RGANIZATION KEKKEK High Availability iRODS System (HAIRS) Yutaka Kawai, KEK Adil Hasan, ULiv December 2nd, 20091Interoperability.

Slides:



Advertisements
Similar presentations
Peter Berrisford RAL – Data Management Group SRB Services.
Advertisements

Welcome to Middleware Joseph Amrithraj
Data Management Expert Panel - WP2. WP2 Overview.
The SMART Way to Migrate Replicated Stateful Services Jacob R. Lorch, Atul Adya, Bill Bolosky, Ronnie Chaiken, John Douceur, Jon Howell Microsoft Research.
KX-NS1000 Initial Set Up For step by step : 16 May,
Copyright 2009 FUJITSU TECHNOLOGY SOLUTIONS PRIMERGY Servers and Windows Server® 2008 R2 Benefit from an efficient, high performance and flexible platform.
Copyright © 2002 Wensong Zhang. Page 1 Free Software Symposium 2002 Linux Virtual Server: Linux Server Clusters for Scalable Network Services Wensong Zhang.
Microsoft Load Balancing and Clustering. Outline Introduction Load balancing Clustering.
Hands-On Microsoft Windows Server 2008 Chapter 8 Managing Windows Server 2008 Network Services.
Presented by : Ran Koretzki. Basic Introduction What are VM’s ? What is migration ? What is Live migration ?
10/02/2004ELFms meeting1 Linux Virtual Server Miroslav Siket FIO-FS.
Experience of a low-maintenance distributed data management system W.Takase 1, Y.Matsumoto 1, A.Hasan 2, F.Di Lodovico 3, Y.Watase 1, T.Sasaki 1 1. High.
1 The Google File System Reporter: You-Wei Zhang.
An easy way to manage Relational Databases in the Globus Community Sandro Fiore ISUFI/ Center for Advanced Computational Technologies Director: prof. Giovanni.
Module 13: Network Load Balancing Fundamentals. Server Availability and Scalability Overview Windows Network Load Balancing Configuring Windows Network.
IRODS performance test and SRB system at KEK Yoshimi KEK Building data grids with iRODS 27 May 2008.
INSTALLING MICROSOFT EXCHANGE SERVER 2003 CLUSTERS AND FRONT-END AND BACK ‑ END SERVERS Chapter 4.
NAREGI WP4 (Data Grid Environment) Hideo Matsuda Osaka University.
High-Availability Linux.  Reliability  Availability  Serviceability.
The Data Grid: Towards an Architecture for the Distributed Management and Analysis of Large Scientific Dataset Caitlin Minteer & Kelly Clynes.
EGEE-II INFSO-RI Enabling Grids for E-sciencE Data Grid Services/SRB/SRM & Practical Hai-Ning Wu Academia Sinica Grid Computing.
Production Data Grids SRB - iRODS Storage Resource Broker Reagan W. Moore
Storage Tank in Data Grid Shin, SangYong(syshin, #6468) IBM Grid Computing August 23, 2003.
INFSO-RI Enabling Grids for E-sciencE Distributed Metadata with the AMGA Metadata Catalog Nuno Santos, Birger Koblitz 20 June 2006.
Frontiers in Massive Data Analysis Chapter 3.  Difficult to include data from multiple sources  Each organization develops a unique way of representing.
Intro – Part 2 Introduction to Database Management: Ch 1 & 2.
OGSA-DAI in OMII-Europe Neil Chue Hong EPCC, University of Edinburgh.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
INFSO-RI Enabling Grids for E-sciencE AMGA Metadata Server - Metadata Services in gLite (+ ARDA DB Deployment Plans with Experiments)
Enabling Grids for E-sciencE EGEE-III INFSO-RI I. AMGA Overview What is AMGA Metadata Catalogue of EGEE’s gLite 3.1 Middleware Main Feature of.
Remote Controller & Presenter Make education more efficiently
MapReduce and GFS. Introduction r To understand Google’s file system let us look at the sort of processing that needs to be done r We will look at MapReduce.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks AMGA PHP API Claudio Cherubino INFN - Catania.
Cloning Windows NT Systems Mainly based on experiences at RAL and Oxford.
Replica Management Services in the European DataGrid Project Work Package 2 European DataGrid.
GFS. Google r Servers are a mix of commodity machines and machines specifically designed for Google m Not necessarily the fastest m Purchases are based.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Implementation and performance analysis of.
© 2008 Open Grid Forum File Catalog Development in Japan e-Science Project GFS-WG, OGF24 Singapore Hideo Matsuda Osaka University.
Windows Azure. Azure Application platform for the public cloud. Windows Azure is an operating system You can: – build a web application that runs.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
Managing Data DIRAC Project. Outline  Data management components  Storage Elements  File Catalogs  DIRAC conventions for user data  Data operation.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
Linux Operations and Administration
Grid Technology CERN IT Department CH-1211 Geneva 23 Switzerland t DBCF GT Upcoming Features and Roadmap Ricardo Rocha ( on behalf of the.
AliEn central services Costin Grigoras. Hardware overview  27 machines  Mix of SLC4, SLC5, Ubuntu 8.04, 8.10, 9.04  100 cores  20 KVA UPSs  2 * 1Gbps.
AFS/OSD Project R.Belloni, L.Giammarino, A.Maslennikov, G.Palumbo, H.Reuter, R.Toebbicke.
AMGA-Bookkeeping Carmine Cioffi Department of Physics, Oxford University UK Metadata Workshop Oxford, 05 July 2006.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Mario Reale – GARR NetJobs: Network Monitoring Using Grid Jobs.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
H IGH E NERGY A CCELERATOR R ESEARCH O RGANIZATION KEKKEK Learning System to Classify Batch Jobs Yutaka Kawai (*1), Adil Hasan (*2), Wataru Takase (*1),
VIRTUAL SERVERS Chapter 7. 2 OVERVIEW Exchange Server 2003 virtual servers Virtual servers in a clustering environment Creating additional virtual servers.
An Introduction to GPFS
Presented by Deepak Varghese Reg No: Introduction Application S/W for server load balancing Many client requests make server congestion Distribute.
Domain Name System The Technology Context Presentation.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
1 © 2004, Cisco Systems, Inc. All rights reserved. CCNA 2 v3.1 Module 2 Introduction to Routers.
Sanjay Ghemawat, Howard Gobioff, Shun-Tak Leung
LCG Storage Management Workshop, CERN, 7th April 2005
BDII Performance Tests
Introduction to Data Management in EGI
CERN-Russia Collaboration in CASTOR Development
NAREGI at KEK and GRID plans
Introduction to Networks
Storage Virtualization
DUCKS – Distributed User-mode Chirp-Knowledgeable Server
Interoperability of Digital Repositories
First Level Incident Handling FAQ (For EAL)
The SMART Way to Migrate Replicated Stateful Services
Presentation transcript:

H IGH E NERGY A CCELERATOR R ESEARCH O RGANIZATION KEKKEK High Availability iRODS System (HAIRS) Yutaka Kawai, KEK Adil Hasan, ULiv December 2nd, 20091Interoperability of Digital London, UK -- Yutaka Kawai, KEK

Outline ▸Introduction ▸iRODS HA system with Director ▸Large File Transfer ▸Speed Performance ▸Future works (apply to RNS application) ▸Summary December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK2

Introduction ▸Replication enables high availability (HA) system for catalog service ▹Replicate by back-end, i.e. iRODS ▹Replicate by front-end; ▪i.e. AMGA (ARDA [1] Metadata Grid Application) ▫Metadata Catalogue of EGEE’s gLite 3.1 Middleware ▫Back-end : Oracle, PostgreSQL, MySQL, SQLite ▫ ▸The current iRODS HA is implemented by replicating ICAT DB with PgPool tool [2] ▹A problem when iRODS server fails ▹Solve the problem by using Director December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK3

The Current iRODS HA ▸ICAT DB replication by Pgpool December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK4 ICAT Pgpool iRODS ServerPostgreSQL iRODS Client Change the server info in.irodEnv A B

Problem of the current HA ▸Even if the iRODS server fails, clients still continue to access the same server without noticing the failure. December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK5 ICAT Pgpool iRODS Server PostgreSQL iRODS Client ? Need to change server info in.irodEnv A B

Solution by using Director ▸Place a Director between Client and Server ▹Monitor the iRODS server statuses ▹Load balance to the iRODS servers December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK6 ICAT Pgpool iRODS Server PostgreSQL iRODS Client Director A B

How to Implement Director? ▸UltraMonkey [3] ▹Linux based director ▹Low cost but not so high speed ▹Need some steps to setup ▸Hardware Director ▹High cost and high speed ▹Easy to setup (?) ▹Cisco, HP, etc. December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK7

UltraMonkey ▸UltraMonkey consists of 3 components ▹Linux Virtual Server (LVS) : Load balancing ▹ldirectord : Monitoring real servers ▹Linux-HA (LHA) : Monitoring directors ▸LVS and ldirectord are used here ▹LVS : Provide Virtual IP for load balance ▹ldirectord : Monitoring iRODS service ▹LHA : Future use for director redundancy December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK8

Virtual IP for load balance December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK9 iRODS Real Severs / Linux Director /24 VIP iRODS Client.100 iRODS Client can specify only this VIP in.irodsEnv Gateway of Real Servers is Director

Monitoring iRODS service ▸ldirector monitors iRODS real servers ▹Polling server status via iRODS control port December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK10 RODS_VERSION rods2.1 d 0 0 Director iRODS Server Port# 1247 req: any string ack: iRODS MsgHeader

Outline ▸Introduction ▸iRODS HA system with Director ▸Large File Transfer ▸Speed Performance ▸Future works (apply to RNS application) ▸Summary December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK11

Large File Transfer ▸iRODS uses parallel ports to transfer a large file. ▹Smaller than 32MB file is transferred through iRODS control port #1247. ▸iRODS catalog server directs a server to open parallel ports to transfer a large file ▹iRODS clients can directly connect with the server through the parallel ports. December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK12

Process of Large File Transfer ▸Steps to transfer a large file in iRODS December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK13 ICAT iRODS ClientiRODS Server w/o ICAT Physical Data Physical Data Start service for Parallel I/O (4) PostgreSQL (1) (3) iput a large file Find physical location to store (2) A C File Transfer via Parallel I/O

Large File Transfer w/ Director ▸Need to confirm whether Director interferes in transferring a large file or not ▸The physical storage should be located out of the local network of iRODS real servers ▹Director handles only iRODS catalog server IP ▹Director cannot manage all of the parallel ports December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK14

Process using Director ▸Works as same as normal case ▹Only one additional step between (1) and (2) December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK15 ICAT Pgpool iRODS Server PostgreSQL iRODS Client Director iRODS Server w/o ICAT Physical Data Physical Data (1) (3) (4) A B C (1)’ (2)

Outline ▸Introduction ▸iRODS HA system with Director ▸Large File Transfer ▸Speed Performance ▸Future works (apply to RNS application) ▸Summary December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK16

Speed Performance ▸Test Program ▹concurrent-test in iRODS package ▹iput, imeta, iget, imv ▹1000 entries ▹Servers are VMs (Xen) on same physical machine ▪Client is located on the different machine ▸No Director ▹552.2sec = sec/entry ▸Use Director ▹618.4 sec = sec/entry ▹About 10% slower December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK17 This result is reasonable to consider tradeoff between speed and availability

Speed Performance (cont’d) ▸Use Director and Load balance to 2 iRODS severs ▹697.8sec= sec/entry ▹The concurrent-test is not suitable under such a Load balanced system. ▹Need a program using multi- clients/threading. December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK18

Outline ▸Introduction ▸iRODS HA system with Director ▸Large File Transfer ▸Speed Performance ▸Future works (apply to RNS application) ▸Summary December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK19

What is RNS ? ▸RNS : Resource Namespace Service ▹RNS offers a simple standard way of mapping names to endpoints within a grid or distributed network [4] ▹The latest version is available here; ▸Java based RNS application is being developed by Osaka University and Tsukuba University ▹This application is similar to iRODS ▹The other kind of RNS application is Grid Shell of Genesis II by The Virginia Center for Grid Research (VCGR) [5]. December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK20

Apply to RNS application?? ▸Derby can do replication? ▹ ▹No load-sharing in the above example December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK21 DB Replication RNS ServerDerby RNS Client Director ?

Issues in RNS application ▸Several issues to be solved ▹Derby is not enough to work replication as same as using PostgreSQL w/Pgpool ▹Need some developments to replace Derby by PostgreSQL ▹The catalog implementation in the current RNS application has specific IP addresses December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK22

Opinions in this study ▸Network limitation ▹Director works as NAT. Difficult to place iRODS catalog servers in different subnets. ▹But the problem depends on NAT technology. We hope some NAT vender can implement extensions. ▸Speed Performance ▹The “concurrent-test” consumes overhead. The result 10% slow is in one of the worst cases. We may see less than 10% in actual uses. ▸PostgreSQL only? ▹How about other DB services? They have the same tools as PgPool? ▹Back-end replication is enough? Front-end replication should be considered for iRODS? December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK23

Summary ▸iRODS HA system ▹The current approach using only PgPool ▹The new approach using Director ▹The new one can solve the current problem ▸Large File Transfer ▹iRODS large file transfer works well when using Director ▸Speed Performance ▹Director results in the speed performance of concurrent- test getting slower 10% ▸Future works ▹Apply this solution to other catalog services December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK24

References ▸[1] : ARDA is A Realization of Distributed Analysis for LHC, ▸[2] : iRODS High Avaliability, ▸[3] : Ultra Monkey project, ▸[4] : citation from abstract of “Resource Namespace Service Specification”, ▸[5] : nding_Your_Genesis_II_Distribution#RNS_Namespace nding_Your_Genesis_II_Distribution#RNS_Namespace December 2nd, 2009Interoperability of Digital London, UK -- Yutaka Kawai, KEK25