National Institute of Advanced Industrial Science and Technology Experiences through Grid Challenge Event Yoshio Tanaka.

Slides:



Advertisements
Similar presentations
PRAGMA – TeraGrid – AIST Interoperation Testing Philip Papadopoulos.
Advertisements

Kento Aida, Tokyo Institute of Technology Grid Working Group Meeting Aug. 27 th, 2003 Tokyo Institute of Technology Kento Aida.
Resource WG Breakout. Agenda How we will support/develop data grid testbed and possible applications (1 st day) –Introduction of Gfarm (Osamu) –Introduction.
CSF4 Meta-Scheduler Tutorial 1st PRAGMA Institute Zhaohui Ding or
Resources WG Update PRAGMA 9 Hyderabad. Status (in 1 slide) Applications QMMD (AIST) Savannah (MU) iGAP (SDSC, AIST) Middleware Gfarm (AIST) Community.
Reports from Resource Breakout PRAGMA 16 KISTI, Korea.
Resource WG Update PRAGMA 14 Mason Katz, Yoshio Tanaka, Cindy Zheng.
Resource WG Update PRAGMA 8 Singapore. Routine Use - Users make a system work.
National Institute of Advanced Industrial Science and Technology Status report on the large-scale long-run simulation on the grid - Hybrid QM/MD simulation.
Resource WG Update PRAGMA 14 Mason Katz, Yoshio Tanaka, Cindy Zheng.
Resource WG PRAGMA Mason Katz, Yoshio Tanaka, Cindy Zheng.
Cindy Zheng, PRAGMA 8, Singapore, 5/3-4/2005 Status of PRAGMA Grid Testbed & Routine-basis Experiments Cindy Zheng Pacific Rim Application and Grid Middleware.
Experiences of Grid Enabled MPI Implementation named MPICH-GX with Atmospheric Applications Oh-kyoung Kwon, KISTI Salvador Castañeda, CICESE PRAGMA 11.
National Institute of Advanced Industrial Science and Technology Meta-scheduler based on advanced reservation Grid Technology Research Center, AIST Atsuko.
Resource/data WG Summary Yoshio Tanaka Mason Katz.
National Institute of Advanced Industrial Science and Technology Running flexible, robust and scalable grid application: Hybrid QM/MD Simulation Hiroshi.
Resource WG Summary Mason Katz, Yoshio Tanaka. Next generation resources on PRAGMA Status – Next generation resource (VM-based) in PRAGMA by UCSD (proof.
Resource WG Report. Projects Applications EOL Ninf-G Climate model GridBlast GOC Gangla / SCMSWeb => Uniform Database Goodness Status map (e.g. IVDGL)
Introduction of Grid Security
National Institute of Advanced Industrial Science and Technology Advance Reservation-based Grid Co-allocation System Atsuko Takefusa, Hidemoto Nakada,
A Proposal of Capacity and Performance Assured Storage in The PRAGMA Grid Testbed Yusuke Tanimura 1) Hidetaka Koie 1,2) Tomohiro Kudoh 1) Isao Kojima 1)
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks MyProxy and EGEE Ludek Matyska and Daniel.
TeraGrid Deployment Test of Grid Software JP Navarro TeraGrid Software Integration University of Chicago OGF 21 October 19, 2007.
National Institute of Advanced Industrial Science and Technology Ninf-G - Core GridRPC Infrastructure Software OGF19 Yoshio Tanaka (AIST) On behalf.
Does the implementation give solutions for the requirements? Flexibility GridRPC enables dynamic join/leave of QM servers. GridRPC enables dynamic expansion.
Severs AIST Cluster (50 CPU) Titech Cluster (200 CPU) KISTI Cluster (25 CPU) Climate Simulation on ApGrid/TeraGrid at SC2003 Client (AIST) Ninf-G Severs.
Three types of remote process invocation
Andrew McNab - Manchester HEP - 17 September 2002 Putting Existing Farms on the Testbed Manchester DZero/Atlas and BaBar farms are available via the Testbed.
GXP in nutshell You can send jobs (Unix shell command line) to many machines, very fast Very small prerequisites –Each node has python (ver or later)
EXAMPLE 4 Solve a multi-step problem SHOPPING
© 2007 IBM Corporation IBM Global Engineering Solutions IBM Blue Gene/P Job Submission.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Tutorial Getting started with GILDA.
PRAGMA19, Sep. 15 Resources breakout Migration from Globus-based Grid to Cloud Mason Katz, Yoshio Tanaka.
National Institute of Advanced Industrial Science and Technology Auditing, auditing template and experiences on being audited Yoshio Tanaka
Minimum intrusion GRID. Build one to throw away … So, in a given time frame, plan to achieve something worthwhile in half the time, throw it away, then.
Minimum intrusion GRID. Build one to throw away … So, in a given time frame, plan to achieve something worthwhile in half the time, throw it away, then.
AustrianGrid, LCG & more Reinhard Bischof HPC-Seminar April 8 th 2005.
UNICORE UNiform Interface to COmputing REsources Olga Alexandrova, TITE 3 Daniela Grudinschi, TITE 3.
Evaluation of the Globus GRAM Service Massimo Sgaravatto INFN Padova.
December 8 & 9, 2005, Austin, TX SURA Cyberinfrastructure Workshop Series: Grid Technology: The Rough Guide Configuring Resources for the Grid Jerry Perez.
GRACE Project IST EGAAP meeting – Den Haag, 25/11/2004 Giuseppe Sisto – Telecom Italia Lab.
The Glidein Service Gideon Juve What are glideins? A technique for creating temporary, user- controlled Condor pools using resources from.
Chapter 8 Implementing Disaster Recovery and High Availability Hands-On Virtual Computing.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Kento Aida, Tokyo Institute of Technology Grid Challenge - programming competition on the Grid - Kento Aida Tokyo Institute of Technology 22nd APAN Meeting.
Matthew Palmer, Cambridge University01/10/2015 First Use of the UK e-Science Grid Overview The Physics Experiences Looking forward Conclusions Matthew.
Grid Computing - AAU 14/ Grid Computing Josva Kleist Danish Center for Grid Computing
PRAGMA: Cyberinfrastructure, Applications, People Yoshio Tanaka (AIST, Japan) Peter Arzberger (UCSD, USA)
National Institute of Advanced Industrial Science and Technology Introduction of PRAGMA routine-basis experiments Yoshio Tanaka
Scalable Systems Software Center Resource Management and Accounting Working Group Face-to-Face Meeting October 10-11, 2002.
N*Grid – Korean Grid Research Initiative Funded by Government (Ministry of Information and Communication) 5 Years from 2002 to million US$ Including.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
Grid Operations Centre LCG Accounting Trevor Daniels, John Gordon GDB 8 Mar 2004.
Institute For Digital Research and Education Implementation of the UCLA Grid Using the Globus Toolkit Grid Center’s 2005 Community Workshop University.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
EGEE-II INFSO-RI Enabling Grids for E-sciencE The GILDA training infrastructure.
GRID Zhen Xie, INFN-Pisa, on DataGrid WP6 meeting1 Globus Installation Toolkit Zhen Xie On behalf of grid-release team INFN-Pisa.
National Institute of Advanced Industrial Science and Technology APGrid PMA: Stauts Yoshio Tanaka Grid Technology Research Center,
Chapter 11 Working with Credit Card Methods of Processing Credit Cards Preparing for Cyber Cash Authoring a Credit card Transaction.
National Institute of Advanced Industrial Science and Technology ApGrid: Asia Pacific Partnership for Grid Computing - Introduction of testbed development.
Portal Update Plan Ashok Adiga (512)
| nectar.org.au NECTAR TRAINING Module 5 The Research Cloud Lifecycle.
National Institute of Advanced Industrial Science and Technology GGF12 Workshop on Operational Security for the Grid Cross-site authentication and access.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
INRNE's participation in LCG Elena Puncheva Preslav Konstantinov IT Department.
Microsoft BackOffice Applications
LQCD Computing Operations
1Z0-477 Exam Dumps
Presentation transcript:

National Institute of Advanced Industrial Science and Technology Experiences through Grid Challenge Event Yoshio Tanaka

Grid Challenge A competition for programming on a Grid Main objectives For participants: To provide opportunities to use real Grid (for participants) For us: To understand obstacles/problems to make a Grid production level (1000cpus are shared by many users) To have an opportunity to encourage participants to use our software (e.g. Ninf-G, GXP) 30 students/graduates were participated in this event Provide 960cpus testbed for participants Schedule Preliminary: Feb. 1 ~ Feb. 28 Final round: Mar. 5 ~ Mar. 20

Grid Challenge Two categories Regular routine A problem is provided Graphic image analysis count the number of objects Ranked by the performance, i.e. which is the fastest program? Free routine Can do anything interesting Could have experiences on running his own software on real Grid

Software Software provided by the organizer ssh GXP GT2 & batch & jobmanager MPICH (p4) Ninf-G2 Other software can be installed by participants

Contributed resources Sites#nodes/#cpus IP addresses Administrated by TITECH / Matsuoka 100/200PublicTanaka-san TITECH / Aida30/60Private Prof. Aida + students Tokushima U. 50/100Private Prof. Ono + students U. Tsukua 20/40Publicstudents UEC50/100Private AIST Support U. Tokyo 40/40Public AIST Support U. Tokyo 63/126Private AIST Support U. Tokyo 107/214Public AIST Support AIST40/80public Total500/960

Preparation ( ~ Feb. 1) Administrators installed software in every site Participants sent ssh public key Administrators created accounts for all participants Participants tested each cluster login compile test run Participants obtained Globus certificates from AIST GTRC CA (if necessary) Participants sent Subject DN and administrators added their entries to grid-mapfile

Preparation (~ Feb. 1) (cont d) AIST provided A document for obtaining Globus certificate Test script for Globus A how-to document and sample programs of Ninf-G2 How to develop Ninf-G apps step-by-step Obtain certificate Test globus Develop and run Ninf-G apps client configuration file for the Grid challenge environment

Problems 30 participants shared 960 cpus for one month Some used ssh for process invocation Some used GXP for process invocation Some used Ninf-G2 for process invocation Need to take care (many) trouble shooting Some nodes went down pbs daemon died students usually made experiments in midnight Interactive use of backend nodes (via ssh/GXP) was allowed F32 prohibits interactive use AIST could not provide F32

Problems (cont d) Participants expected that all processes would be launched immediately (co-allocation) ssh/GXP enables it Ninf-G2 could not expect In order to keep fairness, we decided to change the configuration of batch queuing system For each processor, set the max number of processes per user to 1 Increased the max number of processes per processor to the number of participants (30) This is an unusual configuration!!

Insights valuable for PRAGMA Mixture of batch and interactive use introduce a problem batch is expected to provide dedicated environment load balancing Interactive use (via ssh) may disturb batch But some middleware/apps require interactive use co-allocation / grid-level scheduler is hard to solve (basically) Applications should not expect all resources are available Application developers need extra work for this feature Possible solutions Make application capable for using only available resources in as-is strategy Implement co-allocation based on reservation No grid-level reservation system yet Should be done manually Do we have the same problem in PRAGMA routine-basis experiments?