IU TeraGrid Gateway Support Marlon Pierce Community Grids Lab Indiana University.

Slides:



Advertisements
Similar presentations
Common Instrument Middleware Architecture and Federation of Instrument Resources for X-ray Crystallography Rick McMullen Indiana University.
Advertisements

PRAGMA BioSciences Portal Raj Chhabra Susumu Date Junya Seo Yohei Sawai.
Open Grid Computing Environments Marlon Pierce (IU) & Gopi Kandaswamy (RENCI)
Legacy code support for commercial production Grids G.Terstyanszky, T. Kiss, T. Delaitre, S. Winter School of Informatics, University.
Reusable Components for Grid Computing Portals Marlon Pierce Community Grids Lab Indiana University.
VLab: A Collaborative Cyberinfrastructure for Computations of Materials Properties at High Pressures and Temperatures Cesar R. S. da Silva 1 Pedro R. C.
MTA SZTAKI Hungarian Academy of Sciences Grid Computing Course Porto, January Introduction to Grid portals Gergely Sipos
Seminar Grid Computing ‘05 Hui Li Sep 19, Overview Brief Introduction Presentations Projects Remarks.
Indiana University QuakeSim Activities Marlon Pierce, Geoffrey Fox, Xiaoming Gao, Jun Ji, Chao Sun.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
GRID Workload Management System Massimo Sgaravatto INFN Padova.
Grid Services at NERSC Shreyas Cholia Open Software and Programming Group, NERSC NERSC User Group Meeting September 17, 2007.
Workload Management Massimo Sgaravatto INFN Padova.
First steps implementing a High Throughput workload management system Massimo Sgaravatto INFN Padova
Simo Niskala Teemu Pasanen
Evaluation of the Globus GRAM Service Massimo Sgaravatto INFN Padova.
Core Services I & II David Hart Area Director, UFP/CS TeraGrid Quarterly Meeting December 2008.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
Integrating Geographical Information Systems and Grid Applications Marlon Pierce Contributions: Yili Gong,
OSG End User Tools Overview OSG Grid school – March 19, 2009 Marco Mambelli - University of Chicago A brief summary about the system.
Building service testbeds on FIRE D5.2.5 Virtual Cluster on Federated Cloud Demonstration Kit August 2012 Version 1.0 Copyright © 2012 CESGA. All rights.
Future Grid Future Grid User Portal Marlon Pierce Indiana University.
National Center for Supercomputing Applications The Computational Chemistry Grid: Production Cyberinfrastructure for Computational Chemistry PI: John Connolly.
Towards a Javascript CoG Kit Gregor von Laszewski Fugang Wang Marlon Pierce Gerald Guo
Long Term Ecological Research Network Information System LTER Grid Pilot Study LTER Information Manager’s Meeting Montreal, Canada 4-7 August 2005 Mark.
Software for Science Gateways: Open Grid Computing Environments Marlon Pierce, Suresh Marru Pervasive Technology Institute Indiana University
WP9 Resource Management Current status and plans for future Juliusz Pukacki Krzysztof Kurowski Poznan Supercomputing.
GT Components. Globus Toolkit A “toolkit” of services and packages for creating the basic grid computing infrastructure Higher level tools added to this.
Grids and Portals for VLAB Marlon Pierce Community Grids Lab Indiana University.
INFSO-RI Enabling Grids for E-sciencE Logging and Bookkeeping and Job Provenance Services Ludek Matyska (CESNET) on behalf of the.
Grid Resource Allocation and Management (GRAM) Execution management Execution management –Deployment, scheduling and monitoring Community Scheduler Framework.
TRASC Globus Application Launcher VPAC Development Team Sudarshan Ramachandran.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
GRAM5 - A sustainable, scalable, reliable GRAM service Stuart Martin - UC/ANL.
1 Overview of the Application Hosting Environment Stefan Zasada University College London.
INFSO-RI Module 01 ETICS Overview Etics Online Tutorial Marian ŻUREK Baltic Grid II Summer School Vilnius, 2-3 July 2009.
CSF4 Meta-Scheduler Name: Zhaohui Ding, Xiaohui Wei
Evaluation of Agent Teamwork High Performance Distributed Computing Middleware. Solomon Lane Agent Teamwork Research Assistant October 2006 – March 2007.
GEM Portal and SERVOGrid for Earthquake Science PTLIU Laboratory for Community Grids Geoffrey Fox, Marlon Pierce Computer Science, Informatics, Physics.
Resource Brokering in the PROGRESS Project Juliusz Pukacki Grid Resource Management Workshop, October 2003.
FutureGrid Cyberinfrastructure for Computational Research.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
Holding slide prior to starting show. A Portlet Interface for Computational Electromagnetics on the Grid Maria Lin and David Walker Cardiff University.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Grid Security: Authentication Most Grids rely on a Public Key Infrastructure system for issuing credentials. Users are issued long term public and private.
1 Media Grid Initiative By A/Prof. Bu-Sung Lee, Francis Nanyang Technological University.
INFSO-RI JRA 1 Testbed Management Technologies Alain Roy (University of Wisconsin-Madison, USA) ETICS 2 Final Review Brussels - 11 May 2010.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Applications.
OGCE Components for Enhancing UltraScan Job Management. Suresh Marru,Raminder Singh, Marlon Pierce.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
Biomedical and Bioscience Gateway to National Cyberinfrastructure John McGee Renaissance Computing Institute
Building Grid Portals with OGCE: Big Red Portal and GTLAB Mehmet A. Nacar, Jong Youl Choi, Marlon Pierce, Geoffrey Fox Community Grids Lab Indiana University.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
The Gateway Computational Web Portal Marlon Pierce Indiana University March 15, 2002.
Status of Globus activities Massimo Sgaravatto INFN Padova for the INFN Globus group
Application Web Service Toolkit Allow users to quickly add new applications GGF5 Edinburgh Geoffrey Fox, Marlon Pierce, Ozgur Balsoy Indiana University.
Grid Execution Management for Legacy Code Architecture Exposing legacy applications as Grid services: the GEMLCA approach Centre.
Creating Grid Resources for Undergraduate Coursework John N. Huffman Brown University Richard Repasky Indiana University Joseph Rinkovsky Indiana University.
Open Grid Computing Environment Summary
OGCE Short Summary Marlon Pierce Community Grids Lab
OGCE OGCE The Open Grid Computing Environments Collaboratory
DAT381 Team Development with SQL Server 2005
Middleware for Grid Portal Development
OGCE Portal Software for Big Red and the TeraGrid
Open Grid Computing Environments
Gordon Erlebacher Florida State University
Presentation transcript:

IU TeraGrid Gateway Support Marlon Pierce Community Grids Lab Indiana University

Personnel Marlon Pierce: project leader Sangmi Lee Pallickara: senior developer –Lead on VLAB and File Agent Service development Yu “Marie” Ma: senior developer –Lead on CIMA support Rishi Verma: student intern –Software release and testing technician

Team Strategy Provide general purpose gateway support through software delivered through the NSF- funded Open Grid Computing Environments project. Provide short term development support to new TeraGrid gateway projects to help them integrate with resources. –IU’s CIMA instrument project (R. McMullen) –Minnesota’s VLAB project (R. Wentzcovitch) –IU School of Medicine’s Docking and Scoring Portal (S. Meroueh –IU/ECSU Polar Grid project

CIMA Project Overview Common Instrument Middleware Architecture project –NSF funded project to develop software and portals for instruments. –Dr. Rick McMullen PI Flagship project: crystallography portal and services for collecting and archiving real time crystallography data. –Gateway for data collected at 10 crystallography labs in the US, UK, and Australia. Problems: –Much of the collected data is private and should only be accessed by the owners. –Data must be stored on large, highly available file systems. –Services must be highly reliable.

Gateway Team Support for CIMA (Y. Ma) Existing CIMA project was converted into a TeraGrid gateway. CIMA now provides secure access to CIMA archives –Using IU Data Capacitor for storage –Security through GridFTP and a TeraGrid community credential. Marie Ma also led the development work for CIMA High Availability testing. –SC07 demo Future Work: –Support follow-on NSF funded Crystal Grid project. –Use CIMA as a test case to explore virtual hosting and other data grid strategies.

Users get gateway credentials with normal login. Experiments are grouped into samples with private access. Sample data (images, metadata) are securely retrieved from TeraGrid storage.

High Availability CIMA Prototypes fail-over services and portals for TeraGrid gateways. Demonstrated resilient services for multiple scenarios: –Application (Web service) failures. –Operating system failures –Partial and complete network failures –WAN file system failures

Future CIMA/Crystal Grid Support IU is setting up a virtual hosting environment for Gateways and TeraGrid Web services. –Dave Hancock will describe this in an upcoming talk. We are prototyping this for CIMA. –Provide the Gateway perspective. –Dave will provide the integrator perspective/

VLAB Project Overview U-Minn VLAB project is an NSF ITR funded project for investigating properties of planetary materials under extreme conditions. –Prof. Renata Wenztcovitch, PI Very computationally intense (determinng phase diagrams of materials) –Potentially 1000’s of jobs medium-large parallel jobs. VLAB also develops services and portals for managing the complicated runs. Problem: existing VLAB services for task management needed to be integrated with the TeraGrid. –The service also needed to be more easily extensible to many different scheduling/queuing systems.

Gateway Team Support for VLAB (S. Pallickara) Modified VLAB’s Task Executor Web Service to work with TeraGrid GRAM servers. –New Task Executor code built around Condor-G and Condor Birdbath Web Service Java language clients. –Testing with both serial and parallel versions of VLAB workhorse code (“PWSCF”) on TACC’s Lonestar, NCSA’s various metal machines, ORNL’s cluster. This code also formed the basis of support for the NASA-funded QuakeSim project and will be packaged and released for general use. Next step: integrate more complicated of VLAB’s major codes (“Phonon”).

Project Executor Portal Project Interaction Task Dispatcher Task Executor Databases (Metadata, Session registry, etc.) Databases (Metadata, Session registry, etc.) Auxiliary Services (Phonon Input prep, High T post processing, etc.) TeraGrid Task Interface Lonestar(TACC) Dell PowerEdge Linux Cluster 5840 CPUs 62.6 Peak TFlops Lonestar(TACC) Dell PowerEdge Linux Cluster 5840 CPUs 62.6 Peak TFlops Tungsten(NCSA) Dell Xeon IA32 Linux Cluster 2560 CPUs Peak TFlops Tungsten(NCSA) Dell Xeon IA32 Linux Cluster 2560 CPUs Peak TFlops Cobalt(NCSA) SGI Altix 1024 CPUs 6.55 Peak TFlops Cobalt(NCSA) SGI Altix 1024 CPUs 6.55 Peak TFlops NSTG(ORNL) IBM IA Peak TFlops NSTG(ORNL) IBM IA Peak TFlops TeraGrid Information Service TeraGrid Information Service

GRAM JobManager GRAM JobManager LSF Batch system LSF Batch system GRAM JobManager GRAM JobManager PBS Batch system PBS Batch system GRAM JobManager GRAM JobManager PBS Batch system PBS Batch system GRAM JobManager GRAM JobManager LSF Batch system LSF Batch system Task Executor TeraGrid Task Interface Using Condor Birdbath WebService API TeraGrid Task Interface Using Condor Birdbath WebService API Lonestar(TACC) Dell PowerEdge Linux Cluster 5840 CPUs 62.6 Peak TFlops Lonestar(TACC) Dell PowerEdge Linux Cluster 5840 CPUs 62.6 Peak TFlops Tungsten(NCSA) Dell Xeon IA32 Linux Cluster 2560 CPUs Peak TFlops Tungsten(NCSA) Dell Xeon IA32 Linux Cluster 2560 CPUs Peak TFlops Cobalt(NCSA) SGI Altix 1024 CPUs 6.55 Peak TFlops Cobalt(NCSA) SGI Altix 1024 CPUs 6.55 Peak TFlops NSTG(ORNL) IBM IA Peak TFlops NSTG(ORNL) IBM IA Peak TFlops Condor G Job Submission Condor G Job Submission Modified VLAB’s Task Executor Web Service to work with TeraGrid GRAM servers. - New Task Executor code built around Condor-G and Condor Birdbath Web Service Java language clients. - Testing with both serial and parallel versions of VLAB workhorse code (“PWSCF”) on TACC’s Lonestar, NCSA’s various metal machines, ORNL’s cluster.

Lessons from the VLAB Job Submission Example For the VLAB application, multiple input files and multiple output files were required to be transferred between the Teragrid clusters and the TaskExecutor service. Using CondorG provided us a reasonably unified mechanism. However, each of the TeraGrid clusters provides a batch system, which requires different setups for the executables. Some of the system environments were not setup properly –Scripts generated by jobmanager-lsf on Lonestar, for example, override custom $PATH Tackling each of these problems were not trivial, but we did get enthusiastic support from all the TeraGrid sites that we dealt with.

Scoring and Docking Gateway Users develop scoring functions for the ability of drug-like molecules to dock to proteins. Then need Quantum Chemistry techniques to refine technique. –AMBER We are adapting our Condor-G based Web services to build an AMBER Grid Service. Samy Meroueh, IU School of Medicine

General Purpose Gateway Software (S. Pallickara) TeraGrid community credentials are used with GridFTP to access community archives. –Ex: Data Capacitor, HPSS mass storage Problem: We need a way to enforce additional community restrictions on these files. –Users should have restricted file spaces. Solution: express and enforce access restrictions to community files through the Web gateway Software: File Agent Service and updated File Manager portlet developed and released through the OGCE web site. –Targeted for the DC and HPSS

Portlet (modified from OGCE code base) allows file system views of DC, HPSS, and other GridFTP accessible resources. Portlet enforces additional restrictions on community users to keep their data separate and private from other users.

PolarGrid: Microformats, KML, and GeoRSS feeds used to deliver SAR data to multiple clients. PolarGrid: Microformats, KML, and GeoRSS feeds used to deliver SAR data to multiple clients.

Out of Scope Items We do not currently develop, deploy, or maintain general purpose services for TG resource providers. –TG Information Services (J. P. Navarro) and TG User Portal do this. –We do collaborate with these groups through the OGCE project. –This could change if we have clear requirements for this. We rely on existing resource provider infrastructure such as Globus GRAM and GridFTP. –We don’t install or maintain these.

Project Blogs Get a snapshot of what we are working on: –Sangmi: –Marie: –Rishi: –Marlon: