TEXAS ADVANCED COMPUTING CENTER Deployment of NMI Components on the UT Grid Shyamal Mitra.

Slides:



Advertisements
Similar presentations
CSF4 Meta-Scheduler Tutorial 1st PRAGMA Institute Zhaohui Ding or
Advertisements

Test harness and reporting framework Shava Smallen San Diego Supercomputer Center Grid Performance Workshop 6/22/05.
FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
TEXAS ADVANCED COMPUTING CENTER Grids: TACC Case Study Ashok Adiga, Ph.D. Distributed & Grid Computing Group Texas Advanced Computing Center The University.
CSF4, SGE and Gfarm Integration Zhaohui Ding Jilin University.
A Computation Management Agent for Multi-Institutional Grids
Dr. David Wallom Use of Condor in our Campus Grid and the University September 2004.
USING THE GLOBUS TOOLKIT This summary by: Asad Samar / CALTECH/CMS Ben Segal / CERN-IT FULL INFO AT:
Collaborative Campus Grid - Practices and experiences in Leiden University Campus Grid (LUCGrid) Hui Li Feb 4, 2005.
Slides for Grid Computing: Techniques and Applications by Barry Wilkinson, Chapman & Hall/CRC press, © Chapter 1, pp For educational use only.
The Cactus Portal A Case Study in Grid Portal Development Michael Paul Russell Dept of Computer Science The University of Chicago
1-2.1 Grid computing infrastructure software Brief introduction to Globus © 2010 B. Wilkinson/Clayton Ferner. Spring 2010 Grid computing course. Modification.
Grid Services at NERSC Shreyas Cholia Open Software and Programming Group, NERSC NERSC User Group Meeting September 17, 2007.
Grids and Globus at BNL Presented by John Scott Leita.
Status of Globus activities within INFN (update) Massimo Sgaravatto INFN Padova for the INFN Globus group
Condor Overview Bill Hoagland. Condor Workload management system for compute-intensive jobs Harnesses collection of dedicated or non-dedicated hardware.
Simo Niskala Teemu Pasanen
Globus Computing Infrustructure Software Globus Toolkit 11-2.
Grid Toolkits Globus, Condor, BOINC, Xgrid Young Suk Moon.
Cluster Computing through an Application-oriented Computational Chemistry Grid Kent Milfeld and Chona Guiang, Sudhakar Pamidighantam, Jim Giuliani Supported.
OSG End User Tools Overview OSG Grid school – March 19, 2009 Marco Mambelli - University of Chicago A brief summary about the system.
Michigan Grid Testbed Report Shawn McKee University of Michigan UTA US ATLAS Testbed Meeting April 4, 2002.
UT Grid Project Jay Boisseau, Texas Advanced Computing Center SURA Grid Application Planning & Implementations Workshop December 7, 2005.
University of Virginia Experiences with NMI at the University of Virginia NMI Integration Testbed: Experiences in Middleware Deployment Spring 2003 Internet2.
National Center for Supercomputing Applications The Computational Chemistry Grid: Production Cyberinfrastructure for Computational Chemistry PI: John Connolly.
INFN-GRID Globus evaluation (WP 1) Massimo Sgaravatto INFN Padova for the INFN Globus group
High Performance Louisiana State University - LONI HPC Enablement Workshop – LaTech University,
Slide 1 Experiences with NMI R2 Grids Software at Michigan Shawn McKee April 8, 2003 Internet2 Spring Meeting.
Grid Computing - AAU 14/ Grid Computing Josva Kleist Danish Center for Grid Computing
Grid Computing, B. Wilkinson, b.1 National Science Foundation Middleware Initiative (NMI) Started in 2001 initially over 3 years “to create and deploy.
Grids and Portals for VLAB Marlon Pierce Community Grids Lab Indiana University.
Computational grids and grids projects DSS,
CHEP 2003Stefan Stonjek1 Physics with SAM-Grid Stefan Stonjek University of Oxford CHEP th March 2003 San Diego.
GRAM5 - A sustainable, scalable, reliable GRAM service Stuart Martin - UC/ANL.
CSF4 Meta-Scheduler Name: Zhaohui Ding, Xiaohui Wei
National Computational Science National Center for Supercomputing Applications National Computational Science NCSA-IPG Collaboration Projects Overview.
Evaluation of Agent Teamwork High Performance Distributed Computing Middleware. Solomon Lane Agent Teamwork Research Assistant October 2006 – March 2007.
Laboratório de Instrumentação e Física Experimental de Partículas GRID Activities at LIP Jorge Gomes - (LIP Computer Centre)
Experiences with the Globus Toolkit on AIX and deploying the Large Scale Air Pollution Model as a grid service Ashish Thandavan Advanced Computing and.
Resource Brokering in the PROGRESS Project Juliusz Pukacki Grid Resource Management Workshop, October 2003.
December 8 & 9, 2005, Austin, TX SURA Cyberinfrastructure Workshop Series: Grid Technology: The Rough Guide User Interfaces to Grids Patrick Hurley Texas.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
Kurt Mueller San Diego Supercomputer Center NPACI HotPage Updates.
1 Grid Portal for VN-Grid Cu Nguyen Phuong Ha. 2 Outline Some words about portals in principle Overview of OGCE GridPortlets.
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
Institute For Digital Research and Education Implementation of the UCLA Grid Using the Globus Toolkit Grid Center’s 2005 Community Workshop University.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
NATIONAL PARTNERSHIP FOR ADVANCED COMPUTATIONAL INFRASTRUCTURE Capability Computing - User Environment Anke Kamrath Associate Director, SDSC
Grid Security: Authentication Most Grids rely on a Public Key Infrastructure system for issuing credentials. Users are issued long term public and private.
July 11-15, 2005Lecture3: Grid Job Management1 Grid Compute Resources and Job Management.
Part Five: Globus Job Management A: GRAM B: Globus Job Commands C: Laboratory: globusrun.
GRIDS Center Middleware Overview Sandra Redman Information Technology and Systems Center and Information Technology Research Center National Space Science.
Condor Project Computer Sciences Department University of Wisconsin-Madison Grids and Condor Barcelona,
The DZero/PPDG Test Bed Test bed composition as of Feb 2002: 3 PC at Fermilab (sammy, samadams, sameggs) Contact: Gabriele Garzoglio 1 PC at Imperial College.
Introduction to Grids By: Fetahi Z. Wuhib [CSD2004-Team19]
Portal Update Plan Ashok Adiga (512)
Alain Roy Computer Sciences Department University of Wisconsin-Madison Condor & Middleware: NMI & VDT.
National Computational Science National Center for Supercomputing Applications National Computational Science Integration of the MyProxy Online Credential.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
LSF Universus By Robert Stober Systems Engineer Platform Computing, Inc.
Miron Livny Computer Sciences Department University of Wisconsin-Madison Condor and (the) Grid (one of.
Status of Globus activities Massimo Sgaravatto INFN Padova for the INFN Globus group
MGRID Architecture Andy Adamson Center for Information Technology Integration University of Michigan, USA.
A System for Monitoring and Management of Computational Grids Warren Smith Computer Sciences Corporation NASA Ames Research Center.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
CNAF - 24 September 2004 EGEE SA-1 SPACI Activity Italo Epicoco.
GPIR GridPort Information Repository
From Prototype to Production Grid
Presentation transcript:

TEXAS ADVANCED COMPUTING CENTER Deployment of NMI Components on the UT Grid Shyamal Mitra

2 Outline TACC Grid ProgramTACC Grid Program NMI Testbed ActivitiesNMI Testbed Activities Synergistic ActivitiesSynergistic Activities –Computations on the Grid –Grid Portals

3 TACC Grid Program Building GridsBuilding Grids –UT Campus Grid –State Grid (TIGRE) Grid ResourcesGrid Resources –NMI Components –United Devices –LSF Multicluster Significantly leveraging NMI Components and experienceSignificantly leveraging NMI Components and experience

4 Resources at TACC IBM Power 4 System (224 processors, 512 GB Memory, 1.16 TF)IBM Power 4 System (224 processors, 512 GB Memory, 1.16 TF) IBM IA-64 Cluster (40 processors, 80 GB Memory, 128 GF)IBM IA-64 Cluster (40 processors, 80 GB Memory, 128 GF) IBM IA-32 Cluster (64 processors, 32 GB Memory, 64 GF)IBM IA-32 Cluster (64 processors, 32 GB Memory, 64 GF) Cray SV1 (16 processors, 16 GB Memory, 19.2 GF)Cray SV1 (16 processors, 16 GB Memory, 19.2 GF) SGI Origin 2000 (4 processors, 2 GB Memory, 1 TB storage)SGI Origin 2000 (4 processors, 2 GB Memory, 1 TB storage) SGI Onyx 2 (24 processors, 25 GB Memory, 6 Infinite Reality-2 Graphics pipes)SGI Onyx 2 (24 processors, 25 GB Memory, 6 Infinite Reality-2 Graphics pipes) NMI components Globus and NWS installed on all systems save the Cray SV1NMI components Globus and NWS installed on all systems save the Cray SV1

5 Resources at UT Campus Individual clusters belonging to professors inIndividual clusters belonging to professors in –engineering –computer sciences –NMI components Globus and NWS installed on several machines on campus Computer laboratories having ~100s of PCs in the engineering and computer sciences departmentsComputer laboratories having ~100s of PCs in the engineering and computer sciences departments

6 Campus Grid Model “Hub and Spoke” Model“Hub and Spoke” Model Researchers build programs on their clusters and migrate bigger jobs to TACC resourcesResearchers build programs on their clusters and migrate bigger jobs to TACC resources –Use GSI for authentication –Use GridFTP for data migration –Use LSF Multicluster for migration of jobs Reclaim unused computing cycles on PCs through United Devices infrastructure.Reclaim unused computing cycles on PCs through United Devices infrastructure.

7 UT Campus Grid Overview LSF

8 NMI Testbed Activities Globus – GSI, GRAM, MDS, GridFTPGlobus – GSI, GRAM, MDS, GridFTP –Robust software –Standard Grid middleware –Need to install from source code to link to other components like MPICH-G2, Simple CA Condor-G – submit jobs using GRAM, monitor queues, receive notification, and maintain Globus credentials. LacksCondor-G – submit jobs using GRAM, monitor queues, receive notification, and maintain Globus credentials. Lacks –scheduling capability of Condor –checkpointing

9 NMI Testbed Activities Network Weather Service 2.2.1Network Weather Service –name server for directory services –memory server for storage of data –sensors to gather performance measurements –useful for predicting performance that can be used for a scheduler or “virtual grid” GSI-enabled OpenSSH 1.7GSI-enabled OpenSSH 1.7 –modified version of OpenSSH that allows login to remote systems and transfer files between systems without entering a password – requires replacing native sshd file with GSI-enabled OpenSSH

10 Computations on the UT Grid Components used – GRAM, GSI, GridFTP, MPICH-G2Components used – GRAM, GSI, GridFTP, MPICH-G2 Machines involved – Linux RH (2), Sun (2), Linux Debian (2), Alpha Cluster (16 processors)Machines involved – Linux RH (2), Sun (2), Linux Debian (2), Alpha Cluster (16 processors) Applications run – PI, Ring, SeismicApplications run – PI, Ring, Seismic Successfully ran a demo at SC02 using NMI R2 componentsSuccessfully ran a demo at SC02 using NMI R2 components Relevance to NMIRelevance to NMI –must build from source to link to MPICH-G2 –should be easily configured to submit jobs to schedulers like PBS, LSF, or Loadleveler

11 Computations on the UT Grid Issues to be addressed on clustersIssues to be addressed on clusters –must submit to local scheduler: PBS, LSF or Loadleveler –compute nodes on subnet; cannot communicate with compute nodes on another cluster –must open ports through firewall for communication –version incompatibility – affects source code that are linked to shared libraries

12 Grid Portals HotPage – web page to obtain information on the status of grid resourcesHotPage – web page to obtain information on the status of grid resources –NPACI HotPage ( –TIGRE Testbed portal ( Grid Technologies EmployedGrid Technologies Employed –Security: GSI, SSH, MyProxy for remote proxies –Job Execution: GRAM Gatekeeper –Information Services: MDS (GRIS + GIIS), NWS, Custom information scripts –File Management: GridFTP

13

14 GridPort 2.0 Multi-Application Arch. (Using Globus as Middleware)

15 Future Work Use NMI components where possible in building gridsUse NMI components where possible in building grids Use Lightweight Campus Certificate Policy for instantiating a Certificate Authority at TACCUse Lightweight Campus Certificate Policy for instantiating a Certificate Authority at TACC Build portals and deploy applications on the UT GridBuild portals and deploy applications on the UT Grid

16 Collaborators Mary ThomasMary Thomas Dr. John BoisseauDr. John Boisseau Rich ToscanoRich Toscano Jeson MartajayaJeson Martajaya Eric RobertsEric Roberts Maytal DahanMaytal Dahan Tom UrbanTom Urban