Science Gateways on the TeraGrid Nancy Wilkins-Diehr Area Director for Science Gateways San Diego Supercomputer Center

Slides:



Advertisements
Similar presentations
Scaling TeraGrid Access A Testbed for Attribute-based Authorization and Leveraging Campus Identity Management
Advertisements

Xsede eXtreme Science and Engineering Discovery Environment Ron Perrott University of Oxford 1.
1 US activities and strategy :NSF Ron Perrott. 2 TeraGrid An instrument that delivers high-end IT resources/services –a computational facility – over.
1 WeCAN Works Presentation Monday, December 14, 2009.
High Performance Computing Course Notes Grid Computing.
Ian Foster Computation Institute Argonne National Lab & University of Chicago Education in the Science 2.0 Era.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
O C I October 31, 2006Office of CyberInfrastructure1 Enhancing Virtual Organizations Abhi Deshmukh Office of Cyberinfrastructure & Engineering Directorate.
TeraGrid Science Gateway AAAA Model: Implementation and Lessons Learned Jim Basney NCSA University of Illinois Von Welch Independent.
April 2009 OSG Grid School - RDU 1 Open Science Grid John McGee – Renaissance Computing Institute University of North Carolina, Chapel.
A History of the TeraGrid Science Gateway Program: A Personal View Nancy Wilkins-Diehr GCE11, November 18, 2011.
NOS Objectives, YR 4&5 Tony Rimovsky. 4.2 Expanding Secure TeraGrid Access A TeraGrid identity management infrastructure that interoperates with campus.
GCE06, Tampa, FL November 12-13, 2006 Science Gateways on the TeraGrid Charlie Catlett, Sebastien Goasguen, Jim Marsteller, Stuart Martin, Don Middleton,
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
GIG Software Integration: Area Overview TeraGrid Annual Project Review April, 2008.
SAN DIEGO SUPERCOMPUTER CENTER Science Gateways on the TeraGrid Nancy Wilkins-Diehr TeraGrid Area Director for Science Gateways SDSC Director of Consulting,
April 2006 Science Gateways on the TeraGrid Nancy Wilkins-Diehr Area Director for Science Gateways San Diego Supercomputer Center
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
National Center for Supercomputing Applications The Computational Chemistry Grid: Production Cyberinfrastructure for Computational Chemistry PI: John Connolly.
TeraGrid Science Gateways: Scaling TeraGrid Access Aaron Shelmire¹, Jim Basney², Jim Marsteller¹, Von Welch²,
Advancing Scientific Discovery through TeraGrid Scott Lathrop TeraGrid Director of Education, Outreach and Training University of Chicago and Argonne National.
August 2007 Advancing Scientific Discovery through TeraGrid Scott Lathrop TeraGrid Director of Education, Outreach and Training University of Chicago and.
August 2007 Advancing Scientific Discovery through TeraGrid Adapted from S. Lathrop’s talk in SC’07
SAN DIEGO SUPERCOMPUTER CENTER NUCRI Advisory Board Meeting November 9, 2006 Science Gateways on the TeraGrid Nancy Wilkins-Diehr TeraGrid Area Director.
ESP workshop, Sept 2003 the Earth System Grid data portal presented by Luca Cinquini (NCAR/SCD/VETS) Acknowledgments: ESG.
TeraGrid Overview Cyberinfrastructure Days Internet2 10/9/07 Mark Sheddon Resource Provider Principal Investigator San Diego Supercomputer Center
Grid Resource Allocation and Management (GRAM) Execution management Execution management –Deployment, scheduling and monitoring Community Scheduler Framework.
NEES Cyberinfrastructure Center at the San Diego Supercomputer Center, UCSD George E. Brown, Jr. Network for Earthquake Engineering Simulation Science.
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
SAN DIEGO SUPERCOMPUTER CENTER Impact Requirements Analysis Team Co-Chairs: Mark Sheddon (SDSC) Ann Zimmerman (University of Michigan) Members: John Cobb.
1 PY4 Project Report Summary of incomplete PY4 IPP items.
The Grid System Design Liu Xiangrui Beijing Institute of Technology.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
NanoHUB.org and HUBzero™ Platform for Reproducible Computational Experiments Michael McLennan Director and Chief Architect, Hub Technology Group and George.
TeraGrid CTSS Plans and Status Dane Skow for Lee Liming and JP Navarro OSG Consortium Meeting 22 August, 2006.
Introduction to Grid Computing Ed Seidel Max Planck Institute for Gravitational Physics
Renaissance Computing Institute: An Overview Lavanya Ramakrishnan, John McGee, Alan Blatecky, Daniel A. Reed Renaissance Computing Institute.
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
Institute For Digital Research and Education Implementation of the UCLA Grid Using the Globus Toolkit Grid Center’s 2005 Community Workshop University.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
NEES Cyberinfrastructure Center at the San Diego Supercomputer Center, UCSD George E. Brown, Jr. Network for Earthquake Engineering Simulation NEES TeraGrid.
Biomedical and Bioscience Gateway to National Cyberinfrastructure John McGee Renaissance Computing Institute
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
Leveraging the InCommon Federation to access the NSF TeraGrid Jim Basney Senior Research Scientist National Center for Supercomputing Applications University.
SC06, Tampa FL November 11-17, 2006 Science Gateways on the TeraGrid Powerful Beyond Imagination! Nancy Wilkins-Diehr TeraGrid Area Director for Science.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Riding the Crest: High-End Cyberinfrastructure Experiences and Opportunities on the NSF TeraGrid A Panel Presentation by Laura M c GinnisRadha Nandkumar.
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Biomedical and Bioscience Gateway to National Cyberinfrastructure John McGee Renaissance Computing Institute
Cyberinfrastructure: Many Things to Many People Russ Hobby Program Manager Internet2.
2005 GRIDS Community Workshop1 Learning From Cyberinfrastructure Initiatives Grid Research Integration Development & Support
Network, Operations and Security Area Tony Rimovsky NOS Area Director
TeraGrid Overview John-Paul “JP” Navarro TeraGrid Area Co-Director for Software Integration University of Chicago/Argonne National Laboratory March 25,
AT LOUISIANA STATE UNIVERSITY CCT: Center for Computation & Technology Introduction to the TeraGrid Daniel S. Katz Lead, LONI as a TeraGrid.
October 2007 TeraGrid : Advancing Scientific Discovery and Learning Diane A. Baxter, Ph.D. Education Director San Diego Supercomputer Center University.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
SAN DIEGO SUPERCOMPUTER CENTER Science Gateways on the TeraGrid Nancy Wilkins-Diehr TeraGrid Area Director for Science Gateways SDSC Director of Consulting,
Grid Deployment Technical Working Groups: Middleware selection AAA,security Resource scheduling Operations User Support GDB Grid Deployment Resource planning,
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
TeraGrid’s Process for Meeting User Needs. Jay Boisseau, Texas Advanced Computing Center Dennis Gannon, Indiana University Ralph Roskies, University of.
Gateways security Aashish Sharma Security Engineer National Center for Supercomputing Applications (NCSA) University of Illinois at Urbana-Champaign.
Bob Jones EGEE Technical Director
Office of CyberInfrastructure
Leigh Grundhoefer Indiana University
Data Management Components for a Research Data Archive
Presentation transcript:

Science Gateways on the TeraGrid Nancy Wilkins-Diehr Area Director for Science Gateways San Diego Supercomputer Center

Today’s Outline What are Gateways? Why TeraGrid and Gateways? Initial Strategy Implementation Details –Issues to address when using TG Future growth

Gateways are part of TG’s 3-pronged strategy to further science DEEP Science: Enabling Terascale Science –Make science more productive through an integrated set of very- high capability resources ASTA projects WIDE Impact: Empowering Communities –Bring TeraGrid capabilities to the broad science community Science Gateways OPEN Infrastructure, OPEN Partnership –Provide a coordinated, general purpose, reliable set of services and resources Grid interoperability working group

Science Gateways A new initiative for the TeraGrid Increasing investment by communities in their own cyberinfrastructure, but heterogeneous: Resources Users – from expert to K-12 Software stacks, policies Science Gateways –Provide “TeraGrid Inside” capabilities –Leverage community investment Three common forms: –Web-based Portals –Application programs running on users' machines but accessing services in TeraGrid –Coordinated access points enabling users to move seamlessly between TeraGrid and other grids. Workflow Composer

Initial Focus on 10 Gateways

National Virtual Observatory Facilitating Scientific Discovery Astronomy is increasingly a data-rich science New science enabled by enhancing access to data and computing resources Ease of use in locating, retrieving, and analyzing data from archives and catalogs worldwide NVO is a set of tools used to exploit the data avalanche

NanoHub Harnesses TeraGrid for Education Nanohub is used to complete coursework by undergraduate and graduate students in dozens of courses at 10 universities. Currently serves over 1000 users.

NanoHUB Middleware infrastructure Campus Grids Purdue, GLOW Grid Capability Computing Science Gateway Workspaces Research apps Virtual backends Virtual Cluster with VIOLIN VM Capacity Computing nanoHUB VO Middleware

spruce.teragrid.org Special Priority and Urgent Computing Environment

Biomedical and Biology Gateway Led by Dan Reed, Renaissance Computing Institute, North Carolina Supports –Distributed collaboration –Multi-site data access –Computational tools for local or remote execution –Grid and cluster interoperability Will provides access to –Common sequence and protein structure databases –Over 140 software packages

Linked Environments for Atmospheric Discovery LEAD Providing tools that are needed to make accurate predictions of tornados and hurricanes Data exploration and Grid workflow

Gateways are growing in numbers 10 initial projects as part of TG proposal >20 Gateway projects today No limit on how many gateways can use TG resources –Prepare services and documentation so developers can work independently Open Science Grid (OSG) Special PRiority and Urgent Computing Environment (SPRUCE) National Virtual Observatory (NVO) Linked Environments for Atmospheric Discovery (LEAD) Computational Chemistry Grid (GridChem) Computational Science and Engineering Online (CSE- Online) GEON(GEOsciences Network) Network for Earthquake Engineering Simulation (NEES) SCEC Earthworks Project Network for Computational Nanotechnology and nanoHUB GIScience Gateway (GISolve) Biology and Biomedicine Science Gateway Open Life Sciences Gateway The Telescience Project Grid Analysis Environment (GAE) Neutron Science Instrument Gateway TeraGrid Visualization Gateway, ANL BIRN Gridblast Bioinformatics Gateway Earth Systems Grid Cornell Many others interested –SID Grid –HASTAC

NCAR Earth System Grid ESG originally a distributed data management/access system but it has evolved into more. User registration, authorization controls, and metrics tracking CCSM model source, initialization datasets, post- processing codes, and analysis and visualization tools. Prototypes of model- submission environments, eventually real- time tracking of model status along with references to available output datasets. "science gateway" for climate research. Expect to see more model runs at higher- resolution and with greater component scope.

So how will we meet all these needs? With RATS! (Requirements Analysis Teams) Collection, analysis and consolidation of requirements to jump start the work –Interviews with 10 Gateways –Common user models, accounting needs, scheduling needs Summarized requirements for each TeraGrid working group –Accounting, Security, Web Services, Software Areas for more study identified Primer outline for new Gateways in progress And milestones

Implications for TeraGrid working groups Accounting –Support for accounts with differing capabilities –Ability to associate compute job to a individual portal user –Scheme for portal registration and usage tracking –Support for OSG’s Grid User Management System (GUMS) –Dynamic accounts Security –Community account privileges –Need to identify human responsible for a job for incident response –Acceptance of other grid certificates –TG-hosted web servers, cgi-bin code Web Services –Initial analysis completed 12/05 –Some Gateways (LEAD, Open Life Sciences) have immediate needs –Many will build on capabilities offered by GT4, but interoperability could be an issue –Web Service security –Interfaces to scheduling and account management are common requirements Software –Interoperability of software stacks between TG and peer grids –Software installations for gateways across all TG sites –Community software areas –Management (pacman, other options)

Gateway Web Services Needs Interfaces provided by the TeraGrid The list of services that have been identified by the gateways developers includes: –Resource Status Service (both polling and pub/sub) –Job Submission Interface The gateways expect this to be provided by WS-GRAM –Job Tracking Interface (Both polling and pub/sub) –File/Data Staging Interface –Retrieve Usage Information –Retrieve Inca Info –Advanced Reservation Interface –Cross-site Run interface –Pushing DN to an RP interface Interfaces provided by the Gateways The list of services that have been identified by the gateways developers and the TeraGrid Security group includes: –Retrieve user information for a job –Retrieve accounting information/statistics –Provides the necessary means to track down problem job submissions, identify malicious users, and tabulate accounting and logging information for reporting needs by the RPs. It is expected that the information provided for the first interface is simply the (resource, job id) that is known by both parties at job submission time. This interface provides sufficient user information for the RPs to deal with the situation at hand, and possibly identifies another interface that should be provided by the gateways: –Don't submit jobs from the user who submitted job (resource, job id), until we say it's Ok. –The accounting interface requires no information, but returns sufficient accounting information and statistics to report to funding agencies, program managers, etc.

Gateway primer and “getting started” documentation by end of summer 1. Introduction 2. Science Gateway in Context a. Science Gateway (SGW) Definition(s) b. Science Gateway user modes c. Distinction between SGW and other TeraGrid user modes 3. Components of a Science Gateway a. User Model b. Gateway targeted community c. Gateway Services d. Integration with TeraGrid external resources (data collections, services, …) e. Organizational and administrative structure 4. TeraGrid services and policies available for Science Gateways a. Portal middleware tools (user portal and other portal tools) b. Account Management (user models, community accounts, ) c. Security environment (security models) d. Web Services e. Scheduling services (and meta-scheduling) f. Community accounts and allocations g. Community Software Areas h. All traditional TeraGrid services and resources i. Ability to propose additional services and how that would interact with TeraGrid operations 5. Responsibilities and Requirements for Science Gateways a. Interaction with and compatibility with TeraGrid communities b. Control procedures i. Community user identification and tracking (map TeraGrid usage to Portal user) ii. Use monitoring and reporting iii. Security and trust iv. Appropriate use 6. How to get started a. Existing resources i. Publication references ii. Web areas with more details iii. Online tutorials iv. Upcoming presentations and tutorials b. Who to contact for initial discussions c. How to propose a new Gateway d. How to integrate with TeraGrid Gateways efforts. e. How to obtain a resource allocation

Want to be involved? mailing – – in body Biweekly telecons to get advice from others. Current focus –Auditing strategy –Mini-tutorial at April Lariat workshop, “Accelerating Research Through Grid Computing” –Hands on tutorial at June conference Overview of Gateways In depth presentations by LEAD, nanoHUB, RENCI, GIScience –Transition to GT4 –Scheduling requirements As original gateways move into production, we will be able to provide short term support to new projects that would benefit from utilizing TeraGrid resources Nancy Wilkins-Diehr,