Software Integration Highlights CY2008 Lee Liming, JP Navarro GIG Area Directors for Software Integration University of Chicago, Argonne National Laboratory.

Slides:



Advertisements
Similar presentations
TeraGrid Community Software Areas (CSA) JP (John-Paul) Navarro TeraGrid Grid Infrastructure Group Software Integration University of Chicago and Argonne.
Advertisements

Xsede eXtreme Science and Engineering Discovery Environment Ron Perrott University of Oxford 1.
1 US activities and strategy :NSF Ron Perrott. 2 TeraGrid An instrument that delivers high-end IT resources/services –a computational facility – over.
TeraGrid Quarterly Meeting Dec 6-7, 2007 DVS GIG Project Year 4&5 Project List Kelly Gaither, DVS Area Director.
(e)Science-Driven, Production- Quality, Distributed Grid and Cloud Data Infrastructure for the Transformative, Disruptive, Revolutionary, Next-Generation.
Tools and Services for the Long Term Preservation and Access of Digital Archives Joseph JaJa, Mike Smorul, and Sangchul Song Institute for Advanced Computer.
TeraGrid Science Gateway AAAA Model: Implementation and Lessons Learned Jim Basney NCSA University of Illinois Von Welch Independent.
Simo Niskala Teemu Pasanen
TG QM Arlington: GIG User Support Coordination Plan Sergiu Sanielevici, GIG Area Director for User Support Coordination
Network, Operations and Security Area Tony Rimovsky NOS Area Director
Chapter : Software Process
Attribute-based Authentication for Gateways Jim Basney Terry Fleury Stuart Martin JP Navarro Tom Scavo Jon Siwek Von Welch Nancy Wilkins-Diehr.
TeraGrid’s Integrated Information Service “IIS” Grid Computing Environments 2009 Lee Liming, JP Navarro, Eric Blau, Jason Brechin, Charlie Catlett, Maytal.
NOS Objectives, YR 4&5 Tony Rimovsky. 4.2 Expanding Secure TeraGrid Access A TeraGrid identity management infrastructure that interoperates with campus.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
GIG Software Integration: Area Overview TeraGrid Annual Project Review April, 2008.
TeraGrid Information Services December 1, 2006 JP Navarro GIG Software Integration.
GIG Software Integration Project Plan, PY4-PY5 Lee Liming Mary McIlvain John-Paul Navarro.
TeraGrid Information Services John-Paul “JP” Navarro TeraGrid Grid Infrastructure Group “GIG” Area Co-Director for Software Integration and Information.
TeraGrid Information Services JP Navarro, Lee Liming University of Chicago TeraGrid Architecture Meeting September 20, 2007.
CTSS 4 Strategy and Status. General Character of CTSSv4 To meet project milestones, CTSS changes must accelerate in the coming years. Process –Process.
1 TeraGrid ‘10 August 2-5, 2010, Pittsburgh, PA State of TeraGrid in Brief John Towns TeraGrid Forum Chair Director of Persistent Infrastructure National.
Flexibility and user-friendliness of grid portals: the PROGRESS approach Michal Kosiedowski
Configuration Management (CM)
Data Area Report Chris Jordan, Data Working Group Lead, TACC Kelly Gaither, Data and Visualization Area Director, TACC April 2009.
Contents 1.Introduction, architecture 2.Live demonstration 3.Extensibility.
Coordinating the TeraGrid’s User Interface Areas Dave Hart, Amit Majumdar, Tony Rimovsky, Sergiu Sanielevici.
1 PY4 Project Report Summary of incomplete PY4 IPP items.
Kelly Gaither Visualization Area Report. Efforts in 2008 Focused on providing production visualization capabilities (software and hardware) Focused on.
TeraGrid Quarterly Meeting Dec 5 - 7, 2006 Data, Visualization and Scheduling (DVS) Update Kelly Gaither, DVS Area Director.
TeraGrid CTSS Plans and Status Dane Skow for Lee Liming and JP Navarro OSG Consortium Meeting 22 August, 2006.
Rochester Institute of Technology Cyberaide Shell: Interactive Task Management for Grids and Cyberinfrastructure Gregor von Laszewski, Andrew J. Younge,
TeraGrid Advanced Scheduling Tools Warren Smith Texas Advanced Computing Center wsmith at tacc.utexas.edu.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Biomedical and Bioscience Gateway to National Cyberinfrastructure John McGee Renaissance Computing Institute
Leveraging the InCommon Federation to access the NSF TeraGrid Jim Basney Senior Research Scientist National Center for Supercomputing Applications University.
TeraGrid Extension Gateway Activities Nancy Wilkins-Diehr TeraGrid Quarterly, September 24-25, 2009 The Extension Proposal!
Rob Davidson, Partner Technology Specialist Microsoft Management Servers: Using management to stay secure.
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Biomedical and Bioscience Gateway to National Cyberinfrastructure John McGee Renaissance Computing Institute
Distributed Data for Science Workflows Data Architecture Progress Report December 2008.
User-Facing Projects Update David Hart, SDSC April 23, 2009.
Education, Outreach and Training (EOT) and External Relations (ER) Scott Lathrop Area Director for EOT Extension Year Plans.
2005 GRIDS Community Workshop1 Learning From Cyberinfrastructure Initiatives Grid Research Integration Development & Support
Data Area Report Chris Jordan, Data Working Group Lead, TACC Kelly Gaither, Data and Visualization Area Director, TACC April 2009.
TeraGrid Program Year 5 Overview John Towns Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing Applications University.
Data, Visualization and Scheduling (DVS) TeraGrid Annual Meeting, April 2008 Kelly Gaither, GIG Area Director DVS.
Network, Operations and Security Area Tony Rimovsky NOS Area Director
Attribute-based Authentication for Gateways Jim Basney Terry Fleury Stuart Martin JP Navarro Tom Scavo Nancy Wilkins-Diehr.
CTSS Rollout update Mike Showerman JP Navarro April
CTSS Version 4 User Support Documentation Mike Dwyer, Kerry Hagan, Diana Diehl.
TeraGrid’s Common User Environment: Status, Challenges, Future Annual Project Review April, 2008.
TeraGrid Program Year 5 Overview John Towns Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing Applications University.
TG ’08, June 9-13, State of TeraGrid John Towns Co-Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing.
Common User Environments - Update Shawn T. Brown, PSC CUE Working Group Lead TG Quartely 1.
Building PetaScale Applications and Tools on the TeraGrid Workshop December 11-12, 2007 Scott Lathrop and Sergiu Sanielevici.
TeraGrid Capability Discovery John-Paul “JP” Navarro TeraGrid Area Co-Director for Software Integration University of Chicago/Argonne National Laboratory.
Data Infrastructure in the TeraGrid Chris Jordan Campus Champions Presentation May 6, 2009.
Purdue RP Highlights TeraGrid Round Table November 5, 2009 Carol Song Purdue TeraGrid RP PI Rosen Center for Advanced Computing Purdue University.
TeraGrid’s Process for Meeting User Needs. Jay Boisseau, Texas Advanced Computing Center Dennis Gannon, Indiana University Ralph Roskies, University of.
TeraGrid Software Integration: Area Overview (detailed in 2007 Annual Report Section 3) Lee Liming, JP Navarro TeraGrid Annual Project Review April, 2008.
NSF TeraGrid Review January 10, 2006
TeraGrid Information Services
NSF TeraGrid Review January 10, 2006
TeraGrid Information Services: Building on Globus MDS4
TeraGrid Information Services Developer Introduction
Information Services Discussion TeraGrid ‘08
Leigh Grundhoefer Indiana University
Presentation transcript:

Software Integration Highlights CY2008 Lee Liming, JP Navarro GIG Area Directors for Software Integration University of Chicago, Argonne National Laboratory April

Expanding TeraGrid Capabilities Moving capabilities from working groups to production –Help working groups define new TG-wide capabilities (SGW support, Lustre & GPFS WAN, scheduling, etc.) –Formally document new/enhanced capabilities and work out integration, testing, support details –Prepare software binaries and installers for TG systems Operated central services –Information service –Software build & test service –Speed page (data movement performance monitor) –DMOVER and Lustre WAN Initiated the Quality Assurance activity –Predecessor to QA/CUE working group 2

Capability Model In 2005, we retooled our SW coordination process –Emphasis on use cases, user scenarios enabled by SW –Bottom up, user-driven capability model –Open processes for community input into system def Original TeraGrid (DTF) was aimed at a narrow set of distributed HPC applications –Single platform, narrow user base and target uses (distributed HPC) –Heavy emphasis on identical software environment By 2004 commissioning, TeraGrid had expanded in scope to cover all NSF HPC applications –Very diverse user community, resources –Very wide diversity of user scenarios and use patterns 3

CTSS 4 – A New Paradigm Significant change in how we define CTSS –CTSS 1 thru 3: Monolithic software stack –CTSS 4: Modular user capabilities –Improved many aspects of capability delivery Better descriptions of the capabilities (esp. for users) Better documentation Clearer availability information More focused delivery process (package, deploy, and configure) Improved process for RPs to select and publish their offerings Delivery timeline –Designed in 2006 –Capability kits defined in Q1-Q –Capabilities rolled out in Q2-Q April

5 CTSS Capability Kits (April 2009) Capability Kit (14)Description TeraGrid Core IntegrationMinimal components that integrate RP resources Remote LoginRemote login using TeraGrid credentials to a coordinated Unix environment Remote ComputeRemote job submission Application Development and Runtime Compile and execute applications Data ManagementCollaborative data management capabilities Data MovementData movement to/from RP resources Parallel Application SupportIdentify and configure MPI runtime environment Science Gateway Support new!End-user counting, improved security for gateways On-demand Computation new!On-demand (little or no wait) computing Co-scheduling new!Reserving a set of resources for use at a specific time Science WorkflowRun an orchestrated collection of interdependent jobs Wide Area GPFSLocal access to TeraGrid wide GPFS filesystems Wide Area Lustre new!Local access to TeraGrid wide Lustre filesystems VisualizationCompile and execute visualization applications

2008 Availability & Usage Key idea: Capability usage vs. component usage Most CTSS capabilities were available on all (or nearly all) TG systems and were used heavily or frequently everywhere –Remote compute was used heavily on some systems (like those appropriate for SGW usage) and not on others –Visualization capability was used heavily at UC/Argonne and TACC (other TG resources offer diverse visualization capabilities) –Science workflow capability was used less than once/day, but each use generated 100s or 1000s of jobs Heavy use means more than 100 uses/day on a single system. Frequent use means 1 – 100 uses/day on a single system. Infrequent use means less than 1 use/day on a single system. 6

2008 Operational Issues In 2008, CTSS comprised 10 separate capabilities, with ~80 software components on 19 platforms 16 issues reported by RPs –Installation docs incorrect/incomplete –A GIG-provided installer doesn’t fit well with a system –Issues with specific components (as provided by developers) –Inca test not accurate in all situations –Enhancement requests from admins 7

Capability Development & Expansion VM hosting services supports science teams that utilize highly tailored environments or service-oriented applications –Provided by IU Quarry and Purdue Wispy Science gateway support enables end-user tracking and improved security for gateways –Defined and on track for PY4 availability Client software distribution supports campus champions and related initiatives –Released for evaluation Public build/test system supports NSF SDCI/STCI and CISE program awardees –on track for PY4 availability 8

Advanced Scheduling Capabilities Documented designs and implementations for TeraGrid advanced scheduling capabilities –On-demand computation –Advance reservation –Co-scheduling Broadened availability of new capabilities –On-demand at IU, NCAR, NCSA, SDSC, TACC, and UC/Argonne –Advance reservation and co-scheduling at LONI, NCSA, SDSC Automatic resource selection –In development, still on schedule for end of PY4 9

Information Services Enhancements TeraGrid’s Integrated Information Service is a vital communication channel for system-wide functions –Used by Inca to plan verification tests –Helps keep user documentation up-to-date –Provides queue status data for user portal monitors –Provides data for automatic resource selection –Configures speed page test runs –In general, enables automation of many routine housekeeping tasks Expanded content –Local HPC software registry, SGW-available science tools, resource descriptions Expanded access methods –REST application framework, multiple data formats 10

Questions? Moving capabilities from working groups to operations –Helping WGs move from ideas to production support –Capability-oriented software coordination model –Integration, testing, support planning –Preparing software for deployment on TG resources Specific capabilities –Advanced scheduling capabilities –Information services enhancements –Enhanced science gateway security, end user tracking –VM hosting for highly specialized or service-oriented applications –Software for campuses –Helping SDCI/STCI and CISE awardees prepare software for TG 11