(e)Science-Driven, Production- Quality, Distributed Grid and Cloud Data Infrastructure for the Transformative, Disruptive, Revolutionary, Next-Generation.

Slides:



Advertisements
Similar presentations
A Workflow Engine with Multi-Level Parallelism Supports Qifeng Huang and Yan Huang School of Computer Science Cardiff University
Advertisements

Xsede eXtreme Science and Engineering Discovery Environment Ron Perrott University of Oxford 1.
1 US activities and strategy :NSF Ron Perrott. 2 TeraGrid An instrument that delivers high-end IT resources/services –a computational facility – over.
TeraGrid Quarterly Meeting Dec 6-7, 2007 DVS GIG Project Year 4&5 Project List Kelly Gaither, DVS Area Director.
Unveiling ProjectWise V8 XM Edition. ProjectWise V8 XM Edition An integrated system of collaboration servers that enable your AEC project teams, your.
Kathy Benninger, Pittsburgh Supercomputing Center Workshop on the Development of a Next-Generation Cyberinfrastructure 1-Oct-2014 NSF Collaborative Research:
Sergiu January 2007 TG Users’ Data Transfer Needs SDSC NCAR TACC UC/ANL NCSA ORNL PU IU PSC.
Data Gateways for Scientific Communities Birds of a Feather (BoF) Tuesday, June 10, 2008 Craig Stewart (Indiana University) Chris Jordan.
XSEDE 13 July 24, Galaxy Team: PSC Team:
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
Data Grids: Globus vs SRB. Maturity SRB  Older code base  Widely accepted across multiple communities  Core components are tightly integrated Globus.
Simo Niskala Teemu Pasanen
Core Services I & II David Hart Area Director, UFP/CS TeraGrid Quarterly Meeting December 2008.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
GIG Software Integration: Area Overview TeraGrid Annual Project Review April, 2008.
GRID job tracking and monitoring Dmitry Rogozin Laboratory of Particle Physics, JINR 07/08/ /09/2006.
TeraGrid Information Services December 1, 2006 JP Navarro GIG Software Integration.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
Open Science Grid Software Stack, Virtual Data Toolkit and Interoperability Activities D. Olson, LBNL for the OSG International.
Flexibility and user-friendliness of grid portals: the PROGRESS approach Michal Kosiedowski
The Data Grid: Towards an Architecture for the Distributed Management and Analysis of Large Scientific Dataset Caitlin Minteer & Kelly Clynes.
Grids and Portals for VLAB Marlon Pierce Community Grids Lab Indiana University.
Data Area Report Chris Jordan, Data Working Group Lead, TACC Kelly Gaither, Data and Visualization Area Director, TACC April 2009.
GRAM5 - A sustainable, scalable, reliable GRAM service Stuart Martin - UC/ANL.
BOSCO Architecture Derek Weitzel University of Nebraska – Lincoln.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
The Future of the iPlant Cyberinfrastructure: Coming Attractions.
CYBERINFRASTRUCTURE FOR THE GEOSCIENCES Data Replication Service Sandeep Chandra GEON Systems Group San Diego Supercomputer Center.
DIRAC Review (13 th December 2005)Stuart K. Paterson1 DIRAC Review Exposing DIRAC Functionality.
Application portlets within the PROGRESS HPC Portal Michał Kosiedowski
TeraGrid Quarterly Meeting Dec 5 - 7, 2006 Data, Visualization and Scheduling (DVS) Update Kelly Gaither, DVS Area Director.
Apache Airavata (Incubating) Gateway to Grids & Clouds Suresh Marru Nov 10 th 2011.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
Grid Architecture William E. Johnston Lawrence Berkeley National Lab and NASA Ames Research Center (These slides are available at grid.lbl.gov/~wej/Grids)
Rochester Institute of Technology Cyberaide Shell: Interactive Task Management for Grids and Cyberinfrastructure Gregor von Laszewski, Andrew J. Younge,
TeraGrid Advanced Scheduling Tools Warren Smith Texas Advanced Computing Center wsmith at tacc.utexas.edu.
Institute For Digital Research and Education Implementation of the UCLA Grid Using the Globus Toolkit Grid Center’s 2005 Community Workshop University.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
The Global Land Cover Facility is sponsored by NASA and the University of Maryland.The GLCF is a founding member of the Federation of Earth Science Information.
CEOS Working Group on Information Systems and Services - 1 Data Services Task Team Discussions on GRID and GRIDftp Stuart Doescher, USGS WGISS-15 May 2003.
SAN DIEGO SUPERCOMPUTER CENTER Inca Control Infrastructure Shava Smallen Inca Workshop September 4, 2008.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
7. Grid Computing Systems and Resource Management
1 Overall Architectural Design of the Earth System Grid.
Distributed Data for Science Workflows Data Architecture Progress Report December 2008.
Data Area Report Chris Jordan, Data Working Group Lead, TACC Kelly Gaither, Data and Visualization Area Director, TACC April 2009.
Data, Visualization and Scheduling (DVS) TeraGrid Annual Meeting, April 2008 Kelly Gaither, GIG Area Director DVS.
Grid Interoperability Update on GridFTP tests Gregor von Laszewski
Tool Integration with Data and Computation Grid “Grid Wizard 2”
AHM04: Sep 2004 Nottingham CCLRC e-Science Centre eMinerals: Environment from the Molecular Level Managing simulation data Lisa Blanshard e- Science Data.
LSF Universus By Robert Stober Systems Engineer Platform Computing, Inc.
Bulk Data Transfer Activities We regard data transfers as “first class citizens,” just like computational jobs. We have transferred ~3 TB of DPOSS data.
TeraGrid-Wide Operations Von Welch Area Director for Networking, Operations and Security NCSA, University of Illinois April, 2009.
IU Site Update TeraGrid Round Table Craig Stewart, Stephen Simms, Kurt Seiffert November 4, 2010.
NeST: Network Storage John Bent, Venkateshwaran V Miron Livny, Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau.
ATLAS Distributed Analysis Dietrich Liko IT/GD. Overview  Some problems trying to analyze Rome data on the grid Basics Metadata Data  Activities AMI.
© Geodise Project, University of Southampton, Workflow Support for Advanced Grid-Enabled Computing Fenglian Xu *, M.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
TG ’08, June 9-13, State of TeraGrid John Towns Co-Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing.
Common User Environments - Update Shawn T. Brown, PSC CUE Working Group Lead TG Quartely 1.
Data Infrastructure in the TeraGrid Chris Jordan Campus Champions Presentation May 6, 2009.
TeraGrid Software Integration: Area Overview (detailed in 2007 Annual Report Section 3) Lee Liming, JP Navarro TeraGrid Annual Project Review April, 2008.
Onedata Eventually Consistent Virtual Filesystem for Multi-Cloud Infrastructures Michał Orzechowski (CYFRONET AGH)
OpenMosix, Open SSI, and LinuxPMI
GWE Core Grid Wizard Enterprise (
Cyberinfrastructure and PolarGrid
Saravana Kumar CEO/Founder - Kovai Atomic Scope – Product Update.
Presentation transcript:

(e)Science-Driven, Production- Quality, Distributed Grid and Cloud Data Infrastructure for the Transformative, Disruptive, Revolutionary, Next-Generation TeraGrid (now with free ponies) Data Architecture Progress Report December 11, 2008 Chris Jordan

Goals for the Data Architecture Improve the experience of working with data in the TeraGrid for the user community Reliability, Ease of use, Performance Integrate data management into the user workflow Balance performance goals against usability Avoid overdependence on data location Support the most common use cases as transparently as possible Move data in, run job, move data out as basic pattern Organize, search, and retrieve data from large “collections”

Some Realities Cannot address the issue of available storage Limited opportunity to improve data transfer performance at the high end Cannot introduce drastic changes to TG infrastructure at this stage of the project Remain dependent on the availability of technology and resources for wide-area file systems

Areas of Effort Simplifying command-line data movement Extending the reach of WAN file systems Develop unified data replication and management infrastructure Extend and unify user portal interfaces to data Integrate data into scheduling and workflows Provide common access mechanisms to diverse, distributed data resources

Extending Wide-Area File Systems A “Wide-Area” file system is available on multiple resources A “Global” file system is available on all TeraGrid resources Indiana and SDSC each have a WAN-FS in production now PSC has promising technology for distributed storage and Kerberos integration, but need testing to understand best management practices Point of emphasis: going production

Data Capacitor-WAN (DC-WAN) IU has this in production on BigRed, PSC Pople Can be mounted on any cluster running Lustre 1.4 or Lustre 1.6 Ready for testing and move to production Sites and resources committed: –TACC Lonestar, Ranger, Spur –NCSA Abe, possibly Cobalt and/or Mercury –LONI Queen Bee (testing, possible production) –Purdue Steele? –This presentation is an excellent opportunity to add your site to this list.

PSC “Josephine-WAN” Two major new design features: –Kerberos-based identity mapping –Distributed data and metadata Kerberos is likely to work well OOTB Distributed data/”storage pools” will need careful configuration and management Technology working well, but needs to be actively investigated and tested in various configurations Want to work on integration with TG User Portal

Getting to Global No single file system technology will be compatible/feasible to deploy on every system Will require hybrid solutions TGUP helps, but … Need to understand limit on simultaneous mounts, and … Once production DC-WAN reaches the technical limit, look at technologies to extend the FS: –pNFS –FUSE/SSHFS

Command-line tools Many users are still oriented towards shell access GridFTP is complicated to use via globus-url-copy –Long URLS, many often inconsistent options SSH/SCP is almost universally available and familiar to users Limited usefulness for data transfer in current configuration Simple changes to SSH/SCP configuration: –Support SCP-based access to data mover nodes –Support simpler addressing of data resources –Provide resource specific “default” configuration

Unified Data Management Management of both data and metadata, distributed across storage resources Multiple sites support data collections using SRB, iRODS, databases, web services, etc. This diversity is good in the abstract, but also confusing to new users Extend current iRODS-based data management infrastructure to additional sites Expand REDDNET “cloud storage” availability Integrate access to as many collections as possible through the User Portal

Interfaces to Data SSH and “ls” are not effective interfaces to large, complex datasets Portal and Gateway interfaces to data have proven useful and popular, but: They may not be able to access all resources, may require significant gateway developer effort Extend user portal to support WAN file systems and distributed data management Possible to expose user portal and other APIs to ease development of gateways?

Integrating Data into Workflows Almost all tasks run on TeraGrid require some data management and multiple storage resources Users should be able to include these steps as part of a job or workflow submission Port DMOVER to additional schedulers, deploy across TeraGrid –Working on BigBen, ready for Kraken –Working on SGE and LSF Evaluate PetaShare, other “Data Scheduling” systems (Stork?)

Gratuitous end slide #42 Data-WG has many attendees, but few participants We need: –More sites committed to deploying DC-WAN in production –More sites committed to testing “Josephine-WAN” –More sites contributing to Data Collections infrastructure –Help porting DMOVER, testing PetaShare and REDDNET –Users and projects to exercise the infrastructure –Select one or more If not you, who? If not now, when?