The Grid as Future Scientific Infrastructure Ian Foster Argonne National Laboratory University of Chicago Globus Alliance www.mcs.anl.gov/~foster.

Slides:



Advertisements
Similar presentations
National e-Science Centre Glasgow e-Science Hub Opening: Remarks NeSCs Role Prof. Malcolm Atkinson Director 17 th September 2003.
Advertisements

NG-CHC Northern Gulf Coastal Hazards Collaboratory Simulation Experiment Integration Sandra Harper 1, Manil Maskey 1, Sara Graves 1, Sabin Basyal 1, Jian.
Towards a Virtual European Supercomputing Infrastructure Vision & issues Sanzio Bassini
EInfrastructures (Internet and Grids) US Resource Centers Perspective: implementation and execution challenges Alan Blatecky Executive Director SDSC.
1 Cyberinfrastructure Framework for 21st Century Science & Engineering (CIF21) NSF-wide Cyberinfrastructure Vision People, Sustainability, Innovation,
1 Cyberinfrastructure Framework for 21st Century Science & Engineering (CF21) IRNC Kick-Off Workshop July 13,
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
EInfrastructures (Internet and Grids) - 15 April 2004 Sharing ICT Resources – “Think Globally, Act Locally” A point-of-view from the United States Mary.
Knowledge Environments for Science: Representative Projects Ian Foster Argonne National Laboratory University of Chicago
The Challenges of Grid Computing Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer Science The.
The Grid as Infrastructure and Application Enabler Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer.
Assessment of Core Services provided to USLHC by OSG.
Computing in Atmospheric Sciences Workshop: 2003 Challenges of Cyberinfrastructure Alan Blatecky Executive Director San Diego Supercomputer Center.
Peer to Peer & Grid Computing Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer Science The University.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
CI Days: Planning Your Campus Cyberinfrastructure Strategy Russ Hobby, Internet2 Internet2 Member Meeting 9 October 2007.
High level Knowledge-based Grid Services for Bioinformaticans Carole Goble, University of Manchester, UK myGrid project
Ian Foster Argonne National Lab University of Chicago Globus Project The Grid and Meteorology Meteorology and HPN Workshop, APAN.
Standards and Ontologies to Enable Discovery Data and Information Integration Robin McEntire GlaxoSmithKline 19 Nov, 2002.
What is Internet2? Ted Hanss, Internet2 5 March
Some Grid Experiences Laura Pearlman USC Information Sciences Institute ICTP Advanced Training Workshop on Scientific Instruments on the Grid *Most of.
What is Cyberinfrastructure? Russ Hobby, Internet2 Clemson University CI Days 20 May 2008.
Miron Livny Computer Sciences Department University of Wisconsin-Madison Welcome and Condor Project Overview.
KAROLINSKA INSTITUTET International Biobank and Cohort Studies: Developing a Harmonious Approch February 7-8, 2005, Atlanta; GA Standards The P 3 G knowledge.
MyGrid and the Semantic Web Phillip Lord School of Computer Science University of Manchester.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
Perspectives on Grid Technology Ian Foster Argonne National Laboratory The University of Chicago.
Data and storage services on the NGS Mike Mineter Training Outreach and Education
SEEK Welcome Malcolm Atkinson Director 12 th May 2004.
Commodity Grid Kits Gregor von Laszewski (ANL), Keith Jackson (LBL) Many state-of-the-art scientific applications, such as climate modeling, astrophysics,
Policy Based Data Management Data-Intensive Computing Distributed Collections Grid-Enabled Storage iRODS Reagan W. Moore 1.
High level Grid Services for Bioinformaticans Carole Goble, University of Manchester, UK Robin McEntire, GSK.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
GRID ARCHITECTURE Chintan O.Patel. CS 551 Fall 2002 Workshop 1 Software Architectures 2 What is Grid ? "...a flexible, secure, coordinated resource- sharing.
1 ARGONNE  CHICAGO Grid Introduction and Overview Ian Foster Argonne National Lab University of Chicago Globus Project
Authors: Ronnie Julio Cole David
The Globus Toolkit®: The Open Source Solution for Grid Computing
My Grid and Taverna: Now and in the Future Dr. K. Wolstencroft University of Manchester.
Cyberinfrastructure What is it? Russ Hobby Internet2 Joint Techs, 18 July 2007.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Middleware Camp NMI (NSF Middleware Initiative) Program Director Alan Blatecky Advanced Networking Infrastructure and Research.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Internet2 AdvCollab Apps 1 Access Grid Vision To create virtual spaces where distributed people can work together. Challenges:
Globus online Software-as-a-Service for Research Data Management Steve Tuecke Deputy Director, Computation Institute University of Chicago & Argonne National.
Applications and Requirements for Scientific Workflow May NSF Geoffrey Fox Indiana University.
Cyberinfrastructure Overview Russ Hobby, Internet2 ECSU CI Days 4 January 2008.
Cyberinfrastructure: Many Things to Many People Russ Hobby Program Manager Internet2.
December 10, 2003Slide 1 International Networking and Cyberinfrastructure Douglas Gatchell Program Director International Networking National Science Foundation,
Internet2 Applications Group: Renater Group Presentation T. Charles Yun Internet2 Program Manager, Applications Group 30 October 2001.
Fire Emissions Network Sept. 4, 2002 A white paper for the development of a NSF Digital Government Program proposal Stefan Falke Washington University.
2005 GRIDS Community Workshop1 Learning From Cyberinfrastructure Initiatives Grid Research Integration Development & Support
Securing the Grid & other Middleware Challenges Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer.
U.S. Grid Projects and Involvement in EGEE Ian Foster Argonne National Laboratory University of Chicago EGEE-LHC Town Meeting,
NSF Middleware Initiative Purpose To design, develop, deploy and support a set of reusable, expandable set of middleware functions and services that benefit.
Data and storage services on the NGS.
© Copyright AARNet Pty Ltd PRAGMA Update & some personal observations James Sankar Network Engineer - Middleware.
Internet2 Spring Meeting NSF Middleware Initiative Purpose To design, develop, deploy and support a set of reusable, expandable set of middleware functions.
Realizing the Promise of Grid Computing Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer Science.
All Hands Meeting 2005 BIRN-CC: Building, Maintaining and Maturing a National Information Infrastructure to Enable and Advance Biomedical Research.
Northwest Indiana Computational Grid Preston Smith Rosen Center for Advanced Computing Purdue University - West Lafayette West Lafayette Calumet.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
Toward High Breakthrough Collaboration (HBC) Susan Turnbull Program Manager Advanced Scientific Computing Research March 4, 2009.
Katy Wolstencroft University of Manchester
Bob Jones EGEE Technical Director
RDA US Science workshop Arlington VA, Aug 2014 Cees de Laat with many slides from Ed Seidel/Rob Pennington.
Grid Introduction and Overview
The Globus Toolkit™: Information Services
Grid Application Model and Design and Implementation of Grid Services
Presentation transcript:

The Grid as Future Scientific Infrastructure Ian Foster Argonne National Laboratory University of Chicago Globus Alliance

2 ARGONNE  CHICAGO Overview l Evolution of scientific infrastructure u Technology drivers & new requirements l Building cyberinfrastructure u Standards, software, and facilities u Examples l Futures u Virtual data u Discipline-specific cyberinfrastructures u Challenges

3 ARGONNE  CHICAGO Technology Drivers l Internet revolution: 100M+ hosts u Collaboration & sharing the norm l Universal Moore’s law: x10 3 /10 yrs u Sensors as well as computers l Petascale data tsunami u Gating step is analysis l & our old infrastructure? 114 genomes 735 in progress You are here

4 ARGONNE  CHICAGO If We Can’t Do it All in our Lab … l Then we need to be able to reach out into the network to obtain needed resources u Clusters, archives, instrumentation l And integrate those resources & services into our data processing workflows u Operating at scale (not Web browsing) l Which means we need u Community facilities u Standard means of accessing those facilities u Means of coordinating across facilities u New methodologies and tools

5 ARGONNE  CHICAGO Earth Simulator Atmospheric Chemistry Group LHC Exp. Astronomy Grav. Wave Nuclear Exp. Current accelerator Exp. Scale Metrics: Participants, Data, Tasks, Performance, Interactions, …

6 ARGONNE  CHICAGO l Resources u Computing, storage, data l Communities u Operational procedures, … Cyberinfrastructure—AKA Grid A A A l Services u Authentication, discovery, … l Connectivity u Reduce tyranny of distance l Technologies u Build services & applications

7 ARGONNE  CHICAGO Overview l Evolution of scientific infrastructure u Technology drivers & new requirements l Building cyberinfrastructure u Standards, software, and facilities u Examples l Futures u Virtual data u Discipline-specific cyberinfrastructures u Challenges

8 ARGONNE  CHICAGO Building Cyberinfrastructure l Open standards: absolutely vital u Define how we describe, discover, access, secure, monitor, & manage services u E.g., Open Grid Services Architecture l Open facilities: enabled by standards u Standards-based access, community services (authentication, registry), operational support u E.g., TeraGrid, Grid3, NEESgrid, ESG, DOE Science Grid, LHC Computing Grid, NASA IPG l Open software: accelerate standards adoption u Accelerates creation of facilities & applications u Globus Toolkit, NSF Middleware Initiative, etc.

9 ARGONNE  CHICAGO Increased functionality, standardization Custom solutions Open Grid Services Arch Real standards Multiple implementations Web services, etc. Managed shared virtual systems Computer science research Globus Toolkit Defacto standard Single implementation Internet standards The Emergence of Open Grid Standards 2010

10 ARGONNE  CHICAGO Grid-Web Services Convergence Web Services Messaging, Security, Etc. Open Grid Services Infrastructure Domain-Specific Services Core Services Program Execution Data Services

11 ARGONNE  CHICAGO Grid-Web Services Convergence Completed: A Major Milestone! Web Services Messaging, Security, Etc. Open Grid Services Infrastructure Domain-Specific Services Core Services Program Execution Data Services WS-Resource Framework

Deploying Cyberinfrastructure

13 ARGONNE  CHICAGO

14 ARGONNE  CHICAGO NEESgrid Earthquake Engineering Collaboratory U.Nevada Reno

15 ARGONNE  CHICAGO NEESgrid Multisite Online Simulation Test (July 2003) Illinois Colorado Illinois (simulation) W

16 ARGONNE  CHICAGO Earth System Grid (ESG) Goal: address technical obstacles to the sharing & analysis of high-volume data from advanced earth system models

17 ARGONNE  CHICAGO Grid2003: Towards a Persistent U.S. Open Science Grid Status on 11/19/03 ( P

18 ARGONNE  CHICAGO Virtual Observatories No. & sizes of data sets as of mid-2002, grouped by wavelength 12 waveband coverage of large areas of the sky Total about 200 TB data Doubling every 12 months Largest catalogues near 1B objects Data and images courtesy Alex Szalay, John Hopkins

19 ARGONNE  CHICAGO Resource Center (Processors, disks) Grid server Nodes Resource Center Resource Center Resource Center Operations Center Regional Support Center (Support for Applications Local Resources) Regional Support Regional Support Regional Support EGEE: Enabling Grids for E-Science in Europe

20 ARGONNE  CHICAGO Overview l Evolution of scientific infrastructure u Technology drivers & new requirements l Building cyberinfrastructure u Standards, software, and facilities u Examples l Futures u Virtual data u Discipline-specific cyberinfrastructures u Challenges

21 ARGONNE  CHICAGO Science as Workflow: Integrating & Evolving Data and Computation “Science advances through tentative answers to a series of more & more subtle questions which reach deeper & deeper into the essence of natural phenomena.” (L. Pasteur) l In a networked petascale world, this means leveraging distributed data & computation to u Discover relevant data/hypotheses u Derive new tentative data/hypotheses from old u Publish new data/hypotheses for use by others l In a collaborative framework in which many can contribute to, & draw from, the whole

Integrating Across Different Types of Information ID MURA_BACSU STANDARD; PRT; 429 AA. DE PROBABLE UDP-N-ACETYLGLUCOSAMINE 1-CARBOXYVINYLTRANSFERASE DE (EC ) (ENOYLPYRUVATE TRANSFERASE) (UDP-N-ACETYLGLUCOSAMINE DE ENOLPYRUVYL TRANSFERASE) (EPT). GN MURA OR MURZ. OS BACILLUS SUBTILIS. OC BACTERIA; FIRMICUTES; BACILLUS/CLOSTRIDIUM GROUP; BACILLACEAE; OC BACILLUS. KW PEPTIDOGLYCAN SYNTHESIS; CELL WALL; TRANSFERASE. FT ACT_SITE BINDS PEP (BY SIMILARITY). FT CONFLICT S -> A (IN REF. 3). SQ SEQUENCE 429 AA; MW; 02018C5C CRC32; MEKLNIAGGD SLNGTVHISG AKNSAVALIP ATILANSEVT IEGLPEISDI ETLRDLLKEI GGNVHFENGE MVVDPTSMIS MPLPNGKVKK LRASYYLMGA MLGRFKQAVI GLPGGCHLGP RPIDQHIKGF EALGAEVTNE QGAIYLRAER LRGARIYLDV VSVGATINIM LAAVLAEGKT IIENAAKEPE IIDVATLLTS MGAKIKGAGT NVIRIDGVKE LHGCKHTIIP DRIEAGTFMI [source: GlaxoSmithKline via Carole Goble]

Virtual Data Research: Sharing Recipes As Well as Data

24 ARGONNE  CHICAGO Discipline Cyberinfrastructure: E.g., Life Sciences l Leverage emerging “eScience infrastructure” u Storage, data archives, networks, cluster l Create an extensible service-rich environment u Workflows (standard protocols) u Experimental facilities: MRI, EM, arrays, … u Data update and access services u Computational services: BLAST, etc., etc. u Tools (informatics) suites u Ontologies & conceptual discovery frameworks l Bio-applns-neutral: genomics, proteomics, …

25 ARGONNE  CHICAGO Other Observations l Building & applying cyberinstructure is hard u Need sustained effort and commitment l Not clear that current funding models apply u Must support a critical mass of experts to build & operate infrastructure u Software is a vital part of the whole l Discipline sciences must obtain experience u Start multidisciplinary exploratory projects l International cooperation is vital u Our partners “get it” and are hard at work

26 ARGONNE  CHICAGO For More Information l The Globus Alliance u l Global Grid Forum u l GriPhyN u l Background information u l GlobusWorld, Feb 2005 u 2nd Edition: November 2003