Empowering Distributed Science Ian Foster Argonne National Laboratory University of Chicago Globus Alliance eScience [n]: Large-scale science carried out.

Slides:



Advertisements
Similar presentations
The Access Grid Ivan R. Judson 5/25/2004.
Advertisements

GridWorld 2006 Use of MyProxy for the FusionGrid Mary Thompson Monte Goode GridWorld 2006.
ASCR Data Science Centers Infrastructure Demonstration S. Canon, N. Desai, M. Ernst, K. Kleese-Van Dam, G. Shipman, B. Tierney.
High Performance Computing Course Notes Grid Computing.
Data Grids Darshan R. Kapadia Gregor von Laszewski
1 Cyberinfrastructure Framework for 21st Century Science & Engineering (CF21) IRNC Kick-Off Workshop July 13,
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
The DOE Science Grid Computing and Data Infrastructure for Large-Scale Science William Johnston, Lawrence Berkeley National Lab Ray Bair, Pacific Northwest.
Where do we go from here? “Knowledge Environments to Support Distributed Science and Engineering” Symposium on Knowledge Environments for Science and Engineering.
The LHC Computing Grid Project Tomi Kauppi Timo Larjo.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
Computing and Data Infrastructure for Large-Scale Science Deploying Production Grids: NASA’s IPG and DOE’s Science Grid William E. Johnston
Sergey Belov, LIT JINR 15 September, NEC’2011, Varna, Bulgaria.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
4b.1 Grid Computing Software Components of Globus 4.0 ITCS 4010 Grid Computing, 2005, UNC-Charlotte, B. Wilkinson, slides 4b.
Grid Services at NERSC Shreyas Cholia Open Software and Programming Group, NERSC NERSC User Group Meeting September 17, 2007.
Knowledge Environments for Science: Representative Projects Ian Foster Argonne National Laboratory University of Chicago
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
Computing in Atmospheric Sciences Workshop: 2003 Challenges of Cyberinfrastructure Alan Blatecky Executive Director San Diego Supercomputer Center.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
Presented by The Earth System Grid: Turning Climate Datasets into Community Resources David E. Bernholdt, ORNL on behalf of the Earth System Grid team.
GT Components. Globus Toolkit A “toolkit” of services and packages for creating the basic grid computing infrastructure Higher level tools added to this.
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
What is Internet2? Ted Hanss, Internet2 5 March
What is Cyberinfrastructure? Russ Hobby, Internet2 Clemson University CI Days 20 May 2008.
Service Oriented Science Ian Foster Argonne National Laboratory University of Chicago Univa Corporation.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
Supporting further and higher education The Akenti Authorisation System Alan Robiette, JISC Development Group.
Introduction to Grid Computing Ed Seidel Max Planck Institute for Gravitational Physics
1 Abstract  The term "Grid" tends to generate enthusiasm, skepticism, or perplexity. Enthusiasts speak of the potential for integrating services and resources.
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
1 ARGONNE  CHICAGO Grid Introduction and Overview Ian Foster Argonne National Lab University of Chicago Globus Project
Authors: Ronnie Julio Cole David
NSDL Collections Based on DOE User Facilities Christopher Klaus 10/05/03.
The Earth System Grid (ESG) Computer Science and Technologies DOE SciDAC ESG Project Review Argonne National Laboratory, Illinois May 8-9, 2003.
Building the e-Minerals Minigrid Rik Tyer, Lisa Blanshard, Kerstin Kleese (Data Management Group) Rob Allan, Andrew Richards (Grid Technology Group)
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
GRIDS Center Middleware Overview Sandra Redman Information Technology and Systems Center and Information Technology Research Center National Space Science.
Cyberinfrastructure What is it? Russ Hobby Internet2 Joint Techs, 18 July 2007.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Middleware Camp NMI (NSF Middleware Initiative) Program Director Alan Blatecky Advanced Networking Infrastructure and Research.
GridFTP GUI: An Easy and Efficient Way to Transfer Data in Grid
ESFRI & e-Infrastructure Collaborations, EGEE’09 Krzysztof Wrona September 21 st, 2009 European XFEL.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
May 6, 2002Earth System Grid - Williams The Earth System Grid Presented by Dean N. Williams PI’s: Ian Foster (ANL); Don Middleton (NCAR); and Dean Williams.
Leveraging the InCommon Federation to access the NSF TeraGrid Jim Basney Senior Research Scientist National Center for Supercomputing Applications University.
Introduction to Grids By: Fetahi Z. Wuhib [CSD2004-Team19]
Internet2 AdvCollab Apps 1 Access Grid Vision To create virtual spaces where distributed people can work together. Challenges:
Globus online Software-as-a-Service for Research Data Management Steve Tuecke Deputy Director, Computation Institute University of Chicago & Argonne National.
Cyberinfrastructure Overview Russ Hobby, Internet2 ECSU CI Days 4 January 2008.
U.S. Department of Energy’s Office of Science Dr. Raymond L. Orbach Director, Office of Science April 29, 2004 PRESENTATION FOR THE BIOLOGICAL AND ENVIRONMENTAL.
Securing the Grid & other Middleware Challenges Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer.
U.S. Grid Projects and Involvement in EGEE Ian Foster Argonne National Laboratory University of Chicago EGEE-LHC Town Meeting,
By David P. Schissel and Reza Shakoori Presented at DOE Office of Science High-Performance Network Research PI Meeting Brookhaven National Lab September.
© Copyright AARNet Pty Ltd PRAGMA Update & some personal observations James Sankar Network Engineer - Middleware.
An Architectural Approach to Managing Data in Transit Micah Beck Director & Associate Professor Logistical Computing and Internetworking Lab Computer Science.
The National Grid Service Mike Mineter.
Holding slide prior to starting show. Lessons Learned from the GECEM Portal David Walker Cardiff University
Realizing the Promise of Grid Computing Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer Science.
PARALLEL AND DISTRIBUTED PROGRAMMING MODELS U. Jhashuva 1 Asst. Prof Dept. of CSE om.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
The Helmholtz Association Project „Large Scale Data Management and Analysis“ (LSDMA) Kilian Schwarz, GSI; Christopher Jung, KIT.
Presented by SciDAC-2 Petascale Data Storage Institute Philip C. Roth Computer Science and Mathematics Future Technologies Group.
EGI-InSPIRE EGI-InSPIRE RI The European Grid Infrastructure Steven Newhouse Director, EGI.eu Project Director, EGI-InSPIRE 29/06/2016CoreGrid.
ANL Controls Activities
Recap: introduction to e-science
Presentation transcript:

Empowering Distributed Science Ian Foster Argonne National Laboratory University of Chicago Globus Alliance eScience [n]: Large-scale science carried out through distributed collaborations—often leveraging access to large-scale data & computing

2 10 Years of DOE Collaboratories Research l Distributed Collaboratory Experimental Environments ( ) u Focus on interpersonal collaboration l DOE2000 National Collaboratories ( ) u Somewhat broader focus l Next Generation Internet program (1998) u A broad set of application & technology projects l SciDAC Collaboratory program ( ) u Technology R&D u PPDG, ESG, CMCS, Fusion; DOE Science Grid See:

3 It’s Amazing How Much We Have Achieved in 10 Years l Applications u Production services: Grid3, ESG, Fusion, CMCS; also NEESgrid, many others that use DOE tech) l Infrastructure u Broadly deployed PKI and single sign on u Access Grid at 300+ institutions worldwide l Leadership and technology u Grid concepts & software used worldwide u Global Grid Forum: standards & community u GridFTP: California  Illinois at 27 Gbit/s l Multicast almost works

4 There’s Still Much to Do: Where We Should Be vs. Where We Are l Goal: Any DOE scientist can access any DOE computer, software, data, instrument u ~25,000 scientists* (vs. ~1000 DOE certs) u ~1000 instruments** (vs. maybe 10 online?) u ~1000 scientific applns** (vs. 2 Fusion services) u ~10 PB of interesting data** (vs. 100TB on ESG) u ~100,000 computers* (vs. ~3000 on Grid3) l Not to mention many external partners I.e., we need to scale by 2-3 orders of magnitude to have DOE-wide impact! * Rough estimate; ** WAG

5 “25,000 Scientists”: The Many Aspects of Scaling l Data & computational services integrated into the fabric of science communities u Used not by a handful but by 1000s u Part of everyday science workflows  Scale load on services by factors of 100+ u 100,000 requests annually to fusion codes u 1000 concurrent users for ESG services u 25,000 users to authenticate & authorize l Manageability as a key new challenge u Resource management and provisioning u Automation of management functions

6 “25,000 Scientists”: Authentication & Authorization l User-managed PKI credentials l Single sign on & delegation (GSI) l DOEGrids CA: 1250 users l MyProxy & related tools l WS-Security & SAML-based authentication/authorization

7 Authentication & Authorization: Next Steps l Integration with campus infrastructures u “Authenticate locally, act globally” u E.g., KX509, GridLogon, GridShib, etc. l Enabling access while enhancing security u Create secure virtual national laboratories u Tech & policy solns to risk/benefit tradeoffs l Evolving what we mean by “trust” u Colleagues  collaboration  community l Scaling to the ultrascale u Data volumes, data rates, transaction rates

8 “1000 Instruments”: The Scale of the Problem Lawrence Berkeley National Lab Advanced Light Source National Center for Electron Microscopy National Energy Research Scientific Computing Facility Los Alamos Neutron Science Center Univ. of IL Electron Microscopy Center for Materials Research Center for Microanalysis of Materials MIT Bates Accelerator Center Plasma Science & Fusion Center SC User Facilities Institutions that Use SC Facilities Fermi National Accelerator Lab Tevatron Stanford Linear Accelerator Center B-Factory Stanford Synchrotron Radiation Laboratory Princeton Plasma Physics Lab General Atomics - DIII-D Tokamak SC Laboratories Pacific Northwest National Lab Environmental Molecular Sciences Lab Argonne National Lab Intense Pulsed Neutron Source Advanced Photon Source Argonne Tandem Linac Accelerator System Brookhaven National Lab Relativistic Heavy Ion Collider National Synchrotron Light Source Oak Ridge National Lab High-Flux Isotope Reactor Surface Modification & Characterization Center Spallation Neutron Source (under construction) Thomas Jefferson National Accelerator Facility Continuous Electron Beam Accelerator Facility Physics Accelerators Synchrotron Light Sources Neutron Sources Special Purpose Facilities Large Fusion Experiments Sandia Combustion Research Facility James R. MacDonald Laboratory

9 For Example: NSF Network for Earthquake Engineering Simulation Links instruments, data, computers, people

10 Resources implement standard access & management interfaces Collective services aggregate &/or virtualize resources Users work with client applications Application services organize VOs & enable access to other services NEESgrid: How it Really Happens (A Simplified View) Web Browser Compute Server Globus MCS/RLS Data Viewer Tool Certificate Authority CHEF Chat Teamlet MyProxy CHEF Compute Server Database service Database service Database service Simulation Tool Camera Telepresence Monitor Globus Index Service Globus GRAM OGSA DAI Application Developer 2 Off the Shelf 9 Globus Toolkit 5 Grid Community 3

11 Scaling to 1000 Instruments: Challenges l Common teleoperation control interfaces u NEESgrid Network Telecontrol Protocol (NTCP) provides service-oriented interface: nice start? l Major social & organizational challenges u Operating instruments as shared facilities u Data sharing policies and mechanisms l Basic technological challenges also u Provisioning/QoS for multi-modal experiments u Hierarchical/latency tolerant control algorithms u Reliability, health, and safety

12 “1000 Applications”: Software as Service l Software is increasingly central to almost every aspect of DOE science l Service interfaces are needed for broad adoption: “shrink wrap” isn’t the answer TransP production service: 1662 runs in FY03

13 Software as Service: What If You Have 1000s of Users? l Service-oriented applications u Wrapping applications as Web services u Composing applications into workflows l Service-oriented infrastructure u Provisioning physical resources to support application workloads Appln Service Users Workflows Composition Invocation Provisioning

14 “10 PB Data”: Distributed Data Integration l Major challenges in four dimensions u Number & distribution of data sources u Volume of data u Diversity in data format, quality, semantics u Sophistication & scale of data analysis Experiments & Instruments Simulations facts answers questions ? Literature Other Archives facts

15 Distributed Data Integration: Examples of Where We Are Today Earth System Grid: O(100TB) online data STAR: 5 TB transfer (SRM, GridFTP) NASA/NVO: Mosaics from multiple sources Bertram Ludäscher’s examples

16 Distributed Data Integration: Enabling Automated Analysis l Data ingest l Managing many petabytes l Common schema and ontologies l How to organize petabytes? Reorganize it? l Interactive & batch analysis performance l Universal annotation infrastructure l Query, analysis, visualization tools

17 Tomorrow? “100,000 Computers”: A Healthy Computing Pyramid Supercomputer Cluster Desktop Today Supercomputers USE SPARINGLY Desktop 100,000 SERVINGS Specialized computers 2-3 SERVINGS Clusters 100s of SERVINGS

18 Grid2003: An Operational Grid  28 sites ( CPUs) & growing  concurrent jobs  8 substantial applications + CS experiments  Running since October 2003 Korea

19 Example Grid2003 Workflows Genome sequence analysis Physics data analysis Sloan digital sky survey

20 Example Grid2003 Application: NVO Mosaic Construction NVO/NASA Montage: A small (1200 node) workflow Construct custom mosaics on demand from multiple data sources User specifies projection, coordinates, size, rotation, spatial sampling Work by Ewa Deelman et al., USC/ISI and Caltech

21 Invocation Provenance Completion status and resource usage Attributes of executable transformation Attributes of input and output files

22 “100,000 Computers”: Future Challenges l New modes of working that are driven by (& drive) massive increases in computing u Enabling massive data analysis, simulation- driven problem solving, application services u These make massively parallel computing essential, not an academic curiousity l More pesky security & policy challenges l Technological challenges u Reliability, performance, usability as infrastructure, workflows, data volumes, user community scale by 2+ orders of magnitude u Manageability again

23 Cross-cutting Challenges l Institutionalize infrastructure u Broad deployment & support at sites u Software as infrastructure u Legitimate (& challenging) security concerns l Expand range of resource sharing modalities u Research aimed at federating not just data & computers, but workflow and semantics u Scale data size, community sizes, etc., etc. l Reach new application domains u Sustain current collaboratory pilots, and start new ones of similar or greater ambition

24 Summary: It’s Amazing How Much We Have Achieved in 10 Years l Applications u Production services: Grid3, ESG, Fusion, CMCS; also NEESgrid, many others that use DOE tech) l Infrastructure u Broadly deployed PKI and single sign on u Access Grid at 300+ institutions worldwide l Leadership and technology u Grid concepts & software used worldwide u Global Grid Forum: standards & community u GridFTP: California  Illinois at 27 Gbit/s l Multicast almost works

25 But Over Those Same 10 Years: Dramatic Change l Exponential growth in network speed, data volume, computer speed, collaboration size u E.g., 155 Mb/s  10 Gb/s (ESnet backbone)  eScience methods no longer optional but now vital to scientific competitiveness l We’ve demonstrated feasibility of eScience, but we are far from DOE-wide adoption  We have moved forward, but we’ve also fallen behind

26 The $3.4B Question l Future science will be dominated by “eScience” l Europe is investing heavily in eScience u EU: ~$70M/yr for “Grid” infrastructure, tech u UK: ~$60M/yr for eScience apps and tech u German, Italian, Dutch, etc., programs l Asia Pacific is investing heavily in eScience u Japan, China, South Korea, Singapore, Australia all have programs  How does DOE stay competitive?

27 We Have Done Much, But Have Much More to Do l Any DOE scientist can access any DOE computer, software, data, instrument u ~25,000 scientists* (vs. ~1000 DOE certs) u ~1000 instruments** (vs. maybe 10 online?) u ~1000 scientific applns** (vs. 2 Fusion services) u ~10 PB of interesting data** (vs. 100TB on ESG) u ~100,000 computers* (vs. ~3000 on Grid3) l Not to mention many external partners We need to scale by 2-3 orders of magnitude to have DOE-wide impact * Rough estimate; ** WAG

28 Staff costs - Grid Resources Computers & Network funded separately EPSRC Breakdown UK e-Science Budget ( ) Source: Science Budget 2003/4 – 2005/6, DTI(OST) Total: £213M + Industrial Contributions £25M + £100M via JISC