Download presentation
Presentation is loading. Please wait.
Published byBranden Sherman Modified over 9 years ago
1
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery1 University of Florida avery@phys.ufl.edu Open Science Grid Linking Universities and Laboratories in National CyberInfrastructure SURA Infrastructure Workshop Austin, TX December 7, 2005
2
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery2 Bottom-up Collaboration: “Trillium” Trillium = PPDG + GriPhyN + iVDGL PPDG:$12M (DOE)(1999 – 2006) GriPhyN:$12M (NSF)(2000 – 2005) iVDGL:$14M (NSF)(2001 – 2006) ~150 people with large overlaps between projects Universities, labs, foreign partners Strong driver for funding agency collaborations Inter-agency: NSF – DOE Intra-agency: Directorate – Directorate, Division – Division Coordinated internally to meet broad goals CS research, developing/supporting Virtual Data Toolkit (VDT) Grid deployment, using VDT-based middleware Unified entity when collaborating internationally
3
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery3 Common Middleware: Virtual Data Toolkit Sources (CVS) Patching GPT src bundles NMI Build & Test Condor pool 22+ Op. Systems Build Test Package VDT Build Many Contributors Build Pacman cache RPMs Binaries Test A unique laboratory for testing, supporting, deploying, packaging, upgrading, & troubleshooting complex sets of software!
4
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery4 VDT Growth Over 3 Years (1.3.8 now) # of components VDT 1.0 Globus 2.0b Condor 6.3.1 VDT 1.1.7 Switch to Globus 2.2 VDT 1.1.11 Grid3 VDT 1.1.8 First real use by LCG www.griphyn.org/vdt/
5
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery5 Components of VDT 1.3.5 Globus 3.2.1 Condor 6.7.6 RLS 3.0 ClassAds 0.9.7 Replica 2.2.4 DOE/EDG CA certs ftsh 2.0.5 EDG mkgridmap EDG CRL Update GLUE Schema 1.0 VDS 1.3.5b Java Netlogger 3.2.4 Gatekeeper-Authz MyProxy1.11 KX509 System Profiler GSI OpenSSH 3.4 Monalisa 1.2.32 PyGlobus 1.0.6 MySQL UberFTP 1.11 DRM 1.2.6a VOMS 1.4.0 VOMS Admin 0.7.5 Tomcat PRIMA 0.2 Certificate Scripts Apache jClarens 0.5.3 New GridFTP Server GUMS 1.0.1
6
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery6 VDT Collaborative Relationships Computer Science Research Virtual Data Toolkit Partner science projects Partner networking projects Partner outreach projects Science, ENG, Education Communities Globus, Condor, NMI, TeraGrid, OSG EGEE, WLCG, Asia, South America QuarkNet, CHEPREO, Digital Divide Deployment, Feedback Tech Transfer Techniques & software Requirements Prototyping & experiments Other linkages Work force CS researchers Industry U.S.Grids International Outreach
7
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery7 Search for Origin of Mass New fundamental forces Supersymmetry Other new particles 2007 – ? TOTEM LHCb ALICE 27 km Tunnel in Switzerland & France CMS ATLAS Major Science Driver: Large Hadron Collider (LHC) @ CERN
8
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery8 LHC: Petascale Global Science Complexity:Millions of individual detector channels Scale:PetaOps (CPU), 100s of Petabytes (Data) Distribution:Global distribution of people & resources CMS Example- 2007 5000+ Physicists 250+ Institutes 60+ Countries BaBar/D0 Example - 2004 700+ Physicists 100+ Institutes 35+ Countries
9
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery9 CMS Experiment LHC Global Data Grid (2007+) Online System CERN Computer Center USA Korea Russia UK Maryland 150 - 1500 MB/s >10 Gb/s 10-40 Gb/s 2.5-10 Gb/s Tier 0 Tier 1 Tier 3 Tier 2 Physics caches PCs Iowa UCSDCaltech U Florida 5000 physicists, 60 countries 10s of Petabytes/yr by 2008 1000 Petabytes in < 10 yrs? FIU Tier 4
10
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery10 Grid3 and Open Science Grid
11
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery11 Grid3: A National Grid Infrastructure October 2003 – July 2005 32 sites, 3,500 CPUs: Universities + 4 national labs Sites in US, Korea, Brazil, Taiwan Applications in HEP, LIGO, SDSS, Genomics, fMRI, CS Brazil www.ivdgl.org/grid3
12
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery12 Grid3 Lessons Learned How to operate a Grid as a facility Tools, services, error recovery, procedures, docs, organization Delegation of responsibilities (Project, VO, service, site, …) Crucial role of Grid Operations Center (GOC) How to support people people relations Face-face meetings, phone cons, 1-1 interactions, mail lists, etc. How to test and validate Grid tools and applications Vital role of testbeds How to scale algorithms, software, process Some successes, but “interesting” failure modes still occur How to apply distributed cyberinfrastructure Successful production runs for several applications
13
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery13 http://www.opensciencegrid.org
14
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery14 Sao Paolo Taiwan, S.Korea Production Grid: 50+ sites, 15,000 CPUs “present” (available but not at one time) Sites in US, Korea, Brazil, Taiwan Integration Grid: 10-12 sites Open Science Grid: July 20, 2005
15
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery15 OSG Operations Snapshot November 7: 30 days
16
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery16 OSG Participating Disciplines Computer Science Condor, Globus, SRM, SRBTest and validate innovations: new services & technologies Physics LIGO, Nuclear Physics, Tevatron, LHC Global Grid: computing & data access Astrophysics Sloan Digital Sky SurveyCoAdd: multiply-scanned objects Spectral fitting analysis Bioinformatics Argonne GADU project Dartmouth Psychological & Brain Sciences BLAST, BLOCKS, gene sequences, etc Functional MRI University campus Resources, portals, apps CCR(U Buffalo) GLOW(U Wisconsin) TACC(Texas Advanced Computing Center) MGRID(U Michigan) UFGRID(U Florida) Crimson Grid(Harvard) FermiGrid(FermiLab Grid)
17
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery17 OSG Grid Partners TeraGrid “DAC2005”: run LHC apps on TeraGrid resources TG Science Portals for other applications Discussions on joint activities: Security, Accounting, Operations, Portals EGEE Joint Operations Workshops, defining mechanisms to exchange support tickets Joint Security working group US middleware federation contributions to core- middleware gLITE Worldwide LHC Computing Grid OSG contributes to LHC global data handling and analysis systems Other partners SURA, GRASE, LONI, TACC Representatives of VOs provide portals and interfaces to their user groups
18
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery18 Example of Partnership: WLCG and EGEE
19
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery19 OSG Technical Groups & Activities Technical Groups address and coordinate technical areas Propose and carry out activities related to their given areas Liaise & collaborate with other peer projects (U.S. & international) Participate in relevant standards organizations. Chairs participate in Blueprint, Integration and Deployment activities Activities are well-defined, scoped tasks contributing to OSG Each Activity has deliverables and a plan … is self-organized and operated … is overseen & sponsored by one or more Technical Groups TGs and Activities are where the real work gets done
20
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery20 OSG Technical Groups (deprecated!) GovernanceCharter, organization, by-laws, agreements, formal processes PolicyVO & site policy, authorization, priorities, privilege & access rights SecurityCommon security principles, security infrastructure Monitoring and Information Services Resource monitoring, information services, auditing, troubleshooting StorageStorage services at remote sites, interfaces, interoperability Support CentersInfrastructure and services for user support, helpdesk, trouble ticket Education / OutreachTraining, interface with various E/O projects Networks (new)Including interfacing with various networking projects
21
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery21 OSG Activities BlueprintDefining principles and best practices for OSG DeploymentDeployment of resources & services ProvisioningConnected to deployment Incidence responsePlans and procedures for responding to security incidents IntegrationTesting & validating & integrating new services and technologies Data Resource Management (DRM) Deployment of specific Storage Resource Management technology DocumentationOrganizing the documentation infrastructure AccountingAccounting and auditing use of OSG resources InteroperabilityPrimarily interoperability between OperationsOperating Grid-wide services
22
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery22 OSG Integration Testbed: Testing & Validating Middleware Brazil Taiwan Korea
23
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery23 Networks
24
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery24 Evolving Science Requirements for Networks (DOE High Performance Network Workshop) Science Areas Today End2End Throughput 5 years End2End Throughput 5-10 Years End2End Throughput Remarks High Energy Physics 0.5 Gb/s100 Gb/s 1000 Gb/s High bulk throughput Climate (Data & Computation) 0.5 Gb/s160-200 Gb/s N x 1000 Gb/s High bulk throughput SNS NanoScience Not yet started 1 Gb/s1000 Gb/s + QoS for Control Channel Remote control and time critical throughput Fusion Energy0.066 Gb/s (500 MB/s burst) 0.2 Gb/s (500MB/ 20 sec. burst) N x 1000 Gb/s Time critical throughput Astrophysics0.013 Gb/s (1 TB/week) N*N multicast 1000 Gb/s Computational steering and collaborations Genomics Data & Computation 0.091 Gb/s (1 TB/day) 100s of users1000 Gb/s + QoS for Control Channel High throughput and steering See http://www.doecollaboratory.org/meetings/hpnpw /
25
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery25 UltraLight 10 Gb/s+ network Caltech, UF, FIU, UM, MIT SLAC, FNAL Int’l partners Level(3), Cisco, NLR http://www.ultralight.org Integrating Advanced Networking in Applications
26
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery26 Education Training Communications
27
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery27 iVDGL, GriPhyN Education/Outreach Basics $200K/yr Led by UT Brownsville Workshops, portals, tutorials Partnerships with QuarkNet, CHEPREO, LIGO E/O, …
28
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery28 Grid Training Activities June 2004: First US Grid Tutorial (South Padre Island, Tx) 36 students, diverse origins and types July 2005: Second Grid Tutorial (South Padre Island, Tx) 42 students, simpler physical setup (laptops) Reaching a wider audience Lectures, exercises, video, on web Students, postdocs, scientists Coordination of training activities “Grid Cookbook” (Trauner & Yafchak) More tutorials, 3-4/year CHEPREO tutorial in 2006?
29
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery29 QuarkNet/GriPhyN e-Lab Project http://quarknet.uchicago.edu/elab/cosmic/home.jsp
30
CHEPREO: Center for High Energy Physics Research and Educational Outreach Florida International University Physics Learning Center CMS Research iVDGL Grid Activities AMPATH network (S. America) Funded September 2003 $4M initially (3 years) MPS, CISE, EHR, INT
31
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery31 Grids and the Digital Divide Background World Summit on Information Society HEP Standing Committee on Inter- regional Connectivity (SCIC) Themes Global collaborations, Grids and addressing the Digital Divide Focus on poorly connected regions Brazil (2004), Korea (2005)
32
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery32 Science Grid Communications Broad set of activities (Katie Yurkewicz) News releases, PR, etc. Science Grid This Week OSG Newsletter Not restricted to OSG www.interactions.org/sgtw
33
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery33 Grid Timeline 20002001200320042005200620072002 GriPhyN, $12M PPDG, $9.5M UltraLight, $2M CHEPREO, $4M DISUN, $10M Grid Communications Grid Summer Schools Grid3 operations OSG operations VDT 1.0 First US-LHC Grid Testbeds Digital Divide Workshops LIGO Grid Start of LHC iVDGL, $14M
34
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery34 Future of OSG CyberInfrastructure OSG is a unique national infrastructure for science Large CPU, storage and network capability crucial for science Supporting advanced middleware Long-term support of the Virtual Data Toolkit (new disciplines & international collaborations OSG currently supported by a “patchwork” of projects Collaborating projects, separately funded Developing workplan for long-term support Maturing, hardening facility Extending facility to lower barriers to participation Oct. 27 presentation to DOE and NSF
35
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery35 Sao Paolo Taiwan, S.Korea OSG Consortium Meeting: Jan 23-25 University of Florida (Gainesville) About 100 – 120 people expected Funding agency invitees Schedule Monday Morning: Applications plenary (rapporteurs) Monday Afternoon:Partner Grid projects plenary Tuesday Morning:Parallel Tuesday Afternoon:Plenary Wednesday Morning:Parallel Wednesday Afternoon:Plenary Thursday:OSG Council meeting
36
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery36 Disaster Planning Emergency Response
37
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery37 Grids and Disaster Planning / Emergency Response Inspired by recent events Dec. 2004 tsunami in Indonesia Aug. 2005 Katrina hurricane and subsequent flooding (Quite different time scales!) Connection of DP/ER to Grids Resources to simulate detailed physical & human consequences of disasters Priority pooling of resources for a societal good In principle, a resilient distributed resource Ensemble approach well suited to Grid/cluster computing E.g., given a storm’s parameters & errors, bracket likely outcomes Huge number of jobs required Embarrassingly parallel
38
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery38 DP/ER Scenarios Simulating physical scenarios Hurricanes, storm surges, floods, forest fires Pollutant dispersal: chemical, oil, biological and nuclear spills Disease epidemics Earthquakes, tsunamis Nuclear attacks Loss of network nexus points (deliberate or side effect) Astronomical impacts Simulating human responses to these situations Roadways, evacuations, availability of resources Detailed models (geography, transportation, cities, institutions) Coupling human response models to specific physical scenarios Other possibilities “Evacuation” of important data to safe storage
39
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery39 DP/ER and Grids: Some Implications DP/ER scenarios are not equally amenable to Grid approach E.g., tsunami vs hurricane-induced flooding Specialized Grids can be envisioned for very short response times But all can be simulated “offline” by researchers Other “longer term” scenarios ER is an extreme example of priority computing Priority use of IT resources is common (conferences, etc) Is ER priority computing different in principle? Other implications Requires long-term engagement with DP/ER research communities (Atmospheric, ocean, coastal ocean, social/behavioral, economic) Specific communities with specific applications to execute Digital Divide: resources to solve problems of interest to 3 rd World Forcing function for Grid standards? Legal liabilities?
40
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery40 Grid Project References Open Science Grid www.opensciencegrid.org Grid3 www.ivdgl.org/grid3 Virtual Data Toolkit www.griphyn.org/vdt GriPhyN www.griphyn.org iVDGL www.ivdgl.org PPDG www.ppdg.net CHEPREO www.chepreo.org UltraLight www.ultralight.org Globus www.globus.org Condor www.cs.wisc.edu/condor WLCG www.cern.ch/lcg EGEE www.eu-egee.org
41
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery41 Extra Slides
42
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery42 Grid3 Use by VOs Over 13 Months
43
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery43 CMS: “Compact” Muon Solenoid Inconsequential humans
44
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery44 LHC: Beyond Moore’s Law Moore’s Law (2000) LHC CPU Requirements
45
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery45 Grids and Globally Distributed Teams Non-hierarchical: Chaotic analyses + productions Superimpose significant random data flows
46
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery46 Sloan Digital Sky Survey (SDSS) Using Virtual Data in GriPhyN Galaxy cluster size distribution Sloan Data
47
SURA Infrastructure Workshop (Dec. 7, 2005)Paul Avery47 The LIGO Scientific Collaboration (LSC) and the LIGO Grid LIGO Grid: 6 US sites + 3 EU sites (Cardiff/UK, AEI/Germany) * LHO, LLO: LIGO observatory sites * LSC: LIGO Scientific Collaboration Cardiff AEI/Golm Birmingham
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.