1 Open Science Grid.. An introduction Ruth Pordes Fermilab.

Slides:



Advertisements
Similar presentations
Dec 14, 20061/10 VO Services Project – Status Report Gabriele Garzoglio VO Services Project WBS Dec 14, 2006 OSG Executive Board Meeting Gabriele Garzoglio.
Advertisements

 Contributing >30% of throughput to ATLAS and CMS in Worldwide LHC Computing Grid  Reliant on production and advanced networking from ESNET, LHCNET and.
1 IWLSC, Kolkata India 2006 Jérôme Lauret for the Open Science Grid consortium The Open-Science-Grid: Building a US based Grid infrastructure for Open.
May 9, 2008 Reorganization of the OSG Project The existing project organization chart was put in place at the beginning of It has worked very well.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Open Science Grid June 28, 2006 Bill Kramer Chair of the Open Science Grid Council NERSC Center General Manager, LBNL.
Open Science Grid Use of PKI: Wishing it was easy A brief and incomplete introduction. Doug Olson, LBNL PKI Workshop, NIST 5 April 2006.
R. Pordes, I Brazilian LHC Computing Workshop 1 What is Open Science Grid?  High Throughput Distributed Facility  Shared opportunistic access to existing.
Jan 2010 Current OSG Efforts and Status, Grid Deployment Board, Jan 12 th 2010 OSG has weekly Operations and Production Meetings including US ATLAS and.
Open Science Grid Frank Würthwein UCSD. 2/13/2006 GGF 2 “Airplane view” of the OSG  High Throughput Computing — Opportunistic scavenging on cheap hardware.
April 2009 OSG Grid School - RDU 1 Open Science Grid John McGee – Renaissance Computing Institute University of North Carolina, Chapel.
Assessment of Core Services provided to USLHC by OSG.
Open Science Ruth Pordes Fermilab, July 17th 2006 What is OSG Where Networking fits Middleware Security Networking & OSG Outline.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
Key Project Drivers - FY11 Ruth Pordes, June 15th 2010.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
1 Open Science Grid.. An introduction Ruth Pordes Fermilab.
May 8, 20071/15 VO Services Project – Status Report Gabriele Garzoglio VO Services Project – Status Report Overview and Plans May 8, 2007 Computing Division,
Mar 28, 20071/9 VO Services Project Gabriele Garzoglio The VO Services Project Don Petravick for Gabriele Garzoglio Computing Division, Fermilab ISGC 2007.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks EGEE – paving the way for a sustainable infrastructure.
Key Project Drivers - an Update Ruth Pordes, June 14th 2008, V2: June 23 rd. These slides are in addition to the information available in
Interoperability Grids, Clouds and Collaboratories Ruth Pordes Executive Director Open Science Grid, Fermilab.
Data Intensive Science Network (DISUN). DISUN Started in May sites: Caltech University of California at San Diego University of Florida University.
Partnerships & Interoperability - SciDAC Centers, Campus Grids, TeraGrid, EGEE, NorduGrid,DISUN Ruth Pordes Fermilab Open Science Grid Joint Oversight.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
The Open Science Grid OSG Ruth Pordes Fermilab. 2 What is OSG? A Consortium of people working together to Interface Farms and Storage to a Grid and Researchers.
Job and Data Accounting on the Open Science Grid Ruth Pordes, Fermilab with thanks to Brian Bockelman, Philippe Canal, Chris Green, Rob Quick.
Open Science Grid An Update and Its Principles Ruth Pordes Fermilab.
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
Middleware Camp NMI (NSF Middleware Initiative) Program Director Alan Blatecky Advanced Networking Infrastructure and Research.
OSG Consortium Meeting (January 23, 2006)Paul Avery1 University of Florida Open Science Grid Progress Linking Universities and Laboratories.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
1 Open Science Grid.. An introduction Ruth Pordes Fermilab.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Open Science Grid & its Security Technical Group ESCC22 Jul 2004 Bob Cowles
Status Organization Overview of Program of Work Education, Training It’s the People who make it happen & make it Work.
April 26, Executive Director Report Executive Board 4/26/07 Things under control Things out of control.
June 15, PMG Ruth Pordes Status Report US CMS PMG July 15th Tier-1 –LCG Service Challenge 3 (SC3) –FY05 hardware delivery –UAF support Grid Services.
Sep 25, 20071/5 Grid Services Activities on Security Gabriele Garzoglio Grid Services Activities on Security Gabriele Garzoglio Computing Division, Fermilab.
Open Science Grid in the U.S. Vicky White, Fermilab U.S. GDB Representative.
OSG Report for DOE/NSF Joint Oversight Group U.S. Large Hadron Collider Program OSG Report for DOE/NSF Joint Oversight Group U.S. Large Hadron Collider.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
1 An update on the Open Science Grid for IHEPCCC Ruth Pordes, Fermilab.
Towards deploying a production interoperable Grid Infrastructure in the U.S. Vicky White U.S. Representative to GDB.
Victoria A. White Head, Computing Division, Fermilab Fermilab Grid Computing – CDF, D0 and more..
All Hands Meeting 2005 BIRN-CC: Building, Maintaining and Maturing a National Information Infrastructure to Enable and Advance Biomedical Research.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
Summary of OSG Activities by LIGO and LSC LIGO NSF Review November 9-11, 2005 Kent Blackburn LIGO Laboratory California Institute of Technology LIGO DCC:
Ruth Pordes Executive Director University of Washingon Seattle OSG Consortium Meeting 21st August University of Washingon Seattle.
OSG Facility Miron Livny OSG Facility Coordinator and PI University of Wisconsin-Madison Open Science Grid Scientific Advisory Group Meeting June 12th.
Defining the Technical Roadmap for the NWICG – OSG Ruth Pordes Fermilab.
EGI-InSPIRE EGI-InSPIRE RI The European Grid Infrastructure Steven Newhouse Director, EGI.eu Project Director, EGI-InSPIRE 29/06/2016CoreGrid.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
1 Open Science Grid Progress & Vision Keith Chadwick, Fermilab
Grid Colombia Workshop with OSG Week 2 Startup Rob Gardner University of Chicago October 26, 2009.
What is OSG? (What does it have to do with Atlas T3s?) What is OSG? (What does it have to do with Atlas T3s?) Dan Fraser OSG Production Coordinator OSG.
Open Science Grid Interoperability
Bob Jones EGEE Technical Director
Open Science Grid Progress and Status
JRA3 Introduction Åke Edlund EGEE Security Head
Ian Bird GDB Meeting CERN 9 September 2003
Connecting the European Grid Infrastructure to Research Communities
Leigh Grundhoefer Indiana University
LHC Data Analysis using a worldwide computing grid
Open Science Grid at Condor Week
Presentation transcript:

1 Open Science Grid.. An introduction Ruth Pordes Fermilab

2 OSG Provenance PPDG GriPhyN iVDGL TrilliumGrid3 OSG (DOE) (DOE+NSF) (NSF)

3 Introducing myself at Fermilab for 25 years (well and 2 years in the “pioneer” ‘70s), started on data acquisition for High Energy Physics experiments, a “builder” of the Sloan Digital Sky Survey, led development of a common data acquisition system for 6 experiments at Fermilab (DART), coordinator of the CDF/D0 Joint Run II offline projects (with Dane), coordinator of the Particle Physics Data Grid SciDAC I collaboratory, founder of Trillium collaboration of iVDGL, GridPhyN, PPDG, and GLUE interoperability between US and EU. Now I am variously: Executive Director of the Open Science Grid, an Associate Head of the Computing Division at Fermilab, and US CMS Grid Services and Interfaces Coordinator.

4 A Common Grid Infrastructure

5 Overlaid by community computational environments of single to large groups of researchers located locally to worldwide

6 Grid of Grids - from Local to Global Community Campus National

7 Current OSG deployment 96 Resources across production & integration infrastructures 20 Virtual Organizations +6 operations Includes 25% non-physics. ~20,000 CPUs (from 30 to 4000 shared between OSG and local use) ~6 PB Tapes ~4 PB Shared Disk Jobs Running on OSG over 9 months Sustaining through OSG submissions: 3,000-4,000 simultaneous jobs. ~10K jobs/day ~50K CPUhours/day. Peak short validation jobs ~15K Using production & research networks

8 Examples of Sharing SiteMax # Jobs ASGC_OSG9 BU_ATLAS_Tier2154 CIT_CMS_T299 FIU-PG58 FNAL_GPFARM17 OSG_LIGO_PSU1 OU_OCHEP_SWT282 Purdue-ITaP3 UC_ATLAS_MWT288 UFlorida-IHEPA1 UFlorida-PG (CMS)1 UMATLAS UWMadisonCMS594 UWMilwaukee2 osg-gw-2.t2.ucsd.edu2 CPUHours55,000 VOMax # Jobs ATLAS 2 CDF279 CMS559 COMPBIOGRID10 GADU1 LIGO75 Average # of Jobs (~300 batch slots) 253 CPUHours30,000 #Jobs Completed50,000 last week at UCSD -- CMS Site last week of ATLAS

9 OSG Consortium Contributors Project

10 OSG Project

11 OSG & its goals Project receiving ~$6/M/Year for 5 years from DOE and NSF for effort to sustain and evolve the distributed facility, bring on board new communities and capabilities and EOT. Hardware resources contributed by OSG Consortium members. Goals: Support data storage, distribution & computation for High Energy, Nuclear & Astro Physics collaborations, in particular delivering to the needs of LHC and LIGO science. Engage and benefit other Research & Science of all scales through progressively supporting their applications. Educate & train students, administrators & educators. Provide a petascale Distributed Facility across the US with guaranteed & opportunistic access to shared compute & storage resources. Interface, Federate and Collaborate with Campus, Regional, other national & international Grids, in particular with EGEE & TeraGrid. Provide an Integrated, Robust Software Stack for Facility & Applications, tested on a well provisioned at-scale validation facility. Evolve the capabilities by deploying externally developed new technologies through joint projects with the development groups.

12 Middleware Stack and Deployment OSG Middleware is deployed on existing farms and storage systems. OSG Middleware interfaces to the existing installations of OS, utilities and batch systems. VOs have VO scoped environments in which they deploy applications (and other files), execute code and store data. VOs are responsible for and have control over their end-to-end distributed system using the OSG infrastructure. End-to-end s/w StackDeployment into Production Integration Grid has ~15 sites

13 OSG will support Global Data Transfer, Storage & Access at GBytes/sec 365 days a year e.g. CMS Data To / From Tape at Tier-1 Need to triple in ~1 year. Data To / From Tape at Tier-1 Need to triple in ~1 year. Data to Disk Caches - Data Samples 200MB/sec 600MB/sec Tier-2 sites data distributed to ~7 Tier-1s, CERN + Tier-2s OSG must enable data placement, disk usage, resource managament policies, of 10s Gbit/Sec data movement, 10s Petabyte tape stores, local shared disk caches of 100sTBs across 10s of sites for >10 VOs. Data distribution will depend on & integrate to advanced network infrastructures:  Internet 2 will provide "layer 2” connectivity between OSG University Sites and peers in Europe.  ESNET will provide "layer 2" connectivity between OSG DOE Laboratory sites and EU GEANT network.  Both include the use of the IRNC link (NSF) from the US to Amsterdam

14 Security Infrastructure Identity: X509 Certificates. Authentication and Authorization using VOMS extended attribute certficates. Security Process modelled on NIST procedural controls - management, operational, technical, starting from an inventory of the OSG assets. User and VO Management:  VO Registers with Operations Center  User registers through VOMRS or VO administrator  Site Registers with the Operations Center  Each VO centrally defines and assigns roles  Each Site provides role to access mappings based on VO/VOGroup.  Can reject individuals. Heterogeneous identity management systems – OSG vs TeraGrid/EGEE, grid vs. local, compute vs. storage, head-node vs., old-version vs. new-version. Issues include:  Cross domain right management  Right/identity management of software modules and resources  Error/rejection propagation  Solutions/approaches that work end-to-end

15 Education, Outreach, Training Training Workshops - for Administrators and Application Developers e.g. Grid Summer Workshop (in 4th year) Outreach - e.g. Science Grid This Week -> International Science Grid This Week Education through e-Labs

16 Integrated Network Management OSG Initial Timeline & Milestones - Summary LHC Simulations Support 1000 Users; 20PB Data Archive Contribute to Worldwide LHC Computing Grid LHC Event Data Distribution and Analysis Contribute to LIGO Workflow and Data Analysis +1 Community Additional Science Communities +1 Community Facility Security : Risk Assessment, Audits, Incident Response, Management, Operations, Technical Controls Plan V11st AuditRisk Assessment AuditRisk Assessment AuditRisk Assessment AuditRisk Assessment VDT and OSG Software Releases: Major Release every 6 months; Minor Updates as needed VDT 1.4.0VDT 1.4.1VDT 1.4.2………… Advanced LIGO LIGO Data Grid dependent on OSG CDF Simulation STAR, CDF, D0, Astrophysics D0 Reprocessing STAR Data Distribution and Jobs 10KJobs per Day D0 Simulations CDF Simulation and Analysis LIGO data run SC5 Facility Operations and Metrics: Increase robustness and scale; Operational Metrics defined and validated each year. Interoperate and Federate with Campus and Regional Grids Project startEnd of Phase I End of Phase II VDT Incremental Updates dCache with role based authorization OSG 0.6.0OSG 0.8.0OSG 1.0OSG 2.0OSG 3.0… AccountingAuditing VDS with SRM Common S/w Distribution with TeraGrid EGEE using VDT 1.4.X Transparent data and job movement with TeraGrid Transparent data management with EGEE Federated monitoring and information services Data Analysis (batch and interactive) Workflow Extended Capabilities & Increase Scalability and Performance for Jobs and Data to meet Stakeholder needs SRM/dCache Extensions “Just in Time” Workload Management VO Services Infrastructure Improved Workflow and Resource Selection Work with SciDAC-2 CEDS and Security with Open Science +1 Community Community