Open Science Grid For CI-Days NYSGrid Meeting Sebastien Goasguen, John McGee, OSG Engagement Manager School of Computing.

Slides:



Advertisements
Similar presentations
PRAGMA Application (GridFMO) on OSG/FermiGrid Neha Sharma (on behalf of FermiGrid group) Fermilab Work supported by the U.S. Department of Energy under.
Advertisements

1 US activities and strategy :NSF Ron Perrott. 2 TeraGrid An instrument that delivers high-end IT resources/services –a computational facility – over.
High Performance Computing Course Notes Grid Computing.
Campus High Throughput Computing (HTC) Infrastructures (aka Campus Grids) Dan Fraser OSG Production Coordinator Campus Grids Lead.
The Open Science Grid: Bringing the power of the Grid to scientific research
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Open Science Grid June 28, 2006 Bill Kramer Chair of the Open Science Grid Council NERSC Center General Manager, LBNL.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
April 2009 OSG Grid School - RDU 1 Open Science Grid John McGee – Renaissance Computing Institute University of North Carolina, Chapel.
Copyright James Kent Blackburn This work is the intellectual property of the author. Permission is granted for this material to be shared for non-commercial,
Assessment of Core Services provided to USLHC by OSG.
Open Science Ruth Pordes Fermilab, July 17th 2006 What is OSG Where Networking fits Middleware Security Networking & OSG Outline.
OSG Campus Grids Dr. Sebastien Goasguen, Clemson University ____________________________.
Open Science Grid Software Stack, Virtual Data Toolkit and Interoperability Activities D. Olson, LBNL for the OSG International.
Key Project Drivers - FY11 Ruth Pordes, June 15th 2010.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
CI Days: Planning Your Campus Cyberinfrastructure Strategy Russ Hobby, Internet2 Internet2 Member Meeting 9 October 2007.
SAN DIEGO SUPERCOMPUTER CENTER NUCRI Advisory Board Meeting November 9, 2006 Science Gateways on the TeraGrid Nancy Wilkins-Diehr TeraGrid Area Director.
INFSO-RI Enabling Grids for E-sciencE The US Federation Miron Livny Computer Sciences Department University of Wisconsin – Madison.
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
SG - OSG Improving Campus Research CI Through Leveraging and Integration: Developing a SURAgrid-OSG Collaboration John McGee, RENCI/OSG Engagement Coordinator.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
What is Cyberinfrastructure? Russ Hobby, Internet2 Clemson University CI Days 20 May 2008.
Use of Condor on the Open Science Grid Chris Green, OSG User Group / FNAL Condor Week, April
Open Science Grid  Consortium of many organizations (multiple disciplines)  Production grid cyberinfrastructure  80+ sites, 25,000+ CPU.
Condor Team Welcome to Condor Week #10 (year #25 for the project)
Mar 28, 20071/9 VO Services Project Gabriele Garzoglio The VO Services Project Don Petravick for Gabriele Garzoglio Computing Division, Fermilab ISGC 2007.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
Russ Hobby Program Manager Internet2 Cyberinfrastructure Architect UC Davis.
Interoperability Grids, Clouds and Collaboratories Ruth Pordes Executive Director Open Science Grid, Fermilab.
Data Intensive Science Network (DISUN). DISUN Started in May sites: Caltech University of California at San Diego University of Florida University.
Purdue Campus Grid Preston Smith Condor Week 2006 April 24, 2006.
Introduction to Grid Computing Ed Seidel Max Planck Institute for Gravitational Physics
Partnerships & Interoperability - SciDAC Centers, Campus Grids, TeraGrid, EGEE, NorduGrid,DISUN Ruth Pordes Fermilab Open Science Grid Joint Oversight.
Turning Software Projects into Production Solutions Dan Fraser, PhD Production Coordinator Open Science Grid OU Supercomputing Symposium October 2009.
The Open Science Grid OSG Ruth Pordes Fermilab. 2 What is OSG? A Consortium of people working together to Interface Farms and Storage to a Grid and Researchers.
08/05/06 Slide # -1 CCI Workshop Snowmass, CO CCI Roadmap Discussion Jim Bottum and Patrick Dreher Building the Campus Cyberinfrastructure Roadmap Campus.
Cyberinfrastructure What is it? Russ Hobby Internet2 Joint Techs, 18 July 2007.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Biomedical and Bioscience Gateway to National Cyberinfrastructure John McGee Renaissance Computing Institute
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre 14 February 2006.
Open Science Grid (OSG) Introduction for the Ohio Supercomputer Center Open Science Grid (OSG) Introduction for the Ohio Supercomputer Center February.
4/25/2006Condor Week 1 FermiGrid Steven Timm Fermilab Computing Division Fermilab Grid Support Center.
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
Cyberinfrastructure Overview Russ Hobby, Internet2 ECSU CI Days 4 January 2008.
Cyberinfrastructure: Many Things to Many People Russ Hobby Program Manager Internet2.
CMS Usage of the Open Science Grid and the US Tier-2 Centers Ajit Mohapatra, University of Wisconsin, Madison (On Behalf of CMS Offline and Computing Projects)
Eileen Berman. Condor in the Fermilab Grid FacilitiesApril 30, 2008  Fermi National Accelerator Laboratory is a high energy physics laboratory outside.
Sept 2008 OSG Engagement VO, RENCI 1 Open Science Grid Embedded Immersive Engagement for Cyberinfrastructure on the Open Science Grid John McGee –
Open Science Grid in the U.S. Vicky White, Fermilab U.S. GDB Representative.
An Introduction to Campus Grids 19-Apr-2010 Keith Chadwick & Steve Timm.
April 25, 2006Parag Mhashilkar, Fermilab1 Resource Selection in OSG & SAM-On-The-Fly Parag Mhashilkar Fermi National Accelerator Laboratory Condor Week.
Northwest Indiana Computational Grid Preston Smith Rosen Center for Advanced Computing Purdue University - West Lafayette West Lafayette Calumet.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Campus Grids Working Meeting Report Rob Gardner University of Chicago OSG All Hands March 10, 2010.
April 18, 2006FermiGrid Project1 FermiGrid Project Status April 18, 2006 Keith Chadwick.
HTCondor-CE. 2 The Open Science Grid OSG is a consortium of software, service and resource providers and researchers, from universities, national laboratories.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
What is OSG? (What does it have to do with Atlas T3s?) What is OSG? (What does it have to do with Atlas T3s?) Dan Fraser OSG Production Coordinator OSG.
FermiGrid The Fermilab Campus Grid 28-Oct-2010 Keith Chadwick Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
] Open Science Grid Ben Clifford University of Chicago
FermiGrid - PRIMA, VOMS, GUMS & SAZ Keith Chadwick Fermilab
Clouds , Grids and Clusters
f f FermiGrid – Site AuthoriZation (SAZ) Service
Grid Laboratory Of Wisconsin (GLOW)
GLOW A Campus Grid within OSG
Presentation transcript:

Open Science Grid For CI-Days NYSGrid Meeting Sebastien Goasguen, John McGee, OSG Engagement Manager School of Computing Clemson University, Clemson, SC Renaissance Computing Institute University of North Carolina, Chapel Hill, NC

2 21st Century Discovery The three fold way –theory –experiment –computational analysis Supported by –multimodal collaboration systems –distributed, multi-petabyte data archives –leading edge computing systems –distributed experimental facilities –distributed multidisciplinary teams Socialization and community –multidisciplinary groups –geographic distribution –new enabling technologies –creation of 21st century IT infrastructure sustainable, multidisciplinary communities Theory Experiment Simulation

Shift from Single User, Single Resource To: Multiple Users, Multiple Resources Any Combination of users and resources forms a Virtual Organization (VO) Grid computing is solving the problem of sharing resources among VO

Cyberinfrastructure : “Information Technology infrastructure to support a Virtual Organization” Therefore there are many Cyberinfrastructures not a single on The IT infrastructure is not only about HPC, but also software and applications The CI is put together to meet the needs of the VO members There are many re-usable components Leveraging existing assets is encouraged CI follows basic principles of service orientation and grid architecture The Open Science Grid aims at supporting VO to enable science, it can be a component of the CI you build for a particular VO. Disclaimer: This slide is the view of the author…

The Open Science Grid OSG is a consortium of software, service and resource providers and researchers, from universities, national laboratories and computing centers across the U.S., who together build and operate the OSG project. The project is funded by the NSF and DOE, and provides staff for managing various aspects of the OSG. Brings petascale computing and storage resources into a uniform grid computing environment Integrates computing and storage resources from over 50 sites in the U.S. and beyond A framework for large scale distributed resource sharing addressing the technology, policy, and social requirements of sharing

Principal Science Drivers High energy and nuclear physics –100s of petabytes (LHC)2007 –Several petabytes2005 LIGO (gravity wave search) –0.5 - several petabytes2002 Digital astronomy –10s of petabytes2009 –10s of terabytes2001 Other sciences emerging –Bioinformatics (10s of petabytes) –Nanoscience –Environmental –Chemistry –Applied mathematics –Materials Science

Virtual Organizations (VOs) The OSG Infrastructure trades in Groups not Individuals VO Management services allow registration, administration and control of members of the group. Facilities trust and authorize VOs. Storage and Compute Services prioritize according to VO group. Set of Available Resources VO Management Service OSG and WAN VO Management & Applications VO Management & Applications Campus Grid Image courtesy: UNM

Current OSG Resources OSG has more than 50 participating institutions, including self-operated research VOs, campus grids, regional grids and OSG-operated VOs Provides about 10,000 CPU-days per day in processing Provides 10 Terabytes per day in data transport CPU usage averages about 75% OSG is starting to offer support for MPI

What The OSG Offers that you may need to support your VO(s) Low-threshold access to many distributed computing and storage resources A combination of dedicated, scheduled, and opportunistic computing The Virtual Data Toolkit software packaging and distributions Grid Operations, including facility-wide monitoring, validation, information services and system integration testing Operational security Troubleshooting of end-to-end problems Education and Training

Date range: :00:00 GMT :59:59 GMT

OSG Bottom line: Framework to support VOs: VO of users only VO of resources VO of users and resources Can help you with: Supporting your VO Making your resources available inside and outside campus Enable science through user engagement

Campus Grids to the Rescue

Why should my University facilitate (or drive) resource sharing? Because it’s the right thing to do –Enables new modalities of collaboration –Enables new levels of scale –Democratizes large scale computing –Sharing locally leads to sharing globally –Better overall resource utilization –Funding agencies At the heart of the cyberinfrastructure vision is the development of a cultural community that supports peer-to- peer collaboration and new modes of education based upon broad and open access to leadership computing; data and information resources; online instruments and observatories; and visualization and collaboration services. - Arden Bement CI Vision for 21 st Century introduction At the heart of the cyberinfrastructure vision is the development of a cultural community that supports peer-to- peer collaboration and new modes of education based upon broad and open access to leadership computing; data and information resources; online instruments and observatories; and visualization and collaboration services. - Arden Bement CI Vision for 21 st Century introduction

Campus Grids They are a fundamental building block of the OSG –The multi-institutional, multi-disciplinary nature of the OSG is a macrocosm of many campus IT infrastructure coordination issues. Currently OSG has three operational campus grids on board: –Fermilab, Purdue, Wisconsin –Working to add Clemson, Harvard, Lehigh Elevation of jobs from Campus CI to OSG is transparent Campus scale brings value through –Richness of common software stack with common interfaces –Higher common denominator makes sharing easier –Greater collective buying power with venders –Synergy through common goals and achievements

Simplified View

Submitting jobs through OSG to UW Campus Grid (Dan Bradley, UW Madison) schedd (Job caretaker) startd (Job Executor) HEP matchmaker CS matchmaker GLOW matchmaker flocking schedd (Job caretaker) condor_submit condor gridmanager Open Science Grid User Globus gatekeeper GUMS

FermiGrid - Current Architecture (Keith Chadwick) CMS WC1 CDF OSG1 CDF OSG2 D0 CAB1 GP Farm VOMS Server SAZ Server GUMS Server Step 1 - user issues voms-proxy-init user receives voms signed credentials Step 2 – user submits their grid job via globus-job-run, globus-job-submit, or condor-g Step 4 – Gateway requests GUMS Mapping based on VO & Role Step 3 – Gateway checks against Site Authorization Service clusters send ClassAds via CEMon to the site wide gateway Step 5 - Grid job is forwarded to target cluster BlueArc Periodic Synchronization D0 CAB2 Site Wide Gateway Exterior Interior

Clemson Campus Condor Pool Machines in 27 different locations on Campus ~1,700 job slots >1.8M hours served in 6 months users from Industrial and Chemical engineering, and Economics Fast ramp up of usage Accessible to the OSG through a gatekeeper

Campuses and Regional Grids Campus Condor pool backfills idle nodes in PBS clusters - provided 5.5 million CPU- hours in 2006, all from idle nodes in clusters Use on TeraGrid: 2.4 million hours in 2006 spent Building a database of hypothetical zeolite structures; 2007: 5.5 million hours allocated to TG

“What impressed me most was how quickly we were able to access the grid and start using it. We learned about it at RENCI, and we were running jobs about two weeks later,” says Kuhlman. For each protein we design, we consume about 3,000 CPU hours across 10,000 jobs,” says Kuhlman. “Adding in the structure and atom design process, we’ve consumed about 100,000 CPU hours in total so far.” Engaging Users (more this afternoon)

What can we do together? Clemson’s OSG team is looking for a few partners to help deploy campus wide grid infrastructure that integrates with local enterprise infrastructure and the national CI RENCI’s OSG team is available to help scientists get their applications running on OSG –low impact starting point –Help your researchers gain significant compute cycles while exploring OSG as a framework for your own campus CI mailto:

E N D Sebastien Goasguen, John McGee, Questions ?