GLOW A Campus Grid within OSG

Slides:



Advertisements
Similar presentations
Jaime Frey Computer Sciences Department University of Wisconsin-Madison OGF 19 Condor Software Forum Routing.
Advertisements

Current methods for negotiating firewalls for the Condor ® system Bruce Beckles (University of Cambridge Computing Service) Se-Chang Son (University of.
Dan Bradley Computer Sciences Department University of Wisconsin-Madison Schedd On The Side.
Overview of Wisconsin Campus Grid Dan Bradley Center for High-Throughput Computing.
Building Campus HTC Sharing Infrastructures Derek Weitzel University of Nebraska – Lincoln (Open Science Grid Hat)
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
Condor and GridShell How to Execute 1 Million Jobs on the Teragrid Jeffrey P. Gardner - PSC Edward Walker - TACC Miron Livney - U. Wisconsin Todd Tannenbaum.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Workload Management Massimo Sgaravatto INFN Padova.
Condor Overview Bill Hoagland. Condor Workload management system for compute-intensive jobs Harnesses collection of dedicated or non-dedicated hardware.
Jim Basney Computer Sciences Department University of Wisconsin-Madison Managing Network Resources in.
Jaeyoung Yoon Computer Sciences Department University of Wisconsin-Madison Virtual Machines in Condor.
Zach Miller Computer Sciences Department University of Wisconsin-Madison What’s New in Condor.
SCD FIFE Workshop - GlideinWMS Overview GlideinWMS Overview FIFE Workshop (June 04, 2013) - Parag Mhashilkar Why GlideinWMS? GlideinWMS Architecture Summary.
S. Dasu, CHEP04, Interlacken, Switzerland1 Use of Condor and GLOW for CMS Simulation Production What are Condor & GLOW? What is special about Condor &
Miron Livny Computer Sciences Department University of Wisconsin-Madison Harnessing the Capacity of Computational.
Miron Livny Computer Sciences Department University of Wisconsin-Madison Submit locally and run globally – The GLOW and OSG Experience.
Alain Roy Computer Sciences Department University of Wisconsin-Madison An Introduction To Condor International.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
Grid Laboratory Of Wisconsin (GLOW) Sridhara Dasu, Dan Bradley, Steve Rader Department of Physics Miron Livny, Sean.
1 Evolution of OSG to support virtualization and multi-core applications (Perspective of a Condor Guy) Dan Bradley University of Wisconsin Workshop on.
Grid Computing I CONDOR.
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
Use of Condor on the Open Science Grid Chris Green, OSG User Group / FNAL Condor Week, April
ETICS All Hands meeting Bologna, October 23-25, 2006 NMI and Condor: Status + Future Plans Andy PAVLO Peter COUVARES Becky GIETZEL.
The Owner Share scheduler for a distributed system 2009 International Conference on Parallel Processing Workshops Reporter: 李長霖.
Data Intensive Science Network (DISUN). DISUN Started in May sites: Caltech University of California at San Diego University of Florida University.
Purdue Campus Grid Preston Smith Condor Week 2006 April 24, 2006.
Open Science Grid For CI-Days NYSGrid Meeting Sebastien Goasguen, John McGee, OSG Engagement Manager School of Computing.
Dan Bradley University of Wisconsin-Madison Condor and DISUN Teams Condor Administrator’s How-to.
Derek Wright Computer Sciences Department University of Wisconsin-Madison Condor and MPI Paradyn/Condor.
Condor Week 2004 The use of Condor at the CDF Analysis Farm Presented by Sfiligoi Igor on behalf of the CAF group.
Peter Couvares Associate Researcher, Condor Team Computer Sciences Department University of Wisconsin-Madison
High Energy FermiLab Two physics detectors (5 stories tall each) to understand smallest scale of matter Each experiment has ~500 people doing.
CMS Usage of the Open Science Grid and the US Tier-2 Centers Ajit Mohapatra, University of Wisconsin, Madison (On Behalf of CMS Offline and Computing Projects)
Condor Services for the Global Grid: Interoperability between OGSA and Condor Clovis Chapman 1, Paul Wilson 2, Todd Tannenbaum 3, Matthew Farrellee 3,
Eileen Berman. Condor in the Fermilab Grid FacilitiesApril 30, 2008  Fermi National Accelerator Laboratory is a high energy physics laboratory outside.
Parag Mhashilkar Computing Division, Fermi National Accelerator Laboratory.
Douglas Thain, John Bent Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau, Miron Livny Computer Sciences Department, UW-Madison Gathering at the Well: Creating.
April 25, 2006Parag Mhashilkar, Fermilab1 Resource Selection in OSG & SAM-On-The-Fly Parag Mhashilkar Fermi National Accelerator Laboratory Condor Week.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
Campus Grid Technology Derek Weitzel University of Nebraska – Lincoln Holland Computing Center (HCC) Home of the 2012 OSG AHM!
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Campus Grids Working Meeting Report Rob Gardner University of Chicago OSG All Hands March 10, 2010.
FermiGrid The Fermilab Campus Grid 28-Oct-2010 Keith Chadwick Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
UCS D OSG Summer School 2011 Intro to DHTC OSG Summer School An introduction to Distributed High-Throughput Computing with emphasis on Grid computing.
UCS D OSG Summer School 2011 Overlay systems OSG Summer School An introduction to Overlay systems Also known as Pilot systems by Igor Sfiligoi University.
Workload Management Workpackage
Condor A New PACI Partner Opportunity Miron Livny
Quick Architecture Overview INFN HTCondor Workshop Oct 2016
Dynamic Deployment of VO Specific Condor Scheduler using GT4
Examples Example: UW-Madison CHTC Example: Global CMS Pool
Outline Expand via Flocking Grid Universe in HTCondor ("Condor-G")
Building Grids with Condor
Condor: Job Management
Accounting, Group Quotas, and User Priorities
Semiconductor Manufacturing (and other stuff) with Condor
Condor and Multi-core Scheduling
Basic Grid Projects – Condor (Part I)
Brian Lin OSG Software Team University of Wisconsin - Madison
Genre1: Condor Grid: CSECCR
The Condor JobRouter.
Open Science Grid at Condor Week
Condor Administration in the Open Science Grid
Grid Laboratory Of Wisconsin (GLOW)
Condor-G Making Condor Grid Enabled
Job Submission Via File Transfer
Condor-G: An Update.
Presentation transcript:

GLOW A Campus Grid within OSG University of Wisconsin, Madison Dan Bradley Representing UW Madison CMS, GLOW, and Condor Teams

Overview: GLOW & OSG What is the value of a local campus-level grid? Why are we a part of OSG? How do we make them work together?

Why have a campus or enterprise grid? very high utilization more diverse users = less wasted cycles simplicity All we need is Condor at campus level. Plus, we get the full feature-set rather than lowest common denominator. collective buying power We speak to vendors with one voice. consolidated administration Fewer chores for scientists. Fewer holes for hackers. synergy Face-to-face technical meetings between members. Mailing list scales well at campus level.

Why is GLOW part of OSG? We can always use more resources. But we want to share when we have a surplus. Our users want to collaborate outside the bounds of the campus (e.g. Atlas and CMS). Others may join that trend. OSG does not infringe on our local control. The OSG grid interface does not limit our choice of technology within the campus grid, because it strives to remain independent of it.

What is the UW Campus Grid? Condor pools at various departments, made accessible via Condor ‘flocking’ Users submit jobs to their own private or department Condor scheduler. Jobs are dynamically matched to available machines. No cross-campus NFS for file access. People use Condor remote I/O, sandboxes, AFS, dCache, etc.

How big is the UW campus grid? GLOW Condor pool is distributed across the campus at the sites of the machine owners. 1800 cores 100 TB disk Over 25 million CPU-hours served Machine owner always has highest priority (via machine rank expression). Computer Science Condor pool 1000 cores Other private pools serve as submission and execution points for some users. Their excess jobs flock to GLOW and CS pools.

Who Uses UW Campus Grid? Computational Genomics, Chemistry High Energy Physics (CMS, Atlas) Materials by Design, Chemical Engineering Radiation Therapy, Medical Physics Computer Science Amanda, Ice-cube, Physics/Space Science Plasma Physics OSG VOs: nanohub, DZero, CDF, Zeus, … Diverse users with different conference deadlines, and usage patterns.

Submitting Jobs within UW Campus Grid HEP matchmaker CS matchmaker GLOW matchmaker schedd (Job caretaker) condor_submit job ClassAd flocking startd (Job Executor) machine ClassAd Supports full feature-set of Condor: matchmaking remote system calls checkpointing MPI universe suspension VMs preemption policies job runs

Submitting jobs through OSG to UW Campus Grid HEP matchmaker CS matchmaker GLOW matchmaker flocking Globus gatekeeper schedd (Job caretaker) condor_submit schedd (Job caretaker) job ClassAd startd (Job Executor) machine ClassAd condor gridmanager job runs

Routing Jobs from UW Campus Grid to OSG HEP matchmaker CS matchmaker GLOW matchmaker schedd (Job caretaker) condor_submit Grid JobRouter globus gatekeeper transform to a grid job condor gridmanager Best of both worlds: simple, feature-rich local mode transformable to standard OSG job for traveling globally

Conclusions: GLOW & OSG Our UW Condor grid is not a collection of OSG mini-sites, nor do we intend it to be. However, interoperability with the OSG has been an increasingly important activity for us, because it brings real benefits to our users. This model is emerging on many campuses today. We believe tomorrow’s researches will expect and demand to be well connected to both local and global computing resources.