Purdue Campus Grid Preston Smith Condor Week 2006 April 24, 2006.

Slides:



Advertisements
Similar presentations
Grid Deployments and Cyberinfrastructure Andrew J. Younge 102 Lomb Memorial Drive Rochester, NY 14623
Advertisements

DOSAR Workshop VI April 17, 2008 Louisiana Tech Site Report Michael Bryant Louisiana Tech University.
Building Campus HTC Sharing Infrastructures Derek Weitzel University of Nebraska – Lincoln (Open Science Grid Hat)
Campus High Throughput Computing (HTC) Infrastructures (aka Campus Grids) Dan Fraser OSG Production Coordinator Campus Grids Lead.
NSF Site Visit HYDRA Using Windows Desktop Systems in Distributed Parallel Computing.
PRESTON SMITH ROSEN CENTER FOR ADVANCED COMPUTING PURDUE UNIVERSITY A Cost-Benefit Analysis of a Campus Computing Grid Condor Week 2011.
April 2009 OSG Grid School - RDU 1 Open Science Grid John McGee – Renaissance Computing Institute University of North Carolina, Chapel.
Edge Based Cloud Computing as a Feasible Network Paradigm(1/27) Edge-Based Cloud Computing as a Feasible Network Paradigm Joe Elizondo and Sam Palmer.
TG RoundTable, Purdue RP Update October 11, 2008 Carol Song Purdue RP PI Rosen Center for Advanced Computing.
SCD FIFE Workshop - GlideinWMS Overview GlideinWMS Overview FIFE Workshop (June 04, 2013) - Parag Mhashilkar Why GlideinWMS? GlideinWMS Architecture Summary.
OSG Campus Grids Dr. Sebastien Goasguen, Clemson University ____________________________.
Implementing a Central Quill Database in a Large Condor Installation Preston Smith Condor Week April 30, 2008.
April 2006 Science Gateways on the TeraGrid Nancy Wilkins-Diehr Area Director for Science Gateways San Diego Supercomputer Center
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
Purdue RP Highlights TeraGrid Round Table September 23, 2010 Carol Song Purdue TeraGrid RP PI Rosen Center for Advanced Computing Purdue University.
Ways to Connect to OSG Tuesday afternoon, 3:00 pm Lauren Michael Research Computing Facilitator University of Wisconsin-Madison.
Welcome to CW 2007!!!. The Condor Project (Established ‘85) Distributed Computing research performed by.
Open Science Grid Software Stack, Virtual Data Toolkit and Interoperability Activities D. Olson, LBNL for the OSG International.
A. Mohapatra, HEPiX 2013 Ann Arbor1 UW Madison CMS T2 site report D. Bradley, T. Sarangi, S. Dasu, A. Mohapatra HEP Computing Group Outline  Infrastructure.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
Grids and the Home Institution A Campus Grids Overview by Laura F M c Ginnis Pittsburgh Supercomputing Center.
Gurcharan S. Khanna Director of Research Computing RIT
Information Technology at Purdue Presented by: Dr. Gerry McCartney Vice President and CIO, ITaP HPC User Forum September 8-10, 2008 Using SiCortex SC5832.
PCGRID ‘08 Workshop, Miami, FL April 18, 2008 Preston Smith Implementing an Industrial-Strength Academic Cyberinfrastructure at Purdue University.
Kento Aida, Tokyo Institute of Technology Grid Challenge - programming competition on the Grid - Kento Aida Tokyo Institute of Technology 22nd APAN Meeting.
OSG Site Provide one or more of the following capabilities: – access to local computational resources using a batch queue – interactive access to local.
Condor to Every Corner of Campus Condor Week 2010 Preston Smith Purdue University.
1 Evolution of OSG to support virtualization and multi-core applications (Perspective of a Condor Guy) Dan Bradley University of Wisconsin Workshop on.
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
Use of Condor on the Open Science Grid Chris Green, OSG User Group / FNAL Condor Week, April
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
Edinburgh Investment in e-Science Infrastructure Dr Arthur Trew.
Kelly Gaither Visualization Area Report. Efforts in 2008 Focused on providing production visualization capabilities (software and hardware) Focused on.
Grid Computing at The Hartford Condor Week 2008 Robert Nordlund
HYDRA: Using Windows Desktop Systems in Distributed Parallel Computing Arvind Gopu, Douglas Grover, David Hart, Richard Repasky, Joseph Rinkovsky, Steve.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
São Paulo Regional Analysis Center SPRACE Status Report 22/Aug/2006 SPRACE Status Report 22/Aug/2006.
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
Interoperability Grids, Clouds and Collaboratories Ruth Pordes Executive Director Open Science Grid, Fermilab.
Purdue Tier-2 Site Report US CMS Tier-2 Workshop 2010 Preston Smith Purdue University
Open Science Grid For CI-Days NYSGrid Meeting Sebastien Goasguen, John McGee, OSG Engagement Manager School of Computing.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
Evolution of a High Performance Computing and Monitoring system onto the GRID for High Energy Experiments T.L. Hsieh, S. Hou, P.K. Teng Academia Sinica,
Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre 14 February 2006.
GLIDEINWMS - PARAG MHASHILKAR Department Meeting, August 07, 2013.
Farms User Meeting April Steven Timm 1 Farms Users meeting 4/27/2005
Accelerating Campus Research with Connective Services for Cyberinfrastructure Rob Gardner Steve Tuecke.
An Introduction to Campus Grids 19-Apr-2010 Keith Chadwick & Steve Timm.
Parag Mhashilkar Computing Division, Fermi National Accelerator Laboratory.
Purdue RP Highlights TeraGrid Round Table May 20, 2010 Preston Smith Manager - HPC Grid Systems Rosen Center for Advanced Computing Purdue University.
A. Mohapatra, T. Sarangi, HEPiX-Lincoln, NE1 University of Wisconsin-Madison CMS Tier-2 Site Report D. Bradley, S. Dasu, A. Mohapatra, T. Sarangi, C. Vuosalo.
INRNE's participation in LCG Elena Puncheva Preslav Konstantinov IT Department.
Western Tier 2 Site at SLAC Wei Yang US ATLAS Tier 2 Workshop Harvard University August 17-18, 2006.
Northwest Indiana Computational Grid Preston Smith Rosen Center for Advanced Computing Purdue University - West Lafayette West Lafayette Calumet.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
Campus Grid Technology Derek Weitzel University of Nebraska – Lincoln Holland Computing Center (HCC) Home of the 2012 OSG AHM!
Purdue RP Highlights TeraGrid Round Table November 5, 2009 Carol Song Purdue TeraGrid RP PI Rosen Center for Advanced Computing Purdue University.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Creating Grid Resources for Undergraduate Coursework John N. Huffman Brown University Richard Repasky Indiana University Joseph Rinkovsky Indiana University.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
3 Compute Elements are manageable By hand 2 ? We need middleware – specifically a Workload Management System (and more specifically, “glideinWMS”) 3.
UCS D OSG Summer School 2011 Intro to DHTC OSG Summer School An introduction to Distributed High-Throughput Computing with emphasis on Grid computing.
Deploying Regional Grids Creates Interaction, Ideas, and Integration
Introduce yourself Presented by
GLOW A Campus Grid within OSG
Presentation transcript:

Purdue Campus Grid Preston Smith Condor Week 2006 April 24, 2006

Overview RCAC –Community Clusters Grids at Purdue –Campus –Regional NWICG –National OSG CMS Tier-2 NanoHUB Teragrid Future Work

Purdue’s RCAC Rosen Center for Advanced Computing –Division of Information Technology at Purdue (ITaP) –Wide variety of systems: shared memory and clusters 352 CPU IBM SP Five 24-processor Sun F6800s, Two 56-processor Sun E10ks Five Linux clusters

Linux clusters in RCAC Recycled clusters –Systems retired from student labs –Nearly 1000 nodes of single-CPU PIII, P4, and 2-CPU Athlon MP and EM64T Xeons for general use by Purdue researchers

Community Clusters Federate resources at a low level Separate researchers buy sets of nodes to federate into larger clusters –Enables larger clusters than a scientist could support on his own –Leverage central staff and infrastructure No need to sacrifice a grad student to be a sysadmin!

Community Clusters Hamlet  308 nodes dual Xeon (3.6 Tflops)  3.06 GHz to 3.2 GHz  2 GB and 4 GB RAM  GigE, Infiniband  5 owners (EAS, BIOx2, CMS, EE) Macbeth  126 nodes dual Opteron (~1 Tflops)  1.8 GHz  4-16GB RAM  Infiniband, GigE for IP traffic  7 owners (ME, Biology, HEP Theory) Lear  512 nodes dual Xeon 64 bit (6.4 Tflops)  3.2 GHz  4GB and 6 GB RAM  GigE  6 owners (EEx2, CMS, Provost, VPR, Teragrid)

Community Clusters Primarily scheduled with PBS –Contributing researchers are assigned a queue that can run as many “slots” as they have contributed. Condor co-schedules alongside PBS –When PBS is not running a job, a node is fair game for Condor! But Condor work is subject to preemption if PBS assigns work to the node.

Condor on Community Clusters All in all, Condor joins together 4 clusters (~2500 CPU) within RCAC.

Grids at Purdue - Campus Instructional computing group manages a 1300-node Windows Condor pool to support instruction. –Mostly used by computer graphics classes for rendering animations Maya, etc. –Work in progress to connect Windows pool with RCAC pools.

Grids at Purdue - Campus Condor pools around campus –Physics department: 100 nodes, flocked –Envision Center: 48 nodes, flocked Potential collaborations –Libraries: ~200 nodes on Windows terminals –Colleges of Engineering: 400 nodes in existing pool Or any department interested in sharing cycles!

Grids at Purdue - Regional Northwest Indiana Computational Grid –Purdue West Lafayette –Purdue Calumet –Notre Dame –Argonne Labs Condor pools available to NWICG today. Partnership with OSG?

Open Science Grid Purdue active in Open Science Grid –CMS Tier-2 Center –NanoHUB –OSG/Teragrid Interoperability Campus Condor pools accessible to OSG –Condor used for access to extra, non-dedicated cycles for CMS and is becoming the preferred interface for non-CMS VOs.

CMS Tier-2 - Condor –MC production from UW-HEP ran this spring on RCAC Condor pools. Processed 23% or so of entire production. High rates of preemption, but that’s expected! –2006 will see addition of dedicated Condor worker nodes to Tier-2, in addition to PBS clusters. Condor running on resilient dCache nodes.

NanoHUB Campus Grids Purdue, GLOW Grid Capability Computing Science Gateway Workspaces Research apps Virtual backends Virtual Cluster with VIOLIN VM Capacity Computing nanoHUB VO Middleware

Teragrid Teragrid Resource Provider Resources offered to Teragrid –Lear cluster –Condor pools –Data collections

Teragrid Two current projects active in Condor pools via Teragrid allocations –Database of Hypothetical Zeolite Structures –CDF Electroweak MC Simulation Condor-G Glide-in Great exercise in OSG/TG Interoperability –Identifying other potential users

Teragrid TeraDRE - Distributed Rendering on the Teragrid –Globus, Condor, and IBRIX FusionFS enables Purdue’s Teragrid site to serve as a render farm Maya and other renderers available

Grid Interoperability “Lear”

Grid Interoperability Tier-2 to Tier-2 connectivity via dedicated Teragrid WAN (UCSD->Purdue) Aggregating resources at low level makes interoperability easier! –OSG stack available to TG users and vice versa “Bouncer” Globus job forwarder

Future of Condor at Purdue Add resources –Continue growth around campus RCAC Other departments Add Condor capabilities to resources –Teragrid data portal adding on-demand processing with Condor now Federation –Aggregate Condor pools with other institutions?

Condor at Purdue Questions?

PBS/Condor Interaction PBS Prologue # Prevent new Condor jobs and push any existing ones off # /opt/condor/bin/condor_config_val -rset -startd \ PBSRunning=True > /dev/null /opt/condor/sbin/condor_reconfig -startd > /dev/null if ( condor_status -claimed -direct $(hostname) 2>/dev/null \ | grep -q Machines ) then condor_vacate > /dev/null sleep 5 fi

PBS/Condor Interaction PBS Epilogue /opt/condor/bin/condor_config_val -rset -startd \ PBSRunning=False > /dev/null /opt/condor/sbin/condor_reconfig -startd > /dev/null Condor START Expression in condor_config.local PBSRunning = False # Only start jobs if PBS is not currently running a job PURDUE_RCAC_START_NOPBS = ( $(PBSRunning) == False ) START = $(START) && $(PURDUE_RCAC_START_NOPBS)