1 The Roadmap to New Releases Todd Tannenbaum Department of Computer Sciences University of Wisconsin-Madison

Slides:



Advertisements
Similar presentations
Jaime Frey Computer Sciences Department University of Wisconsin-Madison OGF 19 Condor Software Forum Routing.
Advertisements

Current methods for negotiating firewalls for the Condor ® system Bruce Beckles (University of Cambridge Computing Service) Se-Chang Son (University of.
Dan Bradley Computer Sciences Department University of Wisconsin-Madison Schedd On The Side.
1 Concepts of Condor and Condor-G Guy Warner. 2 Harvesting CPU time Teaching labs. + Researchers Often-idle processors!! Analyses constrained by CPU time!
Setting up of condor scheduler on computing cluster Raman Sehgal NPD-BARC.
A Computation Management Agent for Multi-Institutional Grids
Condor and GridShell How to Execute 1 Million Jobs on the Teragrid Jeffrey P. Gardner - PSC Edward Walker - TACC Miron Livney - U. Wisconsin Todd Tannenbaum.
Jaime Frey Computer Sciences Department University of Wisconsin-Madison Condor-G: A Case in Distributed.
GRID workload management system and CMS fall production Massimo Sgaravatto INFN Padova.
GRID Workload Management System Massimo Sgaravatto INFN Padova.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment Chapter 8: Implementing and Managing Printers.
70-290: MCSE Guide to Managing a Microsoft Windows Server 2003 Environment, Enhanced Chapter 8: Implementing and Managing Printers.
Condor Overview Bill Hoagland. Condor Workload management system for compute-intensive jobs Harnesses collection of dedicated or non-dedicated hardware.
Evaluation of the Globus GRAM Service Massimo Sgaravatto INFN Padova.
Distributed Computing Overviews. Agenda What is distributed computing Why distributed computing Common Architecture Best Practice Case study –Condor –Hadoop.
Jaeyoung Yoon Computer Sciences Department University of Wisconsin-Madison Virtual Machines in Condor.
Derek Wright Computer Sciences Department, UW-Madison Lawrence Berkeley National Labs (LBNL)
Zach Miller Condor Project Computer Sciences Department University of Wisconsin-Madison Flexible Data Placement Mechanisms in Condor.
Cheap cycles from the desktop to the dedicated cluster: combining opportunistic and dedicated scheduling with Condor Derek Wright Computer Sciences Department.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Zach Miller Computer Sciences Department University of Wisconsin-Madison What’s New in Condor.
Miron Livny Computer Sciences Department University of Wisconsin-Madison Harnessing the Capacity of Computational.
Alain Roy Computer Sciences Department University of Wisconsin-Madison An Introduction To Condor International.
Todd Tannenbaum Computer Sciences Department University of Wisconsin-Madison What’s New in Condor.
Todd Tannenbaum Computer Sciences Department University of Wisconsin-Madison What’s New in Condor.
1 Todd Tannenbaum Department of Computer Sciences University of Wisconsin-Madison
Track 1: Cluster and Grid Computing NBCR Summer Institute Session 2.2: Cluster and Grid Computing: Case studies Condor introduction August 9, 2006 Nadya.
LARK Bringing Distributed High Throughput Computing to the Network Todd Tannenbaum U of Wisconsin-Madison Garhan Attebury
Research Achievements Kenji Kaneda. Agenda Research background and goal Research background and goal Overview of my research achievements Overview of.
The Glidein Service Gideon Juve What are glideins? A technique for creating temporary, user- controlled Condor pools using resources from.
PCGRID ‘08 Workshop, Miami, FL April 18, 2008 Preston Smith Implementing an Industrial-Strength Academic Cyberinfrastructure at Purdue University.
Installing and Managing a Large Condor Pool Derek Wright Computer Sciences Department University of Wisconsin-Madison
Hao Wang Computer Sciences Department University of Wisconsin-Madison Security in Condor.
Peter Keller Computer Sciences Department University of Wisconsin-Madison Quill Tutorial Condor Week.
Grid Computing I CONDOR.
Condor Birdbath Web Service interface to Condor
Greg Thain Computer Sciences Department University of Wisconsin-Madison cs.wisc.edu Interactive MPI on Demand.
Miron Livny Computer Sciences Department University of Wisconsin-Madison Welcome and Condor Project Overview.
Condor Project Computer Sciences Department University of Wisconsin-Madison Condor-G Operations.
Hao Wang Computer Sciences Department University of Wisconsin-Madison Authentication and Authorization.
Condor: High-throughput Computing From Clusters to Grid Computing P. Kacsuk – M. Livny MTA SYTAKI – Univ. of Wisconsin-Madison
1 The Roadmap to New Releases Todd Tannenbaum Department of Computer Sciences University of Wisconsin-Madison
Todd Tannenbaum Computer Sciences Department University of Wisconsin-Madison Condor RoadMap.
The Roadmap to New Releases Derek Wright Computer Sciences Department University of Wisconsin-Madison
Derek Wright Computer Sciences Department University of Wisconsin-Madison MPI Scheduling in Condor: An.
Condor Usage at Brookhaven National Lab Alexander Withers (talk given by Tony Chan) RHIC Computing Facility Condor Week - March 15, 2005.
Todd Tannenbaum Computer Sciences Department University of Wisconsin-Madison Quill / Quill++ Tutorial.
1 Condor BirdBath SOAP Interface to Condor Charaka Goonatilake Department of Computer Science University College London
Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre 14 February 2006.
Derek Wright Computer Sciences Department University of Wisconsin-Madison Condor and MPI Paradyn/Condor.
Derek Wright Computer Sciences Department University of Wisconsin-Madison New Ways to Fetch Work The new hook infrastructure in Condor.
Pilot Factory using Schedd Glidein Barnett Chiu BNL
Alain Roy Computer Sciences Department University of Wisconsin-Madison Condor & Middleware: NMI & VDT.
Nick LeRoy Computer Sciences Department University of Wisconsin-Madison Hawkeye.
Greg Thain Computer Sciences Department University of Wisconsin-Madison Configuring Quill Condor Week.
Jaime Frey Computer Sciences Department University of Wisconsin-Madison What’s New in Condor-G.
Dan Bradley Condor Project CS and Physics Departments University of Wisconsin-Madison CCB The Condor Connection Broker.
Miron Livny Computer Sciences Department University of Wisconsin-Madison Condor and (the) Grid (one of.
Matthew Farrellee Computer Sciences Department University of Wisconsin-Madison Condor and Web Services.
Todd Tannenbaum Computer Sciences Department University of Wisconsin-Madison Condor NT Condor ported.
Jaime Frey Computer Sciences Department University of Wisconsin-Madison Condor and Virtual Machines.
HTCondor’s Grid Universe Jaime Frey Center for High Throughput Computing Department of Computer Sciences University of Wisconsin-Madison.
A System for Monitoring and Management of Computational Grids Warren Smith Computer Sciences Corporation NASA Ames Research Center.
Condor on Dedicated Clusters Peter Couvares and Derek Wright Computer Sciences Department University of Wisconsin-Madison
Condor: Job Management
Basic Grid Projects – Condor (Part I)
The Condor JobRouter.
Condor-G Making Condor Grid Enabled
STATEL an easy way to transfer data
Condor-G: An Update.
Presentation transcript:

1 The Roadmap to New Releases Todd Tannenbaum Department of Computer Sciences University of Wisconsin-Madison

2 Stable vs. Development Series › Much like the Linux kernel, Condor provides two different releases at any time:  Stable series  Development series › Allows Condor to be both a research project and a production-ready system

3 Stable series › Series number in version is even (e.g ) › Releases are heavily tested › Only bug fixes and ports to new platforms are added on a stable series

4 Stable series (cont.) › A given stable release is always compatible with other releases from the same series › Recommended for production pools

5 Development Series › Series number in the version is odd (e.g , 6.3.0) › New features and new technology are added frequently › Versions from the same development series are not always compatible with each other

6 Development Series (cont.) › Releases are not as heavily tested › Generally not recommended for production pools  … unless new features are required  … unless we recommend otherwise :^)

7 Where is Condor Today? › New stable version recently released › Previous development series cycle, ver 6.1.*, was far too long › This was our first stable/development series: we're hopefully learning...

8 New Ports in › Full support (with checkpointing and remote system calls):  All current versions of Linux (x86) Kernel: 2.2.* and 2.0.* C Library: glibc-2.[01] and libc-5 NOTE: libc5 support dropped  Solaris 2.7 and 2.8(Sparc and x86)  Irix 6.5

9 New Ports in (cont.) › ”Clipped" support (no checkpointing or remote system calls, but all other functionality is available)  Windows NT  Alpha Linux

10 What is New in 6.2.0? › Personal Condor and Grid Support  Flocking  Proof-of-concept support for Globus Job Universe Globus Glide-In › Full, integrated support for Symmetric Multi-Processor (SMP) machines

11 What's New in 6.2.0? (cont.) › PVM support › DAGMan (for managing inter-job dependencies) › Use of machine ClassAd attributes in submit file  Example: Executable = foo.$$(ARCH).$$(OPSYS)

12 What's New in 6.2.0? (cont.) › Greatly expanded I/O support for STANDARD universe jobs  Condor can automatically buffer I/O requests from jobs  Compression/decompression on-the-fly!  Users get much more information about the kinds of I/O their jobs are performing  Users can "remap" files to alternate locations

13 What's New in 6.2.0? (cont.) › CondorVersion and CondorPlatform strings included in all binaries and libraries  Helps identify and avoid problems with having the wrong version installed  Different parts of the Condor protocol automatically check for version incompatibilities

14 What's New in 6.2.0? (cont.) › Better accounting  Historical usage logging in collector Use condor_stats web interface called CondorView  Accountant stores usage information per user › Better control over user priorities  "Priority factors"

15 What's New in 6.2.0? (cont.) › More powerful administration tools  Setting configuration values remotely  Querying daemons directly for status › Other misc new tools  condor_findhost, condor_hold/release, condor_qedit, condor_userlog, condor_run, … › Lots of performance and bug fixes › A complete list is online in the manual (

16 The 6.3 Development Series › Major focus areas for ver 6.3.x Series  Future-aware scheduling for dedicated resources MPI support  Security enhancements –Secure network communication »Kerberos, PKI (GSI/OpenSSL) –Limitations on executable, remote syscall sandbox  Condor-G production version Integrate Bypass technology  Enhanced Windows support

17 What else will be added in 6.3.*? › And also several easy-to-add, user- visible features  ”Master agents" - helper programs spawned by the condor_master to aid in administration Retrieving remote log, history and/or configuration files Remote "top", "ps" and other monitoring functions

18 What will be added in 6.3.*? (cont.) › Solution for Condor clusters on  Private network  Behind a firewall › New condor_starter and condor_shadow for the VANILLA universe on Unix  "NT version" will be used for Unix, too  Lots of new features, like transferring files automatically for "vanilla" jobs (no need for a shared filesystem)

19 What will be added in 6.3.*? (cont.) › condor_startd will enforce resource limits dynamically (like RAM usage) › More logic put into the condor_shadow to  Detect temporary problems with a job's execution, put the job on hold, and notify the user  Enforce job submitting user’s policy › Better propagation of error messages

20 The 6.5.* Development Series › 6.5.* will be for adding fundamentally new technology to Condor › Being developed in parallel with 6.3.* › Will ultimately be v7.0 › Will be incompatible with previous versions of Condor

21 New Technology in 6.5.* › New version of ClassAds  Conditionals !! if/then/else  Aggregates (lists, nested classads)  Built-in functions String operations, pattern matching, time operators, unit conversions  Clean implementations in C++ and Java  ClassAd collections

22 Other changes for 6.5.* › Re-write of the condor_schedd  Performance enhancements and lowered resource requirements (particularly RAM) › Re-write of the checkpoint server  NEST technology infusion  Enhanced support for multiple servers  Will meta-data along with checkpoint files

23 Planned Future Ports › Full support  Intel Linux 2.4 kernel (RedHat 7.x) › Clipped support  Windows 2000 / Whistler

24 Possible Future Ports  Full Support (I.e. checkpointing, etc) Windows 2000 Alpha Linux DUX 5.x  Clipped-support {free,open,net}BSD MacOS X PowerPC Linux HPUX 11.0 AIX 4.2

25 Thank you for coming to Paradyn/Condor Week!