The UCL Condor Pool Experience John Brodholt 1, Paul Wilson 3, Wolfgang Emmerich 2 and Clovis Chapman 2. 1. Department of Earth Sciences, University College.

Slides:



Advertisements
Similar presentations
Overview of local security issues in Campus Grid environments Bruce Beckles University of Cambridge Computing Service.
Advertisements

Condor use in Department of Computing, Imperial College Stephen M c Gough, David McBride London e-Science Centre.
Overview of Wisconsin Campus Grid Dan Bradley Center for High-Throughput Computing.
A Grid approach to Environmental Molecular Simulations: Deployment and use of Condor within the eMinerals Mini Grid. Paul Wilson 1, Mark Calleja 2, John.
Experience of the SRB in support of collaborative grid computing Martin Dove University of Cambridge.
Building Campus HTC Sharing Infrastructures Derek Weitzel University of Nebraska – Lincoln (Open Science Grid Hat)
Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
What is e-Science? e-Science refers to large scale science that will increasingly be carried out through distributed global collaborations enabled by the.
Wahid Bhimji Andy Washbrook And others including ECDF systems team Not a comprehensive update but what ever occurred to me yesterday.
Dr. David Wallom Use of Condor in our Campus Grid and the University September 2004.
Many kinds of clients and servers This work is licensed under a Creative Commons Attribution-Noncommercial- Share Alike 3.0 License. Skills: none IT concepts:
OxGrid, A Campus Grid for the University of Oxford Dr. David Wallom.
Computational Steering on the GRID Using a 3D model to Interact with a Large Scale Distributed Simulation in Real-Time Michael.
A quick introduction to CamGrid University Computing Service Mark Calleja.
Simulation of natural organic matter adsorption to soils: A preliminary report Indiana Biocomplexity Symposium, Notre Dame, IN, April 2003 Leilani Arthurs.
EMinerals and the Condor Pool John Brodholt UCL Arnaud Marmier Zhimei Du Maria Alfredsson Clovis Chapman Marc Blanchard Presenting the work of :
Enabling Grids for E-sciencE Medical image processing web portal : Requirements analysis. An almost end user point of view … H. Benoit-Cattin,
14 July 2004GridPP Collaboration BoardSlide 1 GridPP Dissemination Sarah Pearce Dissemination Officer
Cloud Computing for the Enterprise November 18th, This work is licensed under a Creative Commons.
ISNE 101 Dr. Ken Cosh. Welcome to ISNE 101  Dr. Ken Cosh   Facebook – “Ken Cosh”   Don’t be Grengjai!
AS Computing Software definitions.
CS 0004 –Lecture 1 Wednesday, Jan 5 th, 2011 Roxana Gheorghiu.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
INTRODUCTION TO COMPUTING
National Center for Supercomputing Applications The Computational Chemistry Grid: Production Cyberinfrastructure for Computational Chemistry PI: John Connolly.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
NATIONAL INSTITUTE OF SCIENCE & TECHNOLOGY Presented by: Santosh kumar Swain Technical Seminar Presentation by SANTOSH KUMAR SWAIN Roll # CS
30-Jun-04UCL HEP Computing Status June UCL HEP Computing Status April DESKTOPS LAPTOPS BATCH PROCESSING DEDICATED SYSTEMS GRID MAIL WEB WTS.
John Kewley e-Science Centre CCLRC Daresbury Laboratory 28 th June nd European Condor Week Milano Heterogeneous Pools John Kewley
1 Evolution of OSG to support virtualization and multi-core applications (Perspective of a Condor Guy) Dan Bradley University of Wisconsin Workshop on.
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Grid tool integration within the eMinerals project Mark Calleja.
Grid MP at ISIS Tom Griffin, ISIS Facility. Introduction About ISIS Why Grid MP? About Grid MP Examples The future.
Neil Geddes GridPP-10, June 2004 UK e-Science Grid Dr Neil Geddes CCLRC Head of e-Science Director of the UK Grid Operations Support Centre.
Cliff Addison University of Liverpool Campus Grids Workshop October 2007 Setting the scene Cliff Addison.
Experiences with the Globus Toolkit on AIX and deploying the Large Scale Air Pollution Model as a grid service Ashish Thandavan Advanced Computing and.
WP8 Meeting Glenn Patrick1 LHCb Grid Activities in UK Grid WP8 Meeting, 16th November 2000 Glenn Patrick (RAL)
ARGONNE NATIONAL LABORATORY Climate Modeling on the Jazz Linux Cluster at ANL John Taylor Mathematics and Computer Science & Environmental Research Divisions.
Leveraging HTC for UK eScience with Very Large Condor Pools: Demand for transforming untapped power into results. Paul Wilson 1, John Brodholt 1, and Wolfgang.
09/02 ID099-1 September 9, 2002Grid Technology Panel Patrick Dreher Technical Panel Discussion: Progress in Developing a Web Services Data Analysis Grid.
“Grids and eScience” Mark Hayes Technical Director - Cambridge eScience Centre GEFD Summer School 2003.
The UK eScience Grid (and other real Grids) Mark Hayes NIEeS Summer School 2003.
The eMinerals minigrid and the national grid service: A user’s perspective NGS169 (A. Marmier)
Building the e-Minerals Minigrid Rik Tyer, Lisa Blanshard, Kerstin Kleese (Data Management Group) Rob Allan, Andrew Richards (Grid Technology Group)
Tevfik Kosar Computer Sciences Department University of Wisconsin-Madison Managing and Scheduling Data.
ELEMENTS OF A COMPUTER SYSTEM HARDWARE SOFTWARE PEOPLEWARE DATA.
Creating and running an application.
UK Grid Meeting Glenn Patrick1 LHCb Grid Activities in UK Grid Prototype and Globus Technical Meeting QMW, 22nd November 2000 Glenn Patrick (RAL)
E-Science Technologies in the Simulation of Complex Materials L. Blanshard, R. Tyer, K. Kleese S. A. French, D. S. Coombes, C. R. A. Catlow B. Butchart,
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
Utility Computing: Security & Trust Issues Dr Steven Newhouse Technical Director London e-Science Centre Department of Computing, Imperial College London.
Timeshared Parallel Machines Need resource management Need resource management Shrink and expand individual jobs to available sets of processors Shrink.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Condor Services for the Global Grid: Interoperability between OGSA and Condor Clovis Chapman 1, Paul Wilson 2, Todd Tannenbaum 3, Matthew Farrellee 3,
AHM04: Sep 2004 Nottingham CCLRC e-Science Centre eMinerals: Environment from the Molecular Level Managing simulation data Lisa Blanshard e- Science Data.
John Kewley e-Science Centre CCLRC Daresbury Laboratory 15 th March 2005 Paradyn / Condor Week Madison, WI Caging the CCLRC Compute Zoo (Activities at.
Stephen Burke – Sysman meeting - 22/4/2002 Partner Logo The Testbed – A User View Stephen Burke, PPARC/RAL.
Group # 14 Dhairya Gala Priyank Shah. Introduction to Grid Appliance The Grid appliance is a plug-and-play virtual machine appliance intended for Grid.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Hopper The next step in High Performance Computing at Auburn University February 16, 2016.
Creating Grid Resources for Undergraduate Coursework John N. Huffman Brown University Richard Repasky Indiana University Joseph Rinkovsky Indiana University.
SCI-BUS project Pre-kick-off meeting University of Westminster Centre for Parallel Computing Tamas Kiss, Stephen Winter, Gabor.
Types of computers Module 1.10 AS identifies and distinguishes between computer types and associated software. AS identifies the main hardware.
VCL Best practices Lee Toderick, Department of Technology Systems
Chapter 1 – Introduction to Computers, the Internet, and the Web
US CMS Testbed.
Office of Information Technology February 16, 2016
GLOW A Campus Grid within OSG
Overview of Computer system
Presentation transcript:

The UCL Condor Pool Experience John Brodholt 1, Paul Wilson 3, Wolfgang Emmerich 2 and Clovis Chapman Department of Earth Sciences, University College London, Gower Street, London WC1E 6BT, UK 2. Department of Computer Science, University College London, Gower Street, London WC1E 6BT, UK 3. Anvil Software, London, UK. Environment from the Molecular Level A NERC eScience testbed project

Environment from the Molecular Level A NERC eScience testbed project Approximately 946 Windows machines (yesterday) 1 to 2.4 GHtz Intel processors 256 to 512 MBytes memory (a few are more) They are in “open access” student cluster rooms PCs are all thin client “WTS” machines with network bootable operating systems. (Citrix/Bpbatch - hit spacebar to upload new operating systems image) The pool is very simple – one manager, one submit machine (via ssh). The UCL Condor Pool

Environment from the Molecular Level A NERC eScience testbed project Q. Why was it someone from an Earth Science Dept. who got it going? 1.Because three years ago, the eScience grants call made me look up “the Grid” on the web and by chance I came across the Condor web site. 2.I also happened to know how Information Systems at UCL managed their student PCs. 3.Persuaded the Director of UCL’s Education and Information Systems Division that I could put it in our eMinerals grant (I think he assumed it wouldn’t get funded).

Environment from the Molecular Level A NERC eScience testbed project Key Political Issues Even though the Director of EISD had agreed for us to put it in the grant, we had to convince Information Systems themselves. Numerous meetings …. IS produced a five page document outlining what they thought their policy on a large Condor cluster would be – i.e. the primary purpose of the student cluster rooms must not be compromised. Nor should IS staff use their time on the project … etc. Needed testing (one cluster, then one image type). Perhaps the key moment was when the UCL presented its eScience projects to Tony Hey and the UCL Provost.

Environment from the Molecular Level A NERC eScience testbed project Timescale Desktop - June 2002 (2 nodes) Earth Science Student Cluster Room - Oct 2002 (18 nodes) Physics Department (one WTS image) – Jan 2003 (150 nodes) Campus – October 16th 2003 (930 nodes) 1 millionth hour of CPU – April 2 nd, 2004 This matched exactly the timescale we outlined in the eMinerals grant

Environment from the Molecular Level A NERC eScience testbed project Other Issues Some needed to compile their codes for Windows machines – “It’s simple, just convert them to Java..” Wolfgang Emmerich, 2002! Difficult to persuade the scientists to get involved for just a few machines. Our central manager died a few times when a user submitted a few thousand jobs all at the same time (took 24 hours to repair disk with fsck). Now have a manager and a submit machine. Students will do anything to reserve a machine – steal the mouse, put out of order signs on them, and UNPLUG them. Also, IS themselves briefly turn machines in some clusters off in order to clear the room. This restricts the length of job.

Environment from the Molecular Level A NERC eScience testbed project

Environment from the Molecular Level A NERC eScience testbed project Spikes in user demand: a) Not many users b) Most are using simple schemes to produce lots of initial input files and send off to pool. Get results back and spend a long time processing them/extracting data/planning next set of inputs.

Jobs Database Analysis Database Model/Configuration Generator Distributed Computing Portal Distributed resources (Condor pools etc.) Steering Database Improve generation / model strategy User Input: Structural model Si/Al, cation types, [H 2 O] etc. User Input: Diffraction data, chemical analysis, building units Si/Al, cation types, [H 2 O] etc. Analysis (geometry, energy, fit) Existing e-science technology Drip feeding and interactive steering of a Condor pool using relational databases Dewi Lewis, Rosie Coates and Sam French UCL Chemistry / RI

THE Science. 1. Simulation of pollutants in the environment Binding of heavy metals and organic molecules in soils. 2. Studies of materials for long-term nuclear waste encapsulation Radiocactive waste leaching through ceramic storage media. 3. Studies of weathering and scaling Mineral/water interface simulations, e.g oil well scaling. also 4. The Earth’s core and mantle Many codes: DL-POLY, GULP, METADISE, CRYSTAL, CASTEP, SIESTA, … Environment from the Molecular Level A NERC eScience testbed project

Environment from the Molecular Level A NERC eScience testbed project Now what? Expand pool to include staff WTS machines ~ 1500 machines (received 3 page from IS - who owns them?). UCL Staff machines at hospitals ~ ???? machines. Federate with other pools: hopefully make it more flexible smooth spikes in demand.