Practical aspects of multi-core job submission at CERN

Slides:



Advertisements
Similar presentations
Copyright © 2010, SAS Institute Inc. All rights reserved. SAS and all other SAS Institute Inc. product or service names are registered trademarks or trademarks.
Advertisements

Operating Systems Review.
© 2007 IBM Corporation IBM Global Engineering Solutions IBM Blue Gene/P Job Submission.
Parallel Session B2 - CPU and Resource Allocation " Panelists: – Charles Young (BaBar) – David Bigagli " Seed Questions: – Batch queuing system in use?
Alastair Dewhurst, Dimitrios Zilaskos RAL Tier1 Acknowledgements: RAL Tier1 team, especially John Kelly and James Adams Maximising job throughput using.
A Dynamic World, what can Grids do for Multi-Core computing? Daniel Goodman, Anne Trefethen and Douglas Creager
Pre-GDB on Batch Systems (Bologna)11 th March Torque/Maui PIC and NIKHEF experience C. Acosta-Silva, J. Flix, A. Pérez-Calero (PIC) J. Templon (NIKHEF)
K.Harrison CERN, 23rd October 2002 HOW TO COMMISSION A NEW CENTRE FOR LHCb PRODUCTION - Overview of LHCb distributed production system - Configuration.
Appendix B Planning a Virtualization Strategy for Exchange Server 2010.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
Scientific Computing Division Juli Rew CISL User Forum May 19, 2005 Scheduler Basics.
Bigben Pittsburgh Supercomputing Center J. Ray Scott
Tier-1 Batch System Report Andrew Lahiff, Alastair Dewhurst, John Kelly, Ian Collier 5 June 2013, HEP SYSMAN.
1 Evolution of OSG to support virtualization and multi-core applications (Perspective of a Condor Guy) Dan Bradley University of Wisconsin Workshop on.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Using the BYU Supercomputers. Resources Basic Usage After your account is activated: – ssh You will be logged in to an interactive.
Workshop KEK - CC-IN2P3 KEK new Grid system 27 – 29 Oct. CC-IN2P3, Lyon, France Day2 14: :55 (40min) Koichi Murakami, KEK/CRC.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
The Owner Share scheduler for a distributed system 2009 International Conference on Parallel Processing Workshops Reporter: 李長霖.
Research Computing Environment at the University of Alberta Diego Novillo Research Computing Support Group University of Alberta April 1999.
CERN - IT Department CH-1211 Genève 23 Switzerland Castor External Operation Face-to-Face Meeting, CNAF, October 29-31, 2007 CASTOR2 Disk.
M. Schott (CERN) Page 1 CERN Group Tutorials CAT Tier-3 Tutorial October 2009.
Lecture 8: 9/19/2002CS149D Fall CS149D Elements of Computer Science Ayman Abdel-Hamid Department of Computer Science Old Dominion University Lecture.
Operating Systems COMP 4850/CISG 5550 Basic Memory Management Swapping Dr. James Money.
Batch Scheduling at CERN (LSF) Hepix Spring Meeting 2005 Tim Bell IT/FIO Fabric Services.
MROrder: Flexible Job Ordering Optimization for Online MapReduce Workloads School of Computer Engineering Nanyang Technological University 30 th Aug 2013.
Faucets Queuing System Presented by, Sameer Kumar.
Cluster Configuration Update Including LSF Status Thorsten Kleinwort for CERN IT/PDP-IS HEPiX I/2001 LAL Orsay Tuesday, December 08, 2015.
CERN IT Department CH-1211 Genève 23 Switzerland t Load Testing Dennis Waldron, CERN IT/DM/DA CASTOR Face-to-Face Meeting, Feb 19 th 2009.
Issues on the operational cluster 1 Up to 4.4x times variation of the execution time on 169 cores Using -O2 optimization flag Using IBM MPI without efficient.
Oracle for Physics Services and Support Levels Maria Girone, IT-ADC 24 January 2005.
GangLL Gang Scheduling on the IBM SP Andy B. Yoo and Morris A. Jette Lawrence Livermore National Laboratory.
Western Tier 2 Site at SLAC Wei Yang US ATLAS Tier 2 Workshop Harvard University August 17-18, 2006.
Lecture 4 Page 1 CS 111 Summer 2013 Scheduling CS 111 Operating Systems Peter Reiher.
Hyukjin Kwon ESLAB SKKU Increasing memory density by using KSM.
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
PACMan: Coordinated Memory Caching for Parallel Jobs Ganesh Ananthanarayanan, Ali Ghodsi, Andrew Wang, Dhruba Borthakur, Srikanth Kandula, Scott Shenker,
Modularization of Geant4 Dynamic loading of modules Configurable build using CMake Pere Mato Witek Pokorski
Slide 1 User-Centric Workload Analytics: Towards Better Cluster Management Saurabh Bagchi Purdue University Joint work with: Subrata Mitra, Suhas Javagal,
Online Parameter Optimization for Elastic Data Stream Processing Thomas Heinze, Lars Roediger, Yuanzhen Ji, Zbigniew Jerzak (SAP SE) Andreas Meister (University.
HPC need and potential of ANSYS CFD and mechanical products at CERN A. Rakai EN-CV-PJ2 5/4/2016.
Slide 1 Cluster Workload Analytics Revisited Saurabh Bagchi Purdue University Joint work with: Subrata Mitra, Suhas Javagal, Stephen Harrell (Purdue),
CERN IT Department CH-1211 Genève 23 Switzerland Benchmarking of CPU servers Benchmarking of (CPU) servers Dr. Ulrich Schwickerath, CERN.
Gridengine Configuration review ● Gridengine overview ● Our current setup ● The scheduler ● Scheduling policies ● Stats from the clusters.
Large-scale cluster management at Google with Borg
Getting the Most out of Scientific Computing Resources
Auburn University
Title of the Poster Supervised By: Prof.*********
Condor DAGMan: Managing Job Dependencies with Condor
Getting the Most out of Scientific Computing Resources
Connecting LRMS to GRMS
Welcome to Indiana University Clusters
Component 2 6G, H, I, J, K.
WLCG Memory Requirements
CS 425 / ECE 428 Distributed Systems Fall 2016 Nov 10, 2016
THE OPERATION SYSTEM The need for an operating system
PES Lessons learned from large scale LSF scalability tests
CS 425 / ECE 428 Distributed Systems Fall 2017 Nov 16, 2017
Computing Resource Allocation and Scheduling in A Data Center
Virtualization in the gLite Grid Middleware software process
Developments in Batch and the Grid
FCT Follow-up Meeting 31 March, 2017 Fernando Meireles
NCSA Supercluster Administration
Condor and Multi-core Scheduling
Support for ”interactive batch”
Chapter 1 Introduction to Operating System Part 2
Uniprocessor Process Management & Process Scheduling
Uniprocessor Process Management & Process Scheduling
Necessary Background for OS
Presentation transcript:

Practical aspects of multi-core job submission at CERN Ricardo Silva CERN IT-FIO-FS

The batch system at CERN Platform LSF In general no free capacity in the public SLC4 resources (> 10k pending jobs) But there are significant unused dedicated resources Practical aspects of multi-core job submission at CERN - 2

Job slot model As a general rule we have one job slot per core + 1: in a 8 core machine 9 job slots (2 CPUs, 4 cores per CPU) Resource allocations (i.e. memory) are based on a job slot model: 2 GB RAM / core Practical aspects of multi-core job submission at CERN - 3

Submitting multi-core jobs (1) Without telling the system about it: bsub my_8thread_job.sh Only 1 job slot is allocated to the job 1 core / 2GB of RAM There are other 7 jobs running on the same 8 cores, the machine will be quickly overloaded LSF will stop scheduling new jobs to any machine if the utilization (load) is above a certain threshold No new jobs will be started on the machine until the utilization is below the threshold again, but these job slots will be reported as “free” (implications for WLCG) The jobs will be killed based on the resource limits defined for 1 job slot (at CERN this happens at 4GB at the moment) Practical aspects of multi-core job submission at CERN - 4

Submitting multi-core jobs (2) And telling the system about it: bsub –n 8 –span[hosts=1] my_8thr_job.sh 8 job slots are allocated to the job 8 cores / 16 GB of RAM The –span option ensures the 8 job slots/CPUs are in the same node) LSF knows the job wants to use the 8 cores and will start reserving cores on the same node before starting the job. The job slots will be marked as being used Practical aspects of multi-core job submission at CERN - 5

Conclusions The system is already capable of correctly handling multi-core jobs, if it knows how many cores the job needs Our experience supporting these jobs is very limited We need to understand the effects of large numbers of multi-core jobs on the scheduling (there is the possibility of wasting resources). We will need to tune the scheduling parameters. For effective backfilling we need accurate runtime estimation (users should use -We option) LSF is in principle also capable of supporting MPI applications but the system is not optimized (ex: infrastructure, GB ethernet) Practical aspects of multi-core job submission at CERN - 6

Questions? Practical aspects of multi-core job submission at CERN - 7