The INFN Grid Project Successful Grid Experiences at Catania Roberto Barbera University of Catania and INFN Workshop CCR Rimini, 08.05.2007.

Slides:



Advertisements
Similar presentations
Alastair Dewhurst, Dimitrios Zilaskos RAL Tier1 Acknowledgements: RAL Tier1 team, especially John Kelly and James Adams Maximising job throughput using.
Advertisements

Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Workload Management Massimo Sgaravatto INFN Padova.
1 Status of the ALICE CERN Analysis Facility Marco MEONI – CERN/ALICE Jan Fiete GROSSE-OETRINGHAUS - CERN /ALICE CHEP Prague.
DONVITO GIACINTO (INFN) ZANGRANDO, LUIGI (INFN) SGARAVATTO, MASSIMO (INFN) REBATTO, DAVID (INFN) MEZZADRI, MASSIMO (INFN) FRIZZIERO, ERIC (INFN) DORIGO,
Enabling Grids for E-sciencE Medical image processing web portal : Requirements analysis. An almost end user point of view … H. Benoit-Cattin,
Quantitative Methodologies for the Scientific Computing: An Introductory Sketch Alberto Ciampa, INFN-Pisa Enrico Mazzoni, INFN-Pisa.
FP6−2004−Infrastructures−6-SSA E-infrastructure shared between Europe and Latin America Pilot Test-bed Operations and Support Work.
Sun Grid Engine. Grids Grids are collections of resources made available to customers. Compute grids make cycles available to customers from an access.
Southgrid Technical Meeting Pete Gronbech: 16 th March 2006 Birmingham.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
INFSO-RI Enabling Grids for E-sciencE Logging and Bookkeeping and Job Provenance Services Ludek Matyska (CESNET) on behalf of the.
GridKa SC4 Tier2 Workshop – Sep , Warsaw Tier2 Site.
CHEP'07 September D0 data reprocessing on OSG Authors Andrew Baranovski (Fermilab) for B. Abbot, M. Diesburg, G. Garzoglio, T. Kurca, P. Mhashilkar.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
Cracow Grid Workshop October 2009 Dipl.-Ing. (M.Sc.) Marcus Hilbrich Center for Information Services and High Performance.
Enabling Grids for E-sciencE EGEE-III INFSO-RI Using DIANE for astrophysics applications Ladislav Hluchy, Viet Tran Institute of Informatics Slovak.
Chapter 10 Chapter 10: Managing the Distributed File System, Disk Quotas, and Software Installation.
Certification and test activity IT ROC/CIC Deployment Team LCG WorkShop on Operations, CERN 2-4 Nov
EGEE-III INFSO-RI Enabling Grids for E-sciencE Overview of STEP09 monitoring issues Julia Andreeva, IT/GS STEP09 Postmortem.
FESR Consorzio COMETA - Progetto PI2S2 The COMETA consortium and its activities for Grid adoption by Industry in the context of.
EGEE-II INFSO-RI Enabling Grids for E-sciencE The GILDA training infrastructure.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
Conference name Company name INFSOM-RI Speaker name The ETICS Job management architecture EGEE ‘08 Istanbul, September 25 th 2008 Valerio Venturi.
INFSO-RI Enabling Grids for E-sciencE EGEE is a project funded by the European Union under contract INFSO-RI Grid Accounting.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks The GILDA t-Infrastructure Roberto Barbera.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
IHEP(Beijing LCG2) Site Report Fazhi.Qi, Gang Chen Computing Center,IHEP.
Peter Couvares Associate Researcher, Condor Team Computer Sciences Department University of Wisconsin-Madison
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) Giuseppe Andronico INFN Sez. CT / Consorzio COMETA Beijing,
Tier 3 Status at Panjab V. Bhatnagar, S. Gautam India-CMS Meeting, July 20-21, 2007 BARC, Mumbai Centre of Advanced Study in Physics, Panjab University,
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Grid2Win : gLite for Microsoft Windows Roberto.
FESR Consorzio COMETA - Progetto PI2S2 GRID Success Stories in Sicily (The TriGrid VL and PI2S2 Projects) Roberto Barbera University.
6 march Building the INFN Grid Proposal outline a.ghiselli,l.luminari,m.sgaravatto,c.vistoli INFN Grid meeting, milano.
INFSO-RI Enabling Grids for E-sciencE Policy management and fair share in gLite Andrea Guarise HPDC 2006 Paris June 19th, 2006.
IAG – Israel Academic Grid, EGEE and HEP in Israel Prof. David Horn Tel Aviv University.
BaBar Cluster Had been unstable mainly because of failing disks Very few (
ROC managers meeting, Barcelona, Luciano Gaido (thanks to Paolo Veronesi for the slides) ROC-IT status.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
Latest Improvements in the PROOF system Bleeding Edge Physics with Bleeding Edge Computing Fons Rademakers, Gerri Ganis, Jan Iwaszkiewicz CERN.
INFN GRID Production Infrastructure Status and operation organization Cristina Vistoli Cnaf GDB Bologna, 11/10/2005.
FESR Consorzio COMETA Giuseppe Andronico INFN Sez. CT & Consorzio COMETA Workshop Grids vs. Clouds Beijing, Consorzio.
First South Africa Grid Training June 2008, Catania (Italy) GILDA t-Infrastructure Valeria Ardizzone INFN Catania.
Lecture 4 Page 1 CS 111 Summer 2013 Scheduling CS 111 Operating Systems Peter Reiher.
A Computing Tier 2 Node Eric Fede – LAPP/IN2P3. 2 Eric Fede – 1st Chinese-French Workshop Plan What is a Tier 2 –Context and definition To be a Tier 2.
II EGEE conference Den Haag November, ROC-CIC status in Italy
Daniele Cesini - INFN CNAF. INFN-CNAF 20 maggio 2014 CNAF 2 CNAF hosts the Italian Tier1 computing centre for the LHC experiments ATLAS, CMS, ALICE and.
1/3/2006 Grid operations: structure and organization Cristina Vistoli INFN CNAF – Bologna - Italy.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
INFN/IGI contributions Federated Clouds Task Force F2F meeting November 24, 2011, Amsterdam.
DGAS Distributed Grid Accounting System INFN Workshop /05/1009, Palau Giuseppe Patania Andrea Guarise 6/18/20161.
FESR Trinacria Grid Virtual Laboratory The Trinacria Grid Virtual Laboratory Roberto Barbera University of Catania and INFN Grid Open Day.
NERSC/LBNL at LBNL in Berkeley October 2009 Site Report Roberto Gomezel INFN 1.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Enabling Grids for E-sciencE Claudio Cherubino INFN DGAS (Distributed Grid Accounting System)
Dynamic Extension of the INFN Tier-1 on external resources
Workload Management Workpackage
Belle II Physics Analysis Center at TIFR
INFN Computing infrastructure - Workload management at the Tier-1
Towards GLUE Schema 2.0 Sergio Andreozzi INFN-CNAF Bologna, Italy
Technical Board Meeting, CNAF, 14 Feb. 2004
Roberto Barbera (a nome di Livia Torterolo)
CREAM-CE/HTCondor site
QoS in the Tier1 batch system(LSF)
湖南大学-信息科学与工程学院-计算机与科学系
Wide Area Workload Management Work Package DATAGRID project
Sun Grid Engine.
Presentation transcript:

The INFN Grid Project Successful Grid Experiences at Catania Roberto Barbera University of Catania and INFN Workshop CCR Rimini,

Enabling Grids for E-sciencE Workshop CCR, Rimini, Outline Catania –Network connection –Catania in the Grid Infrastructures –Production site (ALICE Tier-2) –GILDA –TriGrid VL –PI2S2 Management of the site resources –Goals –Configuration & Policies –Monitoring –Usage statistics Summary & Conclusions

Enabling Grids for E-sciencE Workshop CCR, Rimini, Catania Catania Network Connection (2/2)

Enabling Grids for E-sciencE Workshop CCR, Rimini, Half of the total available bandwidth exploited Usage of the dedicated 1 Gb/s link Rather continuous usage (mostly by ALICE)

Enabling Grids for E-sciencE Workshop CCR, Rimini, Catania in the EGEE Grid Map

Enabling Grids for E-sciencE Workshop CCR, Rimini, Catania in other Grid Infrastructures INFN (Catania, CNAF, Roma3) Poland GRNet CNIC IHEP INFN -Catania MA-GRID INFN -Roma3 INFN -Catania INFN - CNAF ULAKBIM CYNET IUCC GRNET Univ. of Tunis UoM ERI

Enabling Grids for E-sciencE ~ € in 3 years ! ~350 FTEs! (2/3 new employees) More than 2000 CPUs More than 300 TBytes The Sicilian e-Infrastructure

Enabling Grids for E-sciencE Workshop CCR, Rimini, Catania in the TriGrid VL Grid Map

Enabling Grids for E-sciencE Workshop CCR, Rimini, Catania in the GILDA t-Infrastructure

Enabling Grids for E-sciencE Workshop CCR, Rimini, Catania Computing Room (1/2) 3D Model of Catania Data Center Full Area: ~200 m 2 Area # 1 10 racks / 40 kW UPS/PDU Area # 2 80 kW UPS/PDU Area # 2 13 racks Area # 2 80 kW Air Cond. with ~100 kW external chiller

Enabling Grids for E-sciencE Workshop CCR, Rimini, Catania Computing Room (2/2) Area # 2 Area # 1 Security system Area # 2 Fire estinguisher system

Enabling Grids for E-sciencE Workshop CCR, Rimini, INFNGRID Catania Computing and Storage 1. ~270 cores 2. ~280 GB of memory 3. LSF 6.1 as LRMS 1. ~75 TB of raw disk storage FC-2-SATA 2.DPM over GPFS This includes the recent delivery of 40 cores and 50 TB for ALICE

Enabling Grids for E-sciencE Workshop CCR, Rimini, GILDA Computing and Storage cores GB of memory 3. Torque+MAUI as LRMS 1. ~ 3 TB of raw disk storage 2.DPM over GPFS

Enabling Grids for E-sciencE Workshop CCR, Rimini, TriGrid VL Computing and Storage (Catania site only) cores AMD Opteron GB of memory 3. LSF 6.1 HPC as LRMS TB of raw disk storage FC-2-SATA 2.DPM over GPFS New tender for ~260 cores and ~55 TB (for all TriGrid VL sites) expected to start by the end of May

Enabling Grids for E-sciencE Workshop CCR, Rimini, PI2S2 Project Computing and Storage (Catania site only) 2 IBM BladeCenter H enclosures 19 IBM LS21 “blades” 76 cores AMD Opteron 2218 rev. F 152 GB of RAM (2 GB/core) ~ 48.8 mW/SpecInt2000 at full load ! G-Ethernet service network CISCO Topspin Infiniband-4X additional low-latency network for HPC applications LSF 6.1 HPC included ! 1 IBM DS4200 Storage System FC-2-SATA technology 25 TB (raw) of storage Expandability up to ~100 TB GPFS distributed/parallel file sytem included ! New tender for ~1500 cores and ~130 TB (for all PI2S2 sites) expected to start by the end of May

Enabling Grids for E-sciencE Workshop CCR, Rimini, “Owners”, “Users”, and “Goals” Owners of Catania resources: –INFN (ALICE, CMS, GILDA, Theory Group, TriGrid VL Project) –Consorzio COMETA (PI2S2 Project) 31 Virtual Organizations authorized at Catania ! –alice, atlas, babar, bio, biomed, cdf, cms, cometa, compchem, dteam, edteam, eela, egrid, enea, esr, euchina, euindia, eumed, gilda, gridit, inaf, infngrid, ingv, lhcb, magic, planck, theophys, trigrid, virgo, zeus + Local Users (~20) Goals: –Give access to everybody in order to maximize the usage of the site –Let “owners” use “their” resources with zero wait time

Enabling Grids for E-sciencE Workshop CCR, Rimini, Wisdom Rules (by direct experience) 1.Be “open” to new technologies (usually they are not the problem, they are rather the solution). 2.Centralize system administration: create a core team of experts from the beginning and carry on with it. Changes can induce disruption. 3.Share experience and knowledge: no single super-expert knowing crucial element(s) of site administration. 4.Foster bi-directional know-how transfer to ensure long-term sustainability: from permanent staff to temporary younger personnel and viceversa. 5.Be wise and far-sighted in choices: don’t adopt things (especially LRMS) simply because you know them but because their ensure you the biggest possibile adaptability, configurability and scalability (see item 1.). 6.“Remotize” the control/monitoring of your site as much as possible in order to have it managed by a smaller number of persons.

Enabling Grids for E-sciencE The LRMS choice at Catania: LSF Schedules the processing of the job based on rules, policies and priorities (e.g., holding a job until a certain license is available, preempting the queue, fairsharing, event driven, calendar driven, in parallel, etc). Manages, monitors and balances all jobs until completion (e.g., restarting and requeuing jobs if system fails or resources become unavailable). Notifies users when done and logs statistics (for accounting & reporting). All this is done transparently to the user!

Enabling Grids for E-sciencE Workshop CCR, Rimini, LSF and the Catania “goals” Centralization of management as a commodity –LSF Administrators users can do every administration task Fault tolerance Rationalization of the coexistence of different groups of machines belonging to different owners: –Grid Users with jobs submitted through gLite –Local Users with jobs submitted through LSF from frontend nodes Maximize the use of every CPU around the clock with as less empty slots as possible

Enabling Grids for E-sciencE Workshop CCR, Rimini, Partitioning the Catania INFNGRID Farm Group Nodes –ALICE Production (alifarm11-67) –CMS Group (cmsfarm01-09) –GR1 Group(gr1farm1-4) –TheoPhys Local Group (dz01-06 dt05-16 ar01-12) From lsb.hosts file Begin Host default ! () () () () () # Default limits dz / () () # theogroup frontend limits End Host Begin HostGroup all_nodes ( alifarm10 alifarm11... alifarm67) gr1farm ( gr1farm2 gr1farm3 gr1farm4 ) cmsfarm ( cmsfarm1 cmsfarm cmsfarm9 ) theofarm ( dz dt16 ) theo4gridfarm ( dz02 ar07 ar08 ar09 dt05 )..... End HostGroup

Enabling Grids for E-sciencE Workshop CCR, Rimini, Addressing scalabilty and fault tolerance Centralized managed with LSF Batch admin commands performed by LSF admin users –Commands can be run from every host transparently –Immediate activation after a reconfig action –Unique configuration directory where config files are shared Fault tolerance –NFS server host, for LSF binaries and config files shared, not belonging to any cluster host (external) –Master List contains 3 hosts for elections in case of fault  LSF_MASTER_LIST="grid012 alifarm14 alifarm15" –License fault tolerance thanks to CNAF FlexLM hosts redundance: 

Enabling Grids for E-sciencE Workshop CCR, Rimini, Addressing optmization and “ownership” issues Rationalization and Job Optimization implemented with LSF Scheduling Policies –Hosts defined in groups in order to have associations to “owners” –Users defined in groups having distinct submission policies –Queues defined for different kinds of Grid jobs and Users’ Groups with direct relation with the ownership of the hosts –Concept of “owned” and “guest” queues implemented

Enabling Grids for E-sciencE Workshop CCR, Rimini, Scheduling Policies FCFS –Jobs are dispatched on a first-come-first-served basis FAIRSHARE –Jobs are dispatched on a fair-share basis PRE-EMPTIVE/PRE-EMPTABLE –May stop/be stopped jobs in lower/higher priority queues SLA –Service Level Agreements EXCLUSIVE –Job has the exclusive use of an execution host

Enabling Grids for E-sciencE Workshop CCR, Rimini, Begin Queue QUEUE_NAME = short PRIORITY = 70 HOSTS = hostGroupC # potential conflict PREEMPTION = PREEMPTIVE[normal] End Queue Begin Queue QUEUE_NAME = normal PRIORITY = 40 HOSTS = hostGroupC # potential conflict PREEMPTION = PREEMPTABLE[short] End Queue Preemption Scheduling Example: Preemption has been added in this way:

Enabling Grids for E-sciencE Workshop CCR, Rimini, short 70 normal Job submitted 5. Job dispatched & running 4. Job suspended 6. Job completed 7. Job resumed 2. Job dispatched & running 1. Job submitted hostA with 1 job slot available Preemption has been added to the definition of queues PREEMPTION=PREEMPTIVE[normal]PREEMPTION=PREEMPTABLE[short] How does Preemption Scheduling work ?

Enabling Grids for E-sciencE Workshop CCR, Rimini, Begin Queue QUEUE_NAME = lhcb PRIORITY = 15 NICE = 5 #PJOB_LIMIT = #HJOB_LIMIT = 2 QJOB_LIMIT = 30 PREEMPTION = PREEMPTABLE[workq alice gr1cmsq theoqs theoqm theoql theoqi] INTERACTIVE = NO CPULIMIT = 60:00 RUNLIMIT = 96:00 RES_REQ = type==LINUX86 USERS = lhcb_pool JOB_STARTER = /sw/lsf/scripts/jobstarter-lsf-lcg.sh HOSTS = all_nodes+2 theofarm cmsfarm DESCRIPTION = LHCb dedicated Grid infinite queue End Queue Jobs coming from EGEE users belonging to the LHCb VO Implementation of Preemption Scheduling (1/3)

Enabling Grids for E-sciencE Workshop CCR, Rimini, Begin Queue QUEUE_NAME = gr1cmsq PRIORITY = 55 NICE = 5 #PJOB_LIMIT = HJOB_LIMIT = 2 PREEMPTION = PREEMPTIVE RES_REQ = type==LINUX86 USERS = cms_group INTERACTIVE = NO HOSTS = gr1farm cmsfarm DESCRIPTION = Dedicated Queue for jobs of gr1-cms group End Queue Jobs coming from LOCAL users belonging to the CMS group Implementation of Preemption Scheduling (2/3)

Enabling Grids for E-sciencE Workshop CCR, Rimini, Begin Queue QUEUE_NAME = alice PRIORITY = 40 NICE = 5 PREEMPTION = PREEMPTIVE[lhcb] PREEMPTABLE[gr1cmsq theoqs theoqm theoql theoqi] #PJOB_LIMIT = #HJOB_LIMIT = 2 #QJOB_LIMIT = 65 INTERACTIVE = NO CPULIMIT = 48:00 RUNLIMIT = 72:00 RES_REQ = type==LINUX86 USERS = alice_pool JOB_STARTER = /sw/lsf/scripts/jobstarter-lsf-lcg.sh HOSTS = all_nodes theofarm cmsfarm DESCRIPTION = Alice dedicated Grid infinite queue End Queue Jobs coming from EGEE users belonging to the ALICE VO Implementation of Preemption Scheduling (3/3)

Enabling Grids for E-sciencE Workshop CCR, Rimini, Preemption vs. SLA PreemptionSLA Pros Cons Every job slot is filled with suspending actions based on priorities in case of host competition Automatic Restart of the suspended Jobs No need Admin actions after conf The Farm is full & hot 24/7 Remote possible conflicts over host memory caused by not correct memory management of (new) running job Particular config to be adapted for parallel jobs [done!] Service Level Agreements based on projects and timelines (what & when) and priorities Easy project accounting Deadline, Velocity and Throughput goals can be combined Monitoring of progress & tracking historical behavior of SLA Needs LSF Administrator to configure new SLA for new projects (may be frequent) If a VO has reaches its SLA slot limit, jobs can not run any more even if the farm is empty After or during project completion the entire Farm may be not fully used SLA jobs cannot be preempted A goal can be missed because of a misconfiguration

Enabling Grids for E-sciencE Workshop CCR, Rimini, ALICE jobs running on TriGrid VL With the same preemption policies an ALICE “guest” queue has been created in the Catania site of the TriGrid VL infrastructure: –When the TriGrid clusters/queues are empty ALICE jobs can run –If TriGrid jobs are coming in host competition the ALICE jobs running are temporarily suspended to be restarted after TriGrid jobs have finished

Enabling Grids for E-sciencE Workshop CCR, Rimini, LSF Monitoring Tool Monitoring tools available as packages!

Enabling Grids for E-sciencE Workshop CCR, Rimini, Usage of the Catania site (1/3) (All supported VO’s – last year)

Enabling Grids for E-sciencE Workshop CCR, Rimini, Usage of the Catania site (2/3) (All supported VO’s – last month) TriGrid VL CE INFNGRID CE TriGrid VL CE

Enabling Grids for E-sciencE Workshop CCR, Rimini, Usage of the Catania site (3/3) (ALICE only - last 6 months) Globally, Catania Tier-2 (red circles) is contributing to ALICE like a Tier-1 (black circles)

Enabling Grids for E-sciencE Workshop CCR, Rimini, Summary & Conclusions Catania is a multi-environment Grid site The ALICE Tier-2 is one of the most important services offered by the site but it is NOT the only one. We support tens of other VOs as well as many local users. With the implementation of some “wisdom rules” the site is able to exploit as much as possible all the resources, even if they belong to different “owners”, and maximizes its usage.

Enabling Grids for E-sciencE Workshop CCR, Rimini, Questions…