Daniele Cesini - INFN CNAF. INFN-CNAF 20 maggio 2014 CNAF 2 CNAF hosts the Italian Tier1 computing centre for the LHC experiments ATLAS, CMS, ALICE and.

Slides:



Advertisements
Similar presentations
National Grid's Contribution to LHCb IFIN-HH Serban Constantinescu, Ciubancan Mihai, Teodor Ivanoaica.
Advertisements

LNL CMS M.Biasotto, Bologna, 29 aprile LNL Analysis Farm Massimo Biasotto - LNL.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
HIGH PERFORMANCE COMPUTING ENVIRONMENT The High Performance Computing environment consists of high-end systems used for executing complex number crunching.
ATLAS computing in Geneva Szymon Gadomski, NDGF meeting, September 2009 S. Gadomski, ”ATLAS computing in Geneva", NDGF, Sept 091 the Geneva ATLAS Tier-3.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
22/04/2005Donatella Lucchesi1 CDFII Computing Status OUTLINE:  New CDF-Italy computing group organization  Usage status at FNAL and CNAF  Towards GRID:
Luca dell’Agnello INFN-CNAF FNAL, May
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Status Report on Tier-1 in Korea Gungwon Kang, Sang-Un Ahn and Hangjin Jang (KISTI GSDC) April 28, 2014 at 15th CERN-Korea Committee, Geneva Korea Institute.
Status of WLCG Tier-0 Maite Barroso, CERN-IT With input from T0 service managers Grid Deployment Board 9 April Apr-2014 Maite Barroso Lopez (at)
08/06/00 LHCb(UK) Meeting Glenn Patrick LHCb(UK) Computing/Grid: RAL Perspective Glenn Patrick Central UK Computing (what.
CCR GRID 2010 (Catania) Daniele Gregori, Stefano Antonelli, Donato De Girolamo, Luca dell’Agnello, Andrea Ferraro, Guido Guizzunti, Pierpaolo Ricci, Felice.
Status of the DESY Grid Centre Volker Guelzow for the Grid Team DESY IT Hamburg, October 25th, 2011.
Computing for HEP in the Czech Republic Jiří Chudoba Institute of Physics, AS CR, Prague.
INFN-T1 site report Andrea Chierici On behalf of INFN-T1 staff HEPiX Spring 2014.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Federico Ruggieri INFN-CNAF GDB Meeting 10 February 2004 INFN TIER1 Status.
LCG Service Challenge Phase 4: Piano di attività e impatto sulla infrastruttura di rete 1 Service Challenge Phase 4: Piano di attività e impatto sulla.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks WMSMonitor: a tool to monitor gLite WMS/LB.
UKI-SouthGrid Update Hepix Pete Gronbech SouthGrid Technical Coordinator April 2012.
1 INFN-T1 site report Andrea Chierici On behalf of INFN-T1 staff 28 th October 2009.
Tier2 Centre in Prague Jiří Chudoba FZU AV ČR - Institute of Physics of the Academy of Sciences of the Czech Republic.
INFN TIER1 (IT-INFN-CNAF) “Concerns from sites” Session LHC OPN/ONE “Networking for WLCG” Workshop CERN, Stefano Zani
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
Fabric Monitoring at the INFN Tier1 Felice Rosso on behalf of INFN Tier1 Joint OSG & EGEE Operations WS, Culham (UK)
INFN-T1 site report Andrea Chierici On behalf of INFN-T1 staff HEPiX Fall 2013.
The CMS CERN Analysis Facility (CAF) Peter Kreuzer (RWTH Aachen) - Stephen Gowdy (CERN), Jose Afonso Sanches (UERJ Brazil) on behalf.
IHEP(Beijing LCG2) Site Report Fazhi.Qi, Gang Chen Computing Center,IHEP.
Computing Jiří Chudoba Institute of Physics, CAS.
Outline: Status: Report after one month of Plans for the future (Preparing Summer -Fall 2003) (CNAF): Update A. Sidoti, INFN Pisa and.
INFN-T1 site report Luca dell’Agnello On behalf ot INFN-T1 staff HEPiX Spring 2013.
Scientific Computing in PPD and other odds and ends Chris Brew.
ROC managers meeting, Barcelona, Luciano Gaido (thanks to Paolo Veronesi for the slides) ROC-IT status.
CERN IT Department CH-1211 Genève 23 Switzerland t SL(C) 5 Migration at CERN CHEP 2009, Prague Ulrich SCHWICKERATH Ricardo SILVA CERN, IT-FIO-FS.
IT-INFN-CNAF Status Update LHC-OPN Meeting INFN CNAF, December 2009 Stefano Zani 10/11/2009Stefano Zani INFN CNAF (TIER1 Staff)1.
PADME Kick-Off Meeting – LNF, April 20-21, DAQ Data Rate - Preliminary estimate Tentative setup: all channels read with Fast ADC 1024 samples, 12.
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
WNoDeS – a Grid/Cloud Integration Framework Elisabetta Ronchieri (INFN-CNAF) for the WNoDeS Project
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
Farming Andrea Chierici CNAF Review Current situation.
Elastic CNAF Datacenter extension via opportunistic resources INFN-CNAF.
KIT – Universität des Landes Baden-Württemberg und nationales Forschungszentrum in der Helmholtz-Gemeinschaft Steinbuch Centre for Computing
Tier2 Centre in Prague Jiří Chudoba FZU AV ČR - Institute of Physics of the Academy of Sciences of the Czech Republic.
Introduction to Data Analysis with R on HPC Texas Advanced Computing Center Feb
Mass Storage Systems for the Large Hadron Collider Experiments A novel approach based on IBM and INFN software A. Cavalli 1, S. Dal Pra 1, L. dell’Agnello.
Validation tests of CNAF storage infrastructure Luca dell’Agnello INFN-CNAF.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
Academia Sinica Grid Computing Centre (ASGC), Taiwan
Dynamic Extension of the INFN Tier-1 on external resources
Extending the farm to external sites: the INFN Tier-1 experience
Experiments and User Support
HPC usage and software packages
Daniele Cesini – INFN-CNAF - 19/09/2017
Heterogeneous Computation Team HybriLIT
INFN Computing infrastructure - Workload management at the Tier-1
Andrea Chierici On behalf of INFN-T1 staff
The INFN TIER1 Regional Centre
Infrastructure for testing accelerators and new
FCT Follow-up Meeting 31 March, 2017 Fernando Meireles
Bernd Panzer-Steindel CERN/IT
Luca dell’Agnello Daniele Cesini GDB - 13/12/2017
Vladimir Sapunenko On behalf of INFN-T1 staff HEPiX Spring 2017
Luca dell’Agnello, Daniele Cesini – INFN-CNAF CCR - 23/05/2017
This work is supported by projects Research infrastructure CERN (CERN-CZ, LM ) and OP RDE CERN Computing (CZ /0.0/0.0/1 6013/ ) from.
Presentation transcript:

Daniele Cesini - INFN CNAF

INFN-CNAF 20 maggio 2014 CNAF 2 CNAF hosts the Italian Tier1 computing centre for the LHC experiments ATLAS, CMS, ALICE and LHCb.... … but also one of the main Italian processing facilities for several other experiments: CDF, SuperB, KLOE, LHCf, NA62 AMS, ARGO, AUGER, FERMI, MAGIC, PAMELA, Icarus, Xenon100, Borexino, Gerda, CTA, Opera, Darkside, Cuore, Virgo Involved in a number of Grid and Cloud national and international projects ─ WLCG ─ EGI-Inspire ─ MCLOUD ─ Open City Platform ─ !CHAOS ─ many other in preparation towards H2020  collaboration with other disciplines

Daniele Cesini - INFN CNAF INFN-T1 in Numbers 24 June 2014 CNAF 3 CPU (HS06) Disk (TB-N) Tape (TB) For the LHC VOs is the second WLCG computing centre, after CERN, in terms of CPU time Normalised KSI2k

Daniele Cesini - INFN CNAF HPC at CNAF In 2014 CNAF started to operate an HPC cluster but remains mainly an High Throughput Center –In the short term is not going to become a supecomputing center The HPC cluster was created as a collaboration between –Physics Dept. Bologna University –INFN-Bologna –INFN-CNAF –INFN-COKA project To support local HPC users and their close collaborators To offer an HPC testbed for developers before moving to other facilities that operates at a greater scale To acquire expertise at CNAF in a going-parallel world 24 June 2014 CNAF 4

Daniele Cesini - INFN CNAF HPC Cluster at CNAF Status and Perspectives Daniele Cesini - INFN-CNAF

Daniele Cesini - INFN CNAF People and Collaboration The HPC cluster at CNAF was created thanks to a collaboration between: –Physics Dept. Bologna University –INFN-Bologna –INFN-CNAF –INFN-COKA project Many operative contributions and support from: –INFN-BO: Vincenzo Vagnoni (ib, storage, gpfs and main screwdriver interventions ) –UNIBO: Stefano Sinigardi and Francesco Rossi (gpu, ib, cabling, main testers) –CNAF Farming: Stefano Dal Pra (lsf), Massimo Donatelli (ldap and user account) –CNAF Storage: Vladimir Sapunenko (gpfs) –CNAF Network: Stefano Zani, Lorenzo Chiarelli (switch and uplink) –CNAF Infrastructure: Michele Onofri, Andrea Ferraro (power and cooling) –CNAF R&D: Matteo Manzali (mic) 24 June 2014 CNAF 6

Daniele Cesini - INFN CNAF The Cluster Status 13 Worker Nodes –CPU: HT cores 320 HT cores E HT cores X HT cores E –15 GPUs: 8 Tesla K Tesla K20 2x(4GRID K1) –2 MICs: 2 x Xeon Phi June 2014 CNAF 7 2 disks server 60 TB shared disk space 4 TB shared home 1 Infiniband QLOGIC switch 18 ports 1 Eth switch BROCADE 48x1Gb/s + 8x10Gb/s CPUGPUMICTOT TFLOPS (DP - PEAK) Disk server 60 TB 1Gb/s 10Gb/s 2x10Gb/s Worker Nodes IB ETH

Daniele Cesini - INFN CNAF Software Stack LSF9.1 to access the cluster GPFS for the shared file systems Compilers –CUDA5.5, CUDA6.0 –GCC4.6 (system) Compiled by us 4.7/4.8 – working on 4.9 –icc, ifort –Python 2.7/3.3, ipython –Openmpi –cmake, cmake28, gdb Others libs –HDF5, FFTW, BLAS, CUBLAS, GSL Tools –Virtualenv, matplotlib, git, hg, paraview, numpy, scipy, pylab, gnuplot 24 June 2014 CNAF 8

Daniele Cesini - INFN CNAF Usage Statistics LSF –2 batch queues (short, infinite) –1 interactive –Flat configuration no priorities, no shares –About 1k jobs (last 2 months) –26 registered users 13 active (last month) 24 June 2014 CNAF 9 90MB/s (aggregated)

Daniele Cesini - INFN CNAF Open Issues Improve availability and storage performance –Fix the redundant disk controller problems (this week, no down) –Upgrade the storage-switch link to 2x10Gb/s Study the possibility to use infiniband also for storage data –Add monitoring and alarm systems Accounting for CPU and GPU/MIC usage Dedicated GPUs queue with reserved cores Fix the Infiniband problems on 3 nodes –Add a new 24 core 2xTeslaK20 machine Need the IB adapter –Change the IB adapters of the two X5650 servers (24x2 core) 24 June 2014 CNAF 10

Daniele Cesini - INFN CNAF Cluster Upgrade The physical infrastructure is not the limiting factor it depends on budget For further upgrades in the short term: –4 working ports are still available on the Infiniband switch –6x3 new ports can be enabled with just a license upgrade: 3x480 euro+VAT 24 June 2014 CNAF 11

Daniele Cesini - INFN CNAF Links and Contacts hpc-support lists.cnaf.infn.it hpc-users lists.cnaf.infn.it 24 June 2014 CNAF 12