Andrea Chierici On behalf of INFN-T1 staff

Slides:



Advertisements
Similar presentations
Tony Doyle - University of Glasgow GridPP EDG - UK Contributions Architecture Testbed-1 Network Monitoring Certificates & Security Storage Element R-GMA.
Advertisements

Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
INFN-T1 site report Giuseppe Misurelli On behalf of INFN-T1 staff HEPiX Spring 2015.
March 27, IndiaCMS Meeting, Delhi1 T2_IN_TIFR of all-of-us, for all-of-us, by some-of-us Tier-2 Status Report.
Status Report on Tier-1 in Korea Gungwon Kang, Sang-Un Ahn and Hangjin Jang (KISTI GSDC) April 28, 2014 at 15th CERN-Korea Committee, Geneva Korea Institute.
April 2001HEPix/HEPNT1 RAL Site Report John Gordon CLRC, UK.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
INFN-T1 site report Andrea Chierici On behalf of INFN-T1 staff HEPiX Spring 2014.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
28 April 2003Imperial College1 Imperial College Site Report HEP Sysman meeting 28 April 2003.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
1 INFN-T1 site report Andrea Chierici On behalf of INFN-T1 staff 28 th October 2009.
INFN TIER1 (IT-INFN-CNAF) “Concerns from sites” Session LHC OPN/ONE “Networking for WLCG” Workshop CERN, Stefano Zani
KOLKATA Grid Site Name :- IN-DAE-VECC-02Monalisa Name:- Kolkata-Cream VO :- ALICECity:- KOLKATACountry :- INDIA Shown many data transfers.
INFN-T1 site report Andrea Chierici On behalf of INFN-T1 staff HEPiX Fall 2013.
6. Juli 2015 Dietrich Liko Physics Computing 114. Vorstandssitzung.
INFN-T1 site report Andrea Chierici, Vladimir Sapunenko On behalf of INFN-T1 staff HEPiX spring 2012.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
Network infrastructure at FR-CCIN2P3 Guillaume Cessieux – CCIN2P3 network team Guillaume. cc.in2p3.fr On behalf of CCIN2P3 network team LHCOPN.
Tier-1 Andrew Sansum Deployment Board 12 July 2007.
RAL PPD Tier 2 (and stuff) Site Report Rob Harper HEP SysMan 30 th June
INFN-T1 site report Luca dell’Agnello On behalf ot INFN-T1 staff HEPiX Spring 2013.
A. Mohapatra, T. Sarangi, HEPiX-Lincoln, NE1 University of Wisconsin-Madison CMS Tier-2 Site Report D. Bradley, S. Dasu, A. Mohapatra, T. Sarangi, C. Vuosalo.
INFN-T1 site report Andrea Chierici On behalf of INFN-T1 staff HEPiX Fall 2015.
Development of a Tier-1 computing cluster at National Research Centre 'Kurchatov Institute' Igor Tkachenko on behalf of the NRC-KI Tier-1 team National.
IT-INFN-CNAF Status Update LHC-OPN Meeting INFN CNAF, December 2009 Stefano Zani 10/11/2009Stefano Zani INFN CNAF (TIER1 Staff)1.
Farming Andrea Chierici CNAF Review Current situation.
Elastic CNAF Datacenter extension via opportunistic resources INFN-CNAF.
NERSC/LBNL at LBNL in Berkeley October 2009 Site Report Roberto Gomezel INFN 1.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
RAL Site Report HEP SYSMAN June 2016 – RAL Gareth Smith, STFC-RAL With thanks to Martin Bly, STFC-RAL.
INFN Site Report R.Gomezel November 5-9,2007 The Genome Sequencing University St. Louis.
Stato del Tier1 Luca dell’Agnello 11 Maggio 2012.
Australia Site Report Lucien Boland Goncalo Borges Sean Crosby
Dynamic Extension of the INFN Tier-1 on external resources
Extending the farm to external sites: the INFN Tier-1 experience
WLCG IPv6 deployment strategy
COMPUTING FOR ALICE IN THE CZECH REPUBLIC in 2016/2017
DPM at ATLAS sites and testbeds in Italy
Partner Advantages of FlexPod Integration with Tech Data (FIT)
Bob Ball/University of Michigan
LCG Service Challenge: Planning and Milestones
INFN CNAF TIER1 Network Service
Mattias Wadenstein Hepix 2012 Fall Meeting , Beijing
Daniele Cesini – INFN-CNAF - 19/09/2017
Operations and plans - Polish sites
INFN Computing infrastructure - Workload management at the Tier-1
Database Services at CERN Status Update
LCG Deployment in Japan
CC - IN2P3 Site Report Hepix Spring meeting 2011 Darmstadt May 3rd
INFN CNAF TIER1 and TIER-X network infrastructure
Update on Plan for KISTI-GSDC
UK GridPP Tier-1/A Centre at CLRC
The INFN TIER1 Regional Centre
Oxford Site Report HEPSYSMAN
Deployment of IPv6-only CPU on WLCG – an update from the HEPiX IPv6 WG
Olof Bärring LCG-LHCC Review, 22nd September 2008
HPEiX Spring RAL Site Report
Update from the HEPiX IPv6 WG
NET2.
The INFN Tier-1 Storage Implementation
GridPP Tier1 Review Fabric
Bernd Panzer-Steindel CERN/IT
Luca dell’Agnello Daniele Cesini GDB - 13/12/2017
Vladimir Sapunenko On behalf of INFN-T1 staff HEPiX Spring 2017
ETHZ, Zürich September 1st , 2016
RHUL Site Report Govind Songara, Antonio Perez,
QMUL Site Report by Dave Kant HEPSYSMAN Meeting /09/2019
Presentation transcript:

Andrea Chierici On behalf of INFN-T1 staff INFN-T1 site report Andrea Chierici On behalf of INFN-T1 staff

New Logo! Andrea Chierici

Outline Network Data management & Storage Farming Projects and activities Andrea Chierici

Network

Current Status WAN Connectivity: LHCOPN+LHCONE shared physical link: 6x10Gb/s LHCOPN: 4x10Gb/s dedicated link to CERN LHCONE: Potentially 60Gb/s (Geant Peering at 100Gb/s in Milan) General IP: 2x10Gb/s INFN TIER-1 WAN upgrade to 100 Gb/s starting from the end of this year (GARR has completed the Milan-Bologna link upgrade). LAN evolution: 2 new CORE Switches (Cisco Nexus 9516) Acquisition of 2 modules (32 x 100Gb Ethernet Ports) Tender completed and delivery in 1 month Network down required for the substitution of the fabric modules and the upgrade of the existing interface modules to insert the new “EX series modules” Andrea Chierici

Network diagram General IP LHC OPN/ONE Desk Resources 2x10Gb/s Desk Resources 3x10Gb/s 3x10Gb/s Cisco7600 VPC Link Nexus 9516 Nexus 9516 New Storage disk servers will be connected at 2x 100Gb/s. 4x40Gb/s Most Recent Computing resources Every disk server or farming switch connected to both core switches 3x40Gb/s (6x40Gb/s in total) Nexus 7018 Old “single homed” resources Andrea Chierici

Developments 100Gb/s connection should allow us to reduce the number of disk servers Today this solution is still expensive: Using 100G Base-SR4 (multimode)  expensive MPO cabling expansion Using 100G Base-LR (single mode)  expensive transceivers IPv6 deployment ongoing Dual Stack on CEs working GridFTP and XRootD coming next Andrea Chierici

Data Management & Storage

Storage: state of the art 23 PB usable (29 PB raw) on disks 56 PB on tapes Provided to Tier-1 services for all 4 LHC experiments Tier-1 services for several non-LHC experiments General data and computing services for ~30 HEP and Astrophysics experiments Andrea Chierici

Latest news and issues 2016 tender finally in Production All non-LHC experiments at pledge 2017 tender behind schedule  Should be ready by Nov. 2017, 100Gbit servers Had to refurbish in-house 2 old DDN storage appliances Replaced old disks with bigger ones No maintenance, working in mirror-mode In production for ATLAS 2 weeks ago In production for LHCb last week Andrea Chierici

IBM Licensing issues GPFS and TSM licenses are a major concern Working on settling a new deal Licenses are very expensive We haven’t found a convenient and affordable alternative For TSM we moved from active-active configuration to active-stand-by for all HSM nodes reducing by a half the cost of the support Andrea Chierici

Farming

Computing resources Farm power: 220K HS06 2017 tender: still to be assigned Probably AMD EPYC Increased requests compared to the past ssd disks 10Gbit network At least 16 physical cores Sadly only one bidder Andrea Chierici

Farm external extension In 2018 many internal resources will go out-of-warranty We will not have enough money to replace the whole bunch We will borrow some pledges from CINECA “Marconi” cluster dual 2697v4 CPUs (2x 18 cores), 128 GB RAM, 2x 1TB SAS disks, 2x 10GbE links Adopt BARI ReCaS approach (these nodes should be phased-out too) Andrea Chierici

Issues and developments LSF “Job exit threshold exceeded” issue suffered during summer Investigating grid middleware on centos7 BDII seems ready WNs released on august, lack of documentation singularity “Rumors” on other sites upgrading Are experiments ready for centos7? Andrea Chierici

Projects and activities

Major Projects ISO 27001 certification for a subset of racks in the computing room Several regulations to respect, strict activity control Got “informal” certification notice yesterday! Will be used to host experiments with sensitive data Cloud deployment We are still finalizing our configuration, delay due to uncertainties in Italian laws Who is responsible in case a cloud machine is used improperly? Andrea Chierici

Major Projects (2) Deployment of condor pilot Condor CE to follow Migration from puppet v3 to v5 While working on v4, v5 came out, decided to skip directly to latest Waiting for foreman 1.16 to come out with direct support to puppet v5 Testbed ready next month Foreseen full upgrade within next meeting Andrea Chierici

Questions? Andrea Chierici