Soluzioni HW per il Tier 1 al CNAF Luca dell’Agnello Stefano Zani (INFN – CNAF, Italy) III CCR Workshop May 24-27 2004.

Slides:



Advertisements
Similar presentations
Tony Doyle - University of Glasgow GridPP EDG - UK Contributions Architecture Testbed-1 Network Monitoring Certificates & Security Storage Element R-GMA.
Advertisements

Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
Tier 1 Luca dell’Agnello INFN – CNAF, Bologna Workshop CCR Paestum, 9-12 Giugno 2003.
Site Report HEPHY-UIBK Austrian federated Tier 2 meeting
“A prototype for INFN TIER-1 Regional Centre” Luca dell’Agnello INFN – CNAF, Bologna Workshop CCR La Biodola, 8 Maggio 2002.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
INFN – Tier1 Site Status Report Vladimir Sapunenko on behalf of Tier1 staff.
INFN Tier1 Status report Spring HEPiX 2005 Andrea Chierici – INFN CNAF.
INFN Tier1 Andrea Chierici INFN – CNAF, Italy LCG Workshop CERN, March
Amsterdam May 19-23,2003 Site Report Roberto Gomezel INFN - Trieste.
FZU Computing Centre Jan Švec Institute of Physics of the AS CR, v.v.i
October, Site Report Roberto Gomezel INFN.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Federico Ruggieri INFN-CNAF GDB Meeting 10 February 2004 INFN TIER1 Status.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
Tier1 Status Report Martin Bly RAL 27,28 April 2005.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
INDIACMS-TIFR Tier 2 Grid Status Report I IndiaCMS Meeting, April 05-06, 2007.
LCG-2 Plan in Taiwan Simon C. Lin and Eric Yen Academia Sinica Taipei, Taiwan 13 January 2004.
Sandor Acs 05/07/
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
RAL Site Report John Gordon IT Department, CLRC/RAL HEPiX Meeting, JLAB, October 2000.
Tier1 status at INFN-CNAF Giuseppe Lo Re INFN – CNAF Bologna Offline Week
10/22/2002Bernd Panzer-Steindel, CERN/IT1 Data Challenges and Fabric Architecture.
1 PRAGUE site report. 2 Overview Supported HEP experiments and staff Hardware on Prague farms Statistics about running LHC experiment’s DC Experience.
HEPix April 2006 NIKHEF site report What’s new at NIKHEF’s infrastructure and Ramping up the LCG tier-1 Wim Heubers / NIKHEF (+SARA)
Storage and Storage Access 1 Rainer Többicke CERN/IT.
Tier2 Centre in Prague Jiří Chudoba FZU AV ČR - Institute of Physics of the Academy of Sciences of the Czech Republic.
INFN TIER1 (IT-INFN-CNAF) “Concerns from sites” Session LHC OPN/ONE “Networking for WLCG” Workshop CERN, Stefano Zani
October, HEPiX Fall 2005 at SLACSLAC Site Report Roberto Gomezel INFN.
KOLKATA Grid Site Name :- IN-DAE-VECC-02Monalisa Name:- Kolkata-Cream VO :- ALICECity:- KOLKATACountry :- INDIA Shown many data transfers.
Tier1 Andrew Sansum GRIDPP 10 June GRIDPP10 June 2004Tier1A2 Production Service for HEP (PPARC) GRIDPP ( ). –“ GridPP will enable testing.
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
Fabric Monitoring at the INFN Tier1 Felice Rosso on behalf of INFN Tier1 Joint OSG & EGEE Operations WS, Culham (UK)
CASPUR Site Report Andrei Maslennikov Lead - Systems Amsterdam, May 2003.
Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft Implementation of a reliable and expandable on-line storage for compute clusters Jos van Wezel.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
SA1 operational policy training, Athens 20-21/01/05 Presentation of the HG Node “Isabella” and operational experience Antonis Zissimos Member of ICCS administration.
CASTOR CNAF TIER1 SITE REPORT Geneve CERN June 2005 Ricci Pier Paolo
Materials for Report about Computing Jiří Chudoba x.y.2006 Institute of Physics, Prague.
CNAF Database Service Barbara Martelli CNAF-INFN Elisabetta Vilucchi CNAF-INFN Simone Dalla Fina INFN-Padua.
The Italian Tier-1: INFN-CNAF 11-Oct-2005 Luca dell’Agnello Davide Salomoni.
Database CNAF Barbara Martelli Rome, April 4 st 2006.
INRNE's participation in LCG Elena Puncheva Preslav Konstantinov IT Department.
The Italian Tier-1: INFN-CNAF Andrea Chierici, on behalf of the INFN Tier1 3° April 2006 – Spring HEPIX.
IT-INFN-CNAF Status Update LHC-OPN Meeting INFN CNAF, December 2009 Stefano Zani 10/11/2009Stefano Zani INFN CNAF (TIER1 Staff)1.
Storage at TIER1 CNAF Workshop Storage INFN CNAF 20/21 Marzo 2006 Bologna Ricci Pier Paolo, on behalf of INFN TIER1 Staff
PADME Kick-Off Meeting – LNF, April 20-21, DAQ Data Rate - Preliminary estimate Tentative setup: all channels read with Fast ADC 1024 samples, 12.
Tier2 Centre in Prague Jiří Chudoba FZU AV ČR - Institute of Physics of the Academy of Sciences of the Czech Republic.
CNAF - 24 September 2004 EGEE SA-1 SPACI Activity Italo Epicoco.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
status, usage and perspectives
Luca dell’Agnello INFN-CNAF
INFN CNAF TIER1 Network Service
LCG 3D Distributed Deployment of Databases
Andrea Chierici On behalf of INFN-T1 staff
The INFN TIER1 Regional Centre
HIGH-PERFORMANCE COMPUTING SYSTEM FOR HIGH ENERGY PHYSICS
NIKHEF Data Processing Fclty
The INFN Tier-1 Storage Implementation
Luca dell’Agnello Daniele Cesini GDB - 13/12/2017
QMUL Site Report by Dave Kant HEPSYSMAN Meeting /09/2019
Presentation transcript:

Soluzioni HW per il Tier 1 al CNAF Luca dell’Agnello Stefano Zani (INFN – CNAF, Italy) III CCR Workshop May

Tier1 INFN computing facility for HEP community  Ending prototype phase last year, now fully operational  Location: INFN-CNAF, Bologna (Italy) o One of the main nodes on GARR network  Personnel: ~ 10 FTE’s o ~ 3 FTE's dedicated to experiments Multi-experiment  LHC experiments(Alice, Atlas, CMS, LHCb), Virgo, CDF, BABAR, AMS, MAGIC,...  Resources dynamically assigned to experiments according to their needs 50% of the Italian resource for LCG  Participation to experiments data challenge  Integrated with Italian Grid  Resources accessible also in traditional way

Logistics Recently moved to a new location (last January)  Hall in the basement (-2 nd floor)  ~ 1000 m 2 of total space o Computing Nodes o Storage Devices o Electric Power System (UPS) o Cooling and Air conditioning system o Garr GPop  Easily accessible with lorries from the road  Not suitable for office use (remote control needed)

Electric Power Electric Power Generator  1250 KVA (~ 1000 KW)  up to 160 racks Uninterruptible Power Supply (UPS)  Located into a separate room (conditioned and ventilated)  800 KVA (~ 640 KW) 380 V three-phase distributed to all racks (Blindo)  Rack power controls output 3 independent 220 V lines for computers  Rack power controls sustain burden up to 16 or 32 A o 32 A power controls needed for Xeon 36 bi-processors racks  3 APC power distribution modules (24 outlets each) o Completely programmable (allows gradual servers switching on) o Remotely manageable via web 380 V three-phase for other devices (tape libraries, air conditioning, etc…)

Cooling & Air Conditioning RLS (Airwell) on the roof  ~ 700 KW  Water cooling  Need “booster pump” (20 mts T1  roof)  Noise insulation 1 Air Conditioning Unit (uses 20% of RLS refreshing power and controls humidity) 12 Local Cooling Systems (Hiross) in the computing room

WN typical Rack Composition Power Controls (3U) 1 network switch (1-2U) –48 FE copper interfaces –2 GE fiber uplinks U WNs –Connected to network switch via FE –Connected to KVM system

Remote console control Paragon UTM8 (Raritan)  8 Analog (UTP/Fiber) output connections  Supports up to 32 daisy chains of 40 nodes (UKVMSPD modules needed)  Costs: 6 KEuro Euro/server (UKVMSPD module)  IP-reach (expansion to support IP transport) evaluted but not used Autoview 2000R (Avocent)  1 Analog + 2 Digital (IP transport) output connections  Supports connections up to 16 nodes o Optional expansion to 16x8 nodes  Compatible with Paragon (“gateway” to IP) Evaluating Cyclades Alterpath KVM via serial line (cheaper)

Networking (1) Main Network infrastructure based on optical fibres (~ 20 Km)  To ease adoption of new (High Performances) transmission technologies  To insure a better electrical insulation on long distances  Local (Rack wide) links with UTP (copper) cables LAN has a “classical” star topology  GE core switch (Enterasys ER16)  NEW core switch is going to be shipped (Next july) o 120 Gbit Fiber (Scale up to 480 ports) o Gbit Ethernet (Scale up to max 48 ports)  Farms up-link via GE trunk (Channel) to core switch  Disk Servers directly connected to GE switch (mainly fibre)

Networking (2) WN's connected via FE to rack switch (1 switch per rack)  Not a single brand for switches (as for wn's) o 3 Extreme Summit 48 FE + 2 GE ports o Cisco 48 FE + 2 GE ports o 8 Enterasys 48 FE 2GE ports o 7 switch Summit GE copper + 2 GE ports + (2x10Gb ready)  Homogeneous characteristics o 48 Copper Ethernet ports o Support of main standards (e.g q) o 2 Gigabit up-links (optical fibers) to core switch CNAF interconnected to GARR-G backbone at 1 Gbps.  Giga-PoP co-located  2 x 1 Gbps test links to CERN, Karlsruhe

FarmSW3(IBM) NAS4 FarmSWG1 SSR8600 Bo 12KGP FarmSW1 FarmSW2(Dell) LHCBSW1 NAS2 NAS3 S.Zani FarmSW4(IBM3) Catalyst3550 FarmSW5(3Com) DELL AXUS SAN Disk Server F.C. FarmSW9FarmSW FarmSW6 FarmSW7 FarmSW8 FarmSW10 FarmSW11 FarmSWG2 STK F.C. 1 st Floor Internal services T1 Babar SW NAS1 Network Configuration I nfortrend F.C.

L2 Configuration Each Experiment has its own VLAN Solution adopted for complete granularity  Port based VLAN  VLAN identifiers are propagated across switches (802.1q)  Avoid recabling (or physical moving) of machines to change farm topology Level 2 isolation of farms Possibility to define multi-tag (Trunk) ports (for servers)

Power Switches 2 models used at Tier1: “Old” APC MasterSwitch Control Unit AP9224 controlling 3x8 outlets 9222 PDU from 1 Ethernet “New” APC PDU Control Unit AP7951 controlling 24 outlets from 1 Ethernet “zero” Rack Unit (vertical mount) Access to the configuration/control menu via serial/telnet/web/snmp 1 Dedicated machine running APC Infrastruxure Manager Software (in progress) See also: doc/CD0044.doc

Remote Power Distribution Unit Screenshot of APC Infrastruxure Manager Software with the status of all TIER1 PDU

Computing units ~ 400 1U rack-mountable Intel dual processor servers  800 MHz – 2.4 GHz  ~ 240 wn’s (~ 480 CPU’s) available for LCG To be shipped June 2004:  32 1U bi-processors Pentium 2.4 GHz  350 1U bi-processors Pentium IV 3.06 GHz o 2 x 120 GB HDs o 4 GB RAM o 2159 euro each Tendering:  HPC farm with MPI o Servers interconnected via Infiniband  Opteron farm (near future) o To allow experiments to test their software on AMD architecture

Storage Resources ~50 TB RAW Disk Space ON LINE.  NAS o NAS1+NAS4 (3Ware low cost) Tot 4.2 TB o NAS2+NAS3 (Procom)Tot 13.2 TB  SAN o Dell Powervault 660fTot 7TB o Axus (Brownie)Tot 2TB o STK BladestoreTot 9TB o Infortrend ES A16F-R Tot12TB o IBM Fast-T 900(in few weeks)Tot 150 TB See also:

STORAGE resource CLIENT SIDE WAN or TIER1 LAN PROCOM NAS2 Nas2.cnaf.infn.it 8100 Gbyte VIRGO ATLAS Fileserver CMS diskserv-cms-1 PROCOM NAS3 Nas3.cnaf.infn.it 4700 Gbyte ALICE ATLAS IDE NAS1,NAS4 Nas4.cnaf.infn.it Gbyte CDF LHCB AXUS BROWIE Circa 2200 GByte 2 FC interface DELL POWERVAULT 7100 GByte 2 FC interface FAIL-OVER support Gadzoox Slingshot FC Switch 18 port RAIDTEC 1800 Gbyte 2 SCSI interfaces CASTOR Server+staging STK180 with 100 LTO (10Tbyte Native) Fileserver Fcds2 Alias diskserv-ams-1 diskserv-atlas-1 STK BladeStore Circa GByte 4 FC interface STK L5500 robot (max 5000) 6 LTO-2 Infortrend ES A16F-R 12 TB

Storage management and access (1) Tier1 storage resources accessible as classical storage or via grid Non grid disk storage accessible via NFS Generic WN’s also have AFS client NFS mount volumes configured via autofs and ldap  unique configuration repository eases maintenance  in progress: integration of ldap configuration with Tier1 db data Scalability issues with NFS  Experienced stalled mount points  Recent nfs versions use synchronous export: needed to revert to async and use reduced rsize and wsize to avoid huge amount of retransmissions

Storage management and access (2) Part of disk storage used as front-end to CASTOR  Balance between disk and CASTOR according to experiments needs 1 stager for each experiment (installation in progress) CASTOR accessible both directly or via grid  CASTOR SE available ALICE Data Challenge used CASTOR architecture  Feedback to CASTOR team  Need optimization for file restaging

Tier1 Database Resource database and management interface  Postgres database as back end  Web interface (apache+mod_ssl+php)  Hw servers characteristics  Sw servers configuration  Servers allocation Possible direct access to db for some applications  Monitoring system  Nagios Interface to configure switches and interoperate with installation system.  Vlan tags  dns  dhcp

Installation issues Centralized installation system  LCFG (EDG WP4)  Integration with a central Tier1 db  Moving from a farm to another implies just changes in IP address (not name)  Unique dhcp server for all VLANs  Support for DDNS (cr.cnaf.infn.it) Investigating Quattor for future needs

Our Desired Solution for Resource Access SHARED RESOURCES among all experiments  Priorities and reservations managed by the scheduler Most of Tier1 computing machines installed as LCG Worker Nodes, with light modifications to support more VOs Application Software not directly installed on WNs but accessed from outside (NFS, AFS, …) One or more Resource Manager to manage all the WNs in a centralized way Standard way to access Storage for each application