GRID OPERATIONS IN ROMANIA

Slides:



Advertisements
Similar presentations
National Grid's Contribution to LHCb IFIN-HH Serban Constantinescu, Ciubancan Mihai, Teodor Ivanoaica.
Advertisements

CURRENT AND FUTURE HPC SOLUTIONS. T-PLATFORMS  Russia’s leading developer of turn-key solutions for supercomputing  Privately owned  140+ employees.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
IFIN-HH LHCB GRID Activities Eduard Pauna Radu Stoica.
Site Report US CMS T2 Workshop Samir Cury on behalf of T2_BR_UERJ Team.
March 27, IndiaCMS Meeting, Delhi1 T2_IN_TIFR of all-of-us, for all-of-us, by some-of-us Tier-2 Status Report.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
Green technology used for ATLAS processing Dr. Ing. Fărcaş Felix NATIONAL INSTITUTE FOR RESEARCH AND DEVELOPMENT OF ISOTOPIC AND MOLECULAR.
COMSATS Institute of Information Technology, Islamabad PK-CIIT Grid Operations in Pakistan COMSATS Ali Zahir Site Admin / Faculty Member ALICE T1/2 March.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
BINP/GCF Status Report BINP LCG Site Registration Oct 2009
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
INDIACMS-TIFR Tier 2 Grid Status Report I IndiaCMS Meeting, April 05-06, 2007.
Romanian Tier-2 Federation One site for all: RO-07-NIPNE Mihai Ciubancan on behalf of IT Department.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
PDSF at NERSC Site Report HEPiX April 2010 Jay Srinivasan (w/contributions from I. Sakrejda, C. Whitney, and B. Draney) (Presented by Sandy.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
UKI-SouthGrid Update Hepix Pete Gronbech SouthGrid Technical Coordinator April 2012.
ITEP computing center and plans for supercomputing Plans for Tier 1 for FAIR (GSI) in ITEP  8000 cores in 3 years, in this year  Distributed.
Company LOGO “ALEXANDRU IOAN CUZA” UNIVERSITY OF IAŞI” Digital Communications Department Status of RO-16-UAIC Grid site in 2013 System manager: Pînzaru.
KOLKATA Grid Site Name :- IN-DAE-VECC-02Monalisa Name:- Kolkata-Cream VO :- ALICECity:- KOLKATACountry :- INDIA Shown many data transfers.
IHEP(Beijing LCG2) Site Report Fazhi.Qi, Gang Chen Computing Center,IHEP.
Computing Jiří Chudoba Institute of Physics, CAS.
UKI-SouthGrid Overview and Oxford Status Report Pete Gronbech SouthGrid Technical Coordinator HEPSYSMAN – RAL 10 th June 2010.
RAL PPD Tier 2 (and stuff) Site Report Rob Harper HEP SysMan 30 th June
Evangelos Markatos and Charalampos Gkikas FORTH-ICS Athens, th Mar Institute of Computer Science - FORTH Christos.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
INRNE's participation in LCG Elena Puncheva Preslav Konstantinov IT Department.
Development of a Tier-1 computing cluster at National Research Centre 'Kurchatov Institute' Igor Tkachenko on behalf of the NRC-KI Tier-1 team National.
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Mihnea Dulea, IFIN-HH R-ECFA Meeting, National Physics Library IFIN-HH, Magurele Romanian participation in WLCG M. Dulea Elementary Particles.
Pledged and delivered resources to ALICE Grid computing in Germany Kilian Schwarz GSI Darmstadt ALICE Offline Week.
RECENT DEVELOPMENTS IN THE CONTRIBUTION OF DFCTI/IFIN-HH TO THE WLCG COLLABORATION Department of Computational Physics and Information Technologies (DFCTI)
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Grid Operations in Germany T1-T2 workshop 2015 Torino, Italy Kilian Schwarz WooJin Park Christopher Jung.
13 January 2004GDB Geneva, Milos Lokajicek Institute of Physics AS CR, Prague LCG regional centre in Prague
KOLKATA Grid Kolkata Tier-2 Status and Plan Site Name :- IN-DAE-VECC-02 Gocdb Name:- IN-DAE-VECC-02 VO :- ALICE City:- KOLKATA Country :-
Kilian Schwarz ALICE Computing Meeting GSI, October 7, 2009
Brief introduction about “Grid at LNS”
Dynamic Extension of the INFN Tier-1 on external resources
WLCG IPv6 deployment strategy
COMPUTING FOR ALICE IN THE CZECH REPUBLIC in 2015/2016
COMPUTING FOR ALICE IN THE CZECH REPUBLIC in 2016/2017
Cluster Status & Plans —— Gang Qin
Title of the Poster Supervised By: Prof.*********
The Beijing Tier 2: status and plans
Virtualization and Clouds ATLAS position
Grid site as a tool for data processing and data analysis
Operations and plans - Polish sites
ATLAS Cloud Operations
Heterogeneous Computation Team HybriLIT
LCG Deployment in Japan
Kolkata Status and Plan
Update on Plan for KISTI-GSDC
Experience of Lustre at a Tier-2 site
5th DOSAR Workshop Louisiana Tech University Sept. 27 – 28, 2007
Luca dell’Agnello INFN-CNAF
HIGH-PERFORMANCE COMPUTING SYSTEM FOR HIGH ENERGY PHYSICS
Southwest Tier 2.
Статус ГРИД-кластера ИЯФ СО РАН.
PK-CIIT Grid Operations in Pakistan
Small site approaches - Sussex
WLCG Collaboration Workshop;
Vladimir Sapunenko On behalf of INFN-T1 staff HEPiX Spring 2017
Romanian Sites Current Status
This work is supported by projects Research infrastructure CERN (CERN-CZ, LM ) and OP RDE CERN Computing (CZ /0.0/0.0/1 6013/ ) from.
QMUL Site Report by Dave Kant HEPSYSMAN Meeting /09/2019
Romanian Sites Current Status
Presentation transcript:

GRID OPERATIONS IN ROMANIA ALICE T1-T2 workshop 2017 Strasbourg, France Ionel STAN – ISS Mihai CIUBANCAN – IFIN-HH (NIPNE) Mihai CARABAS - UPB Claudiu Schiaua – IFIN-HH (NIHAM)

Table of contents Overview Sites capabilities Sites status Status of Networking - IPv6 readiness EOS Sites planning

Overview UPB - University Politehnica of Bucharest (UPB) NIHAM, NIPNE - Horia Hulubei National Institute for R&D in Physics and Nuclear Engineering (IFIN-HH) ISS, ISS_LCG – Institute Of Space Science (ISS)

Table of contents Overview Sites capabilities Sites status Status of Networking - IPv6 readiness EOS Sites planning

New ISS Computing Infrastructure Sites capabilities - ISS New ISS Computing Infrastructure Designed for high density computing (Hot Aisle, InRow cooling) Scalable solution for future investments UPS Power : 48 kVA (with N+1 redundancy power units) Cooling capacity : 80 kW installed (2N capacity redundancy) new computing resources purchased at the end of 2016 192 cores (Broadwell, 14 nm, 2.2 GHz base freq) - 8 nodes memory 5.3 GB/core - DDR4 2400 MHz ECC 2 x 10 Gb network/server; 4x40 QSFP uplinks enclosure expandable to 28 nodes storage capacity upgraded from 220 to 460 TB

Sites capabilities - ISS HARDWARE AND TOPOLOGY OF COMPUTING FACILITY Our hardware is mainly comprised of SuperMicro machines that were chosen for the great resource density/price ratio. For computing nodes we use Twin servers, Blade servers which give us very good densities and for the storage we use servers with 24, 36 drives and JBOD cases with 45 drives in 4U of rack space. In present we have 550 cores and 460 TB Generic schematic of ISS computing facility :

Sites capabilities - ISS HARDWARE AND TOPOLOGY OF COMPUTING FACILITY The AliEn cluster has at his core a 10 Gbps aggregating switch which is connected to the top-of-rack switch of the computing nodes. In the aggregating switch are connected the interfaces of the storage node, a topology which give a high bandwidth connection between worker nodes and storage with very little oversubscribing.

Sites capabilities – RO-07-NIPNE Computing infrastructure APC InRow Chilled Water Cooling 160KVA UPS More then 3100 CPU(~230 nodes) 8, 16, 20, 32 cores/server Computing and storage resources dedicated to 3 LHC VOs: Alice, ATLAS, LHCb 2 different resource managers: PBS/Torque+Maui, SLURM 4 subclusters , 6 queues ,2 multicore queues Storage access for romanian ATLAS diskless sites Member in FAX(Federated ATLAS storage systems using XRootD) Part of LHCONE network (10Gbps connectivity)

Sites capabilities – RO-07-NIPNE Storage infrastructure 4x80KVA UPS Emerson 10 servers ~1,5PB total capacity >1PB used capacity Network infrastructure: 120Gbps connectivity between DC1 and DC2 2x40Gbps, 4x10Gbps RO-07-NIPNE Software: Scientific Linux 6, UMD3 middleware 3 CREAM + 1 ARC-CE as job management service 12 queues (PBS/Torque + MAUI, SLURM) Disk Pool Manager(DPM – for ATLAS, LHCb) with 9 disk storage EOS (Alice) – 1 FST server Top BDII, Site BDII, VOBOX, CVMFS for all VOs Alice VO: VOBOX, Dedicated CREAM, 688 cores, 30 nodes;

HW Computing Infrastructure Sites capabilities - UPB HW Computing Infrastructure 32 dual quad-core Xeon 20 dual hex-core Opteron 28 dual quad-core Nehalem 4 dual PowerXCell 8i 8 dual octo-core Power7 4 dual Xeon – 2 x NVidia Tesla M2070 60 dual octo-core Haswell 3 dual octo-core Haswell – 2 x NVidia Tesla K40m 1Gb-10Gb Ethernet / 40Gb-56Gb Infiniband Interconnect Total storage of 120TB (small capacity disks SAS/Fibre Channel)

Grid Nodes as OpenStack VMs Sites capabilities - UPB Grid Nodes as OpenStack VMs Worker Grid Nodes running on top of OpenStack Prepared Cloud Image with all the necesarry packages Able to Run Scripts after VM creation Provide elasticity easily increase or decrease the capacity of the Grid No performance issues jobs are running with no cost of performance in terms of CPU performant virtualized I/O operations using Virt I/O RO-03-UPB in Alice Started from November 2016 Pilot Test with168 cores From January 2017 increased at 448 cores on top of OpenStack At any time we can scale-up up very fast E.g.: during the summer the resources aren’t used by students and can be shared in the Grid

Table of contents Overview Sites capabilities Sites status Status of Networking - IPv6 readiness EOS Sites planning

Sites status – Romanian computing contribution - 4.56M jobs (2.46%) - 22.9 M CPU hours (3.99%) - 49.9 M kSI2k hous (2.42%)

Sites status – running jobs profile

Sites status - Job Efficiency

Sites status – SE Availability

Sites status - Aggregated network traffic pe SE

Table of contents Overview Sites capabilities Sites status Status of Networking - IPv6 readiness EOS Sites planning

Status of Networking - IPv6 readiness NIHAM, NIPNE - infrastructure is IPv6 ready - servers not dual stacked yet ISS - work in progress: IPv6 class already assigned, next step basic implementation in central routers (ISS + upstream RoEduNet) UPB - at any time we can switch grid services to IPv6

Table of contents Overview Sites capabilities Sites status Status of Networking - IPv6 readiness EOS Sites planning

EOS NIHAM - no plan to migrate to EOS NIPNE - already use EOS ISS - high initial effort for new storage cluster (upgrading hdds brings more space than purchasing a single server or at most two) UPB - no storage at the moment

Table of contents Overview Sites capabilities Sites status Status of Networking - IPv6 readiness EOS Sites planning

Sites planning ISS - ~ 55000 Euro (from RO-CERN 2 projects : physics + GRID) - purchase new hard drives (+320 TB net gain) to replace the obsolete ones - purchase new worker nodes (+96 Broadwell 2.2 Ghz cores) NIHAM - no funding info - present efforts are concentrated into replacing hardware while preserving the overall capacity. From Claudiu email: “We have new workers to replace the oldest ones. By the end of this year we will also buy new storage machines, so next year we will replace the present storage hardware. Increased storage capacity is desirable, as the storage is now full, but I cannot say now if we will have it.”

Sites planning NIPNE - no funding info - double the storage(EOS) capacity for Alice - upgrade the network bandwidth for RO-07-NIPNE from 10Gbps to 20Gbps - keep the amount of dedicated computing resources for Alice UPB - no funding dedicated for Alice - add storage capacity from internal funding (initial 72TB with potential increase to 200TB)

Thank you for your attention! Contacts for site admins: ISS: ionel.stan@cern.ch, adrian.sevcenco@cern.ch NIHAM: claudiu.schiaua@cern.ch NIPNE: mihai.ciubancan@cern.ch UPB: mihai.carabas@cs.pub.ro