KISTI-GSDC SITE REPORT Jeongheon Kim, Duseok Jin On the behalf of KISTI GSDC 18 April 2016 HEPiX Spring 2016 Workshop DESY Zeuthen, Germany.

Slides:



Advertisements
Similar presentations
Update on Status of Tier-1 GSDC, KISTI Sang-Un Ahn, for the GSDC Tier-1 Team
Advertisements

Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
IHEP Site Status Jingyan Shi, Computing Center, IHEP 2015 Spring HEPiX Workshop.
20 Sept Kunsu Oh*, Jonghu Lee** *Dept. of Physics Pusan National Univ., **KISTI GSDC.
Tier-1 experience with provisioning virtualised worker nodes on demand Andrew Lahiff, Ian Collier STFC Rutherford Appleton Laboratory, Harwell Oxford,
1 1 WLCWLCG workshop G workshop. Introduction to KISTI Introduction to NSDC Project Activities in 2009 System architecture Management plan for Alice tier-1.
Status Report on Tier-1 in Korea Gungwon Kang, Sang-Un Ahn and Hangjin Jang (KISTI GSDC) April 28, 2014 at 15th CERN-Korea Committee, Geneva Korea Institute.
Global Science experiment Data hub Center Oct. 13, 2014 Seo-Young Noh Status Report on Tier 1 in Korea.
11 ALICE Computing Activities in Korea Beob Kyun Kim e-Science Division, KISTI
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
November 16, 2012 Seo-Young Noh Haengjin Jang {rsyoung, Status Updates on STAR Computing at KISTI.
Welcome to KISTI KISTI-CCIN2P3 FKPPL Workshop December 1, 2008 Minsun LEE.
1 1APCTP LHC Konkuk University. Introduction to GSDC Project Activities in 2009 Strategies and Plans in 2010 GSDC office opening ceremony CERN.
International Workshop on HEP Data Grid Nov 9, 2002, KNU Data Storage, Network, Handling, and Clustering in CDF Korea group Intae Yu*, Junghyun Kim, Ilsung.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Sejong STATUS Chang Yeong CHOI CERN, ALICE LHC Computing Grid Tier-2 Workshop in Asia, 1 th December 2006.
Status Report of WLCG Tier-1 candidate for KISTI-GSDC Sang-Un Ahn, for the GSDC Tier-1 Team GSDC Tier-1 Team 12 th CERN-Korea.
KISTI-GSDC SITE REPORT Sang-Un Ahn, Jin Kim On the behalf of KISTI GSDC 24 March 2015 HEPiX Spring 2015 Workshop Oxford University, Oxford, UK.
KISTI-GSDC SITE REPORT Asia Tier Center KISTI, Daejeon, South Korea 22 Sep – 24 Sep 2015 Sang-Un Ahn on the behalf of KISTI-GSDC.
Status Report of WLCG Tier-1 candidate for KISTI-GSDC Sang-Un Ahn, for the GSDC Tier-1 Team GSDC Tier-1 Team ATHIC2012, Busan,
IHEP(Beijing LCG2) Site Report Fazhi.Qi, Gang Chen Computing Center,IHEP.
STATUS OF KISTI TIER1 Sang-Un Ahn On behalf of the GSDC Tier1 Team WLCG Management Board 18 November 2014.
Eygene Ryabinkin, on behalf of KI and JINR Grid teams Russian Tier-1 status report May 9th 2014, WLCG Overview Board meeting.
The Worldwide LHC Computing Grid Frédéric Hemmer IT Department Head Visit of INTEL ISEF CERN Special Award Winners 2012 Thursday, 21 st June 2012.
Development of a Tier-1 computing cluster at National Research Centre 'Kurchatov Institute' Igor Tkachenko on behalf of the NRC-KI Tier-1 team National.
LHC Computing in Korea Seyong Kim For the 1 st Korea-CERN meeting.
Data Center Network Jin Kim GSDC Data Grid Computing School1.
Status of GSDC, KISTI Sang-Un Ahn, for the GSDC Tier-1 Team
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
KISTI activities and plans Global experiment Science Data hub Center Jin Kim LHCOPN-ONE Workshop in Taipei1.
Global Science experimental Data hub Center April 25, 2016 Seo-Young Noh Status Report on KISTI’s Computing Activities.
HEPiX IPv6 Working Group David Kelsey (STFC-RAL) GridPP33 Ambleside 22 Aug 2014.
KISTI-GSDC SITE REPORT KoALICE National ARPINA, Busan, South Korea 8 April – 10 April 2016 Sang-Un Ahn on the behalf of KISTI-GSDC.
Grid Operations in Germany T1-T2 workshop 2015 Torino, Italy Kilian Schwarz WooJin Park Christopher Jung.
GSDC: A Unique Data Center in Korea for Fundamental Research Global Science experimental Data hub Center Korea Institute of Science and Technology Information.
KEK CC - present and future - Mitsuaki NOZAKi (KEK)
PI: Kihyeon Cho & Soonwook Hwang Sponsor: Creative project by KISTI
Academia Sinica Grid Computing Centre (ASGC), Taiwan
Australia Site Report Lucien Boland Goncalo Borges Sean Crosby
NIIF HPC services for research and education
Dynamic Extension of the INFN Tier-1 on external resources
WLCG IPv6 deployment strategy
COMPUTING FOR ALICE IN THE CZECH REPUBLIC in 2015/2016
COMPUTING FOR ALICE IN THE CZECH REPUBLIC in 2016/2017
Status of WLCG FCPPL project
Ian Bird WLCG Workshop San Francisco, 8th October 2016
100G KREONET/KREONet2 Updates
Report from WLCG Workshop 2017: WLCG Network Requirements GDB - CERN 12th of July 2017
(Role Expansion to National Data Center)
INFN Computing infrastructure - Workload management at the Tier-1
October 28, 2013 at 14th CERN-Korea Committee, Geneva
Yaodong CHENG Computing Center, IHEP, CAS 2016 Fall HEPiX Workshop
CC - IN2P3 Site Report Hepix Spring meeting 2011 Darmstadt May 3rd
Vanderbilt Tier 2 Project
Grid status ALICE Offline week Nov 3, Maarten Litmaath CERN-IT v1.0
Update on Plan for KISTI-GSDC
KISTI-GSDC Tier-1 SITE REPORT
HPEiX Spring RAL Site Report
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Update from the HEPiX IPv6 WG
Статус ГРИД-кластера ИЯФ СО РАН.
Computing Infrastructure for DAQ, DM and SC
Particle Physics at KISTI
20409A 7: Installing and Configuring System Center 2012 R2 Virtual Machine Manager Module 7 Installing and Configuring System Center 2012 R2 Virtual.
ExaO: Software Defined Data Distribution for Exascale Sciences
e-Science for High Energy Physics
ETHZ, Zürich September 1st , 2016
This work is supported by projects Research infrastructure CERN (CERN-CZ, LM ) and OP RDE CERN Computing (CZ /0.0/0.0/1 6013/ ) from.
WLCG Tier-2 site at NCP, Status Update and Future Direction
Presentation transcript:

KISTI-GSDC SITE REPORT Jeongheon Kim, Duseok Jin On the behalf of KISTI GSDC 18 April 2016 HEPiX Spring 2016 Workshop DESY Zeuthen, Germany

CONTENTS KISTI GSDC Overview Tier-1 operations GSDC System & Plan HEPiX Spring 2016 Workshop 2

KISTI GSDC OVERVIEW HEPiX Spring 2016 Workshop 3

KISTI Location South Korea KISTI 30 Government Research Institutes 11 Public Research Institutes 29 Non-profit Organizations 7 Universities Daedeok R&D Innopolis HEPiX Spring 2016 Workshop 4 Rare Isotope Accelerator (To be constructed)

KISTI GSDC Government funding research institute for IT founded in people working for National Information Service (distribution & analysis), Supercomputing and Networking Operating Supercomputing and NREN Infrastructure Supercomputer: TFlops at peak (14 th ranked at Top500 in 2009; 201 st now) NREN Infrastructure: KREONet2 Domestic: Seoul ←(100G)→ Daejeon International: HK ←(100G)→ Chicago/Seattle (Member of GLORIAD) 5 KISTI (Korea Institute of Science and Technology Information) GSDC (Global Science experiment Data hub Center) Government funding project to promote research experiment providing computing power and storage HEP: ALICE, CMS, Belle, RENO Others: LIGO, Bioinformatics Running Data-Intensive Computing Facility 13 staffs: sysadmin, experiment support, external-relation, administration Total 7,900 cores, 6,000 TB disk and 1,500 TB tape storage GSDC Facility HEPiX Spring 2016 Workshop History of GSDC Formation of GSDC ALICE T2 Test-bed ALICE T1 Test-bed ALICE T1 candidate 2015 ALICE T2 operation start KISTI Analysis Facility Full T1 for ALICE Start CMS T3 10Gbps LHCOPN Established

GSDC System Overview HEPiX Spring 2016 Workshop PB  Torque/MAUI  4,500 slots  ALICE T1, Belle, RENO  HTCondor  2,500 slots  CMS T3, LIGO, KIAF Public Private 1.5 PB 3.5 PB IBM TSM/GPFS HITACHI USP/VSP EMC Clariion/VNX HITACHI HNAS EMC ICILON 4 Spine switches 74 Leaf switches 500+ Servers in 22 racks 14 Storage racks 4 tape frames

TIER-1 OPERATIONS HEPiX Spring 2016 Workshop 7

KISTI, 3.19% = 2.6M jobs for the last 6 months Jobs Sep 2015 Mar 2016 Maximum 3,402 concurrent jobs = 37 kH06 (84 nodes x 32 (logical) cores + 20 nodes x 40 (logical) cores) HEPiX Spring 2016 Workshop 8 ~ 1300 Maximum 1,344 concurrent jobs (84 nodes x 16 cores) > 5GB memory available per job For HI Reco Run ~ Feb 2016 Stable and smooth running Participating in HI Reco Run at the end of M jobs done for the last 6 months ~ 2500 ~ 3300

Storage 3 Yrs Use History ← May 2013 Run2 Data Taking 400TB Disk buffer for helping fast access to data archived HEPiX Spring 2016 Workshop 9 1,465 TB Used (Tape)855.6 TB Used (Disk) 1.465PB ALICE RAW Data Transferred 99% Storage Availability for the last 6 months Tape Disk 99% Storage Availability of R/W for last 6 months Jun 2013 May 2015

99.8% Reliable for the last 6 month (from Sep-2015 to Feb-2016 ) Monthly Target for Reliability of ALICE test: 97% Less than 10 days of yearly downtime On track for a stable and reliable site Participating in weekly WLCG operations meetings (1 times (Mon) per week) : reporting operation-related issues 24/7 monitoring & maintenance contract 2 persons responsible for on-call Site Availability/Reliability SepOctNovDecJanFeb Reliability Availability HEPiX Spring 2016 Workshop 10

SKB(EAC+PC-1) SKB(EAC+UNITY) SURFnet (CANARI E) SURFnet (SURFnet) SKB(Level3) Seattle LA New York SK Broadband (~’16.8) SURFnet SURFnet (SURFnet) KREONET KISTI-CERN Network (LHCOPN) 10Gbps Upgrade done by 31 st April HEPiX Spring 2016 Workshop 11

Performance HEPiX Spring 2016 Workshop 12 CERN IT Gateway Multi-stream: 500 Max peak: 1GB/s 10G enabled KISTI-GSDC CERN→KISTI (5 min) CERN→KISTI Average: 65 MB/s > 9Gbps peak (~ 1GB/s) observed CERN IT provided a gateway, 500 parallel transfers xrd3cp crashed with Xrootd v3.3.4 v4 or later) Max 1GB/s alimonitor.cern.ch Confirmed full capacity MX960 alimonitor.cern.ch

GSDC SYSTEM HEPiX Spring 2016 Workshop 13

Controlled by Puppet and Foreman Profile implemented by the Puppet Role defined in the Host Group of Foreman HEPiX Spring 2016 Workshop 14

Central Authentication by IPA User DNS Kerberos SUDO rule Automount Host based access control HEPiX Spring 2016 Workshop 15

Log Collecting by ELKStack Messages IPTables SNMP trap Torque job HEPiX Spring 2016 Workshop 16

Plan Additional resources will be procured ~720 CPU cores ~2 PB disk storage (NAS) ~1.5 PB tape storage Linux container based provisioning system will be deployed for supporting several researches HEPiX Spring 2016 Workshop 17

Experimental Setup HEPiX Spring 2016 Workshop 18 Atomic Host Kubernates Cluster etcd Docker Flanneld hyperkube kubelet k8s apiserver monit k8s podmaster k8s schedulerk8s controller manager k8s proxy Gluster k8s UIk8s Kubedash k8s Dashboard Fluentd Elasticsearch InFluxDBGrafanaKibanaoVirt EngineoVirt Node

Persistence Data Storage HEPiX Spring 2016 Workshop 19

Install Script HEPiX Spring 2016 Workshop 20

Running Screenshot HEPiX Spring 2016 Workshop 21

Mesos Cluster Node OS Bare metal Node OS VM ClusterExternal Cloud TorqueHTCondor Docker Torque Docker HTCondor Monitoring MPI Analyzing Logging Docker ETC Services Datacenter OS (Apache Mesos with Kubernetes) Master Plan for Container based Provisioning System

HEPiX Spring 2016 Workshop 23 감사합니다. آپ کا شکریہ. धन्यवाद। ขอบคุณ 谢谢。 ありがとうございます。 Terima kasih. Благодаря. Dank u. Ευχαριστώ. Dziękuję. Grazie. Vielen Dank. Dank je. Merci. Thank you.

Mesos Cluster VM Cluster for Bare Metal Admin Provisioning System Monitoring System Logging System Bare Metal Cluster Analyzing System Overview of Datacenter

HEPiX Spring 2016 Workshop 25

Experimental Setup Atomic host for OS HEPiX Spring 2016 Workshop 26

Experimental Setup HEPiX Spring 2016 Workshop 27

HEPiX Spring 2016 Workshop 28