20 Sept. 2010 Kunsu Oh*, Jonghu Lee** *Dept. of Physics Pusan National Univ., **KISTI GSDC.

Slides:



Advertisements
Similar presentations
Tony Doyle - University of Glasgow GridPP EDG - UK Contributions Architecture Testbed-1 Network Monitoring Certificates & Security Storage Element R-GMA.
Advertisements

CBPF J. Magnin LAFEX-CBPF. Outline What is the GRID ? Why GRID at CBPF ? What are our needs ? Status of GRID at CBPF.
DOSAR Workshop VI April 17, 2008 Louisiana Tech Site Report Michael Bryant Louisiana Tech University.
Contact: Hirofumi Amano at Kyushu 40 Years of HPC Services In this memorable year, the.
Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
HEP Prospects, J. Yu LEARN Strategy Meeting Prospects on Texas High Energy Physics Network Needs LEARN Strategy Meeting University of Texas at El Paso.
Regional Computing Centre for Particle Physics Institute of Physics AS CR (FZU) TIER2 of LCG (LHC Computing Grid) 1M. Lokajicek Dell Presentation.
March 27, IndiaCMS Meeting, Delhi1 T2_IN_TIFR of all-of-us, for all-of-us, by some-of-us Tier-2 Status Report.
1 1 WLCWLCG workshop G workshop. Introduction to KISTI Introduction to NSDC Project Activities in 2009 System architecture Management plan for Alice tier-1.
ScotGrid: a Prototype Tier-2 Centre – Steve Thorn, Edinburgh University SCOTGRID: A PROTOTYPE TIER-2 CENTRE Steve Thorn Authors: A. Earl, P. Clark, S.
Status Report on Tier-1 in Korea Gungwon Kang, Sang-Un Ahn and Hangjin Jang (KISTI GSDC) April 28, 2014 at 15th CERN-Korea Committee, Geneva Korea Institute.
Global Science experiment Data hub Center Oct. 13, 2014 Seo-Young Noh Status Report on Tier 1 in Korea.
HEP GRID CHEP, KNU 11/9/2002 Youngjoon Kwon (Yonsei Univ.) 1 Belle Computing / Data Handling  What is Belle and why we need large-scale computing?
Miracle Consortium: Progress Report Nick Achilleos (UCL), Consortium Chair, Miracle Co-I; Jeremy Yates (UCL), Miracle Co-I Brief report prepared according.
23 Sept Jonghu Lee ( 李鍾厚 ) KISTI GSDC jlee206ATkisti.re.kr.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
Preparation of KIPT (Kharkov) computing facilities for CMS data analysis L. Levchuk Kharkov Institute of Physics and Technology (KIPT), Kharkov, Ukraine.
Charm Research in Heavy Ion Experiment Pusan National Univeristy HIPEx Lab. In-Kwon YOO For the HIM_SRC (Center for Quark Matter)
11 ALICE Computing Activities in Korea Beob Kyun Kim e-Science Division, KISTI
October 2010CERN-Korea J. Schukraft1 ALICE Status 8 th CERN-Korea meeting Collaboration News Korean participation General ALICE status report on Wednesday.
12th November 2003LHCb Software Week1 UK Computing Glenn Patrick Rutherford Appleton Laboratory.
November 16, 2012 Seo-Young Noh Haengjin Jang {rsyoung, Status Updates on STAR Computing at KISTI.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
1 1APCTP LHC Konkuk University. Introduction to GSDC Project Activities in 2009 Strategies and Plans in 2010 GSDC office opening ceremony CERN.
International Workshop on HEP Data Grid Nov 9, 2002, KNU Data Storage, Network, Handling, and Clustering in CDF Korea group Intae Yu*, Junghyun Kim, Ilsung.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Sejong STATUS Chang Yeong CHOI CERN, ALICE LHC Computing Grid Tier-2 Workshop in Asia, 1 th December 2006.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
October 2002 INFN Catania 1 The (LHCC) Grid Project Initiative in Prague Dagmar Adamova INP Rez near Prague.
Status Report of WLCG Tier-1 candidate for KISTI-GSDC Sang-Un Ahn, for the GSDC Tier-1 Team GSDC Tier-1 Team 12 th CERN-Korea.
KISTI-GSDC SITE REPORT Sang-Un Ahn, Jin Kim On the behalf of KISTI GSDC 24 March 2015 HEPiX Spring 2015 Workshop Oxford University, Oxford, UK.
KISTI-GSDC SITE REPORT Asia Tier Center KISTI, Daejeon, South Korea 22 Sep – 24 Sep 2015 Sang-Un Ahn on the behalf of KISTI-GSDC.
KOLKATA Grid Site Name :- IN-DAE-VECC-02Monalisa Name:- Kolkata-Cream VO :- ALICECity:- KOLKATACountry :- INDIA Shown many data transfers.
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
CASPUR Site Report Andrei Maslennikov Lead - Systems Amsterdam, May 2003.
National HEP Data Grid Project in Korea Kihyeon Cho Center for High Energy Physics (CHEP) Kyungpook National University CDF CAF & Grid Meeting July 12,
11 November 2010 Natascha Hörmann Computing at HEPHY Evaluation 2010.
Status Report of WLCG Tier-1 candidate for KISTI-GSDC Sang-Un Ahn, for the GSDC Tier-1 Team GSDC Tier-1 Team ATHIC2012, Busan,
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
Gungwon Kang & Jiwoong Kim (KISTI) LIGO Data Grid and KISTI June 27, 2015 at 8 th J-K Joint Workshop on KAGRA, Gwangju in Korea.
International Workshop on HEP Data Grid Aug 23, 2003, KNU Status of Data Storage, Network, Clustering in SKKU CDF group Intae Yu*, Joong Seok Chae Department.
January 30, 2016 RHIC/USATLAS Computing Facility Overview Dantong Yu Brookhaven National Lab.
Final Implementation of a High Performance Computing Cluster at Florida Tech P. FORD, X. FAVE, K. GNANVO, R. HOCH, M. HOHLMANN, D. MITRA Physics and Space.
Materials for Report about Computing Jiří Chudoba x.y.2006 Institute of Physics, Prague.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
MC Production in Canada Pierre Savard University of Toronto and TRIUMF IFC Meeting October 2003.
Status of Tokyo LCG tier-2 center for atlas / H. Sakamoto / ISGC07 Status of Tokyo LCG Tier 2 Center for ATLAS Hiroshi Sakamoto International Center for.
Ismayilov Ali Institute of Physics of ANAS Creating a distributed computing grid of Azerbaijan for collaborative research NEC'2011.
Evangelos Markatos and Charalampos Gkikas FORTH-ICS Athens, th Mar Institute of Computer Science - FORTH Christos.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
INRNE's participation in LCG Elena Puncheva Preslav Konstantinov IT Department.
UTA Site Report Jae Yu UTA Site Report 7 th DOSAR Workshop Louisiana State University Apr. 2 – 3, 2009 Jae Yu Univ. of Texas, Arlington.
Status of GSDC, KISTI Sang-Un Ahn, for the GSDC Tier-1 Team
KISTI activities and plans Global experiment Science Data hub Center Jin Kim LHCOPN-ONE Workshop in Taipei1.
Global Science experimental Data hub Center April 25, 2016 Seo-Young Noh Status Report on KISTI’s Computing Activities.
Grid activities in Czech Republic Jiri Kosina Institute of Physics of the Academy of Sciences of the Czech Republic
LHC collisions rate: Hz New PHYSICS rate: Hz Event selection: 1 in 10,000,000,000,000 Signal/Noise: Raw Data volumes produced.
GSDC: A Unique Data Center in Korea for Fundamental Research Global Science experimental Data hub Center Korea Institute of Science and Technology Information.
Belle II Physics Analysis Center at TIFR
October 28, 2013 at 14th CERN-Korea Committee, Geneva
LCG Deployment in Japan
Kolkata Status and Plan
Update on Plan for KISTI-GSDC
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Christof Hanke, HEPIX Spring Meeting 2008, CERN
Simulation use cases for T2 in ALICE
Particle Physics at KISTI
Presentation transcript:

20 Sept Kunsu Oh*, Jonghu Lee** *Dept. of Physics Pusan National Univ., **KISTI GSDC

Content STAR Collaboration and Computing 1 1 Current Status of SACC 2 2 Future Plan of SACC 3 3 2

1. STAR Collaboration & Computing RHIC (Relativistic Heavy Ion Collider) Au+Au, Cu+Cu, d+Au, p+p collision 5~200GeV/nucleon International Collaboration 594 scientists 54 institutes 12 countries STAR detector is doing major upgrade on Beam Luminosity X 10 DAA rates X 10 3

STAR Computing Produced Huge Data: ~1PB/year  Needed many Computing resources (CPU and storage) Increase Asian Collaborations: Increase Asian Computing Needs Produced Huge Data: ~1PB/year  Needed many Computing resources (CPU and storage) Increase Asian Collaborations: Increase Asian Computing Needs 4

SACC Project 5 STAR Asian Computing Center Creating and Developing a resource and facility in Asia Goal Hub for Asian Collaborators A new STAR computing center for data production Hub for Asian Collaborators A new STAR computing center for data production Progress Network Optimization STAR software installation Production and Analysis Network Optimization STAR software installation Production and Analysis

MISSION Physics 2009 ~Next ~ ALICE STAR Belle CDF Various Bio-informatics Earth Science Astrophysics 6

Current Status ALICE Tier-1 Test-bed ALICE Tier-2 KIAF (Kisti Analysis Farm) RAW Data Tier-1 RAW Data Computing and Storage Infrastructure Technology Development Apply Grid Technology to legacy app. support Korean Gov. 7

Members NoNameRole 1Dr. Haengjin JangHead of GSDC 2Dr. Hyungwoo ParkProject Management 3Mr. Jinseung YuTechnical Staff (Network) 4Mr. Heejun YoonTechnical Staff (DBA) 5Dr. Beokyun KimTechnical Staff (Grid) 6Dr. Christophe BonnaudTechnical Staff (Admin) 7Dr. Seokmyun KwonTechnical Staff 8Dr. Seo-Young NohTechnical Staff 9Dr. Jonghu LeeTechnical Staff (STAR contact) 10Dr. Seungyun YuPlanning 11Mr. Kyungyun KimTechnical Staff 12Mr. Seunghee LeeTechnical Staff 13Ms. Tajia HanTechnical Staff 8

Activities ALICE Tier-1 Test-bed Completed set-up ALICE Tier-1 test-bed this year Will provide official service in a few years Completed set-up ALICE Tier-1 test-bed this year Will provide official service in a few years ALICE Tier-2 Site availability: 98% since Feb Belle Providing computing resources for Belle MC production (Grid) Bell to provide their data to KISTI GSDC Providing computing resources for Belle MC production (Grid) Bell to provide their data to KISTI GSDC CDF Providing computing resources under NAMCAF Supporting CDFSoft development Providing computing resources under NAMCAF Supporting CDFSoft development LIGO Set-up LIGO cluster test-bed GBrain Planning to cooperate with global brain research project (mainly from McGill Univ. Canada) 9

Current Computing Resources: Cluster Server ClusterSpecMem.NodeCorekSI2k Ce-alice Dell Intel Xeon E GHz Quad 2 CPU 16GB648 Ce01 HP Intel Xeon E GHz Quad 2 CPU 16GB Ce02 IBM Intel Xeon E GHz Quad 2 CPU 16GB This Year IBM Intel Xeon X GHz 6 Core 2 CPU 24GB364321,2 10

Current Computing Resources: Storage YearModelDisk/TapePhysical SizeUsable Size 2008 NetApp FAS2050 Disk(SAN)48TB30TB 2009 NetApp FAS6080 Disk(SAN, NAS)334TB200TB This Year Hitachi USP-V Disk(SAN)960TB600TB This Year-Tape100TB Total1442TB930TB 11

Some Pictures Newly delivered storage system in June 12

Network Architecture 8Gb SAN S/W …. 10GbE S/W … SAN U600TB FC Controller #1Controller #2 PFS Servers 10GbE S/W GLORIAD FC Controller 1,2 SAN U160TB 2009 : 8Gb FC : 1GbEthernet : 10GbEthernet 1G/10GbE S/W FC Controller 1,2 NAS U40TB NAS WN Servers … 1G Grid Servers 10G Grid Servers 1G Grid Servers …. Public Private 2010 IBRIX Fusion Cluster File System 13

2. Current Status SACC Collaboration Provide Service Provide Computing Resources STAR Asian Users (China, India, etc.) 14

Network Architecture 1F 신규 2 전산실 S/W 1F 신규 2 전산실 S/W DJ-F10 대전 OME6500 대전 OME6500 시애틀 OME6500 시애틀 OME6500 시애틀 OSR 시애틀 OSR DJ_7609 Cisco7609 DJ_7609 Cisco7609 1F –FDF(3,4) Foundry 2402CF Foundry 2402CF 2F 대용량 데이터팀 1G FDF(3,4) VLAN 124 IP : / 26 VLAN 233 IP : / 24 G3/18 G2/5 G8/21 G0/26 T2/3 G2/11 1G Starproject to BNL KISTI 2F 대용량데이터팀 LAB 3F –FDF(3,4) Foundry SuperX Foundry SuperX Gloriad-KR 10G Trunk Vlan 123, Vlan124 VLAN 123 IP : / 24 10G 1G Lightpath Vlan124 10G Trunk Vlan G 시애틀 F-10 10G 1G 15

Current Resources CPUs Intel Xeon E cores ≈ 960kSI2K (1core = 3.2kSI2K) shared queue Intel Xeon E cores ≈ 960kSI2K (1core = 3.2kSI2K) shared queue Storages Disk: 100TB Tape: 100TB Disk: 100TB Tape: 100TB Network 10G GLORIAD: Active Software SL10i on the Scientific Linux bits 16

Resource Plan CPU (kSI2K)9602,0803,100 Storage Disk (TB) Tape (TB) Network~10G (GLORID) 17

3. Future Plan of SACC Before 2010 Jan Summer 2011 Before 2010 Jan Summer 2011 Installation of additional CPUs (+430 cores) & Tests MOU between KISTI GSDC and BNL STAR Collaboration Start user service for Asian users MOU between KISTI GSDC and Asian institutes RUN11 – online data transfer from BNL to KISTI GSDC Installation of additional storages (100TB disk + 100TB tape) 18

Our Vision 19

Do You Have Any Questions? 20