Star Asian Computing Center Pusan National University In-Kwon YOO On Behalf of SACC Team J.Lauret, D.Yu, W. Bett, J. Packard S.D.Lee, D.K.Kim, H.W.KimE.

Slides:



Advertisements
Similar presentations
The RHIC-ATLAS Computing Facility at BNL HEPIX – Edinburgh May 24-28, 2004 Tony Chan RHIC Computing Facility Brookhaven National Laboratory.
Advertisements

Distributed IT Infrastructure for U.S. ATLAS Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
BNL Oracle database services status and future plans Carlos Fernando Gamboa RACF Facility Brookhaven National Laboratory, US Distributed Database Operations.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
Data oriented job submission scheme for the PHENIX user analysis in CCJ Tomoaki Nakamura, Hideto En’yo, Takashi Ichihara, Yasushi Watanabe and Satoshi.
Computing/Tier 3 Status at Panjab S. Gautam, V. Bhatnagar India-CMS Meeting, Sept 27-28, 2007 Delhi University, Delhi Centre of Advanced Study in Physics,
Status Report on Tier-1 in Korea Gungwon Kang, Sang-Un Ahn and Hangjin Jang (KISTI GSDC) April 28, 2014 at 15th CERN-Korea Committee, Geneva Korea Institute.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
Charm Research in Heavy Ion Experiment Pusan National Univeristy HIPEx Lab. In-Kwon YOO For the HIM_SRC (Center for Quark Matter)
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
100G R&D at Fermilab Gabriele Garzoglio (for the High Throughput Data Program team) Grid and Cloud Computing Department Computing Sector, Fermilab Overview.
An Overview of PHENIX Computing Ju Hwan Kang (Yonsei Univ.) and Jysoo Lee (KISTI) International HEP DataGrid Workshop November 8 ~ 9, 2002 Kyungpook National.
GridKa SC4 Tier2 Workshop – Sep , Warsaw Tier2 Site.
November 16, 2012 Seo-Young Noh Haengjin Jang {rsyoung, Status Updates on STAR Computing at KISTI.
Introduction to U.S. ATLAS Facilities Rich Baker Brookhaven National Lab.
BNL Facility Status and Service Challenge 3 Zhenping Liu, Razvan Popescu, Xin Zhao and Dantong Yu USATLAS Computing Facility Brookhaven National Lab.
Network Tests at CHEP K. Kwon, D. Han, K. Cho, J.S. Suh, D. Son Center for High Energy Physics, KNU, Korea H. Park Supercomputing Center, KISTI, Korea.
TRIUMF Site Report for HEPiX/HEPNT, Vancouver, Oct20-24/2003 – Corrie Kost TRIUMF SITE REPORT Corrie Kost Head Scientific Computing.
GStore: GSI Mass Storage ITEE-Palaver GSI Horst Göringer, Matthias Feyerabend, Sergei Sedykh
Data transfer over the wide area network with a large round trip time H. Matsunaga, T. Isobe, T. Mashimo, H. Sakamoto, I. Ueda International Center for.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
PDSF at NERSC Site Report HEPiX April 2010 Jay Srinivasan (w/contributions from I. Sakrejda, C. Whitney, and B. Draney) (Presented by Sandy.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
JLAB Computing Facilities Development Ian Bird Jefferson Lab 2 November 2001.
O AK R IDGE N ATIONAL L ABORATORY U.S. D EPARTMENT OF E NERGY Facilities and How They Are Used ORNL/Probe Randy Burris Dan Million – facility administrator.
STAR Off-line Computing Capabilities at LBNL/NERSC Doug Olson, LBNL STAR Collaboration Meeting 2 August 1999, BNL.
BNL Wide Area Data Transfer for RHIC & ATLAS: Experience and Plans Bruce G. Gibbard CHEP 2006 Mumbai, India.
BNL Facility Status and Service Challenge 3 HEPiX Karlsruhe, Germany May 9~13, 2005 Zhenping Liu, Razvan Popescu, and Dantong Yu USATLAS/RHIC Computing.
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Dr Ian Bird LCG Project Leader 6 th October 2009 Telecom 2009 Youth Forum.
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory Review of U.S. LHC Software and Computing Projects Fermi National Laboratory November.
ATLAS Great Lakes Tier-2 (AGL-Tier2) Shawn McKee (for the AGL Tier2) University of Michigan US ATLAS Tier-2 Meeting at Harvard Boston, MA, August 17 th,
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
National HEP Data Grid Project in Korea Kihyeon Cho Center for High Energy Physics (CHEP) Kyungpook National University CDF CAF & Grid Meeting July 12,
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
U.S. ATLAS Computing Facilities Bruce G. Gibbard GDB Meeting 16 March 2005.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
State of LSC Data Analysis and Software LSC Meeting LIGO Hanford Observatory November 11 th, 2003 Kent Blackburn, Stuart Anderson, Albert Lazzarini LIGO.
Star Asian Computing Center
US ATLAS Western Tier 2 Status Report Wei Yang Nov. 30, 2007 US ATLAS Tier 2 and Tier 3 workshop at SLAC.
RCF Status - Introduction PHENIX and STAR Counting Houses are connected to RCF at a Network Bandwidth of 20 Gbits/sec each –Redundant (Bandwidth-wise and.
ATLAS Midwest Tier2 University of Chicago Indiana University Rob Gardner Computation and Enrico Fermi Institutes University of Chicago WLCG Collaboration.
January 30, 2016 RHIC/USATLAS Computing Facility Overview Dantong Yu Brookhaven National Lab.
Office of Science U.S. Department of Energy NERSC Site Report HEPiX October 20, 2003 TRIUMF.
Tier 1 at Brookhaven (US / ATLAS) Bruce G. Gibbard LCG Workshop CERN March 2004.
BNL Oracle database services status and future plans Carlos Fernando Gamboa, John DeStefano, Dantong Yu Grid Group, RACF Facility Brookhaven National Lab,
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
U.S. ATLAS Computing Facilities U.S. ATLAS Physics & Computing Review Bruce G. Gibbard, BNL January 2000.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Pathway to Petaflops A vendor contribution Philippe Trautmann Business Development Manager HPC & Grid Global Education, Government & Healthcare.
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
Global Science experimental Data hub Center April 25, 2016 Seo-Young Noh Status Report on KISTI’s Computing Activities.
100G R&D at Fermilab Gabriele Garzoglio (for the High Throughput Data Program team) Grid and Cloud Computing Department Computing Sector, Fermilab Overview.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
Belle II Physics Analysis Center at TIFR
Vanderbilt Tier 2 Project
5th DOSAR Workshop Louisiana Tech University Sept. 27 – 28, 2007
ATLAS Sites Jamboree, CERN January, 2017
Southwest Tier 2.
Computing Infrastructure for DAQ, DM and SC
Nuclear Physics Data Management Needs Bruce G. Gibbard
Expanding the PHENIX Reconstruction Universe
The LHCb Computing Data Challenge DC06
Presentation transcript:

Star Asian Computing Center Pusan National University In-Kwon YOO On Behalf of SACC Team J.Lauret, D.Yu, W. Bett, J. Packard S.D.Lee, D.K.Kim, H.W.KimE. Dart, E. Hjort

Star Asian Computing Center In-Kwon YOO Tsukuba2/25 Outline

Star Asian Computing Center 1. Motivationa. STAR Computing In-Kwon YOO Tsukuba STAR S&C Structure Our data processing path is - DAQ to DST - DST to MuDST : Size reduction is ~ 1 to 5 STAR Plan - FY 08: 370 Mevts - FY 09: 700 Mevts - FY 10:1200 Mevts - FY 11:1700 Mevts - FY 12:1700 Mevts DAQ DST/event MuDST production Analysis Raw Data JLauret

Star Asian Computing Center Tier 0 BY DEFINITION: RHIC Computing Facility (RCF) at BNL –Acquisition, recording and processing of Raw data ; main resource Services –Long term (permanent) archiving and serving of all data … but NOT sized for Monte Carlo generation –Allow re-distribution of ANY data to minimally Tier1 Functionality –Production reconstruction of most (all) Raw data, provide resource for data analysis –Has full support structure for users, code, … 1. Motivationa. STAR Computing In-Kwon YOO Tsukuba STAR Computing Sites Data transfer should flow from Tier0 to Tier2 Tier 1 Services –Provide some persistent storage, source from Tier0 –MUST Allow redistribution of data to Tier2 –Provide backup (additional copies) for Tier0 –May requires a support structure (user account, ticket, response, grid operations,…) Functionality –Provides a significant added value to processing capabilities of STAR –Half of the analysis power AND/OR Provide support for running embedding, simulation, … –Redistribution of data to Tier2 Tier2 Would host transient datasets – requires only several 100 GB Mostly for local groups, provide analysis power for specific topics MUST provide cycles (opportunistic) for at least simulation –Low requirement of Grid operation support or common project JLauret

Star Asian Computing Center 1. Motivationa. STAR Computing In-Kwon YOO Tsukuba STAR Computing Sites 6 main dedicated sites (STAR software fully installed) BNLTier0 NERSC/PDSF Tier1 WSU (Wayne State University)Tier2 SPU (Sao Paulo U.)Tier2 BHAM (Birmingham, England)Tier2 UIC (University of Illinois, Chicago)Tier2 Incoming PragueTier2 KISTITier1 JLauret

Star Asian Computing Center Universidade de São Paulo PDSF Berkeley LAB Brookhaven National Lab Fermi Lab University of Birmingham Wayne State University

Star Asian Computing Center Amazon.com MIT X-grid SunGrid NPI, Czech Republic Interoperability / outreachVirtualizationVDT extensionSRM / DPM / EGEE STAR is also outreaching other grid resources & projects

Star Asian Computing Center Tsukuba FY05FY06FY07FY08FY09FY10FY11FY12 STAR Requirement Real Data Volume (TB) Reco CPU (KSI2K) Analys CPU (KSI2K) Dist. Disk (TB) Cent. Disk (TB) Annual Tape Volume (TB) Tape bandwidth (MB/sec) WAN bandwidth (Mb/sec) Simulation CPU (KSI2K) Simulation Data Volume (TB) /25 In-Kwon YOO Motivationa. STAR Computing JLauret

Star Asian Computing Center Tsukuba 9/25 In-Kwon YOO 1. Motivationa. STAR Computing JLauret

Star Asian Computing Center Tsukuba 10/25 In-Kwon YOO 1. Motivationa. STAR Computing Observation: Cost seem to go into CPU BUT this folds distributed disks (1/2 cost) Reduced used of centralized disk is nonetheless third in cost Cost is clearly Storage (~ ½) CPU (1/3 rd ) HPSS & LAN is second K$ JLauret

Star Asian Computing Center Tsukuba 11/25 In-Kwon YOO 1. Motivationb. KISTI Resources Cluster system Item Cluster system Phase 1Phase 2 Manufacturer & ModelSUN C48SUN Fusion ArchitectureCluster Processor AMD Opteron 2GHz (Barcelona) Intel Xeon 3.3GHz+ (Gainestown) Operating SystemCent OS Nodes1882,688 CPU cores 3,008 (16/node) 21,504 (8/node) Rpeak24TFlops286TFlops Memory6TB64.5TB Disk storage207TB1PB Tape storage422TB2PB Interconnection networkInfiniband 4X DDR CoolingChilled water cooling Delivery dateJan, 20082Q, 2009 SDLee, HWKim

Star Asian Computing Center Tsukuba 12/25 In-Kwon YOO 1. Motivationb. KISTI Resources Item SMP system Phase 1Phase 2 Manufacturer & ModelIBM p595IBM p6H ArchitectureSMP ProcessorPOWER5+ 2.3GHzPOWER6 5GHz+ Operating systemAIX 5.3AIX 5.3+ Nodes1024 CPU cores 640 (64/node) 1,536 (64/node) Rpeak5.9TFlops30.7TFlops Memory2.6TB9.2TB Disk storage63TB273TB Tape storage- Interconnection networkHPSInfiniband 4X DDR CoolingAir-cooling Delivery dateSept, 20071Q, 2009 SMP system SDLee, HWKim

Star Asian Computing Center Tsukuba 13/25 In-Kwon YOO 1. Motivationb. KISTI Resources Research Networks KREONET SDLee, HWKim

Star Asian Computing Center Tsukuba 14/25 In-Kwon YOO 1. Motivationb. KISTI Resources GLORIAD SDLee, HWKim

Star Asian Computing Center Tsukuba 15/25 In-Kwon YOO 2. Project a. STAR Asian HUB STAR Asian Hub

Star Asian Computing Center Tsukuba 16/25 In-Kwon YOO 2. Project a. STAR Asian HUB Star Asian Computing Center Computing Infrastructure with massive data from STAR –Frontier Research –Maximum Use of IT resources in Korea Data Transfer Cluster Computing with Supercomputer Mass Storage Korean Institute for Science and Technology Informations Daejoen) – Korean HUB for GLORIAD + KREONET – Super Computing Resources – Mass Storage Management  Asian Supercomputing HUB : - BNL – NERSC – KISTI – SSC etc.

Star Asian Computing Center In-Kwon YOO Tsukuba17/25 SACC Working Group PNU –IKYoo et al. KISTI –SDLee, DKKim, HWKim BNL (STAR) –JLauret, DYu, Wbett, Edart, JPackard SSC + Tsinghua Univ. ? –ZXiao et al. ? 2. Project b. Working Group

Star Asian Computing Center In-Kwon Tsukuba 2. Project c. KISTI STAR Computing KISTI STAR Computing Configuration of a testbed(16 cores) under way Eventually SUN cluster 1 st shipment (~3,000 cores) will be dedicated to STAR early next year! Initial Network status : below 1 Mbps (over 10Gbps line) Network Optimization between KISTI and BNL : since Target Throughput : over 2 Gbps (over LP) KISTI’s effort - Installed 10Gbps NIC equipped server in Seattle - Local optimization SDLee, HWKim

Star Asian Computing Center In-Kwon YOO Tsukuba 2. Project c. KISTI-BNL Networking Dantong YU

Star Asian Computing Center BNL to KISTI Data Transfer Network Tuning improved transfer from BNL to KISTI –Identified bottleneck with Kreonet2 peering point with Esnet. 1Gpbs => 10Gbps –Network and TCP stack was tuned at KISTI hosts. Before Tuning After Tuning In-Kwon YOO Tsukuba 2. Project c. KISTI-BNL Networking Dantong YU

Star Asian Computing Center Network Research There are two kinds of network events that re quire some further ex amination. 1.Delayed ramp up. 2.Sudden Drop in Ban dwidth. In-Kwon YOO Tsukuba 2. Project c. KISTI-BNL Networking Dantong YU

Star Asian Computing Center STAR Data Transfer Status Performance between BNL and KISTI : not symmetrical. A bottleneck from KISTI back to BNL. –Packet drops at BNL receiving host. (will be replaced). –Old Data Transfer Nodes at BNL: being replaced –Findings being corrected: High performance TCP parameters –Findings are still under investigation TCP slow ramp up, and performance sudden drop. test/tune GridFtp tools : in next 4 weeks In-Kwon YOO Tsukuba 2. Project c. KISTI-BNL Networking Dantong YU

Star Asian Computing Center STAR Data Transfer Plan Replace the old data transfer nodes –1 Gbps per node, with expansion slots for 10Gbps. –Large local disk for intermediate cache. Deploy OSG BestMan for these nodes. RACF firewall will be rearchitectured. Data transfer performance should be only limited by the local disk buffer at both ends. In-Kwon YOO Tsukuba 2. Project c. KISTI-BNL Networking Dantong YU

Star Asian Computing Center To do list KISTI needs to finish the testbed preparation STAR Software should be installed and tested KISTI net people need to set up lightpath between KISTI and BNL eventually BNL net people need to set up a host for the end- to-end test and measure the throughput We need to complete “a real mass data-transfer” from BNL to KISTI sooner or later! Start Production test at KISTI In-Kwon YOO Tsukuba 2. Project d. To-Do List

Star Asian Computing Center In-Kwon YOO Tsukuba25/25 Outlook towards HACC STAR Asian Computing Center (SACC) (2008 – 2011) – Experimental Data from International Facility Computational Infrastructure for LHC / Galaxity – Asian Hub for International Coworking – Frontier Research Heavy ion Analysis Computing Center (HACC) (2011-) –Extend to Other project (HIM) ? –Extend to ATHIC ? –Dedicated Resources for Heavy ion Analysis Computing 3. Outlook HACC

Star Asian Computing Center BNL PHENIX WAN Data Transfer Appendix PHENIX Dantong YU

Star Asian Computing Center PHENIX Data Transfer Infrastructure 27 Appendix PHENIX Dantong YU

Star Asian Computing Center 28 Computer Platforms on Both Ends  BNL: Multiple 3.0Ghz dual CPU nodes with I ntel copper gigabit network. Local drives co nnected by Raid Controller. There are PHENI X on-line hosts.  CCJ Site: 8 dual-core AMD Opteron based h osts, each with multiple Tera bytes SATA dri ves connected with a RAID controller. Each o ne has one gigabit broadcom network card.  The LAN on both ends are 10Gbps.  The data transfer tool is GridFtp. In-Kwon YOO Tsukuba Appendix PHENIX Dantong YU

Star Asian Computing Center PHENIX to CCJ Data Transfer Test at the beginning of 2008 (Mega Byte/second) We can transfer up to 340Mby te/second from BNL to CCJ, it hits the BNL firewall limitation. In-Kwon YOO Tsukuba Appendix PHENIX Dantong YU

Star Asian Computing Center 30 Data Transfer to CCJ, 2005  2005 RHIC run ended on Jun e 24, Above shows the last day of RHIC Run.  Total data transfer to CCJ (Co mputer Center in Japan) is 260 TB (polarized p+p raw data)  100% data transferred via WA N, Tool used here: GridFtp. No 747 involved.  Average Data Rate: 60~90MB /second, Peak Performance: 10 0 Mbytes/second recorded in Ganglia Plot! About 5TB/day! Courtesy of Y. Watanabe Appendix PHENIX Dantong YU

Star Asian Computing Center Data Transfer to CCJ, In-Kwon YOO Tsukuba Appendix PHENIX Dantong YU