Academia Sinica Grid Computing Centre (ASGC), Taiwan

Slides:



Advertisements
Similar presentations
Condor use in Department of Computing, Imperial College Stephen M c Gough, David McBride London e-Science Centre.
Advertisements

National Grid's Contribution to LHCb IFIN-HH Serban Constantinescu, Ciubancan Mihai, Teodor Ivanoaica.
SLA-Oriented Resource Provisioning for Cloud Computing
ASGC Updated HSIN-YEN CHEN LHCONE workshop CERN 10 Feb
Site Report US CMS T2 Workshop Samir Cury on behalf of T2_BR_UERJ Team.
March 27, IndiaCMS Meeting, Delhi1 T2_IN_TIFR of all-of-us, for all-of-us, by some-of-us Tier-2 Status Report.
1 INDIACMS-TIFR TIER-2 Grid Status Report IndiaCMS Meeting, Sep 27-28, 2007 Delhi University, India.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Status Report on Tier-1 in Korea Gungwon Kang, Sang-Un Ahn and Hangjin Jang (KISTI GSDC) April 28, 2014 at 15th CERN-Korea Committee, Geneva Korea Institute.
Green technology used for ATLAS processing Dr. Ing. Fărcaş Felix NATIONAL INSTITUTE FOR RESEARCH AND DEVELOPMENT OF ISOTOPIC AND MOLECULAR.
A. Mohapatra, HEPiX 2013 Ann Arbor1 UW Madison CMS T2 site report D. Bradley, T. Sarangi, S. Dasu, A. Mohapatra HEP Computing Group Outline  Infrastructure.
23 Oct 2002HEPiX FNALJohn Gordon CLRC-RAL Site Report John Gordon CLRC eScience Centre.
Oxford Update HEPix Pete Gronbech GridPP Project Manager October 2014.
INDIACMS-TIFR Tier 2 Grid Status Report I IndiaCMS Meeting, April 05-06, 2007.
 Cloud computing is the use of computing resources (hardware and software) that are delivered as a service over a network (typically the Internet). 
Looking Ahead: A New PSU Research Cloud Architecture Chuck Gilbert - Systems Architect and Systems Team Lead Research CI Coordinating Committee Meeting.
Finding the minimum number of Sudoku clues Yu-Ting CHEN Academia Sinica Grid Computing.
1 PRAGUE site report. 2 Overview Supported HEP experiments and staff Hardware on Prague farms Statistics about running LHC experiment’s DC Experience.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
Evolution of a High Performance Computing and Monitoring system onto the GRID for High Energy Experiments T.L. Hsieh, S. Hou, P.K. Teng Academia Sinica,
Cluster Configuration Update Including LSF Status Thorsten Kleinwort for CERN IT/PDP-IS HEPiX I/2001 LAL Orsay Tuesday, December 08, 2015.
IHEP(Beijing LCG2) Site Report Fazhi.Qi, Gang Chen Computing Center,IHEP.
Andrea Manzi CERN On behalf of the DPM team HEPiX Fall 2014 Workshop DPM performance tuning hints for HTTP/WebDAV and Xrootd 1 16/10/2014.
Susanna Guatelli Geant4 in a Distributed Computing Environment S. Guatelli 1, P. Mendez Lorenzo 2, J. Moscicki 2, M.G. Pia 1 1. INFN Genova, Italy, 2.
ASCC Site Report Eric Yen & Simon C. Lin Academia Sinica 20 July 2005.
Data transfers and storage Kilian Schwarz GSI. GSI – current storage capacities vobox LCG RB/CE GSI batchfarm: ALICE cluster (67 nodes/480 cores for batch.
Development of a Tier-1 computing cluster at National Research Centre 'Kurchatov Institute' Igor Tkachenko on behalf of the NRC-KI Tier-1 team National.
Activities and Perspectives at Armenian Grid site The 6th International Conference "Distributed Computing and Grid- technologies in Science and Education"
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
CNAF - 24 September 2004 EGEE SA-1 SPACI Activity Italo Epicoco.
HEPiX IPv6 Working Group David Kelsey (STFC-RAL) GridPP33 Ambleside 22 Aug 2014.
Scientific Data Processing Portal and Heterogeneous Computing Resources at NRC “Kurchatov Institute” V. Aulov, D. Drizhuk, A. Klimentov, R. Mashinistov,
UNM SCIENCE DMZ Sean Taylor Senior Network Engineer.
HedEx Lite Obtaining and Using Huawei Documentation Easily
Shaopeng, Ho Architect of Chinac Group
Brief introduction about “Grid at LNS”
Energy Aware Network Operations
WLCG IPv6 deployment strategy
Computing Operations Roadmap
The Beijing Tier 2: status and plans
Grid site as a tool for data processing and data analysis
Dag Toppe Larsen UiB/CERN CERN,
Dag Toppe Larsen UiB/CERN CERN,
ATLAS Cloud Operations
WLCG Manchester Report
ALICE Monitoring
Andrea Chierici On behalf of INFN-T1 staff
Yaodong CHENG Computing Center, IHEP, CAS 2016 Fall HEPiX Workshop
Jan 12, 2005 Improving CMS data transfers among its distributed Computing Facilities N. Magini CERN IT-ES-VOS, Geneva, Switzerland J. Flix Port d'Informació.
Academia Sinica Grid Computing Centre
Kolkata Status and Plan
David Cameron ATLAS Site Jamboree, 20 Jan 2017
Support for IPv6-only CPU – an update from the HEPiX IPv6 WG
Oxford Site Report HEPSYSMAN
Welcome! Thank you for joining us. We’ll get started in a few minutes.
Update from the HEPiX IPv6 WG
Thoughts on Computing Upgrade Activities
Regional Software Defined Science DMZ (SD-SDMZ)
Статус ГРИД-кластера ИЯФ СО РАН.
PL 7.2 The Times They are A-Changin’
Deployment & Advanced Regular Testing Strategies
FCT Follow-up Meeting 31 March, 2017 Fernando Meireles
Integration of Singularity With Makeflow
OffLine Physics Computing
Grid Canada Testbed using HEP applications
Big-Data around the world
BusinessObjects IN Cloud ……InfoSol’s story
The Problem ~6,000 PCs Another ~1,000 boxes But! Affected by:
IPv6 update Duncan Rand Imperial College London
Exploring Multi-Core on
Presentation transcript:

Academia Sinica Grid Computing Centre (ASGC), Taiwan 1 ASGC site report Eric Yen, Felix Lee Academia Sinica Grid Computing Centre (ASGC), Taiwan HEPiX Fall 2016 in LBNL

ASGC site report Facility update Network(WAN) update 2 ASGC site report Facility update Network(WAN) update Activities/events update for Grid/Cloud/HPC Software collaboration

3 Facility update Still targeting high denisty solutions because the rack space is pretty tight in our DC. Battery solution is becoming our basic requirement for new server procurement. Both Supermicro and HP offers such solution. IB clusters are growing. QDR(40Gb) : 3200 cores FDR(40/56 GB): 2160 cores Another 2208 cores are coming with Supermicro battery solution: SYS-6028TP- HTR Over 1152 cpu cores(dual core platform) have been decommissioned:

IPv6 Deployment We finally made it 5 IPv6 Deployment We finally made it After a long fight with our new Brocade router... Have established the BGP peering sessions CERN, ESNET, GEANT, APAN-JP Enabling the perfSONAR Ipv6 Just need to configure ipv6 DNS.

SDN Deployment Have deployed SDN enabled router around the world. 6 SDN Deployment Have deployed SDN enabled router around the world. Brocade MLXe-4 & MLXe-8 One collector server directly connected to the each router Design to collect the s-flow network information Try to make use of those s-flow data as reference for SDN controller. Deploy the SDN controller OpenDaylight at ASGC DC. SDN is still in R&D state.

Grid Activites Network is still our major problem for world wide grid. 7 Grid Activites Network is still our major problem for world wide grid. Cisco N7K core switch has potential issue which would cause network performance drop to everywhere but local data center. Its design is grouping 4 ports to share one buffer, which is not enough in certain condition. e.g. long distance requires big tcp buffer, but big tcp buffer will make port buffer full quickly. It's a bit hard to identify this issue, because everything is just fine but long distance transfer. Thanks ESnet people for helping us to figure it out. Trying to reconnect disk server directly to router Brocade router line card offers dedicated port buffer.

Cloud service update Openstack Kilo 128 hypervisors (2256 cores). Upgrade plan to Newton is coming. The usage of docker container is growing It's good for CI. It's also convenient for some batch jobs. In most of cases, user doesn't have to install their applications to our cluster. And they can get bare-metal performance. Couple production services are hosted by docker container. e.g. ELK, some condor head nodes.

HPC Power saving There are some old and energy hungry IB clusters 9 HPC Power saving There are some old and energy hungry IB clusters Jobs will only go there when the other queues are full. There are also some clusters not always busy. Some of them are dedicated to certain applications, so, it's shame that we can't contribute them to WLCG... Of course, they are still consuming some power when there is no job running. It wasn't a big deal until these days, as we are now in green DC era. We are trying to deploy a mechanism to automatically: Turn off such worker node when there is no job coming Turn on them when jobs are coming. PoC is done, deploying it as pre-production service.

Software collaboration 1010 Software collaboration Keep collaboration with Atlas panda team to evolve our DiCOS software. Collaborating with STARLight on SDN and DTN(data transfer node) We hope DTN can be a little bit helpful to solving long distance transfer issue. Integrating openstack Ironic + SDN Don't follow openstack neutron. :)

Software collaboration 1111 Software collaboration Collaborating with University of Chicago on Parcel and BioInfometic activities. Parcel is based on UDT which can be a solution to solve long distance transfer issue... We can easily rearch 700MB/s to U.C by single 10Gb link . Also make use of our DiCOS system to support their bioinfo activities: RNAseq Mutation analysis for Cacner. Collaborating with our local Institute of Statistic on Financial risk assessment. Modeling and MC.

ISGC 2017 International Symposium on Grids and Clouds (ISGC) 2017 1212 ISGC 2017 International Symposium on Grids and Clouds (ISGC) 2017 CALL FOR PAPERS 5 ~ 10 March 2017, Academia Sinica, Taipei, Taiwan http://event.twgrid.org/isgc2017 Looking forward to seeing you in Taipei. :)

1313 Thanks a lot!