IHEP Computing Site Report Shi, Jingyan Computing Center, IHEP.

Slides:



Advertisements
Similar presentations
Report of Liverpool HEP Computing during 2007 Executive Summary. Substantial and significant improvements in the local computing facilities during the.
Advertisements

CURRENT AND FUTURE HPC SOLUTIONS. T-PLATFORMS  Russia’s leading developer of turn-key solutions for supercomputing  Privately owned  140+ employees.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
IHEP Site Status Jingyan Shi, Computing Center, IHEP 2015 Spring HEPiX Workshop.
S. Gadomski, "ATLAS computing in Geneva", journee de reflexion, 14 Sept ATLAS computing in Geneva Szymon Gadomski description of the hardware the.
Bill Wrobleski Director, Technology Infrastructure ITS Infrastructure Services.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Tier 2 Prague Institute of Physics AS CR Status and Outlook J. Chudoba, M. Elias, L. Fiala, J. Horky, T. Kouba, J. Kundrat, M. Lokajicek, J. Svec, P. Tylka.
Agenda Network Infrastructures LCG Architecture Management
1 INDIACMS-TIFR TIER-2 Grid Status Report IndiaCMS Meeting, Sep 27-28, 2007 Delhi University, India.
Tier 3g Infrastructure Doug Benjamin Duke University.
Computing/Tier 3 Status at Panjab S. Gautam, V. Bhatnagar India-CMS Meeting, Sept 27-28, 2007 Delhi University, Delhi Centre of Advanced Study in Physics,
Status Report on Tier-1 in Korea Gungwon Kang, Sang-Un Ahn and Hangjin Jang (KISTI GSDC) April 28, 2014 at 15th CERN-Korea Committee, Geneva Korea Institute.
A. Mohapatra, HEPiX 2013 Ann Arbor1 UW Madison CMS T2 site report D. Bradley, T. Sarangi, S. Dasu, A. Mohapatra HEP Computing Group Outline  Infrastructure.
SLAC National Accelerator Laboratory Site Report A National Lab in Transition Randy Melen, Deputy CIO Computing Division, Operations Directorate SLAC National.
FZU Computing Centre Jan Švec Institute of Physics of the AS CR, v.v.i
David Hutchcroft on behalf of John Bland Rob Fay Steve Jones And Mike Houlden [ret.] * /.\ /..‘\ /'.‘\ /.''.'\ /.'.'.\ /'.''.'.\ ^^^[_]^^^ * /.\ /..‘\
BINP/GCF Status Report BINP LCG Site Registration Oct 2009
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Oxford Update HEPix Pete Gronbech GridPP Project Manager October 2014.
Progress Energy Corporate Data Center Rob Robertson February 17, 2010 of North Carolina.
Spending Plans and Schedule Jae Yu July 26, 2002.
Site Report BEIJING-LCG2 Wenjing Wu (IHEP) 2010/11/21.
A.Golunov, “Remote operational center for CMS in JINR ”, XXIII International Symposium on Nuclear Electronics and Computing, BULGARIA, VARNA, September,
ITEP computing center and plans for supercomputing Plans for Tier 1 for FAIR (GSI) in ITEP  8000 cores in 3 years, in this year  Distributed.
Panoptic Capacity Planning Presented by. "Scotty, I need warp speed in 3 minutes or we're all dead!” (William Shatner - Star Trek II ‘The Wrath of Khan’)
Summer 2012 A Brief Review of Technology Upgrades at Mead Hall Technology Committee Meeting.
IHEP Computing Center Site Report Shi, Jingyan Computing Center, IHEP.
ATLAS Tier 1 at BNL Overview Bruce G. Gibbard Grid Deployment Board BNL 5-6 September 2006.
Nikhef/(SARA) tier-1 data center infrastructure
11 November 2010 Natascha Hörmann Computing at HEPHY Evaluation 2010.
Gareth Smith RAL PPD RAL PPD Site Report. Gareth Smith RAL PPD RAL Particle Physics Department Overview About 90 staff (plus ~25 visitors) Desktops mainly.
IHEP(Beijing LCG2) Site Report Fazhi.Qi, Gang Chen Computing Center,IHEP.
Computing Jiří Chudoba Institute of Physics, CAS.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
Tier 3 Status at Panjab V. Bhatnagar, S. Gautam India-CMS Meeting, July 20-21, 2007 BARC, Mumbai Centre of Advanced Study in Physics, Panjab University,
Power and Cooling at Texas Advanced Computing Center Tommy Minyard, Ph.D. Director of Advanced Computing Systems 42 nd HPC User Forum September 8, 2011.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Final Implementation of a High Performance Computing Cluster at Florida Tech P. FORD, X. FAVE, K. GNANVO, R. HOCH, M. HOHLMANN, D. MITRA Physics and Space.
RAL PPD Tier 2 (and stuff) Site Report Rob Harper HEP SysMan 30 th June
BaBar Cluster Had been unstable mainly because of failing disks Very few (
Eygene Ryabinkin, on behalf of KI and JINR Grid teams Russian Tier-1 status report May 9th 2014, WLCG Overview Board meeting.
BEIJING-LCG Network Yan Xiaofei
ASGC Site Report Felix Lee HEPiX 2011 Fall In Vancouver 24 Oct, 2011.
Scientific Computing in PPD and other odds and ends Chris Brew.
IHEP Site Status Qiulan Huang Computing Center, IHEP,CAS HEPIX FALL 2015.
A. Mohapatra, T. Sarangi, HEPiX-Lincoln, NE1 University of Wisconsin-Madison CMS Tier-2 Site Report D. Bradley, S. Dasu, A. Mohapatra, T. Sarangi, C. Vuosalo.
The Worldwide LHC Computing Grid Frédéric Hemmer IT Department Head Visit of INTEL ISEF CERN Special Award Winners 2012 Thursday, 21 st June 2012.
Development of a Tier-1 computing cluster at National Research Centre 'Kurchatov Institute' Igor Tkachenko on behalf of the NRC-KI Tier-1 team National.
PIC port d’informació científica First operational experience from a compact, highly energy efficient data center module V. Acín, R. Cruz, M. Delfino,
CD-doc-650 Fermilab Computing Division Physical Infrastructure Requirements for Computers FY04 – 07 (1/4/05)
Dominique Boutigny December 12, 2006 CC-IN2P3 a Tier-1 for W-LCG 1 st Chinese – French Workshop on LHC Physics and associated Grid Computing IHEP - Beijing.
IHEP Computing Center Site Report Gang Chen Computing Center Institute of High Energy Physics 2011 Spring Meeting.
1 1 – Statistical information about our resource centers; ; 2 – Basic infrastructure of the Tier-1 & 2 centers; 3 – Some words about the future.
The status of IHEP Beijing Site WLCG Asia-Pacific Workshop Yaodong CHENG IHEP, China 01 December 2006.
CD-doc-650 Fermilab Computing Division Physical Infrastructure Requirements for Computers FY04 – 07 (1/4/05)
IHEP Computing Center Site Report Shi, Jingyan Computing Center, IHEP.
Status of BESIII Computing
SuperB – INFN-Bari Giacinto DONVITO.
Bob Ball/University of Michigan
The Beijing Tier 2: status and plans
Paul Kuipers Nikhef Site Report Paul Kuipers
Belle II Physics Analysis Center at TIFR
CERN Data Centre ‘Building 513 on the Meyrin Site’
Experience of Lustre at a Tier-2 site
Luca dell’Agnello INFN-CNAF
HIGH-PERFORMANCE COMPUTING SYSTEM FOR HIGH ENERGY PHYSICS
ATLAS Sites Jamboree, CERN January, 2017
AGLT2 Site Report Shawn McKee/University of Michigan
BEIJING-LCG2 Site Report
Presentation transcript:

IHEP Computing Site Report Shi, Jingyan Computing Center, IHEP

Outline Local Cluster Local Cluster LCG Site LCG Site Network Network Infrastructure Upgrade Infrastructure Upgrade Summary Summary Shi,Jingyan CC--IHEP

Local Cluster users, 200+ active ones users, 200+ active ones For BES,YBJ,DYB experiments For BES,YBJ,DYB experiments job slots (include 1500 slots new added)6500+ job slots (include 1500 slots new added) Storage:Storage:  3PB+ lustre  5PB+ tape lib Scheduler: Torque + mauiScheduler: Torque + maui Shi,Jingyan CC--IHEP

Trouble in Lustre Lustre had been running well with high performance Lustre had been running well with high performance MDS problem happened by the end of Sept. MDS problem happened by the end of Sept. Big task to save data from the damaged Lustre Big task to save data from the damaged Lustre Shi,Jingyan CC--IHEP

Trouble in Lustre (cont.) New rules need to be established to regulate storage usage New rules need to be established to regulate storage usage Limit users’ small size filesLimit users’ small size files Data file and user file should be kept in separated storageData file and user file should be kept in separated storage Any suggestion are warmly welcomeAny suggestion are warmly welcome Shi,Jingyan CC--IHEP

LCG Tier II Site For CMS, ATLAS experiments For CMS, ATLAS experiments Job slots Job slots Storage: Storage: 320TB dCache320TB dCache 320TB dpm320TB dpm 1T disks will be replaced by 2T disks1T disks will be replaced by 2T disks 50T extra space will be added50T extra space will be added Shi,Jingyan CC--IHEP

BEIJING-LCG2 Site report Shi,Jingyan CC--IHEP

BEIJING-LCG2 Site report Shi,Jingyan CC--IHEP

Reliability and Availability Shi,Jingyan CC--IHEP

10 * IHEP Campus(Office) Network Star structure 10G Backbone WIFI Covered Over 3000 Users IPv4/IPv6 available for Users 10G IPv4 & IPv6 Link to CSTNet Shi,Jingyan CC--IHEP

Orient + Network connection Daya Bay Beijing CSTNet Hong Kong IHEP USA GLORIAD 10G ASGC IPv4 10G IPv6 Beijing Tsinghua YBJ EUR. 2.5G 155M 2.5G Others EDU.CN 10G Shi,Jingyan CC--IHEP

Two hosts for perfsonar Two hosts for perfsonar Perfsonar.ihep.ac.cn for Bandwidth testPerfsonar.ihep.ac.cn for Bandwidth test Perfsonar2.ihep.ac.cn for Latency testPerfsonar2.ihep.ac.cn for Latency test Network performance tuning is in progress between IHEP and Eur. Sites Network performance tuning is in progress between IHEP and Eur. Sites iew/InternationalConnectivity/IHE P-CCIN2P3http://twiki.ihep.ac.cn/twiki/bin/v iew/InternationalConnectivity/IHE P-CCIN2P3http://twiki.ihep.ac.cn/twiki/bin/v iew/InternationalConnectivity/IHE P-CCIN2P3http://twiki.ihep.ac.cn/twiki/bin/v iew/InternationalConnectivity/IHE P-CCIN2P3 Discussing with related people about the possibility in connecting IHEP to LHCONE Discussing with related people about the possibility in connecting IHEP to LHCONE Shi,Jingyan CC--IHEP

Upgrade for Data Center Network Device Expansion Device Expansion PerformancePerformance  The 10G(2Gbps->10Gbps) firewall is ready(based on Linux & iptables ) The lack of 10G portsThe lack of 10G ports  Some devices are under test FROCE10 Z9000/4810 FROCE10 Z9000/4810 Arista 7148/7508 Arista 7148/7508 Topology Upgrade Topology Upgrade The Grid Area is isolatedThe Grid Area is isolated  Arista 7148: for the area core switch Shi,Jingyan CC--IHEP

Before the Upgrade Before the Upgrade Power consumption reached 90% of total capacityPower consumption reached 90% of total capacity Power supply of per rack can not support high density blade serversPower supply of per rack can not support high density blade servers Single-phase supply can not meet the needs of power systemSingle-phase supply can not meet the needs of power system Infrastructure Upgrade ——Power System Upgrade Shi,Jingyan CC--IHEP

Add one power transformer Add one power transformer Power Capacity: 800kw -> 1800kwPower Capacity: 800kw -> 1800kw Increase the power supply of one rack Increase the power supply of one rack Power Supply Mode : Single-phase supply  Three-phase supplyPower Supply Mode : Single-phase supply  Three-phase supply Power supply for one rack : 10kw  28kwPower supply for one rack : 10kw  28kw Power System Upgrade Shi,Jingyan CC--IHEP

Before the Upgrade Before the Upgrade Reached 80% of its total capacity Reached 80% of its total capacity Limited space limit the increase of cooling system Limited space limit the increase of cooling system Air cooling conditioner can not support high density blade servers Air cooling conditioner can not support high density blade servers ——Overheated island caused by high density blade servers ——Overheated island caused by high density blade servers Infrastructure Upgrade -- Cooling System Upgrade Shi,Jingyan CC--IHEP

Water cooling rack Water cooling rack Inter-row air conditioning Inter-row air conditioning Cooling capacity per rack reaches 28kw Cooling capacity per rack reaches 28kw Cooling System Upgrade Shi,Jingyan CC--IHEP

Sound barrier screen of outdoor unit Sound barrier screen of outdoor unit Reduce running noise Reduce running noise Cooling air partition need to be built Cooling air partition need to be built Improve cooling efficiency Improve cooling efficiency Monitoring System Monitoring System Infrastructure Upgrade ——Unfinished Work Shi,Jingyan CC--IHEP

Outdoor Unit installation Shi,Jingyan CC--IHEP

Outdoor Pipeline Installation Shi,Jingyan CC--IHEP

Power Distribution Cabinet Installation Shi,Jingyan CC--IHEP

Water cooling Rack Installation Shi,Jingyan CC--IHEP

System Tuning Shi,Jingyan CC--IHEP

Most part of computing environment running well Most part of computing environment running well Trouble in Storage Trouble in Storage Infrastructure upgrade meet its aim Infrastructure upgrade meet its aim Power supply 800kw  1800kwPower supply 800kw  1800kw Eliminate overheated islandEliminate overheated island Outlet air temperature of servers : 40 ℃  27 ℃Outlet air temperature of servers : 40 ℃  27 ℃ Summary Shi,Jingyan CC--IHEP

Thank you! Questions? Shi,Jingyan– Kan, Bowen/CC/IHEP