Data oriented job submission scheme for the PHENIX user analysis in CCJ Tomoaki Nakamura, Hideto En’yo, Takashi Ichihara, Yasushi Watanabe and Satoshi.

Slides:



Advertisements
Similar presentations
Site Report: The Linux Farm at the RCF HEPIX-HEPNT October 22-25, 2002 Ofer Rind RHIC Computing Facility Brookhaven National Laboratory.
Advertisements

Computing Infrastructure
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
Southgrid Status Pete Gronbech: 27th June 2006 GridPP 16 QMUL.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
March 27, IndiaCMS Meeting, Delhi1 T2_IN_TIFR of all-of-us, for all-of-us, by some-of-us Tier-2 Status Report.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
1 A Basic R&D for an Analysis Framework Distributed on Wide Area Network Hiroshi Sakamoto International Center for Elementary Particle Physics (ICEPP),
CCJ Computing Center in Japan for spin physics at RHIC T. Ichihara, Y. Watanabe, S. Yokkaichi, O. Jinnouchi, N. Saito, H. En’yo, M. Ishihara,Y.Goto (1),
Operating System. Architecture of Computer System Hardware Operating System (OS) Programming Language (e.g. PASCAL) Application Programs (e.g. WORD, EXCEL)
The Application of DAQ-Middleware to the J-PARC E16 Experiment E Hamada 1, M Ikeno 1, D Kawama 2, Y Morino 1, W Nakai 3, 2, Y Obara 3, K Ozawa 1, H Sendai.
 Introduction to Operating System Introduction to Operating System  Types Of An Operating System Types Of An Operating System  Single User Single User.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
RIKEN CCJ Project Regional computing center in Japan for BNL-RHIC experiment especially for PHENIX collaboration. CCJ serves for RHIC physics activity.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
CERN - IT Department CH-1211 Genève 23 Switzerland t Tier0 database extensions and multi-core/64 bit studies Maria Girone, CERN IT-PSS LCG.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
An Overview of PHENIX Computing Ju Hwan Kang (Yonsei Univ.) and Jysoo Lee (KISTI) International HEP DataGrid Workshop November 8 ~ 9, 2002 Kyungpook National.
The Computing System for the Belle Experiment Ichiro Adachi KEK representing the Belle DST/MC production group CHEP03, La Jolla, California, USA March.
Site Report: Tokyo Tomoaki Nakamura ICEPP, The University of Tokyo 2013/12/13Tomoaki Nakamura ICEPP, UTokyo1.
INDIACMS-TIFR Tier 2 Grid Status Report I IndiaCMS Meeting, April 05-06, 2007.
Data transfer over the wide area network with a large round trip time H. Matsunaga, T. Isobe, T. Mashimo, H. Sakamoto, I. Ueda International Center for.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Presented by Leadership Computing Facility (LCF) Roadmap Buddy Bland Center for Computational Sciences Leadership Computing Facility Project.
SLAC Site Report Chuck Boeheim Assistant Director, SLAC Computing Services.
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
Optimisation of Grid Enabled Storage at Small Sites Jamie K. Ferguson University of Glasgow – Jamie K. Ferguson – University.
PHENIX Computing Center in Japan (CC-J) Takashi Ichihara (RIKEN and RIKEN BNL Research Center ) Presented on 08/02/2000 at CHEP2000 conference, Padova,
Nov. 8, 2000RIKEN CC-J RIKEN CC-J (PHENIX Computing Center in Japan) Report N.Hayashi / RIKEN November 8, 2000 PHENIX Computing
USATLAS dCache System and Service Challenge at BNL Zhenping (Jane) Liu RHIC/ATLAS Computing Facility, Physics Department Brookhaven National Lab 10/13/2005.
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
GRID activities in Wuppertal D0RACE Workshop Fermilab 02/14/2002 Christian Schmitt Wuppertal University Taking advantage of GRID software now.
BNL Service Challenge 3 Status Report Xin Zhao, Zhenping Liu, Wensheng Deng, Razvan Popescu, Dantong Yu and Bruce Gibbard USATLAS Computing Facility Brookhaven.
CERN Computer Centre Tier SC4 Planning FZK October 20 th 2005 CERN.ch.
Status of the Bologna Computing Farm and GRID related activities Vincenzo M. Vagnoni Thursday, 7 March 2002.
ISG We build general capability Introduction to Olympus Shawn T. Brown, PhD ISG MISSION 2.0 Lead Director of Public Health Applications Pittsburgh Supercomputing.
November 10, 1999PHENIX CC-J Updates in Nov.991 PHENIX CC-J Updates in Nov New Hardware - N.Hayashi / RIKEN November 10, 1999 PHENIX Computing Meeting.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
May 10, 2000PHENIX CC-J Updates1 PHENIX CC-J Updates - Preparation For Opening - N.Hayashi / RIKEN May 10, 2000 PHENIX Computing
Final Implementation of a High Performance Computing Cluster at Florida Tech P. FORD, X. FAVE, K. GNANVO, R. HOCH, M. HOHLMANN, D. MITRA Physics and Space.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, R. Brock,T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina,
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
CCJ introduction RIKEN Nishina Center Kohei Shoji.
PADME Kick-Off Meeting – LNF, April 20-21, DAQ Data Rate - Preliminary estimate Tentative setup: all channels read with Fast ADC 1024 samples, 12.
September 26, 2003K User's Meeting1 CCJ Usage for Belle Monte Carlo production and analysis –CPU time: 170K hours (Aug.1, 02 ~ Aug.22, 03)
Compute and Storage For the Farm at Jlab
Operating System.
LCG Deployment in Japan
Bernd Panzer-Steindel, CERN/IT
Luca dell’Agnello INFN-CNAF
Architecture & System Overview
Southwest Tier 2.
Computing Infrastructure for DAQ, DM and SC
Storage Virtualization
Welcome to our Nuclear Physics Computing System
湖南大学-信息科学与工程学院-计算机与科学系
Interoperability of Digital Repositories
Near Real Time Reconstruction of PHENIX Run7 Minimum Bias Data From RHIC Project Goals Reconstruct 10% of PHENIX min bias data from the RHIC Run7 (Spring.
Welcome to our Nuclear Physics Computing System
K computer RIKEN Advanced Institute for Computational Science
K computer RIKEN Advanced Institute for Computational Science
Presentation transcript:

Data oriented job submission scheme for the PHENIX user analysis in CCJ Tomoaki Nakamura, Hideto En’yo, Takashi Ichihara, Yasushi Watanabe and Satoshi Yokkaichi RIKEN Nishina Center for Accelerator-Based Science 2010/10/20Tomoaki Nakamura 1

PHENIX Computing Center in Japan (CCJ) 2010/10/20 Tomoaki Nakamura 2 Tape robot Disk storage and PC node Disk storage and PC node

PHENIX experiment at RHIC 2010/10/20 Tomoaki Nakamura 3 arXiv: (2010 ) QGP physics by heavy-ion beam Spin physics by polarized proton beam Phys. Rev. Lett. 104, (2010)

CCJ specification 2010/10/20 Tomoaki Nakamura 4 For the data production: Network: 10Gbps from BNL to RIKEN (SINET3) GridFTP servers: 4+2 servers (directly connected with buffer box in PHENIX counting room) Tape storage: HPSS 7.1 TS3500 library 2.5PB (Maxim 8PB) For the user analysis: PC cluster: ~252core 160core (Joint operation with RIKEN Integrated Cluster of Clusters) Disk storage: ~320TB Analysis environments: Offline library Data bases Maintaining equivalent resources since the start of operation (2000). Raw or Reconstructed Reconstructed

Data transfer record by GridFTP 2010/10/20 Tomoaki Nakamura 5 RHIC-Run5 RHIC-Run6 RHIC-Run8 RHIC-Run9 1.5PB of raw and reconstructed data was stored as of 2010.

Effective utilization for the user analysis 2010/10/20 Tomoaki Nakamura 6 Preparing special nodes with large capacity of local disk to store the read only data in advance. Silicon VTX

Disk configuration and I/O performance 2010/10/20 Tomoaki Nakamura 7 HP ProLiant DL180 G5 Chassis: 2U, 12HDD-bay CPU: Quad-core 2 2.66GHz Xeon E5430 Memory: 16GB (RDIM) HDD: 146GB SAS 2 for the system (RAID1) 1TB SATA 10 for the data storage NIC: 1Gbps Uplinked by 10Gbps switch in rack. System OS: Scientific Linux 5.3 x86_64 File system: ext3 XFS has no gain for the read only usage. HDD configuration RAID5 configuration: Advantage is only for 1 process. Good performance without RAID: Total ~550MB/sec at 8 parallel jobs.

Data allocation 2010/10/20 Tomoaki Nakamura 8 PC Cluster Total 18 nodes Total 144 CPU cores Total 180TB for the data storage Data sets in HDD Every HDD has a part of each data set segmented by run number. User jobs can use all of CPU cores without time loss at data distribution. ~40GB work area is allocated in each HDD for the non data analysis jobs e.g. simulation. Work area 1TB Run5 Run6 Run8 Run9 Work area 1TB Run5 Run6 Run8 Run9 Work area 1TB Run5 Run6 Run8 Run9 Work area 1TB Run5 Run6 Run8 Run9

Job submission scheme 2010/10/20 Tomoaki Nakamura 9 Job submission script Submit the user job to the appropriate node automatically, which has the required data. Batch Job scheduler (LSF 7.0) Fair-share among users. Set the path to the data directory and work area for environmental variables in the job. Lock file and directory ACL Create at the pre-process for the exclusive use of a physical disk. Second job, which will use the data in occupied disk by the first job, stay in queue until the end of the first job. LSF User got to be able to analyze all of the existing data in local HDD (150TB) within 9 hours.

Summary 2010/10/20 Tomoaki Nakamura calculating nodes with 180 TB local disks were introduced for effectively analyzing huge amounts of PHENIX data. A data-oriented batch queuing system was developed as a wrapper of the LSF system to increase the total computing throughput. Indeed, the total throughput was improved by roughly 10 times as compared to that in the existing clusters; CPU power and I/O performance are increased threefold and tenfold, respectively. Thus, users can analyze data of 150TB within 9 hours. We have experienced just 2 times disk failure out of 180 disks in 1.5 years operation. Therefore, we could conclude that this method is highly effective for the I/O bound data analysis.

Next step 2010/10/20 Tomoaki Nakamura 11 High availability: Reducing down time by the redundant data. Balance of job distribution with respect to any kind of jobs. Easy to expand for the growing data size: Without any other special software. No special servers and no additional HW components except for the cluster. By low cost: 144CPU core, 2GB memory/core, 180TB disk for the local data storage. ~$150K as of The capacity of a HDD is growing and its price is going down steadily. Preparing the same set of cluster with same data set,

Backup 2010/10/20 Tomoaki Nakamura 12

CCJ specification 2010/10/20 Tomoaki Nakamura 13 Riken Integrated Cluster of Clusters:

CCJ history 2010/10/20 Tomoaki Nakamura 14

HPSS data amount 2010/10/20 Tomoaki Nakamura 15

RIKEN Integrated Cluster of Clusters 2010/10/20 Tomoaki Nakamura 16

Luminosity at hadron collider 2010/10/20 Tomoaki Nakamura 17

Integrated luminosity A+A at RHIC 2010/10/20 Tomoaki Nakamura 18

PHENIX DAQ upgrade 2010/10/20 Tomoaki Nakamura 19 DAQ2010 SuperDAQ (i.e. sDAQ) James Nagle: Next Decade Plans for the PHENIX Experiment Workshop on Saturation, the Color Glass Condensate and Glasma