G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th The Naples' testbed for the SuperB computing model: first tests G. Russo, D. Del Prete, S.

Slides:



Advertisements
Similar presentations
Overview of MapReduce and Hadoop
Advertisements

4/11/2017 © 2014 Microsoft Corporation. All rights reserved. Microsoft, Windows, and other product names are or may be registered trademarks and/or trademarks.
Alastair Dewhurst, Dimitrios Zilaskos RAL Tier1 Acknowledgements: RAL Tier1 team, especially John Kelly and James Adams Maximising job throughput using.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
Real Parallel Computers. Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra, Meuer, Simon Parallel.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
Gordon: Using Flash Memory to Build Fast, Power-efficient Clusters for Data-intensive Applications A. Caulfield, L. Grupp, S. Swanson, UCSD, ASPLOS’09.
CPP Staff - 30 CPP Staff - 30 FCIPT Staff - 35 IPR Staff IPR Staff ITER-India Staff ITER-India Staff Research Areas: 1.Studies.
Outline Network related issues and thinking for FAX Cost among sites, who has problems Analytics of FAX meta data, what are the problems  The main object.
Real Parallel Computers. Modular data centers Background Information Recent trends in the marketplace of high performance computing Strohmaier, Dongarra,
1 INDIACMS-TIFR TIER-2 Grid Status Report IndiaCMS Meeting, Sep 27-28, 2007 Delhi University, India.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
© 2013 Mellanox Technologies 1 NoSQL DB Benchmarking with high performance Networking solutions WBDB, Xian, July 2013.
Edge Based Cloud Computing as a Feasible Network Paradigm(1/27) Edge-Based Cloud Computing as a Feasible Network Paradigm Joe Elizondo and Sam Palmer.
A Workflow-Aware Storage System Emalayan Vairavanathan 1 Samer Al-Kiswany, Lauro Beltrão Costa, Zhao Zhang, Daniel S. Katz, Michael Wilde, Matei Ripeanu.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
1 A Basic R&D for an Analysis Framework Distributed on Wide Area Network Hiroshi Sakamoto International Center for Elementary Particle Physics (ICEPP),
Corporate Partner Overview and Update September 27, 2007 Gary Crane SURA Director IT Initiatives.
CERN - European Laboratory for Particle Physics HEP Computer Farms Frédéric Hemmer CERN Information Technology Division Physics Data processing Group.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
UTA Site Report Jae Yu UTA Site Report 4 th DOSAR Workshop Iowa State University Apr. 5 – 6, 2007 Jae Yu Univ. of Texas, Arlington.
Hadoop Hardware Infrastructure considerations ©2013 OpalSoft Big Data.
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Sandor Acs 05/07/
Data transfer over the wide area network with a large round trip time H. Matsunaga, T. Isobe, T. Mashimo, H. Sakamoto, I. Ueda International Center for.
HPC system for Meteorological research at HUS Meeting the challenges Nguyen Trung Kien Hanoi University of Science Melbourne, December 11 th, 2012 High.
Wide Area Network Access to CMS Data Using the Lustre Filesystem J. L. Rodriguez †, P. Avery*, T. Brody †, D. Bourilkov *, Y.Fu *, B. Kim *, C. Prescott.
RAC parameter tuning for remote access Carlos Fernando Gamboa, Brookhaven National Lab, US Frederick Luehring, Indiana University, US Distributed Database.
LOGO PROOF system for parallel MPD event processing Gertsenberger K. V. Joint Institute for Nuclear Research, Dubna.
ITEP computing center and plans for supercomputing Plans for Tier 1 for FAIR (GSI) in ITEP  8000 cores in 3 years, in this year  Distributed.
On High Performance Computing and Grid Activities at Vilnius Gediminas Technical University (VGTU) dr. Vadimas Starikovičius VGTU, Parallel Computing Laboratory.
KIT – The cooperation of Forschungszentrum Karlsruhe GmbH und Universität Karlsruhe (TH) Hadoop on HEPiX storage test bed at FZK Artem Trunov.
Monte Carlo Data Production and Analysis at Bologna LHCb Bologna.
Challenges of deploying Wide-Area-Network Distributed Storage System under network and reliability constraints – A case study
HADOOP DISTRIBUTED FILE SYSTEM HDFS Reliability Based on “The Hadoop Distributed File System” K. Shvachko et al., MSST 2010 Michael Tsitrin 26/05/13.
December 26, 2015 RHIC/USATLAS Grid Computing Facility Overview Dantong Yu Brookhaven National Lab.
A Silvio Pardi on behalf of the SuperB Collaboration a INFN-Napoli -Campus di M.S.Angelo Via Cinthia– 80126, Napoli, Italy CHEP12 – New York – USA – May.
January 30, 2016 RHIC/USATLAS Computing Facility Overview Dantong Yu Brookhaven National Lab.
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
PROOF tests at BNL Sergey Panitkin, Robert Petkus, Ofer Rind BNL May 28, 2008 Ann Arbor, MI.
Scaling up from local DB to distributed DB Cristiano Bozza European Emulsion Group Nagoya, Jan 2004 Presented by Giuseppe Grella.
1 Farm Issues L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
15-Feb-02Steve Traylen, RAL WP6 Test Bed Report1 RAL/UK WP6 Test Bed Report Steve Traylen, WP6 PPGRID/RAL, UK
January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 1 SIMULATION OF DAILY ACTIVITITIES AT REGIONAL CENTERS MONARC Collaboration Alexander Nazarenko.
FroNtier Stress Tests at Tier-0 Status report Luis Ramos LCG3D Workshop – September 13, 2006.
HADOOP Dr. Silvio Pardi INFN-Naples.
16 th IEEE NPSS Real Time Conference 2009 IHEP, Beijing, China, 12 th May, 2009 High Rate Packets Transmission on 10 Gbit/s Ethernet LAN Using Commodity.
Performance analysis comparison Andrea Chierici Virtualization tutorial Catania 1-3 dicember 2010.
S. Pardi Frascati, 2012 March GPGPU Evaluation – First experiences in Napoli Silvio Pardi.
Storage at SMU OSG Storage 9/22/2010 Justin Ross Southern Methodist University.
HELMHOLTZ INSTITUT MAINZ Dalibor Djukanovic Helmholtz-Institut Mainz PANDA Collaboration Meeting GSI, Darmstadt.
1062m0656 between 10692m2192 DS/ICI/CIF EqualLogic PS6510E
S. Pardi Computing R&D Workshop Ferrara 2011 – 4 – 7 July SuperB R&D on going on storage and data access R&D Storage Silvio Pardi
Testing the Zambeel Aztera Chris Brew FermilabCD/CSS/SCS Caveat: This is very much a work in progress. The results presented are from jobs run in the last.
SuperB – Naples Site Dr. Silvio Pardi. Right now the Napoli Group is employed in 3 main tasks relate the computing in SuperB Fast Simulation Electron.
Dynamic Extension of the INFN Tier-1 on external resources
TYBIS IP-Matrix Virtualized Total Video Surveillance System Edge Technology, World Best Server Virtualization.
Video Security Design Workshop:
Distributed storage, work status
Experience of Lustre at QMUL
Solid State Disks Testing with PROOF
A testbed for the SuperB computing model
Diskpool and cloud storage benchmarks used in IT-DSS
Silvio Pardi R&D Storage Silvio Pardi
Christine Leroy Storage performance Christine Leroy
Experience of Lustre at a Tier-2 site
Introduction to HDFS: Hadoop Distributed File System
The Basics of Apache Hadoop
QMUL Site Report by Dave Kant HEPSYSMAN Meeting /09/2019
Presentation transcript:

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th The Naples' testbed for the SuperB computing model: first tests G. Russo, D. Del Prete, S. Pardi

Frascati, 2011 april 4th-7th The rationale Following the Ferrara and Frascati meeting, we decided to setup a testbed for SuperB applications The testbed is based on 12 servers (8 core, 32 GB, 8 disks), connected with a 2x10 GbE A further server (48 core) under order The tests will be based on: – RoCE, MPI, data access, dataset distribution, monitoring tools 2

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th The configuration of each node 3 10 Gbps 1 Gbps 2x CPU 2,13GHz 4Core, 4,80GT/s + 4,80GT/s QPI 8x 4GB Dual Rank 1066MHz 8x HDD 500GB 7200rpm SATA Hot-Plug 3,5” UTP Twinax... 8x one node (12 nodes in this test) 96 core (8x 12), 384 GB Ram (32x 12), 48 TB HD (4x 12)

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th The configuration of the shared storage 4 48 core (12x cpu), 64 GB Ram (16x cpu), 18 TB HD Storage Server for the redundant and shared data

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th 5 Twinax 10Gbps connection Gigabit Ethernet connection Optical fiber Lan2 10Gbps Optical fiber Lan1 10Gbps 4x Twinax 10Gbps connection Server 48 Core 64 GB di RAM Storage 18 TB Rack n

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th WN1/DataNode 10 GB Switch WN2/DataNode Shared/Redundant data Server WN1/DataNode 10 GB Switch WN2/DataNode Shared/Redundant data Server WN12/DataNode Rack 1 Rack n 2x 10 GB 4x 10 GB 2x 10 GB Computing Model 6 4x CPU 12 Core 8x 8GB di RAM Storage 6x 3TB 2x CPU 2,13GHz 4Core 8x 4GB 8x HDD 500GB

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th 7 Network Optimization Parameters: MTU 9000 # TCP buffer sizes net.ipv4.tcp_rmem = net.ipv4.tcp_wmem = net.ipv4.tcp_mem = net.core.rmem_default = net.core.wmem_default = net.core.rmem_max = net.core.wmem_max = # SACK and timestamps - turn off net.ipv4.tcp_dsack = 0 net.ipv4.tcp_sack = 0 net.ipv4.tcp_timestamps = 0 # Network backlog net.core.netdev_max_backlog = Basic: perfomance tests Test Latency : ms Test Iperf : 5.6 Gbits/s Computing Model

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th DOUBLE FILE SYSTEM CONFIGURATION CURRENT SETUP: OS SL 5.5 with gLite 3.2 Middleare. The Nodes are ready to be include in the Atlas TIER2 and available for SuperB FastSim Program (to do in the next weeks) IMPLEMENTATION OF THE HADOOP DISTRIBUTED FILE SYSTEM ON 8 NODE IMPLEMENTATION OF GLUSTERFS DISTRIBUTED FILE SYSTEM ON 9 NODE 8

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th HDSF-HADOOP IMPLEMENTATION HDSF Name Node WN1/DataNode WN2/DataNode WN../DataNode WN8/DataNode 10 GIGABIT SW HDFS FILE SYSTEM 1 NAME NODE 8 WORKER NODE / DATANODES that share the local disk All the DataNode mount HDSF Posix Inferface through Fuse module 9

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th GLUSTER IMPLEMENTATION WN1/DataNode WN2/DataNode WN../DataNode WN9/DataNode 10 GIGABIT SW GLUSTER NO NAME NODE 9 WORKER NODE / DATANODES that share the local disk All the DataNode are client and server. The Gluster FS is configured with replica 3 GLUSTERFS IS A FILE CLUSTER FILE SYSTEM THAT WORK WITHOUT DATA-NODES. IT CAN WORK IN 3 DIFFERENT CONFIGURATION: DISTRIBUTED, STRIPED AND REPLICA. IT USES THE POSIX EXENSION TO GUARANTEE THE CONSISTENCY OF FS. 10

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th First Benchmark Perfomance test executed through a SuperB Analysis job provided by Alejandro Perez. It executes a real analysis test on a set of simulated root files previously replicated in the gluster-fs volume. The following tests show the bandwidth occupation during a set of different job execution setup 11

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th Network Occupation in a Single Node Single Job max rate 59MB/s 12

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th Network Occupation in a Single Node 13 2 Jobs max rate 113MB/s - 1 Gigabit nick cart is still enough but the CPU power is used at 25% - can code parallelization enhance this ratio? 113MB/s

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th Network Occupation in a Single Node 8 Jobs (max box size) max rate 333MB/s 100% CPU Used but Network is at 30% more CPU can support from more 10Gbit/s Nic Cards MB/s

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th Network Occupation over the box core size 16 Job on a 8 CPU Node Whithout multi threading 24 Job on a 8 CPU Node Without multi threading Max Rate 422 MB/s Need to extimate the GLUSTERFS Overhead MB/s 422MB/s

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th Network Occupation 8 job in each node In this test we sent 8 analysis jobs for each 8 core server We Compute an Aggregate Network occupation of 10 Gbit/s In the Cluster 16

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th Analysis The Network Trend during data analysis Jobs show that: We have a set of SuperB Analysis that use 20-50MB/s To support future multicore architecture (24/48) we must move to a 10-20Gbit/s Network (nics now offer two ports); cost of switch is dominant (if using more than 24 ports) We have some option in the open source File System that can support 400MB/s (GLUSTER File System, Hadoop) A correct configuration and tuning of 10Gbit/s Nic cards is needed 17

G. Russo, D. Del Prete, S. Pardi Frascati, 2011 april 4th-7th CONCLUSION Preliminary tests on a full 10Gbit/s Farm show that the possibility to have data and computing in the same node is strictly correlate with the available network (for a 48Core two or four 10 Gbit/s nick card will be needed). Interesting issue: Comparison between different cluster file-system are interesting in terms of overhead, reliability and network usage and services offered. Comparison with different setup or number of replica with the same FS. Can Code Parallelization between core improve the ratio CPU used by Job vs Network requirement for data access? 18