HEPiX Fall Lincoln, Nebraska, Oct 13-17, 2014

Slides:



Advertisements
Similar presentations
Remus: High Availability via Asynchronous Virtual Machine Replication
Advertisements

Distributed Data Processing
Operating System.
Ceph: A Scalable, High-Performance Distributed File System
Ceph: A Scalable, High-Performance Distributed File System Sage Weil Scott Brandt Ethan Miller Darrell Long Carlos Maltzahn University of California, Santa.
Ceph: A Scalable, High-Performance Distributed File System Priya Bhat, Yonggang Liu, Jing Qin.
Copyright 2009 FUJITSU TECHNOLOGY SOLUTIONS PRIMERGY Servers and Windows Server® 2008 R2 Benefit from an efficient, high performance and flexible platform.
Wide-area cooperative storage with CFS
PRASHANTHI NARAYAN NETTEM.
NFS. The Sun Network File System (NFS) An implementation and a specification of a software system for accessing remote files across LANs. The implementation.
Module – 7 network-attached storage (NAS)
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
VMware vCenter Server Module 4.
System Design/Implementation and Support for Build 2 PDS Management Council Face-to-Face Mountain View, CA Nov 30 - Dec 1, 2011 Sean Hardman.
© 2013 Mellanox Technologies 1 NoSQL DB Benchmarking with high performance Networking solutions WBDB, Xian, July 2013.
Status Report on Ceph Based Storage Systems at the RACF
Hands-On Microsoft Windows Server 2008 Chapter 1 Introduction to Windows Server 2008.
U.S. Department of the Interior U.S. Geological Survey David V. Hill, Information Dynamics, Contractor to USGS/EROS 12/08/2011 Satellite Image Processing.
Server Load Balancing. Introduction Why is load balancing of servers needed? If there is only one web server responding to all the incoming HTTP requests.
Cloud MapReduce : a MapReduce Implementation on top of a Cloud Operating System Speaker : 童耀民 MA1G Authors: Huan Liu, Dan Orban Accenture.
Database Services for Physics at CERN with Oracle 10g RAC HEPiX - April 4th 2006, Rome Luca Canali, CERN.
Paper on Best implemented scientific concept for E-Governance projects Virtual Machine By Nitin V. Choudhari, DIO,NIC,Akola.

CSC 456 Operating Systems Seminar Presentation (11/13/2012) Leon Weingard, Liang Xin The Google File System.
 Cloud computing  Workflow  Workflow lifecycle  Workflow design  Workflow tools : xcp, eucalyptus, open nebula.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
An Introduction to Software Architecture
A BigData Tour – HDFS, Ceph and MapReduce These slides are possible thanks to these sources – Jonathan Drusi - SCInet Toronto – Hadoop Tutorial, Amir Payberah.
Ceph Storage in OpenStack Part 2 openstack-ch,
BINP/GCF Status Report BINP LCG Site Registration Oct 2009
Guide to Linux Installation and Administration, 2e1 Chapter 2 Planning Your System.
Large Scale Test of a storage solution based on an Industry Standard Michael Ernst Brookhaven National Laboratory ADC Retreat Naples, Italy February 2,
Building a Parallel File System Simulator E Molina-Estolano, C Maltzahn, etc. UCSC Lab, UC Santa Cruz. Published in Journal of Physics, 2009.
CEPH: A SCALABLE, HIGH-PERFORMANCE DISTRIBUTED FILE SYSTEM S. A. Weil, S. A. Brandt, E. L. Miller D. D. E. Long, C. Maltzahn U. C. Santa Cruz OSDI 2006.
Introduction to dCache Zhenping (Jane) Liu ATLAS Computing Facility, Physics Department Brookhaven National Lab 09/12 – 09/13, 2005 USATLAS Tier-1 & Tier-2.
7. CBM collaboration meetingXDAQ evaluation - J.Adamczewski1.
Fast Crash Recovery in RAMCloud. Motivation The role of DRAM has been increasing – Facebook used 150TB of DRAM For 200TB of disk storage However, there.
Ceph: A Scalable, High-Performance Distributed File System
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
11 CLUSTERING AND AVAILABILITY Chapter 11. Chapter 11: CLUSTERING AND AVAILABILITY2 OVERVIEW  Describe the clustering capabilities of Microsoft Windows.
70-293: MCSE Guide to Planning a Microsoft Windows Server 2003 Network, Enhanced Chapter 12: Planning and Implementing Server Availability and Scalability.
HADOOP DISTRIBUTED FILE SYSTEM HDFS Reliability Based on “The Hadoop Distributed File System” K. Shvachko et al., MSST 2010 Michael Tsitrin 26/05/13.
Architecture View Models A model is a complete, simplified description of a system from a particular perspective or viewpoint. There is no single view.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
A Silvio Pardi on behalf of the SuperB Collaboration a INFN-Napoli -Campus di M.S.Angelo Via Cinthia– 80126, Napoli, Italy CHEP12 – New York – USA – May.
 Introduction  Architecture NameNode, DataNodes, HDFS Client, CheckpointNode, BackupNode, Snapshots  File I/O Operations and Replica Management File.
RHIC/US ATLAS Tier 1 Computing Facility Site Report Christopher Hollowell Physics Department Brookhaven National Laboratory HEPiX Upton,
Cloud Computing Lecture 5-6 Muhammad Ahmad Jan.
PROOF tests at BNL Sergey Panitkin, Robert Petkus, Ofer Rind BNL May 28, 2008 Ann Arbor, MI.
Silberschatz, Galvin and Gagne ©2011 Operating System Concepts Essentials – 8 th Edition Chapter 2: The Linux System Part 1.
Awesome distributed storage system
Database CNAF Barbara Martelli Rome, April 4 st 2006.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Embedded Real-Time Systems Processing interrupts Lecturer Department University.
SOFTWARE DEFINED STORAGE The future of storage.  Tomas Florian  IT Security  Virtualization  Asterisk  Empower people in their own productivity,
An Introduction to GPFS
File system: Ceph Felipe León fi Computing, Clusters, Grids & Clouds Professor Andrey Y. Shevel ITMO University.
Bernd Panzer-Steindel CERN/IT/ADC1 Medium Term Issues for the Data Challenges.
Pilot Kafka Service Manuel Martín Márquez. Pilot Kafka Service Manuel Martín Márquez.
Course: Cluster, grid and cloud computing systems Course author: Prof
DSS-G Configuration Bill Luken – April 10th , 2017
Dag Toppe Larsen UiB/CERN CERN,
Section 6 Object Storage Gateway (RADOS-GW)
Section 7 Erasure Coding Overview
Oracle Solaris Zones Study Purpose Only
Real IBM C exam questions and answers
Chapter 2: System Structures
Chapter 2: The Linux System Part 1
Ceph Appliance – SAFE Storage Appliance For Enterprise
Presentation transcript:

HEPiX Fall 2014 @ Lincoln, Nebraska, Oct 13-17, 2014 Ceph Based Storage Systems for the RHIC & ATLAS Computing Facility at BNL Alexandr Zaytsev alezayt@bnl.gov BNL, USA RHIC & ATLAS Computing Facility HEPiX Fall 2014 @ Lincoln, Nebraska, Oct 13-17, 2014

HEPiX Fall 2014 @ Lincoln, Nebraska, Oct 13-17, 2014 Outline Ceph Project Overview History & Status Architecture & Scalability Features yet to be added Ceph @ RACF Early tests (2012Q4-2013Q3) Pre-production test systems (2013Q4-2014Q2) Production system (since 2014Q3) Summary & Conclusions Q & A Oct 16, 2014 HEPiX Fall 2014 @ Lincoln, Nebraska, Oct 13-17, 2014

Ceph Project: General Overview & Status Ceph is the rapidly developing open source all-in-one clustered storage solution that is partly originated from the DoE environment: “Ceph started off as a part of a research grant from the Department of Energy in cooperation with Los Alamos, Lawrence Livermore, and National Labs… the project is the culmination of years of research by several professors and many graduate students in the University of California at Santa Cruz School of Engineering” Oct 16, 2014 http://community.redhat.com/blog/2014/06/ceph-turns-10-a-look-back/

Ceph Project: General Overview & Status Ceph provides all three layers of clustered storage one could possibly (reasonably) want in a typical HTC environment: Object storage layer Native API storage cluster APIs (librados) Amazon S3 / Openstack Swift compatible APIs Object storage layer does not require special kernel modules, just needs TCP/IP connectivity between the components of the Ceph cluster Block storage layer (Rados Block Device, RBD) Mainly designed to be used by the hypervisors (QEMU/KVM , XEN, LXC, etc.) It can also be used as a raw device via ‘rbd’ kernle module fist introduced in the kernel 2.6.34 (RHEL 6.x is still on 2.6.32; RHEL 7 derived OS distributions need to get though to the majority of HEP/NP resources in order to get Ceph RBD supported everywhere) Current recommended kernel is 3.14.x (not 3.15.x at the moment!) (Nearly) POSIX-compliant distributed filesystem layer (CephFS) Just recently becoming ready for the large scale production Requires kernel modules (‘ceph’/’libceph’) also first introduced in kernel 2.6.34 Current recommended kernel is 3.14.x Oct 16, 2014 http://ceph.com/docs/v0.86/dev/differences-from-posix/

Ceph Architecture: Protocol Stack Oct 16, 2014 HEPiX Fall 2014 @ Lincoln, Nebraska, Oct 13-17, 2014

Ceph Project: General Overview & Status Ceph project has strong industry drivers behind it: OpenStack Cinder and Swift components of Openstack are becoming interfaces around Ceph/RBD and Ceph object storage layer in many large scale installations since 2012 DreamHost One of the first largest users of Ceph/RBD (as a block storage backend for their Openstack based Cloud) Inktank >> RedHat “On April 30, 2014, Red Hat announced the acquisition of Inktank, the company behind Ceph, a popular open source, freely available storage system… Together, Ceph and Gluster provide the open source community with a complementary set of software-defined storage options for their object, block, and file storage needs.” Number of code authors behind the Ceph project (max. scale is 60) Oct 16, 2014 http://community.redhat.com/ceph-faq/

HEPiX Fall 2014 @ Lincoln, Nebraska, Oct 13-17, 2014 Ceph Architecture: Components / Scalability http://storageconference.us/2012/Presentations/M05.Weil.pdf Oct 16, 2014 HEPiX Fall 2014 @ Lincoln, Nebraska, Oct 13-17, 2014

Ceph Architecture: Networking / NTP NTP synchronization is critical for the Ceph cluster Monitors can be severely affected by significant clock skews across the MON nodes By default, the maximum tolerated clock skew is 50 ms

HEPiX Fall 2014 @ Lincoln, Nebraska, Oct 13-17, 2014 Ceph Architecture: CRUSH Algorithm http://www.ssrc.ucsc.edu/Papers/weil-sc06.pdf Ceph internal object placement / replication mechanism: “…CRUSH (Controlled Replication Under Scalable Hashing), a pseudo-random data distribution algorithm that efficiently and robustly distributes object replicas across a heterogeneous, structured storage cluster. CRUSH is implemented as a pseudo-random, deterministic function that maps an input value, typically an object or object group identifier, to a list of devices on which to store object replicas.” “The cluster map is composed of devices and buckets, both of which have numerical identifiers and weight values associated with them. Buckets can be composed arbitrarily to construct a hierarchy representing available storage.” “Calculating a CRUSH mapping is designed to be fast O(log n) for a cluster with n OSDs, so that devices can quickly locate any object or re-evaluate the proper storage targets for the objects that they already store after a cluster map change“ “Mapping calculations have O(log n) running time, requiring only tens of microseconds to execute with thousands of devices“ Oct 16, 2014 HEPiX Fall 2014 @ Lincoln, Nebraska, Oct 13-17, 2014

Ceph Architecture: Replication / Recovery http://storageconference All data replicated N times (N ≤ 10) Ceph OSD cluster handles replication Client writes to first replica (unlike GPFS) Replication factor can be dynamically changed on per pool basis Replicas can be steered to selected groups of OSDs Erasure (forward error correction) codes are supported by CRUSH algorithm since Ceph v0.80 Firefly release (though it is still as an experimental feature) Dynamic cluster Nodes are added, removed (reboot, fail, recover) Recovery is a normal state of a Ceph cluster: Ceph cluster map records cluster state at point in time ceph-osd node status (up/down, weight, IP) CRUSH function specifying desired data distribution ceph-osds cooperatively migrate data to achieve that Any map update potentially triggers data migration ceph-osds monitor peers for failure New nodes register with monitor Configuration adjustments by administrator All these features ensure a remarkable level of survivability of a Ceph cluster with data replica-tion factor N > 1

Ceph Architecture: Authentication / Authorization Ceph Authentication (cephx protocol) Ceph Authorization (through Ceph CAPS) A key scalability feature of Ceph is to avoid a centralized interface to the Ceph object store Ceph clients must be able to interact with OSDs directly Ceph provides its cephx authenti-cation system, which authenticates users operating Ceph clients cephx behaviour similar to the one of Kerberos The cephx protocol authenticates Ceph clients / servers to each other It is not intended to handle authenti-cation of human users or applications If that effect is required to handle your access control needs, you must have another mechanism Ceph uses the term “capabilities” (CAPS) to describe authorizing an authenticated user to exercise the functionality of the MONs, OSDs and MDSes Capabilities can also restrict access to data within one or more pools Capabilities are in effect regardless of what type of client accesses the Ceph object store CephFS uses a different type of capability for files and directories internal to the CephFS filesystem CephFS filesystem access controls are relevant to CephFS, but not block devices or the RESTful gateway (Rados GW) ceph-authtool -n <user-id> --cap osd \ 'allow rwx pool=<pool-id>' cephx

Ceph as an Alternative to Existing Products Cleversafe (hardware/software object storage systems) Commercial product; its functionality is going to be completely covered by Ceph once the erasure codes become a production feature MapR M3/M5/M7 (Hadoop-derived software systems) Commercial product; Ceph can be used as a backend for HDFS with existing feature set IBM General Parallel File System (GPFS) Commercial product; used by RACF in production starting from 2014 (see James Pryor’s talk “BNL RACF Site Report” for more details) Lustre – even with Distributed Application Object Storage (DAOS) extensions for Exascale computing systems Ceph over RDMA features are expected to appear in the next (Hammer) release in 2015Q1 (implementation base on Accelio a high-performance asynchronous reliable messaging and RPC library) RDMA support in Ceph is going to be a game changer with respect to using low latency interconnect (such as Infiniband: switching from IPoIB to RDMA may result in factor of 30x latency drop) as a cluster network and may trigger spreading of Ceph among many HPC sites

Ceph @ RACF: Early Tests (2012Q4-2013Q3) First Tests in the virtualized environment (2012Q4, v0.48) Ceph/RBD testbed using refurbished dCache hardware (2013Q1, v0.61) A single hypervisor based virtual testbed (64 GB RAM): 4x OSD SL 6.x VMs (data disks in RAM, OSDs deployed with XFS underneath) 3x MON SL 6.x VMs 12x Fedora 17 client VMs (mapping RBD volumes & mounting CephFS) Main results: Confirm basic survivability features of the Ceph cluster Confirm the stability of Ceph/RBD layer (try operations with 106 objects in 103 pools) Generate up to 2 GB/s of internal (virtual) traffic in replicas re-balancing / RBD tests Observe instability of CephFS (clearly not production ready at that point) OSD/BTRFS was also attempted: turned out to be highly unstable Small real hardware testbed: 5x Sun Thor (4540) storage servers retired from BNL ATLAS dCache “Symmetrical” cluster configuration: MON + MDS + OSD on every node; the same nodes used as clients; OSDs deployed on top of Ext4; OSD journals are on the same devices 235 HDDs / 20 CPU cores in total Thors are re-installed from Solaris to Fedora 18 Single 10 GbE Myricom NICs on every node (delivering only 3.2 Gbps between the nodes as measured with iperf) Main results: Stability features confirmed on real hardware for both Ceph object storage and Ceph/RBD layers Networking and CPU are the performance killers for this setup: only 0.5 GB/s of throughput is reached in the RBD tests (CPU/NIC saturation)

Ceph @ RACF: Pre-production Test Systems (2013Q4-2014Q2) Ceph in the PHENIX Infiniband Testbed (2013Q4, Ceph v0.72) 26x compute nodes from the 2013 purchase for RACF PHENIX farm (brand new, 32x HT CPU cores, 64 GB RAM, 12x 2 TB 7.2krpm SATA HDDs in HW RAID5 on each) Provided with IPoIB/4X FDR IB interconnect solution(56 Gbps, 70 µs latency) based on Mellanox hardware (factor 3.5x oversubscribed 2-level tree topology) Symmetrical configuration of the cluster again: 1x MON+OSD on every node (no MDSes since CephFS was not tested in this setup) This layout corresponded to a possible configuration of a Ceph cluster providing backend storage for PHENIX dCache 9.4 TB of raw capacity is given to an OSD on every node (with XFS underneath); 0.25 PB of raw capacity over 312 HDDs 18 GB/s of internal Ceph cluster throughput is demonstrated with object replica re-balancing operations 11 GB/s of throughput is demonstrated with read tests over the rbd-fuse mountpoints

Ceph @ RACF: Pre-production Test Systems (2013Q4-2014Q2) Hitachi HUS 130 Storage System (2014Q2, Ceph v0.72.2) Please see Alexandr Zaytsev’s talk “Evaluating Infiniband Based Networking Solutions for HEP/NP Data Processing Applications” for more information on Infiniband

Ceph @ RACF: Pre-production Test Systems (2013Q4-2014Q2) This configuration is close to ideal, though it had an additional layer of disk capacity aggregation on the storage head nodes, which is not needed for Ceph Hitachi HUS 130 Storage System (2014Q2, Ceph v0.72.2) – cont. Half of the new storage backend purchased for the BNL ATLAS dCache: 3x storage head nodes (each holding a redundant 1+1 RAID controller serving 240x 4 TB HDDs each, organized into 8x RAID-6 sets) 6x server nodes (32x HT CPU cores, 64 GB RAM; 1x MON, 6x OSDs on each, OSDs deployed over XFS) 2x bonded 10 GbE & 1x IPoIB/4X FDR IB network uplinks on every node (FDR IB for internal cluster network) 2x SSDs on every node for the OSD journals (1 GB/s of I/O throughput guaranteed on the SSD subsystem) 2.2 PB of raw capacity / 720x HDDs in total (4 racks of equipment) Results obtained: 2.4 GB/s (sustained) / 3.7 GB/s (peak) internal Ceph cluster throughput was reached in the replica re-balancing / RBD tests Moving internal Ceph cluster communications from 2x 10 GbE bonded uplinks to IPoIB/4X FDR IB delivered +56% increase to the sustained I/O rate Furthermore, moving the OSD journals to dedicated SSDs delivered additional +70% These two factors together give a factor of 2.7x improvement of the cluster performance Our tests were saturating CPUs on the RAID controllers up to 95%

Ceph @ RACF: CephFS on the HP Moonshot System CephFS evaluation with HP Moonshot test system (2014Q2, Ceph v0.72.2) 30x HP ProLiant m300 cartridges in the HP Moonshot 1500 chassis Cartridge specs: 4x 8 GB DDR3 1600 DIMMs 1x Intel Atom C2750 CPU @ 2.40 GHz (8 physical cores) 1x 1 TB 6.0 Gb/s SATA Drive 1x 1 GbE uplink Self referring NTP installation was needed within the chassis Deployed under SL6 with custom built 3.14 kernel for CephFS tests 8x MONs + 16x OSDs + 6 client nodes (in non-overlapping groups) Only one Ceph component per cartridge (mixing OSDs with CephFS was creating stability issues) CephFS stability is observed for the first time on the timescale of 1 week with 3.10+ kernels

Ceph @ RACF: Production System (Since 2014Q3) The first large production installation of Ceph for RACF The main purpose is to serve as a temporary object data storage in front of the BNL ATLAS dCache (to be used mainly by the ATLAS Event Service) Base on the refurbished storage hardware being removed from the BNL ATLAS dCache system Deployment is done along the following steps: Step 1, Jun 2014 (2 rack): 4x Ceph cluster nodes 3x Sun Thors + 9x Nexsan disk arrays Step 2, Jul-Aug 2014 (7 racks): 8x Ceph cluster nodes, 16x Thors + 45x Nexsans as raw capacity providers; 4X FDR IB interconnect for internal communications between the OSDs. 0.6 PB of usable capacity / 2.5k HDDs: 6x MONs + 45x OSDs + 2x RadosGW 4.2 GB/s of internal throughput, 3.5k objects/s creation/deletion rate Step 3, expected in Nov 2014 (13 racks): 30x Thors + 74x Nexsans (a temporary configuration with > 1 PB of usable capacity with factor of 3x data replication still using ISCSI exports from Thors) Step 4, expected in 2014Q4-2015Q1 (11-12 racks): 22x Thors + 74x Nexsans, dropping iSCSI exports from the system completely; 1 PB of usable capacity, 4.6k HDDs, 40 GB/s throughput in all cross-section of the system starting from the disk arrays up to Rados GWs and external uplinks

Current Layout (since Aug 2014) SX6018 19

Proposed Upgrade 1 (Nov 2014) SX6018 20

≈ 1 PB of usable space 4.6k HDDs / 11 racks Proposed Upgrade 2 (2014Q4-2015Q1) ≈ 1 PB of usable space 4.6k HDDs / 11 racks 40 GB/s max throughput SX6018 21

Ceph @ RACF: Amazon S3 / Rados GW Interfaces We provide a redundant pair of Ceph Rados Gateways over the httpd/FastCGI wrapper around the RGW CLI that facing both into BNL intranet and to the outside world Allows one to use Amazon S3 compliant tools to access the content of the RACF Ceph object storage system (externally mostly used by CERN for initial ATLAS Event Service tests) We do not support an Amazon style bucket name resolution via subdomains such as http://<bucket_id>.cephgw01/<object_id>, but have an easier custom implemented schema instead that supports URLs like http://cephgw02/<bucket_id>/<object_id> We also provide an extended functionality for viewing the compressed logs stored in our Ceph instance (unpacking in flight, etc.) Additional RACF / ATLAS Event Service Specific Object Retrieval Web Interfaces

Ceph @ RACF: Amazon S3 / Rados GW Interfaces Comparing the performance of Ceph APIs on different levels for the large bulks of small objects (less than 1 kB in size) Using the low level librados Python API to write a bulk of up to 20k objects of identical size per thread to a newly created pool in a synchronous mode (waiting for an object to fully propagate through the cluster before handling the next one) 1.2k objects/s (maximum achieved with 32 threads / 4 client hosts running in parallel) Using the same low level librados Python API to write the same set of objects in an asynchronous mode (no waiting for object propagation through the cluster) 3.5k objects/s (maximum achieved in the same conditions) Using the AWS/S3 Ruby API (ruby 2.1.2p95 & aws-s3-0.6.3) to access both Rados GW instances (FastCGI/httpd/RGW CLI) of the Ceph cluster simultaneously and perform object creation in a synchronous mode 0.43k objects/s (maximum achieved with 128 threads / 4 hosts accessing both Rados GW instances in parallel) Further scaling of the Rados GW subsystem is needed in order to ensure the efficient use of the lower level Ceph API performance capabilities (to be addressed during the proposed upgrade 2 period) Oct 16, 2014 HEPiX Fall 2014 @ Lincoln, Nebraska, Oct 13-17, 2014

Summary & Conclusions Ceph is a rapidly developing open source clustered storage solution that provides object storage, block storage and distributed filesystem layers Ceph is spreads fast among the HTC community in general and HEP/NP community in particular With additional features such as Ceph over RDMA (expected to be available in 2015Q1), it may get into many HPC sites soon enough With the OpenStack community and RedHat backing it, and with the proportion of kernel 3.x based Linux distributions used in production steadily increasing, Ceph is likely to become a de facto standard for open source clustered storage solutions in a few years A long sequence of Ceph functionality and performance tests was carried out in RACF during the period of 2012Q4-2014Q2 that is finally resulted in deployment of a large (1 PB) scale RACF Ceph object storage system that started in 2014Q3 and should be completed by the end of the year Scaling from 1x 1U server to 7 racks of equipment in less than 2 years (and growing) Various storage and networking layouts were compared in the process 4X FDR Infiniband technology was successfully adopted within this installation as the internal cluster interconnect solution We are going to continue evaluating Ceph/RBD and CephFS for possible future applications, e.g. as a dCache storage backend and a distributed filesystem solution alternative to IBM GPFS We are looking forward to use this system as an object storage backend for the ATLAS Event Service starting from 2014Q4

Q & A Total Solar Eclipse of Aug 21, 2017 73.1 sec