Use case of RDMA in Symantec storage software stack Om Prakash Agarwal Symantec.

Slides:



Advertisements
Similar presentations
Middleware Support for RDMA-based Data Transfer in Cloud Computing Yufei Ren, Tan Li, Dantong Yu, Shudong Jin, Thomas Robertazzi Department of Electrical.
Advertisements

The Development of Mellanox - NVIDIA GPUDirect over InfiniBand A New Model for GPU to GPU Communications Gilad Shainer.
NERCS Users’ Group, Oct. 3, 2005 Interconnect and MPI Bill Saphir.
Head-to-TOE Evaluation of High Performance Sockets over Protocol Offload Engines P. Balaji ¥ W. Feng α Q. Gao ¥ R. Noronha ¥ W. Yu ¥ D. K. Panda ¥ ¥ Network.
Evaluation of ConnectX Virtual Protocol Interconnect for Data Centers Ryan E. GrantAhmad Afsahi Pavan Balaji Department of Electrical and Computer Engineering,
Uncovering Performance and Interoperability Issues in the OFED Stack March 2008 Dennis Tolstenko Sonoma Workshop Presentation.
Performance Analysis of Virtualization for High Performance Computing A Practical Evaluation of Hypervisor Overheads Matthew Cawood University of Cape.
Ethernet Unified Wire November 15, 2006 Kianoosh Naghshineh, CEO Chelsio Communications.
Leader in Next Generation Ethernet. 2 Outline Where is iWARP Today? Some Proof Points Conclusion Questions.
Institute of Computer Science Foundation for Research and Technology – Hellas Greece Computer Architecture and VLSI Systems Laboratory Exploiting Spatial.
Remigius K Mommsen Fermilab A New Event Builder for CMS Run II A New Event Builder for CMS Run II on behalf of the CMS DAQ group.
VIA and Its Extension To TCP/IP Network Yingping Lu Based on Paper “Queue Pair IP, …” by Philip Buonadonna.
A Comparative Study of Network Protocols & Interconnect for Cluster Computing Performance Evaluation of Fast Ethernet, Gigabit Ethernet and Myrinet.
Embedded Transport Acceleration Intel Xeon Processor as a Packet Processing Engine Abhishek Mitra Professor: Dr. Bhuyan.
An overview of Infiniband Reykjavik, June 24th 2008 R E Y K J A V I K U N I V E R S I T Y Dept. Computer Science Center for Analysis and Design of Intelligent.
Sockets vs. RDMA Interface over 10-Gigabit Networks: An In-depth Analysis of the Memory Traffic Bottleneck Pavan Balaji  Hemal V. Shah ¥ D. K. Panda 
IWARP Ethernet Key to Driving Ethernet into the Future Brian Hausauer Chief Architect NetEffect, Inc.
P. Balaji, S. Bhagvat, D. K. Panda, R. Thakur, and W. Gropp
Practical TDMA for Datacenter Ethernet
© 2013 Mellanox Technologies 1 NoSQL DB Benchmarking with high performance Networking solutions WBDB, Xian, July 2013.
SRP Update Bart Van Assche,.
Windows RDMA File Storage
Protocols for Wide-Area Data-intensive Applications: Design and Performance Issues Yufei Ren, Tan Li, Dantong Yu, Shudong Jin, Thomas Robertazzi, Brian.
Reliable Datagram Sockets and InfiniBand Hanan Hit NoCOUG Staff 2010.
1 March 2010 A Study of Hardware Assisted IP over InfiniBand and its Impact on Enterprise Data Center Performance Ryan E. Grant 1, Pavan Balaji 2, Ahmad.
Towards a Common Communication Infrastructure for Clusters and Grids Darius Buntinas Argonne National Laboratory.
HPCS Lab. High Throughput, Low latency and Reliable Remote File Access Hiroki Ohtsuji and Osamu Tatebe University of Tsukuba, Japan / JST CREST.
Open Fabrics BOF Supercomputing 2008 Tziporet Koren, Gilad Shainer, Yiftah Shahar, Bob Woodruff, Betsy Zeller.
Boosting Event Building Performance Using Infiniband FDR for CMS Upgrade Andrew Forrest – CERN (PH/CMD) Technology and Instrumentation in Particle Physics.
© 2012 MELLANOX TECHNOLOGIES 1 The Exascale Interconnect Technology Rich Graham – Sr. Solutions Architect.
The NE010 iWARP Adapter Gary Montry Senior Scientist
Gilad Shainer, VP of Marketing Dec 2013 Interconnect Your Future.
Remote Direct Memory Access (RDMA) over IP PFLDNet 2003, Geneva Stephen Bailey, Sandburst Corp., Allyn Romanow, Cisco Systems,
InfiniBand in the Lab Erik 1.
Srihari Makineni & Ravi Iyer Communications Technology Lab
OFED Usage in VMware Virtual Infrastructure Anne Marie Merritt, VMware Tziporet Koren, Mellanox May 1, 2007 Sonoma Workshop Presentation.
© 2012 MELLANOX TECHNOLOGIES 1 Disruptive Technologies in HPC Interconnect HPC User Forum April 16, 2012.
Storage Interconnect Requirements Chen Zhao, Frank Yang NetApp, Inc.
7. CBM collaboration meetingXDAQ evaluation - J.Adamczewski1.
Integrating New Capabilities into NetPIPE Dave Turner, Adam Oline, Xuehua Chen, and Troy Benjegerdes Scalable Computing Laboratory of Ames Laboratory This.
Infiniband Bart Taylor. What it is InfiniBand™ Architecture defines a new interconnect technology for servers that changes the way data centers will be.
University of Mannheim1 ATOLL ATOmic Low Latency – A high-perfomance, low cost SAN Patrick R. Haspel Computer Architecture Group.
IT/EE Palaver FAIR DAQ - J.Adamczewski, S.Linev1.
Windows OpenFabrics (WinOF) Update Gilad Shainer, Mellanox Technologies November 2007.
The influence of system calls and interrupts on the performances of a PC cluster using a Remote DMA communication primitive Olivier Glück Jean-Luc Lamotte.
CMS week, June 2002, CERN 1 First P2P Measurements on Infiniband Luciano Berti INFN Laboratori Nazionali di Legnaro.
Sockets Direct Protocol Over InfiniBand in Clusters: Is it Beneficial? P. Balaji, S. Narravula, K. Vaidyanathan, S. Krishnamoorthy, J. Wu and D. K. Panda.
Using Heterogeneous Paths for Inter-process Communication in a Distributed System Vimi Puthen Veetil Instructor: Pekka Heikkinen M.Sc.(Tech.) Nokia Siemens.
Mellanox Connectivity Solutions for Scalable HPC Highest Performing, Most Efficient End-to-End Connectivity for Servers and Storage April 2010.
iSER update 2014 OFA Developer Workshop Eyal Salomon
Interconnect Networks Basics. Generic parallel/distributed system architecture On-chip interconnects (manycore processor) Off-chip interconnects (clusters.
Mellanox Connectivity Solutions for Scalable HPC Highest Performing, Most Efficient End-to-End Connectivity for Servers and Storage September 2010 Brandon.
Datacenter Fabric Workshop NFS over RDMA Boris Shpolyansky Mellanox Technologies Inc.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
L1/HLT trigger farm Bologna setup 0 By Gianluca Peco INFN Bologna Genève,
Remigius K Mommsen Fermilab CMS Run 2 Event Building.
A Practical Evaluation of Hypervisor Overheads Matthew Cawood Supervised by: Dr. Simon Winberg University of Cape Town Performance Analysis of Virtualization.
Advisor: Hung Shi-Hao Presenter: Chen Yu-Jen
FaRM: Fast Remote Memory Aleksandar Dragojević, Dushyanth Narayanan, Orion Hodson and Miguel Castro, Microsoft Research NSDI’14 January 5 th, 2016 Cho,
Status from Martin Josefsson Hardware level performance of Status from Martin Josefsson Hardware level performance of by Jesper Dangaard Brouer.
Sangwook Ma, Joongi Kim, Sue Moon School of Computing, KAIST
Disruptive Storage Workshop Lustre Hardware Primer
Balazs Voneki CERN/EP/LHCb Online group
SPIDAL Analytics Performance February 2017
Infiniband Architecture
Toward Effective and Fair RDMA Resource Sharing
Microsoft Virtual Academy
ECE 671 – Lecture 8 Network Adapters.
Cluster Computers.
Presentation transcript:

Use case of RDMA in Symantec storage software stack Om Prakash Agarwal Symantec

Agenda About Symantec Storage Software Stack Need for the faster interconnects Write operation without and with RDMA Implementation details: Write operation with RDMA Perf comparison (without and with RDMA) 2Symantec Proprietary: Use case of RDMA in Symantec storage stack

About Symantec Storage Software Stack Symantec Proprietary: Use case of RDMA in Symantec storage stack3 CVM/CFS Kernel Hardware CVM/CFS LLT NIC LLT NIC Applications User HBA Shared storage

Need for the faster interconnects Symantec Proprietary: Use case of RDMA in Symantec storage stack4 CVM/CFS Kernel Hardware CVM/CFS LLT NIC LLT NIC Applications User Local storage Shared Nothing Architecture

RDMA over InfiniBand or Ethernet Write operation without and with RDMA Symantec Proprietary: Use case of RDMA in Symantec storage stack5 CVM/CFS Kernel Hardware CVM/CFS LLT NIC LLT Buffer NIC RNIC Buffer Bufptr

Write operation with RDMA 6 Symantec Proprietary: Use case of RDMA in Symantec storage stack CVM/CFS LLT sender receiver Buffer advertisement: Allocate buffers and send Meta data to sender node. 4k RX map Actual allocation 4k TX map Destination buffer info 4k4k4k1k 8k8k Message 4k RX map Actual Allocation 4k4k4k2k2k 4k4k Message 4k TX map Destination buffer info 4k TX map Destination buffer info 4k RX map Actual Allocation 4k RX map Actual allocation 4k TX map Destination buffer info

Perf comparison (without and with RDMA) 7Symantec Proprietary: Use case of RDMA in Symantec storage stack 1)Throughput 2)Latency Setup configuration  Two node Linux (RHEL6.3) cluster  ConnectX3 VPI NIC card from Mellanox IB -56 Gbps RoCE-40 Gbps Eth-40 Gbps  3.3GHz Processor

LLT perf comparison : Throughput 8Symantec Proprietary: Use case of RDMA in Symantec storage stack 193% 125% Higher is better

LLT perf comparison : Latency 9Symantec Proprietary: Use case of RDMA in Symantec storage stack 62% Lower is better

#OFADevWorkshop Thank You