Presentation is loading. Please wait.

Presentation is loading. Please wait.

HPCS Lab. High Throughput, Low latency and Reliable Remote File Access Hiroki Ohtsuji and Osamu Tatebe University of Tsukuba, Japan / JST CREST.

Similar presentations


Presentation on theme: "HPCS Lab. High Throughput, Low latency and Reliable Remote File Access Hiroki Ohtsuji and Osamu Tatebe University of Tsukuba, Japan / JST CREST."— Presentation transcript:

1 HPCS Lab. High Throughput, Low latency and Reliable Remote File Access Hiroki Ohtsuji and Osamu Tatebe University of Tsukuba, Japan / JST CREST

2 HPCS Lab. Motivation and Background Data-intensive computing is a one of the most important issue in many areas Storage systems for Exa-byte (10 18 ) 2 Need a fast and reliable remote file access system

3 HPCS Lab. Motivation and Background(cont’d) Data sharing – Distributed file system – Clients access the data via Network Bottlenecks – Wide-area network Long latency – Storage cluster Overhead of network Fault tolerance – Suggestion: Congestion avoidance 3 Meta data server Storage node Client

4 HPCS Lab. Remote file access with RDMA Latency of Ethernet is at least 50 microseconds – Overhead of software – Protocol – Memory copy Flash memory based storage devices – 25μs latency (e.g. Fusion-io ioDrive), (HDD=5ms) Network becomes a bottleneck of the system 4

5 HPCS Lab. Usage of Infiniband IP over IB – Use the IP Protocol stack of operating systems Pros – Can use as a network adapter Cons – Inefficient SDP (Socket Direct Protocol) – Pros Easy to use – Specify the LD_PRELOAD Cons – Performance RDMA (Verbs API) – Pros Low-latency – Cons No compatibility with socket APIs Performance Implementation cost 5 better

6 HPCS Lab. Structure of OFED Verbs API SDP IPoIB From ©Mellanox document OFED : Drive and libraries for Infiniband 6

7 HPCS Lab. Remote file access with RDMA Architecture Memory Client Server Storage Infiniband Memory Client Application RDMA Infiniband FDR (54.3Gbps) Storage:Fusion-io ioDrive 7 RC Low overhead remote file access with Verbs API CPU bypass

8 HPCS Lab. Preliminary Evaluation: Throughput A client accesses the file on the file server via Infiniband w/ Verbs API – Sequential (Verbs API) 8

9 HPCS Lab. Preliminary Evaluation of IOPS Stride access from 2KB-64KB (seek 1MB) (Verbs API) Stride = 1MB 9

10 HPCS Lab. Congestion avoidance by using redundant data Concentration of access – There are hotspots (files) on the storage node Redundant data – Fault tolerance – Can be use to avoid congestion

11 HPCS Lab. Redundant data Basic structure 11 012 File1d 0 d 1 p 0 012 d 1 xor p 0 = d 0

12 HPCS Lab. RAID: connected with SCSI / SATA →connected with network 12 SCSI / SATA Network

13 HPCS Lab. Performance deterioration 13 012 File1d 0 d 1 p 0 File2 d 0 d 1 p 0 3 012 3 Storage nodes Clients 012 File1d 0 d 1 p 0 File2 p 0 d 0 d 1 3 012 3 Storage nodes Clients

14 HPCS Lab. Performance deterioration(cont’d) 14 Clients RAM Disk , 16GB , Sequential No congestion w/ congestion

15 HPCS Lab. Congestion avoidance 15 012 File1d 0 d 1 p 0 File2 d 0 d 1 p 0 3 012 3 Clients Storage nodes

16 HPCS Lab. Performance evaluation Compare the cases 16 復号あり w/ decode Congestion avoided +28% +15% Clients 復号あり Rebuild the striped data No congestion w/ congestion

17 HPCS Lab. Related work Stephen C. Simms et al, Wide Area Filesystem Performance using Lustre on the TeraGrid, 2007. Wu, J., Wyckoff, P. and Panda, D.: PVFS over InfiniBand: Design and Performance Evaluation Erasure Coding in Windows Azure Storage, USENIX ATC ‘12 – Shorten the latency by using redundant data HDFS RAID 17

18 HPCS Lab. Conclusion and Future work Remote file access with Infiniband RDMA Congestion avoidance Future work – How to detect the congestion – Writing of data(in progress) w/o performance degradation 18 012 3 Client Storage nodes


Download ppt "HPCS Lab. High Throughput, Low latency and Reliable Remote File Access Hiroki Ohtsuji and Osamu Tatebe University of Tsukuba, Japan / JST CREST."

Similar presentations


Ads by Google