Research Introduction

Slides:



Advertisements
Similar presentations
RAID Redundant Arrays of Independent Disks Courtesy of Satya, Fall 99.
Advertisements

CSCE430/830 Computer Architecture
SDN + Storage.
Implementing A Simple Storage Case Consider a simple case for distributed storage – I want to back up files from machine A on machine B Avoids many tricky.
Distributed Multimedia Systems
Serverless Network File Systems. Network File Systems Allow sharing among independent file systems in a transparent manner Mounting a remote directory.
Lecture 36: Chapter 6 Today’s topic –RAID 1. RAID Redundant Array of Inexpensive (Independent) Disks –Use multiple smaller disks (c.f. one large disk)
Distributed Storage March 12, Distributed Storage What is Distributed Storage?  Simple answer: Storage that can be shared throughout a network.
Cache Coherent Distributed Shared Memory. Motivations Small processor count –SMP machines –Single shared memory with multiple processors interconnected.
Network Coding for Large Scale Content Distribution Christos Gkantsidis Georgia Institute of Technology Pablo Rodriguez Microsoft Research IEEE INFOCOM.
Efficient Support for Interactive Browsing Operations in Clustered CBR Video Servers IEEE Transactions on Multimedia, Vol. 4, No.1, March 2002 Min-You.
RAID-x: A New Distributed Disk Array for I/O-Centric Cluster Computing Kai Hwang, Hai Jin, and Roy Ho.
Storage System: RAID Questions answered in this lecture: What is RAID? How does one trade-off between: performance, capacity, and reliability? What is.
Copyright © 2012 Cleversafe, Inc. All rights reserved. 1 Combining the Power of Hadoop with Object-Based Dispersed Storage.
Take An Internal Look at Hadoop Hairong Kuang Grid Team, Yahoo! Inc
Storage management and caching in PAST PRESENTED BY BASKAR RETHINASABAPATHI 1.
Presented by: Alvaro Llanos E.  Motivation and Overview  Frangipani Architecture overview  Similar DFS  PETAL: Distributed virtual disks ◦ Overview.
Research on cloud computing application in the peer-to-peer based video-on-demand systems Speaker : 吳靖緯 MA0G rd International Workshop.
Chapter 6 RAID. Chapter 6 — Storage and Other I/O Topics — 2 RAID Redundant Array of Inexpensive (Independent) Disks Use multiple smaller disks (c.f.
DISKS IS421. DISK  A disk consists of Read/write head, and arm  A platter is divided into Tracks and sector  The R/W heads can R/W at the same time.
Redundant Array of Independent Disks
RAID: High-Performance, Reliable Secondary Storage Mei Qing & Chaoxia Liao Nov. 20, 2003.
CSC 456 Operating Systems Seminar Presentation (11/13/2012) Leon Weingard, Liang Xin The Google File System.
Parity Logging O vercoming the Small Write Problem in Redundant Disk Arrays Daniel Stodolsky Garth Gibson Mark Holland.
1 Configurable Security for Scavenged Storage Systems NetSysLab The University of British Columbia Abdullah Gharaibeh with: Samer Al-Kiswany, Matei Ripeanu.
Hadoop Hardware Infrastructure considerations ©2013 OpalSoft Big Data.
Distributed File System By Manshu Zhang. Outline Basic Concepts Current project Hadoop Distributed File System Future work Reference.
Swapping to Remote Memory over InfiniBand: An Approach using a High Performance Network Block Device Shuang LiangRanjit NoronhaDhabaleswar K. Panda IEEE.
Data in the Cloud – I Parallel Databases The Google File System Parallel File Systems.
Fast Crash Recovery in RAMCloud. Motivation The role of DRAM has been increasing – Facebook used 150TB of DRAM For 200TB of disk storage However, there.
INTRODUCTION TO DBS Database: a collection of data describing the activities of one or more related organizations DBMS: software designed to assist in.
Active Storage Processing in Parallel File Systems Jarek Nieplocha Evan Felix Juan Piernas-Canovas SDM CENTER.
HADOOP DISTRIBUTED FILE SYSTEM HDFS Reliability Based on “The Hadoop Distributed File System” K. Shvachko et al., MSST 2010 Michael Tsitrin 26/05/13.
Distributed File Systems 11.2Process SaiRaj Bharath Yalamanchili.
6.894: Distributed Operating System Engineering Lecturers: Frans Kaashoek Robert Morris
History & Motivations –RDBMS History & Motivations (cont’d) … … Concurrent Access Handling Failures Shared Data User.
GPFS: A Shared-Disk File System for Large Computing Clusters Frank Schmuck & Roger Haskin IBM Almaden Research Center.
LIOProf: Exposing Lustre File System Behavior for I/O Middleware
Enhanced Availability With RAID CC5493/7493. RAID Redundant Array of Independent Disks RAID is implemented to improve: –IO throughput (speed) and –Availability.
Seminar On Rain Technology
Taeho Kgil, Trevor Mudge Advanced Computer Architecture Laboratory The University of Michigan Ann Arbor, USA CASES’06.
RAID TECHNOLOGY RASHMI ACHARYA CSE(A) RG NO
Presenter: Yue Zhu, Linghan Zhang A Novel Approach to Improving the Efficiency of Storing and Accessing Small Files on Hadoop: a Case Study by PowerPoint.
DISTRIBUTED FILE SYSTEM- ENHANCEMENT AND FURTHER DEVELOPMENT BY:- PALLAWI(10BIT0033)
CWG12: Filesystems for TDS U. FUCHS / CERN. O2 Data Flow Schema FLPs Data Network 1500 EPNs ~3o PB, 10 9 files, ~150 GBps Data Management facilities,
Lustre File System chris. Outlines  What is lustre  How does it works  Features  Performance.
Parallel Virtual File System (PVFS) a.k.a. OrangeFS
rain technology (redundant array of independent nodes)
The demonstration of Lustre in EAST data system
Distributed File Systems
Reddy Mainampati Udit Parikh Alex Kardomateas
RAID Redundant Arrays of Independent Disks
The University of Adelaide, School of Computer Science
The University of Adelaide, School of Computer Science
Introduction to Networks
Storage Virtualization
Understanding System Characteristics of Online Erasure Coding on Scalable, Distributed and Large-Scale SSD Array Systems Sungjoon Koh, Jie Zhang, Miryeong.
Replication Middleware for Cloud Based Storage Service
CMSC 611: Advanced Computer Architecture
RAID RAID Mukesh N Tekwani
The University of Adelaide, School of Computer Science
Lock Ahead: Shared File Performance Improvements
UNIT IV RAID.
The University of Adelaide, School of Computer Science
Lecture 17 Multiprocessors and Thread-Level Parallelism
DISTRIBUTED SYSTEMS Principles and Paradigms Second Edition ANDREW S
RAID RAID Mukesh N Tekwani April 23, 2019
Lecture 17 Multiprocessors and Thread-Level Parallelism
The University of Adelaide, School of Computer Science
Lecture 17 Multiprocessors and Thread-Level Parallelism
Presentation transcript:

Research Introduction Taeuk Kim Laboratory for Advanced System Software Sogang University 1

Contents Background Introduction of Parallel File System LADS FT-LADS What I will do

Background Large amount of Scientific data Satellite image Nuclear data Big data era Cooperative research and analyze => transferring large scale data ↑

Background Network bandwidth scale ↑ 10G ethernet, Infiniband on private network Relatively file system bandwidth ↓ Bottleneck on storage server ↑ A solution for bottleneck on file system is PFS (Parallel File System)

Parallel File System I/O processing consumes time ↑ Idea for distributing I/O workload Managing file by chunking -> object Maintaining and managing multiple storage server and parallely processing file objects with multi-threading

Parallel File System - Lustre MDS/MGS : Metadata Server, Management Server MDT : Metadata Target OSS : Object Storage Server OST : Object Storage Target

LADS There are still problems -> I/O contention on PFS Many clients use same PFS -> QoS ↓ Disk failure -> rebuilding disk from RAID Unbalanced I/O load => performance ↓ If we can avoid I/O congestion, the performance will be better LADS (Layout-Aware Data Scheduling) is motivated from above -> (Data transfer) application level solution It helps you to avoid I/O contention in OSS/OST

LADS Avoiding congestions of cluster storage OSS’s or OST’s data arrival rate > data service rate -> overflow occurs (In Lustre)

LADS In PFS, we can know the layout of objects -> Possible to distributing objects well and avoiding congeston With knowing the layout, it’s possible to schedule I/O not to be congested

LADS Transfer protocol Step1 – (src) NEW_FILE to comm’s Q, NEW_BLOCK to OST’s Q Step 2 – (sink) FILE_ID to comm’s Q Step 3 – (src) wake up N I/O threads and read data from OST and send that to sink Step 4 – (sink) receive req and read from src’s RMA buffer Step 5 – (sink) Store data objects in sink’s OST (layout-aware) Step 6 – (src) If BLOCK_DONE, release RMA buffer and goto step 3 Scheduling Layout-aware : knowing about which OST stores which data objects Congestion-aware : Defualt is RR, but when a thread is accessing an OST, another thread pass that OST and move on to next Object caching When RMA buffer is full, it uses NVM buffer

FT-LADS Disk failure situation (of facebook) New data : 500TB~900TB/day Failure : lose 100k blocks/day Reconstruction traffic : 180TB/day Recovery of disk failure is important Currently LADS doesn’t have fault tolerance mechanism

FT-LADS FT-LADS is a poster accepted by IEEE CLUSTER’16 It suggests fault tolerance mechanism for LADS File logging : remaining sent block’s log file per every files Char type logging Encoding type logging Int type logging Binary type logging : block number as a binary value Bit binary logging : n-th bit position as n-th block Transaction logging : remaining multiple files’ logs in one log file Universal logging : remaining all files’ logs in one log file

What I will do is To Support FT-LADS improvement Implementing some issues, etc To Contribute with a new idea Thinking about core-aware in NUMA

Q&A Thank you