Persistent Memory over Fabrics

Slides:



Advertisements
Similar presentations
Threads, SMP, and Microkernels
Advertisements

Priority Research Direction (I/O Models, Abstractions and Software) Key challenges What will you do to address the challenges? – Develop newer I/O models.
RDS and Oracle 10g RAC Update Paul Tsien, Oracle.
VIA and Its Extension To TCP/IP Network Yingping Lu Based on Paper “Queue Pair IP, …” by Philip Buonadonna.
The SNIA NVM Programming Model
IWARP Ethernet Key to Driving Ethernet into the Future Brian Hausauer Chief Architect NetEffect, Inc.
Basics of Operating Systems March 4, 2001 Adapted from Operating Systems Lecture Notes, Copyright 1997 Martin C. Rinard.
CERN openlab Open Day 10 June 2015 KL Yong Sergio Ruocco Data Center Technologies Division Speeding-up Large-Scale Storage with Non-Volatile Memory.
NVM Programming Model. 2 Emerging Persistent Memory Technologies Phase change memory Heat changes memory cells between crystalline and amorphous states.
New Direction Proposal: An OpenFabrics Framework for high-performance I/O apps OFA TAC, Key drivers: Sean Hefty, Paul Grun.
Cloud Computing for the Enterprise November 18th, This work is licensed under a Creative Commons.
Discussing an I/O Framework SC13 - Denver. #OFADevWorkshop 2 The OpenFabrics Alliance has recently undertaken an effort to review the dominant paradigm.
Challenges of Storage in an Elastic Infrastructure. May 9, 2014 Farid Yavari, Storage Solutions Architect and Technologist.
Low-Latency Datacenters John Ousterhout Platform Lab Retreat May 29, 2015.
1 Public DAFS Storage for High Performance Computing using MPI-I/O: Design and Experience Arkady Kanevsky & Peter Corbett Network Appliance Vijay Velusamy.
Next Generation Operating Systems Zeljko Susnjar, Cisco CTG June 2015.
1 Threads, SMP, and Microkernels Chapter Multithreading Operating system supports multiple threads of execution within a single process MS-DOS.
Business and Partnering Opportunities: “Windows Server 8” Continuous Availability Designing Systems for Continuous Availability and Scalability Session.
OpenFabrics Interface WG A brief introduction Paul Grun – co chair OFI WG Cray, Inc.
Local and Remote byte-addressable NVDIMM High-level Use Cases
Datacenter Fabric Workshop NFS over RDMA Boris Shpolyansky Mellanox Technologies Inc.
OFA OpenFabrics Interfaces Project Data Storage/Data Access subgroup November 2015 Summarizing NVM Usage Models for DS/DA.
© 2013 IBM Corporation 1 Title of presentation goes Elisa Martín Garijo IBM Distinguish Engineer and CTO for IBM Spain. Global Technology.
Under the Hood with NVMe over Fabrics
2015 Storage Developer Conference. © Intel Corporation. All Rights Reserved. RDMA with PMEM Software mechanisms for enabling access to remote persistent.
Progress in Standardization of RDMA technology Arkady Kanevsky, Ph.D Chair of DAT Collaborative.
OFA OpenFabrics Interfaces Project Data Storage/Data Access subgroup October 2015 Summarizing NVM Usage Models for DS/DA.
Advisor: Hung Shi-Hao Presenter: Chen Yu-Jen
OFA OpenFabrics Interfaces Project Data Storage/Data Access subgroup October 2015 Summarizing NVM Usage Models for DS/DA.
Clouding with Microsoft Azure
Extreme Scale Infrastructure
Persistent Memory (PM)
Virtualization for Cloud Computing
Agenda Hardware Virtualization Concepts
Data Center Infrastructure
Distributed Cache Technology in Cloud Computing and its Application in the GIS Software Wang Qi Zhu Yitong Peng Cheng
Application Access to Persistent Memory – The State of the Nation(s)!
Alternative system models
Using non-volatile memory (NVDIMM-N) as block storage in Windows Server 2016 Tobias Klima Program Manager.
Persistent Memory over Fabrics An Application-centric view
Briefing: Leverage HPE Storage Solutions in Windows/Hyper-V
HPE Persistent Memory Microsoft Ignite 2017
Microsoft Build /12/2018 5:05 AM Using non-volatile memory (NVDIMM-N) as byte-addressable storage in Windows Server 2016 Tobias Klima Program Manager.
RDMA Extensions for Persistency and Consistency
Chet Douglas – DCG NVMS SW Architecture
Software Defined Networking (SDN)
Understanding Wi-Fi Direct in Windows 8
Enabling the NVMe™ CMB and PMR Ecosystem
A Case for Virtualizing Persistent Memory
Rob Davis, Mellanox Ilker Cebeli, Samsung
PASTE: A Networking API for Non-Volatile Main Memory
Microsoft Ignite NZ October 2016 SKYCITY, Auckland.
Software Defined Networking (SDN)
ETSI Multi-access Edge Computing:
OpenFabrics Alliance An Update for SSSI
Objective Understand the concepts of modern operating systems by investigating the most popular operating system in the current and future market Provide.
More Exploration to Composable Infrastructure The Application and Analysis of Composable Memory D 阮渥豪
Lecture 4- Threads, SMP, and Microkernels
IBM Power Systems.
Application taxonomy & characterization
Software interoperability in the NGN Service layer
Open Source Activity Showcase Computational Storage SNIA SwordfishTM
Accelerating Applications with NVM Express™ Computational Storage 2019 NVMe™ Annual Members Meeting and Developer Day March 19, 2019 Prepared by Stephen.
NVMe.
Objective Understand the concepts of modern operating systems by investigating the most popular operating system in the current and future market Provide.
© 2016 Global Market Insights, Inc. USA. All Rights Reserved Ethernet Storage Market Size Growth During Forecast Period.
Utilizing the Network Edge
Parallel I/O for Distributed Applications (MPI-Conn-IO)
Factors Driving Enterprise NVMeTM Growth
Presentation transcript:

Persistent Memory over Fabrics Rob Davis, Mellanox Technologies Chet Douglas, Intel Paul Grun, Cray, Inc Tom Talpey, Microsoft Flash Memory Summit 2017 Santa Clara, CA

Agenda The Promise of Persistent Memory over Fabrics Driving new application models How to Get There An Emerging Ecosystem Encouraging Innovation Describing an API API Implementations Flush/commit is the way to go But how to optimize these operations? Flash Memory Summit 2017 Santa Clara, CA

Persistent Memory (PM) Persistent Memory (PM) Benefits Considerably faster than NAND Flash Performance accessible via PCIe or DDR interfaces Lower cost/bit than DRAM Significantly denser than DRAM PM Leading Edge PM Technologies

PM Needs High Availability (HA) NAND Read Latency ~100ns ~20,000ns Write Latency ~500ns ~50,000ns Flash Memory Summit 2017 Santa Clara, CA

Applications Enabled by PM Compute Disaggregation Scale out file/object systems In-Memory Database Machine learning Hyperconverged Infrastructure Flash Memory Summit 2017 Santa Clara, CA

Compute Disaggregation Flash Memory Summit 2017 Santa Clara, CA

Core Requirements for Networked PM PM is really fast Needs ultra low-latency networking PM has very high bandwidth Needs ultra efficient, transport offload, high bandwidth networks Remote accesses must not add significant latency PM networking requires: Predictability, Fairness, Very Low Packet Loss PM NAND Read Latency ~100ns ~20,000ns Write Latency ~500ns ~50,000ns Flash Memory Summit 2017 Santa Clara, CA

We Can Get There HGST live demo at FMS 2015 PM based Key-Value store over 100Gb/s fabric Key-Value Store fetch from non-volatile memory in 5usec Flash Memory Summit 2017 Santa Clara, CA

How to get there – the PM Ecosystem Memory Storage object, file, block… SNIA, OSVs. Language writers, App developers, middleware providers APIs libfabric, Verbs, NVMF, ND, NDK IBTA, OpenFabrics Alliance, NVMe Inc, Linux, Windows… Today’s Focus provider provider provider … network H/W network H/W network H/W IETF, IBTA, OS drivers, N/W Vendors… … This slide needs to convey the overlap between OFIWG as a developer of an API and (OFA+IBTA) as a developer of the verbs API. Persistent Memory Media, Form Factors Vendors, JEDEC…

No one argues with this… Persistent memory changes the way that applications store, access, communicate and share information local mem remote PM remote storage latency pS uS mS Data Access, Data Sharing… Data Storage What happens when apps can access shared information at near-memory speeds? How will apps evolve to take advantage of pools of global, shared, PM? What happens when apps evolve in the ways that they access, share and consume data? Persistent memory is (or soon will) change the landscape. We in the networking community are preparing for the ground shift. What should network architects do to prepare? Flash Memory Summit 2017 Santa Clara, CA

Focus on memory operations Storage - object, file, block - storage class mem Memory - persistent memory Think about accessing PM using memory paradigms, as opposed to storage paradigms… …even if the data at the far end is stored as a memory mapped file. storage client user or kernel app POSIX I/F load/store, memcopy… file system virtual switch DIMM DIMM NVDIMM NVDIMM NVM devices (SSDs…) provider NIC Taxonomy – let’s be clear on the different access methods. For this talk today, we’re not talking about NVMe/F – we’re talking about PM remote storage class memory, persistent memory Flash Memory Summit 2017 Santa Clara, CA

Simultaneous Equations to Solve user app, middleware, languages… Let consumers drive the requirements Consumer of network services Define the functions to be exported upward Fabric Independent API Provider Provider Provider Implement APIs for the fabrics of interest … Different Providers for different fabrics H/W device H/W device H/W device Flash Memory Summit 2017 Santa Clara, CA

Representative Use Cases global shared memory service NVDIMM app Shared PM Remote Access, Storage use cases NVDIMM app Replication High Availability use case We don’t yet know much about how PM is going to be deployed, but these two use cases are a good start Flash Memory Summit 2017 Santa Clara, CA Flash Memory Summit 2017 Santa Clara, CA

How they work High Availability Use Case Remote Shared Memory Use Case store, commit store, flush mem ctrl, library High Availability Use Case user Local NVDIMM write, commit Remote NVDIMM completion user user Remote Shared Memory Use Case get put Remote NVDIMM Remote memory use case. Remote persistent memory appears as a fabric-attached service which is shared by number of clients. completion notice Flash Memory Summit 2017 Santa Clara, CA

Shared PM - A Closer Look Remote NVDIMM valid user put commit get notice persistent node cntlr API Requirement: Distinguish between data which is globally valid, and data which has been made persistent Remote memory use case. Remote persistent memory appears as a fabric-attached service which is shared by number of clients. Flash Memory Summit 2017 Santa Clara, CA

The Beginnings of a List of API Reqmts user user user Applications Requirements – a starting point Ability to force flushes of local caches to remote persistent memory Perform non-cached writes to remote to persistent memory Identify the persistence characteristics of a remote memory device Distinguish between data which is globally valid and data which is remotely persistent Remote memory use case. Remote persistent memory appears as a fabric-attached service which is shared by number of clients. Flash Memory Summit 2017 Santa Clara, CA

Basic API Features (example) provider, fabric adapter provider, fabric adapter persistence domain Consumer PUT w/ Commit Plenty of work to be done to define the semantics of the API Lots more to figure out how to implement the semantics over a given wire Write ack completion Commit Commit Completion completion From those two use cases, there are some things we already know we’re going to need to do. Flash Memory Summit 2017 Santa Clara, CA API Magic Happens Flash Memory Summit 2017 Santa Clara, CA

Fabric Implementations consumer provider, fabric adapter provider, fabric adapter remote mem subsystem remote pmem PUT Write Utilize FLUSH to force previous writes on the connection to be made durable before the flush is completed PUT w/ Commit PCIe Wrt Flush Asynch Mem Wrt NP Write PCIe Flushing Rd Flushing Rd Rsp Flush Response NP PCIe Wrt Completion Synch Mem Wrt NP PCIe Write Rsp API fabric specific platform specific Flash Memory Summit 2017 Santa Clara, CA

Alternate Fabric Implementation consumer provider, fabric adapter provider, fabric adapter remote mem subsystem remote pmem An alternative approach: combine the FLUSH and following WRITE into a WRITE w/ COMMIT PUT Write PUT w/ Commit PCIe Wrt Write w/ Commit Asynch Mem Wrt PCIe Flushing Rd Flushing Rd Rsp NP PCIe Wrt Synch Mem Wrt NP PCIe Write Rsp Flush Response Completion API fabric specific platform specific Flash Memory Summit 2017 Santa Clara, CA

Fabric Implementations Challenge to the industry is to extend the existing fabric protocols to provide: Efficient, low overhead PMEM semantics Simple provider fabric adapter implementation Flexible, extensible across common fabric types Flash Memory Summit 2017 Santa Clara, CA

Q & A, Discussion Rob Davis, Mellanox Technologies Chet Douglas, Intel Paul Grun, Cray, Inc Tom Talpey, Microsoft

Thanks!