10
Zivan Ori Co-founder and CEO May 16 th 2018 Shared NVMe for High Performance Spectrum Scale Clusters

High Availability for Centralized NVMefiles.gpfsug.org/presentations/2018/USA/E8 for GPFS UG...•Extract all performance from NVMe SSDs •Use off-the-shelf hardware •Scalable in

  • Upload
    others

  • View
    12

  • Download
    0

Embed Size (px)

Citation preview

Page 1: High Availability for Centralized NVMefiles.gpfsug.org/presentations/2018/USA/E8 for GPFS UG...•Extract all performance from NVMe SSDs •Use off-the-shelf hardware •Scalable in

Zivan Ori

Co-founder and CEO

May 16th 2018

Shared NVMe for High Performance Spectrum Scale Clusters

Page 2: High Availability for Centralized NVMefiles.gpfsug.org/presentations/2018/USA/E8 for GPFS UG...•Extract all performance from NVMe SSDs •Use off-the-shelf hardware •Scalable in

The E8 Storage Difference

• Extract all performance from NVMe SSDs

• Use off-the-shelf hardware

• Scalable in multiple dimensions• Scale hosts for more computing power• Scale storage for higher capacity

• Simple, centralized management

• High reliability and availability

©2018 E8 Storage, Proprietary and Confidential2

A new architecture built specifically for high performance NVMe™

Page 3: High Availability for Centralized NVMefiles.gpfsug.org/presentations/2018/USA/E8 for GPFS UG...•Extract all performance from NVMe SSDs •Use off-the-shelf hardware •Scalable in

What is NVMe™? (Non-Volatile Memory Express)

• High performance, low latency• Efficient protocol with lower stack overhead• Exponentially more queues / commands than SAS• Parallel processing for SSDs vs serial for HDDs

• Support for fabrics (NVMe-oF™)• Originally designed for PCIe (internal to servers)• Expands support for other transport media

• RDMA Based: RoCE, iWARP, Infiniband• Non-RMDA: FC, TCP

• Maintains NVMe protocol end to end

©2018 E8 Storage, Proprietary and Confidential3

Communication protocol designed specifically for flash storage

Connections I/O Queues Commands per Queue

SAS 2 1 256

NVMe 128 65,535 64,000

Page 4: High Availability for Centralized NVMefiles.gpfsug.org/presentations/2018/USA/E8 for GPFS UG...•Extract all performance from NVMe SSDs •Use off-the-shelf hardware •Scalable in

Architected for High Performance NVMe

• Centralized control operations• E8 Controllers manage all volumes, RAID config• Monitoring, management functions

• Distributed data operations• Built for IB or RDMA over Converged Ethernet (RoCE)• E8 Agents offload 90% of data path operations• Auto-discover provisioned volumes

• Leveraging the performance of RDMA• Enables direct access to flash as memory via network• Bypasses CPU / memory for fast reads

©2018 E8 Storage, Proprietary and Confidential4

Separation of data and control; no controller bottleneck

Page 5: High Availability for Centralized NVMefiles.gpfsug.org/presentations/2018/USA/E8 for GPFS UG...•Extract all performance from NVMe SSDs •Use off-the-shelf hardware •Scalable in

Designed for Availability and Reliability

Hardware

• High-availability off-the-shelf appliances• Redundant controllers with auto-failover• Redundant power, cooling• All parts hot-swappable

Software

• Host agents operate independently• Failure of one agent (or more) does not affect other agents• Access to shared storage is not impacted

• RAID-6 data protection

©2018 E8 Storage, Proprietary and Confidential5

No single point of failure anywhere in the architecture

Host Servers with E8 Host Agents

Page 6: High Availability for Centralized NVMefiles.gpfsug.org/presentations/2018/USA/E8 for GPFS UG...•Extract all performance from NVMe SSDs •Use off-the-shelf hardware •Scalable in

Using E8 with IBM Spectrum Scale

• Standalone pool

• Local Read Only Cache (LROC)

• High Availability Write Cache (HAWC)

• Metadata repository

©2018 E8 Storage, Proprietary and Confidential6

Multiple Deployment Options

Page 7: High Availability for Centralized NVMefiles.gpfsug.org/presentations/2018/USA/E8 for GPFS UG...•Extract all performance from NVMe SSDs •Use off-the-shelf hardware •Scalable in

Deployment – All Clients Connected Directly to Storage

• Scales to over 100 clients

• Direct access to clients, lowest latency

• Standalone pool• Shared LUNs

• LROC• Non-shared LUNs

• HAWC• Non-shared LUNs model

©2018 E8 Storage, Proprietary and Confidential7

IB/RoCE

SSDSSDSSDSSDSSDSSD

E8-D24 (Dual-port NVMe)E8 MDS

RAID-6

E8 MDS

GPFS Client

E8 Agent

GPFS Client

E8 Agent

GPFS Client

E8 Agent

Page 8: High Availability for Centralized NVMefiles.gpfsug.org/presentations/2018/USA/E8 for GPFS UG...•Extract all performance from NVMe SSDs •Use off-the-shelf hardware •Scalable in

Deployment – Some Nodes Connected via NSD to Storage

• Scalable to larger configurations• Can mix connectivity depending on requirements

• Standalone pool• Shared LUNs

• LROC• Non-shared LUNs (direct connect clients only)

• HAWC• Shared LUNs model

©2018 E8 Storage, Proprietary and Confidential8

IB/ RoCE

SSDSSDSSDSSDSSDSSD

E8-D24 (Dual-port NVMe)E8 MDS

RAID-6

E8 MDS

NSDE8 Agent

NSDE8 Agent

GPFS ClientGPFS ClientGPFS ClientE8 Agent

Page 9: High Availability for Centralized NVMefiles.gpfsug.org/presentations/2018/USA/E8 for GPFS UG...•Extract all performance from NVMe SSDs •Use off-the-shelf hardware •Scalable in

Performance - E8 Storage and GPFS

©2018 E8 Storage, Proprietary and Confidential9From IBM Research Performance Benchmarks of GPFS over E8

Page 10: High Availability for Centralized NVMefiles.gpfsug.org/presentations/2018/USA/E8 for GPFS UG...•Extract all performance from NVMe SSDs •Use off-the-shelf hardware •Scalable in

SPEC SFS®2014_swbuild Performance*

• #1 performance for GPFS!• 2.5x more builds vs IBM all flash array• 8x lower latency vs previous record

• The only sub-millisecond ORT!• 0.69ms overall response time (ORT)

• More performance, less hardware

©2018 E8 Storage, Proprietary and Confidential10

8x lower latency!

* As of SPEC SFS®2014_swbuild results published January 2018. SPEC SFS2014 is the industry standard benchmark for file storage performance. See all published results at https://www.spec.org/sfs2014/results/

E8 Storage 24 NVMe SSDs 2U

NetApp FAS8200 144 HDDs, 8 NVMe SSDs 20U