erez cohen & aviram bar haim, mellanox - enhancing your openstack cloud with advanced network...

55
© 2014 Mellanox Technologies 1 - Mellanox Confidential - Erez Cohen, Sr. Director, Cloud Program Aviram Bar-Haim, Cloud Solutions Engineer, SW CloudX: The most efficient and scalable cloud solution OpenStack Berlin, June 2015

Upload: openstack-israel

Post on 14-Aug-2015

190 views

Category:

Technology


2 download

TRANSCRIPT

Page 1: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 1- Mellanox Confidential -

Erez Cohen, Sr. Director, Cloud ProgramAviram Bar-Haim, Cloud Solutions Engineer, SW

CloudX: The most efficient and scalable cloud solutionOpenStack Berlin, June 2015

Page 2: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 2- Mellanox Confidential -

We Live in a World of Data

Data Needs to be Accessible Always and in Real-Time

Exponential Data Growth – The Best Platforms Are Needed

More DataMore ApplicationsMore Devices

Page 3: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 3- Mellanox Confidential -

Exponential Data Growth – Few Facts

▪More than a 1.44B active Facebook users ~ 1/5 of the earth

population.

▪Every second 1157 people start watching YouTube videos

(100,000,000 videos a day).

▪People currently create 1.2 trillion GB of data a year (Equals to

10 iPODs for each person on earth).

▪Average US household has 5.7 internet connected devices (More

than human bodies).

Page 4: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 4- Mellanox Confidential -

Exponential Data Growth – Best Interconnect Required

0.8 Zettabyte

200935 Zettabyte

2020

44X

Source: IDC

Page 5: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 5- Mellanox Confidential -

The Future Depends on the Fastest and Efficient Interconnects

10Gb/s 40/56/100Gb/s1Gb/s

Page 6: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 6- Mellanox Confidential -

Exponential Data Growth – Conclusion

▪We can infer two main conclusions from this data growth:

1. Fast network interconnect is critical for efficient and scalable cloud

2. Efficient storage solutions are necessary to handle all future data

Page 7: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 7- Mellanox Confidential -

Agenda

▪Interconnect technology evolution in virtualized environments

• User-space networking overview (TCP/UDP -> RDMA)

• Virtual interfaces overview (emulation -> paravirt -> SRIOV)

- Using SR-IOV in Openstack Neutron project

• Storage protocols and performance (FC -> iSCSI -> iSER)

- Using iSER in Openstack Cinder project

• Overlay network scalability and performance (VLAN -> VXLAN)

- Using VXLAN offloading in Openstack Neutron project

Page 8: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 8- Mellanox Confidential -

Agenda

▪Interconnect technology evolution in virtualized environments

• User-space networking overview (TCP/UDP -> RDMA)

• Virtual interfaces overview (emulation -> paravirt -> SRIOV)

- Using SR-IOV in Openstack Neutron project

• Storage protocols and performance (FC -> iSCSI -> iSER)

- Using iSER in Openstack Cinder project

• Overlay network scalability and performance (VLAN -> VXLAN)

- Using VXLAN offloading in Openstack Neutron project

Page 9: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 9- Mellanox Confidential -

TCP/UDP -> Remote Direct Memory Access (RDMA)

RDMA over IB / Ethernet

Page 10: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 10- Mellanox Confidential -

Application / Middleware

Thread

User-space Networking using RDMA - Overview

▪HW managed by kernel• Applications create resource via system calls, e.g.- Queues (descriptor rings)- Registering memory buffers

▪Data-path bypasses kernel• Shared queues between application and HW• HW accesses registered buffers directly• Direct signaling mechanism (“doorbells”)• Direct completion detection- In memory polling- Can also register for event (interrupt)

▪Multiple HW resources• No need for locking if resources are accessed by a

single thread

▪Efficient• Asynchronous progress• Zero copy

HW kernel driver

User

Kernel

NIC

RegisteredMemory

ThreadS

end

Q

Recv Q

Co

mp

Q

HW user-space driver

Access library

Page 11: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 11- Mellanox Confidential -

User-space Networking using RDMA – Packet Interfaces

▪Verbs Raw Ethernet Queues as an example▪Basic data path operations

• Post packet buffers to be sent out• Post buffers to receive packets• Poll for completions

▪Checksum offloads for TCP/UDP• Insert checksum on TX• Validate checksum on RX

▪VLAN insertion/stripping▪Receive-side scaling (RSS)

• Distribute incoming packets into multiple queues• Distribution is semi-random (hash based)

▪Flow steering• Deterministic steering of specific flows to specific RQs

▪Deliver very high packet rate to the application• E.g., 25Mpps for 64b packets

Application / Middleware

Thread

HW kernel driver

User

Kernel

NIC

RegisteredMemory

ThreadS

end

Q

Recv Q

Co

mp

Q

HW user-space driver

Access library

Page 12: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 12- Mellanox Confidential -

User-Space Networking – RDMA Infinband / Ethernet Interfaces

▪Pass messages instead of packets• Up to 2GB in size

▪Semantics• Channel (Message passing)- Requestor provides source buffer- Responder provides receive buffer

• Remote Direct Memory Access (RDMA)- Requestor provides both source and target buffers- Both RDMA-read and –write are supported

▪Advanced RDMA operations• Atomics- Compare & swap- Fetch & add- Multi-field

• Data integrity

▪Extreme performance• 700ns one-way latency between applications• 40GE BW at negligible CPU utilization• Packet rate > 35Mpps

Send Queue

Receive Queue

Send Queue

Page 13: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 13- Mellanox Confidential -

Agenda

▪Interconnect technology evolution in virtualized environments

• User-space networking overview (TCP/UDP -> RDMA)

• Virtual interfaces overview (emulation -> paravirt -> SRIOV)

- Using SR-IOV in Openstack Neutron project

• Storage protocols and performance (FC -> iSCSI -> iSER)

- Using iSER in Openstack Cinder project

• Overlay network scalability and performance (VLAN -> VXLAN)

- Using VXLAN offloading in Openstack Neutron project

Page 14: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 14- Mellanox Confidential -

Host

Virtual Interfaces – Device Emulation

▪Host emulates a complete HW device• E.g., Intel e1000 NIC

▪Guest runs unmodified driver

▪Pros• No need to install special drivers in guests• Transparent migration• Unlimited virtual interfaces

▪Cons• Slow• Emulation exists only for very simple devices• High overhead

Qemu process

VM

e1000 driver

e1000 emulator

SW Switch

macvtapnetdev

physnetdev

User

Kernel

User

Kernel

NIC

Page 15: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 15- Mellanox Confidential -

Virtual Interfaces – Para-Virtualization

▪Host exposes a virtual “SW-friendly” device• E.g., virtio-net

▪VM runs special device driver

▪Host emulates device back-end

▪Pros• Decent performance• Transparent migration• Unlimited virtual interfaces

▪Cons• Simple devices only

Host Qemu process

VM

virtio-net

virt-io emulator

SW Switch

macvtapnetdev

physnetdev

User

Kernel

User

Kernel

NIC

Page 16: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 16- Mellanox Confidential -

Virtual Interfaces – Accelerated Para-Virtualization

▪Same para-virtual control interface

▪Fast path offloaded to host kernel• vhost_net

Host Qemu process

VM

virtio-net

virt-io control

SW switch

macvtapnetdev

physnetdev

User

Kernel

User

Kernel

vhost-net

NIC

Page 17: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 17- Mellanox Confidential -

Virtual Interfaces – Physical Device Pass-Through

▪Hosts grants guest direct access to a physical device• Security and isolation still maintained- PCI configuration space is virtualized- IOMMU governs DMA access

▪VM runs standard device driver

▪Pros• Near-native performance• VMs can use any device that is passed to them

▪Cons• No transparent migration• Very limited scalability (physical devices are not

shared)

Host Qemu process

VM

HW driver

User

Kernel

User

Kernel

HW driver

NIC 1 NIC 2

Switch

Page 18: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 18- Mellanox Confidential -

NIC

Virtual Interfaces – Virtual Device Pass-Through (SR-IOV)

▪Single Root I/O Virtualization (SR-IOV)

▪Hosts grants guest direct access to a virtual device• Security and isolation still maintained- PCI configuration space is virtualized- IOMMU governs DMA access

▪VM runs device driver for virtual function

▪Pros• Near-native performance• High scalability (128-256 VFs)

▪Cons• No transparent migration

Host Qemu process

User

Kernel

PF driver

Phyiscal Function

Virtual Function

Embedded switch

Switch

VM

VF driver

User

Kernel

Page 19: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 19- Mellanox Confidential -

Agenda

▪Interconnect technology evolution in virtualized environments

• User-space networking overview (TCP/UDP -> RDMA)

• Virtual interfaces overview (emulation -> paravirt -> SRIOV)

- Using SR-IOV in Openstack Neutron project

• Storage protocols and performance (FC -> iSCSI -> iSER)

- Using iSER in Openstack Cinder project

• Overlay network scalability and performance (VLAN -> VXLAN)

- Using VXLAN offloading in Openstack Neutron project

Page 20: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 20- Mellanox Confidential -

▪The Modular Layer 2 (ML2) Plugin is a framework allowing OpenStack Neutron to simultaneously utilize the variety of layer 2 networking technologies found in complex real-world data centers

▪Mellanox ML2 Mechanism Driver• Adds support for SRIOV in a transparent manner• Upon port creation and binding, the driver will allocate VF and connect it to the VM• The driver will also configure the internal embedded switch

▪The Mellanox Mechanism Driver can also support InfiniBand network for OpenStack in a transparent manner • Convert MAC/VLAN to GUID/PKEY

Mellanox Neutron Plugin – Icehouse Release

Page 21: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 21- Mellanox Confidential -

Architecture Block Diagram – ML2 based neutron

OpenStackCloud Manager

OpenvSwitch Agent

ConnectX

Controller

Compute

Mellanox neutron Agent

RPC

eswitchDOVS

ConnectX

Mellanox neutron Agent

eswitchD

RPC

Neutron

ML2

MLNXOVS

Compute

Page 22: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 22- Mellanox Confidential -

SR-IOV in Openstack Juno/Kilo

▪Starting from Juno, a standard SRIOV NIC switch neutron mechanism driver is supported

upstream

• Offers a standard way to create a direct SRIOV neutron port from a virtual function

▪Two ways to configure SR-IOV in OpenStack environment:

• SRIOVNICSwitch Mechanism Driver: https://wiki.openstack.org/wiki/SR-IOV-Passthrough-For-Networking

• MLNX mechanism Driver (support Eth & IB): https://wiki.openstack.org/wiki/Mellanox-Neutron-ML2

Page 23: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 23- Mellanox Confidential -

OpenStack SRIOV demo with 100G or similar

Page 24: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 24- Mellanox Confidential -

Agenda

▪Interconnect technology evolution in virtualized environments

• User-space networking overview (TCP/UDP -> RDMA)

• Virtual interfaces overview (emulation -> paravirt -> SRIOV)

- Using SR-IOV in Openstack Neutron project

• Storage protocols and performance (FC -> iSCSI -> iSER)

- Using iSER in Openstack Cinder project

• Overlay network scalability and performance (VLAN -> VXLAN)

- Using VXLAN offloading in Openstack Neutron project

Page 25: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 25- Mellanox Confidential -

Storage Area Networks – Fibre Channel (1997) VS. iSCSI (2000)

▪Fibre Channel provides low latency and high availability. But Fibre Channel can be expensive and complex to manage.

▪iSCSI (IP based SCSI) uses Ethernet as its underlying communications fabric, so it is much less expensive than Fibre Channel.

▪As Ethernet continues to advance, iSCSI advances right along with it.

▪In 2007, Fibre Channel support for Ethernet networks (FCoE), proposed a more standard way to use Fibre Channel (with reduced performance..).

Page 26: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 26- Mellanox Confidential -

What is iSER?

▪iSER – iSCSI Extensions for RDMA▪Maps the iSCSI protocol over RDMA fabrics ▪Leverages iSCSI management infrastructure▪The transport layer iSER and/or iSCSI/TCP are transparent

to the user. Just need a simple configurable to decideEthernet (RoCE) / InfiniBand

SCSI Application Layer

iSCSI

iSER

SCSI Layer

Verbs APINet IF (Ethernet/IPoIB)

iSCSI

TCPiSNS/SLP

IB HCARNICNIC

Management

TOE iSER

Datamover API

Page 27: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 27- Mellanox Confidential -

iSER Protocol Overview (Read)

▪ SCSI Reads • Initiator Send Command PDU (Protocol data unit) to Target • Target return data using RDMA Write• Target send Response PDU back when completed transaction• Initiator receives Response and complete SCSI operation

iSC

SI

Init

iato

r

iSE

R

HC

A

HC

A

iSE

R T

arge

t

Tar

get

Sto

rage

Send_Control (SCSI Read Cmd)

RDMA Write for Data

Send_Control + Buffer advertisement

Control_Notify

Data_Put (Data-In PDU) for Read

Control_NotifySend_Control (SCSI Response)

Send_Control (status, sense data)

Page 28: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 28- Mellanox Confidential -

iSER Protocol Overview (Write)

▪SCSI Writes

• Send Command PDU (optionally with Immediate Data to improve latency)

• Map R2T to RDMA Read operation (retrieve data)

• Target send Response PDU back when completed transaction

iSC

SI

Init

iato

r

iSE

R

HC

A

HC

A

iSE

R T

arge

t

Tar

get

Sto

rage

Send_Control (SCSI Write Cmd)

RDMA Read for Data (Optional)

Send_Control + Buffer advertisement Control_Notify

(SCSI Command)

Get_Data (R2T PDU)

Control_NotifySend_Control (SCSI Response)

Send_Control (status, sense data)

Page 29: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 29- Mellanox Confidential -

Performance

Page 30: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 30- Mellanox Confidential -

Conclusion: iSER Is the FASTEST Networked Block Protocol

▪What it is: iSCSI With RDMA Transport• Runs over Ethernet or InfiniBand at 10, 40, 56Gb/s &100Gb/s

• Works with all applications that support SCSI/iSCSI

▪Benefits• Most bandwidth, Highest IOPs, Lowest Latency, Lowest CPU utilization

• iSCSI storage features, management and tools (security, HA, discovery...)

• Faster than iSCSI, FC, FCoE; Easier to manage than SRP

▪Ideal For• SSD / nvmE

• Latency-sensitive workloads; Small, random I/O

• Databases, Virtualization (especially VDI)

Page 31: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 31- Mellanox Confidential -

Need FibreChannel Protocol

Need Highest Performance

Want Ethernet

Protocol / Transport Comparison

InfiniBand FCoE Fibre ChannelRoCE

Transport InfiniBand RoCE Std. Ethernet FCoE FC

Speed Up to 100 Gb/s Up to 100 Gb/s Up to 100 Gb/s Up to 100 Gb/s 8/16 Gb/s

RDMA Yes Yes No No No

Routable Yes Yes Yes No No

SMB Direct

iSER

SMB, NFS, or iSCSI on TCP

FCP

NFSoRDMA

Page 32: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 32- Mellanox Confidential -

Storage Protocol Comparison

Storage Protocol

iSER FC FCoE iSCSI TCP

SMB Direct NFSRDMA

NVMe over fabrics

Access Block Block Block Block File File Block

Transports RoCE, IB FC DCBx Ethernet Ethernet RoCE, IB RoCE, IB RoCE, IB

RDMA Yes No No No Yes1 Yes2 Yes3

1. SMB Direct is an in-box option for SMB 3.0 when the network supports RDMA with InfiniBand, RoCE, or iWARP2. NFS over RDMA has limited ecosystem support but is improving rapidly. It supports InfiniBand and is expected to support RoCE and iWARP3. NVMe over Fabrics is a proposed standard (expected to be finalized in 2015) that will support InfiniBand, RoCE, and iWARP

BlockFile

RDMA

FC

SMB Direct

NFSoRDMA

SMB/CIFS

NFS iSER

NVMe over Fabrics

iSCSI over TCP

FCoE

Page 33: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 33- Mellanox Confidential -

Agenda

▪Interconnect technology evolution in virtualized environments

• User-space networking overview (TCP/UDP -> RDMA)

• Virtual interfaces overview (emulation -> paravirt -> SRIOV)

- Using SR-IOV in Openstack Neutron project

• Storage protocols and performance (FC -> iSCSI -> iSER)

- Using iSER in Openstack Cinder project

• Overlay network scalability and performance (VLAN -> VXLAN)

- Using VXLAN offloading in Openstack Neutron project

Page 34: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 34- Mellanox Confidential -

Mellanox Accelerate OpenStack Cinder Storage

▪iSER driver using TGT / LIO based targets

integrated in Cinder since Havana release

Hypervisor (KVM)

OS

VMOS

VMOS

VM

Adapter

Open-iSCSI w iSER

Compute Servers

Switching Fabric

iSCSI/iSER Target (tgt)

Adapter Local Disks

RDMA Cache

Storage Servers

OpenStack (Cinder)

Utilizing OpenStack Built-in components and management Tools to accelerate storage access

>4X Faster

Page 35: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 35- Mellanox Confidential -

Requirements to Deploy iSER

▪Application(s) that can use SCSI/iSCSI• All applications that use SCSI-based block storage work with iSER

▪with OS or Hypervisor that Supports an iSER initiator• Today: Linux & VMware ESXi, Oracle Solaris

• Expected soon: Windows, FreeBSD

▪iSER Storage Target• Oracle ZFS, Violin Memory, Zadara, Saratoga Speed, HP SL4500 (More coming!)

• Create in Linux using LIO, TGT, or SCST target

▪Network that supports RDMA• Adapters support InfiniBand or RoCE

• Switches support InfiniBand or DCBx with PFC

Mellanox Switches support DCBx Ethernet (with PFC) and/or InfiniBand

Mellanox ConnectX-3 Pro and ConnectX-4 Adapters support DCBx Ethernet (with PFC) and/or InfiniBand

Page 36: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 36- Mellanox Confidential -

[DEFAULT]enabled_backends = default, iser, iser_oldiscsi_helper = tgtadm # This is the default..

[default]volume_group = stack-volumes-defaultvolume_driver = cinder.volume.drivers.lvm.LVMVolumeDrivervolume_backend_name = default

[iser]iscsi_ip_address = 1.1.1.1 #supports RDMAiscsi_protocol = iservolume_group = stack-volumes-defaultvolume_driver = cinder.volume.drivers.lvm.LVMVolumeDrivervolume_backend_name = iser

[iser_old]iser_ip_address = 1.1.1.1 #supports RDMAvolume_driver = cinder.volume.drivers.lvm.LVMISERDrivervolume_group = stack-volumes-defaultvolume_backend_name = iser_old

iSER configurations in Openstack - TGT

Page 37: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 37- Mellanox Confidential -

[DEFAULT]enabled_backends = default, iser..

[default]iscsi_helper=lioadmvolume_group = stack-volumes-defaultvolume_driver = cinder.volume.drivers.lvm.LVMVolumeDrivervolume_backend_name = default

[iser]iscsi_ip_address=1.1.1.1iscsi_protocol = iseriscsi_helper=lioadmvolume_group = stack-volumes-defaultvolume_driver = cinder.volume.drivers.lvm.LVMVolumeDrivervolume_backend_name = iser

iSER configurations in Openstack - LIO

Page 38: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 38- Mellanox Confidential -

Cinder Demo Video – iSER Performance

Page 39: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 39- Mellanox Confidential -

Agenda

▪Interconnect technology evolution in virtualized environments

• User-space networking overview (TCP/UDP -> RDMA)

• Virtual interfaces overview (emulation -> paravirt -> SRIOV)

- Using SR-IOV in Openstack Neutron project

• Storage protocols and performance (FC -> iSCSI -> iSER)

- Using iSER in Openstack Cinder project

• Overlay network scalability and performance (VLAN -> VXLAN)

- Using VXLAN offloading in Openstack Neutron project

Page 40: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 40- Mellanox Confidential -

Server

VM1 VM2 VM3 VM4

Overlay Networks (VXLAN/NVGRE/GENEVE) Acceleration

Overlay Network Virtualization: Isolation, Simplicity, Scalability

Virtual Domain 3

Virtual Domain 2

Virtual Domain 1

Physical View

Server

VM5 VM6 VM7 VM8

Mellanox SDN Switches & Routers

VirtualView

NVGRE/VXLAN Overlay Networks Virtual Overlay Networks Simplifies Management and VM Migration

ConnectX-3 Pro Overlay Accelerators Enable

Bare Metal Performance

Page 41: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 41- Mellanox Confidential -

VxLAN = Virtual Extensible LAN

▪What is VxLAN• “Virtual Extensible LAN (VXLAN) is a network virtualization technology that attempts to ameliorate the

scalability problems associated with large cloud computing deployments. It uses a VLAN-like encapsulation technique to encapsulate MAC-based OSI layer 2 Ethernet frames within layer 3 UDP packets.” Wikipedia

▪Enable creating millions of virtual L2 networks over traditional IP networks• Can serve tenants in a cloud provider infrastructure

▪Can span local or wide area networks • Can migrate the entire network between cloud providers/sites and in case of a disaster • Can create logical L2 networks which span multiple locations (like VPNs)

▪Can run over routers • Leverage L3 network scalability and protocols (OSPF, BGP, ECMP)

Page 42: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 42- Mellanox Confidential -

The VxLAN and NV-GRE overlay challenge

▪Hypervisor IP stack and standard NICs are not aware of the client TCP/IP traffic

▪Common offload techniques such as hardware segmentation/re-assembly, checksum offload, and CPU core scaling (RSS/TSS) do not operate on the VM TCP/IP packets (inner payload)

▪Leading to significant CPU overhead and much lower performance

▪Solution: Overlay aware Network Interface Cards• Overlay Network Accelerators• Penalty free overlays at bare-metal speed

Generated by the VM

Generated by the Hypervisor

VXLAN Packet Format

Page 43: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 43- Mellanox Confidential -

Turbocharge Overlay Networks with ConnectX-3/4 NICs

“Mellanox is the Only Way to Scale Out Overlay Networks”

Saving 35% of total cores while doubling the throughput!

Page 44: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 44- Mellanox Confidential -

Agenda

▪Interconnect technology evolution in virtualized environments

• User-space networking overview (TCP/UDP -> RDMA)

• Virtual interfaces overview (emulation -> paravirt -> SRIOV)

- Using SR-IOV in Openstack Neutron project

• Storage protocols and performance (FC -> iSCSI -> iSER)

- Using iSER in Openstack Cinder project

• Overlay network scalability and performance (VLAN -> VXLAN)

- Using VXLAN offloading in Openstack Neutron project

Page 45: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 45- Mellanox Confidential -

VXLAN in Openstack

▪In order to use VXLAN in Openstack, the OVS Mechanism drivers has to be used (one of them),

tunneling type has to be set to VXLAN and L2 population enabled.

▪VXLAN is usually being used in environments with large amount of tenants or as a tunneling

solution.

▪When using VXLAN in Openstack, neutron creates tunnels between every two hypervisors (Mesh),

and creates a virtual endpoint with the remote IP and VXLAN Network Identifier with OVS/LB.

Page 46: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 46- Mellanox Confidential -

Mellanox NIC VXLAN offload in Openstack

▪VXLAN offload is done in Hardware (Mellanox NIC).

▪In order to enable VXLAN offload, the following configurations has to be used:

https://community.mellanox.com/docs/DOC-1446

Offloading VXLAN Overhead

Page 47: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 47- Mellanox Confidential -

Edit /etc/neutron/plugins/ml2/ml2_conf.ini:

 [ovs]bridge_mappings = default:br-eth5enable_tunneling = Truelocal_ip = 192.168.215.1

[agent]vxlan_udp_port = 4789tunnel_types = vxlanl2_population = Trueroot_helper = sudo /usr/local/bin/neutron-rootwrap /etc/neutron/rootwrap.conf

VXLAN configurations in Openstack

Page 48: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 48- Mellanox Confidential -

VXLAN Demo

Page 49: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 49- Mellanox Confidential -

Mellanox high performance features in Openstack

Summary

Page 50: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 50- Mellanox Confidential -

Comprehensive OpenStack Integration for Switch and Adapter

Neturon-ML2 support for mixed environment with

SR-IOV

Hardware support for performance and

security(VXLAN offload, PV,

SRIOV)

Accelerating storage access by

up to 5x

Neutron ML2 plugin

Page 51: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 51- Mellanox Confidential -

Vast OpenStack distributions support

Ethernet

Integrated with Major OpenStack

Distributions

Provided In-Box Since Openstack Havana

release

Can be used transparently over

Ethernet or Infiniband Networks

Page 52: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 52- Mellanox Confidential -

Accelerating Openstack Cloud Performance - Summary

6X Faster

Storage

Overlay Networks

Virtualization

2.5X

20X

6X

20X Faster

Fibre Channel 8Gb

iSER 40GbE VMs Write

2.5X Faster

Page 53: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 53- Mellanox Confidential -

▪CloudX is a group of reference architectures which allow

building the most efficient, high performance and scalable

Infrastructure As A Service (IaaS) clouds based on

Mellanox superior interconnect and off the shelf building

blocks

▪Supports the most popular cloud software

• Windows Azure Pack (WAP)

• OpenStack

• VMware

http://www.mellanox.com/solutions/cloud/reference.php

CloudX: Optimized Cloud Platform

Page 54: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 54- Mellanox Confidential -

Mellanox Advantages - Summary

• Support more VMs per server

• Offload hypervisor CPU

• Overlay networks

• Unlimited scalability

• Record braking throughput

• Record braking IOPS

• Higher storage density

• Centralized management (e.g. SDN)

• I/O consolidation (one wire)

• 100Gb/s per port with RDMA

• 2us for VM to VM connectivity

• Low CPU utilization

Higher Performance

Cost EffectiveStorage

Simplified and Integrated Solutions

Higher Infrastructure

Efficiency

Maximizing Cloud Return on Investment

Page 55: Erez Cohen & Aviram Bar Haim, Mellanox - Enhancing Your OpenStack Cloud With Advanced Network and Storage Interconnect Technologies, OpenStack Israel 2015

© 2014 Mellanox Technologies 55- Mellanox Confidential -

Thank YouThank You