drorgoldenberg infiniband technology overview

39
Dror Goldenberg, Mellanox Technologies Dror Goldenberg, Mellanox Technologies InfiniBand Technology Overview

Upload: techone

Post on 26-Dec-2015

7 views

Category:

Documents


1 download

DESCRIPTION

SNIA overview of Infiniband

TRANSCRIPT

Page 1: DrorGoldenberg InfiniBand Technology Overview

Dror Goldenberg, Mellanox TechnologiesDror Goldenberg, Mellanox Technologies

InfiniBand Technology

Overview

Page 2: DrorGoldenberg InfiniBand Technology Overview

2InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

SNIA Legal Notice

The material contained in this tutorial is copyrighted by the SNIA. Member companies and individuals may use this material in presentations and literature under the following conditions:

Any slide or slides used must be reproduced without modificationThe SNIA must be acknowledged as source of any material used in the body of any document containing material from these presentations.

This presentation is a project of the SNIA Education Committee.

Page 3: DrorGoldenberg InfiniBand Technology Overview

3InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Abstract

InfiniBand Technology OverviewThe InfiniBand architecture brings fabric consolidation to the data center. Storage networking can concurrently run with clustering, communication and management fabrics over the same infrastructure, preserving the behavior of multiple fabrics. The tutorial provides an overview of the InfiniBand architecture including discussion of High Speed – Low Latency, Channel I/O, QoS scheduling, partitioning, high availability and protocol offload. InfiniBand based storage protocols, iSER (iSCSI RDMA Protocol), NFS over RDMA and SCSI RDMA Protocol (SRP), are introduced and compared with alternative storage protocols, such as iSCSI and FCP. The tutorial further enumerates value-add features that the InfiniBand brings to clustered storage, such as atomic operations and end to end data integrity.

Learning Objectives:– Understand the InfiniBand architecture and feature set.– Understand the benefits of InfiniBand for networked storage.– Understand the standard InfiniBand storage protocols.

Page 4: DrorGoldenberg InfiniBand Technology Overview

4InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Agenda

Motivation and General OverviewProtocol Stack Layers Storage Protocols over InfiniBandBenefits

Page 5: DrorGoldenberg InfiniBand Technology Overview

5InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

The Need for Better I/O

Datacenter trendsMulti-core CPUsBladed architectureFabric consolidationServer virtualization & consolidationIncreasing storage demand

Better I/O is requiredHigh capacityEfficient

Low latencyCPU Offload

ScalableVirtualization friendlyHigh availabilityPerformanceLow powerTCO reduction

Com

pute

Nod

e

CPU CPU

OSApp App

I/O

CPUCore

CPUCore

OSI/O

Com

pute

Nod

e

CPUCore

OS OS OS OS

CPUCore

I/O

Com

pute

Nod

e

Page 6: DrorGoldenberg InfiniBand Technology Overview

6InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

The InfiniBand Architecture

Industry standard defined by the InfiniBand Trade AssociationDefines System Area Network architecture

Comprehensive specification: from physical to applications

Architecture supportsHost Channel Adapters (HCA)Target Channel Adapters (TCA)SwitchesRouters

Facilitated HW design for Low latency / high bandwidthTransport offload

Rev1.0

2000 2001 2004

Rev1.0a

Rev1.2

2007

Rev1.2.1

2002

Rev1.1

… …

ProcessorNode

InfiniBandSubnet

Gateway

HCA

SwitchSwitch

SwitchSwitch

ProcessorNode

ProcessorNode

HCA

HCA

TCA

StorageSubsystem

Consoles

TCA

RAID

Ethernet

Gateway

FibreChannel

HCA

Subnet Manager

Page 7: DrorGoldenberg InfiniBand Technology Overview

7InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

A Comparison of Fabric TechnologiesFeatures and Price

Fibre Channel Standard 10 GbE InfiniBand

Bandwidth 4Gb/s (4GFC)8Gb/s (8GFC)

10Gb/s 20Gb/s (4x DDR)

Raw Bandwidth (unidirectional)

400MB/s (4GFC)800MB/s (8GFC)

1,250 MB/s 2,000 MB/s* (4x DDR)4,000 MB/s (4x QDR)

Reliable Service Yes No Yes

Fabric Consolidation

Practically no Practically partial** Yes

Copper Distance 15m 10GBase-CX4 15m10GBase-T 100m

Passive SDR 20m/ DDR 10mActive DDR 25m

Optical Distance 100m 10GBase-SR 300m10GBase-LRM 220m

300m (SDR)150m (DDR)

* 1,940 MB/s measured

Check outSNIA Tutorial:

Fibre Channelover Ethernet

Check outSNIA Tutorial:

Fibre Channelover Ethernet

**

Page 8: DrorGoldenberg InfiniBand Technology Overview

8InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

InfiniBand Topologies

Example topologies commonly usedArchitecture does not limit topologyModular switches are based on fat tree architecture

Back to Back

2 Level Fat Tree

……

……

… ………

3D Torus

Dual StarHybrid

Page 9: DrorGoldenberg InfiniBand Technology Overview

9InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

InfiniBand Node

ULP

NetworkLayer

LinkLayer

PhysicalLayer

TransportLayer

Application

InfiniBand Node

ULP

NetworkLayer

LinkLayer

PhysicalLayer

TransportLayer

Application

InfiniBand Protocol Layers

Packet relay

PHY

PHY

InfiniBand Router

Packet relay

PHY

Link

PHY

Link

InfiniBand Switch

Page 10: DrorGoldenberg InfiniBand Technology Overview

10InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Physical Layer

Width (1X, 4X, 8X, 12X) including auto-negotiationSpeed (SDR/DDR/QDR) including auto-negotiation

4X DDR HCAs are currently shippingPower management

Polling / SleepingConnector

Board: MicroGiGaCNPluggable: QSFP

8/10 encodingMaintain DC BalanceLimited run length of 0’s or 1’s

Control symbols (Kxx.x)Lane de-skew, auto negotiation, training, clock tolerance, framing

Lane Speed →

Link Width ↓

SDR (2.5GHz)

DDR (5GHz)

QDR (10GHz)

1X 2.5 5 10

4X 10 20 40

8X 20 40 80

12X 30 60 120

Link Speed (10Link Speed (1099 bit/sec)bit/sec)

* MicroGiGaCN is a trademark of Fujitsu Components Limited

Page 11: DrorGoldenberg InfiniBand Technology Overview

11InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Physical Layer – Cont’d

* currently deployed

Width Speed Connector Type MinReach

Power

MediaConverter

300m/150m

Optical Cable

0.8-1W

1W100m

Fiber Media

4X SDR/DDR

Micro-GiGaCN

12 strand MPO

4X DDR Micro-GiGaCN

12 strand attached

Fiber Optics*:Fiber Optics*:

4X –MicroGiGaCN→

Width Speed Connector MinReach

Type / Power

4X SDR/DDR

Micro-GiGaCN

20m/10m

4X DDR Micro-GiGaCN

15-25m Active0.5-1.5W

20m/10m

Passive

12X SDR/DDR

24pin Micro-GiGaCN

Passive

Copper Cables*:Copper Cables*:

4X - MicroGiGaCNMPO Media Converter →

12X –24 pair MicroGiGaCN→

4X - MicroGiGaCNOptical Cable →

Page 12: DrorGoldenberg InfiniBand Technology Overview

12InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Link LayerAddressing and Switching

Local Identifier (LID) addressingUnicast LID - 48K addressesMulticast LID – up to 16K addressesEfficient linear lookupCut through switching supported Multi-pathing support through LMC

Independent Virtual LanesFlow control (lossless fabric)Service level VL arbitration for QoS

Congestion controlForward / Backward Explicit Congestion Notification (FECN/BECN)

Data IntegrityInvariant CRCVariant CRC

Independent Virtual Lanes (VLs)

HighPriority

WRR

LowPriority

WRR

PrioritySelect

Packetsto be

Transmitted

H/L Weighted Round Robin (WRR) VL Arbitration

Efficient FECN/BECN Based Congestion Control

Switch

threshold

FECN

BECNBECN

HCA HCA

Per QP/VLinjection rate control

VL A

RB

Page 13: DrorGoldenberg InfiniBand Technology Overview

13InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Network Layer

Global Identifier (GID) addressingBased on IPv6 addressing schemeGID = {64 bit GID prefix, 64 bit GUID}

GUID = Global Unique Identifier (64 bit EUI-64)GUID 0 – assigned by the manufacturerGUID 1..(N-1) – assigned by the Subnet Manager

Optional for local subnet accessUsed for multicast distribution within end nodesEnables routing between IB subnets

Still under definition in IBTAWill leverage IPv6 routing algorithms

Subnet A Subnet BIB Router

Page 14: DrorGoldenberg InfiniBand Technology Overview

14InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Transport - Host Channel Adapter ModelAsynchronous interface

Consumer posts work requestsHCA processesConsumer polls completions

Transport executed by HCAI/O channel exposed to the applicationTransport services

Reliable / UnreliableConnected / Datagram

Port

VL VL VL VL…

Port

VL VL VL VL…

Transport and RDMA Offload Engine

SendQueue

ReceiveQueue

QP

SendQueue

ReceiveQueue

QP

Consumer

CompletionQueue

postingWQEs

pollingCQEs

HCAHCA

InfiniBand

CPU CPU

Chipset

HCA

Mem

PCIe

CPU CPU

Bridge

HCA

Mem

PCIe

InfiniBand

Mem

Page 15: DrorGoldenberg InfiniBand Technology Overview

15InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Transport Layer

Queue Pair (QP) – transport endpointAsynchronous interface

Send Queue, Receive Queue, Completion QueueFull transport offload

Segmentation, reassembly, timers, retransmission, etcOperations supported

Send/Receive – messaging semanticsRDMA Read/Write – enable zero copy operationsAtomics – remote Compare & Swap, Fetch & AddMemory management - Bind/Fast Register/Invalidate

Kernel bypassEnables low latency and CPU offloadEnabled through QPs, Completion Queues (CQs), Protection Domains (PD), Memory Regions (MRs)

Page 16: DrorGoldenberg InfiniBand Technology Overview

16InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Partitions

Logically divide the fabric into isolated domainsPartial and full membership per partitionPartition filtering at switches

Similar toFC Zoning802.1Q VLANs

Host AHost B

InfiniBand fabric

Partition 1 Inter-Host

Partition 2private to host BPartition 3

private to host APartition 4shared

I/O A

I/O BI/O C

I/O D

Page 17: DrorGoldenberg InfiniBand Technology Overview

17InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

InfiniBand Packet Format

LRH GRH BTH ExtHDRs Payload ICRC VCRC

8B 40B 12B var 0..4096B 4B 2B

VL LVer SL rsvdLNH DLID

rsvd SLIDLen

LRH

IPVer

Next HeaderPayload Len

Flow LabelTClass

Hop Lim

SGID[127:96]

SGID[95:64]

SGID[63:32]

SGID[31:0]

DGID[127:96]

DGID[95:64]

DGID[63:32]

DGID[31:0]

GRH (Optional)

Partition Key

Destination QP

TVerOpcode

rsvd

PSNrsvdA

SMPad

BTH

InfiniBand Data Packet

Extended headers:•Reliable Datagram ETH (4B)•Datagram ETH (8B)•RDMA ETH (16B)•Atomic ETH (28B)•ACK ETH (4B)•Atomic ACK ETH (8B)•Immediate Data ETH (4B)•Invalidate ETH (4B)

Page 18: DrorGoldenberg InfiniBand Technology Overview

18InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

InfiniBand Data Integrity

Hop by hopVCRC – 16 bit CRCCRC16 0x100B

End to endICRC – 32 bit CRCCRC32 0x04C11DB7 Same CRC as Ethernet

Application levelT10/DIF Logical Block Guard

Per block CRC16 bit CRC 0x8BB7

InfiniBand Fabric

Fibre

Channel

SAN

VCRC VCRC VCRC

ICRC

T10/DIF

T10/DIF

VCRC VCRC VCRC

ICRC

Switch

Switch

Gateway

Switch

InfiniBandBlock Storage

FC Block Storage

VCRC

Page 19: DrorGoldenberg InfiniBand Technology Overview

19InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Management Model

Subnet Mgt Agent

Subnet Manager

Subnet Management Interface

QP0 (uses VL15)QP1

Baseboard Mgt Agent

Communication Mgt Agent

Performance Mgt Agent

Device Mgt Agent

Vendor-Specific Agent

Application-Specific Agent

SNMP Tunneling Agent

Subnet Administration

General Service Interface

Subnet Manager (SM)Configures/Administers fabric topologyImplemented at an end-node or a switchActive/Passive model when more than one SM is presentTalks with SM Agents in nodes/switches

Subnet AdministrationProvides path recordsQoS management

Communication ManagementConnection establishment processing

Page 20: DrorGoldenberg InfiniBand Technology Overview

20InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Upper Layer Protocols

ULPs connect InfiniBand to common interfacesSupported on mainstream operating systems

ClusteringMPI (Message Passing Interface)RDS (Reliable Datagram Socket)

NetworkIPoIB (IP over InfiniBand)SDP (Socket Direct Protocol)

StorageSRP (SCSI RDMA Protocol)iSER (iSCSI Extensions for RDMA)NFSoRDMA (NFS over RDMA)

Hardware

Device Driver

InfiniBand Core Services

IPoIB

TCP/IP SDP RDS

socket interface

SRP iSERNFSover

RDMA

block storagefile storage

Device Driver

InfiniBand Core Services

MPI

HPC clustering

kern

el b

ypas

s

Kernel

IBApps

IBApps

ClusteringApps

sockets

Socket based Apps

Userstorage

Interfaces(file/block)

Storage Apps

Operating system InfiniBand Infrastructure Applications

Page 21: DrorGoldenberg InfiniBand Technology Overview

21InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

InfiniBand Block Storage Protocols

SRP - SCSI RDMA ProtocolDefined by T10

iSER – iSCSI Extensions for RDMADefined by IETF IP Storage WGInfiniBand specifics (e.g. CM) defined by IBTALeverages iSCSI management infrastructure

Protocol offloadUse IB Reliable ConnectedRDMA for zero copy data transfer

SCSI Application

Layer

SCSI Transport Protocol

Layer

Interconnect Layer

SAM-3

FC-3 (FC-FS, FC-LS)

FC-2 (FC-FS)FC-1 (FC-FS)FC-0 (FC-PI)

SCSI Application

Layer

FC-4 Mapping (FCP-3)

Fibre Channel

InfiniBand

SCSI Application

Layer

SRP

SRP

InfiniBand /iWARP

SCSI Application

Layer

iSCSI

iSCSI

iSER

Page 22: DrorGoldenberg InfiniBand Technology Overview

22InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

SRP - Data Transfer Operations

Send/ReceiveCommandsResponses Task management

RDMA – Zero Copy PathData-InData-Out

iSER uses the same principlesImmediate/Unsolicited data allowed through Send/Receive

Initiator TargetSRP_CMD

RDMA Write

SRP_RSP

Initiator TargetSRP_CMD

RDMA Read

SRP_RSP

RDMA Read Resp

IO ReadIO Read

IO WriteIO Write

Page 23: DrorGoldenberg InfiniBand Technology Overview

23InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Data Transfer Summary

SRP iSER iSCSI FCP

Request SRP_CMD (SEND) SCSI-Command (SEND)

SCSI-Response (SEND)

RDMA Write

RDMA ReadRDMA Read Resp.

Unsolicited Data-Out Delivery

Part of SCSI-Command (SEND)Data-Out (SEND)

Part of SCSI-CommandData-Out

FCP_DATA

Task Management Function Request/Response (SEND)

SCSI-Command FCP_CMND

Response SRP_RSP (SEND) SCSI-Response(or piggybacked on Data-In PDU)

FCP_RSP

Data-In Delivery RDMA Write Data-In FCP_DATA

Data-Out Delivery RDMA ReadRDMA Read Resp.

R2TData-Out

FCP_XFER_RDYFCP_DATA

Task Management SRP_TSK_MGMT (SEND)

Task Management Function Request/Response

FCP_CMND

Page 24: DrorGoldenberg InfiniBand Technology Overview

24InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

SRP Discovery

Discovery methodsPersistent Information {Node_GUID:IOC_GUID}Subnet Administrator (Identify all ports with CapabilityMask.IsDM)Configuration Manager (CFM)

Locate the Device Administrator through Service Record

Boot ManagerBoot Information Service

IdentifiersPer LUN WWN (through INQUIRY VPD)SRP Target Port ID {IdentifierExt[63:0], IOC GUID[63:0]}Service Name – SRP.T10.{PortID ASCII}Service ID – Locally assigned by the IOC/IOU

I/O Controller

I/OController

I/O U

nit

InfiniBand I/O Model

Page 25: DrorGoldenberg InfiniBand Technology Overview

25InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

iSER Discovery

Leverages all iSCSI infrastructureUsing IP over InfiniBand

Same iSCSI mechanisms for discovery (RFC 3721)Static Configuration {IP, port, target name}Send Targets {IP, port}SLPiSNS

Same target naming (RFC 3721/3980)iSCSI Qualified Names (iqn.)IEEE EUI64 (eui.)T11 Network Address Authority (naa.)

Page 26: DrorGoldenberg InfiniBand Technology Overview

26InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

NFS over RDMA

Defined by IETFONC-RPC extensions for RDMANFS mapping

RPC Call/ReplySend/Receive – if smallVia RDMA Read chunk list - if big

Data transferRDMA Read/Write – described by chunk list in XDR messageSend – inline in XDR message

Uses InfiniBand Reliable Connected QPUses IP extensions to CMConnection based on IP address and TCP portZero copy data transfers

Client ServerRPC Call

RDMA Write

RPC Reply

Client ServerRPC Call

RDMA Read

RPC Reply

RDMA Read Resp

NFS READNFS READ

NFS WRITENFS WRITE

Page 27: DrorGoldenberg InfiniBand Technology Overview

27InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

I/O Consolidation

Convergewith InfiniBand

OS

High bandwidth pipe for capacity provisioningDedicated I/O channels enable convergence

For Networking, Storage, ManagementApplication compatibilityQoS - differentiates different traffic typesPartitions – logical fabrics, isolation

Gateways - Share remote Fibre Channel and Eth portsDesign based on average load across multiple serversScale incrementally – add Ethernet/FC/Server bladesScale independently

Slower I/ODifferent service needs –different fabricsNo flexibilityMore ports to manageMore powerMore spaceHigher TCO

StorageApp

NetworkingApp

ManagementApp

OS

NetworkingApp

ManagementApp

IB HCA

StorageApp

One wireGbE NIC

FC HCA GbE NICFC HCA GbE NIC

GbE NICGbE NIC

GbE NIC

Page 28: DrorGoldenberg InfiniBand Technology Overview

28InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

High Availability and Redundancy

Multi-port HCAsCovers link failure

Redundant fabric topologiesCovers link failure

Link layer multi-pathing (LMC)Automatic Path Migration (APM)ULP High Availability

Application level multi-pathing (SRP/iSER)Teaming/Bonding (IPoIB)Covers HCA failure and link failure

Page 29: DrorGoldenberg InfiniBand Technology Overview

29InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Performance Metrics

IB VerbsLatency

RDMA Write 0.99usRDMA Read 1.87us (roundtrip)

Bandwidth1.5-1.9GB/s (unidirectional)3.0-3.4GB/s (bidirectional)Depends on PCIe (2.5-5GT/s)

Clustering (MPI)Latency 1.2usMessage rate 30M msg/sec

Block Storage (SRP)Bandwidth (1MB I/O, no RAID)

I/O Read 1.4GB/sI/O Write 1.2GB/s

File Storage (NFSoRDMA)Read 1.3GB/sWrite 0.59GB/s

Page 30: DrorGoldenberg InfiniBand Technology Overview

30InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

InfiniBand Storage Opportunities & Benefits

Clustering port can connect to storageHigh Bandwidth FabricFabric consolidation (QoS, partitioning)Efficiency – full offload and zero copyGateways

One wire out of the serverShared remote FC ports - scalability

Clustered/Parallel storage, Backend fabric benefits:

Combined with clustering infrastructureEfficient object/block transferAtomic operationsUltra low latency High bandwidth

Parallel / clustered file-systemParallel NFS

Server

OSD/BlockStorage Targets

Servers

InfiniBand

BackendInfiniBand

Native IBJBODs

Direct attach native IB

Block Storage

Native IBFile Server

(NFS RDMA)Native IB

Block Storage(SRP/iSER)

Servers InfiniBand

Gateway

InfiniBand Storage Deployment Alternatives

FibreChannel

Page 31: DrorGoldenberg InfiniBand Technology Overview

31InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Summary

Datacenter developments require better I/OIncreasing compute power per hostServer virtualizationIncreasing storage demand

InfiniBand I/O is a great fit for the datacenterLayered implementationBrings fabric consolidationEnables efficient SAN, Network, IPC and Management trafficPrice/PerformanceGateways provide scalable connectivity to existing fabrics

Existing storage opportunities with InfiniBand

Page 32: DrorGoldenberg InfiniBand Technology Overview

32InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Q&A / Feedback

Please send any questions or comments on this presentation to SNIA: [email protected]

Many thanks to the following individuals for their contributions to this tutorial.

SNIA Education Committee

Bill Lee Howard GoldsteinRon Emerick Sujal DasWalter Dey

Page 33: DrorGoldenberg InfiniBand Technology Overview

Backup

Page 34: DrorGoldenberg InfiniBand Technology Overview

34InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Interconnect: A Competitive Advantage

Enterprise Data CentersEnterprise Data Centers

HighHigh--Performance ComputingPerformance Computing

EmbeddedEmbedded

Clustered DatabaseeCommerce and RetailFinancialSupply Chain ManagementWeb Services

Biosciences and GeosciencesComputer Automated EngineeringDigital Content CreationElectronic Design AutomationGovernment and Defense

CommunicationsComputing and Storage AggregationIndustrialMedicalMilitary

End-UsersServers

And Blades

Embedded

Switches

Storage

Page 35: DrorGoldenberg InfiniBand Technology Overview

35InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Top500 Interconnect Trends

020406080

100120140160180200220240260

Num

ber o

f Clu

ster

s

Jun-05 Nov-05 Jun-06 Nov-06 Jun-07

Interconnect Trends – Top500

Growth rate from Nov 06 to June 07 (6 months)

InfiniBand: +61%Myrinet: -41%GigE: -4%

InfiniBand Myrinet GigE

61% growth for InfiniBand from Nov 2006, 230% growth from June 2006

132

47

207

Growth rate from June 06 to June 07 (year)

InfiniBand: +230%Myrinet: -47%GigE: -19%

Source: http://www.top500.org/list/2007/06/The TOP500 project was started in 1993 to provide a reliable basis for tracking and detecting trends in high-performance computing.

Page 36: DrorGoldenberg InfiniBand Technology Overview

36InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Applicable Markets for InfiniBand

Data CentersClustered database, data warehousing, shorter backups, I/O consolidation, power savings, virtualization, SOA, XTP

FinancialReal-time risk assessment, grid computing and I/O consolidation

Electronic Design Automation (EDA) and Computer Automated Design (CAD)

File system I/O is the bottleneck to shorter job run timesHigh Performance Computing

High throughput I/O to handle expanding datasetsGraphics and Video Editing

HD file sizes exploding, shorter backups, real-time production

Page 37: DrorGoldenberg InfiniBand Technology Overview

37InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

InfiniBand Resources

InfiniBand software is developed under OpenFabrics Open source Alliance http://www.openfabrics.org/index.html

InfiniBand standard is developed by the InfiniBand® Trade Association http://www.infinibandta.org/home

Page 38: DrorGoldenberg InfiniBand Technology Overview

38InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Reference

InfiniBand Architecture Specification Volume 1-2 Release 1.2

www.infinibandta.org

IP over InfiniBandRFCs 4391, 4392, 4390, 4755 (www.ietf.org)

NFS Direct Data Placementhttp://www.ietf.org/html.charters/nfsv4-charter.html

iSCSI Extensions for RDMA Specificationhttp://www.ietf.org/html.charters/ips-charter.html

SCSI RDMA Protocol, DIFwww.t10.org

Page 39: DrorGoldenberg InfiniBand Technology Overview

39InfiniBand Technology Overview© 2007 Storage Networking Industry Association. All Rights Reserved.

Glossary

APM - Automatic Path MigrationBECN - Backward Explicit Congestion NotificationBTH - Base Transport HeaderCFM - Configuration ManagerCQ - Completion QueueCQE - Completion Queue ElementCRC - Cyclic Redundancy CheckDDR - Double Data RateDIF - Data Integrity FieldFC - Fibre ChannelFECN - Forward Explicit Congestion NotificationGbE - Gigabit EthernetGID - Global IDentifierGRH - Global Routing HeaderGUID - Globally Unique IDentifierHCA - Host Channel AdapterIB - InfiniBandIBTA - InfiniBand Trade AssociationICRC - Invariant CRCIPoIB - Internet Protocol Over InfiniBandIPv6 - Internet Protocol Version 6iSER - iSCSI Extensions for RDMALID - Local IDentifierLMC - Link Mask ControlLRH - Local Routing HeaderLUN - Logical Unit Number

MPI - Message Passing InterfaceMR - Memory RegionNFSoRDMA - NFS over RDMAOSD - Object based Storage DeviceOS - Operating SystemPCIe - PCI ExpressPD - Protection DomainQDR - Quadruple Data RateQoS - Quality of ServiceQP - Queue PairRDMA - Remote DMARDS - Reliable Datagram SocketRPC - Remote Procedure CallSAN - Storage Area NetworkSDP - Sockets Direct ProtocolSDR - Single Data RateSL - Service LevelSM - Subnet ManagerSRP - SCSI RDMA ProtocolTCA - Target Channel AdapterULP - Upper Layer ProtocolVCRC - Variant CRCVL - Virtual LaneWQE - Work Queue ElementWRR - Weighted Round Robin