10
CONFIDENTIAL CONFIDENTIAL Mellanox InfiniBand Interconnect The Fabric of Choice for Clustering and Storage September 2008 Shainer – Director of Technical Marketing

Mellanox InfiniBand Interconnect The Fabric of Choice for Clustering and Storage

Embed Size (px)

DESCRIPTION

Mellanox InfiniBand Interconnect The Fabric of Choice for Clustering and Storage. September 2008. Gilad Shainer – Director of Technical Marketing. Company Overview. Silicon-based server and storage interconnect products R&D, Operations in Israel; Business in California - PowerPoint PPT Presentation

Citation preview

Page 1: Mellanox InfiniBand Interconnect The Fabric of Choice for Clustering and Storage

CONFIDENTIALCONFIDENTIAL

Mellanox InfiniBand InterconnectThe Fabric of Choice for Clustering and Storage

September 2008

Gilad Shainer – Director of Technical Marketing

Page 2: Mellanox InfiniBand Interconnect The Fabric of Choice for Clustering and Storage

2 Mellanox ConfidentialMellanox Confidential

Company Overview

Silicon-based server and storage interconnect products• R&D, Operations in Israel; Business in California• Four generations of products since 1999• 250+ employees; worldwide sales & support

InfiniBand and Ethernet leadership• Foundation for the world’s most powerful computer• 3.7M 10/20/40Gb/s ports shipped as of Jun08• Proven execution, high-volume manufacturing & quality

Solid financial position• FY’07 $84.1M, 73% growth from FY’06• Record Revenue in 2Q’08, $28.2M• 1H’08 $53.4M, 3Q’08 est. $28.5M-$29M

Tier-one, diversified customer base• Includes Cisco, Dawning, Dell, Fujitsu, Fujitsu-Siemens, HP, IBM, NEC,

NetApp, QLogic, SGI, Sun, Supermicro, Voltaire

$106M raised in IPO Feb07Ticker MLNX

Page 3: Mellanox InfiniBand Interconnect The Fabric of Choice for Clustering and Storage

3 Mellanox ConfidentialMellanox Confidential

InfiniBand End-to-End Products

High Throughput - 40Gb/s Low latency - 1us Low CPU overhead

Kernel bypass Remote DMA (RDMA) Reliability

Blade/Rack Servers StorageSwitch

ADAPTER ADAPTERSWITCH

Adapter ICs & Cards

Cables

Switch ICs

Software

End-to-End Validation

Maximum Productivity

Cables

Adapter ICs & Cards

Page 4: Mellanox InfiniBand Interconnect The Fabric of Choice for Clustering and Storage

4 Mellanox ConfidentialMellanox Confidential

Virtual Protocol Interconnect

StorageNFS, CIFS, iSCSI

NFS-RDMA, SRP, iSER,Fibre Channel, Clustered

NetworkingTCP/IP/UDP

Sockets

ClusteringMPI, DAPL, RDS, Sockets

ManagementSNMP, SMI-S

OpenView, Tivoli, BMC, Computer Associates

10/20/40 InfiniBand

Consolidated Application Programming Interface

App1 App2 App3 App4 AppX…

Acceleration Engines

10GigE Data CenterEthernet

Any Protocol over Any Convergence Fabric

Protocols

Applications

Networking VirtualizationClustering Storage RDMA

Page 5: Mellanox InfiniBand Interconnect The Fabric of Choice for Clustering and Storage

5 Mellanox ConfidentialMellanox Confidential

The Fastest InfiniBand Technology

InfiniBand 40Gb/s QDR in full productions• Multiple sites already utilized InfiniBand QDR performance

ConnectX InfiniBand - 40Gb/s server and storage adapter• 1usec application latency, zero scalable latency impact

InfiniScale IV - 36 InfiniBand 40Gb/s switch device• 3Tb/s switching capability in a single switch device

Page 6: Mellanox InfiniBand Interconnect The Fabric of Choice for Clustering and Storage

6 Mellanox ConfidentialMellanox Confidential

InfiniBand QDR Switches

1RU 36-port QSFP, QDR switch• Up to 2.88Tb/s switching capacity• Powered connectors for active cables• Available now

19U 18 slot chassis, 324-port QDR switch• Up to 25.9Tb/s switching capacity• 18 QSFP ports per switch blade• Available: Q4 2009

Page 7: Mellanox InfiniBand Interconnect The Fabric of Choice for Clustering and Storage

7 Mellanox ConfidentialMellanox Confidential

InfiniBand Technology Leadership

Industry Standard• Hardware, software, cabling, management• Design for clustering and storage interconnect

Price and Performance• 40Gb/s node-to-node• 120Gb/s switch-to-switch• 1us application latency• Most aggressive roadmap in the industry

Reliable with congestion management Efficient

• RDMA and Transport Offload• Kernel bypass• CPU focuses on application processing

Scalable for Petascale computing & beyond

End-to-end quality of service Virtualization acceleration I/O consolidation Including storage InfiniBand Delivers the Lowest Latency

The InfiniBand Performance Gap is Increasing

Fibre Channel

Ethernet

60Gb/s

20Gb/s

120Gb/s

40Gb/s

240Gb/s (12X)

80Gb/s (4X)

Page 8: Mellanox InfiniBand Interconnect The Fabric of Choice for Clustering and Storage

8 Mellanox ConfidentialMellanox Confidential

InfiniBand 40Gb/s QDR Capabilities

Performance driven architecture • MPI latency 1us, Zero scalable latency• MPI bandwidth 6.5GB/s bi-dir, 3.25GB/s uni-dir

Enhanced communication• Adaptive/static routing, congestion control

Enhanced Scalability• Communication/Computation overlap• Minimizing systems noise effect (DOE funded project)

Mellanox ConnectX MPI Latency - Multi-core Scaling

0

2

4

6

1 2 3 4 5 6 7 8

# of CPU cores (# of processes)

La

ten

cy

(u

se

c)

Mellanox ConnectX MPI Latency - Multi-core Scaling

0

3

6

9

1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16

# of CPU cores (# of processes)

La

ten

cy

(u

se

c)

8-cores 16-cores

Page 9: Mellanox InfiniBand Interconnect The Fabric of Choice for Clustering and Storage

9 Mellanox ConfidentialMellanox Confidential

HPC Advisory Council

Distinguished HPC alliance (OEMs, IHVs, ISVs, end-users) Members activities

• Qualify and optimize HPC solutions • Early access to new technology, mutual development of future solutions• Outreach

A community effort support center for HPC end-users• End-User Cluster Center• End- user support center

For details – [email protected]

Page 10: Mellanox InfiniBand Interconnect The Fabric of Choice for Clustering and Storage

10 Mellanox ConfidentialMellanox Confidential

Thank You

[email protected]

10