Our customers call us superheroes, who are we to argue?

Mellanox ConnectX-3 Pro VPI IC

Single/Dual-Port Adapter Silicon with Virtual Protocol Interconnect

[MT27528A1-FCCF-FV]

ConnectX-3 Pro Adapter Silicons with Virtual Protocol Interconenct (VPI), supporting InfiniBand and Ethernet connectivity with hardware offload engines to Ovelay Networks ("Tunneling"), provide the highest performing and most flexible interconnect solution for PCI Express Gen3 servers used in public and private clouds, Enterprise Data Centres, and High Performance Computing (HPC).


Availability: In stock

For details on how to purchase the Mellanox ConnectX-3 Pro VPI IC, please click the button below to send us your details and brief requirements. We can then quote you accordingly.

Details

ConnectX-3 Pro Single/Dual-Port Adapter Device with Virtual Protocol Interconnect

ConnectX-3 Pro adapter devices with Virtual Protocol Interconnect (VPI), supporting InfiniBand and Ethernet connectivity with hardware offload engines to Overlay Networks ("Tunneling"), provide the highest performing and most flexible interconnect solution for PCI Express Gen3 servers used in public and private clouds, enterprise data centres, and high performance computing.


Public and private cloud clustered databases, parallel processing, transactional services, and high-performance embedded I/O applications will achieve significant performance improvements resulting in reduced completion time and lower cost per operation. ConnectX-3 Pro with VPI also simplifies system development by serving multiple fabrics with one hardware design.

Benefits

  • One design for InfiniBand, Ethernet (10GbE, 40GbE), or Data Centre Bridging fabrics
  • World-class cluster, network, and storage performance
  • Cutting edge performance in virtualised overlay networks (VXLAN and NVGRE)
  • Guaranteed bandwidth and low-latency services
  • I/O consolidation
  • Virtualisation acceleration
  • Power efficient
  • Scales to tens-of-thousands of nodes

Key Benefits
  • Virtual Protocol Interconnect
  • 1us MPI ping latency
  • Up to 56Gb/s InfiniBand or 40 Gigabit Ethernet per port
  • Single- and Dual-Port options available
  • PCI Express 3.0 (up to 8GT/s)
  • CPU offload of transport operations
  • Application
  • GPU communication acceleration
  • Precision Clock Synchronisation
  • HW Offloads for NVGRE and VXLAN encapsulated traffic
  • End-to-end QoS and congestion control
  • Hardware-based I/O virtualisation
  • Ethernet encapsulation (EoIB)
  • RoHS R6

Manufacturer Mellanox
Part No. MT27528A1-FCCF-FV
End of Life? No
Advanced Network Features
  • Hardware-based reliable transport
  • Collective operations offloads
  • GPU communication acceleration
  • Hardware-based reliable multicast
  • Extended Reliable Connected transport
  • Enhanced Atomic operations
Bandwidth Max 56Gb/s
Channels 16 million I/O channels
Host OS Support Citrix, RHEL, CentOS, Novell, Ubuntu, Windows Server, FreeBSD, OFED, WinOF, VMWare
I/O Virtualisation
  • Single Root IOV
  • Address translation and protection
  • Dedicated adapter resources
  • Multiple queues per virtual machine
  • Enhanced QoS for vNICs
  • VMWare NetQueue support
Supported Software

Open MPI, OSU MVAPICH, Intel MPI, MS, MPI, Platform MPI, TCP/UDP, EoIB, IPoIB, RDS, SRP, iSER, NFS RDMA, uDAPL

PCI Slot(s) Gen 3.0, 2.0, 1.1
Ports Up to 56Gb/s connectivity per port
Port-Port Latency 1us MPI ping latency
IEEE Compliance
  • IEEE 802.3ae 10 Gigabit Ethernet
  • IEEE 802.3ba 40 Gigabit Ethernet
  • IEEE 802.3ad Link Aggregation
  • IEEE 802.3az Energy Efficient Ethernet
  • IEEE 802.1Q, 1P VLAN tags and priority
  • IEEE 802.1Qau Congestion Notification
  • IEEE 802.1Qbg
  • IEEE P802.1Qaz DO.2 ETS
  • IEEE P802.1Qbb D1.0 Priority-based Flow Control
  • IEEE 1588v2

Others Products of Interest