Our customers call us superheroes, who are we to argue?


Ultra-efficient deep learning in scale-out servers


More Image(s)

The Tesla P4 accelerates any scale-out server, offering an incredible 60x higher efficiency compared to CPUs.
Availability: In stock

For details on how to purchase the NVIDIA TESLA P4, please click the button below to send us your details and brief requirements. We can then quote you accordingly.


Ultra-efficient deep learning in scale-out servers

In the new era of AI and intelligent machines, deep learning is shaping our world like no other computing model in history.

Interactive speech, visual search, and video recommendations are a few of many AI-based services that we use every day.

Accuracy and responsiveness are key to user adoption for these services. As deep learning models increase in accuracy and complexity, CPUs are no longer capable of delivering a responsive user experience.

The NVIDIA Tesla P4 is powered by the revolutionary NVIDIA Pascal architecture and purpose-built to boost efficiency for scale-out servers running deep learning workloads, enabling smart responsive AI-based services.

It slashes inference latency by 15x in any hyperscale infrastructure and provides an incredible 60x better energy efficiency than CPUs.

This unlocks a new wave of AI services previous impossible due to latency limitations.

Part No. TESLA P4
Manufacturer nvidia
End of Life? No
GPU Architecture NVIDIA Pascal
Maximum Power Consumption 50W/75W
Server-Optimised for Data Center Deployment Yes
Enhanced Programmability with Page Migration Engine Yes
ECC Protection Yes
Integer Operations (INT8) 22 TeraOPS (TOPS)
GPU Memory 8 GB
Memory Bandwidth 192 GB/s
Single-Precision Performance 5.5 TeraFLOPS
PCI Slot(s) Low-Profile PCIe
Hardware-Accelerated Video Engine 1x Decode Engine, 2x Encode Engine