Tesla P40 offers great inference performance, INT8 precision and 24GB of on board memory for an amazing user experience. A Tesla P40 is purpose-built to deliver maximum throughput for deep learning deployment. With 47 TOPS of inference performance and INT8 operations per GPU, a single server with 8 Tesla P40s delivers the performance of over 140 CPU servers.
- World’s fastest processor for inference workloads
- 47 TOPS of INT8 for maximum inference throughput and responsiveness
- Hardware-decode engine capable of transcoding
- Inferencing 35 HD video streams in real time