White Papers

Deep Learning Inference on P40 GPUs
Authors: Rengan Xu, Frank Han and Nishanth Dandapanthu. Dell EMC HPC Innovation Lab. Mar. 2017
Introduction to P40 GPU and TensorRT
Deep Learning (DL) has two major phases: training and inference/testing/scoring. The training phase
builds a deep neural network (DNN) model with the existing large amount of data. And the inference
phase uses the trained model to make prediction from new data. The inference can be done in the data
center, embedded system, auto and mobile devices, etc. Usually inference must respond to user request
as quickly as possible (often in real time). To meet the low-latency requirement of inference, NVIDIA®
launched Tesla® P4 and P40 GPUs. Aside from high floating point throughput and efficiency, both GPUs
introduce two new optimized instructions designed specifically for inference computations. The two new
instructions are 8-bit integer (INT8) 4-element vector dot product (DP4A) and 16-bit 2-element vector dot
product (DP2A) instructions. Deep learning researchers have found using FP16 is able to achieve the same
inference accuracy as FP32 and many applications only require INT8 or lower precision to keep an
acceptable inference accuracy. Tesla P4 delivers a peak of 21.8 INT8 TIOP/s (Tera Integer Operations per
Second), while P40 delivers a peak of 47.0 INT8 TIOP/s. This blog only focuses on P40 GPU.
TensorRT
TM
, previously called GIE (GPU Inference Engine), is a high performance deep learning inference
engine for production deployment of deep learning applications that maximizes inference throughput and
efficiency. TensorRT provides users the ability to take advantage of fast reduced precision instructions
provided in the Pascal GPUs. TensorRT v2 supports the INT8 reduced precision operations that are
available on the P40.
Testing Methodology
This blog quantifies the performance of deep learning inference using TensorRT on Dell’s PowerEdge
C4130 server which is equipped with 4 Tesla P40 GPUs. Since TensorRT is only available for Ubuntu OS, all
the experiments were done on Ubuntu. Table 1 shows the hardware and software details. The inference
benchmark we used was giexec in TensorRT sample codes. The synthetic images which were filled with
random non-zero numbers to simulate real images were used in this sample code. Two classic neural
networks were tested: AlexNet (2012 ImageNet winner) and GoogLeNet (2014 ImageNet winner) which
is much deeper and complicated than AlexNet.
We measured the inference performance in images/sec which means the number of images that can be
processed per second. To measure the performance improvement of the current generation GPU P40, we
also compared its performance with the previous generation GPU M40. The most important goal of this
testing is to measure the inference performance in INT8 mode, compared to FP32 mode. P40 uses the
new Pascal architecture and supports the new INT8 instructions. The previous generation GPU M40 uses
Maxwell architecture and does not support INT8 instructions. The theoretical performance of INT8, FP32
in both M40 and P40 is shown in Table 2. We measured the performance FP32 on both devices and both
FP32 and INT8 on the P40.

Summary of content (6 pages)