Deep Learning GPU Benchmarks

Overview of the benchmarked GPUs

Although we only tested a small selection of all the available GPUs, we think we covered all GPUs that are currently best suited for deep learning training, model fine tuning and inference tasks due to their compute and memory capabilities and their compatibility to current deep learning frameworks, namely Pytorch and Tensorflow.

For reference also the iconic deep learning GPUs: Geforce GTX 1080 Ti, RTX 2080 Ti, RTX 3090 and Tesla V100 are included to visualize the increase of compute performance over the recent years.

GTX 1080TI

Suitable for: Workstations

Launch Date: 2017.03

Architecture: Pascal

VRAM Memory (GB): 11 (GDDR5X)

Cuda Cores: 3584

Tensor Cores: -

Power Consumption (Watt): 250

Memory Bandwidth (GB/s): 484

Geforce RTX 2080TI

Suitable for: Workstations

Launch Date: 2018.09

Architecture: Turing

VRAM Memory (GB): 11 (DDR6)

Cuda Cores: 5342

Tensor Cores: 544

Power Consumption (Watt): 260

Memory Bandwidth (GB/s): 616

QUADRO RTX 5000

Suitable for: Workstations/Servers

Launch Date: 2018.08

Architecture: Turing

VRAM Memory (GB): 16 (GDDR6)

Cuda Cores: 3072

Tensor Cores: 384

Power Consumption (Watt): 230

Memory Bandwidth (GB/s): 448

Geforce RTX 3090

Suitable for: Workstations

Launch Date: 2020.09

Architecture: Ampere

VRAM Memory (GB): 24 (GDDR6X)

Cuda Cores: 10496 Tensor Cores: 328

Power Consumption (Watt): 350

Memory Bandwidth (GB/s): 936

RTX A5000

Suitable for: Workstations/Servers 

Launch Date: 2021.04 

Architecture: Ampere 

VRAM Memory (GB): 24 (GDDR6) 

Cuda Cores: 8192 

Tensor Cores: 256 

Power Consumption (Watt): 230 

Memory Bandwidth (GB/s): 768

RTX A5500

Suitable for: Workstations/Servers 

Launch Date: 2022.03 

Architecture: Ampere 

VRAM Memory (GB): 24 (GDDR6) 

Cuda Cores: 10240 

Tensor Cores: 220 

Power Consumption (Watt): 230 

Memory Bandwidth (GB/s): 768

RTX 5000 Ada

Suitable for: Servers

Launch Date: 2023.08

Architecture: Ada Lovelace

VRAM Memory (GB): 32 (GDDR6)

Cuda Cores: 12800

Tensor Cores: 400

Power Consumption (Watt): 250

Memory Bandwidth (GB/s): 576

RTX A6000

Suitable for: Workstations/Servers 

Launch Date: 2020.10 

Architecture: Ampere 

VRAM Memory (GB): 48 (GDDR6) 

Cuda Cores: 10752 

Tensor Cores: 336 

Power Consumption (Watt): 300 

Memory Bandwidth (GB/s): 768

AMD Instinct MI100

Suitable for: Servers

Launch Date: 2020.11

Architecture: CDNA (1)

VRAM Memory (GB): 32 (HBM2)

Stream Processors: 7680

Power Consumption (Watt): 250

Memory Bandwidth (TB/s): 1.2

Geforce RTX 4090

Suitable for: Workstations 

Launch Date: 2022.10 

Architecture: Ada Lovelace 

VRAM Memory (GB): 24 (GDDR6X) 

Cuda Cores: 16384 

Tensor Cores: 512 

Power Consumption (Watt): 450 

Memory Bandwidth (GB/s): 1008

RTX 6000 Ada

Suitable for: Workstations/Servers 

Launch Date: 2022.09 

Architecture: Ada Lovelace 

VRAM Memory (GB): 48 (GDDR6) 

Cuda Cores: 18176 

Tensor Cores: 568 

Power Consumption (Watt): 300 

Memory Bandwidth (GB/s): 960

NVIDIA L40S

Suitable for: Servers

Launch Date: 2022.09

Architecture: Ada Lovelace

VRAM Memory (GB): 48 (GDDR6)

Cuda Cores: 18176

Tensor Cores: 568

Power Consumption (Watt): 300

Memory Bandwidth (GB/s): 864

Tesla V100

Suitable for: Servers 

Launch Date: 2017.05 

Architecture: Volta 

VRAM Memory (GB): 16 (HBM2) 

Cuda Cores: 5120 

Tensor Cores: 640 

Power Consumption (Watt): 250 

Memory Bandwidth (GB/s): 900

A100

Suitable for: Servers 

Launch Date: 2020.05 

Architecture: Ampere 

VRAM Memory (GB): 40/80 (HBM2) 

Cuda Cores: 6912 

Tensor Cores: 512 

Power Consumption (Watt): 300 

Memory Bandwidth (GB/s): 1935 (80 GB PCIe)

H100

Suitable for: Servers 

Launch Date: 2022.10 

Architecture: Grace Hopper 

VRAM Memory (GB): 80 (HBM2) 

Cuda Cores: 14592 

Tensor Cores: 456 

Power Consumption (Watt): 350 

Memory Bandwidth (GB/s): 2000

The Deep Learning Benchmark

The visual recognition ResNet50 model (version 1.5) is used for our benchmark. As the classic deep learning network with its complex 50 layer architecture with different convolutional and residual layers, it is still a good network for comparing achievable deep learning performance. As it is used in many benchmarks, a close to optimal implementation is available, driving the GPU to maximum performance and showing where the performance limits of the devices are.

The comparison of the GPUs have been made using synthetic random image data, to minimize the influence of external elements like the type of dataset storage (SSD or HDD), data loader and data format.

Regarding the setup used, we have to remark two important points. The first one is the XLA feature. A Tensorflow performance feature that was declared stable a while ago, but is still turned off by default. XLA (Accelerated Linear Algebra) does optimize the network graph by dynamically compiling parts of the network to kernels optimized for the specific device. This can have performance benefits of 10% to 30% compared to the static crafted Tensorflow kernels for different layer types. This feature can be turned on by a simple option or environment flag and maximizes the execution performance.

The second one is the employment of mixed precision. Concerning inference jobs, a lower floating point precision is the standard way to improve performance. For most training situation float 16bit precision can also be applied for training tasks with neglectable loss in training accuracy and can speed-up training jobs dramatically. Applying float 16bit precision is not that trivial as the model layers have to be adjusted to use it. As not all calculation steps should be done with a lower bit precision, the mixing of different bit resolutions for calculation is referred as mixed precision.

The Python scripts used for the benchmark are available on Github here.

The Testing Environment

As AIME offers server and workstation solutions for deep learning tasks, we used our AIME A4000 server and our AIME G400 Workstation for the benchmark.

The AIME A4000 server and AIME G400 workstation are elaborated environments to run high performance multiple GPUs by providing sophisticated power and cooling, necessary to achieve and hold maximum performance and the ability to run each GPU in a PCIe 4.0 x16 slot directly connected to the CPU.

The technical specs to reproduce our benchmarks are:

For server compatible GPUs: AIME A4000, AMD EPYC 7543 (32 cores), 128 GB ECC RAM

For GPUs only available for workstations: G400, AMD Threadripper Pro 5955WX (16 cores), 128 GB ECC RAM

Using the AIME Machine Learning Container (MLC) management framework with the following setup:

  • Ubuntu 20.04
  • NVIDIA driver version 520.61.5
  • CUDA 11.2
  • CUDNN 8.2.0
  • Tensorflow 2.9.0 (official build)

The NVIDIA H100, RTX 6000 Ada, L40S, RTX 4090,  RTX 5000 Ada, RTX A6000, RTX A5500 and RTX A5000 were tested with the following setup:

  • CUDA 11.8
  • CUDNN 8.6.0
  • Tensorflow 2.13.1 (official build)

The AMD GPU in the benchmark, the AMD Instinct MI100 was tested with:

  • ROCM 5.4
  • MIOpen 2.19.0
  • Tensorflow 2.11.0 (AMD build)

Single GPU Performance

The results of our measurements is the average of images per second that could be trained while running for 50 steps at the specified batch size. The average of three runs were taken, the start temperature of all GPUs was below 50° Celsius.

The GPU speed-up compared to a 32-core-CPU rises here several orders of magnitude, making GPU computing not only feasible but mandatory for high performance deep learning tasks.

Next the results using mixed precision.

One can see that using mixed precision option can increase the performance up to three times.

Multi GPU Deep Learning Training Performance

The next level of deep learning performance is to distribute the work and training loads across multiple GPUs. The AIME A4000 and the AIME G400 support up to four server capable GPUs.

Deep learning does scale very well across multiple GPUs. The method of choice for multi GPU scaling is to spread the batch across the GPUs. Therefore the effective (global) batch size is the sum of the local batch sizes of each GPU in use. Each GPU does calculate the back propagation for the applied inputs of the batch slice. The back propagation results of each GPU are then summed and averaged. The weights of the model are adjusted accordingly and have to be distributed back to all GPUs.

Concerning the data exchange, there is a peak of communication happening to collect the results of a batch and adjust the weights before the next batch can be calculated. While the GPUs are working on calculation a batch not much or no communication at all is happening across the GPUs.

In this standard solution for multi GPU scaling one has to make sure that all GPUs run at the same speed, otherwise the slowest GPU will be the bottleneck for which all GPUs have to wait for! Therefore mixing of different GPU types is not useful.

The next two graphs show how well the RTX A5000 scales by using single and mixed precision.

A good linear and constant scale factor of around 0.93 is reached, meaning that each additional GPU add around 93% of its theoretical linear performance. The similar scale factor is obtained employing mixed precision.

With the RTX 4090 we have a different situation though. While the single GPU performance is decent, the multi-GPU performance of the RTX 4090 is falling short of expectations. As shown in the diagram below, the scale factor of the second RTX 4090 is only 0.76 - not enough for a reasonable Multi-GPU setup.

The reason here fore is probably an intended market segmentation by NVIDIA to separate the Pro-GPUs from the cheaper NVIDIA GeForce 'consumer' series not to be used in Multi-GPU setups.

Conclusions

Mixed Precision can speed-up the training by more than factor 2

A feature definitely worth a look in regards of performance is to switch training from float 32 precision to mixed precision training. Getting a performance boost by adjusting software depending on your constraints could probably be a very efficient move to double the performance.

Multi GPU scaling is more than feasible

Deep Learning performance scales well with multi GPUs for at least up to 4 GPUs: 2 GPUs can often outperform the next more powerful GPU model in regards of price and performance.

Mixing of different GPU for Multi-GPU setups is not useful

The slowest GPU will be the bottleneck for which all GPUs have to wait for!

The best GPU for Deep Learning?

As in most cases there is no simple answer to that question. Performance is for sure the most important aspect of a GPU used for deep learning tasks but not the only one.

So it highly depends on your requirements. Here are our assessments for the most promising deep learning GPUs:

RTX A5000

The RTX A5000 is a good entry card for deep learning training and inference tasks. It has a very good energy efficiency with a similar performance as the legendary but more power hungry, graphic card flagship of the NVIDIA Ampere generation, the RTX 3090.

It is also not to underestimate: a 4x RTX A5000 GPU setup can, as seen in the diagrams above, outperform a more than 10 times as expensive single NVIDIA H100 80GB accelerator in performance, possible batch size and acquisition costs!

RTX A6000

The bigger brother of the RTX A5000 and an efficient solution to get 48 GB GDDR6 GPU memory on a single card. Especially for inference jobs needing a larger memory configuration the RTX A6000 is an interesting option.

RTX 4090

The first NVIDIA GPU of the Ada Lovelace generation. The single GPU performance is outstanding due its high power budget of 450 Watts but the multi-GPU performance is falling short of expectations as shown above. Beside inference jobs, where a single GPU can deliver its full performance and no communication between the GPUs is necessary, a multi-GPU setup with more than two RTX 4090s doesn't seem to be an efficient way to scale.

RTX 6000 Ada / L40S

The pro versions of the RTX 4090 with double the GPU memory (48 GB) and very solid performance with moderate power requirements. The fastest scalable all-round GPU with best top performance / price ratio, also capable to be used in large language model setups.

The RTX 6000 Ada is the fastest available card for Multi-GPU workstation setups.

On first look the L40S seems to be the server only passive cooled version of the RTX 6000 Ada, since the GPU processor specifications appear to be the same. But it has a detail disadvantage: about 10% lower memory bandwidth, affecting its possible performance in this benchmark.

RTX 5000 Ada

The RTX 5000 Ada is one of the later additions to the NVIDIA Ada Love Lace series and is positioned as successor of the RTX A5000. It is a good replacement but doesn't offer a significant performance upgrade to the RTX A5000. The biggest disadvantage of the RTX 5000 Ada is its lower memory bandwidth of 576 GB/s. On the plus side it has 32 GB GDDR6 memory which could enable larger batch size in some use cases.

NVIDIA A100 40GB / 80GB

The first dedicated deep learning accelerator from NVIDIA to deliver highest performance density. The 40 GB models are not produced anymore and hardly available but the 80GB model is still going strong.

The 80GB HBM2 memory is the key value for scaling with larger models.

The lower power consumption of 250/300 Watt in the PCIe version compared to the 400 Watt SXM version of DGX/HGX servers with comparable performance under sustained load the difference in energy and cooling costs can become a factor to consider.

An octa (8x) NVIDIA A100 setup, like possible with the AIME A8000, catapults one into multi petaFLOPS HPC computing area.

NVIDIA H100

In case the most performance regardless of price and highest performance density is needed, the NVIDIA H100 is currently the first choice: it delivers high end deep learning performance.

The biggest advantage to the NVIDIA A100, beside its about 40% higher base performance, is the capability of FP8 computing which can double the compute throughput, especially in inference tasks.

It is also the first PCIe 5.0 capable accelerator, so this card is best suited for PCIe 5.0 ready servers like the AIME A4004 and AIME A8004 to have a comfortable 128 GB/s data interchange between accelerators.

This article will be updated with new additions and corrections as soon as available.

Questions or remarks? Please contact us under: hello@aime.info