Skip to content

LambdaLabsML/DeepLearningExamples

Repository files navigation

NVIDIA Deep Learning Examples for Tensor Cores

Introduction

This repository provides State-of-the-Art Deep Learning examples that are easy to train and deploy, achieving the best reproducible accuracy and performance with NVIDIA CUDA-X software stack running on NVIDIA Volta, Turing and Ampere GPUs.

NVIDIA GPU Cloud (NGC) Container Registry

These examples, along with our NVIDIA deep learning software stack, are provided in a monthly updated Docker container on the NGC container registry (https://ngc.nvidia.com). These containers include:

  • The latest NVIDIA examples from this repository
  • The latest NVIDIA contributions shared upstream to the respective framework
  • The latest NVIDIA Deep Learning software libraries, such as cuDNN, NCCL, cuBLAS, etc. which have all been through a rigorous monthly quality assurance process to ensure that they provide the best possible performance
  • Monthly release notes for each of the NVIDIA optimized containers

Computer Vision

ModelsFrameworkA100AMPMulti-GPUMulti-NodeTRTONNXTritonDLCNB
EfficientNet-B0PyTorchYesYesYes----Yes-
EfficientNet-B4PyTorchYesYesYes----Yes-
EfficientNet-WideSE-B0PyTorchYesYesYes----Yes-
EfficientNet-WideSE-B4PyTorchYesYesYes----Yes-
EfficientNetTensorFlow2YesYesYesYes---Yes-
GPUNetPyTorchYesYesYes-YesYesYesYes-
Mask R-CNNPyTorchYesYesYes-----Yes
Mask R-CNNTensorFlowYesYesYes----Yes-
Mask R-CNNTensorFlow2YesYesYes----Yes-
nnUNetPyTorchYesYesYes----Yes-
ResNet-50MXNet-YesYes------
ResNet-50PaddlePaddleYesYesYes-Yes----
ResNet-50PyTorchYesYesYes-Yes-YesYes-
ResNet-50TensorFlowYesYesYes----Yes-
ResNeXt-101PyTorchYesYesYes-Yes-YesYes-
ResNeXt-101TensorFlowYesYesYes----Yes-
SE-ResNeXt-101PyTorchYesYesYes-Yes-YesYes-
SE-ResNeXt-101TensorFlowYesYesYes----Yes-
SSDPyTorchYesYesYes-----Yes
SSDTensorFlowYesYesYes----YesYes
U-Net IndTensorFlowYesYesYes----YesYes
U-Net MedTensorFlowYesYesYes----Yes-
U-Net 3DTensorFlowYesYesYes----Yes-
U-Net MedTensorFlow2YesYesYes----Yes-
V-Net MedTensorFlowYesYesYes----Yes-

Natural Language Processing

ModelsFrameworkA100AMPMulti-GPUMulti-NodeTRTONNXTritonDLCNB
BERTPyTorchYesYesYesYes--YesYes-
TransformerXLPyTorchYesYesYesYes---Yes-
GNMTPyTorchYesYesYes------
TransformerPyTorchYesYesYes------
ELECTRATensorFlow2YesYesYesYes---Yes-
BERTTensorFlowYesYesYesYesYes-YesYesYes
BERTTensorFlow2YesYesYesYes---Yes-
BioBertTensorFlowYesYesYes----YesYes
TransformerXLTensorFlowYesYesYes------
GNMTTensorFlowYesYesYes------
Faster TransformerTensorflow----Yes----

Recommender Systems

ModelsFrameworkA100AMPMulti-GPUMulti-NodeTRTONNXTritonDLCNB
DLRMPyTorchYesYesYes--YesYesYesYes
DLRMTensorFlow2YesYesYesYes---Yes-
NCFPyTorchYesYesYes------
Wide&DeepTensorFlowYesYesYes----Yes-
Wide&DeepTensorFlow2YesYesYes----Yes-
NCFTensorFlowYesYesYes----Yes-
VAE-CFTensorFlowYesYesYes------
SIMTensorFlow2YesYesYes----Yes-

Speech to Text

ModelsFrameworkA100AMPMulti-GPUMulti-NodeTRTONNXTritonDLCNB
JasperPyTorchYesYesYes-YesYesYesYesYes
Hidden Markov ModelKaldi--Yes---Yes--

Text to Speech

ModelsFrameworkA100AMPMulti-GPUMulti-NodeTRTONNXTritonDLCNB
FastPitchPyTorchYesYesYes----Yes-
FastSpeechPyTorch-YesYes-Yes----
Tacotron 2 and WaveGlowPyTorchYesYesYes-YesYesYesYes-

Graph Neural Networks

ModelsFrameworkA100AMPMulti-GPUMulti-NodeTRTONNXTritonDLCNB
SE(3)-TransformerPyTorchYesYesYes------

NVIDIA support

In each of the network READMEs, we indicate the level of support that will be provided. The range is from ongoing updates and improvements to a point-in-time release for thought leadership.

Glossary

Multinode Training
Supported on a pyxis/enroot Slurm cluster.

Deep Learning Compiler (DLC)
TensorFlow XLA and PyTorch JIT and/or TorchScript

Accelerated Linear Algebra (XLA)
XLA is a domain-specific compiler for linear algebra that can accelerate TensorFlow models with potentially no source code changes. The results are improvements in speed and memory usage.

PyTorch JIT and/or TorchScript
TorchScript is a way to create serializable and optimizable models from PyTorch code. TorchScript, an intermediate representation of a PyTorch model (subclass of nn.Module) that can then be run in a high-performance environment such as C++.

Automatic Mixed Precision (AMP)
Automatic Mixed Precision (AMP) enables mixed precision training on Volta, Turing, and NVIDIA Ampere GPU architectures automatically.

TensorFloat-32 (TF32)
TensorFloat-32 (TF32) is the new math mode in NVIDIA A100 GPUs for handling the matrix math also called tensor operations. TF32 running on Tensor Cores in A100 GPUs can provide up to 10x speedups compared to single-precision floating-point math (FP32) on Volta GPUs. TF32 is supported in the NVIDIA Ampere GPU architecture and is enabled by default.

Jupyter Notebooks (NB)
The Jupyter Notebook is an open-source web application that allows you to create and share documents that contain live code, equations, visualizations and narrative text.

Feedback / Contributions

We're posting these examples on GitHub to better support the community, facilitate feedback, as well as collect and implement contributions using GitHub Issues and pull requests. We welcome all contributions!

Known issues

In each of the network READMEs, we indicate any known issues and encourage the community to provide feedback.

About

Deep Learning Examples

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python50.5%
  • Jupyter Notebook41.6%
  • Shell3.2%
  • C++2.9%
  • Cuda1.3%
  • Makefile0.2%
  • Other0.3%