Model Description

GPUNets are a new family of deployment and production ready Convolutional Neural Networks from NVIDIA auto-designed to max out the performance of NVIDIA GPU and TensorRT.

Crafted by NVIDIA AI using novel Neural Architecture Search(NAS) methods, GPUNet demonstrates state-of-the-art inference performance up to 2x faster than EfficientNet-X and FBNet-V3. This notebook allows you to load and test all the the GPUNet model implementation listed in our CVPR-2022 paper. You can use this notebook to quickly load each one of listed models to perform inference runs.


In the example below the pretrained GPUNet-0 model is loaded by default to perform inference on image and present the result. You can switch the default pre-trained model loading from GPUNet-0 to one of these: GPUNet-1, GPUNet-2, GPUNet-P0, GPUNet-P1, GPUNet-D1 or GPUNet-D2.

Install pre-requisites

To run the example you need some extra python packages installed. These are needed for preprocessing images and visualization.

!pip install validators matplotlib
!pip install timm==0.5.4
import torch
from PIL import Image
import torchvision.transforms as transforms
import numpy as np
import json
import requests
import matplotlib.pyplot as plt
import warnings

%matplotlib inline

if torch.cuda.is_available():
    device = torch.device("cuda") 
    device = torch.device("cpu")

print(f'Using {device} for inference')

Load Pretrained model

Loads NVIDIA GPUNet-0 model by default pre-trained on ImageNet dataset. You can switch the default pre-trained model loading from GPUNet-0 to one of the following models listed below.

The model architecture is visible as output of the loaded model. For details architecture and latency info please refer to architecture section in the original repo and Table#3 in the CVPR-2022 paper, respectively.

Please pick and choose one of the following pre-trained models:

TorchHub model Description
GPUNet-0 GPUNet-0 has the fastest measured latency on GV100
GPUNet-1 GPUNet-1 has improved accuracy with one additional layer on GPUNet-0
GPUNet-2 GPUNet-2 has higher accuracy with two additional layers on GPUNet-0
GPUNet-P0 GPUNet-P0 is the distilled model with higher accuracy than GPUNet-0 but similar latency
GPUNet-P1 GPUNet-P1 is distilled model with even higher accuracy than GPUNet-1 but similar latency
GPUNet-D1 GPUNet-D1 has the second highest accuracy amongst all GPUNets
GPUNet-D2 GPUNet-D2 has the highest accuracy amongst all GPUNets
model_type = "GPUNet-0" # select one from above
precision = "fp32" # select either fp32 of fp16 (for better performance on GPU)

gpunet = torch.hub.load('NVIDIA/DeepLearningExamples:torchhub', 'nvidia_gpunet', pretrained=True, model_type=model_type, model_math=precision)
utils = torch.hub.load('NVIDIA/DeepLearningExamples:torchhub', 'nvidia_convnets_processing_utils')

Prepare inference data

Prepare sample input data for inference.

uris = [

batch =
    [utils.prepare_input_from_uri(uri) for uri in uris]

if precision == "fp16":
    batch = batch.half()
print("Ready to run inference...")

Run inference

Use pick_n_best(predictions=output, n=topN) helper function to pick N most probable hypotheses according to the model.

with torch.no_grad():
    output = torch.nn.functional.softmax(gpunet(batch), dim=1)
results = utils.pick_n_best(predictions=output, n=5)

Display result

for uri, result in zip(uris, results):
    img =, stream=True).raw)
    img.thumbnail((256,256), Image.ANTIALIAS)


For detailed information on model input and output, training recipies, inference and performance visit: github