All modules for which code is available
- torch
- torch._C
- torch._tensor_str
- torch._utils
- torch.autograd
- torch.cuda
- torch.distributed
- torch.distributions
- torch.functional
- torch.multiprocessing
- torch.nn.functional
- torch.nn.init
- torch.nn.modules.activation
- torch.nn.modules.batchnorm
- torch.nn.modules.container
- torch.nn.modules.conv
- torch.nn.modules.distance
- torch.nn.modules.dropout
- torch.nn.modules.instancenorm
- torch.nn.modules.linear
- torch.nn.modules.loss
- torch.nn.modules.module
- torch.nn.modules.padding
- torch.nn.modules.pixelshuffle
- torch.nn.modules.pooling
- torch.nn.modules.rnn
- torch.nn.modules.sparse
- torch.nn.modules.upsampling
- torch.nn.parallel.data_parallel
- torch.nn.parallel.distributed
- torch.nn.parameter
- torch.nn.utils.clip_grad
- torch.nn.utils.rnn
- torch.nn.utils.weight_norm
- torch.onnx
- torch.optim.adadelta
- torch.optim.adagrad
- torch.optim.adam
- torch.optim.adamax
- torch.optim.asgd
- torch.optim.lbfgs
- torch.optim.lr_scheduler
- torch.optim.optimizer
- torch.optim.rmsprop
- torch.optim.rprop
- torch.optim.sgd
- torch.optim.sparse_adam
- torch.random
- torch.serialization
- torch.sparse
- torch.storage
- torch.tensor
- torch.utils.data.dataloader
- torch.utils.data.dataset
- torch.utils.data.distributed
- torch.utils.data.sampler
- torch.utils.ffi
- torch.utils.model_zoo