TORCH torch.backends.cudnn.benchmark=True will enable the inbuilt cudnn auto-tuner to find the best algorithm to use for your hardware. This way, cudnn will look for the optimal set of algorithms for that particular configuration (which takes some time). This usually leads to faster runtime. It depends on the task. If your input size is changing a lot, then it might hurt runtime, if not, it should be much faster. https://discuss.pytorch.org/t/what-does-torch-backends-cudnn-benchmark-do/5936/3 [https://hsaghir.github.io/data_science/pytorch_starter/] torch: a general purpose array library similar to Numpy that can do computations on GPU when the tensor type is cast to (torch.cuda.TensorFloat) torch.autograd: a package for building a computational graph and automatically obtaining gradients torch.nn: a neural net library with common layers and cost functions torch.optim: an optimization package with common optimization algorithms like SGD,Adam, etc torch.j...