Automatic Mixed Precision (AMP)

torch.cuda.amp provides convenience methods for mixed precision, where some operations use the torch.float32 (float) datatype and other operations use torch.float16.

Resources

Mixed precision tries to match each op to its appropriate datatype, which can reduce your network’s runtime and memory footprint.