How to train with fp16 in torch.cuda.amp?

Yes, you won’t need to use autocast, since you are explicitly skipping the mixed-precision training util. Note that pure FP16 training is generally not stable for training.