Hi there,
i’m currently trying to perform a quantization-aware fine-tuning of a Resnet50 model on another dataset and experiencing some issues related to the initialization of the training optimizer.
quantize = True returns a quantized model with 8 bit weights. Quantized models only support inference and run on CPUs. So you cannot do QAT on that, since we don’t support training integer weights.
You can follow the tutorial here for details on how to do QAT.
Thank you very much for your response. Does it means that once a model is trained in a quantized-aware manner, it can no longer be modified (say, in a continuous training setting)?