i wanna make a quantization model, so i followed FX GRAPH MODE POST TRAINING STATIC QUANTIZATION tutorial on my resnet50 model.
so I checked my model weight capacity reduced.
here is my question.
After quantizing the trained model, do I need to convert it into int8 and put it in when inference?