Hi all, I need help in understanding some of the changes to the model during quantization.
Pre quantization setup:
- Relu6 converted to Relu
- Quantstub and Dequantstub added to the end of the model (after classifier)
(quant): QuantStub()
(dequant): DeQuantStub()
After QAT and convert - Batch norm layers seem to be removed
First block of the model after preprocessing:
(features): Sequential(
(0): Conv2dNormActivation(
(0): Conv2d(3, 32, kernel_size=(3, 3), stride=(2, 2), padding=(1, 1), bias=False)
(1): BatchNorm2d(32, eps=1e-05, momentum=0.1, affine=True, track_running_stats=True)
(2): ReLU()
)
First block of model after quantization
(features): Sequential(
(0): Conv2dNormActivation(
(0): QuantizedConvReLU2d(3, 32, kernel_size=(3, 3), stride=(2, 2), scale=0.030749445781111717, zero_point=0, padding=(1, 1))
(1): Identity()
(2): Identity()
)
What would the reasoning be behind these changes?