Hello everyone. I recently use the torch.quantization.quantize_dynamic(model, dtype=torch.qint8) to dynamic quantize my model. When use fbgemm（the default engine） to quantize the model, it works. However, when I change the engine from fbgemm to qnnpack, it has some problems. The way I use qnnpack is like this: model.qconfig = torch.quantization.get_default_qconfig(‘qnnpack’)
torch.backends.quantized.engine = ‘qnnpack’.
qnnpack causes the problem like that:
Has anyone ever done any related work or met the same problem?
by the way, the torch version is 1.5.0a0+b336deb
I’d appreciate if anybody can help me! Thanks in advance!