Hello, I’m a beginner in quantization. When we want to quantize a model, we must specify a qconfig for a model to choose scales and zero-points. For example:
I think that the weight param of QConfig is the observer of the weight tensors. But what about activation? Does activation observer watch the output values of a layer or the input? And it is to choose the scales and zero-points for the input of layers, right?
Thank you. I am confused that the term activation here is the output of an arbitrary layer (conv2d, linear, …) right? Or it must be the output of an activation function?