Bloom float16 quantization fail

I try to use quantize_fx to quantize huggingface bloom model to float16.
I use exactly the same way as the torch quantization test/tutorial file. The code runs well except that when I try to print the result type, it still float32 type which means the quantization never happens.
Here is the patch I use:

the result will stay in float32 by default, you will need to check the ops used in the model to see if the model is actually quantized or not, by printing the model itself.