m-chenie
(Miranda Chen)
1
Hi there,
I see in the wiki about Quantized Tensors Introducing Quantized Tensor · pytorch/pytorch Wiki · GitHub, it mentions that there are plans to support int16 quantization. Although, that page was last edited in 2020.
Are there any updates on support for int16 quantization?
Thanks,
Miranda
HDCharles
(Hd Charles)
2
no if you wanted to do that you’d normally just use bf16 or fp16 quantization. There are no plans to support it in pytorch at present.
m-chenie
(Miranda Chen)
4
Is there a way to do custom quantization so I can quantize to int16?
jerryzh168
(Jerry Zhang)
5
yeah please check out our new flow: Quantization — PyTorch main documentation
which backend do you want to deploy to?