its just our old directory. using ao.quantization is the recommended folder. You can directly replace all instances of torch.quantization with torch.ao.quantization. (note the reverse doesn’t work, most new functions/classes developed after the move weren’t duplicated to torch.quantization.
Our team started off doing only quantization and we were granted the top level torch.quantization folder, then we started doing pruning and sparsity and developed various numerical debugging tools. to avoid needing 5 top level folders, we moved everything to the torch.ao (architecture optimization) top level folder with quantization being a subfolder of it. We left the torch.quantization there but it just imports things from torch.ao.quantization to maintain BC, though its temporary and will eventually be removed.
Thank you very much for your reply. well I think what you describe is the difference between torch.quantization and torch.ao.quantization, which is similar in this answer. I am not sure that torch.quantization/torch.ao.quantization and pytoch-quantization whether be two different packages(the former come from pytorch, the letter come from nvidia?), they can both be used to quantization, just api not common?
Oh sorry, yeah i misunderstood. the pytorch-quantization tool in TensorRT/tools is something made by NVIDIA to simulate quantized numerics. It doesn’t look like its been updated recently so i’d assume its a bit outdated. However, the TensorRT library is useful for lowering to different backends and is something that our ao quantization APIs will be using under the hood in certain situations.