Unable to convert a quantized model to TorchScript
|
|
2
|
414
|
November 10, 2023
|
List of operations supported by QuantizedCPU backend
|
|
2
|
297
|
November 10, 2023
|
Is it possible to do int16 qat?
|
|
1
|
374
|
November 10, 2023
|
Is the Static Quantization(PTSQ) model performs Integer-Arithmetic-Only Inference?
|
|
1
|
298
|
November 10, 2023
|
How does quantized conv2d handle scale and zero_point?
|
|
5
|
1289
|
November 9, 2023
|
Accuracy drop after model quantization
|
|
3
|
292
|
November 4, 2023
|
What's the supported datatype for activation in torch.ao.nn.quantized.linear?
|
|
2
|
289
|
November 4, 2023
|
Accuracy drop after prepare_qat_fx with no quantization
|
|
2
|
279
|
November 1, 2023
|
Why does an unsigned torch.quint8 tensor have a sign?
|
|
6
|
415
|
November 1, 2023
|
Quantization of ssdlite from torchvision
|
|
1
|
235
|
October 26, 2023
|
RuntimeError: Unsupported qscheme: per_channel_affine during fx qat
|
|
8
|
880
|
October 17, 2023
|
Static quantization for Transformer block : AttributeError 'function' object has no attribute 'is_cuda'
|
|
5
|
533
|
October 13, 2023
|
Quantization official example
|
|
11
|
489
|
October 13, 2023
|
Quantization of a single tensor
|
|
3
|
357
|
October 12, 2023
|
Custom Quantization using PT2 -> q/dq representation
|
|
1
|
254
|
October 12, 2023
|
Quantization of a pytorch model
|
|
1
|
301
|
October 12, 2023
|
Mixed Precision Training and Quantisation Aware Training together
|
|
1
|
289
|
October 12, 2023
|
How to let the two input nodes of `add` op share the same quantization params(scale and zero point)?
|
|
2
|
396
|
October 9, 2023
|
Quantization of multi_head_attention_forward
|
|
11
|
1021
|
October 4, 2023
|
Error trying to quantize Transformer model
|
|
2
|
456
|
October 3, 2023
|
How to train non quantized layers of quantized model on GPU
|
|
1
|
298
|
October 3, 2023
|
Torchscript with dynamic quantization produces inconsistent model outputs in Python and Java
|
|
9
|
579
|
October 3, 2023
|
Is bias quantized while doing pytorch static quantization?
|
|
18
|
3651
|
September 27, 2023
|
Quantized model profiling
|
|
12
|
655
|
September 26, 2023
|
Pointers to bring quantized models to device
|
|
8
|
1034
|
September 26, 2023
|
Model size reduction
|
|
1
|
301
|
September 26, 2023
|
Understanding differences in the default qconfig for fbgemm and qnnpack
|
|
15
|
917
|
September 22, 2023
|
What do [De]QuantStub actually do?
|
|
8
|
5389
|
September 21, 2023
|
The ONNX exported by my QAT quantization training does not have a fake operator, the code is as follows?
|
|
4
|
455
|
September 19, 2023
|
RFC-0019 Extending PyTorch Quantization to Custom Backends
|
|
2
|
261
|
September 19, 2023
|