About the quantization category
|
|
0
|
1802
|
October 2, 2019
|
Torchscript with dynamic quantization produces inconsistent model outputs in Python and Java
|
|
8
|
201
|
September 29, 2023
|
Quantization of multi_head_attention_forward
|
|
5
|
82
|
September 28, 2023
|
Is bias quantized while doing pytorch static quantization?
|
|
18
|
2494
|
September 27, 2023
|
Quantized model profiling
|
|
12
|
334
|
September 26, 2023
|
Pointers to bring quantized models to device
|
|
8
|
534
|
September 26, 2023
|
Model size reduction
|
|
1
|
29
|
September 26, 2023
|
Understanding differences in the default qconfig for fbgemm and qnnpack
|
|
15
|
366
|
September 22, 2023
|
What do [De]QuantStub actually do?
|
|
8
|
4248
|
September 21, 2023
|
The ONNX exported by my QAT quantization training does not have a fake operator, the code is as follows?
|
|
4
|
80
|
September 19, 2023
|
RFC-0019 Extending PyTorch Quantization to Custom Backends
|
|
2
|
41
|
September 19, 2023
|
Quantizion partially applied to a PyTorch model
|
|
1
|
52
|
September 15, 2023
|
Quantization official example
|
|
5
|
82
|
September 15, 2023
|
LSTM Quantization
|
|
1
|
56
|
September 14, 2023
|
Can the output of operator QuantizedConv2d is fp32?
|
|
4
|
96
|
September 12, 2023
|
Questions about preparing QAT model
|
|
6
|
245
|
September 12, 2023
|
Can we use int8 activation quantization in pytorch
|
|
3
|
385
|
September 1, 2023
|
Cannot import name 'QuantStub' from 'torch.ao.quantization'
|
|
5
|
7675
|
August 31, 2023
|
Expected INT8 Accuracies on ImageNet-1K (ResNet QAT)
|
|
2
|
64
|
August 28, 2023
|
RuntimeError: Quantized cudnn conv2d is currenty limited to groups = 1; received groups =16 , during QAT
|
|
1
|
77
|
August 28, 2023
|
How to export a correct quantized model to onnx format
|
|
3
|
163
|
August 28, 2023
|
PyTorch Dynamic Quantization clarification
|
|
6
|
126
|
August 25, 2023
|
How to quantize a model with both CNN and LSTM
|
|
6
|
864
|
August 22, 2023
|
Qnnpack using activation dtype int8 is not runnable
|
|
16
|
210
|
August 22, 2023
|
RuntimeError: Could not run 'quantized::conv2d_relu.new' with arguments from the 'CPU' backend
|
|
1
|
3415
|
December 17, 2020
|
Could not run 'aten::_slow_conv2d_forward' with arguments from the 'QuantizedCPU' backend
|
|
1
|
83
|
August 14, 2023
|
Error trying to quantize Transformer model
|
|
0
|
103
|
August 12, 2023
|
How to train non quantized layers of quantized model on GPU
|
|
0
|
78
|
August 12, 2023
|
How to implement fp16 quantization on CPU
|
|
2
|
177
|
August 8, 2023
|
Using quantizable model for normal training
|
|
5
|
156
|
July 31, 2023
|