About the quantization category
|
|
0
|
2427
|
October 2, 2019
|
Quantization of depthwise 1d convolution with QAT is slower than non-quantized
|
|
0
|
4
|
January 21, 2025
|
Taylor-series Approximation for Sigmiod in Integer
|
|
1
|
23
|
January 15, 2025
|
Triton kernel to efficiently dequantize int4
|
|
0
|
32
|
January 5, 2025
|
BatchNorm not fusing with Cone and ReLU
|
|
0
|
12
|
December 26, 2024
|
Compile Model with TensorRT
|
|
0
|
25
|
December 25, 2024
|
How to convert a QAT model to ONNX model
|
|
3
|
94
|
December 19, 2024
|
Pytorch 2 Export QAT is training
|
|
0
|
55
|
December 19, 2024
|
Quantized GLU not implemented?
|
|
1
|
56
|
December 17, 2024
|
Kernel Dies When Testing a Quantized ResNet101 Model in PyTorch
|
|
2
|
17
|
December 12, 2024
|
Auto-cast and pytorch 2 export quantization
|
|
8
|
170
|
December 9, 2024
|
RuntimeError: Quantized cudnn conv2d is currenty limited to groups = 1; received groups =16 , during QAT
|
|
3
|
822
|
December 6, 2024
|
Quantization fails for custom backend
|
|
2
|
68
|
December 6, 2024
|
Support for quantization in int16
|
|
5
|
68
|
December 5, 2024
|
Quantize a single tensor obtained from a float32 model
|
|
2
|
35
|
November 29, 2024
|
Can't get dynamic shape with torch.export.export_for_training
|
|
1
|
41
|
November 28, 2024
|
Simple quantisation reproduction - how to convert state dict to int8
|
|
1
|
36
|
November 27, 2024
|
torch.ao.nn.quantizable.modules.activation.MultiheadAttention not loading the pre-trained model weights correctly
|
|
1
|
29
|
November 27, 2024
|
QConfig for Resnet50 with weights dtype quint8
|
|
5
|
95
|
November 27, 2024
|
Load custom trained parameters into quantized model
|
|
1
|
27
|
November 27, 2024
|
Per channel setting for QAT Quantization
|
|
1
|
24
|
November 27, 2024
|
Custom QAT using ao.nn.qat modules, is this a valid approach?
|
|
1
|
23
|
November 27, 2024
|
Absence of qint32 in torch.ao.quantization.utils.weight_is_quantized
|
|
1
|
58
|
November 27, 2024
|
Questions on QAT for Wav2Vec
|
|
1
|
77
|
November 27, 2024
|
QAT QuantizedConv2d converted to ONNX format
|
|
1
|
57
|
November 27, 2024
|
Changing Qconfig to set datatype to int8
|
|
1
|
119
|
November 20, 2024
|
Inserting Unnecessary Fake Quants during QAT?
|
|
2
|
78
|
November 12, 2024
|
Torch.bfloat16 < how does it work? in bf 16 model
|
|
1
|
111
|
November 4, 2024
|
pytorch quantized linear function gives shape invalid error
|
|
3
|
65
|
November 1, 2024
|
How to lower to target backend?
|
|
3
|
95
|
November 1, 2024
|