About INT8 and UINT8 #432
-
Excause me. I think your quantized prune yolo5 model is trained with uint8 weights and int32 bias. |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments
-
Hello, is this problem solved somehow? |
Beta Was this translation helpful? Give feedback.
-
Hi @hunterchenghx @rui-shen-afk weight qconfig properties can be modified directly using Hope this helps, let me know if you have any other questions. |
Beta Was this translation helpful? Give feedback.
Hi @hunterchenghx @rui-shen-afk weight qconfig properties can be modified directly using
weight_config_kwargs
, however for tensorrt support, the current best pathway is to settensorrt: True
for anyQuantizationModifier
in your recipe.sparseml/src/sparseml/pytorch/sparsification/quantization/modifier_quantization.py
Line 142 in d845052
Hope this helps, let me know if you have any other questions.