I am trying to quantize an ONNX model using the onnxruntime quantization tool.
My code is below for quantization:
import onnx
from quantize import quantize, QuantizationMode
# Load the onnx model
model = onnx.load('3ddfa_optimized_withoutflatten.onnx')
# Quantize
quantized_model = quantize(model, quantization_mode=QuantizationMode.IntegerOps)
# Save the quantized model
onnx.save(quantized_model, 'quantized_model.onnx')
After this method the model I am getting has 0 dimensional model. What arguments do I have to pass in quantize function so I can get a proper model?