I trained a model using mxnet framework. The inference time for the model is ~ 9 milliseconds. The model mainly consists of conv layers and uses depthwise separable convolution.
I want to run that model in browser. I converted the model to ONNX format then from
ONNX -> tensorflow -> tensorflowjs.
The inference time for tensorflowjs model ~129 milliseconds.
Any suggestion to improve the performance for the model?
I have also tried ONNXJS but it seems it still has few bugs.