I have 2 Tensorflow Lite models (they are Yolo V2 Tiny models):
- Model A) Downloaded from the internet, detects and classifies objects with 80 classes. The .tflite files weights 44,9mb.
- Model B) Trained by myself using Darknet, detects and classifies objects with 52 classes. The .tflite files weights 20,8mb. The model is converted to TFLite using Darkflow.
However both on a mobile phone and on a computer model B takes 10x more time to predict than model A (even if model B detects within less classes and its file is lighter). Also, models seem to work with input images of size 416x416 and use float numbers.
What could be the reason for model A being faster than model B? How can I find out why model A is faster?
One of the problems I have is that for model A, since I have not trained it myself, I don't have its .cfg file with the whole setup...