0

I successfully converted a ssd_mobilenet_v3 model into a .tflite by retraining the model. (previously trained on coco dataset)

But while using this model for object detection to run inference on a single image using the following code:

interpreter = tf.lite.Interpreter(model_path)

I am getting the following error:

ValueError                                Traceback (most recent call last)
<ipython-input-15-e1c9008b610c> in <module>
----> 1 interpreter = tf.lite.Interpreter("/home/sushanth/Documents      /nuts_poc/tflite_od/nam_model_quantized.lite")
  2 interpreter.allocate_tensors()
  3 input_details = interpreter.get_input_details()
  4 output_details = interpreter.get_output_details()
  5 input_tensor_index = interpreter.get_input_details()[0]["index"]

~/.local/lib/python3.7/site-packages/tensorflow/lite/python/interpreter.py   in __init__(self, model_path, model_content)
 75       self._interpreter = (
 76             _interpreter_wrapper.InterpreterWrapper_CreateWrapperCPPFromFile(
---> 77               model_path))
 78       if not self._interpreter:
 79         raise ValueError('Failed to open {}'.format(model_path))

ValueError: Op builtin_code out of range: 117. Are you using old TFLite binary with newer model?Registration failed.

Kindly explain the error and a possible solution.

Tensorflow Version: 1.1.4

OS: Ubuntu 18.04

Python: 3.7

PS: I converted a classifier model (inception_v2) into tflite and used the above code ("interpreter = tf.lite.Interpreter(model_path)") without any error!

Sushanth
  • 2,224
  • 13
  • 29

1 Answers1

1

Update tensorflow version to >=2.0.0