I'm running the TfLite model analyzer on Google Brain's AutoML EfficientDet model like so:
model_path = "/content/efficientdet.tflite"
tf.lite.experimental.Analyzer.analyze(model_path=model_path, gpu_compatibility=True)
The analyzer output shows that the last layer is not a supported TfLite GPU operation:
GPU COMPATIBILITY WARNING: Not supported custom op TFLite_Detection_PostProcess
GPU COMPATIBILITY WARNING: Subgraph#0 has GPU delegate compatibility issues at nodes 482 with TFLite runtime version 2.8.2
The TFLite_Detection_PostProcess operator is not supported on the GPU. This is causing CPU/GPU synchronization issues and it is increasing the latency.
What do I need to do to get the op running on the GPU in Tensorflow Lite?
I'm open to re-compiling the TfLite binary or modifying the neural network.