I have trained an object detection model to be used in production for real-time applications. I have the following two options. Can anyone suggest what is the best way to run inference on Jetson Xavier for best performance? Any other suggestions are also welcome.
- Convert the model to ONXX format and use with TensorRT
- Save the model as Torchscript and run inference in C++