I would like to experiment with machine learning (especially CNNs) on the aligned RGB and depth stream of either an Intel RealSense or an Orbbec Astra camera. My goal is to do some object recognisation and highlight/mark them in the output video stream (as a starting point).
But after having read many articles I am still confused about the involved frameworks and how the data flows from the camera through the involved software components. I just can't get a high level picture.
This is my assumption regarding the processing flow:
Sensor => Driver => libRealSense / Astra SDK => TensorFlow
Questions
- Is my assumption correct regarding the processing?
- Orbbec provides an additional
Astra OpenNI SDK
besides theAstra SDK
where as Intel has wrappers (?) forOpenCV
andOpenNI
. When or why would I need this additional libraries/support? - What would be the quickest way to get started? I would prefer C# over C++