I am developing a Hololens application that requires me to determine the user's eye position from the front-facing camera's captured image. This information is crucial in identifying the real-world object that the user is looking at. I have tried two existing methods but encountered some challenges:
First Method: I extracted
GazeOrigin
andGazeDirection
from the camera's image and calculated the intersection point of that ray with the viewport. However, I encountered problems due to differences in intrinsic and extrinsic properties between the front-facing camera and the FOV (field of view) camera.Second Method: I calculated the 3D position of the intersection point between
GazeDirection
andSpatialMapping
mesh and projected this point onto the front-facing camera and FOV. This method does not work well with dynamic objects such as humans or books held in front of the user.
I am looking for alternative solutions that can handle dynamic objects or improvements to the above methods that can effectively solve my problem. Any insights or suggestions would be appreciated.