I have an image containing:
- a set of coordinates that act as orientation markers.
- a set of coordinates containing data.
Let's call this image A.
This image is fed to a scanner that returns a copy of the image with certain transformations applied (rotation, scale, translation). Let's call the transformed image B. The transformation values applied are not know to me, of course.
Once I receive the transformed image (B), I can easily track the coordinates of the orientation markers and calculate the angle of rotation, scale (x,y) and translation (x,y).
Now I need to retrieve the data coordinates since I already know the transformed orientation coordinates.
If a data point was at location (10, 10) in image A, where would it be in image B? Given that all three transformations are known.
When I apply a simple matrix transformation the transformed data points I calculate are inaccurate. I tried changing the order of transformations but that seems to have absolutely no effect.
What am I doing wrong? Is it the order/sequence of transformations or something else that I'm missing?
EDIT
Please refer to this question for context.