So, imagine that the camera is looking at the screen of your computer. What I'm trying to do is determine how far this camera is rotated, how far it is on the screen, and where it is relative to the center of the screen. In short, rotation and translation matrices.
I use opencv for this and follow the example of their camera calibration in order to accomplish this task with a checkerboard pattern and a frame from a webcam. I would like to do with any generic images, namely with a screen cover and a frame from a webcam.
I tried using function detection algorithms to get a list of cue points from both images, and then map these cue points to BFMatcher, but ran into problems. In particular, SIFT does not match key points correctly, and SURF does not find key points correctly in a scaled image.
Is there an easier solution to this problem? I feel that this would be a common thing that people did, but did not find much discussion about it on the Internet.
Thanks!!
source share