RE: [Vxl-users] Fundamental matrix & normalized coordinates

 RE: [Vxl-users] Fundamental matrix & normalized coordinates From: Oli Cooper - 2005-05-11 10:13:58 Attachments: Message as HTML ```You seem to be a little confused as to the use of homogeneous coordinates and the difference between the Fundamental and Essential matrices. The Hartley and Zisserman book "Multiple View Geometry" should help. It is fine to compute F by setting w=1.0 Normalised coordinates in the sense of the Essential Matrix refers to the coordinates being in the camera coordinate system as opposed to image coordinates. Just to confuse things further, you should ALWAYS normalise the coordinates when computing F by translating them so the centroid of the points is at the origin and scaling them so they are a mean distance of sqrt(2) away. There are methods in the mvl library for normalising coordinates and extracting camera matrices from a computed F. I'm sure the MVG book will explain better than me... Oli. _____ From: vxl-users-admin@... [mailto:vxl-users-admin@...] On Behalf Of Angel Todorov Sent: 10 May 2005 08:11 To: vxl-users@... Subject: [Vxl-users] Fundamental matrix & normalized coordinates Hi, Suppose we have two uncalibrated cameras, and we are processing some image sequence - in the example contrib\oxl\mvl\examples\compute_FMatrix_example.cxx , the point coordinates are normalized with w=1.0, i.e we are basically computing the essential matrix, but, my question is, given that we have no knowledge of the homogeneous coordinates, is it still possible to obtain a meaningful error by computing F (with w=1.0 set), and then try to find the camera matrices by some other method (like bundle adjustment) ? The purpose is to refine matches by fitting them to epipolar lines in the beginning, and I was wondering if this is a sensible approach provided that the matched points' coordinates are normalized, and we know nothing about the camera calibration matrices as well (and we don't do explicit calibration)? In some publications, unknown cameras are also assumed, but it's not very clear to me if, while first computing F or T, coordinates are normalized or not. (Otherwise, as far as I know, we can obtain the essential matrix, if knowing the calibration matrices: E = K'*F*K.) Thank you very much for your help. Regards, Angel ```

 [Vxl-users] Fundamental matrix & normalized coordinates From: Angel Todorov - 2005-05-10 07:10:41 Attachments: Message as HTML ```Hi, Suppose we have two uncalibrated cameras, and we are processing some image sequence - in the example contrib\oxl\mvl\examples\compute_FMatrix_example.cxx , the point coordinates are normalized with w=1.0, i.e we are basically computing the essential matrix, but, my question is, given that we have no knowledge of the homogeneous coordinates, is it still possible to obtain a meaningful error by computing F (with w=1.0 set), and then try to find the camera matrices by some other method (like bundle adjustment) ? The purpose is to refine matches by fitting them to epipolar lines in the beginning, and I was wondering if this is a sensible approach provided that the matched points' coordinates are normalized, and we know nothing about the camera calibration matrices as well (and we don't do explicit calibration)? In some publications, unknown cameras are also assumed, but it's not very clear to me if, while first computing F or T, coordinates are normalized or not. (Otherwise, as far as I know, we can obtain the essential matrix, if knowing the calibration matrices: E = K'*F*K.) Thank you very much for your help. Regards, Angel ```
 RE: [Vxl-users] Fundamental matrix & normalized coordinates From: Oli Cooper - 2005-05-11 10:13:58 Attachments: Message as HTML ```You seem to be a little confused as to the use of homogeneous coordinates and the difference between the Fundamental and Essential matrices. The Hartley and Zisserman book "Multiple View Geometry" should help. It is fine to compute F by setting w=1.0 Normalised coordinates in the sense of the Essential Matrix refers to the coordinates being in the camera coordinate system as opposed to image coordinates. Just to confuse things further, you should ALWAYS normalise the coordinates when computing F by translating them so the centroid of the points is at the origin and scaling them so they are a mean distance of sqrt(2) away. There are methods in the mvl library for normalising coordinates and extracting camera matrices from a computed F. I'm sure the MVG book will explain better than me... Oli. _____ From: vxl-users-admin@... [mailto:vxl-users-admin@...] On Behalf Of Angel Todorov Sent: 10 May 2005 08:11 To: vxl-users@... Subject: [Vxl-users] Fundamental matrix & normalized coordinates Hi, Suppose we have two uncalibrated cameras, and we are processing some image sequence - in the example contrib\oxl\mvl\examples\compute_FMatrix_example.cxx , the point coordinates are normalized with w=1.0, i.e we are basically computing the essential matrix, but, my question is, given that we have no knowledge of the homogeneous coordinates, is it still possible to obtain a meaningful error by computing F (with w=1.0 set), and then try to find the camera matrices by some other method (like bundle adjustment) ? The purpose is to refine matches by fitting them to epipolar lines in the beginning, and I was wondering if this is a sensible approach provided that the matched points' coordinates are normalized, and we know nothing about the camera calibration matrices as well (and we don't do explicit calibration)? In some publications, unknown cameras are also assumed, but it's not very clear to me if, while first computing F or T, coordinates are normalized or not. (Otherwise, as far as I know, we can obtain the essential matrix, if knowing the calibration matrices: E = K'*F*K.) Thank you very much for your help. Regards, Angel ```