Connection between pose estimation, epipolar geometry and depth map

asked 2020-07-28 09:04:27 -0600

Hi I am an undergraduate student working on a graduate project, and a beginner to computer vision.

After I went through the tutorial "Camera Calibration and 3D Reconstruction" provided by OpenCV (link) :

I failed to see the connections between the second part to the final part. What I understand here is :

  • The intrinsic and extrinsic parameters of a camera is required to estimate the position of the camera and the captured object
  • To reconstruct a 3D model multiple point clouds are needed, and to generate a point cloud a disparity map is required.

What I do not understand is :

  • The importance of estimating the position of the camera or the object to compute the epiline or epipole in either image planes.
  • The importance of epipolar geometry, and finding out the location of epiline and epipole to compute the disparity map.

As far as I am aware, the code below generate a disparity map

stereo = cv2.createStereoBM(numDisparities=16, blockSize=15)
disparity = stereo.compute(imgL,imgR)

and the input includes a pair of stereo images, minDisparities, numDisparities and blockSize, but not the position of the camera nor the epiline/epipole.

Any help would be greatly appreciated.

edit retag flag offensive close merge delete