### Advice on calibration and triangulation for large scale

Hi, so I am looking at tracking and triangulating the position of a flying target (large portions of the camera view is the sky!) that is a couple of kilometres away. This means that the camera pair used for triangulation has to have quite a big baseline, atleast hundreds of meters for any type of accuracy in the triangulation. I am a bit confused however how to perform this task especially the calibration part.

For calibration of the camera pair I thought I could use block matching and the result to estimate the fundamental matrix (OpenCV findFundamental). Since I know the camera intrinsics I could from this determine the essential matrix and since I also know the translation vector I could calculate the rotation matrix. Finally since this would yield the relation between the cameras, but not the world coordinates final step would be using a couple of known points (about 5-7 in world coordinates) with solvePnP to find one of the cameras orientation relative the world frame. With this I could now form my projection matrices for my cameras used in the triangulation step.

My questions are:

Is the general idea correct or is there something I am missing?

I have also locked at using stereoCalibrate, but my understanding is that it requires the use of a calibration pattern, which in this case is a bit unrealistic given the scale. Or could I use this instead for calibration?

Finally a short question on triangulation. There are multiple algorithms out there, [Triangulation
Richard I. Hartley] my understanding is that L2-norm solutions such as the polynomial algorithm are optimal for projection problems, but for my task of 3d-point accuracy then Linear-LS perform better (see GRAPH 2 from paper)?

edit: Considering the distances involved and the complexity in finding good points for matching, would it be unreasonable to consider simply measuring the approximate pose of the cameras with something like a spirit-level + compass?

Best