2018-09-28 16:38:53 -0600 | commented question | OpenCV Manager: Item not Found I have not, but as i understand this is only good for development. Once I want to deploy to the store, I need each user |
2018-09-04 12:05:10 -0600 | received badge | ● Student (source) |
2018-09-04 12:00:48 -0600 | asked a question | OpenCV Manager: Item not Found OpenCV Manager: Item not Found The OpenCV docs says: In most cases OpenCV Manager may be installed automatically f |
2016-09-27 13:59:18 -0600 | commented question | Cross Compile OpenCV 2.4.4 @Angeloc: did you ever figure this out? How did you find the *.pc files, and use them? where would they be on the device? Thanks |
2016-01-13 10:47:19 -0600 | commented question | 3D Reconstruction with 1 camera OK, I got it... I was going to answer my own question but I can't post. I understand why this does not work: the vanishing point of the cameras is supposed to be at infinity, while mine is at the object rotation center and the stereoRctify can't transform that. Also, the disparity map assume a horizontal shift to recalculate Z, and I really only have a rotation. So, I get it, this won't work. I'll work on a rig with translation. I assumed my rotation implied the translation I needed of the camera, but the vanishing point issue makes it unworkable. |
2016-01-12 13:40:33 -0600 | commented question | 3D Reconstruction with 1 camera @Edouardo I do not understand your comment about my translation not being right because I always have a distance of 1: I only have 2 images, so why do you say always? Then, in your flow, you say 'rectify the left and right images to have a fronto-parallel view': This seems to be the step I am missing, so how would I do that? Thanks for help. |
2016-01-12 12:38:33 -0600 | commented question | 3D Reconstruction with 1 camera For stereoRectify, the docs says the Rotation is the rotation between the coordinate systems of the cameras. And translation is translation between the 2 cameras. This is what I am doing. Now if the disparity map process expects the images to be taken from a fronto-parallel setup that's a different story: that means I should account for some 'distortion' when calculating that, is that right? I had calibrated my camera on a chessboard pointed towards the camera, and used that for both distortion in the stereoRectify, but maybe I need to calibrate the 2 views by rotating the chessboard? I'm starting to wonder if this can even work with rotation; I guess I should 'translate' the camera then to simulate a front-parallel setup. |
2016-01-12 12:30:07 -0600 | commented question | 3D Reconstruction with 1 camera You say the views are fronto-parallel, so what does that mean for the Rotation matrix? Does that mean there is NO rotation between the cameras, only translation? In my case since I rotate the object, the 2 cameras point at the same point, which is also the axis of rotation of the object. This is why I used the Rotation matrix as I explained. If this is not correct, what should I do? |
2016-01-12 11:18:03 -0600 | received badge | ● Editor (source) |
2016-01-12 10:56:39 -0600 | commented question | 3D Reconstruction with 1 camera From what I have been reading, the rotation and translation matrices are unit-less. Correct me if I am wrong. So the distance to the object is constant, yes, and unit 1. I did try adding a multiplication factor to account for the distance to the object, but all that seems to do is change the scale (in x,y and z) not change the z depth. If I scale the disparity map, it does change the z scale but does not seem to map right. So i'm confused about what I am supposed to adjust, or even if it is possible to get reasonable z depth with just 2 images. |
2016-01-12 00:21:34 -0600 | commented question | 3D Reconstruction with 1 camera I have only used 2 images right now. I understand that if I wanted to move around the object 360degrees I would need to move each frame by the exact same angle, but would be the same if I was to move the camera. But with 2 images, I only need to know one angle. My question really is: is 2 frame enough to get reasonable Z, or am I supposed to go 360 around? If I have to take multiple frames, how am I supposed to match the Z from one reprojection to another? |
2016-01-11 23:47:38 -0600 | commented question | 3D Reconstruction with 1 camera Well, it's all relative, right? The camera rotate around the object or the object rotates in front of the camera. I use a green screen to remove background, so effectively I rotate the object on itself, but in fact it is as if the camera moved around the object, by the same angle, at a given distance from the object. |
2016-01-11 20:51:48 -0600 | asked a question | 3D Reconstruction with 1 camera Playing with stereo 3D reconstruction using one camera and rotating the object, but I can't seem to get proper Z mapping. EDITED following comments: I follow the typical process:
All this seems fine to me. The images look good. I get the disparity map with There is some black areas but mostly gray, so the Z seems to be computed for most of the image. then I use I use a Rotation matrix which I built using The Rotation matrix seems right as far as I can tell: I tried with trivial angles and I get what is expected. I also need the translation vector, so I used I use When i So, I'm just wondering if this is normal, and expected because it's only 2 images from a ~20degree angle difference, or if I'm just messing up somewhere. Especially I wonder:
|