estimate motion of ROIs in Image
Hello everyone.
I am trying to detect in real time pedestrians with a camera mounted on a moving vehicle. My problem arises in the tracking phase.
Basically, I am using the points in the point cloud obtained from a Lidar sensor to project ROIs in the image, then the detection is actually performed in the image.
The tracking is performed in the ground plane with the point cloud. So , when I get a detection window in the image from the pedestrian detector, the roi which most overlaps that is used as reference point for tracking (I generate a ROI for each point in front of the car so I can recover the corresponding point that generated that roi easily).
The issue I get is that for the Kalman Filter to be efficient and to perform correctly data association I need several detections while my detector provides only few consecutive detections. (For instance a walking pedestrian in front of the still car would lead to three or four detections not consecutive).
For the data association to work properly this is a problem because in that case my Kalman filter, instead of keeping track of the same pedestrian, will instantiate different tracks for every detections if they are too much far one from the others.
To get rid of this problem I thought to use Motion History Images and so, once a detection arises, I start create "dummy detections" to feed the Kalman Filter by using the motion estimated with Motion History Image.
Is that feasible to be done in real time?
Moreover, How could I create the masks that are needed to create the MHI? The only things I am interested in are the motions of the ROIs in the image, not the whole image. The camera I am using is a color camera so I would like to speed up the computation by considering only the rois but the MHI needs the whole image right?
So I am confused about that and it is not clear for me how to use this MHI to do what I want.
Could someone kindly give some hint/feedback?
Thank you