Suppose you find an object in the first two frames. From this information, you can extrapolate where you expect the object in the third frame. Instead of using the universal find-the-object algorithm, you can use a slower, more complex (and therefore reliably more reliable) algorithm, limiting it to checking in the neighborhood that extrapolation predicts. Perhaps this is not quite what you expect (perhaps the speed vector is changing), but you can certainly reduce the area that has been tested.
This should help reduce the number of cases where any other part of the frame is mistakenly identified as an object (because you are looking at a smaller part of the frame and because you are using a more efficient detector).
Update the extrapolations based on what you find and iterate for the next frame.
If the object is out of scope, you return to your common feature detector, as is the case with the first two frames, and try to get a โlockโ when the object returns to the view.
Also, if you can, throw as much light as possible into the physical scene. If the scene is dim, the webcam will use a longer exposure time, which will lead to more motion blur on moving objects. Motion blur can make it very difficult for a function detector (although it can give you information about direction and speed).
source share