Abstract
Previous methods for estimating observer motion in a rigid 3D scene assume that image velocities can be measured at isolated points. When the observer is moving through a cluttered 3D scene such as a forest, however, pointwise measurements of image velocity are more challenging to obtain because multiple depths, and hence multiple velocities, are present in most local image regions. We introduce a method for estimating egomotion that avoids pointwise image velocity estimation as a first step. In its place, the direction of motion parallax in local image regions is estimated, using a spectrum-based method, and these directions are then combined to directly estimate 3D observer motion. There are two advantages to this approach. First, the method can be applied to a wide range of 3D cluttered scenes, including those for which pointwise image velocities cannot be measured because only normal velocity information is available. Second, the egomotion estimates can be used as a posterior constraint on estimating pointwise image velocities, since known egomotion parameters constrain the candidate image velocities at each point to a one-dimensional rather than a two-dimensional space.
© 2005 Optical Society of America
Full Article | PDF ArticleMore Like This
Michael S. Langer and Fahim Mannan
J. Opt. Soc. Am. A 29(9) 1794-1807 (2012)
Steven J. Nowlan and Terrence J. Sejnowski
J. Opt. Soc. Am. A 11(12) 3177-3200 (1994)
M. G. Nagle, M. V. Srinivasan, and D. L. Wilson
J. Opt. Soc. Am. A 14(12) 3233-3241 (1997)