Abstract
As an observer moves and explores the environment, the visual stimulation in his/her eye is constantly changing. Somehow he/she is able to perceive the spatial layout of the scene, and to discern his/her movement through space. Computational vision researchers have been trying to solve this problem for a number of years with only limited success. It is a difficult problem to solve because the optical flow field is nonlinearly related to the 3D motion and depth parameters.
Here, we show that the nonlinear equation describing the optical flow field can be split by an exact algebraic manipulation to form three sets of equations. The first set relates the flow field to only the translational component of 3D motion. Thus, depth and rotation need not be known or estimated prior to solving for translation. Once the translation has been recovered, the second set of equations can be used to solve for rotation. Finally, depth can be estimated with the third set of equations, given the recovered translation and rotation.
The algorithm applies to the general case of arbitrary motion with respect to an arbitrary scene. It is simple to compute, and it is plausible biologically. The results reported in this article demonstrate the potential of our new approach, and show that it performs favorably when compared with two other well-known algorithms.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Adelson, E.H., and Bergen, J.R., 1986. The extraction of spatiotemporal energy in human and machine vision. Proc. IEEE Workshop on Motion: Representation and Analysis, Charleston, S. Carolina, pp. 151–156.
AdivG., 1985. Determining three-dimensional motion and structure from optical flow generated by several moving objects. IEEE Trans. Patt. Anal. Mach. Intell. 7:384–401.
AdivG., 1989. Inherent ambiguities in recovering 3D motion and structure from a noisy flow field. IEEE Trans. Anal. Mach. Intell. 11:477–489.
AggarwalJ.K. and NandhakumarN., 1988. On the computation of motion from sequences of images—a review. Proc. IEEE. 76: 917–935.
AnandanP., 1989. A computational framework and an algorithm for the measurement of visual motion. Intern. J. Comput. Vision. 2:283–310.
BallardD.H. and KimballO.A., 1983. Rigid body motion from depth and optical flow. Comput. Vision, Graph. Image Process 22:95–115.
Barron, J., 1984. A survey of approaches for determining optic flow, environmental layout and egomotion. Techn. Rept. RBCV-TR-84-5, Department of Computer Science, University of Toronto.
BarronJ.L., JepsonA.D., and TsotsosJ.K., 1990. The feasibility of motion and structure from noisy time-varying image velocity information. Intern. J. Comput. Vision 5:239–269.
BollesR.C., BakerH.H., and MarimontD.H., 1987. Epipolarplane image analysis: An approach to determining structure from motion. Intern. J. Comput. Vision 1:7–55.
BroidaT.J., and ChellappaR., 1986. Estimation of object motion parameters from noisy images. IEEE Trans. Patt. Anal. Mach. Intell. 8:90–99.
BrussA.R., and HornB.K.P., 1983. Passive navigation. Comput. Vision, Graph. Image Process. 21:3–20.
Faugeras, O.D., Lustman, F., and Toscani, G., 1987. Motion and structure from motion from point and line matches. Proc. 1st Intern. Conf. Comput. Vision, London, June, pp. 25–34.
FleetD.J. and JepsonA.D., 1990. Computation of component image velocity from local phase information. Intern. J. Comput. Vision. 5:77–104.
GibsonJ.J. 1950. The Perception of the Visual World. Houghton Mifflin, Boston.
GibsonJ.J., and GibsonE.J., 1957. Continuous perspective transformation and the perception of rigid motions. J. Exp. Psychol. 54: 129–138.
GrzywaczN.M., and YuilleA.L., 1990. A model for the estimate of local image velocity by cells in the visual cortex. Proc. Roy. Soc. London A, 239:129–161.
HayJ.C., 1966. Optical motions and space perception: An extension of Gibson's analysis. Psychological Review, 73:550–565.
HeegerD.J., 1987. Model for the extraction of image flow. J. Opt. Soc. Amer. A 4:1455–1471.
HeegerD.J., 1988. Optical flow using spatiotemporal filters. Intern. J. Comput. Vision 1:279–302.
HeegerD.J., and JepsonA., 1990a. Visual perception of three-dimensional motion. Neural Computation 2:129–137.
HeegerD.J., and JepsonA., 1990b. Visual perception of 3D motion and depth. Invest. Opthal. Vis. Sci. Suppl. 31:173.
Heeger, D.J., and Jepson, A., 1990c. Simple method for computing 3D motion and depth. Proc. 3rd. Intern. Conf. Comput. Vision, Osaka, Japan, December, pp. 96–100.
HeegerD.J., and JepsonA., 1991. Recovering observer translation with center-surround motion-opponent mechanisms. Invest. Opthal. Vis. Sci. Suppl. 32:823.
Heel, J., 1989a. Direct estimation of structure and motion for multiple frames. Tech. Rep. 1190, MIT AI Lab.
Heel, J., 1989b. Dynamic motion vision. Proc. SPIE. Philadelphia.
Heel, J., 1990. Direct dynamic motion vision. Proc. IEEE Conf. Robot. Autom. Cincinnati.
HornB.K.P., 1986. Robot Vision. MIT Press: Cambridge, Ma.
HornB.K.P., 1987. Motion fields are hardly ever ambiguous. Intern. J. Comput. Vision, 1:259–274.
HornB.K.P., and NegahdaripourS., 1987. Direct passive navigation: Analytical solution for planes. IEEE Trans. Patt. Anal. Mach. Intell. 9:168–176.
HornB.K.P., and SchunkB.G., 1981. Determining optical flow. Artificial Intelligence 17:185–203.
HornB.K.P., and WeldonE.J., 1988. Direct methods for recovering motion. Intern. J. Comput. Vision 2:51–76.
Jepson, A., and Heeger, D.J., 1989. Egomotion without depth estimation. Optics News 15:A-20.
Jepson, A., and Heeger, D.J., 1990. Subspace methods for recovering rigid motion II: Theory Submitted to International Journal of Computer Vision, available as Tech. Rept. RBCV-TR-90-36, Department of Computer Science, University of Toronto.
Jepson, A. and Heeger, D.J. 1991. A fast subspace algorithm for recovering rigid motion. Proc. IEEE Workshop on Visual Motion, Princeton, N.J., pp. 124–131.
JohanssonG., 1975. Visual motion perception. Scientific American 232: 76–88.
KoenderinkJ.J. and vanDornA.J. 1975. Invariant properties of the motion parallax field due to the movement of rigid bodies relative to an observer. Optica Acta 22: 773–791.
KoenderinkJ.J., and vanDornA.J. 1976. Local structure of movement parallax of the plane. J. Opt. Soc. Amer. 66:717–723.
KoenderinkJ.J., and vanDornA.J. 1981. Exterospecific component of the motion parallax field. J. Opt. Soc. Amer. 71:953–957.
Longuet-HigginsH.C., 1981. A computer algorithm for reconstructing a scene from two projections. Nature 293:133–135.
Longuet-HigginsH.C., 1984. The visual ambiguity of a moving plane. Proc. Roy. Soc. London B 223:165–175.
Longuet-HigginsH.C., 1988. Multiple interpretations of a pair of images of a surface. Proc. Roy. Soc. London B 418:1–15.
Longuet-Higgins, and PrazdnyK. 1980. The interpretation of a moving retinal image. Proc. Roy. Soc. London B 208:385–397.
Lucas, B.D., and Kanade, T., 1981. An iterative image registration technique with an application to stereo vision. Proc. 7th Intern. Joint Conf. Artif. Intell. Vancouver, pp. 674–679.
MaloneyL.T., and WandellB.A., 1986. Color constancy: a method for recovering surface spectral reflectance. J. Opt. Soc. Amer. A 1:29–33.
MatthiesL., SzeliskiR., and KanadeT., 1989. Kalman filter-based algorithms for estimating depth from image sequences. Intern. J. Comput. Vision. 3:209–238.
MaybankS.J., 1985. The angular velocity associated with the optical flow field arising from motion through a rigid environment. Proc. Roy. Soc. London A 410:317–326.
Maybank, S.J., 1987. A Theoretical Study of Optical flow. Ph.D. thesis, University of London.
NagelH.H., 1987. On the estimation of optical flow: relations between different approaches and some new results. Artificial Intelligence 33:299–324.
NakayamaK., 1985. Biological image motion processing: A review. Vision Research 25:625–660.
NegahdaripourS., and HornB.K.P., 1989. A direct method for locating the focus of expansion. Comput. Vision, Graph. Image Process. 46:303–326.
PrazdnyK., 1980. Egomotion and relative depth from optical flow, Biological Cybermetics 36:87–102.
PrazdnyK., 1981. Determining the instantaneous direction of motion from optical flow generated by a curvilinearly moving observer. Comput. Graph. Image Process. 17:238–248.
PrazdnyK., 1983. On the information in optical flows. Comput. Graph. Image Process. 22:239–259.
ReigerJ.H., and LawtonD.T., 1985. Processing differential image motion. J. Opt. Soc. Amer. A. 2:354–359.
RoachJ.W., and AggarwalJ.K., 1980. Determining the movement of objects from a sequence of images. IEEE Trans. Patt. Anal. Mach. Intell. 2:554–562.
SimoncelliE.P., and AdelsonE.H., 1991. Relationship between gradient, spatio-temporal energy, and regression models for motion perception. Invest. Opthal. Vis. Sci. Suppl. 32:893.
Simoncelli, E.R., Adelson, E.H., and Heeger, D.J., 1991. Probability distributions of optical flow. Proc. Comput. Vision Patt. Recog., Maui, HI, June, pp. 310–315.
Southall, J.P.C., editor, 1962. Helmholtz's Treatise on Physiological Optics. Dover Publications: NY Originally published by the Optical Society of America in 1925.
StrangG. 1980. Linear Algebra and Its Applications. Academic Press: New York.
Sundareswaran, V. 1991. Egomotion from global flow field data. Proc. IEEE Workshop on Visual Motion, Princeton, N.J., pp. 140–145.
TsaiR.Y., and HuangT.S., 1984. Uniqueness and estimation of three-dimensional motion parameters of rigid objects with curved surfaces. IEEE, Trans. Patt. Anal. Mach. Intell. 6:13–27.
UllmanS., 1979. The Interpretation of Visual Motion. MIT Press: Cambridge, MA
UllmanS., 1984. Maximizing rigidity: the incremental recovery of 3-D structure from rigid and rubbery motion. Perception 13:255–274.
WallachH., and O'ConnellD.N., 1953. The kinetic depth effect. J. Exp. Psychol. 45:205–217.
WandellB.A., 1987. The synthesis and analysis of color images. IEEE Trans. Patt. Anal. Mach. Intell. 9:2–13.
WarrenW.H., and HannonD.J., 1988. Direction of self-motion is perceived from optical flow. Nature 336:162–163.
WarrenW.H., and HannonD.J., 1990. Eye movements and optical flow. J. Opt. Soc. Amer. A 7:160–169.
WatsonA.B., and AhumadaA.J., 1985. Model of human visualmotion sensing. J. Opt. Soc. Amer. A 2:322–342.
WaxmanA.M., and UllmanS. 1985. Surface structure and three-dimensional motion from image flow kinematics. Intern. J. Robot. Res. 4:72–94.
WaxmanA.M., and WohnK., 1985. Contour evolution, neighborhood deformation, and global image flow: planar surfaces in motion. Intern J. Robot. Res. 4:95–108.
WaxmanA.M., and WohnK., 1988. Image flow theory: A framework for 3-D inference from time-varying imagery. In Advances in Computer Vision. vol. 1, pp. 165–224. Lawrence Erlbaum Assoc.: Hillsdale, NJ.
WaxmanA.M., Kamgar-ParsiB., and SubbaraoM., 1987. Closed-form solutions to image flow equations and 3D structure and motion. Intern. J. Comput. Vision 1:239–258.
WengJ., HuangT.S., and AhujaN., 1989. Motion and structure from two perspective views: Algorithms, error analysis, and error estimation. IEEE Trans. Patt. Anal. Mach. Intell. 11:451–476.
Author information
Authors and Affiliations
Rights and permissions
About this article
Cite this article
Heeger, D.J., Jepson, A.D. Subspace methods for recovering rigid motion I: Algorithm and implementation. Int J Comput Vision 7, 95–117 (1992). https://doi.org/10.1007/BF00128130
Received:
Revised:
Issue Date:
DOI: https://doi.org/10.1007/BF00128130