Dynamic motion analysis using wavelet flow surface images

https://doi.org/10.1016/S0167-8655(99)00006-9Get rights and content

Abstract

We have developed a motion analysis method that combines the wavelet transform with the flow surface image technique. In dynamic navigation environments, whenever objects move, the projections of the objects onto the image plane also move. These projections build over time spatiotemporal surfaces of their movement and volumes created by the surfaces. This paper presents a new method for the interpretation of optical flow for moving objects from a sequence of images. Flow surface images of the moving objects are created within the wavelet-derived space, chosen from seven different directionally sensitive detail images using the 3D wavelet decomposition. The motion estimation algorithm concentrates on the integration of information from the flow surface images, followed by a quadratic patch parameterization and determination of flow paths of the end points of edges on the flow surface images. The results of two experimental studies with an object exhibiting out-of-plane translation and rotation are also reported.

Introduction

An important goal of computer vision is the analysis of the motion characteristics of the observer and the objects in dynamic scenes for the recovery of 3D surface information. There are three approaches that can be used to interpret the motion of objects: gradient approach, correspondence approach, and spatiotemporal method. Good reviews can be found in (Aggarwal and Nandhakumar, 1988; Barron et al., 1994; Beauchemin and Barron, 1995). An approach which uses a fusion of results from all three methods was developed by Singh (1991).

The gradient approach is based on the computation of image changes in space and time (e.g. Horn and Schunck, 1981; Snyder, 1991). The gradient method does not allow for acceleration in the optic flow calculation and is sensitive to noise in the gradients. In addition, the brightness constancy assumption is violated by light source motion and/or shadows. The method has recently been extended by Negahdaripour (1998) to include variations in lighting by decoupling geometric and radiometric sources of scene gradients. Similar studies along these lines can be found in (Cornelius and Kanade, 1983; Schunck, 1985; Nagel, 1989; Verri and Poggio, 1989). A recent experimental study by Brandt (1997) indicated that small-support low-pass prefiltering of the input sequence led to better accuracy in the optic flow computation.

In the correspondence method, points, lines, or regions of objects in the first image are put into correspondence with similar features later in the sequence (e.g. Huntsberger and Jayaramamurthy, 1987; Liu and Huang, 1986; Roach and Aggarwal, 1980). A study by Weng et al. (1993) indicated that even small errors of one or two pixels in image coordinates used for point correspondence cause an ambiguous interpretation of translation or rotation due to violation of the epipolar constraint. Their solution was to use a two-step approach that consisted of a linear solution of the closed-form correspondence equations, followed by a minimization of image errors using a minimum variance estimator technique. A recent analysis by Zhang (1995) demonstrated that 3D motion and structure parameters can be derived from corresponding line segments in just two perspective images. His study uses an epipolar (EPI) geometry constraint on the common part of the corresponding line segments.

Another approach for motion analysis and the method used in our study is to consider a 3D spatiotemporal image (e.g. Allmen and Dyer, 1991; Baker and Bolles, 1989; Rangachar et al., 1990), where a temporal cube is built by stacking together a sequence of images. Work also related to this approach is the spatiotemporal filters of Fleet and Jepson (1990) and Heeger (1988), where a frequency space analysis is performed using tuned filters. A recent study by Otsuka et al. (1997) used a 3D Hough transform for planar surface detection to derive spatiotemporal surfaces (trajectory surfaces). Although robust, the method suffers from a high computational cost from the Hough transform. We use a local region growing approach followed by a Monge patch parameterization to derive the spatiotemporal surfaces in the wavelet coefficient spaces.

In general, 3D motion analysis computes the motion and structure of objects from a sequence of images using two steps. The first step is to compute observables (features or optical flow) in the images; followed by the extraction of the motion and structure of objects using the observables. Closed form solutions for derivation of the structure and motion paramers have been developed by Subbarao (1989) and Waxman et al. (1987). This paper uses a combination of feature extraction and optical flow derived from the flow surface of wavelet-derived edge points in a sequence of images for the first motion analysis step. In a further study, Chang (1997) used the solution of a linear set of equations derived from the optical flow of the surfaces to derive the structure and motion parameters of the objects.

Huntsberger et al. (1994) focused on wavelet-derived epipolar-plane image (EPI) analysis, and alleviated some of the problems in the inter pretation of translation in depth and occlusion/disocclusion. Allmen and Dyer, 1990, Allmen and Dyer, 1991 have studied the spatiotemporal surface flow method for motion estimation. They assumed that the arc length of a segment is constant throughout the temporal slices. Although this assumption is applied to vertical and horizontal movement, out-of-plane translation would be problematic. Chang and Huntsberger (1996) in their study of a wavelet flow surface motion analysis method created spatiotemporal surfaces for pure translation and pure rotation of simple lines. The partial derivatives from the patch parameterization of the flow surface were then used to derive the optical flow. The wavelet method has an advantage over that of Allmen and Dyer (1990) in that directional information is automatically supplied by the wavelet channels, which leads to a reduction in the amount of computational processing.

In this paper, we present a new approach for the derivation of optical flow using wavelet flow surface images. This approach does not require texture for differentiation or correlation, but only uses spatiotemporal flow surfaces. The new method applies factors for noise reduction, feature grouping, and surface smoothness. Using the directional sensitivity of wavelet-derived coefficients, the computational speed is faster than the previous methods. The slope of the wavelet-derived flow surfaces are computed to determine optic flow for out-of-plane translation and rotation of a rigid object.

Section 2contains a brief review of wavelet decomposition and characteristics of 3D wavelet coefficients. This is followed by a discussion of the flow surface method for the extraction of the optical flow. The final section gives the results of some experimental studies of a object undergoing out-of-plane translation and rotation in an industrial environment.

Section snippets

Wavelet analysis

Grossmann and Morlet (1984) developed the continuous wavelet representation for one-dimensional signals to analyze seismic data. The wavelet functions were constructed as continuous translations and dilations of one function with local support in time and frequency domain. The wavelet representation has good localization properties in the frequency and spatial domains. Another benefit of the wavelet representation is that the filters used are orientation sensitive. The horizontal, vertical, or

Flow surface analysis

Spatiotemporal (ST) flow surfaces are the natural extension of optical flow to ST surfaces. The structure of volumetric flow surfaces has a great deal of information about contours in a scene and their motion. Whenever objects move in the real world, the projections of the objects onto the image plane also move. These projections build, over time, ST surfaces of their movement and volumes created by the surfaces. Therefore, the motion of the objects is represented by these surfaces and volumes.

Experimental study

We performed three experimental studies to test our method. The first used a box moving with out-of-plane translation (in depth), and the second and third used a box moving with two different rates of out-of-plane rotation. The box is manually moved for each frame in the sequence which gives us the exact flow since we use calibration marks on the floor to measure distances. Our previous studies using generated sequences of in-plane translating and rotating objects returned errors of less than

Conclusions

We have presented a new method of motion analysis that combines the 3-D wavelet decomposition of an image sequence with a flow surface image analysis of the wavelet-derived volume data. The combination of patch parameterization and surface filtering for the partial derivatives was introduced for velocity field determination. Three factors were used to remove spurious coefficients, and to automatically smooth and cluster points for derivation of the flow surfaces. These factors resulted in

References (40)

  • P. Besl et al.

    Invariant surface characteristics for 3D object recognition in range images

    CVGIP

    (1986)
  • B. Horn et al.

    Determining optical flow

    Artificial Intell.

    (1981)
  • T.L. Huntsberger et al.

    Determination of the optic flow field using the spatiotemporal deformation of region properties

    Pattern Recognition Letters

    (1987)
  • J.K. Aggarwal et al.

    On the computation of motion from sequences of images – A review

    Proceedings of the IEEE

    (1988)
  • Allmen, M., Dyer, C., 1990. Computing spatiotemporal surface flow. In: Proc. 3rd Internat. Conf. Comput. Vision, pp....
  • Allmen, M., Dyer, C., 1991. Long-range spatiotemporal motion understanding using spatiotemporal flow curves. In: Proc....
  • H. Baker et al.

    Generating epipolar–plane image analysis on the spatiotemporal surface

    Internat. J. Comput. Vision

    (1989)
  • J.L. Barron et al.

    Performance of optical flow techniques

    Internat. J. Comput. Vision

    (1994)
  • S.S. Beauchemin et al.

    The computation of optical flow

    ACM Computing Surveys

    (1995)
  • J.W. Brandt

    Improved accuracy in gradient-based optical flow estimation

    Internat. J. Comput. Vision

    (1997)
  • Chang, J., 1997. Motion and structure from optical flow using wavelets and flow surface images. Ph.D. Dissertation,...
  • Chang, J., Huntsberger, T.L., 1996. Determination of the optical flow using wavelets and flow surface images. TR 9603,...
  • Chui, C.K., 1992. An Introduction to Wavelets. Academic Press, New York,...
  • Cornelius, N., Kanade, T., 1983. Adapting optical flow to measure object motion in reflectance and X-ray image...
  • Daubechies, I., 1992. Ten Lectures on Wavelets. CBMS 61, SIAM...
  • D.J. Fleet et al.

    Computation of component image velocity from local phase information

    Internat. J. Comput. Vision

    (1990)
  • G.E. Forsythe

    Generation and use of orthogonal polynomials for data-fitting with a digital computer

    J. SIAM

    (1957)
  • A. Grossmann et al.

    Decomposition of Hardy functions into square integrable wavelets of constant shape

    SIAM J. Math.

    (1984)
  • R. Haralick

    Digital step edges from zero crossing of second directional derivative

    IEEE Trans. PAMI

    (1984)
  • D.J. Heeger

    Optical flow using spatiotemporal filters

    Internat. J. Comput. Vision

    (1988)
  • Cited by (0)

    1

    Now with the Department of Computer Science, Hanshin University, Kyung-Ki Do Osan City, Yang-san Dong 411, S. Korea, 447-791.

    2

    Research supported in part under ONR Grant no. N00014-94-1-1163. Reprints of this paper can be obtained by contacting the author at [email protected]. The authors would like to thank the reviewers whose comments have led to a more readable manuscript.

    View full text