Skip to main content
Log in

Multi-vector feature space based on pseudo-Euclidean space and oblique basis

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

The Earth Mover’s Distance (EMD) and the quadratic-form distance (QFD) are representative distances used in similarity searches of images. Although the QFD greatly outperforms the EMD in speed, the EMD outperform the QFD in performance. The EMD, however, has almost no theoretical justification and requires high computation costs. We propose a feature space model we call a “multi-vector feature space based on pseudo-Euclidean space and an oblique basis (MVPO).” In MVPO, an object such as an image is represented by a vector set (roughly speaking, a solid) and the EMD is reinterpreted as the distance between vector sets while the QFD is reinterpreted as the distance between the centroids of vector sets. Therefore MVPO gives a common geometrical view to these distances. We hypothesized that in MVPO the entity of an image is represented by a vector set (solid) and geometrical reasoning is applicable to MVPO. Our hypothesis explains well that the EMD outperforms the QFD in performance because the centroid of a solid is the simplest approximation of it. Our hypothesis implies that the performance of the QFD should be good when solids are far apart but bad when they are close together. We conjectured that discriminability would decline—that is, dissimilar images would be judged to be similar—when the centroids of solids are very close. Our experiment supported this conjecture. And from our hypothesis we conjectured that by making an original solid simpler, we can make an approximation method that has better performance than the QFD and faster than the EMD. The results of our experiment with this method supported our conjecture and consequently our hypothesis.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Fig. 16
Fig. 17
Fig. 18
Fig. 19

Similar content being viewed by others

Notes

  1. The proof on an image consisting of four colors in Fig. 11 is got from the property of quadratic-form distance that for any vectors v and w, if d(v, w) = 0 then d((v + w) / 2, v) = 0.

References

  1. Castelli V, Bergman LD (eds) (2002) Image databases—search and retrieval of digital imagery. Wiley, New York

  2. Gaede V, Günther O (1998) Multidimensional access methods. ACM Comput Surv 30(2):170–231 (June)

    Article  Google Scholar 

  3. Hafner J, Sawhney HS, Equitz W, Flickner M, Niblack W (1995) Efficient color histogram indexing for quadratic form distance functions. IEEE Trans Pattern Anal Mach Intell 17(7):729–736

    Article  Google Scholar 

  4. Jean JSN (1990) A new distance measure for binary images. Proc. IEEE ICASSP ’90, pp3–6, paper no. M5.19

  5. Katayama N, Satoh S (1997) The SR-tree: an index structure for high-dimensional nearest neighbor queries. In: Proc. SIGMOD 1997, pp369–380

  6. Lee D, Barber R, Niblack W, Flickner M, Hafner J, Petkovic D (1994) Query by image content using multiple objects and multiple features: user interface issues. In: Proc. ICIP 1994, pp76–80

  7. Levina E, Bickel P (2001) The Earth Mover’s Distance is the Mallows Distance: some insights from statistics. In: Proc. ICCV 2001, vol 2, pp251–256

  8. Rubner Y, Tomasi C (2001) Perceptual metrics for image database navigation. Kluwer, Dordrecht, The Netherlands

    MATH  Google Scholar 

  9. Rubner Y, Tomasi C, Guibas LJ (1998) A metric for distribution with applications to image databases. In: Proc. IEEE Intl. Conf. on Computer Vision, pp59–66

  10. Weber R, Schek H, Blott S (1998) A quantitative analysis and performance study for similarity-search methods in high-dimensional spaces. In: Proc. 24th VLDB, pp194–205

Download references

Acknowledgments

We would like to thank senior researcher D. Masumoto, Mr. Y. Uehara, and Mr. T. Baba in Fujitsu Laboratories Ltd. for their helpful advice and discussions during our research.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yasuo Yamane.

Appendix

Appendix

1.1 Determining an Oblique Basis

Instead of using the eigenvalues of the similarity matrix, we show a method based on simultaneous equations taking non-regular matrices into account.

  1. (1)

    Method for Regular Similarity Matrix.

    Let \(T = {\left( {e_{1} ,e_{2} , \ldots ,e_{n} } \right)} = {\left( {\begin{array}{*{20}c} {{e_{{11}} }} & {{e_{{12}} }} & { \cdots } & {{e_{{1n}} }} \\ {0} & {{e_{{22}} }} & { \cdots } & {{e_{{2n}} }} \\ { \vdots } & { \vdots } & { \ddots } & { \vdots } \\ {0} & {0} & { \cdots } & {{e_{{nn}} }} \\ \end{array} } \right)}\)

    Then Conditions (C1) and (C3) are regarded as being simultaneous equations with the components of T, e 11, e 12···, e nn . Given that e ii  ≠ 0, these equations can be solved by determining e 11, e 12,···, e 1n , e 22,···, e 2n ,···, e nn in that order. The value of e ij (i < j) and e ii are as follows:

    $$e_{{ij}} = {{\left( {r_{{ij}} - {\sum\limits_{k = 1}^{i - 1} {e_{{ki}} e_{{kj}} } }} \right)}} \mathord{\left/ {\vphantom {{{\left( {r_{{ij}} - {\sum\limits_{k = 1}^{i - 1} {e_{{ki}} e_{{kj}} } }} \right)}} {e_{{ii}} }}} \right. \kern-\nulldelimiterspace} {e_{{ii}} }{\text{ and }}e_{{ii}} = {\sqrt {1 - {\sum\limits_{k = 1}^{i - 1} {e_{{ki}} ^{2} } }} }.$$
  2. (2)

    Oblique Basis for Any Similarity Matrix

    We assumed a similarity matrix is regular, but using a method similar to the one above, we can get a linearly independent oblique basis for any similarity matrix in 2n-dimensional pseudo-Euclidean space. In fact, the similarity matrix in the circular case is not regular.

    Let \({\mathbf{T}} = {\left( {e_{1} ,e_{2} , \ldots ,e_{n} } \right)} = {\left( {\begin{array}{*{20}c} {1} & {{e_{{12}} }} & { \cdots } & {{e_{{1n}} }} \\ {0} & {1} & { \cdots } & {{e_{{2n}} }} \\ { \vdots } & { \vdots } & { \ddots } & { \vdots } \\ {0} & {0} & { \cdots } & {1} \\ {0} & {0} & { \cdots } & {0} \\ {0} & {{e_{{n + 2,2}} i}} & { \cdots } & {0} \\ { \vdots } & { \vdots } & { \ddots } & { \vdots } \\ {0} & {0} & {0} & {{e_{{2n,n}} i}} \\ \end{array} } \right)}.\)

    This matrix consists of two square sub-matrices. The upper one is an upper triangular matrix, and its diagonal components are 1. Therefore, e 1, e 2,..., e n are obviously linearly independent. And the lower sub-matrix is a diagonal matrix whose diagonal components are imaginary numbers or 0. And using a method similar to the one above, we can determine an oblique basis. Here all the e ij ’s obtained are real numbers.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Yamane, Y., Hoshiai, T., Tsuda, H. et al. Multi-vector feature space based on pseudo-Euclidean space and oblique basis. Multimed Tools Appl 31, 287–308 (2006). https://doi.org/10.1007/s11042-006-0045-z

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-006-0045-z

Keywords

Navigation