Skip to main content

Distorted High-Dimensional Binary Patterns Search by Scalar Neural Network Tree

  • Conference paper
  • First Online:
  • 989 Accesses

Part of the book series: Communications in Computer and Information Science ((CCIS,volume 542))

Abstract

The paper offers an algorithm (SNN-tree) that extends the binary tree search algorithm so that it can deal with distorted input vectors. Perceptrons are the tree nodes. The algorithm features an iterative solution search and stopping criterion. Unlike the SNN-tree algorithm, popular methods (LSH, k-d tree, BBF-tree, spill-tree) stop working as the dimensionality of the space grows (N > 1000). In this paper we managed to obtain an estimate of the upper bound on the error probability for SNN-tree algorithm. The proposed algorithm works much faster than exhaustive search (26 times faster at N = 10000).

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Notes

  1. 1.

    Here, \( {\mathbf{X}} \) is a row-vector.

References

  1. Kryzhanovsky, V., Malsagov, M., Tomas, J.A.C.: Hierarchical classifier: based on neural networks searching tree with iterative traversal and stop criterion. Opt. Mem. Neural Netw. (Inf. Opt.) 22(4), 217–223 (2013)

    Article  Google Scholar 

  2. Friedman, J.H., Bentley, J.L., Finkel, R.A.: An algorithm for finding best matches in logarithmic expected time. ACM Trans. Math. Softw. 3, 209–226 (1977)

    Article  MATH  Google Scholar 

  3. Liu, T., Moore, A.W., Gray, A, Yang, K.: An investigation of practical approximate nearest neighbor algorithms. In: Proceeding of Conference, Neural Information Processing Systems (2004)

    Google Scholar 

  4. Indyk, P., Motwani, R.: Approximate nearest neighbors: towards removing the curse of dimensionality. In: Proceedings of 30th STOC, pp. 604–613 (1998)

    Google Scholar 

  5. Beis, J.S., Lowe, D.G.: Shape indexing using approximate nearest-neighbour search in high-dimensional spaces. In: Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 1000–1006 (1997)

    Google Scholar 

  6. Kryzhanovsky, B., Kryzhanovskiy, V., Litinskii, L.: Machine learning in vector models of neural networks. In: Koronacki, J., Ras, Z.W., Wierzchon, S.T., et al. (eds.) Advances in Machine Learning II. Studies in Computational Intelligence, vol. SCI 263, pp. 427–443. Springer, Berlin (2010). (Dedicated to the memory of Professor Ryszard S. Michalski)

    Chapter  Google Scholar 

  7. Kryzhanovsky, V., Malsagov, M., Zelavskaya, I., Tomas, J.A.C.: High-dimensional binary pattern classification by scalar neural network tree. In: Proceedings of International Conference on Artificial Neural Networks, pp. 169–177 (2014)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Magomed Malsagov .

Editor information

Editors and Affiliations

Appendices

Appendix A

It is necessary to calculate the following probability:

$$ P^{*} = 1 - \Pr \left[ {\bigcap\limits_{m = 1}^{M - 1} {\left| {{\mathbf{XX}}_{m}^{T} } \right| < (1 - 2b_{\hbox{max} } )N} } \right]. $$
(A1)

Let scalar products \( {\mathbf{XX}}_{m}^{T} \) and \( {\mathbf{XX}}_{\mu }^{T} \) be independent random quantities, \( m \ne \mu \).

$$ P^{*} = 1 - \prod\limits_{m = 1}^{M - 1} {\Pr \left[ {\left| {{\mathbf{XX}}_{m}^{T} } \right| < (1 - 2b_{\hbox{max} } )N} \right]} . $$
(A2)

Now, it is necessary to calculate the probability that the product of each pattern by input vector is smaller than the threshold.

Scalar product \( {\mathbf{XX}}_{m}^{T} \) is a discrete quantity, which values lie in \( [ - N;N] \). Let k be the number of components with the opposite sign in vectors \( {\mathbf{X}} \) and \( {\mathbf{X}}_{m} \). Then its probability function is:

$$ \Pr \left[ {{\mathbf{XX}}_{m}^{T} = (N - 2k)} \right] = \frac{{C_{N}^{k} }}{{2^{N} }}. $$
(A3)

Random variable \( {\mathbf{XX}}_{m}^{T} \) is symmetrically distributed with zero mean, so

$$ \Pr \left[ {\left| {{\mathbf{XX}}_{m}^{T} } \right| < (1 - 2b_\text{{max}} )N} \right] = 1 - 2\sum\limits_{k = 0}^{{b_\text{{max}} N}} {\frac{{C_{N}^{k} }}{{2^{N} }}} . $$
(A4)

From A2 and A4 we can conclude that

$$ P^{*} = 1 - \left\{ {1 - 2\sum\limits_{k = 0}^{{b_\text{{max}} N}} {\frac{{C_{N}^{k} }}{{2^{N} }}} } \right\}^{M - 1} . $$
(A5)

Appendix B

Scalar product

$$ \xi = {\mathbf{XX}}_{m}^{T} = \sum\limits_{i = 1}^{N} {x_{i} x_{mi} } $$
(B1)

consists of a large number of random quantities. Therefore, at big dimensions (N > 100) its distribution can be approximated by Gaussian law with the following probability moments:

$$ \bar{\xi } = 0\quad \text{b}\quad \sigma^{2} \left( \xi \right) = N. $$
(B2)

Therefore, probability (A1) can be described by integral expression:

$$ P^{*} \sim 1 - \left\{ {1 - \frac{2}{{\sqrt {2\pi N} }}\int\limits_{ - \infty }^{{ - (1 - 2b_{\hbox{max} } )N}} {e^{{ - \frac{{\xi^{2} }}{2N}}} d\xi } } \right\}^{M - 1} . $$
(B3)

Using the following approximation

$$ \int\limits_{x}^{\infty } {e^{{ - t^{2} }} dt} \approx \frac{{e^{{ - x^{2} }} }}{2x},x \gg 1, $$
(B4)

obtain the final estimation of probability (A1.1):

$$ P^{*} < \frac{2M}{{\sqrt {2\pi \tilde{N}} }}\exp \left( { - \frac{{\tilde{N}}}{2}} \right),\quad \tilde{N} = N(1 - 2b_{\hbox{max} } )^{2} . $$
(B5)

Rights and permissions

Reprints and permissions

Copyright information

© 2015 Springer International Publishing Switzerland

About this paper

Cite this paper

Kryzhanovsky, V., Malsagov, M. (2015). Distorted High-Dimensional Binary Patterns Search by Scalar Neural Network Tree. In: Khachay, M., Konstantinova, N., Panchenko, A., Ignatov, D., Labunets, V. (eds) Analysis of Images, Social Networks and Texts. AIST 2015. Communications in Computer and Information Science, vol 542. Springer, Cham. https://doi.org/10.1007/978-3-319-26123-2_20

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-26123-2_20

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-26122-5

  • Online ISBN: 978-3-319-26123-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics