skip to main content
research-article

Tensorizing Restricted Boltzmann Machine

Authors Info & Claims
Published:07 June 2019Publication History
Skip Abstract Section

Abstract

Restricted Boltzmann machine (RBM) is a famous model for feature extraction and can be used as an initializer for neural networks. When applying the classic RBM to multidimensional data such as 2D/3D tensors, one needs to vectorize such as high-order data. Vectorizing will result in dimensional disaster and valuable spatial information loss. As RBM is a model with fully connected layers, it requires a large amount of memory. Therefore, it is difficult to use RBM with high-order data on low-end devices. In this article, to utilize classic RBM on tensorial data directly, we propose a new tensorial RBM model parameterized by the tensor train format (TTRBM). In this model, both visible and hidden variables are in tensorial form, which are connected by a parameter matrix in tensor train format. The biggest advantage of the proposed model is that TTRBM can obtain comparable performance compared with the classic RBM with much fewer model parameters and faster training process. To demonstrate the advantages of TTRBM, we conduct three real-world applications, face reconstruction, handwritten digit recognition, and image super-resolution in the experiments.

References

  1. N. Andrew. 2011. Sparse autoencoder. CS294A Lecture notes, 72.Google ScholarGoogle Scholar
  2. Y. Bengio. 2009. Learning deep architectures for AI. Foundations and Trends in Machine Learning 2, 1 (2009), 1--127.Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. C. M. Bishop. 2006. Pattern Recognition and Machine Learning. Springer. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. N. Cai, Z. Su, Z. Lin, H. Wang, Z. Yang, and B. Ling. 2017. Blind inpainting using the fully convolutional neural network. Visual Computer 33, 2 (2017), 249--261. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. C. Chen, H. Tong, L. Xie, L. Ying, and Q. He. 2017. Cross-dependency inference in multi-layered networks: A collaborative filtering perspective. ACM Transactions on Knowledge Discovery from Data 2016 Special Issue 11, 4, Article 42 (2017), 1--26.Google ScholarGoogle Scholar
  6. S. He, H. Chen, Z. Zhu, D. G. Ward, H. J. Cooper, M. R. Viant, J. K. Heath, and X. Yao. 2015. Robust twin boosting for feature selection from high-dimensional omics data with label noise. Information Sciences 219, 10 (2015), 1--18. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. G. E. Hinton. 2002. Training products of experts by minimizing contrastive divergence. Neural Computation 14, 8 (2002), 1771--1800. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. G. E. Hinton, S. Osindero, and Y. W. Teh. 2006. A fast learning algorithm for deep belief nets. Neural Computation 18, 7 (2006), 1527--1554. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. G. E. Hinton and R. Salakhutdinov. 2009. Replicated softmax: An undirected topic model. In Proceedings of the Advances in Neural Information Processing Systems. 1607--1614. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. L. K. Kim. 2018. DeepX: Deep learning accelerator for restricted boltzmann machine artificial neural networks. IEEE Transactions on Neural Networks and Learning Systems 29, 5 (2018), 1441--1453.Google ScholarGoogle ScholarCross RefCross Ref
  11. T. G. Kolda and B. W. Bader. 2009. Tensor decompositions and applications. SIAM Review 51, 3 (2009), 455--500. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. A. Krizhevsky and G. E. Hinton. 2010. Factored 3-way restricted Boltzmann machines for modeling natural images. In Proceedings of the International Conference on Artificial Intelligence and Statistics. 621--628.Google ScholarGoogle Scholar
  13. S. Lee, Y. Kim, J. Lee, K. Kim, K. Lee, and J. Noh. 2014. Depth manipulation using disparity histogram analysis for stereoscopic 3D. Visual Computer 30, 4 (2014), 455--465. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Q. Liu, C. Zhang, Q. Guo, H. Xu, and Y. Zhou. 2016. Adaptive sparse coding on PCA dictionary for image denoising. Visual Computer 32, 4 (2016), 535--549. Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. T. D. Nguyen, T. Tran, D. Phung, and S. Venkatesh. 2015. Tensor-variate restricted Boltzmann machines. In Proceedings of the Association for the Advancement of Artificial Intelligence. 2887--2893. Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. A. Novikov, D. Podoprikhin, A. Osokin, and D. P. Vetrov. 2015. Tensorizing neural networks. In Proceedings of the Advances in Neural Information Processing Systems. 442--450.Google ScholarGoogle Scholar
  17. I. V. Oseledets. 2011. Tensor-train decomposition. SIAM Journal on Scientific Computing 33, 5 (2011), 2295--2317. Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. E. Papalexakis, C. Faloutsos, and N. D. Sidiropoulos. 2015. ParCube: Sparse parallelizable CANDECOMP-PARAFAC tensor decomposition. ACM Transactions on Knowledge Discovery from Data 10, 1 (2015), 3:1--3:25.Google ScholarGoogle Scholar
  19. G. Qi, Y. Sun, J. Gao, Y. Hu, and J. Li. 2016. Matrix variate restricted Boltzmann machine. In Proceedings of the IEEE International Joint Conference on Neural Networks. 389--395.Google ScholarGoogle Scholar
  20. R. Socher, D. Chen, C. Manning, and N. Andrew. 2013. Reasoning with neural tensor networks for knowledge base completion. In Proceedings of the Advances in Neural Information Processing Systems. 926--934.Google ScholarGoogle Scholar
  21. S. Soundarajan and J. E. Hopcroft. 2015. Use of local group information to identify communities in networks. ACM Transactions on Knowledge Discovery from Data 9, 3 (2015), 1--27. Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. J. Tang, X. Shu, Z. Li, Y. Jiang, and Q. Tian. 2018. Social anchor-unit graph regularized tensor completion for large-scale image retagging. arxiv:1804.04397.Google ScholarGoogle Scholar
  23. J. Tang, X. Shu, G. Qi, Z. Li, M. Wang, S. Yan, and R. Jain. 2017. Tri-clustered tensor completion for social-aware image tag refinement. IEEE Transactions on Pattern Analysis and Machine Intelligence 39, 8 (2017), 1662--1674.Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. T. Tieleman. 2008. Training restricted Boltzmann machines using approximations to the likelihood gradient. In Proceedings of the International Conference on Machine Learning. 1064--1071. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. W. K. Wong, Z. Lai, Y. Xu, J. Wen, and C. P. Ho. 2015. Joint tensor feature analysis for visual object recognition. IEEE Transactions on Cybernetics 45, 11 (2015), 2425--2436.Google ScholarGoogle ScholarCross RefCross Ref
  26. Y. Wu, S. C. Hoi, T. Mei, and N. Yu. 2017. Large-scale online feature selection for ultra-high dimensional sparse data. ACM Transactions on Knowledge Discovery from Data 11, 4 (2017), 48:1--48:22. Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. J. Yang, J. Wright, T. S. Huang, and Y. Ma. 2010. Image super-resolution via sparse representation. IEEE Transactions on Image Processing 19, 11 (2010), 2861--2873. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. M. Yang, P. Zhu, F. Liu, and L Shen. 2015. Joint representation and pattern learning for robust face recognition. Neurocomputing 168 (2015), 70--80. Google ScholarGoogle ScholarDigital LibraryDigital Library
  29. H. Zhang, Y. Yang, H. Luan, S. Yang, and T. Chua. 2014. Start from scratch: Towards automatically identifying, modeling, and naming visual attributes. In Proceedings of the ACM International Conference on Multimedia. 187--196.Google ScholarGoogle Scholar
  30. J. Zhang, Y. Han, and J. Jiang. 2015. Tensor rank selection for multimedia analysis. Journal of Visual Communication and Image Representation 30, 7 (2015), 376--392.Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. Z. Zhu, S. Jia, S. He, Y. Sun, Z. Ji, and L. Shen. 2015. Three-dimensional Gabor feature extraction for hyperspectral imagery classification using a memetic framework. Information Sciences 298, 20 (2015), 274--287. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Tensorizing Restricted Boltzmann Machine

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in

    Full Access

    • Published in

      cover image ACM Transactions on Knowledge Discovery from Data
      ACM Transactions on Knowledge Discovery from Data  Volume 13, Issue 3
      June 2019
      261 pages
      ISSN:1556-4681
      EISSN:1556-472X
      DOI:10.1145/3331063
      Issue’s Table of Contents

      Copyright © 2019 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 7 June 2019
      • Accepted: 1 March 2019
      • Revised: 1 January 2019
      • Received: 1 October 2017
      Published in tkdd Volume 13, Issue 3

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Research
      • Refereed

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format .

    View HTML Format