skip to main content
10.1145/3628797.3629005acmotherconferencesArticle/Chapter ViewAbstractPublication PagessoictConference Proceedingsconference-collections
research-article

Tensor Analysis Of Convolutional Neural Network For Reducing Network Parameters

Published:07 December 2023Publication History

ABSTRACT

Convolutional Neural Networks have demonstrated excellent ability in image recognition, yet they frequently have significant computational and memory requirements. It is advantageous to create a tensor factorization framework that can effectively compress networks in order to solve this problem. With its ability to express high-dimensional tensors using a smaller set of core tensors and fewer parameters, the Tensor-Train (TT) structure is a viable choice for compressing neural networks. The choice of appropriate TT-ranks, however, is now without theoretical guarantees. In our study, we introduce a method that employs the widely recognized heuristic network slimming structure through batch normalization techniques. This method serves as metrics to gauge the requisite parameter count, encompassing the intricacies of the training data complexity derived from the original CNNs. Our focus remains on the TT-rank, which denotes the scale parameter choice facilitated by the estimated size. This approach provides a comprehensive strategy to ascertain network compression requirements and resource optimization. Our results reveal that our calculated TT-ranks lead to substantial reduction in computational complexity and memory usage, while maintaining competitive accuracy compared to baseline CNNs.

References

  1. Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. 2009. Imagenet: A large-scale hierarchical image database. 2009 IEEE conference on computer vision and pattern recognition (2009), 248–255.Google ScholarGoogle ScholarCross RefCross Ref
  2. Timur Garipov, Dmitry Podoprikhin, Alexander Novikov, and Dmitry Vetrov. 2016. Ultimate tensorization: compressing convolutional and FC layers alike. arXiv preprint arXiv:1611.03214 (2016).Google ScholarGoogle Scholar
  3. Patrick Gelß. 2017. The Tensor-Train Format and Its Applications: Modeling and Analysis of Chemical Reaction Networks, Catalytic Processes, Fluid Flows, and Brownian Dynamics. Ph. D. Dissertation.Google ScholarGoogle Scholar
  4. Patrick Gelß. 2023. The tensor-train format and its applications. Accessed: January 2, 2023.Google ScholarGoogle Scholar
  5. Song Han, Huizi Mao, and William J. Dally. 2016. Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization, and Huffman Coding. arXiv preprint arXiv:1510.00149 (2016). arxiv:1510.00149 [cs.CV]Google ScholarGoogle Scholar
  6. Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. 2015. Distilling the Knowledge in a Neural Network. arXiv preprint arXiv:1503.02531 (2015). arxiv:1503.02531 [stat.ML]Google ScholarGoogle Scholar
  7. S. Houben, J. Stallkamp, J. Salmen, and C. Igel. 2013. GTSRB: German Traffic Sign Recognition Benchmark. http://benchmark.ini.rub.de/?section=gtsrb&subsection=news.Google ScholarGoogle Scholar
  8. A. Krizhevsky and G. Hinton. 2009. Learning multiple layers of features from tiny images. Technical Report. Tech Report.Google ScholarGoogle Scholar
  9. Y. Li, X. Li, Y. Li, X. Sun, Y. Zhang, and Y. Zhang. 2021. HLRTF: Hierarchical Low-Rank Tensor Factorization for Inverse Problems in Multi-Dimensional Imaging. IEEE Transactions on Medical Imaging 40, 4 (2021), 1094–1105. https://doi.org/10.1109/TMI.2020.3030494Google ScholarGoogle ScholarCross RefCross Ref
  10. Z. Liu, J. Li, Z. Shen, G. Huang, S. Yan, and C. Zhang. 2017. Learning Efficient Convolutional Networks through Network Slimming. In 2017 IEEE International Conference on Computer Vision (ICCV). Venice, Italy, 2755–2763. https://doi.org/10.1109/ICCV.2017.298Google ScholarGoogle ScholarCross RefCross Ref
  11. I. V. Oseledets. 2011. Tensor-Train decomposition. SIAM J. Scientific Computing 33, 5 (2011), 2295–2317.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Karen Simonyan and Andrew Zisserman. 2015. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv preprint arXiv:1409.1556 (2015).Google ScholarGoogle Scholar
  13. D. Tian, S. Yamagiwa, and K. Wada. 2022. Heuristic Method for Minimizing Model Size of CNN by Combining Multiple Pruning Techniques. Sensors (Basel) 22, 15 (2022), 5874. https://doi.org/10.3390/s22155874Google ScholarGoogle ScholarCross RefCross Ref
  14. H. Zhou, J. M. Alvarez, and F. Porikli. 2016. Less is more: Towards compact cnns. In ECCV.Google ScholarGoogle Scholar

Index Terms

  1. Tensor Analysis Of Convolutional Neural Network For Reducing Network Parameters

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in
      • Published in

        cover image ACM Other conferences
        SOICT '23: Proceedings of the 12th International Symposium on Information and Communication Technology
        December 2023
        1058 pages
        ISBN:9798400708916
        DOI:10.1145/3628797

        Copyright © 2023 ACM

        Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 7 December 2023

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article
        • Research
        • Refereed limited

        Acceptance Rates

        Overall Acceptance Rate147of318submissions,46%
      • Article Metrics

        • Downloads (Last 12 months)36
        • Downloads (Last 6 weeks)9

        Other Metrics

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      HTML Format

      View this article in HTML Format .

      View HTML Format