skip to main content
10.1145/3486001.3486222acmotherconferencesArticle/Chapter ViewAbstractPublication PagesaimlsystemsConference Proceedingsconference-collections
research-article

Leveraging Prediction Confidence For Versatile Optimizations to CNNs

Published:22 October 2021Publication History

ABSTRACT

Modern convolutional neural networks (CNNs) incur huge computational and energy overheads. In this paper, we propose two techniques for inferring the confidence in the correctness of a prediction in the early layers of a CNN. The first technique uses a statistical approach, whereas the second technique requires retraining. We argue that prediction confidence estimation can enable diverse optimizations to CNNs. We demonstrate two optimizations. First, we predict selected images in early layers. This is possible because in a dataset, many images are easy to predict and they can be predicted in the early layers of a CNN. This reduces the average computation count at the cost of accuracy and parameter count. Second, we propose predicting only selected images for which the prediction-confidence is high. This reduces the coverage; however, the accuracy on the images that are predicted is higher. Our results with VGG16 and ResNet50 CNNs on the Caltech256 dataset show that our techniques are effective. For example, for ResNet, our first technique reduces the accuracy from 71.6% to 69.8% while reducing the computations by 14%. Similarly, with the second technique, on reducing the coverage from 100% to 90%, the accuracy is increased from 71.6% to 75.6%.

Keywords: computer vision, CNN, approximate computing, accuracy-coverage tradeoff, prediction confidence

References

  1. [n.d.]. AI and Compute. https://openai.com/blog/ai-and-compute/.Google ScholarGoogle Scholar
  2. Yunpeng Chen, Yannis Kalantidis, Jianshu Li, Shuicheng Yan, and Jiashi Feng. 2018. A2-Nets: Double Attention Networks. arXiv preprint arXiv:1810.11579(2018).Google ScholarGoogle Scholar
  3. Yonatan Geifman and Ran El-Yaniv. 2017. Selective classification for deep neural networks. arXiv preprint arXiv:1705.08500(2017).Google ScholarGoogle Scholar
  4. Gregory Griffin, Alex Holub, and Pietro Perona. 2007. Caltech-256 object category dataset. (2007).Google ScholarGoogle Scholar
  5. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. 2016. Deep residual learning for image recognition. In CVPR. 770–778.Google ScholarGoogle Scholar
  6. Anitha Kannan, Jason Alan Fries, Eric Kramer, Jen Jen Chen, Nigam Shah, and Xavier Amatriain. 2020. The accuracy vs. coverage trade-off in patient-facing diagnosis models. AMIA Summits on Translational Science Proceedings (2020), 298.Google ScholarGoogle Scholar
  7. Sparsh Mittal. 2016. A Survey Of Techniques for Approximate Computing. Comput. Surveys 48, 4 (2016), 62:1–62:33.Google ScholarGoogle Scholar
  8. Sparsh Mittal. 2019. A Survey on Optimized Implementation of Deep Learning Models on the NVIDIA Jetson Platform. Journal of Systems Architecture 97 (2019), 428–442.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Sparsh Mittal. 2020. A Survey on Modeling and Improving Reliability of DNN Algorithms and Accelerators. Journal of Systems Architecture 104 (March 2020), 101689.Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. Priyadarshini Panda, Abhronil Sengupta, and Kaushik Roy. 2016. Conditional deep learning for energy-efficient and enhanced pattern recognition. In DATE. 475–480.Google ScholarGoogle Scholar
  11. Poonam Rajput 2020. Improving Accuracy and Efficiency of Object Detection Algorithms using Multiscale Feature Aggregation Plugins. In IAPR TC3 Workshop on Artificial Neural Networks in Pattern Recognition (ANNPR). 65–76.Google ScholarGoogle Scholar
  12. Rajat Saini 2020. ULSAM: Ultra-Lightweight Subspace Attention Module for Compact Convolutional Neural Networks. In IEEE/CVF Winter Conference on Applications of Computer Vision. 1627–1636.Google ScholarGoogle Scholar
  13. Karen Simonyan and Andrew Zisserman. 2014. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556(2014).Google ScholarGoogle Scholar
  14. Surat Teerapittayanon, Bradley McDanel, and Hsiang-Tsung Kung. 2016. Branchynet: Fast inference via early exiting from deep neural networks. In Intl. Conf. on Pattern Recognition (ICPR). IEEE, 2464–2469.Google ScholarGoogle ScholarCross RefCross Ref
  15. Yawen Wu, Zhepeng Wang, Zhenge Jia, Yiyu Shi, and Jingtong Hu. 2020. Intermittent inference with nonuniformly compressed multi-exit neural network for energy harvesting powered devices. In DAC. 1–6.Google ScholarGoogle Scholar
  16. Zhilu Zhang and Mert R Sabuncu. 2018. Generalized cross entropy loss for training deep neural networks with noisy labels. arXiv preprint arXiv:1805.07836(2018).Google ScholarGoogle Scholar

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image ACM Other conferences
    AIMLSystems '21: Proceedings of the First International Conference on AI-ML Systems
    October 2021
    170 pages
    ISBN:9781450385947
    DOI:10.1145/3486001

    Copyright © 2021 ACM

    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    • Published: 22 October 2021

    Permissions

    Request permissions about this article.

    Request Permissions

    Check for updates

    Qualifiers

    • research-article
    • Research
    • Refereed limited
  • Article Metrics

    • Downloads (Last 12 months)9
    • Downloads (Last 6 weeks)0

    Other Metrics

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format .

View HTML Format