skip to main content
10.1145/3224207.3224219acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicdpaConference Proceedingsconference-collections
research-article

Research Summary of Convolution Neural Network in Image Recognition

Authors Info & Claims
Published:12 May 2018Publication History

ABSTRACT

Convolution neural networks is a model of deep learning and it is popular in image recognition, object detection and speech recognition. This paper studied Convolution neural networks in detail. Firstly this paper introduced the generation and development of convolution neural networks and illustrated its advantages in image recognition tasks. Then, it summarized the classic structure of convolution neural networks. Next, this paper stated the study trends and summarized five aspects: appropriately simplified networks, reducing over-fitting, increasing gradient signal, deeper networks and randomization. Finally, this paper discussed the problems existed in convolution neural networks and looked forward to the development trends

References

  1. Hubel, D.H. & Wiesel, T.N. Receptive fields and functional architecture of monkey striate cortex. J. Physiol. (Lond.) 195, 215--243.Google ScholarGoogle ScholarCross RefCross Ref
  2. Fukushima, K. Biol. Cybernetics (1980) 36: 193.Google ScholarGoogle Scholar
  3. Y. LeCun and Y. Bengio: Convolutional Networks for Images, Speech, and Time-Series, in Arbib, M. A. (Eds), The Handbook of Brain Theory and Neural Networks, MIT Press, 1995. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Lécun Y, Bottou L, Bengio Y, et al. Gradient-based learning applied to document recognition{J}. Proceedings of the IEEE, 1998, 86(11): 2278--2324.Google ScholarGoogle ScholarCross RefCross Ref
  5. Krizhevsky A, Sutskever I, Hinton G E. ImageNet Classification with Deep Convolutional Neural Networks{J}. Advances in Neural Information Processing Systems, 2012, 25(2):2012. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Nair, Vinod, and G. E. Hinton. "Rectified Linear Units Improve Restricted Boltzmann Machines Vinod Nair." International Conference on Machine Learning 2010: 807--814. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Lin M, Chen Q, Yan S. Network In Network{J}. Computer Science, 2014.Google ScholarGoogle Scholar
  8. Springenberg J T, Dosovitskiy A, Brox T, et al. Striving for Simplicity: The All Convolutional Net{J}. Eprint Arxiv, 2014.Google ScholarGoogle Scholar
  9. Szegedy C, Liu W, Jia Y, et al. Going deeper with convolutions{J}. 2014: 1--9.Google ScholarGoogle Scholar
  10. He K, Zhang X, Ren S, et al. Deep Residual Learning for Image Recognition{J}. Computer Science, 2015.Google ScholarGoogle Scholar
  11. Srivastava R K, Greff K, Schmidhuber J. Highway Networks{J}. Computer Science, 2015.Google ScholarGoogle Scholar
  12. Huang G, Liu Z, Weinberger K Q. Densely Connected Convolutional Networks{J}. 2016.Google ScholarGoogle Scholar
  13. Huang G, Sun Y, Liu Z, et al. Deep Networks with Stochastic Depth{J}. 2016.Google ScholarGoogle Scholar
  14. Yann LeCun, Yoshua Bengio & Geoffrey Hinton. Deep learning.Google ScholarGoogle Scholar
  15. Smith L N, Topin N. Deep Convolutional Neural Network Design Patterns{J}. 2016.Google ScholarGoogle Scholar
  16. He K, Zhang X, Ren S, et al. Deep Residual Learning for Image Recognition{J}. Computer Science, 2015.Google ScholarGoogle Scholar
  17. Nagi J, Ducatelle F, Caro G A D, et al. Max-pooling convolutional neural networks for vision-based hand gesture recognition{C}// IEEE International Conference on Signal and Image Processing Applications. IEEE, 2011: 342--347.Google ScholarGoogle Scholar
  18. Lecun Y, Kavukcuoglu K, Farabet C. Convolutional networks and applications in vision{C}// IEEE International Symposium on Circuits and Systems. IEEE, 2010: 253--256.Google ScholarGoogle Scholar
  19. Jarrett K, Kavukcuoglu K, Ranzato M, et al. What is the best multi-stage architecture for object recognition?{J}. 2009, 30(2): 2146--2153.Google ScholarGoogle Scholar
  20. Hawkins D M. The problem of overfitting.{J}. ChemInform, 2004, 35(19): 1--12.Google ScholarGoogle ScholarCross RefCross Ref
  21. Hariharan B, Arbeláez P, Girshick R, et al. Hypercolumns for object segmentation and fine-grained localization{J}. 2015: 447--456.Google ScholarGoogle Scholar
  22. Hinton G E, Srivastava N, Krizhevsky A, et al. Improving neural networks by preventing co-adaptation of feature detectors{J}. Computer Science, 2012, 3(4):págs. 212--223.Google ScholarGoogle Scholar
  23. Srivastava N, Hinton G, Krizhevsky A, et al. Dropout: a simple way to prevent neural networks from overfitting{J}. Journal of Machine Learning Research, 2014, 15(1): 1929--1958. Google ScholarGoogle ScholarDigital LibraryDigital Library
  24. Wan L, Zeiler M, Zhang S, et al. Regularization of neural networks using dropconnect{C}// International Conference on Machine Learning. 2013: 1058--1066. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Hinton G E, Dayan P, Frey B J, et al. The "wake-sleep" algorithm for unsupervised neural networks.{J}. Science, 1995, 268(5214): 1158--61.Google ScholarGoogle ScholarCross RefCross Ref
  26. Schölkopf, B, Platt, J, Hofmann, T. Greedy Layer-Wise Training of Deep Networks{J}. Advances in Neural Information Processing Systems, 2007, 19: 153--160. Google ScholarGoogle ScholarDigital LibraryDigital Library
  27. Erhan D, Bengio Y, Courville A, et al. Why Does Unsupervised Pre-training Help Deep Learning?{J}. Journal of Machine Learning Research, 2010, 11(3): 625--660. Google ScholarGoogle ScholarDigital LibraryDigital Library
  28. Erhan D, Manzagol P A, Bengio Y, et al. The Difficulty of Training Deep Architectures and the Effect of Unsupervised Pre-Training.{J}. Immunology of Fungal Infections, 2009, 5: 153--160.Google ScholarGoogle Scholar
  29. Erhan D. Understanding Deep Architectures and the Effect of Unsupervised Pre-training{J}. 2011.Google ScholarGoogle Scholar
  30. Wu Y N. Data Augmentation{J}. Computer Vision, 2014: 165--166.Google ScholarGoogle Scholar
  31. Dyk D A V, Meng X L. The Art of Data Augmentation{J}. Journal of Computational & Graphical Statistics, 2012, 10(1): 1--50.Google ScholarGoogle Scholar
  32. Howard A G. Some Improvements on Deep Convolutional Neural Network Based Image Classification{J}. Computer Science, 2013.Google ScholarGoogle Scholar
  33. Wong S C, Gatt A, Stamatescu V, et al. Understanding data augmentation for classification: when to warp?{J}. 2016.Google ScholarGoogle Scholar
  34. Raiko T, Valpola H, Lecun Y. Deep learning made easier by linear transformations in perceptrons{J}. 2012, 22: 924--932.Google ScholarGoogle Scholar
  35. He K, Zhang X, Ren S, et al. Identity Mappings in Deep Residual Networks{J}. 2016.Google ScholarGoogle Scholar
  36. Simonyan K, Zisserman A. Very Deep Convolutional Networks for Large-Scale Image Recognition{J}. Computer Science, 2014.Google ScholarGoogle Scholar
  37. Larsson G, Maire M, Shakhnarovich G. FractalNet: Ultra-Deep Neural Networks without Residuals{J}. 2016.Google ScholarGoogle Scholar
  38. A. L. Maas, A. Y. Hannun. And Y. Bengio. Rectifier nonlinearities improve neural network acoustic models, in ICML 2103.Google ScholarGoogle Scholar
  39. He K, Zhang X, Ren S, et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification{J}. 2015: 1026--1034. Google ScholarGoogle ScholarDigital LibraryDigital Library
  40. Xu B, Wang N, Chen T, et al. Empirical Evaluation of Rectified Activations in Convolutional Network{J}. Computer Science, 2015.Google ScholarGoogle Scholar
  41. Agostinelli F, Hoffman M, Sadowski P, et al. Learning Activation Functions to Improve Deep Neural Networks{J}. Computer Science, 2014.Google ScholarGoogle Scholar
  42. Yu D, Wang H, Chen P, et al. Mixed Pooling for Convolutional Neural Networks{C}// the 9th international conference on rough sets and knowledge technology. 2014: 364--375.Google ScholarGoogle Scholar
  43. Zeiler M D, Fergus R. Stochastic Pooling for Regularization of Deep Convolutional Neural Networks{J}. Computer Science, 2013.Google ScholarGoogle Scholar
  44. Goodfellow I J, Wardefarley D, Mirza M, et al. Maxout Networks{J}. Computer Science, 2013: 1319--1327. Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Research Summary of Convolution Neural Network in Image Recognition

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Other conferences
      ICDPA 2018: Proceedings of the International Conference on Data Processing and Applications
      May 2018
      73 pages
      ISBN:9781450364188
      DOI:10.1145/3224207

      Copyright © 2018 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 12 May 2018

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Research
      • Refereed limited

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader