Abstract
Convolutional neural network is powerful for general object recognition. However, its excellent performance depends largely on huge training set. Facing task like military object recognition in which image samples for training are scarce, its performance will degrade sharply. To solve this problem, a deep transfer learning method is proposed in this paper. The main idea consists of two parts: transfer learning for prior knowledge embedding and mixed layer for better feature extraction. It has been proved that the ability of feature extraction learned in large dataset is helpful to related tasks and can be transferred to a new neural network. The transfer learning process is achieved by fixing the weights of some layers and then retraining the remained layers. The key problem for deep transfer learning is which part should be transferred and which part should be retrained to adapt the network to the new task. This problem is solved by extensive experiments, and it is found that retraining the last three layers and transferring prior to the other layers can reach the best performance. Besides, we used mixed layer scheme to make use of the current information. In each mixed layer, convolution filters in different scales are combined together, helping to adapt features in different scales. By employing these two methods, the proposed method exhibits a large improvement in military object recognition under small training set. Experiments demonstrate that our method can achieve a high recognition precision, superior to many other algorithms compared.




Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Adankon MM, Cheriet M (2009) Support vector machine. In: International conference on intelligent networks and intelligent systems, pp 418–421
Cortes C, Vapnik V (1995) Support-vector networks. Mach Learn 20(3):273–297
Dai W, Yang Q, Xue GR, Yu Y (2007) Boosting for transfer learning. In: International conference on machine learning, pp 193–200
Denoeux T (1995) A k-nearest neighbor classification rule based on Dempster–Shafer theory. IEEE Trans Syst Man Cybern 25(5):804–813
Gao Y, Ma J, Yuille AL (2017) Semi-supervised sparse representation based classification for face recognition with insufficient labeled samples. IEEE Trans Image Process 26(5):2545–2560
García-Laencina PJ, Sancho-Gómez JL, Figueiras-Vidal AR (2010) Pattern classification with missing data: a review. Neural Comput Appl 19(2):263–282
Girshick R (2015) Fast r-cnn. In: IEEE international conference on computer vision, pp 1440–1448
Girshick R, Donahue J, Darrell T, Malik J (2013) Rich feature hierarchies for accurate object detection and semantic segmentation, pp 580–587
Guo X, Li Y, Ling H (2017) Lime: low-light image enhancement via illumination map estimation. IEEE Trans Image Process 26(2):982–993
He K, Sun, J (2014) Convolutional neural networks at constrained time cost. In: Computer vision and pattern recognition, pp 5353–5360
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Computer vision and pattern recognition, pp 770–778
Johnson J, Alahi A, Li FF (2016) Perceptual losses for real-time style transfer and super-resolution. In: European conference on computer vision, pp 694–711
Kasthuriarachchy BH, Zoysa KD, Premaratne HL (2015) Enhanced bag-of-words model for phrase-level sentiment analysis. In: International conference on advances in ICT for emerging regions, pp 210–214
Krizhevsky A, Sutskever I, Hinton GE (2012) Imagenet classification with deep convolutional neural networks. In: International conference on neural information processing systems, pp 1097–1105
Lazebnik S, Schmid C, Ponce J (2006) Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories. In: Computer vision and pattern recognition, 2006 IEEE computer society conference on, pp 2169–2178
Lecun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278–2324
Li Y, Gong S, Sherrah J, Liddell H (2004) Support vector machine based multi-view face detection and recognition. Image Vis Comput 22(5):413–427
Li Y, Zhang Y, Huang X, Zhu H, Ma J (2018) Large-scale remote sensing image retrieval by deep hashing neural networks. IEEE Trans Geosci Remote Sens 56(2):950–965
Liu Z, Li X, Luo P, Loy CC, Tang X (2015) Semantic image segmentation via deep parsing network. In: IEEE international conference on computer vision, pp 1377–1385
Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation. In: IEEE conference on computer vision and pattern recognition, pp 3431–3440
Ma J, Jiang J, Liu C, Li Y (2017) Feature guided gaussian mixture model with semi-supervised em and local geometric constraint for retinal image registration. Inf Sci 417:128–142
Ma J, Ma Y, Li C (2019) Infrared and visible image fusion methods and applications: a survey. Inf Fusion 45:153–178
Ma J, Zhao J (2017) Robust topological navigation via convolutional neural network feature and sharpness measure. IEEE Access 5:20707–20715
Oyedotun OK, Khashman A (2017) Deep learning in vision-based static hand gesture recognition. Neural Comput Appl 28(12):3941–3951
Semwal VB, Mondal K, Nandi GC (2017) Robust and accurate feature selection for humanoid push recovery and classification: deep learning approach. Neural Comput Appl 28(3):565–574
Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556
Szegedy C, Ioffe S, Vanhoucke V, Alemi AA (2017) Inception-v4, inception-resnet and the impact of residual connections on learning. In: AAAI, pp 4278–4284
Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A (2015) Going deeper with convolutions. In: IEEE conference on computer vision and pattern recognition, pp 1–9
Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z (2016) Rethinking the inception architecture for computer vision. In: Computer vision and pattern recognition, pp 2818–2826
Toshev A, Szegedy C (2014) Deeppose: Human pose estimation via deep neural networks. In: Computer vision and pattern recognition, pp 1653–1660
Viola P, Jones M (2003) Rapid object detection using a boosted cascade of simple features. In: Computer vision and pattern recognition, 2001. CVPR 2001. Proceedings of the 2001 IEEE computer society conference on, vol 1, pp I–511–I–518
Wu Y, Ianakiev K, Govindaraju V (2002) Improved k-nearest neighbor classification. Pattern Recognit 35(10):2311–2318
Yosinski J, Clune J, Bengio Y, Lipson H (2014) How transferable are features in deep neural networks? Eprint Arxiv 27:3320–3328
Zeiler MD, Fergus R (2014) Visualizing and understanding convolutional networks. In: European conference on computer vision, pp 818–833
Zhang H, Cao X, Ho JK, Chow TW (2017) Object-level video advertising: an optimization framework. IEEE Trans Ind Inform 13(2):520–531
Zhang H, Li J, Ji Y, Yue H (2017) Understanding subtitles by character-level sequence-to-sequence learning. IEEE Trans Ind Inform 13(2):616–624
Zhang H, Wu QJ, Chow TW, Zhao M (2012) A two-dimensional neighborhood preserving projection for appearance-based face recognition. Pattern Recognit 45(5):1866–1876
Zhao M, Zhan C, Wu Z, Tang P (2015) Semi-supervised image classification based on local and global regression. IEEE Signal Process Lett 22(10):1666–1670
Zhou ZH, Feng J (2017) Deep forest: towards an alternative to deep neural networks. In: IJCAI, pp 3553–3559
Acknowledgements
This work was supported by the National Natural Science Foundation of China under Grant Nos. 61773295, 61503288 and 61572076.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Rights and permissions
About this article
Cite this article
Yang, Z., Yu, W., Liang, P. et al. Deep transfer learning for military object recognition under small training set condition. Neural Comput & Applic 31, 6469–6478 (2019). https://doi.org/10.1007/s00521-018-3468-3
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-018-3468-3