Abstract
Today deep learning has become a supreme tool for machine learning regardless of application field. However, due to a large number of parameters, tremendous amount of data is required to avoid over-fitting. Data acquisition and labeling are done by human one by one, and therefore expensive. In a given situation, the dataset with adequate amount is difficult to acquire. To resolve the problem, transfer learning is adopted (Yosinski et al, Adv Neural Inf Process Syst, 2014, [1]). The transfer learning is delivering the knowledge learned from abundant dataset, e.g. ImageNet, to the dataset of interest. The fundamental way to transfer knowledge is to reuse the weights leaned from huge dataset. The brought weights can be either frozen or fine-tuned with respect to a new small dataset. The transfer learning definitely showed a improvement on target performance. However, one drawback is that the performance depends on the similarity between the source and the target dataset (Azizpour et al, Proceedings of the IEEE conference on computer vision and pattern recognition workshops, 2015, [2]). In other words, the two datasets should be alike to be effective. Then finding the similar source dataset becomes another difficulty. To alleviate the problems, we propose a method that maximizes the effectiveness of the transferred weights regardless of what source data is used. Among the weights pre-trained with source data, only the ones relevant to the target data is transferred. The relevance is measured statistically. In this way, we improved the classification accuracy of downsized 50 sub-class ImageNet 2012 by 2%.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Yosinski, J. et al.: How transferable are features in deep neural networks? Adv. Neural Inf. Process. Syst. (2014)
Azizpour, H. et al.: From generic to specific deep representations for visual recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (2015)
Jia, Y. et al.: Caffe: convolutional architecture for fast feature embedding (2014). arXiv:1408.5093
Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. (2012)
He, K. et al.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016)
Acknowledgements
This work was supported by Institute for Information communications Technology Promotion (IITP) grant funded by the Korea government (MSIT) (2017-0-01780, The technology development for event recognition/relational reasoning and learning knowledge based system for video understanding).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer International Publishing AG, part of Springer Nature
About this paper
Cite this paper
Jung, M., Kim, J. (2019). Selective Filter Transfer. In: Kim, JH., et al. Robot Intelligence Technology and Applications 5. RiTA 2017. Advances in Intelligent Systems and Computing, vol 751. Springer, Cham. https://doi.org/10.1007/978-3-319-78452-6_3
Download citation
DOI: https://doi.org/10.1007/978-3-319-78452-6_3
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-78451-9
Online ISBN: 978-3-319-78452-6
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)