Abstract
The suspension of foreign objects on high-voltage transmission lines is extremely harmful to the safety of the line. If it is not handled in time, it will easily cause phase-to-phase short circuit of the transmission line and even cause forest fires. Foreign object suspension is a small probability event with fewer existing samples. To use CNN to perform target classification detection, there is a problem of insufficient sample or sample imbalance. Aiming at the above problems that often occur in engineering applications of CNN, we propose a data enhancement algorithm based on GANs. The main idea of this algorithm is: Firstly, the pre-training model is used to extract the feature map of sample, and GANs is used to learn the feature map directly. Then, the feature map generated by GANs and the original data are used to train the classification layer of the pre-training model, so as to achieve the purpose of data enhancement and balancing samples, and then enhance the classification ability of the model. The experimental results show that the classification performance of several classical CNN models can be improved significantly by using this method in the case of insufficient sample and sample imbalance.
Foundation item: Supported by The National Natural Science Foundation of China (61701192); Shandong Provincial Key Research and Development Project (2017CXGC0810); Shandong Education Science Plan “Special Subject for Scientific Research of Educational Admission Examination” (ZK1337212B008).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Mirza, M., Osindero, S.: Conditional generative adversarial nets. Comput. Sci., 2672–2680 (2014)
Radford, A., Metz, L., Chintala, S.: Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv:1511.06434 (2015)
Goodfellow, L., Bengio, Y., Courville, A.: Deep Learning, Chinese version, vol. 1, pp. 327–329. Posts & Telecom Press, Beijing (2017)
Abdi, L., Hashemi, S.: To combat multi-class imbalanced problems by means of over-sampling techniques. Soft. Comput. 19, 3369–3385 (2015)
Zhang, H., Wang, Z.: A normal distribution-based over-sampling approach to imbalanced data classification. In: Tang, J., King, I., Chen, L., Wang, J. (eds.) ADMA 2011. LNCS (LNAI), vol. 7120, pp. 83–96. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-25853-4_7
Yosinski, J., Clune, J., Bengio, Y., et al.: How transferable are features in deep neural networks? vol. 27, pp. 3320–3328 (2014)
Pan, S.J., Yang, Q.: A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 22, 1345–1359 (2010)
Yin, X., et al.: CrossMine: efficient classification across multiple database relations. In: International Conference on Data Engineering. IEEE Computer Society (2004)
Chawla, N.V., Bowyer, K.W., Hall, L.O., Kegelmeyer, W.P.: SMOTE: synthetic minority over-sampling technique. JAIR 16, 321–357 (2002)
Han, H., Wang, W.-Y., Mao, B.-H.: Borderline-SMOTE: a new over-sampling method in imbalanced data sets learning. In: Huang, D.-S., Zhang, X.-P., Huang, G.-B. (eds.) ICIC 2005. LNCS, vol. 3644, pp. 878–887. Springer, Heidelberg (2005). https://doi.org/10.1007/11538059_91
Mariani, G., Scheidegger, F., Istrate, R., Bekas, C., Malossi, C.: BAGAN: data augmentation with balancing GAN. arXiv:1803.09655 (2018)
Arjovsky, M., Bottou, L.: Towards principled methods for training generative adversarial networks. In: NIPS 2016 Workshop on Adversarial Training. In review for ICLR (2017)
Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V., Radford, A., Chen, X.: Improved techniques for training gans. In: NIPS (2016)
Gulrajani, I., Ahmed, F., Arjovsky, M., et al.: Improved training of wasserstein GANs. arXiv:1704.00028 (2017)
Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556 (2014)
Krizhevsky, A, Nair, V, Hinton, G: CAFAR-10 (2014). http://www.cs.toronto.edu/kriz/cifar.html
Kingma, D., Ba, J.: Adam: a method for stochastic optimization. arXiv:1412.6980v8 (2014)
Bradley, A.P.: The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognit. 30, 1145–1159 (1997)
He, K., Zhang, X., Ren, S., et al.: Deep residual learning for image recognition, pp. 770–778 (2015)
Szegedy, C., Vanhoucke, V., Ioffe, S., et al.: Rethinking the inception architecture for computer vision. In: Computer Vision and Pattern Recognition, pp. 2818–2826 (2016)
Szegedy, C., Ioffe, S., Vanhoucke, V., et al.: Inception-v4, Inception-ResNet and the impact of residual connections on learning. arXiv:1602.07261 (2016)
Chollet, F.: Xception: deep learning with depthwise separable convolutions. arXiv:1610.02357 (2016)
Lin, M., Chen, Q., Yan, S.: Network in network. arXiv:1312.4400 (2013)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Dou, Y., Yu, X., Li, J. (2019). Feature GANs: A Model for Data Enhancement and Sample Balance of Foreign Object Detection in High Voltage Transmission Lines. In: Vento, M., Percannella, G. (eds) Computer Analysis of Images and Patterns. CAIP 2019. Lecture Notes in Computer Science(), vol 11679. Springer, Cham. https://doi.org/10.1007/978-3-030-29891-3_50
Download citation
DOI: https://doi.org/10.1007/978-3-030-29891-3_50
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-29890-6
Online ISBN: 978-3-030-29891-3
eBook Packages: Computer ScienceComputer Science (R0)