Abstract
This article aims at the problems that exist in the monitoring process of employees’ standard clothing. Adopting the reverse way of thinking to carry out the normative identification work, that is, the classification ability depends on the feature fitting ability. First of all, building an employee work card template library, which is used to realize personnel authentication in combination with template matching method. The recognition result is given as a label to the real captured image. Then, the construction of the auxiliary categorical-generative adversarial network was done with a few real images. The training of generator G and discriminator D is completed through iterative interactive update, so as to realize the fitting of the real clothing image. The trained ACat-GAN is used as a standard clothing monitoring model. Respectively, based on the discriminant image and parameter adjustment data, discriminator D and generator G push out the list of well-dressed people and Image of non-standard. Based on a power supply business hall as the background platform, this article collected personnel clothing images for experiments. The experimental results show the feasibility and practicability of the method which described in this paper.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Kaibao, F., Xiang, Z.: Current situation analysis and policy recommendations of electricity market reform. Macroecon. Manag. 33(01), 49-54 (2018)
Chen, Y., Cai, X., Zeng, Y., Wang, M.: Crossing-scene pedestrian identification method based on twice FAS. In: Zou, B., Han, Q., Sun, G., Jing, W., Peng, X., Lu, Z. (eds.) ICPCSEE 2017. CCIS, vol. 728, pp. 483–491. Springer, Singapore (2017). https://doi.org/10.1007/978-981-10-6388-6_41
Dong, J., Zheng, B., Yang, Z.: License plate character recognition based on convolutional neural network. J. Comput. Appl. 37(07), 2014–2018 (2017)
Zhu, Y., Zhao, J., Wang, Y.: A review of human behavior recognition algorithms based on deep learning. Acta Autom. Sin. 42(06), 848–857 (2016)
Wang, K., Gu, C., Duan, Y., et al.: Generative adversarial networks: the state of the art and beyond. Acta Autom. Sin. 43(03), 321–332 (2017)
Springenberg, J.T.: Unsupervised and Semi-supervised Learning with Categorical Generative Adversarial Networks. In: International Conference on Learning Representations (2016). arXiv:1511.06390v2
Goodfellow, I.J., Pouget-Abadie, J., Mirza, M., et al.: Generative adversarial nets. In: Proceedings of the 2014 Conference on Advances in Neural Information Processing Systems 27, Montreal, Canada, pp. 2672 − 2680. Curran Associates (2014)
Shi, L., Wang, Y., Cao, Y., Wei, L.: Vehicle type recognition based on deep convolution neural network. In: Zou, B., Han, Q., Sun, G., Jing, W., Peng, X., Lu, Z. (eds.) ICPCSEE 2017. CCIS, vol. 728, pp. 492–502. Springer, Singapore (2017). https://doi.org/10.1007/978-981-10-6388-6_42
Dong, Y., Chen, Z., Song, H.: Application of improved feature matching algorithm in bank card number recognition. J. Ji lin Univ. (Sci Edit.) 56(01), 126–129 (2018)
Ma, Y.: Research on Clothing Identification Algorithm for Downstream People in Video Surveillance Scenes. Anhui University (2015)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Qiao, X., Rong, Y., Liu, Y., Jiang, T. (2018). Research on Electricity Personnel Apparel Monitoring Model Based on Auxiliary Categorical-Generative Adversarial Network. In: Zhou, Q., Miao, Q., Wang, H., Xie, W., Wang, Y., Lu, Z. (eds) Data Science. ICPCSEE 2018. Communications in Computer and Information Science, vol 902. Springer, Singapore. https://doi.org/10.1007/978-981-13-2206-8_31
Download citation
DOI: https://doi.org/10.1007/978-981-13-2206-8_31
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-13-2205-1
Online ISBN: 978-981-13-2206-8
eBook Packages: Computer ScienceComputer Science (R0)