Abstract
With the large-scale use of various inspection methods such as drones, helicopters, and robots, the generated power inspection images have increased significantly, which has brought huge pressure on data storage and transmission. At the same time, with the rapid development of artificial intelligence technology in the electric power field, the cost of manual labeling required for model training has become a major pain point. This paper studies a transmission image deduplication technology based on ViT-Siamese cascade network, which reduces the amount of data and the cost of data annotation. This paper first investigates the research status of image similarity at home and abroad, and then studies the transmission image deduplication technology based on the ViT-Siamese cascade network, which greatly reduces the complexity of similarity calculation, and finally trains the model on the transmission image data set. Firstly, this paper investigates the research status of image similarity at domestic and international. Next, transmission image deduplication technology based on ViT-Siamese cascade network is studied, which greatly reduces the complexity of similarity calculation. And finally, the AI model is trained on transmission image dataset, and effectiveness and feasibility of the technology in transmission scene processing are verified by experiments.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Zhaoyun, Z., Guanfeng, H., Shihong, H., Zhi, Z.: Target detection in power checking the application and development trend of. J. Hubei Univ. Nat. (Natural Science edn.) 33(3) 6, 305-314 (2021). https://doi.org/10.13501/j.carolcarrollnki.42-1908/n.2021.09.012
Mingmin, F.: Design and application of cloud storage subsystem for power equipment inspection data. Sci. Technol. Wind 30, 157 (2019). https://doi.org/10.19392/j.cnki.1671-7341.201930136
Zhijian, S.: Research on Massive Image De-Duplication Algorithm. Chongqing Jiaotong University (2016)
Fengqing, H., Zhijian, S., Rui, Y.: Fast deduplication for massive images. J. Comput. Appl. 36(07), 1797–1800 (2016)
Tengfei, P.: The aurora image retrieval based on convolution neural network research. Nanjing University of Posts and Telecommunications (2021). 10.27251
Bromley, J., et al.: Signature verification using a “siamese” time delay neural network. Int. J. Patt. Recogn. Artif. Intell. 07(04), 669–688 (1993)
Chopar, S., Hadsell, R., Lecun, Y.: With application to face verification. In: 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 113–119. IEEE, San Diego(2005)
Nair, V., Geoffrey, E.H.: Rectified linear units improve restricted boltzmann machines. In: Proceedings of the 27th International Conference on Machine Learning, Haifa, Israel, 21–24 Jun 2010
Zagoruyko, S., Komodakis, N.: Learning to compare image patches via convolutional neural networks. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE (2015)
Everingham, M., Van Gool, L., Williams, C.K.I., Winn, J., Zisserman, A.: The pascal visual object classes (VOC) challenge. Int. J. Comput. Vis. 88(2), 303–338 (2010)
Vaswani, A., Shazeer, N., Parmar, N.: Attention is all you need. Adv. Neural Inform. Process. Syst. Long Beach 6000–6010 (2017)
Parmar, N., et al.: Image Transformer. ArXiv: 1802.05751 [cs. CV]. 15 Feb 2018. https://arxiv.org/abs/1802.05751
Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 213–229. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-58452-8_13
Zhu, X.Z., Su, W.J., Lu, L.W., Li, B., Wang, X., Dai J.: Deformable DETR: Deformable Transformers for End-to-End Object Detection. https://arxiv.org/ABS/2010.04159 (2020)
Zheng, M.H., Gao, P.; Wang, X.G., et al.: Adaptive Clustering Transformer with end-to-end Object Detection. ArXiv. https://arxiv.org/ABS/2011.09315 V1 (2022)
Liu, L., et al.: Deep learning for generic object detection: a survey. Int. J. Comput. Vision 128(2), 261–318 (2019). https://doi.org/10.1007/s11263-019-01247-4
Gao, X.Y., Hoi, S.C.H., Zhang, Y.D., Zhou, J.S., Wan, J.: Sparse online learning of image similarity. ACM Trans. Intell. Syst. Technol. 8(5), 641–6422 (2017)
Zhang, Y., Gao, X., Chen, Z., Zhong, H., Xie, H., Yan, C.: Mining spatial-temporal similarity for visual tracking. IEEE Trans. Image Process. 29, 8107–8119 (2020)
Xia, Z.X., Hong, X.P., Gao, X.Y., Feng, X.Y., Zhao, G.Y.: Spatiotemporal recurrent convolutional networks for recognizing spontaneous micro-expressions. IEEE Trans. Multimed. 22(3), 626–640 (2020)
Gao, X.Y., Xie, J.Y., Chen, Z.Y., Liu, A.A., Sun, Z., Lyu, L.: Dilated convolution-based feature refinement network for crowd localization. ACM Trans. Multimedia Comput., Commun. Appl. (2022)
Dosovitskiy, A., et al.: An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv:2010.11929 (2020)
Acknowledgment
This work was funded by the “Research on the key technologies of zero sample knowledge transfer learning and defect recognition for fine-grained goals” program of the Big Data Center, State Grid Corporation of China.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Chen, Z., Chen, S., Peng, X., Bian, J., Jiang, L., Zhang, X. (2023). ViT-Siamese Cascade Network for Transmission Image Deduplication. In: Zhai, G., Zhou, J., Yang, H., Yang, X., An, P., Wang, J. (eds) Digital Multimedia Communications. IFTC 2022. Communications in Computer and Information Science, vol 1766. Springer, Singapore. https://doi.org/10.1007/978-981-99-0856-1_29
Download citation
DOI: https://doi.org/10.1007/978-981-99-0856-1_29
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-0855-4
Online ISBN: 978-981-99-0856-1
eBook Packages: Computer ScienceComputer Science (R0)