Abstract
The application of computer vision technology to detect prohibited items in X-ray security inspection images holds significant practical research value. We have observed that object detection models built upon the Visual Transformer (ViT) architecture outperform those relying on Convolutional Neural Networks (CNNs) when assessed on publicly available datasets. However, the ViT’s attention mechanism, while offering a global response, lacks the CNN model’s inductive bias, which can hinder its performance, demanding more samples and learning parameters. This drawback is particularly problematic for time-sensitive processes like security inspections. This research paper aims to develop a lightweight prohibited item detection model grounded in the ViT framework, utilizing MobileViT as the underlying network for feature extraction. To enhance the model’s sensitivity to small object features, we have established dense connections among various network layers. This design ensures effective integration of both high- and low-level visual features without increasing computational complexity. Additionally, learnable group convolutions are employed to replace traditional convolutions, further reducing model parameters and computational demands. Simulation experiments conducted on the publicly available SIXray dataset validate the effectiveness of the proposed model in this study. The code is publicly accessible at https://github.com/zhg-SZPT/MVray.
This work was supported by the School-level Project of Shenzhen Polytechinc University (No. 6022310006K), Research Projects of Department of Education of Guangdong Province (No. 2020ZDZX3082, 2023ZDZX1081, 2023KCXTD077).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Mery, D.: X-ray testing by computer vision. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 360–367 (2013)
Liang, T., Lv, B., Zhang, N., Yuan, J., Zhang, Y., Gao, X.: Prohibited items detection in x-ray images based on attention mechanism. J. Phys. Conf. Ser. 1986(1), 012087 (6pp) (2021)
An Chang, Yu., Zhang, S.Z., Zhong, L., Zhang, L.: Detecting prohibited objects with physical size constraint from cluttered x-ray baggage images. Knowl.-Based Syst. 237, 107916 (2022)
Tao, R., et al.: Over-sampling de-occlusion attention network for prohibited items detection in noisy x-ray images. arXiv preprint arXiv:2103.00809 (2021)
Zhang, Y., Zhang, H., Zhao, T., Yang, J.: Automatic detection of prohibited items with small size in x-ray images. Optoelectron. Lett. 16(4), 313–317 (2020)
Tao, R., et al.: Towards real-world x-ray security inspection: a high-quality benchmark and lateral inhibition module for prohibited items detection. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 10923–10932 (2021)
Nguyen, H.D., Cai, R., Zhao, H., Kot, A.C., Wen, B.: Towards more efficient security inspection via deep learning: a task-driven x-ray image cropping scheme. Micromachines 13(4), 565 (2022)
Howard, A., et al.: Searching for mobilenetv3. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 1314–1324 (2019)
Ren, Yu., Zhang, H., Sun, H., Ma, G., Ren, J., Yang, J.: LightRay: lightweight network for prohibited items detection in x-ray images during security inspection. Comput. Electr. Eng. 103, 108283 (2022)
Ghosh, S., Srinivasa, S.K.K., Amon, P., Hutter, A., Kaup, A.: Deep network pruning for object detection. In: 2019 IEEE International Conference on Image Processing (ICIP), pp. 3915–3919. IEEE (2019)
Howard, A.G., et al.: MobileNets: efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017)
Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.-C.: MobileNetV 2: Inverted residuals and linear bottlenecks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510–4520 (2018)
Zhang, X., Zhou, X., Lin, M., Sun, J.: ShuffleNet: an extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6848–6856 (2018)
Ma, N., Zhang, X., Zheng, H.-T., Sun, J.: ShuffleNet V2: practical guidelines for efficient CNN architecture design. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) Computer Vision – ECCV 2018. LNCS, vol. 11218, pp. 122–138. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01264-9_8
Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)
Mehta, S., Rastegari, M.: MobileViT: light-weight, general-purpose, and mobile-friendly vision transformer (2022)
Huang, G., Liu, Z., Van Der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700–4708 (2017)
Miao, C., et al.: SIXray: a large-scale security inspection x-ray benchmark for prohibited item discovery in overlapping images. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2119–2128 (2019)
Bochkovskiy, A., Wang, C.-Y., Liao, H.-Y.M.: YOLOv4: optimal speed and accuracy of object detection (2020)
Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of NAACL-HLT, vol. 1, p. 2 (2019)
Han, K., et al.: A survey on vision transformer. IEEE Trans. Pattern Anal. Mach. Intell. 45(1), 87–110 (2022)
Chu, X., et al.: Conditional positional encodings for vision transformers. arXiv preprint arXiv:2102.10882 (2021)
Wu, H., et al.: CVT: introducing convolutions to vision transformers. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 22–31 (2021)
Yuan, K., Guo, S., Liu, Z., Zhou, A., Yu, F., Wu, W.: Incorporating convolution designs into visual transformers. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 579–588 (2021)
Mery, D., Svec, E., Arias, M., Riffo, V., Saavedra, J.M., Banerjee, S.: Modern computer vision techniques for x-ray testing in baggage inspection. IEEE Trans. Syst. Man Cybernet. Syst. 47(4), 682–692 (2016)
Akcay, S., Breckon, T.: Towards automatic threat detection: a survey of advances of deep learning within x-ray security imaging. Pattern Recogn. 122, 108245 (2022)
Aydin, I., Karakose, M., Akin, E.: A new approach for baggage inspection by using deep convolutional neural networks. In: 2018 International Conference on Artificial Intelligence and Data Processing (IDAP), pp. 1–6. IEEE (2018)
Gaus, Y.F.A., Bhowmik, N., Akçay, S., Guillén-Garcia, P.M., Barker, J.W., Breckon, T.P.: Evaluation of a dual convolutional neural network architecture for object-wise anomaly detection in cluttered x-ray security imagery. In: 2019 International Joint Conference on Neural Networks (IJCNN), pp. 1–8. IEEE (2019)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778 (2016)
Han, K., Wang, Y., Tian, Q., Guo, J., Xu, C., Xu, C.: GhostNet: more features from cheap operations. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1580–1589 (2020)
Kim, J.-H., Kim, N., Park, Y.W., Won, C.S.: Object detection and classification based on yolo-v5 with improved maritime dataset. J. Marine Sci. Eng. 10(3), 377 (2022)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Sun, P., Zhang, H., Yang, J., Wei, D. (2023). MobileViT Based Lightweight Model for Prohibited Item Detection in X-Ray Images. In: Lu, H., Blumenstein, M., Cho, SB., Liu, CL., Yagi, Y., Kamiya, T. (eds) Pattern Recognition. ACPR 2023. Lecture Notes in Computer Science, vol 14407. Springer, Cham. https://doi.org/10.1007/978-3-031-47637-2_4
Download citation
DOI: https://doi.org/10.1007/978-3-031-47637-2_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-47636-5
Online ISBN: 978-3-031-47637-2
eBook Packages: Computer ScienceComputer Science (R0)