Abstract
Accessing the venous bloodstream to obtain a blood sample is the most common clinical routine. Nevertheless, due to the reliance of venipuncture on manual technique, first-stick accuracy of venipuncture falls below 50% in difficult cases. A surge of research on robotic guidance for autonomous vascular access have been conducted. With regard to robotic venipuncture, efficiency and accuracy of vein segmentation is of much importance. This paper describes a method to accurately and efficiently segment, localize and track the topology of human veins from near-infrared (NIR) images. Both spatial and color augmentation are implemented on the dataset at first. Next, Mixer-UNet is used for identifying veins that would be hard to find in clinical visual assessment. The Mixer-UNet is developed on the basis of UNet and MLP-Mixer. Through the flexible information exchange through Token-mixing layer and Channel-mixing layer, Mixer-UNet can extract features from NIR images accurately. The performance of Mixer-UNet is validated on 270 NIR images, which are collected from 30 volunteers. Mixer-UNet reaches 93.07% on Accuracy indicator. Compared with the best-performing baseline, the F1-score indicator increases by 2.82%, reaching 78.37% in testing sample. The high accuracy and robustness of Mixer-UNet is expected to improve the vein segmentation of NIR images, and further contributes to the goal of an improved automated venipuncture robot.
This work is supported by the National Natural Science Foundation of China (51905379), Shanghai Science and Technology Development Funds (20QC1400900), the Shanghai Municipal Science and Technology Major Project (2021SHZDZX0100) and the Fundamental Research Funds for the Central Universities.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Chen, A.: Image-guided robotics for autonomous venipuncture. Ph.D. dissertation, Department, BME, Rutgers University, New Brunswick (2016)
De Boer, T., Steinbuch, M., Neerken, S., Kharin, A.: Laboratory study on needle-tissue interaction: towards the development of an instrument for automatic venipuncture. J. Mech. Med. Biol. 07(03), 325–335 (2007)
Lee, S., et al.: A transparent bending-insensitive pressure sensor. Nat. Nanotechnol. 11(5), 472–478 (2016)
Paulo, C., Anurag, K., Sean, W., Patrick, F., Gregory, S.: Robotic assistive device for phlebotomy. In: Proceedings of International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, Boston, Massachusetts, USA, vol. 57106, p. V003T14A012 (2015)
Unger, M., Berger, J., Gerold, B., Melzer, A.: Robot-assisted ultrasound-guided tracking of anatomical structures for the application of focused ultrasound. Curr. Dir. Biomed. Eng. 6(3), 123–126 (2020)
Chen, A., Balter, M., Maguire, T., Yarmush, M.: Deep learning robotic guidance for autonomous vascular access. Nat. Mach. Intell. 2(2), 104–115 (2020)
Balter, M.: Robotic devices for automated venipuncture and diagnostic blood analysis. Ph.D. dissertation, Department, BME, Rutgers University, New Brunswick (2017)
Huang, D., et al.: Autonomous robotic subcutaneous injection under near-infrared image guidance. In: International Design Engineering Technical Conferences and Computers and Information in Engineering Conference (2021)
Chen, Y., et al.: Semi-supervised vein segmentation of ultrasound images for autonomous venipuncture. In: Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, pp. 9475–9481 (2021)
Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234–241. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-24574-4_28
Tolstikhin, I., et al.: MLP-mixer: an all-MLP architecture for vision. In: Advances in Neural Information Processing Systems, vol. 34, pp. 24261–24272 (2021)
Lin, T., Dollar, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of IEEE Conference on Computer Vision And Pattern Recognition, pp. 2117–2125 (2017)
Zhao, H., Shi, J., Qi, X., Wang, X., Jia, J.: Pyramid scene parsing network. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, USA, pp. 2881–2890 (2017)
Chaurasia, A., Culurciello, E.: LinkNet: exploiting encoder representations for efficient semantic segmentation. In: Proceedings of IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, pp. 1–4 (2017)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Ji, J., Zhao, Y., Xie, T., Du, F., Qi, P. (2022). Automated Vein Segmentation from NIR Images Using a Mixer-UNet Model. In: Liu, H., et al. Intelligent Robotics and Applications. ICIRA 2022. Lecture Notes in Computer Science(), vol 13458. Springer, Cham. https://doi.org/10.1007/978-3-031-13841-6_6
Download citation
DOI: https://doi.org/10.1007/978-3-031-13841-6_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-13840-9
Online ISBN: 978-3-031-13841-6
eBook Packages: Computer ScienceComputer Science (R0)