Skip to main content

Vision-Based Automatic Hair Follicular Unit Separation

  • Conference paper

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 9246))

Abstract

In this paper, a vision-based method is proposed to automatically recognize the hair follicles and plan the cutting path to separate them into units. By using color information and machine learning, hair area in the image can be recognized. And the interferences such as adipose shadows and scalpel parts will be eliminated by texture and area information. In order to recognize single piece of hair, a curve detection method is proposed which combine the linear Hough transform and the quadratic curve fitting method to detect hair pieces with follicles on them. After determining the location and distribution of hair follicles, based on the hair growth direction and the minimum external rectangle of hair area, cutting path will be planned to separate each follicular unit. Compared with the traditional artificial hair follicular unit separation, this method not only ensures the fitting accuracy, but also speeds up the processing speed.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Dou, G.: Reasons of hair loss, and methods of hair protection (in Chinese). J. Family Medicine 3, 17 (2014)

    Google Scholar 

  2. Zhao, D.: Reasons and solutions of hair loss (in Chinese). J. Health Expo. 12, 46–47 (2014)

    Google Scholar 

  3. Bernstein, R.M., Rassman, W.R.: The aesthetics of follicular transplantation. J. Dermatol. Surg. 23, 785–799 (1997)

    Article  Google Scholar 

  4. Rassman, W.R., Jones, R., Worton, E., Uyttendaele, H.: Follicular unit extraction: minimally invasive surgery for hair transplantation. J. Dermatologic Surgery 28(8), 720–728 (2002)

    Google Scholar 

  5. Comparing, F.U.T.: Pre-making recipient sites to increase graft survival in manual and robotic FUE procedures. J. Hair Transplant Forum Int. 4, 128–131 (2012)

    Google Scholar 

  6. Ji, W., Li, N., Li, J.: Application progress of Da Vinci surgical robot (in Chinese). J. Southeast of the National Defense Medicine 12, 427–430 (2010)

    Google Scholar 

  7. Paul, M., Rodriguez, E., Chitwood, W.R.: Robotics in cardiac surgery. J. Interactive cardiovascular and thoracic Surgery 9(3), 500–505 (2009)

    Article  Google Scholar 

  8. Lee, E.K., Baack, J., Duchene, D.A.: Survey of practicing urologists: robotic versus open radical prostatectomy. J. The Canadian Journal of Urology 17(2), 5094–5098 (2010)

    Google Scholar 

  9. Cho, J.E., Nezhat, F.R.: Robotics and gynecologic oncology: review of the literature. J. Journal of Minimally Invasive Gynecology 16(6), 669–681 (2009)

    Article  Google Scholar 

  10. Roukos, D.H.: The era of robotic surgery for colorectal cancer. J. Annals of Surgical Oncology 17(1), 338–347 (2010)

    Article  Google Scholar 

  11. Li, J.: Robots plant hair as fast as plant seeds. J. China’s High-Tech Enterprises 1, 72–73 (2004)

    Google Scholar 

  12. Robert, M., Bernstein, M.D., Zingaretti, G.: Robotic Recipient Site Creation in Hair Transplantation. J. Hair Transplant Forum Intl. 24(3), 95–97 (2014)

    Google Scholar 

  13. Shin, J.W., Kwon, S.H., Kim, S.A., Kim, J.Y., Im, N.J., Park, K.C., Huh, C.H.: Characteristics of robotically harvested hair follicles in Koreans. J. Journal of the American Academy of Dermatology 72(1), 146–150 (2015)

    Article  Google Scholar 

  14. Cortes, C., Vapnik, V.: Support-vector network. J. Machine learning 20(3), 273–297 (1995)

    MATH  Google Scholar 

  15. Tamura, H., Mori, S., Yamawaki, T.: Textural features corresponding to visual perception. J. IEEE Transactions on Systems, Man and Cybernetics 8(6), 460–473 (1978)

    Article  Google Scholar 

  16. Huang, Z., Guo, S., Liao, J.: Ellipse Hough transform based detection of red cells in the urinary sediments (in Chinese). J. The Computer System Application 9, 74–77 (2009)

    Google Scholar 

  17. Xu, L., Oja, E., Kultanen, P.: A new curve detection method: randomized Hough transform (RHT). J. Pattern Recognition Letters 11(5), 331–338 (1990)

    Article  Google Scholar 

  18. Parent, P., Zucker, S.W.: Trace inference, curvature consistency, and curve detection. J. IEEE Transactions on Pattern Analysis and Machine Intelligence 11(8), 823–839 (1989)

    Article  Google Scholar 

  19. Akima, H.: A new method of interpolation and smooth curve fitting based on local procedures. J. Journal of the ACM 17(4), 589–602 (1970)

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Hesheng Wang or Weidong Chen .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2015 Springer International Publishing Switzerland

About this paper

Cite this paper

Yang, B., Wang, H., Chen, W., Liang, Y. (2015). Vision-Based Automatic Hair Follicular Unit Separation. In: Liu, H., Kubota, N., Zhu, X., Dillmann, R. (eds) Intelligent Robotics and Applications. Lecture Notes in Computer Science(), vol 9246. Springer, Cham. https://doi.org/10.1007/978-3-319-22873-0_24

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-22873-0_24

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-22872-3

  • Online ISBN: 978-3-319-22873-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics