skip to main content
10.1145/3644116.3644239acmotherconferencesArticle/Chapter ViewAbstractPublication PagesisaimsConference Proceedingsconference-collections
research-article

A New 3D Human Pose Estimation Network for Knee Posture Estimation

Published: 05 April 2024 Publication History

Abstract

In lower extremity bone surgery, the surgeons need to adjust the patient's posture to expose the surgical site, which is usually done manually by the operator. However, manually adjusting is time-consuming and labor-intensive. An automatic positioning system (APS) is more preferable over the manual operation. The APS includes an electromechanic setup, a monocular camera and a computer. A monocular camera is used to detect the patient's lower limb postures, and the positioning device adjusts the patient's knee to the desired postures. This study employs a 3D human pose estimation (HPE) network, which contains a dual-attention module, to estimate the patient's knee joint angle and provide physicians with a measure of the patient's knee pose automatically from monocular images. First, a trained cascaded pyramid network (CPN) is used to obtain the 2D joint points of human images. Then, the detected two-dimensional coordinates, as the input for the proposed 3D HPE network, are defined in the three-dimensional space. The flexion and varus-valgus angles of knee joint are computed by establishing the knee coordinate system eventually. The 3D HPE network proposed in this study considers both spatial and temporal information to obtain the correlation between the spatial and temporal dimensions of the feature map. Subsequently, dilated convolution is used to extract features from the spatial and temporal attention feature maps, which realize a better estimation of the human body in 3D posture. The new model was tested on the Human3.6M dataset and showed that the mean per joint position error (MPJPE) is 48.6mm, and the errors of the flexion and varus-valgus angles are 7.9° and 12.7°, which are acceptable results for knee posture prediction compared to the manually placement of the patient postures.

References

[1]
Burlingame, B. L. 2017. Guideline implementation: positioning the patient. AORN journal, 106(3), 227-237. https://doi.org/10.1016/j.aorn.2017.07.010
[2]
Waelkens, P., van Oosterom, M. N., van den Berg, N. S., Navab, N., & van Leeuwen, F. W. 2016. Surgical navigation: an overview of the state-of-the-art clinical applications. Radioguided Surgery: Current Applications and Innovative Directions in Clinical Practice, 57-73. https://doi.org/10.1007/978-3-319-26051-8_4
[3]
Liu, W., Bao, Q., Sun, Y., & Mei, T. 2022. Recent advances of monocular 2d and 3d human pose estimation: A deep learning perspective. ACM Computing Surveys, 55(4), 1-41. https://doi.org/10.1145/3524497
[4]
Wang, J., Tan, S., Zhen, X., Xu, S., Zheng, F., He, Z., & Shao, L. 2021. Deep 3D human pose estimation: A review. Computer Vision and Image Understanding, 210, 103225. https://doi.org/10.1016/j.cviu.2021.103225
[5]
Pavlakos, G., Zhou, X., Derpanis, K. G., & Daniilidis, K. 2017. Coarse-to-fine volumetric prediction for single-image 3D human pose. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 7025-7034). https://doi.org/10.48550/arXiv.1611.07828
[6]
Sun, X., Shang, J., Liang, S., & Wei, Y. 2017. Compositional human pose regression. In Proceedings of the IEEE international conference on computer vision (pp. 2602-2611). https://doi.org/10.48550/arXiv.1704.00159
[7]
Cao, Z., Simon, T., Wei, S. E., & Sheikh, Y. 2017. Realtime multi-person 2d pose estimation using part affinity fields. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 7291-7299). https://doi.org/10.48550/arXiv.1611.08050
[8]
Chen, Y., Wang, Z., Peng, Y., Zhang, Z., Yu, G., & Sun, J. 2018. Cascaded pyramid network for multi-person pose estimation. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 7103-7112). https://doi.org/10.48550/arXiv.1711.07319
[9]
Sun, K., Xiao, B., Liu, D., & Wang, J. 2019. Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 5693-5703). https://doi.org/10.48550/arXiv.1902.09212
[10]
Martinez, J., Hossain, R., Romero, J., & Little, J. J. 2017. A simple yet effective baseline for 3d human pose estimation. In Proceedings of the IEEE international conference on computer vision (pp. 2640-2649). https://doi.org/10.48550/arXiv.1705.03098
[11]
Tekin, B., Márquez-Neila, P., Salzmann, M., & Fua, P. 2017. Learning to fuse 2d and 3d image cues for monocular body pose estimation. In Proceedings of the IEEE international conference on computer vision (pp. 3941-3950). https://doi.org/10.48550/arXiv.1611.05708
[12]
Hossain, M. R. I., & Little, J. J. 2018. Exploiting temporal information for 3d human pose estimation. In Proceedings of the European conference on computer vision (ECCV) (pp. 68-84). https://doi.org/10.48550/arXiv.1711.08585
[13]
Arnab, A., Doersch, C., & Zisserman, A. 2019. Exploiting temporal context for 3D human pose estimation in the wild. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 3395-3404). https://doi.org/10.48550/arXiv.1905.04266
[14]
Wang, J., Huang, S., Wang, X., & Tao, D. 2019. Not all parts are created equal: 3d pose estimation by modeling bi-directional dependencies of body parts. In Proceedings of the IEEE/CVF international conference on computer vision (pp. 7771-7780). https://doi.org/10.48550/arXiv.1905.07862
[15]
Ma, X., Su, J., Wang, C., Ci, H., & Wang, Y. 2021. Context modeling in 3d human pose estimation: A unified perspective. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 6238-6247). https://doi.org/10.48550/arXiv.2103.15507
[16]
Ionescu, C., Papava, D., Olaru, V., & Sminchisescu, C. 2013. Human3. 6m: Large scale datasets and predictive methods for 3d human sensing in natural environments. IEEE transactions on pattern analysis and machine intelligence, 36(7), 1325-1339. https://doi.org/10.1109/TPAMI.2013.248

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Other conferences
ISAIMS '23: Proceedings of the 2023 4th International Symposium on Artificial Intelligence for Medicine Science
October 2023
1394 pages
ISBN:9798400708138
DOI:10.1145/3644116
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 05 April 2024

Permissions

Request permissions for this article.

Check for updates

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

ISAIMS 2023

Acceptance Rates

Overall Acceptance Rate 53 of 112 submissions, 47%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 12
    Total Downloads
  • Downloads (Last 12 months)12
  • Downloads (Last 6 weeks)1
Reflects downloads up to 17 Jan 2025

Other Metrics

Citations

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

HTML Format

View this article in HTML Format.

HTML Format

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media