Skip to main content
Log in

A Fast Compression Framework Based on 3D Point Cloud Data for Telepresence

  • Research Article
  • Published:
International Journal of Automation and Computing Aims and scope Submit manuscript

Abstract

In this paper, a novel compression framework based on 3D point cloud data is proposed for telepresence, which consists of two parts. One is implemented to remove the spatial redundancy, i.e., a robust Bayesian framework is designed to track the human motion and the 3D point cloud data of the human body is acquired by using the tracking 2D box. The other part is applied to remove the temporal redundancy of the 3D point cloud data. The temporal redundancy between point clouds is removed by using the motion vector, i.e., the most similar cluster in the previous frame is found for the cluster in the current frame by comparing the cluster feature and the cluster in the current frame is replaced by the motion vector for compressing the current frame. The first, the B-SHOT (binary signatures of histograms orientation) descriptor is applied to represent the point feature for matching the corresponding point between two frames. The second, the K-mean algorithm is used to generate the cluster because there are a lot of unsuccessfully matched points in the current frame. The matching operation is exploited to find the corresponding clusters between the point cloud data of two frames. Finally, the cluster information in the current frame is replaced by the motion vector for compressing the current frame and the unsuccessfully matched clusters in the current and the motion vectors are transmitted into the remote end. In order to reduce calculation time of the B-SHOT descriptor, we introduce an octree structure into the B-SHOT descriptor. In particular, in order to improve the robustness of the matching operation, we design the cluster feature to estimate the similarity between two clusters. Experimental results have shown the better performance of the proposed method due to the lower calculation time and the higher compression ratio. The proposed method achieves the compression ratio of 8.42 and the delay time of 1 228 ms compared with the compression ratio of 5.99 and the delay time of 2 163 ms in the octree-based compression method under conditions of similar distortion rate.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. C. G. Yang, Y. H. Ye, X. Y. Li, R. W. Wang. Development of a neuro-feedback game based on motor imagery EEG. Multimedia Tools and Applications, vol. 77, no. 12, pp. 15929–15949, 2018. DOI: https://doi.org/10.1007/s11042-017-5168-x.

    Article  Google Scholar 

  2. F. Nagata, K. Watanabe, M. K. Habib. Machining robot with vibrational motion and 3D printer-like data interface. International Journal of Automation and Computing, vol. 15, no. 1, pp. 1–12, 2018. DOI: https://doi.org/10.1007/s11633-017-1101-z.

    Article  Google Scholar 

  3. C. G. Yang, H. W. Wu, Z. J. Li, W. He, N. Wang, C. Y. Su. Mind control of a robotic arm with visual fusion technology. IEEE Transactions on Industrial Informatics, vol. 14, no. 9, pp. 3822–3830, 2018. DOI: https://doi.org/10.1109/TII.2017.2785415.

    Article  Google Scholar 

  4. X. Y. Wang, C. G. Yang, Z. J. Ju, H. B. Ma, M. Y. Fu. Robot manipulator self-identification for surrounding obstacle detection. Multimedia Tools and Applications, vol. 76, no. 5, pp. 6495–6520, 2017. DOI: https://doi.org/10.1007/s11042-016-3275-8.

    Article  Google Scholar 

  5. J. H. Zhang, M. Li, Y. Feng, C. G. Yang. Robotic grasp detection based on image processing and random forest. Multimedia Tools and Applications, vol. 79, no. 3–4, pp. 2427–2446, 2020. DOI: https://doi.org/10.1007/s11042-019-08302-9.

    Article  Google Scholar 

  6. H. Y. Chen, H. L. Huang, Y. Qin, Y. J. Li, Y. H. Liu. Vision and laser fused slam in indoor environments with multi-robot system. Assembly Automation, vol. 39, no. 2, pp. 297–307, 2019. DOI: https://doi.org/10.1108/AA-04-2018-065.

    Article  Google Scholar 

  7. Y. Yang, F. Qiu, H. Li, L. Zhang, M. L. Wang, M. Y. Fu. Large-scale 3D semantic mapping using stereo vision. International Journal of Automation and Computing, vol. 15, no. 2, pp. 194–206, 2018. DOI: https://doi.org/10.1007/s11633-018-1118-y.

    Article  Google Scholar 

  8. J. Oyekan, A. Fischer, W. Hutabarat, C. Turner, A. Tiwari. Utilising low cost RGB-D cameras to track the real time progress of a manual assembly sequence. Assembly Automation, to be published. DOI: https://doi.org/10.1108/AA-06-2018-078.

  9. G. L. Wang, X. T. Hua, J. Xu, L. B. Song, K. Chen. A deep learning based automatic surface segmentation algorithm for painting large-size aircraft with 6-DOF robot. Assembly Automation, vol. 40, no. 2, pp. 199–210, 2019. DOI: https://doi.org/10.1108/AA-03-2019-0037.

    Article  Google Scholar 

  10. J. W. Li, W. Gao, Y. H. Wu. Elaborate scene reconstruction with a consumer depth camera. International Journal of Automation and Computing, vol. 15, no. 4, pp. 443–453, 2018. DOI: https://doi.org/10.1007/s11633-018-1114-2.

    Article  Google Scholar 

  11. C. G. Yang, Z. R. Wang, W. He, Z. J. Li. Development of a fast transmission method for 3D point cloud. Multimedia Tools and Applications, vol. 77, no. 19, pp. 25369–25387, 2018. DOI: https://doi.org/10.1007/s11042-018-5789-8.

    Article  Google Scholar 

  12. S. M. Prakhya, B. B. Liu, W. S. Lin, V. Jakhetiya, S. C. Guntuku. B-shot: A binary 3D feature descriptor for fast Keypoint matching on 3D point clouds. Autonomous Robots, vol. 41, no. 7, pp. 1501–1520, 2017. DOI: https://doi.org/10.1007/s10514-016-9612-y.

    Article  Google Scholar 

  13. J. H. Hou, L. P. Chau, N. Magnenat-Thalmann, Y. He. Compressing 3-D human motions via Keyframe-based geometry videos. IEEE Transactions on Circuits and Systems for Video Technology, vol. 25, no. 1, pp. 51–62, 2015. DOI: https://doi.org/10.1109/TCSVT.2014.2329376.

    Article  Google Scholar 

  14. J. Wingbermuehle. Towards automatic creation of realistic anthropomorphic models for realtime 3D telecommunication. Journal of VLSI Signal Processing Systems for Signal, Image and Video Technology, vol. 20, no. 1–2, pp. 81–96, 1998. DOI: https://doi.org/10.1023/A:1008018307114.

    Article  Google Scholar 

  15. J. H. Zhang, C. B. Owen. Octree-based animated geometry compression. Computers & Graphics, vol. 31, no. 3, pp. 463–479, 2007. DOI: https://doi.org/10.1016/j.cag.2006.12.002.

    Article  Google Scholar 

  16. Q. H. Yu, W. Yu, J. H. Zheng, X. Z. Zheng, Y. He, Y. C. Rong. A high-throughput and low-complexity decoding scheme based on logarithmic domain. Journal of Signal Processing Systems, vol. 88, no. 3, pp. 245–257, 2017. DOI: https://doi.org/10.1007/s11265-016-1143-4.

    Article  Google Scholar 

  17. C. Loop, C. Zhang, Z. Y. Zhang. Real-time high-resolution sparse voxelization with application to image-based modeling. In Proceedings of the 5th High-performance Graphics Conference, ACM, New York, USA, pp. 73–79, 2013. DOI: https://doi.org/10.1145/2492045.2492053.

    Chapter  Google Scholar 

  18. R. Schnabel, R. Klein. Octree-based point-cloud compression. In Proceedings of the 3rd Symposium on Point-based Graphics, ACM, Boston, USA, pp. 111–121, 2006.

    Google Scholar 

  19. J. Kammerl, N. Blodow, R. B. Rusu, S. Gedikli, M. Beetz, E. Steinbach. Real-time compression of point cloud streams. In Proceedings of IEEE International Conference on Robotics and Automation, IEEE, Saint Paul, USA, pp. 778–785, 2012. DOI: https://doi.org/10.1109/ICRA.2012.6224647.

    Google Scholar 

  20. C. Zhang, D. Florêncio, C. Loop. Point cloud attribute compression with graph transform. In Proceedings of IEEE International Conference on Image Processing, IEEE, Paris, France, pp. 2066–2070, 2014. DOI: https://doi.org/10.1109/ICIP.2014.7025414.

    Google Scholar 

  21. D. Sedlacek, J. Zara. Graph cut based point-cloud segmentation for polygonal reconstruction. In Proceedings of the 5th International Symposium on Visual Computing, Springer, Las Vegas, USA, pp. 218–227, 2009. DOI: https://doi.org/10.1007/978-3-642-10520-3_20.

    Google Scholar 

  22. L. Landrieu, C. Mallet, M. Weinmann. Comparison of belief propagation and graph-cut approaches for contextual classification of 3D lidar point cloud data. In Proceedings of IEEE International Geoscience and Remote Sensing Symposium, IEEE, Fort Worth, USA, pp. 2768–2771, 2017. DOI: https://doi.org/10.1109/IGARSS.2017.8127571.

    Google Scholar 

  23. X. M. Zhang, W. G. Wan, X. D. An. Clustering and DCT based color point cloud compression. Journal of Signal Processing Systems, vol. 86, no. 1, pp. 41–49, 2017. DOI: https://doi.org/10.1007/s11265-015-1095-0

    Article  Google Scholar 

  24. J. Euh, J. Chittamuru, W. Burleson. Power-aware 3D computer graphics rendering. Journal of VLSI Signal Processing Systems for Signal, Image and Video Technology, vol. 39, no. 1–2, pp. 15–33, 2005. DOI: https://doi.org/10.1003/B:VLSI.0000047269.03965.e9.

    Article  Google Scholar 

  25. D. Thanou, P. A. Chou, P. Frossard. Graph-based compression of dynamic 3D point cloud sequences. IEEE Transactions on Image Processing, vol. 25, no. 4, pp. 1765–1778, 2016. DOI: https://doi.org/10.1109/TIP.2016.2529506.

    Article  MathSciNet  Google Scholar 

  26. Y. T. Shao, Q. Zhang, G. Li, Z. Li, L. Li. Hybrid point cloud attribute compression using slice-based layered structure and block-based intra prediction. In Proceedings of the 26th ACM Multimedia Conference on Multimedia Conference, ACM, Istanbul, Turkey, pp. 1199–1207, 2018.

    Chapter  Google Scholar 

  27. P. M. Djuric, J. H. Kotecha, J. Zhang, Y. F. Huang, T. Ghirmai, M. F. Bugallo, J. Miguez. Particle filtering. IEEE Signal Processing Magazine, vol. 20, no. 5, pp. 19–38, 2003. DOI: https://doi.org/10.1109/MSP.2003.1236770.

    Article  Google Scholar 

  28. Z. Chen. Bayesian filtering: From Kalman filters to particle filters, and beyond. Statistics: A Journal of Theoretical and Applied Statistics, vol. 182, no. 1, pp. 1–69, 2003.

    Article  Google Scholar 

  29. J. S. Liu, R. Chen, T. Logvinenko. A theoretical framework for sequential importance sampling with resampling. Sequential Monte Carlo Methods in Practice, A. Doucet, N. de Freitas, N. Gordon, Eds., New York, USA: Springer, pp. 225–246, 2001. DOI: https://doi.org/10.1007/978-1-4757-3437-9_11.

    Chapter  Google Scholar 

  30. X. H. Liu, S. Payandeh. Implementation of levels-of-detail in bayesian tracking framework using single RGB-D sensor. In Proceedings of the 7th IEEE Annual Information Technology, Electronics and Mobile Communication Conference, IEEE, Vancouver, Canada, 2016. DOI: https://doi.org/10.1109/IEMCON.2016.7746290.

    Google Scholar 

  31. S. Salti, F. Tombari, L. Di Stefano. SHOT: Unique signatures of histograms for surface and texture description. Computer Vision and Image Understanding, vol. 125, pp. 251–264, 2014. DOI: https://doi.org/10.1016/j.cviu.2014.04.011.

    Article  Google Scholar 

  32. A. Frome, D. Huber, R. Kolluri, T. Bülow, J. Malik. Recognizing objects in range data using regional point descriptors. In Proceedings of the 8th European Conference on Computer Vision, Springer, Prague, Czech Republic, pp. 224–237, 2004. DOI: https://doi.org/10.1007/978-3-540-24672-5_18.

    Google Scholar 

  33. F. Tombari, S. Salti, L. Di Stefano. Unique signatures of histograms for local surface description. In Proceedings of the 11th European Conference on Computer Vision, Springer, Heraklion, Greece, pp. 356–369, 2010. DOI: https://doi.org/10.1007/978-3-642-15558-1_26.

    Google Scholar 

  34. R. Mekuria, K. Blom, P. Cesar. Design, implementation, and evaluation of a point cloud codec for tele-immersive video. IEEE Transactions on Circuits and Systems for Video Technology, vol. 27, no. 4, pp. 828–842, 2017. DOI: https://doi.org/10.1109/TCSVT.2016.2543039.

    Article  Google Scholar 

Download references

Acknowledgements

This work was supported by National Nature Science Foundation of China (No.61 811 530 281 and 61 861 136 009), Guangdong Regional Joint Foundation (No. 2019B1515120076), and the Fundamental Research for the Central Universities.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Chen-Guang Yang.

Additional information

Recommended by Associate Editor De Xu

Zun-Ran Wang received the B. Eng. degree in automation from the South China University of Technology, China in 2017. He is currently a M.Sc. degree candidate in the South China University of Technology, China.

His research interests include human-robot interaction, intelligent control and image processing.

Chen-Guang Yang received the B. Eng. degree in measurement and control from Northwestern Polytechnical University, China in 2005, the Ph.D. degree in control engineering from the National University of Singapore, Singapore in 2010. He received Best Paper Awards from IEEE Transactions on Robotics and over 10 international conferences. His research interests include robotics and automation.

Shi-Lu Dai received his B. Eng. degree in thermal engineering, the M. Eng. and Ph. D. degrees in control science and engineering, Northeastern University, China in 2002, 2006, and 2010, respectively. He was a visiting student in Department of Electrical and Computer Engineering, National University of Singapore, Singapore from November 2007 to November 2009, and a visiting scholar at Department of Electrical Engineering, University of Notre Dame, USA from October 2015 to October 2016. Since 2010, he has been with the School of Automation Science and Engineering, South China University of Technology, China, where he is currently a professor.

His research interests include adaptive and learning control, distributed cooperative systems.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wang, ZR., Yang, CG. & Dai, SL. A Fast Compression Framework Based on 3D Point Cloud Data for Telepresence. Int. J. Autom. Comput. 17, 855–866 (2020). https://doi.org/10.1007/s11633-020-1240-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11633-020-1240-5

Keywords

Navigation