Abstract
Conventional scene flow containing only translational vectors is not able to model 3D motion with rotation properly. Moreover, the accuracy of 3D motion estimation is restricted by several challenges such as large displacement, noise, and missing data (caused by sensing techniques or occlusion). In terms of solution, there are two kinds of approaches: local approaches and global approaches. However, local approaches can not generate smooth motion field, and global approaches is difficult to handle large displacement motion. In this paper, a completed dense scene flow framework is proposed, which models both rotation and translation for general motion estimation. It combines both a local method and a global method considering their complementary characteristics to handle large displacement motion and enforce smoothness respectively. The proposed framework is applied on the RGB-D image space where the computation efficiency is further improved. According to the quantitative evaluation based on Middlebury dataset, our method outperforms other published methods. The improved performance is further confirmed on the real data acquired by Kinect sensor.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Hadfield, S., Bowden, R.: Scene particles: unregularized particle based scene flow estimation. IEEE Trans. Pattern Anal. Mach. Intell. 36, 564–576 (2014)
Gottfried, J.-M., Fehr, J., Garbe, C.S.: Computing range flow from multi-modal Kinect data. In: Bebis, G., et al. (eds.) ISVC 2011, Part I. LNCS, vol. 6938, pp. 758–767. Springer, Heidelberg (2011)
Herbst, E., Ren, X., Fox, D.: Rgb-d flow: Dense 3-d motion estimation using color and depth. In: IEEE International Conference on Robotics and Automation (ICRA). IEEE (2013)
Quiroga, J., Devernay, F., Crowley, J.L., et al.: Local/global scene flow estimation. In: ICIP-IEEE International Conference on Image Processing (2013)
Zhang, X., Chen, D., Yuan, Z., Zheng, N.: Dense scene flow based on depth and multi-channel bilateral filter. In: Lee, K.M., Matsushita, Y., Rehg, J.M., Hu, Z. (eds.) ACCV 2012, Part III. LNCS, vol. 7726, pp. 140–151. Springer, Heidelberg (2013)
Barnes, C., Shechtman, E., Finkelstein, A., Goldman, D.: Patchmatch: A randomized correspondence algorithm for structural image editing. ACM Transactions on Graphics-TOG 28, 24 (2009)
Barnes, C., Shechtman, E., Goldman, D.B., Finkelstein, A.: The generalized patchmatch correspondence algorithm. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010, Part III. LNCS, vol. 6313, pp. 29–43. Springer, Heidelberg (2010)
Korman, S., Avidan, S.: Coherency sensitive hashing. In: 2011 IEEE International Conference on Computer Vision (ICCV), pp. 1607–1614. IEEE (2011)
Brox, T., Malik, J.: Large displacement optical flow: descriptor matching in variational motion estimation. IEEE Trans. Pattern Anal. Mach. Intell. 33, 500–513 (2011)
Hornacek, M., Fitzgibbon, A., Carsten, R.: Sphereflow: 6 dof scene flow from rgb-d pairs. In: 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE (2014)
Junejo, I.N., Dexter, E., Laptev, I., Perez, P.: View-independent action recognition from temporal self-similarities. IEEE Trans. Pattern Anal. Mach. Intell. 33, 172–185 (2011)
Baker, S., Scharstein, D., Lewis, J., Roth, S., Black, M.J., Szeliski, R.: A database and evaluation methodology for optical flow. Int. J. Comput. Vis. 92, 1–31 (2011)
Vedula, S., Rander, P., Collins, R., Kanade, T.: Three-dimensional scene flow. IEEE Trans. Pattern Anal. Mach. Intell. 27, 475–480 (2005)
Vogel, C., Schindler, K., Roth, S.: Piecewise rigid scene flow. In: 2013 IEEE International Conference on Computer Vision (ICCV). IEEE (2013)
Basha, T., Moses, Y., Kiryati, N.: Multi-view scene flow estimation: a view centered variational approach. Int. J. Comput. Vis. 101, 6–21 (2013)
Huguet, F., Devernay, F.: A variational method for scene flow estimation from stereo sequences. In: IEEE 11th International Conference on Computer Vision, ICCV 2007, pp. 1–7. IEEE (2007)
HaCohen, Y., Shechtman, E., Goldman, D.B., Lischinski, D.: Non-rigid dense correspondence with applications for image enhancement. ACM Trans. Graph. (TOG) 30, 70:1–70:9 (2011)
Eshet, Y., Korman, S., Ofek, E., Avidan, S.: Dcsh-matching patches in rgbd images. In: 2013 IEEE International Conference on Computer Vision (ICCV), pp. 89–96. IEEE (2013)
Lowe, D.G.: Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 60, 91–110 (2004)
Lempitsky, V., Rother, C., Roth, S., Blake, A.: Fusion moves for markov random field optimization. IEEE Trans. Pattern Anal. Mach. Intell. 32, 1392–1405 (2010)
Boros, E., Hammer, P.L.: Pseudo-boolean optimization. Discrete Appl. Math. 123, 155–225 (2002)
Xu, L., Jia, J., Matsushita, Y.: Motion detail preserving optical flow estimation. IEEE Trans. Pattern Anal. Mach. Intell. 34, 1744–1757 (2012)
Willimon, B., Hickson, S., Walker, I., Birchfield, S.: An energy minimization approach to 3d non-rigid deformable surface estimation using rgbd data. In: 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 2711–2717. IEEE (2012)
Acknowledgement
This work was supported by Microsoft Research, Redmond. We also acknowledge Minqi Li for recording the Kinect RGB-D data sequence in our experiment.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Wang, Y. et al. (2015). Completed Dense Scene Flow in RGB-D Space. In: Jawahar, C., Shan, S. (eds) Computer Vision - ACCV 2014 Workshops. ACCV 2014. Lecture Notes in Computer Science(), vol 9008. Springer, Cham. https://doi.org/10.1007/978-3-319-16628-5_14
Download citation
DOI: https://doi.org/10.1007/978-3-319-16628-5_14
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-16627-8
Online ISBN: 978-3-319-16628-5
eBook Packages: Computer ScienceComputer Science (R0)