Abstract
In this paper, we propose two methods for estimating the scales of point clouds to align them. The first method estimates the scale of each point cloud separately: each point cloud has its own scale that is something like the size of a scene. We call it a keyscale, which is a representative scale and is defined for a given 3D point cloud as the minimum of the cumulative contribution rates of PCA of descriptors over different scales. Our second method directly estimates the ratio of scales (scale ratio) of two point clouds. Instead of finding the minimum, this approach registers the two sets of curves of the cumulative contribution rate of PCA by assuming that those differ only in scale. Experimental results with simulated and real scene point clouds demonstrate that the scale alignment of 3D point clouds can be effectively accomplished by our scale ratio estimation.
Similar content being viewed by others
References
Akagunduz, E., Ulusoy, I.: Extraction of 3D transform and scale invariant patches from range scans. In: CVPR, pp. 1–8 (2007)
Baatz, G., Koser, K., Chen, D., Grzeszczuk, R., Pollefeys, M.: Leveraging 3D city models for rotation invariant place-of-interest recognition. IJCV 96(3), 315–334 (2012)
Belongie, S., Malik, J., Puzicha, J.: Shape matching and object recognition using shape contexts. PAMI 24(4), 509–522 (2002)
Besl, P.J., McKay, N.D.: A method for registration of 3-D shapes. PAMI 14(2), 239–256 (1991)
Castellani, U., Cristani, M., Fantoni, S.: Sparse points matching by combining 3D mesh saliency with statistical descriptors. In: Proceedings of Computer Graphics Forum, pp. 643–652 (2008).
Cheung, W., Hamarneh, G.: n-sift: n-dimensional scale invariant feature transform. Trans. Image Process. 18(9), 2012–2021 (2009)
Conde, C., Serrano, A.: 3D Facial normalization with spin images and influence of range data calculation over face verification. In: CVPRW, pp. 115 (2005)
Du, S., Zheng, N., Ying, S., You, Q., Wu, Y.: An extension of the ICP algorithm considering scale factor. ICIP 5, 193–196 (2007)
Dyden, L.L., Mardia, K.V.: Statistical shape analysis. Wiley, New York (1998)
Furukawa, Y., Ponce, J.: Accurate, dense, and robust multi-view stereopsis. In: CVPR, pp. 1362–1376 (2007).
Haker, M., Bohme, M., Martinetz, T., Barth, E.: Scale invariant range features for time-of-flight camera applications. In: CVPRW (2008).
Hurley, J.R., Cattell, R.B.: Producing direct rotation to test a hypothesized factor structure. Behav. Sci. 7(2), 258–262 (1962)
Izadi, S., Kim, D., Hilliges, O., Molyneaux, D., Newcombe, R., Kohli, P., Shotton, J., Hodges, S., Freeman, D., Davison, A., Fitzgibbon, A.: KinectFusion: real-time 3D reconstruction and interaction using a moving depth camera. In Proceedings of the ACM symposium on User interface software and technology (UIST), pp. 559–568 (2011).
Johnson, A.E., Hebert, M.: Surface matching for object recognition in complex three-dimensional scenes. Image Vis. Comput. 16, 635–651 (1998)
Johnson, A.E., Hebert, M.: Using spin images for efficient object recognition in cluttered 3D scene. PAMI 21, 433–449 (1999)
Kaick, O.V., Zhang, H., Hamarneh, G., Cohen-Or, D.: A survey on shape correspondence. In: Proceedings of Computer Graphics Forum, pp. 1681–1707 (2011).
Lin, B., Tamaki, T., Raytchev, B., Kaneda, K., Ichii, K.: Scale ratio ICP for 3D point clouds with different scales. In: ICIP (2013).
Lin, B., Tamaki, T., Slomp, M., Raytchev, B., Kaneda, K., Ichii, K.: 3D Keypoints detection from a 3D point cloud for real-time camera tracking. IEEJ Trans. Electro. Inf. Syst. 133(1), 84–90 (2013)
Makadia, A., Patterson, A., Daniilidis, K.: Fully automatic registration of 3D point clouds. In: cVPR, pp. 1297–1304 (2006).
Mian, A.S., Bennamoun, M., Owens, R.: Keypoint detection and local feature matching for textured 3D face recognition. IJCV 79(1), 1–12 (2008)
Salvi, J., Matabosch, C., Fofi, D., Forest, J.: A review of recent range image registration methods with accuracy evaluation. Image Vis. Comput. 25(5), 578–596 (2007)
Scovanner, P., Ali, S., Shah, M.: A 3-dimensional sift descriptor and its application to action recognition. In: ACM Multimedia, pp. 357–360 (2007).
Snavely, N., Seitz, S.M., Szeliski, R.: Modeling the world from internet photo collections. IJCV 80, 189–210 (2008)
Steder, B., Rusu, R.B., Konolige, K., Burgard, W.: NARF: 3D range image features for object recognition. In: Workshop on Defining and Solving Realistic Perception Problems in Personal Robotics at IROS (2010).
Tam, G.K.L., Cheng, Z., Lai, Y., Langbein, F.C., Liu, Y., Marshall, D., Martin, R.R., Sun, X., Rosin, P.L.: Registration of 3D point clouds and meshes: a survey from rigid to nonrigid. IEEE Trans. Vis. Comput. Graph. 19(7), 1199–1217 (2013)
Tamaki, T., Tanigawa, S., Ueno, Y., Raytchev, B., Kaneda, K.: Scale matching of 3D point clouds by finding keyscales with spin images. In: ICPR, pp. 3480–3483 (2010).
The Stanford 3D Scanning Repository. Available at: http://www.graphics.stanford.edu/data/3Dscanrep/
Unnikrishnan, R., Hebert, M.: Multi-scale interest regions from unorganized point clouds. In: Proceedings of Workshop on Search in 3D, vol. S3D, pp. 1–8 (2008).
Wu, C., Clipp, B., Li, X., Frahm, J. M., Pollefeys, M.: 3D model matching with viewpoint invariant patches (VIPS). In: CVPR (2008).
Zaharescu, A., Boyer, E., Varanasi, K., Horaud, R.: Surface feature detection and description with applications to mesh matching. In: Proceedings of International Conference on Computer Vision and Pattern Recognition (CVPR), pp. 373–380 (2009).
Zinßer, T., Schmidt, J., Niemann, H.: Point set registration with integrated scale estimation. In: Proceedings of International Conference on Pattern Recognition and Image Processing, pp. 116–119 (2005).
Acknowledgments
This work was supported in part by JSPS KAKENHI Grant Number 23700211.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Lin, B., Tamaki, T., Zhao, F. et al. Scale alignment of 3D point clouds with different scales. Machine Vision and Applications 25, 1989–2002 (2014). https://doi.org/10.1007/s00138-014-0633-2
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00138-014-0633-2