Abstract
Image matching is a challenging problem as different views often undergo significant appearance changes caused by illumination changes, scale variations, large displacement, and deformation. Most state-of-the-art algorithms, however, still cannot perform well enough in handling challenging real-world cases, especially in different objects and scenes. In this paper, we explore deep features extracted from pretrained convolutional neural networks to help the estimation of image matching so that dense pixel correspondence can be built. As the deep features are able to describe the image structures and details hierarchically, the matching method based on these features is able to match different scenes and object appearances effectively. We analyze the deep features and compare them with other robust features, e.g., SIFT. Extensive experiments on benchmark datasets demonstrate that the proposed algorithm performs favorably against the state-of-the-art methods in terms of visually matching quality and accuracy.















Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Baker, S., Scharstein, D., Lewis, J.P., Roth, S., Black, M.J., Szeliski, R.: A database and evaluation methodology for optical flow. In: NIPS, pp. 1–31 (2011)
Barnes, C., Shechtman, E., Finkelstein, A., Goldman, D.B.: Patchmatch: a randomized correspondence algorithm for structural image editing. In: ACM TOG, vol. 28, no. 3 (2009)
Black, M.J., Anandan, P.: The robust estimation of multiple motions—parametric and piecewise-smooth flow fields. Comput. Visi. Image Underst. 63(1), 75–104 (1996)
Brox, T., Bregler, C., Malik, J.: Large displacement optical flow. In: CVPR, pp. 41–48 (2009)
Brox, T., Bruhn, A., Papenberg, N., Weickert, J.: High accuracy optical flow estimation based on a theory for warping. In: ECCV, pp. 25–36 (2004)
Brox, T., Malik, J.: Large displacement optical flow: descriptor matching in variational motion estimation. IEEE Trans Pattern Anal. Mach. Intell. 33(3), 500–513 (2011)
Choy, C.B., Gwak, J., Savarese, S., Chandraker, M.K.: Universal correspondence network. In: NIPS (2016)
Datta, R., Joshi, D., Li, J., Wang, J.Z.: Image retrieval—Ideas, influences, and trends of the new age. ACM Comput. Surv. 40(2), 1–60 (2008)
Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Li, F.F.: ImageNet—a large-scale hierarchical image database. In: CVPR, pp. 248–255 (2009)
Dosovitskiy, A., Fischer, P., Ilg, E., Häusser, P., Hazirbas, C., Golkov, V., van der Smagt, P., Cremers, D., Brox, T.: FlowNet—learning optical flow with convolutional networks. In: ICCV (2015)
van den Elsen, P.A., Pol, E.J.D., Viergever, M.A.: Medical image matching—a review with classification. IEEE Eng. Med. Biol. Mag. 12(1), 26–39 (1993)
Fei-Fei, L., Fergus, R., Perona, P.: Learning generative visual models from few training examples: an incremental Bayesian approach tested on 101 object categories. Comput. Vis. Image Underst. 106(1), 59–70 (2007)
Franke, U., Rabe, C., Badino, H., Gehrig, S.K.: 6D-Vision—fusion of stereo and motion for robust environment perception. In: DAGM-Symposium vol. 3663, pp. 216–223 (2005) (Chapter 27)
Han, K., Rezende, R.S., Ham, B., Wong, K.Y.K., Cho, M., Schmid, C., Ponce, J.: SCNet—learning semantic correspondence. In: ICCV, pp. 1849–1858 (2017)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR, pp. 770–778 (2016)
Horn, B.K., Schunck, B.G.: Determining optical flow. Artif. Intell. 17, 185–203 (1981)
Hur, J., Lim, H., Park, C., Ahn, S.C.: Generalized deformable spatial pyramid—geometry-preserving dense correspondence estimation. In: CVPR, pp. 1392–1400 (2015)
Ilg, E., Mayer, N., Saikia, T., Keuper, M., Dosovitskiy, A., Brox, T.: FlowNet 2.0: evolution of optical flow estimation with deep networks. In: CVPR, pp. 1647–1655 (2017)
Kim, J., Liu, C., Sha, F., Grauman, K.: Deformable spatial pyramid matching for fast dense correspondences. In: CVPR, pp. 2307–2314 (2013)
Kim, S., Min, D., Ham, B., Jeon, S., Lin, S., Sohn, K.: FCSS—fully convolutional self-similarity for dense semantic correspondence. In: CVPR, pp. 616–625 (2017)
Lester, H., Arridge, S.R.: A survey of hierarchical non-linear medical image registration. Pattern Recognit. 32(1), 129–149 (1999)
Li, C., Wand, M.: Combining Markov random fields and convolutional neural networks for image synthesis. In: CVPR, pp. 2479–2486 (2016)
Li, W., Cosker, D., Brown, M., Tang, R.: Optical flow estimation using Laplacian mesh energy. In: CVPR, pp. 2435–2442 (2013)
Liu, C., Yuen, J., Torralba, A.: SIFT flow—dense correspondence across scenes and its applications. IEEE Trans. Pattern Anal. Mach. Intell. 33(5), 978–994 (2011)
Liu, Y., Pan, J., Su, Z.: Deep feature matching for dense correspondence. In: ICIP (2017)
Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. In: CVPR, pp. 3431–3440 (2015)
Long, J., Zhang, N., Darrell, T.: Do convnets learn correspondence? In: NIPS (2014)
Lowe, D.G.: Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 60(2), 91–110 (2004)
Ma, C., Huang, J.B., Yang, X., Yang, M.H.: Hierarchical convolutional features for visual tracking. In: ICCV, pp. 3074–3082 (2015)
Nag, S.: Image registration techniques: a survey (2017). arXiv:1712.07540
Revaud, J., Weinzaepfel, P., Harchaoui, Z., Schmid, C.: EpicFlow: edge-preserving interpolation of correspondences for optical flow. In: CVPR, pp. 1164–1172 (2015)
Russell, B.C., Torralba, A., Murphy, K.P., Freeman, W.T.: LabelMe: a database and web-based tool for image annotation. In: NIPS, pp. 157–173 (2007)
Shekhovtsov, A., Kovtun, I., Hlavac, V.: Efficient MRF deformation model for non-rigid image matching. In: CVPR, pp. 1–6 (2007)
Shen, X., Tao, X., Zhou, C., Gao, H., Jia, J.: Regional foremost matching for internet scene images. In: ACM TOG , vol. 35, no. 6, pp. 1–12 (2016)
Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: ICLR (2015)
Sun, D., Roth, S., Black, M.J.: Secrets of optical flow estimation and their principles. In: CVPR, pp. 2432–2439 (2010)
Tau, M., Hassner, T.: Dense correspondences across scenes and scales. IEEE Trans. Pattern Anal. Mach. Intell. 38(5), 875–888 (2016)
Weinzaepfel, P., Revaud, J., Harchaoui, Z., Schmid, C.: DeepFlow—large displacement optical flow with deep matching. In: ICCV, pp. 1385–1392 (2013)
Xu, L., Jia, J., Matsushita, Y.: Motion detail preserving optical flow estimation. IEEE Trans. Pattern Anal. Mach. Intell. 34(9), 1744–1757 (2012)
Yan, Z., Xiang, X.: Scene flow estimation: a survey (2016). arXiv:1612.02590
Yang, H., Lin, W.Y., Lu, J.: DAISY filter flow: a generalized discrete approach to dense correspondences. In: CVPR, pp. 3406–3413 (2014)
Zeiler, M.D., Fergus, R.: Visualizing and understanding convolutional networks. In: ECCV (2014)
Zhou, T., Krähenbühl, P., Aubry, M., Huang, Q., Efros, A.A.: learning dense correspondence via 3D-guided cycle consistency. In: CVPR, pp. 117–126 (2016)
Funding
This work is partially supported by the National Natural Science Foundation of China (Nos. 61872421, 61572099, and 61702243), the National Science and Technology Major Project (Nos. ZX20140419, 2014ZX04001011), and the Natural Science Foundation of Jiangsu Province (No. BK20180471).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Liu, Y., Pan, J., Su, Z. et al. Robust dense correspondence using deep convolutional features. Vis Comput 36, 827–841 (2020). https://doi.org/10.1007/s00371-019-01656-z
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00371-019-01656-z