Efficient guided hypothesis generation for multi-structure epipolar geometry estimation

https://doi.org/10.1016/j.cviu.2016.10.003Get rights and content

Highlights

  • A sampling method EGHG is proposed for multi-structure geometry estimation.

  • EGHG combines the benefits of a global and a local sampling strategy.

  • The global sampling strategy is designed to rapidly obtain promising solutions.

  • The local sampling strategy is designed to efficiently achieve accurate solutions.

  • Experimental results show the effectiveness of EGHG on public real image pairs.

Abstract

We propose an Efficient Guided Hypothesis Generation (EGHG) method for multi-structure epipolar geometry estimation. Based on the Markov Chain Monte Carlo process, EGHG combines two guided sampling strategies: a global sampling strategy and a local sampling strategy. The global sampling strategy, guided by using both spatial sampling probabilities and keypoint matching scores, rapidly obtains promising solutions. The spatial sampling probabilities are computed by using a normalized exponential loss function. The local sampling strategy, guided by using both Joint Feature Distributions (JFDs) and keypoint matching scores, efficiently achieves accurate solutions. In the local sampling strategy, EGHG updates a set of current best hypothesis candidates on the fly, and then computes JFDs between the input data and these candidates. Experimental results on public real image pairs show that EGHG significantly outperforms several state-of-the-art sampling methods on multi-structure data.

Introduction

Epipolar geometry estimation is an important task in computer vision and has been successfully employed in image registration (Kushnir and Shimshoni, 2014), image segmentation (Wang et al., 2012), motion segmentation (Jung et al., 2014), etc. RANSAC (Fischler and Bolles, 1981) has been extensively used in computer vision, especially for recovering the epipolar geometry from a pair of images. The success of RANSAC and its variants (such as Chum, Matas, Kittler, 2003, Lebeda, Matas, Chum, 2012, Rabin, Delon, Gousseau, Moisan, 2010, Raguram, Chum, Pollefeys, Matas, Frahm, 2013, Raguram, Frahm, 2011, Sattler, Leibe, Kobbelt, 2009, Torr, Zisserman, 2000, Wang, Chin, Suter, 2012) depends on whether these methods can sample at least one all-inlier minimal subset (i.e., one clean minimal subset). The sampling strategy used in RANSAC is random sampling. Random sampling has been used in many recently proposed fitting methods (e.g., Isack, Boykov, 2012, Isack, Boykov, 2014, Mittal, Anand, Meer, 2012, Wang, Chin, Suter, 2012) due to its simplicity, so it is a rather ubiquitous element of robust fitting methods.

A main deficiency of random sampling lies in that it is not an effective (in terms of efficiency and accuracy) sampling strategy. A robust fitting method based on random sampling needs to exponentially increase the number of samples to ensure at least one clean minimal subset is sampled when the dimension of the model, or the outlier ratio of the input data, grows. Moreover, in practice, data in computer vision are usually composed of multiple structures (i.e., model instances) (Stewart, 1999). In such a case, the outliers of a structure contain both gross outliers and pseudo-outliers (i.e., inliers belonging to the other structures are outliers to the structure), which often leads to high outlier rates and thus severely affects the performance of random sampling.

To reduce the deficiency of random sampling, many recently proposed sampling methods (e.g., Chum, Matas, 2005, Fragoso, Sen, Rodriguez, Turk, 2013, Tordoff, Murray, 2005) accelerate promising hypothesis generation by exploiting prior information from keypoint matching scores. However, these sampling methods only consider the single structure case: They can efficiently find clean subsets for single-structure data (even with a large fraction of outliers) benefiting from accurate matching scores, but they usually fail to find a “clean” solution for multiple-structure data in a reasonable time (e.g., 10 s),1 where a “clean” solution means that at least one clean minimal or larger-than-minimal subset is sampled for each model instance in data. This is because the data points in the subsets sampled by these methods usually have high matching scores, but they might be from different structures. This may lead to overfitting (the inliers of one fitted hypothesis contain the inliers of multiple structures).

To efficiently sample for multi-structure fitting methods2 (e.g., Lazic, Givoni, Frey, Aarabi, 2009, Magri, Fusiello, 2014, Schindler, Suter, 2006, Thakoor, Gao, 2008, Toldo, Fusiello, 2008, Wang, Chin, Suter, 2012, Yu, Chin, Suter, 2011), which aim to simultaneously estimate all model instances of multi-structure data, some sampling methods (e.g., Chin, Yu, Suter, 2012, Wong, Chin, Yu, Suter, 2013) have been proposed to simultaneously generate promising hypotheses for all the structures of multi-structure data, by residual sorting. However, in a reasonable time budget, these methods may not obtain any accurate solution for data with a low inlier rate, because using all the generated hypotheses, via residual sorting, to guide the subsequent hypothesis generation is time-consuming. Although some other sampling methods (e.g., Kanazawa and Kawakami, 2004) can also obtain clean solutions for multi-structure data (even with a low inlier rate), the clean solutions obtained by these methods may be inaccurate. This is often because the spans of minimal subsets sampled by these methods are small due to the use of local constraints (Tran et al., 2014).

In this paper, we focus on simultaneously and efficiently sampling promising hypotheses for multi-structure data. We propose an Efficient Guided Hypothesis Generation (EGHG) method, which contains two guided sampling strategies (i.e., a global sampling strategy and a local sampling strategy), for epipolar geometry estimation, which aims to simultaneously estimate all model instances of multi-structure data.

Although the global sampling strategy in EGHG can rapidly obtain clean solutions, the obtained solutions may not be accurate enough. The local sampling strategy, on the other hand, can efficiently generate more accurate solutions by using Joint Feature Distributions (JFDs) (Triggs, 2001), which use the epipolar constraint to accurately characterise the dependencies among keypoint matches. However, the local sampling strategy needs the clean solutions (i.e., the clean subsets) to compute JFDs. Thus, we combine the global sampling strategy and the local sampling strategy, based on a Markov Chain Monte Carlo (MCMC) process, to rapidly generate promising hypotheses for multi-structure data.

In detail, the sampling strategies are given as follows:

The global sampling strategy: We propose to sample minimal subsets by using both keypoint matching scores and computationally cheap spatial sampling probabilities in the global sampling process to rapidly “hit” clean solutions.

The local sampling strategy: A set of current best hypothesis candidates (corresponding to different model instances in data) is updated from the generated hypotheses by using a nonparametric kernel density estimate technique, on the fly. Then a hypothesis is selected from these candidates in each local sampling process, where the hypothesis is generated from a sampled subset S. Following this, the JFDs between the input data and S are computed. Finally, we propose to sample subsets guided by using both JFDs and keypoint matching scores in the local sampling process, to effectively generate accurate solutions. The process of updating the hypothesis candidates and computing the JFDs is efficient.

The rest of the paper is constructed as follows: In Section 2, we review related work. In Section 3, we present the global sampling strategy. In Section 4, we describe the local sampling strategy. In Section 5, we propose the complete EGHG method, and in Section 6 we show the experimental results on real image pairs. Finally, we draw conclusions in Section 7.

Section snippets

Related work

Recently, some guided sampling methods (e.g., Brahmachari, Sarkar, 2013, Chin, Yu, Suter, 2012, Chum, Matas, Kittler, 2003, Kanazawa, Kawakami, 2004, Lebeda, Matas, Chum, 2012) have been proposed for accelerating promising hypothesis generation. For example, LO-RANSAC (Chum et al., 2003) inserts a local optimization step into the standard RANSAC to speed up promising hypothesis generation. However, for data with a large number of inliers, the local optimization step may increase the

The global sampling strategy

In this section, we first introduce the notation used in this paper, and then we present the global sampling strategy.

Let X={xi}i=1N={(ui,vi)}i=1N be N input data (i.e., the matches between two input images), where ui=(τλi,υλi,1) and vi=(τλi,υλi,1) are respectively the normalized coordinates of the ith match on the two input images (we assume that putative matches are established by using a similarity metric), where the global coordinate normalization (Hartley, 1997) is performed on putative

The local sampling strategy

The local sampling strategy aims to generate more promising hypotheses by using the matching scores and the JFDs (Triggs, 2001) between input data and the set of current best hypothesis candidates (that are updated from the generated hypotheses, on the fly). In this section, we describe the process of computing the JFDs between input data and a sampled subset in Section 4.1. Following this, we propose an update algorithm for maintaining a set of current best hypothesis candidates (corresponding

The complete method

Similar to Brahmachari and Sarkar (2013), EGHG adopts a MCMC process to combine both the global sampling strategy and the local sampling strategy for hypothesis generation. The proposed EGHG sampling method is summarized in Algorithm 2. The parameter β of MCMC in Algorithm 2 controls the balance between global sampling and local sampling; and its influence on the performance of EGHG will be evaluated in Section 6.3.2. Two crucial components of Algorithm 2: the global sampling strategy (Line 4)

Experiments

We first discuss the size of the subsets sampled in the local sampling strategy of the proposed EGHG in Section 6.1. Then, we describe the dataset and evaluation metrics used in the experiments in Section 6.2. In Section 6.3, we evaluate the sensitivity of the proposed method to the parameters. We assess the influence of different hypothesis selection strategies used to guide the local sampling process in Section 6.4. We compare the proposed method with several state-of-the-art sampling methods

Conclusions

In this paper, we present an efficient sampling method called EGHG to accelerate promising hypothesis generation for multi-structure epipolar geometry estimation. The key idea of EGHG is that, instead of selecting the current best hypothesis, a set of current best hypothesis candidates are updated on the fly to guide the local sampling process. Thus, EGHG can rapidly and accurately generate hypotheses for multi-structure data, and it also works well on single-structure data. We evaluate EGHG on

Acknowledgment

The authors would like to thank Tat-Jun Chin, Aveek S. Brahmachari, Toldo Roberto, Pham Trung-Thanh and Hoi Sim Wong for sharing their codes and datasets. This work was supported by the National Natural Science Foundation of China under Grants 61472334 and 61571379. David Suter acknowledged funding under ARC DPDP130102524

References (41)

  • P.H. Torr et al.

    MLESAC: A new robust estimator with application to estimating image geometry

    Comput. Vision Image Understanding

    (2000)
  • H.S. Wong et al.

    A simultaneous sample-and-filter strategy for robust multi-structure model fitting

    Comput. Vision Image Understanding

    (2013)
  • A.S. Brahmachari et al.

    Hop-diffusion monte carlo for epipolar geometry estimation between very wide-baseline images

    IEEE Trans. Pattern Anal. Mach. Intell.

    (2013)
  • ChinT.-J. et al.

    Accelerated hypothesis generation for multistructure data via preference analysis

    IEEE Trans. Pattern Anal. Mach. Intell.

    (2012)
  • O. Chum et al.

    Matching with PROSAC-progressive sample consensus

    Proceedings of IEEE Conference on Computer Vision and Pattern Recognition

    (2005)
  • O. Chum et al.

    Locally optimized RANSAC

    DAGM-Symp. Pattern Recognit.

    (2003)
  • O. Chum et al.

    Two-view geometry estimation unaffected by a dominant plane

    Proceedings of IEEE Conference on Computer Vision and Pattern Recognition

    (2005)
  • M.A. Fischler et al.

    Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography

    Commun. ACM.

    (1981)
  • V. Fragoso et al.

    EVSAC: accelerating hypotheses generation by modeling matching scores with extreme value theory

    Proceedings of IEEE International Conference on Computer Vision

    (2013)
  • J.-M. Frahm et al.

    RANSAC for (quasi-) degenerate data (QDEGSAC)

    Proceedings of IEEE Conference on Computer Vision and Pattern Recognition

    (2006)
  • L. Goshen et al.

    Balanced exploration and exploitation model search for efficient epipolar geometry estimation

    IEEE Trans. Pattern Anal. Mach. Intell.

    (2008)
  • R. Hartley et al.

    Multiple View Geometry in Computer Vision

    (2004)
  • R. Hartley

    In defense of the eight-point algorithm

    IEEE Trans. Pattern Anal. Mach. Intell.

    (1997)
  • H. Isack et al.

    Energy-based geometric multi-model fitting

    Int. J. Comput. Vis.

    (2012)
  • H. Isack et al.

    Energy based multi-model fitting & matching for 3d reconstruction

    Proceedings of IEEE Conference on Computer Vision and Pattern Recognition

    (2014)
  • H. Jung et al.

    Rigid motion segmentation using randomized voting

    Proceedings of IEEE Conference on Computer Vision and Pattern Recognition

    (2014)
  • Y. Kanazawa et al.

    Detection of planar regions with uncalibrated stereo using distributions of feature points.

    Proceedings of British Machine Vision Conference

    (2004)
  • S. Kirkpatrick et al.

    Optimization by simulated annealing

    Science

    (1983)
  • M. Kushnir et al.

    Epipolar geometry estimation for urban scenes with repetitive structures

    IEEE Trans. Pattern Anal. Mach. Intell.

    (2014)
  • N. Lazic et al.

    Floss: facility location for subspace segmentation

    Proceedings of IEEE International Conference on Computer Vision

    (2009)
  • Cited by (0)

    View full text