A feature-based metric for the quantitative evaluation of pixel-level image fusion

https://doi.org/10.1016/j.cviu.2007.04.003Get rights and content

Abstract

Pixel-level image fusion has been investigated in various applications and a number of algorithms have been developed and proposed. However, few authors have addressed the problem of how to assess the performance of those algorithms and evaluate the resulting fused images objectively and quantitatively. In this study, two new fusion quality indexes are proposed and implemented through using the phase congruency measurement of the input images. Therefore, the feature-based measurements can provide a blind evaluation of the image fusion result, i.e. no reference image is needed. These metrics take the advantage of the phase congruency measurement which provides a dimensionless contrast- and brightness-invariant representation of image features. The fusion quality indexes are compared with recently developed blind evaluation metrics. The validity of the new metrics are identified by the test on the fusion results achieved by a number of multiresolution pixel-level fusion algorithms.

Introduction

The analysis of multisensor images benefits from the technique named image fusion. The applications include but are not limited to military target detection, medical imaging, remote sensing, nondestructive inspection, and intelligent surveillance [1]. The study of fusing multiple images mainly focuses on how to optimize the use of the redundant and complementary information provided by heterogeneous image sensors. With different purposes, the image fusion algorithms can be classified as either combination or classification fusion. In the first case the fusion algorithm consists of combining the complementary features from multiple input images while in the second case the redundant information is mainly used for making a decision through modeling. The output of the first type of image fusion is still an image but comprised of the most salient features captured from the different sensors. Usually such fusion operation is implemented with four major steps at the pixel-level as illustrated in Fig. 1. The success of a specific post-processing or analysis relies largely on the performance of the specific fusion algorithm. The goal of the classification fusion is to derive a thematic map that indicates certain homogeneous characteristics of pixel regions in the image. This process needs a higher-level operation like feature- or decision-level fusion. In order to determine the performance of different fusion algorithms, an objective metric that can quantitatively measure the quality of the results should be introduced. However, such a metric would largely depend on the requirements of the specific application. A straightforward way to implement the evaluation is through the comparison with a reference image, which is assumed to be perfect. The metrics for image comparison are often employed for the quantitative assessment of image fusion. In the case of classification fusion, the resulting thematic map is compared with the ground truth data. The classification results are then used to generate a confusion matrix. Alternatively the classification errors for each of the classes, and for various thresholds, can also be represented by a receiver operating characteristic (ROC) curve. The perfect results can be prepared with the help of experts’ experience. Unfortunately, the reference image is not always perfect or available practically, thus, raising the need for the quantitative and blind evaluation of the fused images.

A typical example for pixel-level image fusion is the fusion of multi-focus images from a digital camera [2], [3]. In such case, a cut and paste operation is applied to obtain the full-focus image that will serve as a reference for evaluating the fusion results. However, such operation does not assure a perfect reference image. In some applications, the ground truth reference can be generated from a more precise measurement. For example, in nondestructive testing the thickness of a multilayer-structured specimen from an aircraft is estimated from the eddy current inspection or the ultrasonic testing. The actual thickness map of the specimen can be obtained through the post-teardown inspection with X-ray, a destructive process that is used to verify the accuracy of the estimation [4]. Such comparison can only be applied after the acquired images are fully registered, i.e. converted to the same resolution, size, and format. The evaluation metric should be optimized for the image feature. Pixel-by-pixel comparison does not meet the requirement, because in the original image pixels are closely related. Moreover, it would be better if the quantitative evaluation can still be achieved without the presence of reference image. This is the case of most practical applications. The evaluation metric should provide a measurement of how well the information of the inputs is integrated into the output.

In this paper, the study is limited to the evaluation of pixel-level image fusion. We propose a measurement of fusion performance based on the phase congruency calculation which was suggested by Kovesi [5], [6], [7]. The phase congruency measurement provides an absolute or dimensionless value ranging from 0 to 1; the larger values correspond to the salient features. Two methods are proposed to identify the availability and quality of input features in the fused image. The first one is based on the modified structural similarity measurement, where phase congruency is employed as the structural component. Similarity map with the fused image is generated for each input image. Then, the larger value at each location is retained for overall assessment. The second one is implemented by computing the local cross-correlation of the phase congruency maps between the fused and input images. The index value is obtained by averaging the similarity or cross-correlation value in each pre-defined region. The proposed schemes achieve a no-reference evaluation of the fused image. Experiments are carried out on a group of fused images obtained by various multiresolution fusion algorithms. The validity of the proposed methods are demonstrated.

The rest of the paper is organized as follows. An overview of the metrics used for image fusion is presented in Section 2. In Section 3, the concept and implementation of the feature-based evaluation are described. Experiments with the proposed approach and comparison with other existing methods can be found in Section 4. The experimental results obtained from both the reference-based assessment and blind evaluation are presented. Thus, the proposed algorithms are validated with these tests. Section 5 presents the discussion. In the final Section 6, the conclusion of whole paper is drawn.

Section snippets

Evaluation of image fusion algorithms

A straightforward approach for fusion performance assessment is to compare the fused image with a reference image. The commonly used methods include the root mean square error (RMSE), normalized least square error (NLSE), the peak signal-to-noise ratio (PSNR), correlation (CORR), difference entropy (DE), and mutual information (MI) [8]. The expressions that correspond to the above treatments are listed below. The meaning of the symbols used in the following equations is listed in Table 1.

Root

A strategy for the feature-based evaluation

This section describes our approaches for the evaluation of fused image. The proposed feature-based strategy proceeds in two steps: first extracting image features and then measuring how those features are integrated in the fused image. The phase congruency is employed to provide an absolute measurement of image feature. Such measurement is incorporated into the SSIM or a local cross-correlation is performed to determine if the features from inputs are available in the fused image. An overall

Experimental results

The major differences between the existing fusion algorithms reside in two aspects: the multiresolution strategy used and the fusion rule to be applied in the transform domain. Various multiresolution representations have been investigated for image fusion applications. The efficiency of the fusion rules largely depends on the applications, i.e. the characteristic of the images in the tests. The group of images considered here consists of multi-focus and simulated multi-focus images from a

Discussion

The image fusion is an application-dependent operation. In other words, the process depends on the type of images or their formats. Because the images acquired by heterogeneous sensors possess a different intensity map, there is no “one size fits all” solution for the evaluation process. Therefore, one fusion algorithm may not necessarily achieve the same performance on distinct images by using certain evaluation metric. One purpose of this study is to identify the feasibility and validity of

Conclusion

In this paper, two new feature-based metrics for image fusion performance are presented. The two metrics are based on a modified SSIM scheme and the local cross-correlations between the feature maps of the fused and the input images. The image features are represented by a dimensionless quantity ranging from zero to one, namely phase congruency, which is invariant to the changes in image illumination and contrast. These metrics provide an objective quality measure of the fused image in the

Acknowledgments

Dr. Andy Goshtasby at Image Fusion Systems Research, Dr. Zhou Wang at University of Texas (Austin), the SPCR Lab at the Lehigh University, and Ginfuku at Kyoto are acknowledged for providing the images used in this work. Some of the images are available from http://www.imagefusion.org and http://www.cns.nyu.edu/~zwang/.

References (25)

  • R.S. Blum et al.

    An overview of image fusion

  • H. Li et al.

    Multisensor image fusion using the wavelet transform

    Graphical Models and Image Processing

    (1995)
  • Z. Zhang, R.S. Blum, Image fusion for a digital camera application, in: Proceedings of 32nd Asilomar Conference on...
  • D.S. Forsyth et al.

    NDT data fusion for improved corrosion detection

  • P. Kovesi

    Image features from phase congruency

    Videre: A Journal of Computer Vision Research

    (1999)
  • P. Kovesi, Image features from phase congruency, University of Western Australia, Tech. Rep.,...
  • P. Kovesi, Invarian measures of image features from phase information, Ph.D. dissertation, University of Western...
  • Y. Wang, B. Lohmann, Multisensor image fusion: concept, method and applications, Institut für Automatisierungstechnik,...
  • Z. Wang et al.

    A universal image quality index

    IEEE Signal Processing Letters

    (2002)
  • Z. Wang et al.

    Image quality assessment: from error measurement to structural similarity

    IEEE Transactions on Image Processing

    (2004)
  • G. Qu et al.

    Information measure for performance of image fusion

    Electronics Letters

    (2002)
  • C.S. Xydeas et al.

    Objective image fusion performance measure

    Electronics Letters

    (2000)
  • Cited by (0)

    View full text