An Integrated Color and Intensity Co-occurrence Matrix

https://doi.org/10.1016/j.patrec.2007.01.004Get rights and content

Abstract

The paper presents a novel approach for representing color and intensity of pixel neighborhoods in an image using a co-occurrence matrix. After analyzing the properties of the HSV color space, suitable weight functions have been suggested for estimating relative contribution of color and gray levels of an image pixel. The suggested weight values for a pixel and its neighbor are used to construct an Integrated Color and Intensity Co-occurrence Matrix (ICICM). We have shown that if the ICICM matrix is used as a feature in an image retrieval application, it is possible to have higher recall and precision compared to other existing methods.

Introduction

Color and texture are two low-level features widely used for image classification, indexing and retrieval. Color is usually represented as a histogram, which is a first order statistical measure that captures global distribution of color in an image (Swain and Ballard, 1991, Gevers and Stokman, 2004). One of the main drawbacks of the histogram-based approaches is that the spatial distribution and local variations in color are ignored. Local spatial variation of pixel intensity is commonly used to capture texture information in an image. Grayscale Co-occurrence Matrix (GCM) is a well-known method for texture extraction in the spatial domain (Haralick et al., 1973). A GCM stores the number of pixel neighborhoods in an image that have a particular grayscale combination. Let I be an image and let p and Np respectively denote any arbitrary pixel and its neighbor in a given direction. If GL denotes the total number of quantized gray levels and gl denotes the individual gray levels, where, gl  {0,  , GL  1}, then each component of GCM can be written as follows:gcm(i,j)=Pr((glp,glNp)=(i,j))gcm(i, j) is the number of times the gray level of a pixel p denoted by glp equals i, and the gray level of its neighbor Np denoted by glNp equals j, as a fraction of the total number of pixels in the image. Thus, it estimates the probability that the gray level of an arbitrary pixel in an image is i, and that of its neighbor is j. One GCM matrix is generated for each possible neighborhood direction, namely, 0°, 45°, 90° and 135°. Average and range of 14 features like Angular Second Moment, Contrast, Correlation, etc., are generated by combining all the four matrices to get a total of 28 features (Haralick et al., 1973). In the GCM approach for texture extraction, color information is completely lost since only pixel gray levels are considered.

To incorporate spatial information along with the color of image pixels, a feature called color correlogram has recently been proposed. It is a three dimensional matrix that represents the probability of finding pixels of any two given colors at a distance ‘d’ apart (Huang et al., 1997). Auto correlogram is a variation of correlogram, which represents the probability of finding two pixels with the same color at a distance ‘d’ apart. This approach can effectively represent color distribution in an image. However, correlogram features do not capture intensity variation. Many image databases often contain both color as well as gray scale images. The color correlogram method does not constitute a good descriptor in such databases.

Another method called Color Co-occurrence Matrix (CCM) has been proposed to capture color variation in an image (Shim and Choi, 2003). CCM is represented as a three-dimensional matrix, where color pair of the pixels p and Np are captured in the first two dimensions of the matrix and the spatial distance ‘d’ between these two pixels is captured in the third dimension. This approach is a generalization of the color correlogram and reduces to the pure color correlogram for d = 1. CCM is generated using only the Hue plane of the HSV (Hue, Saturation and Intensity Value) color space. The Hue axis is quantized into HL number of levels. If individual hue values are denoted by hl, where hl  {0,  , HL  1}, then each component of CCM can be written as follows:ccm(i,j)=Pr((hlp,hlNp)=(i,j))Four matrices representing neighbors at angles 0°, 90°, 180° and 270° are considered. This approach was further extended by separating the diagonal and the non-diagonal components of CCM to generate a Modified Color Co-occurrence Matrix (MCCM). MCCM, thus, may be written as follows:MCCM=(CCMD,CCMND)Here, CCMD and CCMND correspond to the diagonal and off-diagonal components of CCM. The main drawback of this approach is that, like correlogram, it also captures only color information and intensity information is completely ignored.

An alternative approach is to capture intensity variation as a texture feature from an image and combine it with color features like histograms using suitable weights (Manjunath et al., 2001). One of the challenges of this approach is to determine suitable weights since these are highly application-dependent. In certain applications like Content-based Image Retrieval (CBIR), weights are often estimated from relevance feedback given by users (Aksoy and Haralick, 2000, Wu and Zhang, 2002). While relevance feedback is sometimes effective, it makes the process of image retrieval user-dependent and iterative. There is also no guarantee on the convergence of the weight-learning algorithms. In order to overcome these problems, researchers have tried to combine color and texture features together during extraction.

Palm (2004) proposed two approaches for capturing color and intensity variations from an image using the LUV color space. In the Single-channel Co-occurrence Matrix (SCM), variations for each color channel, namely, L, U and V are considered independently. In the Multi-channel Co-occurrence Matrix (MCM), variations are captured taking two channels at a time – UV, LU and LV. Since the LUV color space separates out chrominance (L and U) from luminance (V), SCM in effect, generates one GCM and two CCMs from each image independently. As a result, correlation between the color channels is lost. However, in MCM, the count of pairwise occurrences of the values of different channels of the color space is captured. Thus, each component of MCM can be written as follows:mcmUV(i,j)=Pr((up,vNp)=(i,j))mcmLU(i,j)=Pr((lp,uNp)=(i,j))mcmLV(i,j)=Pr((lp,vNp)=(i,j))Here, mcmUV(i, j) is the number of times the U chromaticity value of a pixel p denoted by up equals i, and the V chromaticity value of its neighbor Np denoted by vNp equals j, as a fraction of the total number of pixels in the image. Similarly, mcmLU(i, j) and mcmLV(i, j) are defined. One MCM matrix is generated for each of the four neighborhood directions, namely, 0°, 45°, 90° and 135°.

Deng and Manjunath (2001) proposed a two-stage method called JSEG, which combines color and texture after image segmentation. In the first stage, colors are quantized to the required levels for differentiating between various regions of an image. Pixel values of the regions are then replaced by their quantized color levels to form a color map. Spatial variation of color levels between different regions in the map is viewed as a type of texture composition of the image. Yu et al. (2002) suggested the use of color texture moments to represent both color and texture of an image. This approach is based on the calculation of Local Fourier Transformation (LFT) coefficients. Eight templates equivalent to LFT are operated over an image to generate a characteristic map of the image. Each template is a 3 × 3 filter that considers eight neighbors of the current pixel for LFT calculation. First and second order moments of the characteristic map are then used to generate a set of features.

In this paper, we propose an integrated approach for capturing spatial variation of both color and intensity levels in the neighborhood of each pixel using the HSV color space. In contrast to the other methods, for each pixel and its neighbor, the amount of color and intensity variation between them is estimated using a weight function. Suitable constraints are satisfied while choosing the weight function for effectively relating visual perception of color and the HSV color space properties. The color and intensity variations are represented in a single composite feature known as Integrated Color and Intensity Co-occurrence Matrix (ICICM). While the existing schemes generally treat color and intensity separately, the proposed method provides a composite view to both color and intensity variations in the same feature. The main advantage of using ICICM is that it avoids the use of weights to combine individual color and texture features. We use ICICM feature in an image retrieval application from large image databases. Early result on this work was reported in (Vadivel et al., 2004a). In the next section, we describe the proposed feature extraction technique after introducing some of the properties of the HSV color space. Choice of quantization levels for color and intensity axes, selection of parameter values and a brief overview of the image retrieval application is given in Section 3. Retrieval performance of the proposed scheme with labeled and unlabelled databases is presented in Section 4, and we conclude in the last section of the paper.

Section snippets

Integrated color and intensity co-occurrence matrix

We propose to capture color and intensity variation around each pixel in a two-dimensional matrix called Integrated Color and Intensity Co-occurrence Matrix (ICICM). This is a generalization of the Grayscale Co-occurrence Matrix and the Color Co-occurrence Matrix techniques. For each pair of neighboring pixels, we consider their contribution to both color perception as well as gray level perception to the human eye. Some of the useful properties of the HSV color space and their relationship to

Application of ICICM

The Integrated Color and Intensity Co-occurrence Matrix can be used in a number of image processing and pattern recognition problems. We have considered Content-based Image Retrieval applications to study the effectiveness of the ICICM matrix. In the following two sub-sections, we show the effect of quantization levels and the choice of parameters r1 and r2 in Eq. (9) on image retrieval performance. We use two standard metrics, namely, recall and precision for measuring performance, which are

Retrieval performance

We have used two separate of image databases for measuring the performance of ICICM in content-based image retrieval applications. The first is a database contains 10,000 images from IMSI3 master clips. The second database was generated by crawling 28,000 images from the World Wide Web. In the next two sub-sections, we present our experimental results using these databases.

Conclusions

We have proposed an integrated color and intensity based co-occurrence matrix and shown its usefulness in image retrieval applications. We effectively use some of the properties of the HSV space and their relationship to human visual perception for representation of color and intensity in the co-occurrence matrix. Each pixel and its neighbor contribute to both color as well as gray level perception in the neighborhood. Four components of the co-occurrence matrix are updated with relative

Acknowledgements

The work done by Shamik Sural is supported by research grants from the Department of Science and Technology, India, under Grant No. SR/FTP/ETA-20/2003 and by a grant from IIT Kharagpur under ISIRD scheme No. IIT/SRIC/ISIRD/2002–2003. Work done by A.K. Majumdar is supported by a research grant from the Department of Science and Technology, India, under Grant No. SR/S3/EECE/024/2003-SERC-Engg.

References (21)

  • C. Palm

    Color texture classification by integrative co-occurrence matrices

    Pattern Recognition

    (2004)
  • Aksoy, S., Haralick, R.M., 2000. A Weight Distance Approach to Relevance Feedback. In: Proc. Internat. Conf. on Pattern...
  • Y. Deng et al.

    Unsupervised segmentation of color-texture regions in images and video

    IEEE Trans. Pattern Anal. Machine Intell.

    (2001)
  • French, J.C., Watson, J.V.S., Jin, X., Martin, W.N., 2003. Integrating multiple multi-channel CBIR systems. In: Proc....
  • T. Gevers et al.

    Robust histogram construction from color invariants for object recognition

    IEEE Trans. Pattern Anal. Machine Intell.

    (2004)
  • R.C. Gonzalez et al.

    Digital Image Processing

    second ed.

    (2002)
  • R.M. Haralick et al.

    Textural features for image classification

    IEEE Trans. Systems Man Cybernat.

    (1973)
  • Huang, J., Ravi Kumar, S., Mandar, M., Wei-Jing, Z., Ramin, Z., 1997. Image indexing using color corellograms. In:...
  • Leow, W.K., Li, R., 2004. The analysis and applications of adaptive-binning color histograms. In: Proc. Computer Vision...
  • B.S. Manjunath et al.

    Color and texture descriptors

    IEEE Trans. Circuits Systems Video Technol.

    (2001)
There are more references available in the full text version of this article.

Cited by (91)

  • A new framework for Person Re-identification: Integrated level feature pattern (ILEP)

    2021, KSII Transactions on Internet and Information Systems
View all citing articles on Scopus
View full text