Abstract
In this work, we propose a bootstrapped generalized mixture estimation algorithm for synthetic aperture radar image segmentation. The Bootstrap sampling reduces the dependence effect of pixels in real images, and reduces segmentation time. Given an original image, we randomly select small representative set of pixels. Then, a generalized expectation maximization algorithm based on optimal Bootstrap sample is released for mixture identification. The generalized aspect comes from the use of distributions from the Pearson system. We validate the proposed algorithm on the classification of SAR images. The results we obtain show that the bootstrap sampling method yield the same accuracy and robustness of image classification as the basic algorithm while reducing time computing. This fact make possible the integration of such technique in real time applications.
You have full access to this open access chapter, Download conference paper PDF
1 Introduction
Synthetic Aperture Radar images (SAR images) of the earth are an important tool for many scientific applications such as high resolution remote sensing for mapping, surface surveillance, search-and-rescue, mine detection, navigation position location, and Automatic Target Recognition. For most of the applications mentioned, besides the importance of segmentation quality in the subsequent analysis for target detection and recognition, the time computing can play a key role. This constraint is difficult to reach because of the large size of SAR images. Amongst several existing works especially designed for SAR data [11, 19], one finds the contribution of the unsupervised statistical approach [17, 20].
Tree main statistical approaches have been suggested depending on the way that neighborhood influences the classification of a given pixel. The local or blind approaches consider that pixels are spatially independent [18]. Contextual methods [14] take into account a neighborhood of limit extent. Global approaches [17] assume that all pixels in an image influence the classification of the pixel of interest. Different estimation algorithms have been applied to the problem of unsupervised segmentation. The most popular one is the expectation-maximization (EM) algorithm [1, 6], with the maximum likelihood (ML) estimation. Many variants of the EM have been presented such as Stochastic EM (SEM) [14], Gibsien EM (GEM) [5], Iterative Conditional Estimation (ICE) [2, 4] and Zhang algorithms [21]. Because of the inherent speckle noise, it is now accepted that the statistics of SAR images can be well modeled by the family of probability distributions known as the Pearson system [8, 12].
We base our work in the field of statistical approaches; the mixture identification will be performed with the generalized mixture expectation maximization algorithm GMEM where the generalized aspect comes from the use of the Pearson system.
In statistical segmentation, the time increases with the size of training data set. Because of the large size of SAR images, the size of the training data is very large. As a result, the time required to segmentation could be prohibitively large which constrain its use in the real-world applications.
In this context, we propose a fast segmentation algorithm based on the principle of Bootstrap sampling. The Bootstrap [9] has not seen much applications in image analysis and classification. The first attempt to apply the bootstrap to image analysis was by Ghorbel and Banga [10] who introduced a Bootstrap scheme in the context of Bayesian image segmentation under the Gaussian assumption. M’hiri et al. [15, 16] extended the technique later and applied them to the segmentation of brain images.
In this paper, we propose to combine the bootstrap technique with the generalized mixture expectation maximization algorithm GMEM for unsupervised SAR image classification. The Bootstrap technique consists on selecting a small representative set a pixels from the original image and mixture identification will be done on bootstrap samples instead of the correlated pixels in the real image. The use of such resampling procedure enables to reduce considerably the computation times while preserving the estimation equality. The reminder of this paper is organized as follow. In Sect. 2, we describe the unsupervised Bayesian segmentation by Bootstrapping GMEM algorithm, Sect. 3 is devoted to experimental results of SAR image classification. Conclusions and future prospects are presented in Sect. 4.
2 Unsupervised Image Segmentation by the BGMEM Algorithm
2.1 Bootsrap Sampling for Image Classification
Efron [9] introduces the Bootstrap term to designate the set of the random re-sampling procedures of the data observed intended to be approached by simulating the statistics of the underlying distribution. The Bootstrap theory is based on the convergence of the empirical law of the sample towards the underlying unknown law when the sample size is sufficiently large.
Given a random sample \(\chi _N=({X_1,X_2,..,X_N}\)) of size N from a population with distribution \(F_X\). The bootstrap approximation was to estimate the distribution of a given unknown statistics \(R(\chi _N,F_X)\) by the bootstrap distribution noted \(F_n^*\) corresponding to the sample \(\chi _N^*={X_1^*,X_2^*,..,X_n^*},(n<N)\) where \( X_1^*,X_2^*,..,X_N^*\) are randomly selected from \(\chi _N\). Since the empirical distribution converges almost surely to the underlying distribution, one can hope that the bootstrap distribution would converge to the true unknown distribution. Details and applications of bootstrap technique can be found in Ref. [3, 22].
Ghorbel and Banga [10] introduce a bootstrap sampling scheme for gray levels image analysis. In this case, the whole image distribution is presented only by a small bootstrap sample size but should be representative of the entire image.
Suppose we have a two-dimensional image of \(N_r*N_c=N\) pixel resolution with \(N_r\) rows and \(N_c\) columns. In statistical segmentation, we suppose that the image is a finished population of N observations. It is then noted \(Y=({y_1, y_2,...,y_N})\) that is to say a sample from an unknown distribution. From this original image, we construct the bootstrap sample \(Y^*=(y_1^*,...,y_n^*)\), \((n<N)\) by randomly selecting n pixels \(y_i^*=(k_i,l_i)^t\) for \(i\in \left\{ 1,..,N\right\} \). \(k_i\) and \(l_i\) are obtained by making independent uniform random trials under the set \(\left\{ 1,..,N_r\right\} \) and the set \(\left\{ 1,..,N_c\right\} \) respectively. The \(Y^*=(y_1^*,...,y_n^*)\) is a resample of size n chosen with replacement from Y.
The optimal size of the bootstrap sample is determined by two representative criteria for the use of this technique in image segmentation.
The sample \(Y^*= (y_1^*,...,y_n^*)\) is representative of an image Y when each image gray level (GL) appears at least one time in the samples. It’s explicitly shown by the C1 and C2 equations.
D is the number of the different gray levels values in the image, \(B(n_0)\) is the sampling characteristic function and \(\epsilon \) is a fixed small value.
The advantage of the C1 criterion is to be able to depart from a minimum sample size \(n_0\). The sampling characteristic function \(B(n_0)\) defined by the criterion C2 takes into account both image pixel distribution and bootstrap sample size . The \(n_0\) is initially computed by the C1 criterion then \(n_0\) is progressively increased. The corresponding bootstrap sample is constructed for each size \(n_0\) by random uniform selecting set of pixels from the image and the C2 criterion is verified . The bootstrap sample is qualify to be representative when the sampling characteristic function value is lower than the expected precision \(\epsilon =10^{-2}\).
An empirical convergence study based on minimizing the mean integrated square error (MISE) between original density and its estimate based on the bootstrap sample show that the fixed precision \(\epsilon =10^{-2}\) ensure the representativity of the bootstrap sample.
2.2 An Overview of the Pearson System of Distributions
The Pearson system [13] is made up of mainly height families of distributions including Gaussian, Gamma and Beta ones and offers a large variety of shapes (symmetrical or not, with finite or semi-finite support, etc.). Each law can be uniquely defined by its mean \(\mu _1\) and its first three centered moments \((\mu _2, \mu _3, \mu _4)\). All of them can be represented in the so-called Pearson diagram (Fig. 1) in which axes \(\beta _1\) and \(\beta _2\) are given by \(\beta _1=\frac{(\mu _3)^2}{(\mu _2)^3}\) and \(\beta _1=\frac{(\mu _4)}{(\mu _2)^3}\).
Gaussian distributions are located at \((\beta _1=0,\beta _2=3)\). Gamma distributions on the straight line \(\beta _2=1.5 \beta _1+3\) and inverse gamma distribution on the curve with the equation \(\beta _2=\frac{3}{\beta _1-32}(-13\beta _1-16-2(\beta _1+4)^\frac{3}{2})\). First kind Beta distributions are located between the lower limit and the gamma line, second kind Beta distributions are located between the gamma and the inverse Gamma distributions, and Type 4 distributions are located between the inverse Gamma distributions and the upper limit. Then it is possible to estimate empirical moments of a distribution from a sample and to assess the family of distributions from coordinates \((\beta _1,\beta _2)\) and determine the parameters that precisely characterize the probability density function within its family.
2.3 Description of the Bootstrapped Generalized Expectation Maximization Algorithm
Before proceeding with mixture identification by the BGMEM algorithm, a pre-processing step is needed to determine the size of the representative bootstrap sample because the performance of the mixture identification depends on the pre-processing step. The sample size drawn from the image fulfills the criteria proposed by Ghorbel and Banga in Sect. 2.1. Let given an available bootstrap sample \(Y^*=(y_1^*,...,y_n^*)\) from the original image Y.
The mixture identification by BGMEM algorithm is performed using only the representative Bootstrap sample instead of using the entire image. After initializing parameters from the histogram, the two following steps are iterated.
-
Expectation Step: For each class, the distribution \(f_k\) will be selected from the Pearson system distribution according to the skewness and kurtosis values [7] then the a posterior probability for a pixel \(y_i^*\) to belong to class k at the iteration q is given by:
$$\begin{aligned} \forall k\in \{1,...,K\} \,\,\,\,\,\,\ P(x_k|y_i^*,\theta ^q)= \frac{\displaystyle {\pi _k^{q-1}f_k(y_i^*|\theta _k^{(q-1)})}}{\displaystyle \sum _{l=1}^K \pi _l^{q-1}f_l(y_i^*|\theta _l^{q-1})} \ \end{aligned}$$(3) -
Maximization Step: the parameters of the mixture are constructed in the following way:
Hight order moments \(\mu _{j}\) with \(j={2,3,4}\) are computed according the equation:
At the \(q^th\) iteration on the sample the skewness and the kurtosis of the class k are computed as:
The algorithm applied to the sample stops when the sample estimated parameters are stagnated.
3 Unsupervised Bayesian Classification of SAR Images
In the unsupervised Bayesian classification, the image is supposed as a realization of mixture distributions and the classification problem consists on mixture identification associated to a Bayesian rule. A large study is done to validate the BGMEM framework on a large data set of both synthetic images and real SAR images. In this section, we will show results only for four SAR images. The original SAR images are \(512*512\) pixels resolutions presented by 256gray levels classified in four classes. Before proceeding with image classification by the BGMEM algorithm, a pre-processing step is needed to determine the size of the representative bootstrap sample according to the representativity criteria presented in Sect. 2.1. The bootstrap sample size must be large enough to ensure a good estimation and small enough to reduce the computation time. The sample size \(n_0=3000\) pixels is validate according to the sampling characteristic function equation \((B(3000)<0.004)\) for the four images.
The Fig. 2a shows the fitting of the estimated densities to the image histogram in the case of classical GMEM algorithm for the four images and the Fig. 2b shows the result in the Bootstapped case. The conditional estimated densities fit the image histogram in the two cases of classical and bootstrapped algorithm.
We present in the Table 1 a comparison of the Mean Square Error MSE between the classical GMEM algorithm and the bootstrapped one. The mean square error shows an improvement of estimation when considering the bootstrapped version of the algorithm. The observations are decorrelated by randomly selecting observations which construct the sample. This fact offer best conditions of application of the GMEM algorithm and so better parameter estimation.
The Fig. 3b and c represent the segmentation results we obtained by applying the classical GMEM and the Bootstrapped BGMEM algorithms. We can observe visually a same segmentation quality by the both algorithms which is consequently of good mixture identification. We dont focus our work in the interpretation of the SAR images so the segmentation quality is showed visually but the contribution here was in accelerating the time computing of image classification. The computational time of the Boostrap algorithm and the classical one are given in Table 2. The programs are done with matlab and turned under a core i5 processor time computing may be lower if the algorithms are running on a more powerful processor. On the basis of the bootstrap sample which represents 0.01 of the image size, we have considerable gain in time computing more than a factor of 87 by the boostrapped algorithm with the same accuracy and robustness of image classification.
4 Conclusion
In this paper, we have proposed a Bootstrap model to the unsupervised Bayesian image segmentation. The Bootstrap technique allows an estimation of parameters of the image from a small sized sample instead of the entire image. The principle advantage of the proposed algorithm is the reduction of time computing which make it useful in real-time applications. As future work, we are interest to the study of multivariate Pearson system for the statistical modeling of images acquired by different sensors.
References
Ambroise, C., Govaert, G.: Convergence of an EM-type algorithm for spatial clustering. Pattern Recognit. Lett. 19, 919–927 (1998)
Braathen, B., Pieczynski, W., Masson, P.: Global and local methods of unsupervised Bayesian segmentation of images. Mach. Graph. Vis. 2, 39–52 (1993)
Basiri, S., Ollila, E., Koivunen, V.: Robust, scalable, and fast bootstrap method for analyzing large scale data equation. IEEE Trans. Signal Process. 64(4), 1007–1017 (2015)
Caillol, H., Pieczynski, W., Hillion, A.: Estimation of fuzzy Gaussian mixture and unsupervised statistical image segmentation. IEEE Trans. Image Process. 6, 425–440 (1997)
Chalmond, B.: An iterative Gibbsian technique for reconstruction of m-ary images. Pattern Recognit. Lett. 22, 747–761 (1989)
Dempester, A.P., Laird, N.M., Rubin, D.B.: Maximum likelihood from incomplete data via the E.M algorithm. J. R. Stat. Soc., Series B 39(1), 1–38 (1977)
Delignon, Y., Marzouki, A., Pieczynski, W.: Estimation of generalised mixture and its application in image segmentation. IEEE Trans. Image Process. 6(10), 1364–1375 (1997)
Delignon, Y., Garello, R., Hillion, A.: Statistical modelling of ocean SAR images. IEE Proc. Radar, Sonar and Navig. 44(66), 348–354 (1997)
Efron, B.: Bootstrap method : another look at the Jackknife. Ann. Stat. 7, 1–26 (1979)
Ghorbel, F., Banga, C.: Bootstrap sampling applied to image analysis, invited paper, special session. IEEE-ICASSP, Adelaide, S. Aust. 6, 81–84 (1994)
Gong, M.G., Zhou, Z.Q., Ma, J.J.: Change detection in synthetic aperture radar images based on image fusion and fuzzy clustering. IEEE Trans. Image Process. 21(4), 2141–2151 (2012)
Inglada, J.: Change detection on SAR images by using a parametric estimation of the Kullback-Leibler divergence. In: Proceeding IEEE International Conference Geoscience and Remote Sensing (IGARSS), pp. 4104–4106, Toulouse (2003)
Johnson, N.L., Kotz, S.: Distribution in Statistics: Continuous Univariate Distributions, vol. 1 and 2. Wiley, New York (1994)
Masson, P., Pieczynski, W.: SEM algorithm and unsupervised statistical segmentation of satellite images. IEEE Trans. Geos. Rem. Sen. 31, 618–633 (1993)
Mhiri, S., Cammoun, L., Ghorbel, F.: Speeding up HMRF EM algorithms for fast unsupervised image segmentation by Bootstrap resampling: Application to the brain tissue segmentation. J. Signal Process. 87, 2544–2559 (2007)
M’hiri, S., Mabrouk, S., Ghorbel, F.: Segmentation des IRM cerebrales par une variante bootstrapee du HMRF-EM : etude preliminaire sur fantomes. IRBM 33(1), 2–10 (2012)
Pal, N.R., Pal, S.K.: A review on image segmentation techniques. Pattern Recogn. 26(9), 1277–1294 (1993)
Peng, A., Pieczynski, W.: Adaptive mixture estimation and unsupervised local Bayesian image segmentation. Graph. Models Image Process. 57, 389–399 (1995)
Yang, D., Wang, L., Hei, X., Gong, M.: An efficient automatic Sar segmentation framework in AIS using kernel clustering index and histogram statistics. App. Soft Comput. 16, 63–79 (2014)
Zhang, Y.J.: Evaluation and comparison of different segmentation algorithms. Pattern Recogn. Lett. 18, 963–974 (1997)
Zhang, Y.J., Modestino, J.W., Langan, D.A.: Maximum likelihood parameter estimation for unsupervised stochastic model-based image segmentation. IEEE Trans. Image Process. 3, 404–420 (1994)
Zoubir, A., Iskander, D.: Bootstrap methods and applications : a tutorial for the signal processing practitioner. IEEE Signal Process. Magazine 24(4), 10–19 (2007)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing Switzerland
About this paper
Cite this paper
Bougarradh, A., Mhiri, S., Ghorbel, F. (2016). Unsupervised Classification of Synthetic Aperture Radar Imagery Using a Bootstrap Version of the Generalized Mixture Expectation Maximization Algorithm. In: Mansouri, A., Nouboud, F., Chalifour, A., Mammass, D., Meunier, J., Elmoataz, A. (eds) Image and Signal Processing. ICISP 2016. Lecture Notes in Computer Science(), vol 9680. Springer, Cham. https://doi.org/10.1007/978-3-319-33618-3_28
Download citation
DOI: https://doi.org/10.1007/978-3-319-33618-3_28
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-33617-6
Online ISBN: 978-3-319-33618-3
eBook Packages: Computer ScienceComputer Science (R0)