Conference PaperPDF Available

Robust Image Segmentation with Mixtures of Student's t-Distributions

Authors:

Abstract and Figures

Gaussian mixture models have been widely used in image segmentation. However, such models are sensitive to outliers. In this paper, we consider a robust model for image segmentation based on mixtures of Student's t -distributions which have heavier tails than Gaussian and thus are not sensitive to outliers. The t -distribution is one of the few heavy tailed probability density functions (pdf) closely related to the Gaussian, that gives tractable maximum likelihood inference via the Expectation-Maximization (EM) algorithm. Numerical experiments that demonstrate the properties of the proposed model for image segmentation are presented.
Content may be subject to copyright.
ROBUST IMAGE SEGMENTATION WITH MIXTURES OF STUDENT’S t-DISTRIBUTIONS
Giorgos Skas Christophoros Nikou Nikolaos Galatsanos
University of Ioannina,
Department of Computer Science,
PO Box 1185, 45110 Ioannina, Greece,
{skas, cnikou, galatsanos}@cs.uoi.gr
ABSTRACT
Gaussian mixture models have been widely used in image
segmentation. However, such models are sensitive to outliers.
In this paper, we consider a robust model for image segmenta-
tion based on mixtures of Student’s t-distributions which have
heavier tails than Gaussian and thus are not sensitive to out-
liers. The t-distribution is one of the few heavy tailed proba-
bility density functions (pdf) closely related to the Gaussian,
that gives tractable maximum likelihood inference via the Ex-
pectation-Maximization (EM) algorithm. Numerical experi-
ments that demonstrate the properties of the proposed model
for image segmentation are presented.
Index Terms Image segmentation, clustering, Student’s
t-distribution, mixture model, EM algorithm, segmentation
evaluation.
1. INTRODUCTION
Image segmentation is the process of grouping image pixels
based on the coherence of certain attributes such as intensity,
color or texture. Many approaches have been proposed to
solve the image segmentation problem. For surveys on this
topic the reader may refer to [1]. In this paper, we will focus
our attention to image segmentation methods based on clus-
tering. Clustering is the process of arranging data into groups
having common characteristics and is a fundamental problem
in many elds of science [2]. Thus, image segmentation can
be viewed a special type of clustering. Usually, in image seg-
mentation, our data, the image pixels have spatial locations
associated with them. Thus, apart from the commonality of
attributes such as intensity, color or texture, commonality of
location is an important characteristic of the grouping that we
are seeking in image segmentation.
More specically, in this paper we will focus our attention
on clustering methods based on the modeling of the probabil-
ity density function (pdf) of the data via nite mixture models
(FMM) [3, 4]. Modeling the pdf of data with FMM is a nat-
ural way to cluster data because it automatically provides a
This work was partially supported by Interreg IIIA (Greece-Italy) grant
I2101005.
grouping of the data based on the components of the mixture
that generated them. More specically, FMM are based on
the assumption that each datum originates from one compo-
nent of the mixture according to some probability. Thus, this
probability can be used to assign each datum to the compo-
nent that has most likely generated it. Furthermore, the like-
lihood of an FMM is a rigorous measure for evaluating the
clustering performance [4].
FMM based pdf modeling with Gaussian components has
been used successfully in a number of applications ranging
from bioinformatics [5] to image retrieval [6]. The parameters
of Gaussian mixture models (GMM) can be estimated very
efciently through maximum likelihood (ML) estimation us-
ing the EM algorithm [7]. Furthermore, it can be shown that
Gaussian components allow efcient representation of any
pdf [4]. However, it is well known that GMM are sensi-
tive to outliers and may lead to excessive sensitivity to small
numbers of data points. The problem of providing protec-
tion against outliers in multivariate data is very difcult and
increases with the dimensionality.
In this paper, we apply to the image segmentation prob-
lem mixture models with Student-t pdf components. This pdf
has heavier tails as compared to the exponentially decaying
tails of a Gaussian [8]. Thus, each component in the mixture
originates from a wider class of elliptically symmetric dis-
tributions with an additional parameter called the degrees of
freedom. Hence a more robust model is used than the classi-
cal normal mixture.
In the remainder of this paper, background on standard
GMM is given in Section 2 and the mixture of multivariate
t-distributions and the EM algorithm for parameter estima-
tion are described in Section 3. Results on image segmenta-
tion and comparisons with the standard GMM are presented
in Section 4 and conclusions are drawn in Section 5.
2. BACKGROUND ON STANDARD GAUSSIAN
MIXTURE MODELS
Let X denote the vector of features representing an image
spatial location (pixel). The GMM assumes that the pdf of
I - 2731-4244-1437-7/07/$20.00 ©2007 IEEE ICIP 2007
the observation x is expressed by
φ(x; Ω)=
K
i=1
π
i
f(x; μ
i
, Σ
i
) (1)
where Ω is the mixture parameter set Ω =[Ω
1
, Ω
2
, ..., Ω
K
]
with Ω
i
=(π
i
i
, Σ
i
). For the mixing proportions of the i
th
component π
i
, we have that
0 π
i
1,i=1, 2, ..., K,
K
i=1
π
i
=1 (2)
For each component of the model in (1), the Gaussian pdf is
expressed by
f(x; μ
i
, Σ
i
)=
1
(2π)
d
2
|Σ
i
|
1
2
exp
1
2
(xμ
i
)
T
Σ
1
i
(xμ
i
)
(3)
where d is the dimensionality of the vector (e.g. intensity,
location, texture features) and μ
i
, Σ
i
are the mean vector and
covariance matrix respectively.
Training of a GMM, or in other words nding its ML so-
lution, can be performed using the EM algorithm [7]. The
EM algorithm is a well-known numerical method used in a
variety of ML problems. In the case of a GMM, each im-
age pixel x, is associated with a binary hidden variable z of
dimension K, whose k
th
component has a value of 1 if the
observation (i.e. the pixel) was produced by that component
and is zero otherwise. In the E-step of the algorithm, the ex-
pected value of the hidden variables conditioned on the obser-
vation is computed. These expected values give the probabil-
ities that a given datum originates from a different component
of the mixture. Thus, they provide a means for segmenting
the data. In the M-step, the model parameters (mean, covari-
ance and mixing proportions) are computed by maximizing
the log-likelihood of the complete data (hidden variables and
observations). This scheme is repeated iteratively until con-
vergence is achieved.
3. MIXTURE OF STUDENT’S t-DISTRIBUTIONS
AND THE EM ALGORITHM
A d-dimensional random variable X follows a multivariate
t-distribution with mean μ, positive denite, symmetric and
real d × d covariance matrix Σ and has ν [0, ) degrees
of freedom when, given the weight u, the variable X has the
multivariate normal distribution with mean μ and covariance
Σ/u:
X|μ, Σ,u N(μ, Σ/u),
and the weight u follows a Gamma distribution parameterized
by ν:
u Gamma(ν/2/2).
Integrating out the weights from the joint density leads to the
density function of the marginal distribution:
p(x; μ, Σ)=
Γ
ν+d
2
|Σ|
1
2
(πν)
d
2
Γ
ν
2
[1 + ν
1
δ(x, μ; Σ)]
ν+d
2
(4)
where δ(x, μ;Σ)=(x μ)
T
Σ
1
(x μ) is the Mahalanobis
squared distance and Γ is the Gamma function. It can be
shown that for ν →∞the Student’s t-distribution tends to
a Gaussian distribution with covariance Σ. Also, if ν>1, μ
is the mean of X and if ν>2, ν(ν 2)
1
Σ is the covariance
matrix of X. Therefore, the family of t-distributions provides
a heavy-tailed alternative to the normal family with mean μ
and covariance matrix that is equal to a scalar multiple of Σ,
if ν>2 (g. 1).
Fig. 1. The Student’s t-distribution for various degrees of
freedom. As ν →∞the distribution tends to a Gaussian.
For small values of ν the distribution has heavier tails than a
Gaussian.
A Student’s t-distribution mixture model (SMM) may also
be trained using the EM algorithm [8]. A K-component mix-
ture of t-distributions is given by
φ(x, Ψ) =
K
i=1
π
i
p(x; μ
i
, Σ
i
i
) (5)
where x =(x
1
, ..., x
N
)
T
denotes the observed-data vector
and
Ψ=(π
1
, ..., π
K
1
, ..., μ
K
, Σ
1
, ..., Σ
K
1
, ..., ν
K
)
T
. (6)
are the parameters of the components of the mixture.
Consider now the complete data vector
x
c
=(x
1
, ...x
N
,z
1
, ..., z
N
,u
1
, ..., u
N
)
T
(7)
where z
1
, ..., z
N
are the component-label vectors and z
ij
=
(z
j
)
i
is either one or zero, according to whether the observa-
tion x
j
is generated or not by the i
th
component. In the light
I - 274
of the denition of the t-distribution, it is convenient to view
that the observed data augmented by the z
j
, j =1, ..., N are
still incomplete because the component covariance matrices
depend on the degrees of freedom. This is the reason that
the complete-data vector also includes the additional missing
data u
1
, ..., u
N
. Thus, the E-step on the (t +1)
th
iteration
of the EM algorithm requires the calculation of the posterior
probability that the datum x
j
belongs to the i
th
component of
the mixture:
z
t+1
ij
=
π
t
i
p(x
j
; μ
t
i
, Σ
t
i
t
i
)
K
m=1
p(x
j
; μ
t
m
, Σ
t
m
t
m
)
(8)
as well as the expectation of the weights for each observation:
u
t+1
ij
=
ν
t
i
+ d
ν
t
i
+ δ(x
j
t
i
t
i
)
(9)
Maximizing the log-likelihood of the complete data pro-
vides the update equations of the respective mixture model
parameters:
π
t+1
i
=
1
N
N
j=1
z
t
ij
t+1
i
=
N
j=1
z
t
ij
u
t
ij
x
j
N
j=1
z
t
ij
u
t
ij
, (10)
Σ
t+1
i
=
N
j=1
z
t
ij
u
t
ij
(x
j
μ
t+1
i
)(x
j
μ
t+1
i
)
T
N
j=1
z
t+1
ij
. (11)
The degrees of freedom for each component are computed as
the solution to the equation:
log
ν
t+1
i
2
ψ
ν
t+1
i
2
+1 log
ν
t
i
+ d
2
+
+
N
j=1
z
t
ij
(log u
t
ij
u
t
ij
)
N
j=1
z
t
ij
+ ψ
ν
t
i
+ d
2
=0 (12)
where ψ(x)=
(lnΓ(x))
∂x
is the digamma function. A detailed
derivation of the EM algorithm for Student’s t-mixtures is pre-
sented in [8].
4. EXPERIMENTAL RESULTS
In this paper, we employed an 8-dimensional vector as a fea-
ture for each image pixel [9]. The rst three components of
Table 1. Number of images (over 30) where the SMM pro-
vides lower quantization error than the GMM for p =1.2.
Noise type K =3 K =5 K =7
noise free 23 17 24
uniform 20 dB 20 19 21
uniform 14 dB 23 17 18
uniform 7 dB 26 15 17
salt pepper 10% 19 15 16
the feature vector are the Lab color coordinates, the next three
components are texture descriptors, namely, the polarity, the
anisotropy and the contrast as described in [9] and the remain-
ing two coordinates are the horizontal and vertical pixel loca-
tions. Prior to model training, each feature vector component
was separately normalized to ensure that no feature dominates
the others.
In order to evaluate the proposed segmentation scheme
and compare it to GMM segmentations we compute a quan-
tization error for 30 images provided by the Berkeley image
segmentation data base [10]. The quantization error, for each
pixel location, is dened as the distance between the image
feature and the mean of the mixture component that gener-
ated the measure (i.e. the component with the larger mixing
proportion). This p-norme distance, between a d-dimensional
feature vector x and the mean vector μ is dened as
D(x, μ)=
d
i=1
(x
i
μ
i
)
p
1
p
(13)
We have experimented with different values of p, namely 0.7,
1.2 and 2 (Euclidean distance). It is well known that norms
close to 1 measure a quantization error that better corresponds
to human perceptual characteristics. These experiments were
performed by degrading the images by uniform and salt-and-
pepper noise of varying strength. Also, the predened number
of kernels varied (K =3, 5, 7).
Let us also notice that the experiments were performed us-
ing a variation of the standard EM algorithm, called Greedy
EM [11] providing a segmentation result independent of the
model initialization. The performance of the model is pre-
sented in table 1. A comparison is shown in table 2 where
one can see that the SMM has a slight yet better performance
than the GMM. Some segmentation results are depicted in g.
2 and 3 where it can be observed that SMM provide smoother
segmentations than the standard GMM.
5. CONCLUSION
We have presented a methodology for image segmentation
based on mixtures of Student’s t-distributions. The model
can account for outliers values and thus provides smoother
I - 275
Tabl e 2. Quantization error statistics for 30 images of the
Berkeley segmentation data base for all the congurations of
the uniform noise (see table 1).
K =3 K =5 K =7
GMM SMM GMM SMM GMM SMM
p =0.7
mean 11.54 11.54 10.27 10.23 9.59 9.51
s. d. 0.85 0.92 0.87 0.89 0.88 0.92
p =1.2
mean 3.85 3.83 3.48 3.47 3.27 3.25
s. d. 0.24 0.25 0.25 0.26 0.27 0.28
p =2.0
mean 2.26 2.25 2.07 2.07 1.96 1.96
s. d. 0.12 0.13 0.14 0.14 0.15 0.15
Original GMM SMM
Fig. 2. Segmentation examples using the GMM and the SMM
methods for K =5components.
Original GMM SMM
Fig. 3. Segmentation of a MRI brain image into K =3
classes (white matter, grey matter and cerebrospinal uid).
segmentations than the standard GMM. However, important
issues for mixture based clustering still need to be addressed.
Such issues are how the number of model components can
be selected automatically and which features should be used.
These are open questions and are subject of current research.
6. REFERENCES
[1] N. Pal and S. Pal, A review of image segmentation
techniques, Pattern Recognition, vol. 26, pp. 1277–
1294, 1993.
[2] R. Xu and D. Wunsch II, “Survey of clustering algo-
rithms, IEEE Transactions on Neural Networks, vol.
16, no. 3, pp. 645–678, 2005.
[3] C. M. Bishop, Pattern Recognition and Machine Learn-
ing, Springer, 2006.
[4] G. McLachlan, Finite mixture models, Wiley-
Interscience, 2000.
[5] K. Blekas, A. Likas, N. Galatsanos, and I. Lagaris, A
spatially constrained mixture model for image segmen-
tation, IEEE Transactions on Neural Networks, vol. 16,
no. 2, pp. 494–498, 2005.
[6] H. Greenspan, G. Dvir, and Y. Rubner, “Context-
dependent segmentation and matching in image data-
bases, Computer Vision and Image Understanding, vol.
93, no. 1, pp. 86–109, 2004.
[7] P. Dempster, N. M. Laird, and D. B. Rubin, “Max-
imum likelihood from incomplete data via EM algo-
rithm, Journal of the Royal Statistical Society, vol. 39,
no. 1, pp. 1–38, 1977.
[8] D. Peel and G. J. McLachlan, “Robust mixture modeling
using the t-distribution, Statistics and Computing, vol.
10, pp. 339–348, 2000.
[9] C. Carson, S. Belongie, H. Greenspan, and J. Ma-
lik, “Blobworld: image segmentation using expectation-
maximization and its application to image querying,
IEEE Transactions on Pattern Analysis and Machine In-
telligence, vol. 24, no. 8, pp. 1026–1038, 2002.
[10] D. Martin, C. Fowlkes, D. Tal, and J. Malik, A database
of human segmented natural images and its application
to evaluating segmentation algorithms and measuring
ecological statistics, in Proceedings of the 8th Inter-
national Conference one Computer Vision, July 2001,
vol. 2, pp. 416–423.
[11] N. Vlassis and A. Likas, A greedy EM algorithm for
Gaussian mixture learning, Neural Processing Letters,
vol. 15, pp. 77–87, 2002.
I - 276
... Here, we further extend this approach to dynamic inputs. tation [19,20,1]. ...
Conference Paper
Full-text available
We propose a family of probabilistic segmentation algorithms for videos that rely on a generative model capturing static and dynamic natural image statistics. Our framework adopts flexibly regularized mixture models (FlexMM) [1], an efficient method to combine mixture distributions across different data sources. FlexMMs of Student-t distributions successfully segment static natural images, through uncertainty-based information sharing between hidden layers of CNNs. We further extend this approach to videos and exploit FlexMM to propagate segment labels across space and time. We show that temporal propagation improves temporal consistency of segmentation, reproducing qualitatively a key aspect of human perceptual grouping. Besides, Student-t distributions can capture statistics of optical flows of natural movies, which represent apparent motion in the video. Integrating these motion cues in our temporal FlexMM further enhances the segmentation of each frame of natural movies. Our probabilistic dynamic segmentation algorithms thus provide a new framework to study uncertainty in human dynamic perceptual segmentation.
... (2) AFPDEFCM-fourth-order PDE-based fuzzy c-means [139]; (3) FCM-PDE-based fuzzy c-means [140]; (4) SMM-Student's-t mixture model [141]; (5) standard GMM [142]; and _6) GMM-SMSI-GMM with spatial pixel relationship extracted using a saliency map [143]. Further improvements were presented in fibrotic plaque detection by Liu et al., who demonstrated that a CNN based on the VGG-16 network outperformed the single-shot detector (SSD) and you only look once (YOLO)-v3 based models, with accuracies of 94.12%, 93.75%, and 64.89%, respectively [144][145][146][147][148][149]. ...
Article
Full-text available
Coronary optical coherence tomography (OCT) is an intravascular, near-infrared light-based imaging modality capable of reaching axial resolutions of 10–20 µm. This resolution allows for accurate determination of high-risk plaque features, such as thin cap fibroatheroma; however, visualization of morphological features alone still provides unreliable positive predictive capability for plaque progression or future major adverse cardiovascular events (MACE). Biomechanical simulation could assist in this prediction, but this requires extracting morphological features from intravascular imaging to construct accurate three-dimensional (3D) simulations of patients’ arteries. Extracting these features is a laborious process, often carried out manually by trained experts. To address this challenge, numerous techniques have emerged to automate these processes while simultaneously overcoming difficulties associated with OCT imaging, such as its limited penetration depth. This systematic review summarizes advances in automated segmentation techniques from the past five years (2016–2021) with a focus on their application to the 3D reconstruction of vessels and their subsequent simulation. We discuss four categories based on the feature being processed, namely: coronary lumen; artery layers; plaque characteristics and subtypes; and stents. Areas for future innovation are also discussed as well as their potential for future translation.
... Image denoising based on Student-t models was addressed in [27] and image deblurring in [12,47]. Further applications include robust image segmentation [3,34,42] and superresolution [19] as well as registration [14,48]. For learning MMs a maximizer of the corresponding log-likelihood has to be computed. ...
Article
Full-text available
Inertial algorithms for minimizing nonsmooth and nonconvex functions as the inertial proximal alternating linearized minimization algorithm (iPALM) have demonstrated their superiority with respect to computation time over their non inertial variants. In many problems in imaging and machine learning, the objective functions have a special form involving huge data which encourage the application of stochastic algorithms. While algorithms based on stochastic gradient descent are still used in the majority of applications, recently also stochastic algorithms for minimizing nonsmooth and nonconvex functions were proposed. In this paper, we derive an inertial variant of a stochastic PALM algorithm with variance-reduced gradient estimator, called iSPALM, and prove linear convergence of the algorithm under certain assumptions. Our inertial approach can be seen as generalization of momentum methods widely used to speed up and stabilize optimization algorithms, in particular in machine learning, to nonsmooth problems. Numerical experiments for learning the weights of a so-called proximal neural network and the parameters of Student-t mixture models show that our new algorithm outperforms both stochastic PALM and its deterministic counterparts.
... The ddCRP has been further extended to perform image segmentation using a region-based hierarchical representation [22]. Similarly in a series of papers, Nikou et al. have proposed to consider the mixing probabilities as Gauss-Markov random fields either directly [51,43,52] or hierarchically [44]. More recently and using comparable ideas, Sun et al. have introduced Location Dependent Dirichlet Processes (LDDP) [57] which use exponential of Gaussian processes in combination with Dirichlet processes to describe class distributions. ...
Article
Full-text available
Probabilistic finite mixture models are widely used for unsupervised clustering. These models can often be improved by adapting them to the topology of the data. For instance, in order to classify spatially adjacent data points similarly, it is common to introduce a Laplacian constraint on the posterior probability that each data point belongs to a class. Alternatively, the mixing probabilities can be treated as free parameters, while assuming Gauss–Markov or more complex priors to regularize those mixing probabilities. However, these approaches are constrained by the shape of the prior and often lead to complicated or intractable inference. Here, we propose a new parametrization of the Dirichlet distribution to flexibly regularize the mixing probabilities of over-parametrized mixture distributions. Using the Expectation-Maximization algorithm, we show that our approach allows us to define any linear update rule for the mixing probabilities, including spatial smoothing regularization as a special case. We then show that this flexible design can be extended to share class information between multiple mixture models. We apply our algorithm to artificial and natural image segmentation tasks, and we provide quantitative and qualitative comparison of the performance of Gaussian and Student-t mixtures on the Berkeley Segmentation Dataset. We also demonstrate how to propagate class information across the layers of deep convolutional neural networks in a probabilistically optimal way, suggesting a new interpretation for feedback signals in biological visual systems. Our flexible approach can be easily generalized to adapt probabilistic mixture models to arbitrary data topologies.
... One way to distinguish one class from the other is to look at their voxels' intensities. The use of Student's t distribution to cluster brain tissues using intensity information has been proposed in some works [31,32]. The main advantage of the Student's t distribution over other distributions, such as Gaussian Mixture Models (GMM), is that it has a "heavy" tail, which makes it more robust to outliers 4 [33] and, therefore, less susceptible to noise. ...
Chapter
Full-text available
Multiple sclerosis (MS) is an autoimmune demyelinating disease that affects one’s central nervous system. The disease has a number lesion states. One of them is known as active, or enhancing, and indicates that a lesion is under an inflammatory condition. This specific case is of interest to radiologists since it is commonly associated with the period of time a patient suffers most from the effects of MS. To identify which lesions are active, a Gadolinium-based contrast is injected in the patient prior to a magnetic resonance imaging procedure. The properties of the contrast medium allow it to enhance active lesions, making them distinguishable from nonactive ones in T1-w images. However, studies from various research groups in recent years indicate that Gadolinium-based contrasts tend to accumulate in the body after a number of injections. Since a comprehensive understanding of this accumulation is not yet available, medical agencies around the world have been restricting its usage to cases only where it is absolutely necessary. In this work we propose a supervised algorithm to distinguish active from nonactive lesions in FLAIR images, thus eliminating the need for contrast injections altogether. The classification task was performed using textural and enhanced features as input to the XGBoost classifier on a voxel level. Our database comprised 54 MS patients (33 with active lesions and 21 with nonactive ones) with a total of 22 textural and enhanced features obtained from Run Length and Gray Level Co-occurrence Matrices. The average precision, recall and F1-score results in a 6-fold cross-validation for active and nonactive classes were 0.892, 0.968, 0.924 and 0.994, 0.987, 0.991, respectively. Moreover, from a lesion perspective, the algorithm misclassified only 3 active lesions out of 157. These results indicate our tool can be used by physicians to get information about active MS lesions in FLAIR images without using any kind of contrast, thus improving one’s health and also reducing the cost of MRI procedures for MS patients.
... Image denoising based on Student t models was addressed in [17] and image deblurring in [6,34]. Further applications include robust image segmentation [4,25,29] as well as robust registration [8,35]. ...
Article
Full-text available
In this paper, we consider maximum likelihood estimations of the degree of freedom parameter ν, the location parameter μ and the scatter matrix Σ of the multivariate Student t distribution. In particular, we are interested in estimating the degree of freedom parameter ν that determines the tails of the corresponding probability density function and was rarely considered in detail in the literature so far. We prove that under certain assumptions a minimizer of the negative log-likelihood function exists, where we have to take special care of the case ν→∞, for which the Student t distribution approaches the Gaussian distribution. As alternatives to the classical EM algorithm we propose three other algorithms which cannot be interpreted as EM algorithm. For fixed ν, the first algorithm is an accelerated EM algorithm known from the literature. However, since we do not fix ν, we cannot apply standard convergence results for the EM algorithm. The other two algorithms differ from this algorithm in the iteration step for ν. We show how the objective function behaves for the different updates of ν and prove for all three algorithms that it decreases in each iteration step. We compare the algorithms as well as some accelerated versions by numerical simulation and apply one of them for estimating the degree of freedom parameter in images corrupted by Student t noise.
... Image denoising based on Student-t models was addressed in [23] and image deblurring in [14,44]. Further applications include robust image segmentation [4,32,39] and superresolution [19] as well as registration [16,45]. ...
Preprint
Inertial algorithms for minimizing nonsmooth and nonconvex functions as the inertial proximal alternating linearized minimization algorithm (iPALM) have demonstrated their superiority with respect to computation time over their non inertial variants. In many problems in imaging and machine learning, the objective functions have a special form involving huge data which encourage the application of stochastic algorithms. While the stochastic gradient descent algorithm is still used in the majority of applications, recently also stochastic algorithms for minimizing nonsmooth and nonconvex functions were proposed. In this paper, we derive an inertial variant of the SPRING algorithm, called iSPRING, and prove linear convergence of the algorithm under certain assumptions. Numerical experiments show that our new algorithm performs better than SPRING or its deterministic counterparts, although the improvement for the inertial stochastic approach is not as large as those for the inertial deterministic one. The second aim of the paper is to demonstrate that (inertial) PALM both in the deterministic and stochastic form can be used for learning the parameters of Student-$t$ mixture models. We prove that the objective function of such models fulfills all convergence assumptions of the algorithms and demonstrate their performance by numerical examples.
Article
The expected patch log-likelihood (EPLL) model is a patch prior-based image restoration method which received extensive attention in image processing in recent years for its outstanding ability to preserve the detail and structure. However, due to using the Gaussian mixture model (GMM) with the noise sensitivity as the local prior, the EPLL model suffers from undesired artifact and poor robustness frequently. In this paper, to restrain the generation of artifact of EPLL model, we replace the GMM with a bounded asymmetrical Student’s-t mixture model (BASMM), which is sufficiently flexible to fit different shapes of image data, such as non-Gaussian, non-symmetric, and bounded support data. Then, the anisotropic nonlocal self-similarity (ANSS) based regularization parameters are designed to improve the robustness of the proposed model. Experimental results demonstrate the competitiveness of our proposed model compared with that of state-of-the-art methods in performance both visually and quantitatively.
Article
Accurately modeling the distributions of spectral intensities is an effective way to obtain accurate segmentation results. Existing mixture models mostly fail to establish accurate statistical models of high-resolution SAR images due to the complex statistical features of their spectral intensities. We propose a Hierarchical Gamma Mixture Model (HGaMM) based high-resolution SAR images segmentation algorithm. In the proposed algorithm, a statistical model for SAR images is built using HGaMM, which can model heavy-tailed, asymmetric, multimodal, or flat distributions with its hierarchical structure. The algorithm allows for the effective utilization of spectral intensity information required for image segmentation. The HGaMM consists of several components that serve as the first layer, and several elements under each component that serve as the second layer. The component weight is constructed using posterior probabilities of local pixels aimed at reducing noise effects. The segmentation model can then be established by the Bayesian theorem. A new Expectation-Maximization/Adding or Deleting Markov Chain Monte Carlo (EM/ADMCMC) is incorporated to implement parameter estimation and determine the optimum number of components. Several experiments were implemented on simulated and real high-resolution SAR images to evaluate the applicability and efficiency of the proposed approach. The experimental results show that the proposed HGaMM algorithm outperforms traditional algorithms and can obtain accurate results.
Article
Full-text available
Normal mixture models are being increasingly used to model the distributions of a wide variety of random phenomena and to cluster sets of continuous multivariate data. However, for a set of data containing a group or groups of observations with longer than normal tails or atypical observations, the use of normal components may unduly affect the fit of the mixture model. In this paper, we consider a more robust approach by modelling the data by a mixture of t distributions. The use of the ECM algorithm to fit this t mixture model is described and examples of its use are given in the context of clustering multivariate data in the presence of atypical observations in the form of background noise.
Article
Full-text available
Gaussian mixture models (GMMs) constitute a well-known type of probabilistic neural networks. One of their many successful applications is in image segmentation, where spatially constrained mixture models have been trained using the expectation-maximization (EM) framework. In this letter, we elaborate on this method and propose a new methodology for the M-step of the EM algorithm that is based on a novel constrained optimization formulation. Numerical experiments using simulated images illustrate the superior performance of our method in terms of the attained maximum value of the objective function and segmentation accuracy compared to previous implementations of this approach.
Article
Many image segmentation techniques are available in the literature. Some of these techniques use only the gray level histogram, some use spatial details while others use fuzzy set theoretic approaches. Most of these techniques are not suitable for noisy environments. Some works have been done using the Markov Random Field (MRF) model which is robust to noise, but is computationally involved. Neural network architectures which help to get the output in real time because of their parallel processing ability, have also been used for segmentation and they work fine even when the noise level is very high. The literature on color image segmentation is not that rich as it is for gray tone images. This paper critically reviews and summarizes some of these techniques. Attempts have been made to cover both fuzzy and non-fuzzy techniques including color image segmentation and neural network based approaches. Adequate attention is paid to segmentation of range images and magnetic resonance images. It also addresses the issue of quantitative evaluation of segmentation results.
Article
S ummary A broadly applicable algorithm for computing maximum likelihood estimates from incomplete data is presented at various levels of generality. Theory showing the monotone behaviour of the likelihood and convergence of the algorithm is derived. Many examples are sketched, including missing value situations, applications to grouped, censored or truncated data, finite mixture models, variance component estimation, hyperparameter estimation, iteratively reweighted least squares and factor analysis.
Article
The content of an image can be summarized by a set of homogeneous regions in an appropriate feature space. When exact shape is not important, the regions can be represented by simple "blobs". Even for similar images, the blob representation of the two images might vary in shape, position, the number of blobs, and the represented features. In addition, separate blobs in one image might correspond to a single blob in the other image and vice versa. In this paper we present the BlobEMD framework as a novel method to compute the dissimilarity of two sets of blobs while allowing for context-based adaptation of the image representation. This results in representation that represent well the original images but at the same time are best aligned with respect to the representation of the context images. We compute the blobs by using Gaussian mixture modeling and use the Earth Mover's Distance (EMD) to compute both the dissimilarity of the images and the flow matrix of the blobs between the images. The BlobEMD flow-matrix is used to find optimal correspondences between source and target image representations and to adapt the representation of the source image to that of the target image. This allows for similarity measures between images that are insensitive to the segmentation process and to dierent levels of details of the representation. We show applications of this method for content-based image retrieval, image segmentation, and matching models of heavily dithered images with models of full resolution images.
Article
A broadly applicable algorithm for computing maximum likelihood estimates from incomplete data is presented at various levels of generality. Theory showing the monotone behaviour of the likelihood and convergence of the algorithm is derived. Many examples are sketched, including missing value situations, applications to grouped, censored or truncated data, finite mixture models, variance component estimation, hyperparameter estimation, iteratively reweighted least squares and factor analysis.
Article
Finite mixture models provide a natural way of modeling continuous or discrete outcomes that are observed from populations consisting of a finite number of homogeneous subpopulations. Applications of finite mixture models are abundant in the social and behavioral sciences, biological and environmental sciences, engineering and finance. Such models have a natural representation of heterogeneity in a finite, usually small, number of latent classes, each of which may be regarded as a type. More generally, the finite mixture model can be shown to approximate any unknown distribution under suitable regularity conditions. The Stata package -fmm- implements a maximum likelihood estimator for a class of finite mixture models. In this talk, I will begin by introducing finite mixture models using a number of examples and discuss issues of estimation, testing and model selection. I will then describe estimation using fmm, calculations of predictions, marginal effects, and posterior class probabilities, and illustrate these using examples from econometrics and finance.