34th Annual International Conference of the IEEE EMBS
San Diego, California USA, 28 August - 1 September, 2012
Spectral Clustering of Shape and Probability Prior Models for
Automatic Prostate Segmentation
S. Ghose1,2 , J. Mitra1,2 , A. Oliver2 , R. Martı́2 , X. Lladó2 ,
J. Freixenet2 , J. C. Vilanova3 , J. Comet4 , D. Sidibé1 and F. Meriaudeau1
Abstract— Imaging artifacts in Transrectal Ultrasound
(TRUS) images and inter-patient variations in prostate shape
and size challenge computer-aided automatic or semi-automatic
segmentation of the prostate. In this paper, we propose to
use multiple mean parametric models derived from principal
component analysis (PCA) of shape and posterior probability
information to segment the prostate. In contrast to traditional
statistical models of shape and intensity priors, we use posterior
probability of the prostate region determined from random
forest classification to build, initialize and propagate our model.
Multiple mean models derived from spectral clustering of combined shape and appearance parameters ensure improvement in
segmentation accuracies. The proposed method achieves mean
Dice similarity coefficient (DSC) value of 0.96±0.01, with a
mean segmentation time of 0.67±0.02 seconds when validated
with 46 images from 23 datasets in a leave-one-patient-out
validation framework.
Index Terms— Prostate segmentation, random forest, statistical shape and posterior probability models, spectral clustering.
I. INTRODUCTION
Accurate prostate segmentation in TRUS may aid in
biopsy needle placement and multimodal image fusion between TRUS and magnetic resonance imaging (MRI) to
improve malignant tissue extraction during biopsy [1]. However, accurate computer aided prostate segmentation from
TRUS images is a challenging task due to low contrast of
TRUS images, speckle, and shadow artifacts, inter-patient
prostate shape, size and deformation variations and heterogeneous intensity distribution inside the prostate gland. To
deal with the prostate segmentation challenges we propose
a novel prostate segmentation method in which appearance
and spatial context based information from the training
images are used to achieve a probabilistic classification of
the prostate. Thereafter, the statistical shape and appearance
model derived from PCA of prostate shape and posterior
probabilistic values of the prostate region of the training
This work was supported by VALTEC 08-1-0039 of Generalitat de
Catalunya, Spanish Science and Innovation grant nb. TIN2011-23704, Spain
and Conseil Régional de Bourgogne, France.
1 S. Ghose, J. Mitra, D. Sidibé & F. Meriaudeau are with
Le2i CNRS-UMR 6306, Université de Bourgogne, Le Creusot,
France. soumyaghose@gmail.com, jhimli.mitra@u-bourgogne.fr, drodesire.sidibe@u-bourgogne.fr, fabrice.meriaudeau@u-bourgogne.fr
2 R. Martı́, A. Oliver, X. Lladó & J. Freixenet are with the
Computer Vision and Robotics Group, University of Girona, Girona,
Spain. marly@eia.udg.edu, aoliver@eia.udg.edu, llado@eia.udg.edu,
jordif@eia.udg.edu
3 J. C. Vilanova is with the Girona Magnetic Resonance Center, Girona,
Spain.
4 J. Comet is with the Hospital Dr. Josep Trueta, Girona, Spain.
978-1-4577-1787-1/12/$26.00 ©2012 IEEE
TRUS images propagates in a multi-resolution framework
to segment a test image. However instead of a single mean
model, multiple mean models are created using spectral clustering from the combined shape and appearance parameters
to capture larger variations. Finally, the mean model that has
the lowest fitting error with the test image provides accurate
segmentation of the prostate. The performance of our method
is compared with the traditional AAM [3] and also with our
previous works [11], [14], [13]. The key contributions of
this work are: 1) The use of random forest classification
framework to obtain a soft classification of the prostate. 2)
Use of spectral clustering on prostate shape and probability
priors to group similar prostates creating multiple mean
models.
The remaining paper is organized in the following manner.
Section II provides a description of the proposed segmentation framework, followed by the results and discussions in
Section III. Finally, the paper is concluded in Section IV
with the conclusions and future works.
II. P ROPOSED S EGMENTATION F RAMEWORK
The proposed method is developed on three major components: A) supervised learning framework of random forest to
determine posterior probability of a pixel being prostate, B)
adapting statistical models of shape and intensity priors to
incorporate the posterior probabilities of the prostate region
for training, initialization and propagation of the parametric
model and C) use of spectral clustering to group prostate
multiple mean models having similar shape and appearances.
These components of the proposed segmentation method are
described in the following subsections.
A. Random forest based probabilistic classification
In traditional AAM [3], the appearance model is built from
the intensity distribution of the prostate region. However,
depending on that ultrasound image acquisition parameters
and the machine manufacturer, the intensities of the prostate
may vary significantly leading to an inaccurate appearance
model. Therefore, to reduce inter-dataset intensity variations
and intensity variations inside the prostate region, we propose
to determine the posterior probability of the image pixels
being prostate in a supervised learning framework of random
forest and use PCA of the posterior probabilities of the
prostate region to build our appearance model. Decision trees
are discriminative classifiers which are known to suffer from
over-fitting. However, a random decision forest or random
2335
forest achieves better generalization by growing an ensemble
of many independent decision trees on a random subset of the
training data and by randomizing the features made available
at each node during training [2].
During training, to minimize the pose and intensity
variations, our datasets are rigidly aligned and the interpatient intensity variations are normalized. The data consists
of a collection of V = (X, F ), each centered at 3 × 3
neighborhood of pixels. Where, X = (x, y) denotes the pixel
position and the feature vector F constitutes of the mean and
standard deviation of the 3×3 pixel neighborhood. Each tree
τi in random forest receives the full set V , along with the
label and the root node and selects a test to split V into two
subsets to maximize the information gain. A test constitutes
of a feature (like the mean of a pixel neighborhood) and
a feature response threshold. The left and the right child
nodes receive their respective subsets of V and the process
is repeated at each child node to grow the next level of
the tree. Growth is terminated when either the information
gain is minimum or the tree has grown to a maximum depth
specified. Each decision tree in the forest is unique as each
tree node selects a random subset of features and threshold.
During testing, the test image is rigidly aligned to the
same frame of the training datasets and its intensities are
normalized. The pixels are routed to one leaf in each tree by
applying the test (selected during training). Each pixel of the
test dataset is propagated through all the trees by successive
application of the relevant binary test to determine the probability of belonging to class c. When reaching a leaf node lτ ,
where τ ∈ [1..., Γ], the posterior probabilities (Pτ (c|V )) are
gathered in order to compute the final
∑Γposterior probability
of the pixel defined by P (c|V ) = Γ1 τ =1 Pτ (c|V ).
B. Statistical Shape and Appearance Model
The process of building the parametric statistical model of
shape and appearance variations involves the task of building
a shape model, an appearance model, and consecutively
a combined model of shape and appearance priors. To
build the shape model, a PDM [3] is built by equal angle
sampling of the prostate contours to determine the landmarks
automatically. The PDM of the contours are aligned to a
common reference frame by generalized Procrustes analysis
[5]. PCA of the aligned PDMs identifies the principal modes
of shape variations. Posterior probabilistic information (of
pixels being prostate) of the segmented region are warped
into correspondence using a piece-wise affine warp and are
sampled from a shape-free reference similar to that of AAM
[3]. PCA of the posterior probabilities from Section II-A
is used to identify their principal modes of variation. The
model may be formalized in the following manner. Let s
and t represent the shape and posterior probability models,
then
(1)
s = s + Φs θs , t = t + Φt θt
where s and t denote the mean shape and posterior probability information respectively, then Φs and Φt contain the
first p eigenvectors (obtained from 98% of total variations) of
the estimated joint dispersion matrix of shape and posterior
probability information and θ represent the corresponding
eigenvalues. The model of shape and posterior probability
variations are combined
in
framework]as,
[
] a linear
[
W θs
W ΦTs (s − s)
(2)
b=
=
θt
ΦTt (t − t)
where W denotes a weight factor (determined as in AAM
[3]) coupling the shape and the probability space. A third
PCA of the combined model ensures the reduction in redundancy of the combined model, and is given as,
b = Eα
(3)
where E is the matrix of eigenvectors and α the appearance
parameters.
In our model, we use the optimization framework similar
to that proposed by Cootes et al. [3]. The objective function
of our model is similar to AAM. However, instead of minimizing the sum-of-squared differences of intensities between
the mean model and target image, we minimize the sumof-squared differences of the posterior probabilities of the
mean model and the target image. The prior knowledge of the
optimization space is acquired by perturbing the combined
model with known model parameters and perturbing the pose
(translation, scale and rotation) parameters. Linear relationships between the perturbation of the combined model (δc)
and the residual posterior probability values (δt) (obtained
from the sum-of-squared differences between the posterior
probabilities of the mean model and the target image), and
between the perturbation of the pose parameters (δp) and
the residual posterior probability values are acquired in
multivariate regression frameworks as,
δc = Rc δt, δp = Rp δt
(4)
Rc and Rp refer to the correlation coefficients. Given a test
image, the posterior probability values of the pixels being
prostate is determined with random forest soft classification.
The sum-of-squared differences of the posterior probability
values with the mean model is used to determine the residual
value δt. The combined model (δc) and the pose parameters
(δp) are then updated using Eq. (4) to generate a new shape,
and combined model and consequently the new posterior
probabilities. The process continues in an iterative manner
until the difference of the mean model with the target image
remains unchanged.
C. Spectral clustering and multiple mean models
Traditionally a statistical shape and appearance model
assumes the shape and the appearance spaces to be Gaussian. However in real cases, inter-patient prostate intensity
distribution, prostate shapes and sizes may vary significantly.
In such circumstances, a single mean model is inefficient to
capture the variations of shape and appearance spaces and
it would be inappropriate to approximate them with a single
Gaussian distribution. To address this problem, we propose
to use multiple mean models. Therefore, the objective is
to employ a grouping scheme that considers the prostates
with similar shape and appearance parameters to build a
set of mean models and produce an accurate optimization
2336
space as described in Section II-B. Furthermore, the number
of mean models should change dynamically depending on
combined shape and appearance parameters of the training
datasets. To address both the issues, we propose to use
spectral clustering in the combined space of shape and
appearance parameters. Spectral clustering relies on the eigen
structure of a similarity matrix to partition points into disjoint
clusters with high intra-cluster similarity and also high intercluster dissimilarity. Moreover, the number of clusters could
be determined dynamically from the principal components
in the eigen space. During training, the combined shape
and appearance eigenvectors are obtained by Eq. (2) and
Eq. (3). Cosine similarities of the combined vectors (E) of
P eigenvectors are used to construct a P × P similarity
or affinity matrix W that measures the similarity between
the P points. Wi,j is large when the points indexed by i
and j are likely to be in the same cluster. The problem
may be defined in terms of a complete graph with vertices
υ = 1, ..., P and an affinity matrix with weights Wi,j , for
i, j ∈ υ. The objective is to∪determine K disjoint clusters
A = (Ak )k∈1,....,K , where k Ak = υ, that optimizes the
cost function of K−way
normalized
cut
defined as,
K
∑
∑
∑
Wi,j
Wi,j /
C(A, W) =
k=1
i∈Ak ,j∈υ\Ak
i∈Ak ,j∈υ
(5)
The K−way normalized cut C(A, W) may be simplified as
D−1/2 WD−1/2 where D = diag(W) [15]. The algorithm
may be summarized as,
1) Input:Similarity matrix W ∈ RP×P
2) Compute first k eigenvectors of D−1/2 WD−1/2 where
D = diag(W) to build the matrix U .
3) Re-normalize the matrix U .
4) Perform k-means clustering on the normalized U .
5) Output: Similar disjoint clusters k.
The number of clusters k is determined by the number
of the largest k-eigen vectors (obtained from 98% of total
variations) of the normalized Laplacian C(A, W) of the
affinity matrix. The k-means clustering therefore groups the
similar prostates to form k mean models.
III. E XPERIMENTAL R ESULTS AND D ISCUSSIONS
We have validated the accuracy and robustness of our
method with 46 axial mid gland TRUS images of the prostate
with a resolution of 348×237 pixels from 23 prostate datasets
in a leave-one-patient-out evaluation strategy. Manual segmentations performed by an expert radiologist were validated
by an experienced urologist to prepare the ground truth.
For the random forest based classification, we have fixed
the number of trees to 100, the tree depth to 30 and the
lower bound of information gain to 10−7 . These parameters
were chosen empirically as they produced promising results
with the test images. During validation, the test dataset is
removed and multiple mean models are built with the aid
of spectral clustering. The number of mean models varied
between 5 to 7 depending on the training datasets. Given an
TABLE I
P ROSTATE SEGMENTATION QUANTITATIVE COMPARISON (HD, AND
MAD ARE IN MM .)
Method
DSC
HD
MAD
AAM [3]
0.94±0.03
4.92±0.96
2.15±0.94
Ghose [14]
0.95±0.06
2.53±1.94
0.87±1.23
RF-AAM
0.95±0.06
3.76±2.03
1.40±0.91
Our Method
0.96±0.01
2.51±0.93
0.84±0.31
(a)
(b)
Fig. 1. Improvement in segmentation accuracies with spectral clustering
and multiple mean models. (a) shows segmentation without multiple mean
models and (b) shows segmentation with multiple mean models.
image of the test dataset, all the mean models are applied
for segmentation in parallel. Fitting or registration errors
between the mean models after final deformation and the test
image are computed by normalizing the pixel-wise intensity
differences. Finally, the segmentation by the mean model
that provided the least fitting error is selected as optimum
segmentation. We have used most of the popular prostate
segmentation evaluation metrics like DSC, 95% Hausdorff
Distance (HD), and mean absolute distance (MAD) to evaluate our method. Furthermore, the results are compared with
the traditional AAM proposed by Cootes et al. [3] and to
our previous work [14]. It is observed from Table I that
a probabilistic representation of the prostate appearance in
TRUS images significantly improves segmentation accuracy
when compared to traditional AAM. We achieved a statistically significant improvement in t-test p-value<0.0001 for
DSC, HD and MAD compared to traditional AAM [3]. As
observed in Table I that [14] and RF-AAM (that uses posterior probalities from random forest classification) provide
better results compared to AAM justifying the use of posterior probability of the prostate region instead of intensity.
However, our proposed method provides excellent results
when compared to [14] and RF-AAM, suggesting that the use
of both posterior probability and multiple mean models are
essential to improve segmentation accuracies as illustrated in
Fig. 1. In Fig. 2 we illustrate the necessity of a deformable
model based segmentation after random forest classification.
We observe that random forest classification produces misclassified regions as shown in Fig. 2(a). Our statistical shape
and probability prior models working on initial segmentation
improve segmentation accuracies as observed in Fig. 2(c).
To provide qualitative results of our method we present a
subset of results in Fig. 3. The first row shows the results
achieved with AAM [3] and the second row shows the
results achieved with our method. A quantitative comparison
of different prostate segmentation methodologies is difficult
in absence of a public dataset and standardized evaluation
metrics. Nevertheless, to have an overall qualitative estimate
of the functioning of our method we have compared with
some of the existing works in Table II. Analyzing the results
we observe that our mean DSC value is better compared
2337
TABLE II
(a)
(b)
Q UALITATIVE COMPARISON OF PROSTATE SEGMENTATION . T IME IS
MEASURED IN SECS , PXS = PIXELS , MM = MILLIMETER
Reference
Area Acc.
Contour Acc.
Datasets
Time
Betrouni
Overlap
Dist-3.77±1.3 pxs
10 images
5
[8]
93±0.9%
Shen [10]
Error
Dist-3.2±0.87 pxs
8 images
64
3.98±0.97%
Ladak [9]
Accuracy
MAD-4.4±1.8 pxs
117 images
90.1±3.2%
Cosio [12]
MAD-1.65±0.67
22 images
660
mm
Yan [1]
MAD-2.10±1.02
19 datasets/
0.3
mm
301 images
Ghose
DSC
MAD-1.26±0.51
6 datasets/
1.50
[13]
0.95±0.02
mm
24 images
Ghose
DSC
MAD-1.82±0.76
23 datasets/
5.97
[11]
0.97±0.006 pxs
46 images
Our
DSC
MAD-3.00±1.22
23 datasets/
0.67
Method
0.96±0.01
pxs/0.84±0.34 mm
46 images
(c)
Fig. 2. Illustration of the requirement of shape and probability prior model.
(a) is the images to be segmented, (b) shows the random forest classification
with mis-classified regions and (c) shows the final segmentation achieved.
Fig. 3. The green contour gives the ground truth and the red contour gives
the obtained result. Row 1 show the results achieved with AAM and row 2
show results of our model for the corresponding prostates.
to area overlap accuracy values of [8], [9], [13], and very
close to the area overlap error value of [10]. However, it
is to be noted that we have validated with more images
compared to [10]. Our MAD value is better compared to [8],
[10], [9] and [13]. From these observations we may conclude
that qualitatively our method performs well in overlap and
contour accuracy measures. Our method is implemented in
Matlab 7 on an Intel Core i5, 2.8 GHz processor and 8
GB RAM. The mean segmentation time of the method is
0.67±0.02 seconds with an unoptimized Matlab code. Even
with an unoptimized Matlab code in Table II we observe
that our mean segmentation time is less when compared to
[8], [10], [12], [13], [11] and only more compared to [1].
However, [1] used an optimized C++ code to achieve their
results.
Comparing our proposed model with our previous work
[11] we observe that we achieve similar contour and area
overlap accuracies. However, our proposed method achieves
prostate segmentation in significantly less time compared to
[11]. The mean segmentation time of [11] with current machine configuration is 4.33±0.21 seconds while the proposed
method using supervised learning with random forest takes
0.67±0.02 seconds without compromising the segmentation
accuracies. Mean segmentation time is often a critical element in selecting one segmentation method over the other for
near real-time multimodal image fusion between TRUS and
MRI to improve malignant tissue sampling during biopsy. In
this context, we may claim that our present method shows
improvement over our previous work in [11]. In contrast to
[11] where the number of mean models was dependent on a
user defined threshold of fitting error, our present work uses
spectral clustering to determine the number of mean models.
IV. C ONCLUSIONS AND F UTURE W ORKS
A novel approach of multiple statistical models of shape
and posterior probability information of prostate region with
the goal of segmenting the prostate in 2D TRUS images
has been proposed. Our approach is accurate, and robust
to significant shape, size and contrast variations in TRUS
images when compared to traditional AAM. While the proposed method is validated with prostate mid gland images,
effectiveness of the method on the base and the apical slices
is yet to be validated.
R EFERENCES
[1] P. Yan et al., “Discrete Deformable Model Guided by Partial Active
Shape Model for TRUS Image Segmentation,” IEEE Trans. on
Biomed. Engg., vol. 57, pp. 1158–1166, 2010.
[2] E. Geremia et al., “Spatial decision forests for MS lesion segmentation
in multi-channel MR images,” in MICCAI, 2010, vol. 6361, pp. 111–
118.
[3] T.F. Cootes et al., “Active Appearance Models,” in Proc. of Europ.
Conf. on Comp. Vision, 1998, pp. 484–498.
[4] T.F. Cootes et al., “The Use of Active Shape Model for Locating
Structures in Medical Images,” Imag. & Vis. Comp., vol. 12, pp. 355–
366, 1994.
[5] J.C. Gower, “Generalized Procrustes Analysis,” Psychometrika, vol.
40, pp. 33–51, 1975.
[6] H. Liu et al., “Automatic Segmentation of Prostate Boundaries in
Transrectal Ultrasound (TRUS) Imaging,” in Proc. of the SPIE Med.
Imag., 2002, pp. 412–423.
[7] K. Diaz and B. Castaneda, “Semi-automated Segmentation of the
Prostate Gland Boundary in Ultrasound Images Using a Machine
Learning Approach,” in Proc. of SPIE Med. Imag., 2008, vol. 69144,
pp. 69144A-1–8.
[8] N. Betrouni et al., “Segmentation of Abdominal Ultrasound Images of
the Prostate Using A priori Information and an Adapted Noise Filter,”
Comp. Med. Imag. & Graph., vol. 29, pp. 43–51, 2005.
[9] H.M. Ladak et al., “Prostate Segmentation from 2D Ultrasound
Images,” in Proc. of IEEE EMBS, 2000, pp. 3188–3191.
[10] D. Shen et al., “Segmentation of Prostate Boundaries from Ultrasound
Images Using Statistical Shape Model,” IEEE Trans. on Med. Imag.,
vol. 22, pp. 539–551, 2003.
[11] S. Ghose et al., “Multiple Mean Models of Statistical Shape and
Probability Priors for Automatic Prostate Segmentation,” in MICCAI
- Prostate Cancer Imaging, 2011, vol. LNCS 6963, pp. 35–46.
[12] F.A. Cosı́o, “Automatic Initialization of an Active Shape Model of the
Prostate,” Med. Imag. Anal., vol. 12, pp. 469–483, 2008.
[13] S. Ghose et al., “Statistical shape and texture model of quadrature
phase information for prostate segmentation,” IJCARS, vol. 7, pp.
43–55, 2012.
[14] S. Ghose et al., “A probabilistic framework for automatic prostate
segmentation with a statistical model of shape and appearance,” IEEE
ICIP, pp. 713–716, 2011.
[15] F. Bach et al., “Learning Spectral Clustering,” in NIPS, 2003.
2338