Vous êtes sur la page 1sur 15

1

Optimal Feature Set for Automatic Detection


and Classification of Underwater Objects in
SAS Images
Raquel Fandos∗ and Abdelhak M. Zoubir
Signal Processing Group, Institute of Telecommunications
Technische Universität Darmstadt
Merckstr. 25, 64283 Darmstadt, Germany
Phone Number: +49 6151-16 70804, Fax Number: +49 6151-16 3778
E-mail: {rfandos, zoubir}@spg.tu-darmstadt.de

Abstract—The problem of automatic detection and clas- complete CAD/CAC system that has been tested with
sification for mine hunting applications is addressed. We a 10,000 square meter database of real SAS images.
propose a set of algorithms which are tested using a large Two main approaches to the problem exist: template
database of real Synthetic Aperture Sonar (SAS) images.
The highlights and shadows of the objects in an SAS fitting and feature description. The former typically
image are segmented using both a Markovian algorithm applies a set of allowed transformations to a given
and the Active Contours algorithm. The comparison template, so that it matches the region of interest. The
of both segmentation results is used as a feature for distance between the deformed template and the object
classification. In addition, other features are considered. is measured in some norm and the object is classified
These include geometrical shape descriptors, not only
of the shadow region, but also of the object highlight, accordingly. Template fitting for sidescan sonar images
which demonstrates a significant improvement of the has been considered in [5]–[8], and in [9]–[11] for SAS
performance. Furthermore, a novel set of features based applications.
on the image statistics is described. Finally, we propose The feature description approach, adopted in this
an optimal feature set that leads to the best classification work, requires the segmentation of the regions of
results for the available database.
interest. That is, the highlight and the shadow of the
object under consideration must be extracted. Then, a
set of significant features are computed and the object is
I. I NTRODUCTION
classified according to the comparison of these features
Both sidescan sonar and Synthetic Aperture Sonar with those of a training set. Sidescan sonar image
(SAS) technologies provide high resolution imagery detection and classification based on feature extraction
for mine-countermeasure applications. While sidescan schemes have been considered in several studies with
sonar has been well developed for several decades, SAS successful results [12]–[16]. So far, the focus has been
is a fairly new technology [1]. Sidescan sonar systems on descriptors of the shadow shape [17]–[19]. In [6],
use antenna arrays to scan the seabed in a narrow beam [20], Fourier descriptors are considered and in [18]
fashion. The reconstructed images present a resolution normalized central moments are used. In this work, we
that is not only limited by the antenna aperture but that have tested several kinds of shape descriptors, not only
also decreases with range. SAS technology overcomes for the shadow but also for the object highlight. While
these limitations: a single antenna system moves along some of them are well-known shape descriptors from
a straight line and pings are sent at different posi- the literature, to the best of our knowledge, this is the
tions. Hence, a synthetic array, theoretically as long first time that they are tested with SAS images.
as desired, is built up and the resolution increases. An alternative approach is to focus on the statistical
Furthermore, it can be demonstrated that the resolution properties of an image. In [21], the mean and vari-
is independent of the range [2]. ance of different regions are considered. The kurtosis
Traditionally, human operators perform the identifi- and skewness are used for detection and classification
cation and classification of objects present on a seabed. purposes in [22], [23]. Also the difference of SNR
However, there is a growing demand for Computer between the different regions has been considered [3].
Aided Detection and Computer Aided Classification In this paper, we propose a novel set of statistical
(CAD/CAC) operations, which were introduced for features based on a Weibull parametric model of the
sidesonar applications in [3] and are also an active field SAS images.
of research in Synthetic Aperture Radar [4]. Due to its This paper is organized as follows. In Sec. II, the
strategic relevance for military applications, very few segmentation of the regions of interest is considered.
studies with real sidescan data exist in the literature, Two algorithms, a Markov Random Field (MRF) model
and even less with SAS images. This paper presents a based algorithm and an Active Contours (AC) algo-
2 Segmentation Feature Extraction Classification
AC Shadow Shape
MRF & AC
Fusion
Features
AC MRF Shadow Shape Shape
SAS Image Description
Statistical
Class
Classifier
MRF Highlight Shape Shape
MRF Description

Segmentation Result
Statistics
Description

Fig. 1. Scheme of the CAD/CAC system for SAS images proposed in this paper. The sonar image is segmented by two algorithms, an
MRF model based algorithm and the AC algorithm. The latter uses the result of the former for its initialization. Several sets of features
are extracted: both statistical and geometrical features for both the shadow and the highlight of the object. Moreover, the comparison of
the shadow shape estimated by both segmentation algorithms is exploited. Each set of features is tested independently with an statistical
classifier. However, it will be shown that an optimal feature set, consisting of some descriptors from each kind, provides the best classification
results.

SAS Sonar Image Label Field x̂


1 1
Crossrange (meters)

Crossrange (meters)
1.5 1.5

2 2

2.5 2.5

3 3
1 2 3 4 5 1 2 3 4 5
Range (meters) Range (meters)

Fig. 2. SAS sonar image of a cylindrical object. The pixels of lower Fig. 3. Estimated label field x̂ associated with the sonar image in
intensity correspond to the shadow of the object, while the object Fig. 2. The label for the background pixels B is depicted in grey,
highlight presents an intensity that is higher than the background. the shadow label S in black and the highlight label H in white.

rithm are described and results are shown. A statistical for more complex environments such as sand ripples.
classifier is proposed in Sec. III. Different shape and To cope with object detection that is robust to such
statistical descriptors for both the shadow and the conditions, we consider two different segmentation
highlight regions are presented and tested for the avail- approaches. A well-known algorithm based on an MRF
able data in Secs. IV-A to IV-E. Besides, we propose model [27] and a second segmentation algorithm, the
an original descriptor based on a comparison of the Active Contours [28], have been used. Both approaches
segmentation results provided by both segmentation are modified for the application at hand and tested using
algorithms. An optimal feature vector is presented, a database of real SAS images.
before we provide conclusions and an outlook for some
future work. A scheme of the CAD/CAC system for A. The Markovian Segmentation Algorithm
SAS applications that we propose is depicted in Fig. 1.
Assuming an MRF model for image segmentation
is a popular approach for many applications, such as
II. S EGMENTATION OF S ONAR I MAGES Synthetic Aperture Radar. It has also been used in
Segmentation is the process of partitioning an image the context of seabed reconstruction from raw sonar
into several regions [24]. In our application, three data [29], [30]. MRF models for sidescan sonar image
different regions occur (see Fig. 2): the highlights H segmentation were introduced in [31], [32] and later
of the objects in the scene, their shadows S and the used in [33], [34], where they were combined with
seabed or background B. The objects cannot only be the Iterative Conditional Estimation (ICE) algorithm to
mines, but also physical features of the terrain such as achieve an unsupervised implementation. To the best
rocks or sand ripples. In this paper, we consider any knowledge of the authors, this is the first time that MRF
element that produces a shadow as being an object. based segmentation is applied to SAS images.
Several approaches have been carried out to segment Let us express the sonar image Y (u, v), u =
sonar images automatically. Some of them are based on 1, · · · , Umax , v = 1, · · · , Vmax , associated with the
simple clustering techniques such as histogram thresh- lattice L, as the vector y = {yi , i ∈ L}, where
olding [18] or fuzzy K-means [25], [26]. These models yi denotes the intensity of pixel i. The label field
perform well for flat seabeds with high SNR, but fail x = {xi , i ∈ L} is the ‘ground truth’ that we want to
1
ηi,3 1
ηi,2 1
ηi,4 0.035 3
S hist
1 2 S pdf
ηi,1 i ηi,1 0.03 B hist
2 B pdf
2 2 ηi,3
ηi,4 ηi,2 0.025 H hist

Histograms / pdf
H pdf

0.02
Fig. 4. Second order neighborhood system and associated clique
type 1, 2, 3 and 4. 0.015

0.01

0.005
recover, where each xi has one of the possible assigned
labels {S, B, H}. The MRF model consists of the two 0
0 50 100 150
Intensity
fields (y, x). According to Bayes theorem, the posterior
probability of the label field x given the sonar image Fig. 5. Histograms and estimated Weibull pdfs for the different
y corresponds to the expression regions of the SAS image in Fig. 2 as segmented in Fig. 3. The
Weibull parametric model is appropriate for SAS images.
P (x|y) = P (x) · P (y|x), (1)
where P (x) is the Markovian a priori probability
and P (y|x) is the likelihood function of the image.
Both are described in the sequel (Secs. II-A1 and
II-A2). The ICE algorithm, employed to estimate the energy function that corresponds to the expression
parameters that define P (x) and P (y|x), is described
 
β1
in Sec. II-A4.  β2 
The optimal estimate for x is the one that corre- U (xi ) = ΘTi Ωx = (Θi,1 , Θi,2 , Θi,3 , Θi,4 ) · 
 β3  .

sponds to the global maximum of Eq. (1). Unfortu- β4
nately, its calculation is a computationally unaffordable (3)
task. In Sec. II-A5, the Iterated Conditional Modes Each βz describes the neighborhood relations of a pixel
(ICM) [27], which converges to a local maximum of with its neighbors of clique type z, and
Eq. (1), is proposed as a reasonable and fast approach. 1 2
Θi,z = 2 − δ[xi − x(ηi,z )] − δ[xi − x(ηi,z )], (4)
For the sonar image y in Fig. 2, the estimate of x, x̂,
is depicted in Fig. 3. where z = 1, · · · , 4, δ(·) is Kronecker’s delta and
1) The Markovian Probability P (x): Given a pixel 1 2
x(ηi,z ) and x(ηi,z ) refer to the labels of the neighbor
i, its neighborhood Mi is formed by a set of pixels 1 2
pixels ηi,z and ηi,z , respectively. While Θi can be
such that i ∈ / Mi , and ∀i′ ∈ Mi , i ∈ Mi′ . computed directly from x̂, the parameter vector Ωx
Each pair {i, i′ } | i′ ∈ Mi is known as a clique. needs to be estimated. Implementation details for the
Several neighborhood systems are commonly used in estimation of Ωx from (y, x̂) can be found in [37].
image modeling [24]. For this application the second
order neighborhood system is chosen. Fig. 4 shows Including a priori knowledge about the topology
such a neighborhood configuration and its associated of the sonar images into the algorithm, improves its
cliques. The first clique type relates the pixel i with performance. First of all, highlight regions are always
1 2 located to the left of shadow regions. Secondly, the
the neighbors to its right and left (ηi,1 and ηi,1 ), the
second one models its dependency with the neighbors typical size of expected highlight regions is known for
1 2 a given range. Both effects can be taken into account
above and below (ηi,2 and ηi,2 ), the third and forth
1 in the Gibbs energy of Eq. (3) [38].
one relates it with the pixels in the first (ηi,3 and
2 1 2
ηi,3 ) and second (ηi,4 and ηi,4 ) diagonals, respectively.
Hence, the neighborhood of pixel i is defined by 2) The Likelihood Function P (y|x): Assuming that
1 2 1 2 1 2 1 2
Mi = {ηi,1 , ηi,1 , ηi,2 , ηi,2 , ηi,3 , ηi,3 , ηi,4 , ηi,4 }. yi are conditionally independent, the likelihood func-
According to the Hammersley-Clifford theorem [35], tion for the image y is given by
[36], there is a one-to-one equivalence between MRF Y
and the so-called Gibbs Random Fields, which have an P (y|x) = P (yi |xi ) =
associated Gibbs distribution. This kind of representa- i∈L
Y Y Y
tion is very convenient for modeling the a priori proba- PB (yi )· PS (yi )· PH (yi ),
bility P (xi ), that is, the dependency of a pixel label xi {yi |xi =B} {yi |xi =S} {yi |xi =H}
with the labels of its neighbors Mi . A random field x (5)
has a Gibbs distribution with respect to a neighborhood
system M if and only if its joint distribution can be where PB , PS and PH are the probability density func-
expressed as tions (pdf) of the corresponding regions. As proposed in
[38], [39], a suitable parametric model for SAS images
1 −U(xi ) is the Weibull distribution W(λ, ξ), where λ and ξ
e
P (xi ) = , (2)
Z are the scale and shape parameters respectively. The
where Z is a normalizing constant and U (xi ) is an parameters ξS and λS are estimated maximizing the
4 Initial Segmentation [0]
1
means of the ICE algorithm [43]. After initializing Ωy
[0]
and Ωx , the algorithm performs the following steps at
each iteration l:
Crossrange (meters)
1.5 [l]
• Use Ωx to calculate the Markovian probability
P (xi ) for the three labels: P (xi = S), P (xi = B)
2
and P (xi = H).
[l]
• Use Ωy to calculate the likelihood functions for
2.5
each pixel: P (yi |xi = S), P (yi |xi = B) and
P (yi |xi = H).
3 • Calculate the a posteriori probability for each
1 2 3 4 5
Range (meters) label:
Fig. 6. Initial segmentation of the sonar image in Fig. 2. The label P (xi |yi ) = P (xi )·P (yi |xi ), xi = S, B, H (7)
for the background pixels B is depicted in grey, the shadow label S
in black and the highlight label H in white. • Use the Gibbs sampling algorithm to obtain n
samples of the label field, x̂(1) , · · · , x̂(n) , accord-
[l] [l]
likelihood function [40] ing to P (x|y) and using Ωy and Ωx .
• For each sample x̂(m) , estimate the parameter
Y ξ  yi ξ−1  y ξ
i
(ξS , λS ) = arg max exp − , vectors Ω̂x (x(m) ) and Ω̂y (x(m) ), 1 ≤ m ≤ n (as
ξ,λ λ λ λ described in Secs. II-A1 and II-A2, respectively).
{yi |xi =S}
(6) [l+1]
• Calculate Ωx as
and analogously for B and H, which yields the pa- n
rameter vector Ωy = (λS , ξS , λB , ξB , λH , ξH ). Fig. 5 1 X
Ω[l+1]
x = Ω̂x (x(m) ) (8)
shows the estimated pdfs of the SAS image in Fig. 2, n m=1
as segmented in Fig. 3. For comparison, the histograms [l+1]
are depicted. and analogously, calculate Ωy from Ω̂y (x(m) ),
3) Initialization: The ICE algorithm requires an 1 ≤ m ≤ n.
initialization of the label field x̂ so that initial values In our implementation, n = 1 has been chosen (see
[0] [0] [38] for details).
for Ωy and Ωx , Ωy and Ωx , can be estimated.
The K-means algorithm [41] with K = 3 has been 5) Segmentation. The ICM Algorithm: In order to
tested, leading to poor results due to the high variability estimate x̂, a local maximum of the a posteriori prob-
of the background region. Instead, we propose the ability P (x|y) is found by the ICM algorithm. First,
following approach [42], which has shown successful P (xi |yi ) is calculated for each pixel and each label
results. A non-overlapping window is shifted along the using the Ωy and Ωx estimates. Then, each pixel is
image y and the mean mw of the pixel intensities is considered in turn and is assigned a label according to
calculated for each position, which results in the vector the Maximum A Posteriori (MAP) criterion:
m = (m1 , m2 , · · · , mW ). The values in m are sorted x̂i = arg max{P (xi |yi )} = arg max{P (xi )·P (yi |xi )}.
in ascending order, and the resulting vector is split into xi xi
two vectors mlow and mhigh of the same length. The K- (9)
means algorithm is then applied to divide each vector This process is repeated until convergence.
into K = 2 clusters. Hence mlow splits into m1low and
m2low , and m1high and m2high stem from mhigh . The pixels B. The Active Contours Algorithm
in m1low are labeled as shadow S. The clusters m2low If the position of the region of interest within the
and m1high , that is, those with intermediate intensity, image is known, the AC segmentation algorithm can
are merged together and assigned to the background B. be applied. Typically used for medical imaging appli-
Finally, the elements that belong to the cluster with the cations [44], the AC algorithm has been successfully
highest intensity, m2high , are assigned to the highlight applied to sidescan sonar image segmentation [34],
H. [45]. We suggest its application to SAS images in
In general, the smaller the window size, the more this contribution. Unlike the Markovian segmentation
accurate is the initial segmentation result. However, approach, the AC algorithm does not assume any a
it also encourages the appearance of small negligible priori probability of the regions, only the intensity
groups of dark pixels being classified as S, or small values of the pixels influence the segmentation result.
groups of light pixels being classified as highlight. An active contour (or statistical snake) n =
For the available data, a window of 6 × 6 pixels {n1 , n2 , · · · , nN }T , is a closed curve defined by a list
has demonstrated to produce good results. The initial of N nodes that has the ability to evolve in order to
segmentation of the sonar image in Fig. 2 is shown in match the contour of an object present in an image
Fig. 6. [28]. Each node, expressed in cartesian coordinates
4) Parameter Estimation. The ICE Algorithm: The nj = (uj , vj )T , corresponds to a pixel i in L. The
parameters Ωy and Ωx are estimated iteratively by image of interest is therefore divided into two regions,
Initial Segmentation Final Segmentation 5
1 1

Crossrange (meters)

Crossrange (meters)
1.5 1.5

2 2

2.5 2.5

3 3
1 2 3 4 5 1 2 3 4 5
Range (meters) Range (meters)

Fig. 7. Initialization of the AC algorithm. The result of the MRF Fig. 8. Final AC segmentation result. A Rayleigh parametric model
segmentation is used. has been adopted.

and analogously for αB .


the target region T inside n and the background
2) Initialization: The initialization of the AC algo-
region B outside n. For segmentation of sonar images,
rithm is crucial. If n is initialized without comprising at
the former corresponds to either the shadow or the
least a piece of a target, the contour will diverge. Given
highlight of an object.
that the Markovian segmentation result is available,
1) The Cost Function: The objective of the AC it can be used to initialize the AC. However, if the
algorithm is to deform n in such a way that a given AC initialization is identical to the Markov based final
cost function F (n) is minimized. Originally, the AC segmentation, the AC final result in general is not
algorithm was based on the gradient so that F (n) is significantly different from it. On the other hand, the
minimum when n coincides with an edge between two purpose of using two segmentation algorithms is their
regions. However, this approach performs poorly for ability to produce different results when the borders
noisy images and is very sensitive to initialization. between regions are not clear (see Sec. IV-F). Hence,
Another approach to the problem consists of using the initialization of n is based on the Markovian result
parametric shape templates [46], [47]. Such templates but is, at the same time, significantly different: a rect-
are in general too stiff for an application like SAS, angle centered in the center of mass of the Markovian
where the target regions present a broad variability. segmented region has been adopted (see Fig. 7).
A polygonal active contour that minimizes a cost 3) The Algorithm: The number of times that nodes
function based on the likelihood function of the image are added to the initial 4-node contour n[0] is counted
is considered here [48]. Assuming independence among by the index l. When the optimal position for n[l] is
the pixels, the cost function associated with a position achieved, n[l+1] is initialized: if two consecutive nodes
of the active contour n is defined as the negative log- are further apart than a certain distance dn , Nn nodes
likelihood function of y are inserted between them. Good results have been
 Y Y  found for dn = 6 and Nn = 1 or 2.
F (n) = − ln PB (yi ) · PT (yi ) , (10) The index k is set to 0 every time that l increases and
yi |i∈B yi |i∈T counts the number of iterations that are performed for a
where PT and PB refer to the pdfs of the target fixed amount of nodes. At each iteration, a single node
and background regions respectively and need to be nj of the current best active contour, n[l] , is shifted by
estimated. An efficient implementation of Eq. (10) is a random distance, resulting in n̂[l] . If F (n̂[l] ) < Fmin ,
possible for certain parametric models of the pdfs n̂[l] is adopted. Furthermore, the so-called crossing test
[48]. However, the Weibull distribution, which is an [48] is performed to prevent intersections between the
appropriate model for SAS images (see Sec. II-A2), different segments of n.
does not lead to this. We have carried out an extensive A common problem of the AC algorithm is that
empirical study to evaluate typical parameter settings n may converge to local minima, leading to a poor
for the target and background regions. It could be segmentation. If a number of iterations Nmax elapses
shown that a Rayleigh distribution is an acceptable after the last time that new nodes were added and no
approximation of the true pdfs and will be used here update was produced in n[l] , we might be stuck in
for the sake of an efficient AC implementation. Thus, a local minimum. Thus, an alternative active contour,
PT and PB are modelled by the Rayleigh distributions ñ[l] , is calculated by shifting all nodes in n[l] by ∆
R(αT ) and R(αB ), respectively. The parameter αT is pixels. If the final cost function F (ñ[l] ) is smaller than
estimated according to the original cost function F (n[l] ), we set n[l] := ñ[l] .
This modification of the standard AC algorithm has
Y yi  yi2 
αT = arg max exp − , (11) demonstrated to improve its performance in more than
α α2 2α2 35 % of the cases. Good results have been obtained for
{yi |i∈T }
6 2 2 1
Cylindrical Object
1
Spherical Object Natural Object
2
1.5
4 4

Crossrange (meters)
Crossrange (meters)

Crossrange (meters)
Crossrange (meters)

Crossrange (meters)
1.5 2
4
2.5
6 6
2 3 6

8 8 3.5
8
2.5 4
10 10
4.5 10

12 12 3
1 1.5 2 2.5 3 3.5 4
5
1 1.5 2 2.5 3 1 2 3 4 5
Range (meters) Range (meters) Range (meters)
14 14

16 16
2 4 6 8 10 2 4 6 8 10
Fig. 10. The first and second images represent the segmentation of
Range (meters) Range (meters)
the highlight and the shadow of a cylindrical and an spherical man
2 2
made object. The last image corresponds to a physical feature of the
terrain that has been segmented as shadow.
Crossrange (meters)

Crossrange (meters)
4 4

6 6

8 8 and Ωy parameters vary within big SAS images, this


10 10 allows a better segmentation result. Hence, the Marko-
12
2 4 6 8 10
12
2 4 6 8 10
vian algorithm is applied to each subimage and the
Range (meters) Range (meters)
resulting label fields are merged. We discard shadow
2 2
and highlight pixel clusters that are too small to belong
Crossrange (meters)

Crossrange (meters)

4 4
to an object of interest. With the rest, the following
6 6

8 8
database is created. It is known from the literature [18],
10 10
[34] that an object laying on the seabed produces a
12 12
shadow but not always a highlight. Hence, an entry
14
2 4 6 8 10
14
2 4 6 8 10
is created in the database for each shadow cluster. If
Range (meters) Range (meters)
there are highlight clusters to the left of a shadow and
Fig. 9. Markovian (first column) and AC (second column) seg- within a reasonable distance they are associated with
mentation result for different SAS images. While the images in the that shadow. Highlights that are not associated to any
two first images contain several man made objects, the last image shadow are discarded. Therefore, an object consists of
corresponds to a seabed area full of sand ripples.
a shadow and, sometimes, one or more highlights. The
AC algorithm is applied to each shadow and highlight
∆ = 1 pixel and Nmax = 200 iterations. of the database.
Hence, the implementation of the algorithm reads as In Fig. 9, some parts of the segmented images
follows: are shown. The top row contains the Markovian seg-
[0] mentation results, while the AC results are shown
• Initialize n with 4 nodes, estimate PB and PT ,
in the bottom row. The two first images correspond
and calculate Fmin := F (n[0] )
[l] to scenarios where plenty of man made objects are
• Iterate while nodes in n are sparse
present. They are all successfully segmented, together
1) k := 0
with several dark areas of the seabed. The last image
2) While k < Nmax
shows a seabed where deep sand ripples are present. In
– k := k + 1
this case, the Markovian algorithm segments as shadow
– Choose a node nj ∈ n[l] (randomly or
the dark areas of the sand ripples, and as highlight the
sequentially) and shift it by a random
lightest parts to the left of the shadow areas.
distance: n̂[l]
A total of 1215 objects were segmented. Out of
– Estimate PB and PT and calculate F (n̂[l] )
them, 1037 correspond to physical natural features of
– If F (n̂[l] ) < Fmin , n[l] := n̂[l] , Fmin :=
the terrain, while 178 are man made objects. There
F (n̂[l] )
are 115 spheres and 63 cylinders. Although the objects
3) If n[l] = n[l−1] (local minimum?), ñ[l] :=
in the available SAS database are not mines, their
n[l] + ∆,
sizes and shapes are comparable, and therefore serve
else n[l+1] := add nodes (n[l] )
our purpose. Ideal spherical objects consist of a more
The AC segmentation result of the example SAS or less circular highlight and an elongated shadow.
image is shown in Fig. 8. Cylindrical objects present an elongated highlight and
a rhomboid like shadow. While spheres are insensitive
C. Segmentation Results to orientation, the position of the cylinder with respect
Both segmentation algorithms have been applied to a to the sonar influences the shadow shape. The natural
set of real SAS images. We have processed more than objects present a great variability. A representative
10,000 square meters, from 6 different images. A broad example of each object kind is shown in Fig. 10.
variability of the seabed is present. The images have
a resolution of 2.5 cm in both range and crossrange III. S TATISTICAL C LASSIFICATION
directions. In this Section, exemplary results are shown. The object database referred in Sec. II-C has been
In order to apply the Markovian segmentation al- used to study the performance of different sets of
gorithm to SAS images of large size, they have been features to classify underwater objects as man made or
split into subimages of 200 x 200 pixels. Since the Ωx natural objects. Visual inspection and classification by
TABLE I 7
E STIMATED MEAN µ̂ AND STANDARD DEVIATION σ̂ OF THE SHADOW GEOMETRICAL FEATURES ( SEE S EC . IV-A), FOR THE S , C AND N
CLASSES .

Feature ρ A χ rS O Γ
Class µ̂ σ̂ µ̂ σ̂ µ̂ σ̂ µ̂ σ̂ µ̂ σ̂ µ̂ σ̂
N 205 112 1.3e3 1.2e3 35 12 2.1 0.6 8.5 66 0.77 0.08
S 305 125 1.8e3 0.9e3 52 18 4.2 1.8 1.6 7 0.67 0.11
C 309 88 3.1e3 1.2e3 31 9 1.9 1 2.3 49 0.82 0.07

several experts was used to generate the ground truth. If The region Rt is then determined by
the available database is too small, resampling schemes
Ψtr (f ) > Qt − Qr , t = S, C, N , r 6= t. (14)
such as the boostrap may be used [49].
Three classes are considered: spheres S, cylinders The constants Qt are chosen such that Pc is maximized
C and natural objects N . Some natural objects are for a given Pfa .
much bigger or smaller than typical man made objects The ‘leave-one-out’ technique has been adopted.
and can be directly classified as N . In our database, For each object, we estimate the parameters f̄t and
22 % of the N objects fall into this category. For the Σt using the complete training set but the object
remaining ones, a statistical three- class classifier based under consideration. Then, Ψtr (f ) for t = S, C, N ,
on the Mahalanobis distance has been implemented. r 6= t are computed. The object is classified as S if
If new classes of objects were to be detected, the ΨSC (f ) > QS − QC and ΨSN (f ) > QS − QN , and
classifier could be modified accordingly, as long as analogously for C and N .
enough samples are available in order to estimate the
features of the new classes. The features are considered IV. F EATURE D ESCRIPTION OF U NDERWATER
in Sec. IV, but we describe the classifier first. O BJECTS
Let us assume a set of M features, f = The literature offers a large number of descriptors for
(f1 , f2 , · · · , fM ), to describe each object. For sim- shape representation [24], [51]. They can be divided
plicity, it is assumed that the pdf of each feature is into two main groups: those that model the contour
Gaussian. Then, the pdf of the feature vector is given of the region of interest, and those that model the
by the multivariate Gaussian distribution [50] region itself. For SAS images, the latter are more
appropriate, since the variability of the contour is not
1 1 meaningful. Several sets of shape descriptors have
Pt (f ) = M |Σt | 2 ·
(2π) 2 been considered. Some simple geometrical descriptors,
1 normalized central moments and principal component
exp − (f − f̄t )′ Σ−1
 
t (f − f̄t ) , t = S, C, N , analysis haven been applied to the shape of shadow
2
(12) region. We have represented the highlight shape by
some geometrical features. Moreover, the statistical
where the mean vector f̄t and the covariance matrix Σt
properties of the objects have been exploited as well as
are estimated from the data.
the comparison of the segmentation results provided by
We need to divide the feature space into three the Markovian and AC algorithms. Finally, an optimal
exclusive regions, RS , RC and RN . A given object set of features is proposed.
is classified as S if its feature vector f belongs to
RS , and analogously for C and N . The probability A. Shadow Geometrical Features
of classifying a natural object as sphere or cylinder
is known as the false alarm rate Pfa . The probability of Two straightforward features are the area A and
misclassifying S or C objects are denoted by PmS and perimeter ρ of the shadow. The perimeter tends to reach
PmC . The misclassification probability Pm is defined as higher values for the S and C classes than for N (see
Pm = max{PmS , PmC }. While classifying a sphere as a Table I). The area of C objects is in average greater
cylinder or a cylinder as a sphere is a minor problem, than the S and N area. Both the perimeter and the
classifying them as natural objects is critical. Although area of the N objects present heavy tailed histograms
this distinction is not considered in the calculation of due to the variability of the natural objects.
Pm , it is shown in the confusion matrices (see Tables The ratio between the squared perimeter and the
VII and VIII). The probability of correct classification area, known as compactness χ [52], has also been con-
equals Pc = 1 − Pm . sidered. The compactness is minimum for a circle and
tends to infinity as the shape approaches a straight line.
Let us define Hence, it reaches high values for the elongated shadow
Pt (f ) of the sphere class S. There are two other features, the
Ψtr (f ) = ln t, r = S, C, N , r 6= t. (13)
Pr (f ) ratio of principal axes rS and the orientation O, which
8

0.25 0.25

Crossrange (meters)
Crossrange (meters)
0.5 0.5

0.75
0.75

1
1
1.25
1.25
1.5
1.5
0.5 1 1.5 2 2.5
0.25 0.5 0.75 1 1.25 1.5
Range (meters) Range (meters)

Fig. 11. Principal axes of the shadow of an S object. The angle Fig. 13. Difference between the orientation of the right part of the
between the major axis and the abscissa is the orientation of the highlight contour (in green) and the orientation of the left part of
region. the shadow contour (in red) of a cylinder. For C objects, this feature
takes smaller values.

TABLE II
1.25
0.25 0.25 N UMBER OF HIGHLIGHTS ASSIGNED TO AN OBJECT AS A
Crossrange (meters)
Crossrange (meters)
Crossrange (meters)

2.5
0.5 0.5 FUNCTION OF THE OBJECT CLASS , IN PERCENTAGE .
0.75
0.75

NH
3.75
1
1
5 1.25
1.25

6.25
0.5 1 1.5 2 2.5 3 3.5
1.5
0.5 1 1.5 2 0.25 0.5 0.75 1 1.25
Class 0 1 2 3 >3
Range (meters) Range (meters) Range (meters)

N 39 48 10 2 1
Fig. 12. Solidity of the shadow of a N , C and S object, respectively. S 14 77 8 1 0
Each shadow region is surrounded by the contour of its minimal
convex hull. C 1 73 20 3 0

also have distinct values for S objects. The principal B. Highlight Geometrical Features
axes of a region are defined as two line segments that While almost 40 % of the N objects lack a highlight,
cross orthogonally in the center of mass and represent only 1 % of the C objects and 14 % of the S objects
the directions with zero cross-correlation (see Fig. 11). have no highlight. Thus, the presence or absence of a
For a given region with contour n and center of mass highlight is already a valuable feature of the object. In
n̄, the covariance matrix of the contour is defined by Table II, the number of highlights associated with each
N
X object, NH , are specified.
Υ= (nj − n̄)(nj − n̄)T . (15) A prominent feature for discerning C objects is the
j=1 ratio of principal axes of the highlight, rH . While for
The ratio of principal axes, rS , equals the coefficient S and N , its mean value is around 2, it is almost 4 in
of the eigenvalues of Υ and is a good measure of the the case of C highlights. Another good feature for the
elongation of the region defined by n. For S objects, C class is the rate between the highlight and shadow
rS is significantly higher than for the C and N classes. widths along the crossrange direction, W . Again, it
The orientation is the angle that the major axis of the reaches higher values for C than for S and N objects.
shadow forms with the abscissa of the image. Its value Finally, the difference between the orientation of the
is comprised in the interval from −10o to 10o for about right part of the highlight contour and the orientation
95 % of the S objects, while it is roughly uniformly of the left part of the shadow contour, ∆O , has been
distributed between −90o and 90o for C and N objects. considered (see Fig. 13). The influence of the highlight
orientation on the orientation of the shadow is stronger
Finally, the solidity Γ is the coefficient between for C objects. The estimated mean and standard devi-
the area of the region and the minimal convex area ation of the highlight descriptors rH , W and ∆O are
that comprises it [52]. Many shadows of N objects included in Table III.
have a very low solidity. Since they have a greater Ideally, the classification should be performed ac-
area cylindrical objects are in general more solid than cording to (fH , fS ), where fH = (rH , W, ∆O ). How-
spheres. The solidity of the N shadow in Fig. 12 equals ever, fH is only defined for those objects with highlight.
0.61. It is 0.85 for the cylinder and 0.72 for the sphere. A possible solution consists of using two different
The estimated mean and standard deviation of all classifiers: one for those objects with highlight and
descriptors, ρ, A, χ, rS , O and Γ, are included in Table another one for the objects without highlight. This
I. In Fig. 14, Pc is represented as a function of Pfa for has the disadvantage that not all the elements of the
the feature set fS = (ρ, A, χ, rS , O, Γ). If Pfa = 0.2, database are used, and the performance consequently
the probability of correct classification is Pc = 0.92. decreases. The following classification procedure has
The classification of S objects as C and vice versa is been derived to overcome this problem. According to
the reason for Pc < 1 when Pfa = 1. Eq. (13), Ψtr (fS ) for t, r = S, C, N , with r 6= t
1 1
9

0.8 0.8

Pc 0.6 0.6

Pc
0.4 0.4
c00

fS c00 ,C[2]
0.2 0.2
fH c00 ,C[2,3]

fS , fH c00 ,C[2,3,4]
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
Pfa Pfa
Fig. 14. Classification performance when the shadow geometrical Fig. 15. Classification performance for the normalized central
descriptors fS , the highlight geometrical descriptors fH , and a moments. For Pfa > 0.12, the feature vector that provides the best
combination of both are used. performance is f = (c00 ,C[2] ).

TABLE III
E STIMATED MEAN µ̂ AND STANDARD DEVIATION σ̂ OF THE
HIGHLIGHT GEOMETRICAL FEATURES ( SEE S EC . IV-B), FOR THE c10 and c01 vanish. For a binary image that takes its
S , C AND N CLASSES . values from {0, 1}, c00 equals the area A of the object
Feature rH W ∆O in Y .
Class µ̂ σ̂ µ̂ σ̂ µ̂ σ̂ The normalized central moments are obtained nor-
malizing cpq with respect to the area
N 2.1 0.8 0.45 0.26 8. 9 11.6
S 1.8 0.5 0.48 0.19 10.9 14.2 cpq p+q
Cpq = κ , where κ = 1 + , (18)
C 3.9 1.2 0.75 0.18 3.8 4.9 c00 2
with p = 0, . . . , pmax , q = 0, . . . , qmax .
The moments of order 2 have a simple geometri-
are computed. In the case where the object under cal interpretation [53]. While C11 is related with the
consideration has associated highlights, Ψtr (fH ) is also covariance of the region, C20 and C02 correspond
calculated, otherwise we assume Ψtr (fH ) = 0. Assum- respectively to the length of the major and minor axes
ing independence between fS and fH , the classification of the ellipse that best fits the object in Y (u, v).
can be done according to Ψtr (fS )+Ψtr (fH ). In Fig. 14 The shadow normalized central moments Cpq up to
the performance of fH , fS and the combination of both order 4 have been tested for classification, taking c00 as
are compared. The highlight feature vector fH does an extra feature. In Table IV, the estimated mean and
not provide a good result on its own, but improves standard deviation of the moments of orders 2 and 3 are
the performance of the shadow feature vector when listed. While C11 is not a good descriptor, C02 and C20
both are combined. For Pfa = 0.2, the probabilities of take distinct values for the different classes, rendering
correct classification are Pc (fS ) = 0.92, Pc (fH ) = 0.42 C20 specially useful for S discrimination. Also C30 ,
and Pc (fS , fH ) = 0.94. C40 and C04 are good descriptors.
In the case where NH > 1, one of the highlights has Let us denote the set of normalized central moments
to be selected to compute fH . The following approach of order s by C[s] so that, for instance, C[2] =
has been adopted. For each highlight h, its area Ah (C11 , C20 , C02 ) . In Fig. 15, Pc vs. Pfa is depicted for
and the rate between its width and the shadow width different combinations of C[s] . The best performance is
along the crossrange direction, Wh , are calculated. We achieved by f = (c00 , C[2] ), with a probability of cor-
select the highlight that obtains the highest value for rect classification Pc = 0.82 at Pfa = 0.2. For Pfa ∼ 0,
the coefficient the highest Pc is obtained by f = (c00 , C[2] , C[3] , C[4] ).
Ah
θh = τ + (1 − τ )Wh , h = 1, · · · , NH . D. Principal Components Analysis
max{Ah }
(16) Principal Components Analysis (PCA) [54] is a
Good results have been obtained for τ = 0.5. popular pattern recognition technique. It has been
studied for sidescan sonar detection in [55], where
C. Normalized Central Moments PCA is applied directly to the sonar images. In the
For an image Y (u, v), u = 1, · · · , Umax , v = current approach we use it on a binary representation
1, · · · , Vmax , the sample central moment of order (p+q) of the shadow region. PCA consists of representing
is defined as a set of correlated variables by a smaller number of
XX uncorrelated ones, called principal components. In this
cpq = (u − ū)p (v − v̄)q Y (u, v). (17) application, the correlated variables are the pixels of
u v each binary image that contains the shape of a shadow.
The origin of coordinates is placed at the center of The first principal component, ζ1 , represents as much
mass of the image, (ū, v̄). Due to this normalization, of the data correlation as possible, and each consecutive
10 TABLE IV
E STIMATED MEAN µ̂ AND STANDARD DEVIATION σ̂ OF THE SAMPLE NORMALIZED CENTRAL MOMENTS OF ORDER 2 AND 3 ( SEE
S EC . IV-C). A LL QUANTITIES ARE NORMALIZED BY 1 E -3.

Feature C11 C02 C20 C12 C21 C30 C03


Class µ̂ σ̂ µ̂ σ̂ µ̂ σ̂ µ̂ σ̂ µ̂ σ̂ µ̂ σ̂ µ̂ σ̂
N 0.9 58 170 61 90 47 -0.4 12 -0.8 9 -0.1 10 0.4 19
S -0.9 48 40 24 450 176 5 12 -1.2 17 41 90 1 11
C -1.3 42 90 47 130 102 -0.5 7.6 0.3 7 1 13 -0.2 12

0.25
0.8
0.5

Crossrange (meters)
0.6 0.75
Pc

1
0.4 1.25
ζ1
1.5
ζ1 , ζ2
0.2
ζ1 , ζ2 , ζ3 1.75
ζ1 , · · · , ζ6
2
0
0 0.2 0.4 0.6 0.8 1
2.25
Pfa 0.5 1 1.5 2 2.5 3
Range (meters)
Fig. 16. Classification performance of the principal components.
The performance improves when the number of ζi increases up to Fig. 17. Estimated label field for the sonar image in Fig.2, where a
3. For illustration purposes, the curve that corresponds to the first 6 stripe around the shadow indicates the pixels used for the background
principal components is included. statistics estimation.

1
ζb accounts for as much of the remaining correlation
as possible. For implementation details see [24], [56]. 0.8
The estimated mean and standard deviation of the
first 6 ζb are included in Table V. It is ζ1 the principal 0.6
Pc

component that takes more distinct values from class


0.4
to class, and also produces better classification results:
Pc = 0.45 if Pfa = 0.2, when tested alone. These
0.2
results are however outperformed when several ζb are W params
included in f (see Fig. 16). If f = (ζ1 , ζ2 ), then Pc ∆W params
0
0 0.2 0.4 0.6 0.8 1
increases to 0.6, and to 0.65 when also ζ3 is taken into Pfa
account. A further increase of the number of principal
components has no effect on the performance. Fig. 18. Classification performance of the statistical features. The
curve in red corresponds the a feature vector formed by the Weibull
parameters f = (λS , λH , λB , ξS , ξH , ξB ), and the green line stems
E. Statistical Features from f = (∆λSB , ∆λHB , ∆λSH , ∆ξSB , ∆ξHB , ∆ξSH ).
As referred in Sec. II-A2, the different regions of an
SAS image, S, C and H, are conveniently modelled
by the Weibull distributions W(λS , ξS ), W(λC , ξC ) tends to be higher for N than for C or S objects. This
and W(λH , ξH ), respectively. The parameters for the is due to the fact that C and S objects tend to have
shadow and the highlight are estimated from the seg- darker shadows and lighter highlights than the natural
mented regions. For the estimation of the background objects N .
Weibull parameters, not the entire background region is An alternative consists of using the distance between
used, since the statistical properties might change due the parameters of the different regions as features, that
to its size. Instead, only a stripe of pixels around the is, f = (∆λSB , ∆λHB , ∆λSH , ∆ξSB , ∆ξHB , ∆ξSH ),
shadow is used. The dilation morphological operation where ∆λSB = λS − λB and equivalently for the other
is applied to the segmented shadow in order to build elements. The classification performance improves, as
this stripe. See Fig. 17 for an illustration. can be observed in Fig. 18. For Pfa = 0.2, Pc = 0.5
The estimated mean and standard deviation for the if the Weibull parameters are used as feature vector
three classes N , S and C are detailed in Table VI. and Pc = 0.62 when the distance between them is
While the shape factor ξ does not significantly differ adopted. The performance of the statistical descriptors
from class to class, the scale factor λ helps to dis- is quite poor compared with other feature sets but it
criminate classes. The value of λH for N objects is improves the performance of the classifier when they
significantly lower than for C or S objects, while λS are all combined (see Sec. IV-G).
TABLE V 11
E STIMATED MEAN µ̂ AND STANDARD DEVIATION σ̂ OF THE PRINCIPAL COMPONENTS ζb , i = 1, · · · , 6 ( SEE S EC . IV-D).

Feature ζ1 ζ2 ζ3 ζ4 ζ5 ζ6
Class µ̂ σ̂ µ̂ σ̂ µ̂ σ̂ µ̂ σ̂ µ̂ σ̂ µ̂ σ̂
N 2.8 7.2 -0.5 7 -0.16 8.2 -0.18 4.2 -0.04 4.2 0.2 3.8
S -14 8.3 0.6 9 -0.18 2.3 1.2 3.8 0.07 2.2 0.4 2
C -3.7 7.6 -2.7 12 -0.28 5.3 0.26 2.9 -0.07 3.5 -0.8 3.8

TABLE VI
E STIMATED MEAN µ̂ AND STANDARD DEVIATION σ̂ OF THE W EIBULL PARAMETERS ξS , λS , ξB , λB , ξH AND λH ( SEE S EC . IV-E).

Feature ξS λS ξB λB ξH λH
Class µ̂ σ̂ µ̂ σ̂ µ̂ σ̂ µ̂ σ̂ µ̂ σ̂ µ̂ σ̂
N 1.7 0.25 34 4.2 2.19 0.33 53 5.6 3.5 0.63 86 8.8
S 1.6 0.14 31.6 2.6 2.16 0.14 57.3 4 3.5 0.64 108 11.5
C 1.5 0.17 27.7 3.2 2.13 0.19 55.8 4.4 3.2 0.44 104 10.2

0.8
0.5
Crossrange (meters)

1 0.6

1.5 Pc 0.4 fS , fH
c00 ,C[2]
2 ζ1 , ζ2 , ζ3
0.2
∆W params
2.5 f∗
0
0 0.2 0.4 0.6 0.8 1
0.5 1 1.5 2 2.5 3 3.5
Pfa
Range (meters)
Fig. 20. Comparison of the classification performance for the
Fig. 19. Segmentation results provided by the Markovian and optimal feature vector f ∗ and other feature vectors. At a false alarm
AC algorithms for an N object. For shadow regions that are not rate of Pfa = 0.07, a probability of correct classification Pc = 0.95
well differentiated, both algorithms produce significantly different is achieved.
segmentations.

of dimensionality [57]. Furthermore, some of the de-


F. MRF & AC Rate scriptors contain redundant information. Therefore, it
The segmentation results provided by the Markovian is necessary to choose an optimal set of features, f ∗ ,
and the AC algorithms are in general quite similar (see that provides the best performance. For the available
Fig. 9). However, when no man made object is present database, the best performance has been observed for
in an image, it can happen that an area that is slightly a feature vector formed by all shape and highlight ge-
darker than the background is segmented as a shadow ometrical descriptors, the normalized central moments
region. In these cases, the AC algorithm often provides C20 and C02 , the second principal component ζ2 , the
a segmentation result that is significantly different statistical descriptors ∆λSB , ∆λHB and ∆λSH , and
from the Markovian segmentation, as illustrated by the the parameter Φ, that is,
example in Fig. 19. The rate Φ between the area of the
region where both segmentation results intersect and f∗ = (ρ, A, χ, rS , O, Γ, rH , W, ∆O , C20 , · · ·
the area of the AC segmented region constitutes a good C02 , ζ2 , ∆λSB , ∆λHB , ∆λSH , Φ). (19)
measure of the segmentation reliability.
Including other features has either no effect on Pc
The mean value of Φ is 0.97 for the C objects, 0.87
or degrades it. The comparison of f ∗ with some of the
for S and 0.73 for the N class. Indeed, while Φ is
feature vectors described above is illustrated in Fig. 20.
lower than 0.5 for more than 20 % of the N objects, a
For Pfa > 0.4, the performance of f ∗ is identical to
negligible amount of S and C objects present Φ < 0.5.
that of (fS , fH ). However, when a low probability of
The estimated standard deviation is smaller than 0.1 for
false alarm is desired, f ∗ clearly outperforms (fS , fH ).
the C and S classes and 0.16 for the N class.
Hence, for Pfa = 0.07, the feature vector f ∗ reaches a
probability of correct classification Pc = 0.95 while for
G. Optimal Feature Set (fS , fH ), Pc = 0.88. The confusion matrices for both
Using all the features described above together does f ∗ and (fS , fH ) at Pfa = 0.07 are included in Tables
not produce good results due to the well-known curse VII and VIII.
12 TABLE VII
C ONFUSION MATRIX OF f ∗ AT PFA = 0.07. T HIS TABLE algorithms are novel to this work. Furthermore, objects
ILLUSTRATES THE CLASSIFICATION RESULT FOR EACH CLASS with and without highlight are compared within the
WITH REFERENCE TO THE GROUND TRUTH .
same framework, avoiding the effective reduction of
Decision Class the training set for those objects with highlight.
Ground Truth N S C Some future work on the characterization of the
N 711 6 50 different features is foreseen. The feature robustness in
S 2 109 4 the presence of noise, navigation errors or complicated
C 1 2 60 seabeds such as sand ripples, has to be addressed.
A relaxation of the Gaussianity assumption for the
feature descriptors might improve the classification per-
TABLE VIII
C ONFUSION MATRIX OF (fS , fH ) AT PFA = 0.07. W HILE (fS , fH ) formance. Moreover, other approaches to classification
REACHES PC = 0.88, PC (f ∗ ) = 0.95. such as support vector machines [58], [59] or fuzzy
Decision Class logic [60] will also be considered. Support vector
Ground Truth N S C machines do not assume any specific distribution of
the features and therefore are more robust to situations
N 711 10 46 where the Gaussianity assumption does not hold. Fuzzy
S 10 102 3 logic techniques can be used to assess the confidence
C 3 4 56 of the classifier, which is extremely useful in real
applications.

In Fig. 21, all the C and S objects that are missclassi- ACKNOWLEDGMENT
fied by f ∗ are shown. Two out of the three misclassified
cylinders are classified as S, and four out of the six The authors would like to thank Dr.-Ing. Robert
misclassified spheres are classified as C, being still Engel, Dipl.-Ing. Ursula H”olscher, Dr. Arne Kraft,
detected as mines. Some of the most challenging C and Dr. rer. nat Konstantinos Siantidis from Atlas
and S objects that are correctly classified are shown in Elektronik GmbH for making available the data and
Fig. 22. Finally, Fig. 23 shows some of the N objects for their support and valuable comments.
that produce false alarms. The first row contains three
N objects classified as S, and the rest are classified as R EFERENCES
C.
[1] M. P. Hayes, “Synthetic aperture sonar: A review of current
status,” IEEE Journal of Oceanic Engineering, vol. 34, no. 3,
V. C ONCLUSION pp. 207 – 223, 2009.
[2] D. Massonnet and J. C. Souyris, Imaging with Synthetic Aper-
In this paper, two segmentation algorithms have been ture Radar. CRC Press, 2008.
described. The first one is based on a Markovian model [3] S. G. Johnson and A. Deaett, “The application of automated
recognition techniques to side-scan sonar imagery,” Journal of
with an optimized initialization scheme. The second is Oceanic Engineering, vol. 19, no. 1, pp. 138 – 144, 1994.
the Active Contours algorithm, whose initialization is [4] C. Debes, M. Amin, and A. Zoubir, “Target detection in
based on the Markovian segmentation result. Moreover, single- and multiple-view through-the-wall radar imaging,”
IEEE Transactions on Geoscience and Remote Sensing, vol.
we propose a novel modification of the standard AC 47(5), pp. 1349 – 1361, 2009.
algorithm that avoids the local minima and significantly [5] M. Mignotte, C. Collet, P. Pérez, and P. Bouthemy, “Hybrid
improves the performance in more than 35 % of the genetic optimization and statistical model based approach for
the classification of shadow shapes in sonar imagery,” IEEE
cases. Several kinds of descriptors for underwater mine Transactions on Pattern Analysis and Machine Intelligence,
classification have been investigated and tested with a vol. 22, no. 2, pp. 129 – 141, 2000.
statistical classifier. An optimal feature vector has been [6] R. Balasubramanian and M. Stevenson, “Pattern recognition for
proposed, obtaining a rate of 95 % correct classification underwater mine detection,” in Proceedings of the CAC/CAD
Conference, 2001.
with a false alarm rate of 7 % for a database of [7] E. Dura, J. Bell, and D. Lane, “Superellipse fitting for the
real SAS images with more than 1000 objects. This classification of mine-like shapes in side-scan sonar images,”
optimal feature vector consists of several geometrical in Proceedings of the OCEANS Conference, vol. 1, 2002, pp.
23 – 28.
shape descriptors for the shadow and the highlight [8] S. Reed, Y. Petillot, and J. Bell, “Automated approach to
shapes, two normalized central moments as well as classification of mine-like objects in sidescan sonar using
a principal component of the shadow shape, three highlight and shadow information,” IEE Proceedings - Radar,
Sonar and Navigation, vol. 151, no. 1, pp. 48 – 56, 2004.
statistical descriptors based on a Weibull parametric [9] J. Groen, E. Coiras, and D. Williams, “Detection rate statistics
model of the image, and a coefficient that accounts in Synthetic Aperture Sonar images,” in Proceedings of the
for the similarity between the segmentation results Underwater Acoustic Measurements Conference, 2009.
[10] E. Coiras and J. Groen, “3D target shape from SAS images
produced by the Markovian and the AC segmentation based on a deformable mesh,” in Proceedings of the Underwa-
algorithms. The shadow and highlight shape descriptors ter Acoustic Measurements Conference, 2009.
are well-known in the literature, but this is the first time [11] H. Midelfart, J. Groen, and O. Midtgaard, “Template matching
methods for object classification in synthetic aperture sonar
that they are tested with SAS images. The statistical images,” in Proceedings of the Underwater Acoustic Measure-
descriptors and the fusion of the two segmentation ments Conference, 2009.
S S S 13

0.25 0.25
0.5

Crossrange (meters)

Crossrange (meters)

Crossrange (meters)
0.5
0.5
0.75 1
0.75
1
1 1.5
1.25
1.25
1.5
2
1.75 1.5

1.25 2.5 3.75 0.5 1 1.5 2 2.5 3 1.25 2.5 3.75


Range (meters) Range (meters) Range (meters)

S S
S
0.25 0.25

Crossrange (meters)

Crossrange (meters)
0.5 0.5
Crossrange (meters)

0.5
0.75
1 0.75
1
1
PSfrag
1.5 1.25
1.25
1.5
2
1.5 2

2.5
0.5 1 1.5 2 2.5 3 0.5 1 1.5 2 0.5 1 1.5 2 2.5
Range (meters) Range (meters) Range (meters)

C C C

0.25 0.25
Crossrange (meters)

Crossrange (meters)

0.5

Crossrange (meters)
0.5 0.5

0.75 1
0.75
1 1.5
1
1.25
1.25 2
1.5
1.5 2.5
2
1.25 2.5 3.75 5 1.25 2.5 3.75 0.25 0.5 0.75 1 1.25 1.5
Range (meters) Range (meters) Range (meters)

Fig. 21. All objects from the C and S classes that are missclassified with the optimal feature set. The four first spheres are classified as
cylinders, and the two last ones as natural objects. The first two cylinders are classified as spheres, and the last one as a natural object.

[12] M. Doherty, J. Landowski, P. Maynard, G. Uber, D. Fries, and [23] F. Maussang, M. Rombaut, J. Chanussot, and M. Amate, “Fu-
F. Maltz, “Side scan sonar object classification algorithms,” in sion of local statistical parameters for buried underwater mine
Proceedings of the 6th International Symposium on Unmanned detection in sonar imaging,” EURASIP Journal on Advances in
Untethered Submersible Technology, 1989, pp. 417 –424. Signal Processing, vol. 2008, 2008.
[13] G. J. Dobeck, J. C. Hyland, and L. Smedley, “Automated [24] R. C. Gonzalez and R. E. Woods, Digital Image Processing.
detection and classification of sea mines in sonar imagery,” Boston, MA, USA: Addison-Wesley Longman Publishing Co.,
in Society of Photo-Optical Instrumentation Engineers (SPIE) Inc., 2001.
Conference Series, vol. 3079, no. 1, 1997, pp. 90–110. [25] S. Guillaudeux, S. Daniel, and E. Maillard, “Optimization of
[14] S. Perry and L. Guan, “Pulse-length-tolerant features and detec- a sonar image processing chain: a fuzzy rules based expert
tors for sector-scan sonar imagery,” IEEE Journal of Oceanic system approach,” in Proceedings of the OCEANS conference,
Engineering, vol. 29, no. 1, pp. 138 – 156, 2004. vol. 3, 1996, pp. 1319 – 1323.
[15] T. Aridgides, M. F. Fernandez, and G. J. Dobeck, “Side-scan [26] S. Daniel, S. Guillaudeux, and E. Maillard, “Adaptation of
sonar imagery fusion for sea mine detection and classification a partial shape recognition approach,” in IEEE International
in very shallow water,” A. C. Dubey, J. F. Harvey, J. T. Broach, Conference on Systems, Man, and Cybernetics: ‘Computational
and V. George, Eds., vol. 4394. SPIE, 2001, pp. 1123 – 1134. Cybernetics and Simulation’, vol. 3, 1997, pp. 2157 – 2162.
[16] C. Ciany and W. Zurawski, “Performance of fusion algorithms [27] J. Besag, “On the statistical analysis of dirty images,” Journal
for computer aided detection and classification of bottom mines of the Royal Statistical Society, vol. B-48, pp. 259 – 302, 1986.
in the shallow water environment,” in Proceedings of the
[28] M. Kass, A. Witkin, and D. Terzopoulos, “Snakes: Active
OCEANS Conference, vol. 4, 2002, pp. 2164 – 2167.
contour models,” International Journal of Computer Vision,
[17] A. Castellano and B. Gray, “Autonomous interpretation of
vol. 1, pp. 321 – 331, 1988.
side scan sonar returns,” in Proceedings of the Symposium on
Autonomous Underwater Vehicle Technology, 1990, pp. 248 – [29] V. Murino, A. Trucco, and C. Regazzoni, “A probabilistic
253. approach to the coupled reconstruction and restoration of
[18] I. Quidu, J. Malkasse, G. Burel, and P. Vilbe, “Mine classifi- underwater acoustic images,” IEEE Transactions on Pattern
cation using a hybrid set of descriptors,” in Proceedings of the Analysis and Machine Intelligence, vol. 20, no. 1, pp. 9 – 22,
OCEANS Conference, vol. 1, 2000, pp. 291 – 297. 1998.
[19] J. C. Delvigne, “Shadow classification using neural networks,” [30] V. Murino and A. Trucco, “Edge/region-based segmentation
in 4th Undersea Defence Conference, 1992, pp. 214 – 221. and reconstruction of underwater acoustic images by Markov
[20] D. Boulinguez and A. Quinquis, “Classification of underwater random fields,” in Proceedings of the IEEE Computer Society
objects using Fourier descriptors,” in Proceedings of the Inter- Conference on Computer Vision and Pattern Recognition, 1998,
national Conference on Image Processing and its Applications, p. 408.
1999, pp. 240 – 244. [31] C. Collet, P. Thourel, P. Pérez, and P. Bouthemy, “Hierarchical
[21] I. Tena Ruiz, D. Lane, and M. Chantler, “A comparison of MRF modeling for sonar picture segmentation,” in Proceedings
inter-frame feature measures for robust object classification in of the International Conference on Image Processing, vol. 3,
sector scan sonar image sequences,” IEEE Journal of Oceanic 1996, pp. 979 – 982.
Engineering, vol. 24, pp. 458 – 469, 1999. [32] P. Thourel, C. Collet, P. Bouthemy, and P. P., “Multiresolution
[22] F. Maussang, J. Chanussot, A. Hétet, and M. Amate, “Higher- analysis and MRF modeling applied to the segmentation of
order statistics for the detection of small objects in a noisy shadows in sonar pictures,” in 2nd Asian Conf. Computer
background application on sonar imaging,” EURASIP Journal Vision, vol. 2, 1996, pp. 81 – 85.
on Advances in Signal Processing, vol. 2007, no. 1, pp. 25 – [33] M. Mignotte, C. Collet, P. Pérez, and P. Bouthemy, “Sonar
25, 2007. image segmentation using an unsupervised hierarchical MRF
14 S S S

20 10

Crossrange (meters)
20

Crossrange (meters)

Crossrange (meters)
40
40 20

60
60 30
80

80 40
100

120 100 50
50 100 150 200 50 100 150 20 40 60 80 100
Range (meters) Range (meters) Range (meters)

S S S

10
20

Crossrange (meters)

Crossrange (meters)
20
Crossrange (meters)

20

40 40 30

40
60
60
50
80
80 60

100 70
20 40 60 80 100 120 140 50 100 150 20 40 60 80 100
Range (meters) Range (meters) Range (meters)

C C C

20 20 20
Crossrange (meters)

Crossrange (meters)

Crossrange (meters)
40
40 40

60
60 60
80
80
80
100
100
100
50 100 150 50 100 150 20 40 60 80 100 120
Range (meters) Range (meters) Range (meters)

C C C

20
20 20
Crossrange (meters)
Crossrange (meters)

Crossrange (meters)
40
40
40 60
60
80
60
80
100
80 100
120
120
100 140
20 40 60 80 100 120 50 100 150 20 40 60 80 100
Range (meters) Range (meters) Range (meters)

Fig. 22. Some of the most challenging C and S objects that are correctly classified with the optimal feature set.

model,” IEEE Transactions on Image Processing, vol. 9, no. 7, application to SAS images,” in Proceedings of the European
pp. 1216 – 1231, 2000. Conference on Underwater Acoustics, 2010.
[34] S. Reed, Y. Petillot, and J. Bell, “An automatic approach to the [43] F. Salzenstein and W. Pieczynski, “Parameter estimation in
detection and extraction of mine features in sidescan sonar,” hidden fuzzy Markov random fields and image segmentation,”
IEEE Journal of Oceanic Engineering, vol. 28, no. 1, pp. 90 – Graphical Models Image Processing, vol. 59, no. 4, pp. 205 –
105, 2003. 220, 1997.
[35] J. Besag, “Spatial interaction and statistical analysis of Lattice [44] T. Cootes, A. Hill, C. Taylor, and J. Haslam, “Use of active
systems,” Journal of the Royal Statistical Society, vol. 36, pp. shape models for locating structures in medical images,” Image
192 – 236, 1974. and Vision Computing, vol. 12, no. 6, pp. 355 – 365, 1994.
[36] F. Spitzer, “Markov random fields and Gibbs ensembles,” The [45] F. Langner, C. Knauer, W. Jans, and A. Ebert, “Side scan sonar
American Mathematical Monthly, vol. 78, no. 2, pp. 142 – 154, image resolution and automatic object detection, classification
1971. and identification,” in Proceedings of the OCEANS Conference,
2009, pp. 1 – 8.
[37] H. Derin and H. Elliott, “Modeling and segmentation of noisy
and textured images using Gibbs random fields,” IEEE Transac- [46] Y. Amit, U. Grenander, and M. Piccioni, “Structural image
tions on Pattern Analysis and Machine Intelligence, vol. PAMI- restoration through deformable templates,” Journal of the Amer-
9, no. 1, pp. 39 – 55, 1987. ican Statistical Association, vol. 86, no. 414, pp. 376 – 387,
1991.
[38] M. Mignotte, C. Collet, P. Pérez, and P. Bouthemy, “Three- [47] C. Kervrann and F. Heitz, “A hierarchical statistical frame-
class Markovian segmentation of high resolution sonar images,” work for the segmentation of deformable objects in image
Computer Vision and Image Understanding, vol. 76, no. 3, pp. sequences,” in Proceeding of the IEEE Computer Society Con-
191 – 204, 1999. ference on Computer Vision and Pattern Recognition, 1994, pp.
[39] F. Maussang, J. Chanussot, A. Hetet, and M. Amate, “Mean / 724 – 728.
standard deviation representation of sonar images for echo de- [48] C. Chesnaud, P. Refregier, and V. Boulet, “Statistical region
tection: Application to SAS images,” IEEE Journal of Oceanic snake-based segmentation adapted to different physical noise
Engineering, vol. 32, no. 4, pp. 956 – 970, 2007. models,” IEEE Transactions on Pattern Analysis and Machine
[40] A. C. Cohen, “Maximum likelihood estimation in the Weibull Intelligence, vol. 21, no. 11, pp. 1145 – 1157, 1999.
distribution based on complete and on censored samples,” [49] A. M. Zoubir and D. R. Iskander, Bootstrap Techniques for
Technometrics, vol. 7, no. 4, pp. 579 – 588, 1965. Signal Processing. Cambridge University Press, 2004.
[41] S. Banks, Signal Processing, Image Processing and Pattern [50] T. W. Anderson, An Introduction to Multivariate Statistical
Recognition. Englewood Cliffs, NJ: Prentice-Hall, 1990. Analysis, 3rd ed. Hoboken, NJ: John Wiley and Sons, Inc.,
[42] R. Fandos and A. M. Zoubir, “Enhanced initialization scheme 2003.
for a three-region markovian segmentation algorithm and its [51] Luciano and R. M. Cesar, Shape Analysis and Classification:
N N
15
N

10

Crossrange (meters)
20

Crossrange (meters)
20

Crossrange (meters)
20

40 30
40

60 40

60 50
80
60
80 100 70
20 40 60 80 20 40 60 80 100 120 140 20 40 60 80 100
Range (meters) Range (meters) Range (meters)

N N N

10 20

Crossrange (meters)
Crossrange (meters)

Crossrange (meters)
20 20 40
30 60
40 40 80

50 100
60
60 120

70 140
80
20 40 60 80 100 20 40 60 80 20 40 60 80 100
Range (meters) Range (meters) Range (meters)

N N N

20
Crossrange (meters)

20
Crossrange (meters)

Crossrange (meters)
50
40
40
100 60

60 80
150
100
80
120
200
20 40 60 80 20 40 60 80 20 40 60 80
Range (meters) Range (meters) Range (meters)

N N N

10 10
20
Crossrange (meters)

Crossrange (meters)

Crossrange (meters)
20 20
40
30 30

40 40 60

50 50 80
60 60
100
70 70
120
20 40 60 20 40 60 20 40 60 80
Range (meters) Range (meters) Range (meters)

Fig. 23. Some of the N objects that produce false alarms with the optimal feature set. The objects in the first row are classified as S and
the rest as C.

Theory and Practice (Image Processing Series). 6000 Broken


Sound Parkway NW, Suite 300: CRC Press, Taylor & Francis
Group, 2000.
[52] D. Zhang and G. Lu, “Review of shape representation and
description techniques,” Pattern Recognition, vol. 37, no. 1, pp.
1 – 19, 2004.
[53] M. Timothy, Signal and image processing with neural networks.
New York, NY, USA: John Wiley & Sons, Inc., 1994.
[54] K. Pearson, “On lines and planes of closest fit to systems of
points in space,” Philosophical Magazine, vol. 2, no. 6, pp. 559
– 572, 1901.
[55] J. A. Fawcett, “Image-based classification of sidescan sonar
detections,” in Proceedings of the CAC/CAD Conference, 2001.
[56] P. M. Mather and T. Brandt, Classification Methods for Re-
motely Sensed Data, Second Edition. 6000 Broken Sound
Parkway NW, Suite 300: CRC Press, Taylor & Francis Group,
2009.
[57] R. E. Bellman, “Adaptive control processes,” Princeton Uni-
versity Press, 1961.
[58] C. Cortes and V. Vapnik, “Support-vector networks,” Machine
Learning, vol. 20, no. 3, pp. 273 – 297, 1995.
[59] C. M. Bishop, Pattern Recognition and Machine Learning
(Information Science and Statistics). Secaucus, NJ, USA:
Springer-Verlag New York, Inc., 2006.
[60] L. A. Zadeh, “Fuzzy sets,” Information Control, vol. 8, pp. 338
– 353, 1965.

Vous aimerez peut-être aussi