Académique Documents
Professionnel Documents
Culture Documents
1 5
the Department of Computer Science, University of Idaho, Idaho Falls, ID the Department of Ultrasound, The affiliated Hospital of Qingdao Univer-
84302 USA. sity, Qingdao, Shandong 266003 China
2 6
the School of Computer Science, Harbin Institute of Technology, Harbin, the Department of General Surgery, The Second Hospital of Hebei Medical
Heilongjiang 150001 China. University, Shijiazhuang, Hebei 050000 China.
3
the Department of Computer Science, Utah State University, Logan, UT
†
84322 USA. Min Xian and Yingtao Zhang are co-first authors.
4
the School of Computer Science, Harbin Institute of Technology, Weihai, * To whom correspondence should be addressed (hengda.cheng@usu.edu).
Shandong 264209 China.
1/9
Article Type Year Category # of images/Availability Metrics
Xian, et al. [3] F 2015 Graph-based 184/private TP, FP, SI, HD, MD
Shao, et al. [18] F 2015 Graph-based 450/private TP, FP, SI
Huang, et al.[4] S 2014 Graph-based 20/private ARE,TPVF, FPVF,FNVF
Pons, et al [5] S 2014 Deformable models 163/private Sensitivity, ROC area
Xian, et al. [6] F 2014 Graph-based 131/private SI, FP, AHE
Kuo, et al.[7] S 2014 Deformable models 98/private DS
Torbati, et al.[8] S 2014 Neural network 30/private JI
Moon, et al. [9] S 2014 Fuzzy C-means 148/private Sensitivity and FP
Jiang, et al.[10] S 2012 Adaboost 112/private Mean overlap ratio
Shan, et al. [11] F 2012 Neural network 60/private TP, FP, FN, HD, MD
Yang, et al.[12] S 2012 Naive Bayes classifier 33/private FP
Shan, et al.[13] F 2012 Neutrosophic L-mean 122/private TP, FP, FN, SI, HD, and MD
Liu, et al. [14] S 2012 Cellular automata 205/private TP, FP,FN, SI
Hao, et al. [57] F 2012 DPM + CRF 480/private JI
Gao, et al. [15] S 2012 Normalized cut 100/private TP, FP, SI, HD, MD
Hao, et al. [38] F 2012 Hierarchical SVM + CRF 261/private JI
Liu, et al. [16] S 2010 Level set-based 79/private TP, FP, SI
Gómez, et al.[17] S 2010 Watershed 50/private Overlap ratio, NRV and PD
Table 1. Recently published approaches. F: fully automatic, S: semi-automatic; SVM: support vector machine, CRF: conditional random field,
DPM: deformable part model, TP: true positive, FP: false positive, SI: similarity, HD: Hausdorff distance, MD: mean distance, DS: Dice similarity,
JI: Jaccard Index, ROC: Receiver operating characteristic, ARE: average radial error, TPVF : true positive volume fraction , FPVF : false positive
volume fraction, FNVF: false negative volume fraction, PR: precision ratio, MR: match rate, NRV: normalized residual value, and PD: proportional
distance.
popular quantitative metrics. [21] utilized the sticks filter [22] to enhance edge and reduce
We also put the BUS dataset and the performances of the five speckle noise before using the PDMs. Huang et al. [23] pro-
approaches on http://cvprip.cs.usu.edu/busbench. To the au- posed an automatic BUS image segmentation approach by us-
thors’ best knowledge, this is the first attempt to benchmarking ing the gradient vector flow (GVF) model [24], and the initial
the BUS image segmentation methods. With the help of this boundary was obtained by using the watershed approach.
benchmark, researchers can compare their methods with other In GDMs-based BUS image segmentation approaches, many
algorithms, and find the primary and essential factors improv- methods focused on dealing with the weak boundary and inho-
ing the segmentation performance. mogeneity of BUS images. Gomez et al. [25] proposed a BUS
The paper is organized as follows: in section II, a brief review image segmentation approach based on the active contour with-
of BUS image segmentation approaches is given; in section III, out edges (ACWE) model [26] which defined the stopping term
the set-up of the benchmark is presented; in section IV, the ex- on Mumford-Shah technique. The initial contour was a five-
periments and discussions are conducted; and in section V, the pixel radius circle centered at a point in the tumor marked by
conclusion is presented. the user. Daoud et al. [27] built a two-fold termination criterion
based on the signal-to-noise ratio and local intensity value. Gao
II. RELATED WORK et al. [28] proposed a level set approach based on the method in
Many BUS segmentation approaches have been studied in [29] by redefining the edge-based stop function using phase
the last two decades, and have been proved to be effective on congruency [30] which was invariant to intensity magnitude,
their private datasets. In this section, we present a brief review and integrated the GVF model into the level set framework. Liu
of automatic BUS image segmentation approaches. For more et al. [16] proposed a GDMs-based approach which enforced
details about BUS image segmentation approaches, refer the priors of intensity distribution by calculating the probability
survey paper [19]. density difference between the observed intensity distributions
We classify the BUS image segmentation approaches in and the estimated Rayleigh distribution.
four categories: (1) deformable models, (2) graph-based ap- Graph-based approaches: graph-based approaches gain
proaches, (3) learning-based approaches, and (4) classical ap- popularity in BUS image segmentation because of their flexi-
proaches (e.g., thresholding, region growing, and watershed). bility and efficient energy-optimization. The Markov random
Deformable models (DMs): According to the ways of repre- field - Maximum a posteriori - Iterated Conditional Mode
senting the curves and surfaces, we can generally classify DMs (MRF-MAP-ICM) and the Graph cuts are the two major
into two subcategories: (1) the parametric DMs (PDMs) and (2) frameworks in graph-based approaches.
the geometric DMs (GDMs). MRF-MAP-ICM: Boukerroui et al. [31] stated that healthy
In PDMs-based BUS image segmentation approaches, the and pathological breast tissues presented different textures on
main work was focused on generating good initial tumor bound- BUS images, and proposed an improved method in [32] by
ary. Madabhushi et al. [20] proposed a fully automatic approach modeling both intensity and texture distributions in the
for BUS tumor segmentation by initializing PDMs using the likelihood energy; they also assumed that the texture features
boundary points produced in tumor localization step; and the represented by using co-occurrence matrix follow the Gaussian
balloon forces were employed in the extern forces. Chang et al. distribution; and the parameters were estimated in a way similar
2/9
to the one in [32]. In [33], the Gaussian parameters in the like- local image features to classify small image lattices (16 × 16)
lihood energy were defined globally and specified manually. into the tumor or non-tumor classes; the radius basis function
[34] proposed a one-click user interaction to estimate Gaussian (RBF) was utilized; and 18 features (16 features from co-occur-
parameters automatically. rence matrix and the mean and variance of the intensities) were
Graph cuts: Xian et al. [3] proposed a fully automatic BUS extracted from a lattice. Jiang et al. [10] trained Adaboost clas-
image segmentation framework in which the graph cuts energy sifier using 24 Haar-like features [43] to generate a set of can-
modeled the information from both the frequency and space do- didate tumor regions and trained SVM to determine the false
mains. The data term (likelihood energy) modeled the tumor positive and true positive regions. [44] proposed an NN-based
pose, position and intensity distribution. [35] built the graph on method to segment 3D BUS images by processing 2D images
image regions, and initialized it by specifying a group of tumor slices using local image features. Othman et al. [45] trained
regions (F) and a group of background regions (B). The weight two ANNs to determine the best-possible threshold. The ANN
of any t-link was set to ∞ if the node belonged to 𝐹 ∩ 𝐵, and all had 3 layers, 60 nodes in hidden layer, one node in the output
the other weights of t-links were set to 0; and the region inten- layer. The first ANN used the Scale Invariant Feature Trans-
sity difference and edge strength discussed in [36] were applied form (SIFT) descriptors as the inputs; and the second employed
to define the weight function of the smoothness term (prior en- the texture features from the Grey Level Co-occurrence Matrix
ergy). [35] proposed a discriminative graph cut approach in (GLCM) as the inputs. [11] trained an ANN to conduct pixel-
which the data term was determined online by a pre-trained level classification by using the joint probability of intensity
Probabilistic Boosting Tree (PBT) [37] classifier. and texture [20] and two new features: the phase in the max-
In [38], hierarchical multiscale superpixel classification energy orientation (PMO) and radial distance (RD). The ANN
framework was proposed to define the data term. The hierar- had 6 hidden nodes and 1 output node.
chical classifier had four layers (20, 50, 200, and 800 superpix- Deep Learning-based approaches have been reported to
els/nodes) built by using the normalized cut and k-means for achieve state-of-the-art performance for many medical tasks
multiscale representation; the histogram difference (Euclidean such as prostate segmentation [46], cell tracking [47], muscle
distance) between adjacent superpixels was used to define the perimysium segmentation [48], brain tissue segmentation [49],
weights in the smoothness term. breast tumor diagnosis [50], etc. Deep learning models have
Low optimization speed and locally optima are the two main great potential to achieve good performance because of their
drawbacks of the MRF-MAP-ICM; while the “shrinking” prob- ability to characterize big image variations and to learn compact
lem is the main disadvantage of Graph cuts-based approaches. image representation using sufficiently large BUS image da-
Learning based approaches: both supervised and unsuper- taset. Deep learning architectures based on convolutional neural
vised learning approaches have been applied to solve the BUS networks (CNNs) and recurrent neural networks (RNNs) are
image segmentation problem. Unsupervised approaches are employed in medical image segmentation [46 - 50].
simple and fast, and commonly utilized as preprocessing to Classical approaches: Three most popular classical ap-
generate candidate image regions. Supervised approaches are proaches were applied to BUS image segmentation: threshold-
good in integrating features at different levels, but not good in ing, region growing and watershed.
applying boundary constraints to generate accurate tumor In BUS image segmentation, thresholding was often used as
boundary. a pre-processing step for tumor localization. Region growing
Clustering: Xu et al. [39] proposed a BUS image segmenta- extracts image regions by starting from a set of pixels (called
tion method applying the spatial FCM (sFCM) [40] to the local seeds) and growing seeds to large regions based on predefined
texture and intensity features. In sFCM, the membership value growth criteria. In [11], Shan et al. proposed an automatic seed
of each point was updated by using its neighbors’ membership generation approach. In [52], Kwak et al. defined the cost of
values. In [39], the number of clusters was set as 2, and the growing a region by modelling common contour smoothness
membership values were assigned by using the modes of image and region similarity (mean intensity and size).
histogram as the initial cluster centers. In [41], FCM was ap- Watershed could produce more stable results than threshold-
plied to pixel intensities for generating image regions in four ing and region growing approaches, and selecting the marker(s)
clusters; then morphology, location and size features were is the key issue in watershed segmentation. Huang et al. [40]
measured for each region; a linear regression model trained on selected the markers based on grey level and connectivity. [54]
the features was employed to produce the tumor likelihoods for applied watershed to determine the boundaries on binary image.
all regions, and the region with the highest likelihood was con- The markers were set as the connected dark regions on the bi-
sidered as a tumor. Moon et al. [9] applied FCM to image re- nary image. [56] applied watershed and post-refinement based
gions produced by using the mean shift method; the number of on grey level and location to generate candidate tumor regions.
clusters was set to 4, and the regions belonging to the darkest In Table 1, we list the brief information of 18 approaches
cluster were extracted as the tumor candidates. Shan et al. [13] published recently.
extended the FCM and proposed the neutrosophic l-means
(NLM) clustering to deal with the weak boundary problem in
BUS image segmentation; and it took the indeterminacy of
membership into consideration.
SVM and NN: Liu et al. [42] trained a SVM classifier using
3/9
Original image Radiologist A Radiologist B
4/9
Figure 2. An example of 4 ROIs with LRs of 1.3, 1.5, 1.7 and 1.9, respectively.
automatic methods; therefore, we will compare the methods in Ar is the pixel set of the tumor region generated by a segmenta-
two categories separately. tion method, and | ∙ | indicates the number of elements of a set.
In the evaluation of semi-automatic approaches, we compare TPR, FPR and AER take values in [0, 1]; and FPR could be
the segmentation performances of the two methods using the greater than 1 and takes value in [0, +∞ ). Furthermore,
same set of ROIs, and evaluate the sensitivity of the methods to Hausdorf error (HE) and mean absolute error (MAE) are used
ROIs with different looseness ratio (LR) defined by to measure the worst possible disagreement and the average
𝐵𝐷 agreement between two boundaries, respectively. Let Cm and Cr
𝐿𝑅 =
𝐵𝐷 be the boundaries of tumor in the ground truth and the segmen-
where 𝐵𝐷 is the size of the bounding box of the ground truth tation result, respectively. The HE is defined by
and is used as the baseline, and BD is the size of a ROI contain-
HE(𝐶 , 𝐶 ) = max{max {𝑑(𝑥, 𝐶 )} , max{𝑑(𝑦, 𝐶 )}} (6)
ing BD0. We produce 10 groups of ROIs with different LRs au- ∈ ∈
tomatically using the approach described in [55]: move the four where x and y are the points on boundaries Cm and Cr , respec-
sides of a ROI toward the image borders to increase the loose- tively; 𝑑(∙, 𝐶) is the distance between a point and a boundary C
ness ratio; and the amount of the move is proportional to the as
margin between the side and the image border. The LR of the
𝑑(𝑧, 𝐶) = min{‖𝑧 − 𝑘‖}
first group is 1.1; and the LR of each of the other groups is 20% ∈
larger than that of its previous group. A BUS image and its four where ‖𝑧 − 𝑘‖ is the Euclidean distance between points z and
ROIs with different LRs are shown in Figure 2. k; and 𝑑(𝑧, 𝐶) is the minimum distance between point z and all
The method in [11] is fully automatic, involves neural net- points on C.
work training and testing, and a 10-fold cross validation strat- MAE is defined by
egy is utilized to evaluate its performance. Methods in [3, 18]
need no training and operator interaction. All experiments are 𝑑(𝑥, 𝐶 ) 𝑑(𝑦, 𝐶 )
MAE(𝐶 , 𝐶 ) = 1/2 + . (7)
performed using a windows-based PC equipped with a dual- 𝑛 𝑛
∈ ∈
core (2.6 GHz) processor and 8 GB memory. The performances
In Eq. (7), 𝑛 and 𝑛 are the numbers of points on bounda-
of these methods are validated by comparing the results with
the ground truths. ries 𝐶 and 𝐶 , respectively.
Both area and boundary error metrics are employed to assess The seven metrics above were discussed in [19]. For the first
two metrics (TPR and FPR), each of them only measures a
the performance of the five approaches. The area error metrics
include the true positive ratio (TPR), false positive ratio (FPR), certain aspect of the segmentation result, and is not suitable for
describing the overall performance; e.g., a high TPR value
Jaccard index (JI), Dice’s coefficient (DSC), and area error ratio
(AER) indicates that most portion of the tumor region is in the
segmentation result; however, it cannot claim an accurate
|𝐴 ∩ 𝐴 | segmentation because it does not measure the ratio of correctly
TPR = (1)
|𝐴 | segmented non-tumor regions. The other five metrics (JI, DSC,
AER, HE and MAE) are more comprehensive and effective to
|𝐴 ∪ 𝐴 − 𝐴 |
FPR = (2) measure the overall performance of segmentation approaches,
|𝐴 | and are commonly applied to tune the parameters of
|𝐴 ∩ 𝐴 | segmentation models [3], e.g., large JI and DSC and small AER,
JI = (3) HE and MAE values indicate the high overall segmentation
|𝐴 ∪ 𝐴 |
performance.
2|𝐴 ∩ 𝐴 | Although JI, DSC, AER, HE and MAE are comprehensive
DSC = (4)
|𝐴 | + |𝐴 | metrics, we still recommend using both TPR and FPR for eval-
|𝐴 ∪ 𝐴 | − |𝐴 ∩ 𝐴 | uating BUS image segmentation; since with these two metrics,
AER = (5) we can discover some hidden characteristics that cannot be
|𝐴 |
found through the comprehensive metrics. Suppose that the al-
where Am is the pixel set of the tumor region of the ground truth,
gorithm has low overall performance (small JI and DSC, and
large AER, HE and MAE); if FPR and TPR are large, we can
5/9
Looseness Ratio (LR) Looseness Ratio (LR)
Figure 3. Segmetation results of [16]. (a) average TPR, FPR, JI, DSC and AER on different ROI sets; (b) average HE and MAE.
conclude that the algorithm has overestimated the tumor region; method in [16] can be achieved by using tight ROIs; however,
if both FPR and TPR are small, the algorithm has underesti- the values of the three metrics increase almost linearly with the
mated the tumor regions. The findings from TPR and FPR can LRs of ROIs when the looseness is greater than 1.3; this
guide the improvement of the BUS segmentation algorithms. observation shows that the overall performance of [16] drops
rapidly by using large ROIs above a certain level of LR. The
IV. APPROACHES COMPARISON AND DISCUSSIONS average MAE values decrease firstly, and then increase and
vary with the LRs in a small range. Four metrics (average JI,
In this section, we evaluate five state-of-the-art approaches DSC, AER and MAE) reach their optimal values at the LR of
[11, 14, 16, 18, 26]. For fully automatic approaches, we com- 1.5 (Table 2).
pare their average performances by using the seven metrics dis- The segmentation results of the approach in [14] are
cussed in section III-C; while for semi-automatic approaches, demonstrated in Figure 4. All average JI values are between 0.7
we also evaluate their sensitivities to different LRs. and 0.8; and all average DSC values are between 0.8 and 0.9.
A. Semi-automatic Segmentation As the method in [16], the average TPR values of the method
Ten ROIs are generated automatically for each BUS image, in [14] are above 0.7; and increase with LRs of ROIs; the
and LRs range from 1.1 to 2.9 (step is 0.2). Totally, 5620 ROIs average JI and DSC values increase firstly, and then decrease;
are generated for the entire BUS dataset, and we run each of the the average FPR values increase with the increasing looseness
semi-automatic segmentation approach 5620 times to produce of ROIs; and the average DSC, HE and MAE decrease firstly,
the results. All the segmentation results on the ROIs with the and then increase. Five metrics (average JI, DSC, AER, HE and
same LR are utilized to calculate the average TRP, FPR, DSC, MAE) reach their optimal values at the LR of 1.9 (Table 2).
AER, HE and MAE, respectively; and the results of the ap- As shown in Figures 3 and 4 and Table 2, the two approaches
proaches in [14] and [16] are shown in Figure 3 and Figure 4, achieve their best performances with different LRs (1.5 and 1.9
respectively. respectively). We can also observe the following facts:
All the average TPR and DSC values of the method in [16] [16] and [14] are quite sensitive to the sizes of ROIs; their
are above 0.7, and its average JI values vary in the range [0.65, performances vary greatly at different LRs.
0.75]. The average TPR values increase with the increasing LR Every approach achieves the best performance at a certain
values of ROIs. Both the average JI and DSC values tend to value of LR; however, not at the lowest looseness level (1.1).
increase firstly, and then decrease with the increasing LRs of
ROIs. FPR, AER and HE have low average values when the
LRs are small, which indicate that the high performance of
6/9
Metrics Area error metrics Boundary error metrics Time
Looseness Ratio
Methods Ave. TPR Ave. FPR Ave. JI Ave. DSC Ave. AER Ave. HE Ave. MAE Ave. Time (s)
1.1 0.73 0.08 0.67 0.78 0.35 45.4 12.6 18
1.3 0.79 0.10 0.72 0.82 0.31 42.2 10.9 22
1.5 0.82 0.13 0.73 0.84 0.31 44.0 10.4 27
1.7 0.83 0.17 0.73 0.83 0.33 48.3 10.9 27
1.9 0.85 0.20 0.72 0.83 0.36 51.3 11.2 30
[16]
2.1 0.86 0.24 0.71 0.82 0.39 54.9 11.7 30
2.3 0.86 0.27 0.70 0.82 0.41 57.0 12.1 36
2.5 0.87 0.32 0.69 0.80 0.46 61.3 13.1 39
2.7 0.87 0.35 0.68 0.79 0.48 62.1 13.4 40
2.9 0.86 0.40 0.66 0.77 0.54 66.2 14.6 44
1.1 0.70 0.01 0.70 0.82 0.31 35.8 11.1 487
1.3 0.76 0.02 0.75 0.85 0.26 32.0 9.1 467
1.5 0.79 0.03 0.77 0.87 0.23 29.9 8.1 351
1.7 0.82 0.05 0.79 0.88 0.23 29.5 7.8 341
1.9 0.84 0.07 0.79 0.88 0.23 29.0 7.6 336
[14]
2.1 0.86 0.10 0.79 0.88 0.24 29.5 7.7 371
2.3 0.87 0.13 0.78 0.87 0.26 31.3 8.3 343
2.5 0.89 0.16 0.77 0.87 0.28 31.9 8.5 365
2.7 0.90 0.20 0.75 0.85 0.31 34.1 9.2 343
2.9 0.90 0.25 0.73 0.84 0.35 36.9 10.2 388
Table 3. Results of three fully automatic tumor segmentation approaches [3, 11, 18] and the average optimal performances of [14] and [16].
The performances of the two approaches drop if the loose- average TPR (62.1%); however, it outperforms the method in
ness level is greater than a certain value; and the perfor- [11] on all the other six metrics.
mance of the method [14] drops much slower than that of In Table 3, we also show the average optimal performances
the method in [11]. of the methods in [16] and [14] at the LRs of 1.5 and 1.9, re-
Set 1.9 as the optimal LR for [14] and 1.5 for [16]; and [14] spectively. Their performances outperform that of all the fully
achieves better optimal average performance than that of automatic approaches, and [14] achieves the best performance
[16]. among them; but the two approaches are much slower than the
The running time of the approach in [16] is proportional to fully automatic approaches, and operators are impossible to
the size of a user specified ROI, while there is no such rela- know what the best ROI sizes are beforehand which can lead to
tionship of the running time of the approach in [14]. the best segmentation results.
The running time of the approach in [14] is slower than that C. Discussions
of the approach in [16] by one order of the magnitude.
As discussed in [19], many semi-automatic segmentation ap-
B. Fully Automatic Segmentation proaches are utilized for BUS image segmentation. User inter-
[3], [11] and [18] are three fully automatic approaches and actions (setting seeds and/or ROIs) are required in these
their segmentation performances are shown in Table 3. The approaches, and could be useful for segmenting BUS images
method in [3] achieves better performance than that of the meth- with extremely low quality. As shown in Table 3, the two inter-
ods in [11] and [18] on all five comprehensive metrics and its active approaches could achieve better performances than many
average FPR is also the lowest. The method in [11] has the fully automatic approaches if the ROI size is set properly.
same average TPR value as the method in [3]; however, its Figures 3 and 4 also demonstrate that the two approaches
average FPR value reaches 106.5% which is almost six times achieve different performances using different sizes of ROIs.
larger than that of the method in [3]; the high average FPR and Therefore, the major issue in semi-automatic approaches is how
AER values of the method in [11] indicate that a large portion to determine the best ROIs/seeds. But such issue has been ne-
of non-tumor regions are misclassified as tumor regions. glected before; most semi-automatic approaches were only fo-
Among the three fully automatic approach, [18] has the lowest cused on improving segmentation performance by designing
7/9
more complex features and segmentation models, and did not The approach modelling knowledge that is more robust can
consider user interaction as an important factor that could affect achieve better performance on image dataset with large var-
the segmentation performance. Hence, we recommend re- iance in image quality.
searchers that they should consider this issue when they develop The approach based on strong constraints such as prede-
semi-automatic approaches. Two possible solutions could be fined reference point, intensity distribution, and ROI size
applied to solve this issue. First, for a given approach, we could cannot handle BUS images from different sources well.
choose the best LR by running experiments on a given BUS im- The quantitative metrics such as JI, DSC, AER, HE and
age training set (like section IV-A) and apply the LR to the test MAE are more comprehensive and effective to measure the
set. Second, like the interactive segmentation approach in [55], overall segmentation performance than TPR and FPR; how-
we could bypass this issue by designing segmentation models ever, TPR and FPR are also useful for developing and im-
less sensitive to user interactions. proving algorithms.
Fully automatic segmentation approaches have many good In addition, the benchmark should be and will be expanded
properties such as operator-independence and reproducibility. continuously.
The key strategy that shared by many successful fully automatic
approaches is to localize the tumor ROI by modeling the do- VI. ACKNOWLEDGEMENT
main knowledge. [11] localizes tumor ROI by formulizing the
empirical tumor location, appearance and size; [22] generates This work is supported, in part, by the Chinese NSF
tumor ROI by finding adaptive reference position; and in [18], (81501477) and by the Livelihood Technology Project of Qing-
the ROI is generated to detect the mammary layer of BUS im- dao city (15-9-2-89-NSH). We would like to acknowledge Dr.
age, and the segmentation algorithm only detects the tumor in Juan Shan, Dr. Bo Liu, Dr. Yan Liu and Mr. Haoyang Shao for
this layer. However, in many fully automatic approaches, some providing the source code of their published BUS image seg-
inflexible constraints are utilized which lower their robustness; mentation approaches. We also acknowledge Dr. Jiawei Tian
e.g., [11] utilizes the fixed reference position to rank the for providing 220 images from their BUS image dataset.
candidate regions in the ROI localization process, and achieves
good performance on the private BUS dataset; however, it fails REFERENCES
to localize 8 images of the benchmark. As discussed in [19], to [1] R. L. Siegel, K. D. Miller, and A. Jemal, “Cancer statistics, 2015,” CA-
Cancer J. Clin., vol. 65, no. 1, pp. 5-29, 2015.
avoid the degradation of segmentation performance for BUS [2] H. D. Cheng, J. Shan, W. Ju et al., “Automated breast cancer detection
images not collected under the same controlled settings, it is and classification using ultrasound images: A survey,” Pattern Recognt.,
important to develop unconstrained segmentation techniques vol. 43, no. 1, pp. 299-317, Jan, 2010.
that are invariant to image settings. [3] M. Xian, Y. Zhang, and H. D. Cheng, “Fully automatic segmentation of
breast ultrasound images based on breast characteristics in space and fre-
As shown in Table 1, many different quantitative metrics ex- quency domains,” Pattern Recognit., vol. 48, no. 2, pp. 485-497, 2015.
ist for evaluating the performances of BUS image segmentation [4] Q. Huang, X. Bai, Y. Li et al., “Optimized graph-based segmentation for
approaches. In this paper, we have applied seven metrics rec- ultrasound images,” Neurocomputing, vol. 129, pp. 216-224, 2014.
ommended in [19] to evaluate BUS image segmentation ap- [5] G. Pons, R. Martí, S. Ganau et al., “Computerized Detection of Breast Le-
sions Using Deformable Part Models in Ultrasound Images,” Ultrasound
proaches. As shown in Figures 3 and 4, average JI, DSC and in medicine & biology, 2014.
AER have the same trend, and each of them is sufficient to eval- [6] M. Xian, H. Cheng, and Y. Zhang, "A Fully Automatic Breast Ultrasound
uate the area error comprehensively. Image Segmentation Approach Based on Neutro-Connectedness." In
ICPR, 2014, pp. 2495-2500.
[7] H.-C. Kuo, M. L. Giger, I. Reiser et al., “Segmentation of breast masses
V. CONCLUSION on dedicated breast computed tomography and three-dimensional breast
In this paper, we establish a BUS image benchmark and pre- ultrasound images,” Journal of Medical Imaging, vol. 1, no. 1, pp.
014501(1-12), 2014.
sent the comparing results of five state-of-the-art BUS segmen- [8] N. Torbati, A. Ayatollahi, and A. Kermani, “An efficient neural network
tation approaches; two of them are semi-automatic and others based method for medical image segmentation,” Computers in Biology
are fully automatic. The BUS dataset contains 562 BUS images and Medicine, vol. 44, pp. 76-87, 2014.
collected using three different ultrasound machines; therefore, [9] W. K. Moon, C.-M. Lo, R.-T. Chen et al., “Tumor detection in automated
breast ultrasound images using quantitative tissue clustering,” Medical
the images have large variance in terms of image contrast, Physics, vol. 41, no. 4, pp. 042901, 2014.
brightness and degree of noise, and can be valuable for testing [10] P. Jiang, J. Peng, G. Zhang et al., "Learning-based automatic breast tu-
the robustness of the algorithms as well. In the five approaches, mor detection and segmentation in ultrasound images." in ISBI, 2012,
two of them [3, 18] are graph-based approaches, [11] is ANN- pp. 1587-1590.
[11] J. Shan, H. D. Cheng, and Y. X. Wang, “Completely Automated Seg-
based approach, [16] is a level set-based segmentation approach, mentation Approach for Breast Ultrasound Images Using Multiple-Do-
and [14] is based on cell competition. main Features,” Ultrasound Med. Biol. , vol. 38, no. 2, pp. 262-275, Feb,
The quantitative analysis of the considered approaches high- 2012.
lights the following important issues. [12] M.-C. Yang, C.-S. Huang, J.-H. Chen et al., “Whole breast lesion detec-
tion using naive bayes classifier for portable ultrasound,” Ultrasound in
By using the benchmark, no approaches in this study can medicine & biology, vol. 38, no. 11, pp. 1870-1880, 2012.
achieve the same performances reported in their original [13] J. Shan, H. Cheng, and Y. Wang, “A novel segmentation method for
papers. breast ultrasound images based on neutrosophic l-means clustering,”
The two semi-automatic approaches are quite sensitive to Medical Physics, vol. 39, no. 9, pp. 5669-5682, 2012.
[14] Y. Liu, H. Cheng, J. Huang et al., “An effective approach of lesion seg-
user interaction (LR). mentation within the breast ultrasound image based on the cellular au-
tomata principle,” Journal of Digital Imaging, vol. 25, no. 5, pp. 580-
590, 2012.
8/9
[15] L. Gao, W. Yang, Z. Liao et al., “Segmentation of ultrasonic breast tu- [40] K.-S. Chuang, H.-L. Tzeng, S. Chen et al., “Fuzzy c-means clustering
mors based on homogeneous patch,” Medical Physics, vol. 39, no. 6, pp. with spatial information for image segmentation,” Computerized Medi-
3299-3318, 2012. cal Imaging and Graphics, vol. 30, no. 1, pp. 9-15, 2006.
[16] B. Liu, H. Cheng, J. Huang et al., “Probability density difference-based [41] C. Lo, Y.-W. Shen, C.-S. Huang et al., “Computer-aided multiview tu-
active contour for ultrasound image segmentation,” Pattern Recognit., mor detection for automated whole breast ultrasound,” Ultrasonic Imag-
vol. 43, no. 6, pp. 2028-2042, 2010. ing, vol. 36, no. 1, pp. 3-17, 2014.
[17] W. Gómez, L. Leija, A. Alvarenga et al., “Computerized lesion segmen- [42] B. Liu, H. D. Cheng, J. H. Huang et al., “Fully automatic and segmenta-
tation of breast ultrasound based on marker-controlled watershed trans- tion-robust classification of breast tumors based on local texture analysis
formation,” Medical Physics, vol. 37, no. 1, pp. 82-95, 2010. of ultrasound images,” Pattern Recognit., vol. 43, no. 1, pp. 280-298,
[18] H. Shao, Y. Zhang, M. Xian et al., "A saliency model for automated tu- Jan, 2010.
mor detection in breast ultrasound images.", in: IEEE ICIP, pp. 1424- [43] P. Viola, and M. J. Jones, “Robust real-time face detection,” IJCV, vol.
1428, 2015. 57, no. 2, pp. 137-154, 2004.
[19] M. Xian, Y. Zhang, H. Cheng et al., “Automatic Breast Ultrasound Im- [44] S. F. Huang, Y. C. Chen, and W. K. Moon, “Neural network analysis ap-
age Segmentation: A Survey,” arXiv preprint arXiv:1704.01472, 2017. plied to tumor segmentation on 3D breast ultrasound images,” in ISBI,
[20] A. Madabhushi, and D. N. Metaxas, “Combining low-, high-level and pp. 1303-1306, 2008.
empirical domain knowledge for automated segmentation of ultrasonic [45] A. A. Othman, and H. R. Tizhoosh, "Segmentation of Breast Ultrasound
breast lesions,” IEEE Trans. Med. Imaging, vol. 22, no. 2, pp. 155-169, Images Using Neural Networks," Engineering Applications of Neural
Feb, 2003. Networks, pp. 260-269: Springer, 2011.
[21] R.-F. Chang, W.-J. Wu, W. K. Moon et al., “Segmentation of breast tu- [46] F. Milletari, N. Navab, and S.-A. Ahmadi, “V-net: Fully convolutional
mor in three-dimensional ultrasound images using three-dimensional neural networks for volumetric medical image segmentation,” in IEEE
discrete active contour model,” Ultrasound in medicine & biology, vol. 3D Vision (3DV), 2016, pp. 565-571.
29, no. 11, pp. 1571-1581, 2003. [47] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional net-
[22] R. N. Czerwinski, D. L. Jones, and W. D. O'Brien Jr, “Detection of lines works for biomedical image segmentation,” in MICCAI, 2015, pp. 234-
and boundaries in speckle images-application to medical ultrasound,” 241.
IEEE Trans. Med. Imaging, vol. 18, no. 2, pp. 126-136, 1999. [48] Y. Xie, Z. Zhang, M. Sapkota et al., “Spatial Clockwork Recurrent Neu-
[23] Y.-L. Huang, and D.-R. Chen, "Automatic contouring for breast tumors ral Network for Muscle Perimysium Segmentation,” in MICCAI, 2016,
in 2-D sonography.", in: IEEE EMBS, 2006, pp. 3225-3228. pp. 185-193.
[24] C. Xu, and J. L. Prince, “Generalized gradient vector flow external [49] W. Zhang, R. Li, H. Deng et al., “Deep convolutional neural networks
forces for active contours,” Signal Processing, vol. 71, no. 2, pp. 131- for multi-modality isointense infant brain image segmentation,” Neu-
139, 1998. roImage, vol. 108, pp. 214-224, 2015.
[25] W. Gómez, A. Infantosi, L. Leija, W. Pereira, Active Contours without [50] J.-Z. Cheng, D. Ni, Y.-H. Chou, J. Qin, C.-M. Tiu, Y.-C. Chang, C.-S.
Edges Applied to Breast Lesions on Ultrasound, in: Springer Huang, D. Shen, C.-M. Chen, Computer-Aided diagnosis with deep
MEDICON, 2010, pp. 292-295. learning architecture: applications to breast lesions in US images and
[26] T.F. Chan, L.A. Vese, Active contours without edges, IEEE TIP 10 pulmonary nodules in CT scans, Scientific Reports, 6 (2016).
(2001) 266-277. [51] M. H. Yap, “A novel algorithm for initial lesion detection in ultrasound
[27] M.I. Daoud, M.M. Baba, F. Awwad, M. Al-Najjar, E.S. Tarawneh, Ac- breast images,” Journal of Applied Clinical Medical Physics, vol. 9, no.
curate Segmentation of Breast Tumors in Ultrasound Images Using a 4, 2008.
Custom-Made Active Contour Model and Signal-to-Noise Ratio Varia- [52] J. I. Kwak, S. H. Kim, and N. C. Kim, "RD-based seeded region growing
tions, in: IEEE SITIS, 2012, pp. 137-141. for extraction of breast tumor in an ultrasound volume," Computational
[28] L. Gao, X. Liu, W. Chen, Phase-and gvf-based level set segmentation of Intelligence and Security, pp. 799-808: Springer, 2005.
ultrasonic breast tumors, J Appl. Math. 2012 (2012) 1-22. [53] Y.-L. Huang, and D.-R. Chen, “Watershed segmentation for breast tu-
[29] C. Li, C. Xu, C. Gui, M.D. Fox, Distance regularized level set evolution mor in 2-D sonography,” Ultrasound in medicine & biology, vol. 30, no.
and its application to image segmentation, IEEE TIP 19 (2010) 3243- 5, pp. 625-632, May, 2004.
3254. [54] M. Zhang, L. Zhang, and H.-D. Cheng, “Segmentation of ultrasound
[30] P. Kovesi, Phase congruency: A low-level image invariant, Psychologi- breast images based on a neutrosophic method,” Optical Engineering,
cal Research Psychologische Forschung, 64 (2000) 136-148. vol. 49, no. 11, pp. 117001-117001-12, 2010.
[31] D. Boukerroui, O. Basset, N. Guerin et al., “Multiresolution texture [55] M. Xian, Y. Zhang, H.-D. Cheng et al., “Neutro-connectedness cut,”
based adaptive clustering algorithm for breast lesion segmentation,” IEEE Transactions on Image Processing, vol. 25, no. 10, pp. 4691-4703,
EJU, vol. 8, no. 2, pp. 135-144, 1998. 2016.
[32] E. A. Ashton, and K. J. Parker, “Multiple resolution Bayesian segmenta- [56] C. M. Lo, R. T. Chen, Y. C. Chang et al., “Multi-Dimensional Tumor
tion of ultrasound images,” Ultrasonic Imaging, vol. 17, no. 4, pp. 291- Detection in Automated Whole Breast Ultrasound Using Topographic
304, 1995. Watershed,” IEEE TMI, vol. 33, no. 7, pp. 1503-1511, 2014.
[33] G. Xiao, M. Brady, J. A. Noble et al., “Segmentation of ultrasound B- [57] Z. Hao, Q. Wang, Y. K. Seong et al., “Combining CRF and multi-hy-
mode images with intensity inhomogeneity correction,” IEEE Trans. pothesis detection for accurate lesion segmentation in breast sono-
Med. Imaging, vol. 21, no. 1, pp. 48-57, 2002. grams,” in MICCAI, 2012, pp. 504-511.
[34] G. Pons, J. Martí, R. Martí et al., "Simultaneous lesion segmentation and
bias correction in breast ultrasound images," Pattern Recognition and
Image Analysis, pp. 692-699: Springer, 2011.
[35] H.-H. Chiang, J.-Z. Cheng, P.-K. Hung et al., “Cell-based graph cut for
segmentation of 2D/3D sonographic breast images,” in IEEE ISBI: From
Nano to Macro,2010, pp. 177-180.
[36] C.-M. Chen, Y.-H. Chou, C. S. Chen et al., “Cell-competition algorithm:
A new segmentation algorithm for multiple objects with irregular bound-
aries in ultrasound images,” Ultrasound in medicine & biology, vol. 31,
no. 12, pp. 1647-1664, 2005.
[37] Z. Tu, “Probabilistic boosting-tree: Learning discriminative models for
classification, recognition, and clustering,” in IEEE ICCV, 2005, pp.
1589-1596.
[38] Z. Hao, Q. Wang, H. Ren et al., “Multiscale superpixel classification for
tumor segmentation in breast ultrasound images,” in IEEE ICIP, 2012,
pp. 2817-2820.
[39] Y. Xu, “A modified spatial fuzzy clustering method based on texture
analysis for ultrasound image segmentation,” in IEEE ISIE, 2009, pp.
746-751.
9/9