Académique Documents
Professionnel Documents
Culture Documents
net/publication/8156037
CITATIONS READS
168 104
5 authors, including:
All content following this page was uploaded by Jianping Hua on 20 March 2015.
Gene expression
© The Author 2004. Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oupjournals.org 1509
J.Hua et al.
into account correlation—as opposed to the less realistic assumption (Braga-Neto and Dougherty, 2004b). Thus, trying numerous feature
of equal marginal distributions. sets and selecting the one with the lowest estimated error presents
Feature-set size was first addressed for multinomial discrimina- a multiple-comparison type problem in which it is likely that some
tion via the histogram rule by finding an expression for the expected feature set will have an estimated error far below its true error, and
error E[εd (Sn )] over all probability models, assuming equally likely therefore appear to provide excellent classification. Since variation
probability models (Hughes, 1968). The drawback of this approach is worse for large feature sets, this could create a bias in favor of
is twofold. First, in practice there is only one probability model and large feature sets, which goes directly into the teeth of the peaking
the behavior of the error can vary substantially for different mod- phenomenon. Hence, gaining an idea of feature-set sizes that provide
els. Second, all models are not equally likely in realistic settings. good classification in various circumstances is of substantial benefit.
Recently, an analytic representation of the distribution of the error
εd (Sn ) for multinomial discrimination has been discovered that is
distribution-specific (as is our simulation procedure here), and from 2 SIMULATION WITH SYNTHETIC DATA
which a distribution-specific expected error E[εd (Sn )] is derived Seven classifiers are considered in our study: 3-nearest-neighbor
(Braga-Neto and Dougherty, 2004a). The optimal feature number (3NN), Gaussian kernel, linear support vector machine (linear SVM),
is at once determined by this expectation. polynomial support vector machine (polynomial SVM), perceptron,
The other case historically addressed is classification with two regular histogram and linear discriminant analysis (LDA). For linear
Gaussian class-conditional distributions. The Bayes classifier d SVM and polynomial SVM, we use the codes provided by LIBSVM
is determined by a discriminant Qd (x), with d (x) = 1 if and 2.4 (Chang and Lin, 2000) with the default setting, except that for
1510
Optimal number of features as sample size function
0.3 0.3
0.45
0.25 0.25
0.4
error rate
error rate
0.2 0.2
error rate
0.15 0.15
0.35
0.1 0.1
0.3
0.05 0.05
0 0 0.25
0 0 0 0 0 0
5 5 5
50 50 50
10 10 10
100 15 100 15 100 15
20 20 20
150 150 150
25 25 25
200 30 200 30 200 30
feature size feature size feature size
sample size sample size sample size
Fig. 1. Optimal feature size versus sample size for LDA classifier. Linear model is tested. σ 2 is set to let Bayes error be 0.05. (a) Uncorrelated features. (b)
Slightly correlated features, G = 5, ρ = 0.125. (c) Highly correlated features, G = 1, ρ = 0.5.
1511
J.Hua et al.
0.45
0.45 0.45
0.4
0.4 0.4
0.35
error rate
error rate
error rate
0.35 0.35
0.3
0.3 0.3
0.25
50 50 50
Fig. 2. Optimal feature size versus sample size for regular histogram classifier. Uncorrelated features. σ 2 is set to let Bayes error be 0.05.
0.3
0.25 0.25
0.25
error rate
error rate
error rate
0.2 0.2
0.15
Perceptron 0.15
Linear SVM 0.2 Polynomial SVM
0.15
0.1 0.1
0.1
0 0 0
0 0 0 0 0 0
5 5 5
50 50 50
10 10 10
100 15 100 15 100 15
20 20 20
150 150 150
25 25 25
200 30 200 30 200 30
feature size feature size feature size
sample size sample size sample size
0.34 0.32
0.35
0.3
0.32
0.28 0.3
error rate
error rate
error rate
0.3
0.26
0.28
0.24 0.25
0.26
0.22
0.2
0.24 0.2
0.36 0.34
0.32 0.4
0.34
0.3
error rate
error rate
error rate
0.32
0.28 0.35
0.3
0.26
0.28
0.24 0.3
0.26 0.22
Fig. 3. Optimal feature size versus sample size for perceptron and SVM classifiers.
1512
Optimal number of features as sample size function
the need for more features to separate three concentrations of mass (a)
as opposed to two.
In Figure 3, we compare the perceptron, linear SVM and poly-
0.35
nomial SVM classifiers. Of practical importance, the linear SVM
shows no peaking phenomenon for up to 30 features, the polyno- 0.3
mial SVM peaks at under 30 only for quite small samples on the
uncorrelated linear model and the polynomial SVM shows no peak- 0.25
error rate
ing at up to 30 features for the correlated linear model. When there
is no peaking, one can safely use a large number of features even for 0.2
30
SVM for the correlated model. 20
40 25
Perhaps the most interesting aspect of Figure 3 is that there are 50 30
feature size
cases in which the optimal number of features is not monotonically (b) sample size
increasing with the sample size (and here we are not referring to the
in all cases and the linear SVM in the correlated cases, in a range
0.2
of sample sizes we do not observe the typical concave behavior of E[ε (S )] at n=30
error
d n
the error as a function of the number of features. On the contrary,
in some feature size range, the classification error will increase and 0.15
E[∆ (S )] at n=30
d n
then decrease with the feature size, thereby forming a ridge across the
error surface. A zoomed plot for the perceptron in the uncorrelated 0.1
1513
J.Hua et al.
3NN
0.34 0.34
0.4
0.32
0.32
0.3 0.35
error rate
error rate
error rate
0.3
0.28
0.28
0.26 0.3
0.26
0.24
0.25
0.22 0.24
Fig. 5. Optimal feature size versus sample size for the 3NN classifier. Correlated features, G = 1, ρ = 0.25. σ 2 is set to let Bayes error be 0.05.
0.28
0.35 0.2
0.26
0.3 0.18
0.24
error rate
error rate
0.2 0.2 0.14
0.18
0.15 0.12
0.16
0.1 0.1
0.14
0.2
0.22
0.19 0.3
0.2
0.18
0.25
0.17 0.18
error rate
error rate
error rate
0.16 0.16
0.2
0.15
0.14
0.14 0.15
0.12
0.13
Fig. 6. Optimal feature size versus sample size for various classifiers on real patient data. Sample size n = 40.
Once again the optimal-feature-number curve is not increasing using more than d = 10 features is counterproductive for the models
as a function of sample size—this being observed in the non-linear considered.
model for both classifiers. The optimal feature size is larger at very
small sample sizes, rapidly decreases, and then stabilizes as the
sample size increases. To check this stabilization, we have tested 4 PATIENT DATA
the 3NN classifier on the non-linear model case in Figure 5 for In addition to the synthetic data, we have conducted experiments
sample size up to 5000. The result shows that the optimal feature size based on real patient data. With real data it is not possible to
increases very slowly with sample size. In particular, for n = 100 perform the kind of systematic study performed on synthetic data
and n = 5000, the optimal feature sizes are 9 and 10, respectively. arising from parameterized distributions. Our purpose in consider-
This suggests a useful property of kNN and Gaussian-kernel classi- ing a particular set of real microarray data is to demonstrate that the
fiers: once we find an optimal feature size for a very modest sized behavior of optimal feature-set sizes for such data can bear similar-
sample, we can use the same number of features for much larger ity to that arising from synthetic data, in particular, with correlated
samples without sacrificing optimality. Based on our simulations, synthetic data.
1514
Optimal number of features as sample size function
The patient data come from a microarray-based cancer- features and therefore one should attempt to use a number close to the
classification study (van de Vijver et al., 2002) that analyzes a optimal number. This means that it can be useful to refer to a database
large number of microarrays prepared with RNA from breast tumor of optimal-feature-size curves to choose a feature size, even if this
samples of 295 patients. Using a previously established 70-gene means making a necessarily very coarse approximation of the distri-
prognosis profile (van’t Veer et al., 2002), a prognosis signa- bution model from the data—even perhaps just a visual assessment of
ture based on gene-expression is proposed in van de Vijver et al. the data. Owing to the roughness of these kinds of approximations,
(2002) that correlates well with patient survival data and other exist- a classifier like the polynomial SVM, which shows strong robust-
ing clinical measures. Of the 295 microarrays, 115 belong to the ness with respect to large feature sets, has inherent advantages over
‘good-prognosis’ class, whereas the remaining 180 belong to the a classifier like LDA, which does not show robustness.
‘poor-prognosis’ class.
All classifiers are tested on various feature sizes from 1 to 30, REFERENCES
except the regular histogram, which is omitted for the patient data
Bowker,A.H. (1961) A representation of Hotelling’s T 2 and Anderson’s classification
because its error surface is too rough with the limited number of statistics. In Solomon,H., ed. Studies in Item Analysis and Prediction. Stanford
replications used. To mitigate the confounding effects of feature University Press, Stanford, pp. 285–292.
selection, for each feature-set size, floating forward selection (Pudil Braga-Neto,U. and Dougherty,E.R. (2004a) Exact performance of error estimators for
et al., 1994) is used to find a (hopefully) close-to-optimal feature discrete classifiers, submitted.
Braga-Neto,U. and Dougherty,E.R. (2004b) Is cross-validation valid for small-sample
subset based on all 295 data points. This will provide ‘population-
microarray classification? Bioinformatics, 20, 374–380.
based’ feature sets whose sample-based performances can then be
1515