Vous êtes sur la page 1sur 2

B O O K R E V I E W

these two examples, basic concepts and idea of kernel trick, kernel-based regres-
Pattern Recognition and Machine
terms in machine learning literature are re- sion using dual representation, and kernel-
Learning
viewed. Machine learning concepts such as based modeling of the probability distribu-
Christopher M. Bishop, 73 pp., ISBN 0- Bayes’ theorem, overfitting phenomena, tion. The role of kernels in probabilistic
387-31073-8, Springer, New York 共2006兲, model selection, the curse of dimensional- discriminative models, leading to the
$74.95 hardcover. ity, decision and information theory are in- framework of Gaussian processes for re-
Reviewed by Nasser M. Nasrabadi, U.S. troduced. Chapter 2 is dedicated to the ex- gression and classification, is also provided
Army Research Laboratory, Adelphi, ploration of some particular examples of in Chapter 6. Sparse kernel machines, also
Maryland probability density functions and their known as maximum margin classifiers,
properties. In this chapter the parametric such as support vector machines 共SVM兲
This book provides an introduction to the probability distributions that are reviewed and relevance vector machines 共RVM兲, are
field of pattern recognition and machine are binomial and multinomial distributions reviewed in Chapter 7. The concept of
learning. It gives an overview of several for discrete random variables and the clas- maximizing the margin in the design of
basic and advanced topics in machine sical Gaussian distribution for continuous SVM is described. However, more recent
learning theory. The book is definitely random variables. The nonparametric den- extensions of SVM such as the least-
valuable to scientists and engineers who sity estimation based on Parzen window or squares SVM1 are not discussed. Probabi-
are involved in developing machine learn- a kernel function is also covered in this listic graphical models represented by di-
ing tools applied to signal and image pro- chapter. The next two chapters, Chapters 3 rected or undirected graph are covered in
cessing applications. This book is also and 4, discuss the classical linear regres- Chapter 8. Chapter 8 is quite long and an
suitable for courses on machine learning sion and binary classification 共logistic re- excellent review of Bayesian networks and
and pattern recognition, designed for ad- gression兲 based on linear models, respec- Markov networks is provided in detail.
vanced undergraduates or PhD students. tively. Least-squares and Bayesian-based Polynomial regression is used as an illus-
No previous knowledge of machine learn- solutions to the parameters of a linear re- tration for Bayesian networks. Image de-
ing concepts or algorithms is assumed, but gression model are given in Chapter 3. The noising is used as an example of Markov
readers need some knowledge of calculus idea of regularized least squares to avoid networks. At the end of this chapter, infer-
and linear algebra. The book is comple- overfitting is also given in Chapter 3. In ence in graphical models is used as an ef-
mented by a great deal of additional sup- addition, the concept of bias-variance de- ficient method for inference. In Chapter 9
ports for instructors and students. The sup- composition is introduced in this chapter in the well-known K-means clustering is first
ports include solutions to the exercises in order to provide flexible models that have reviewed. In addition, the Bayesian treat-
each chapter, the example data sets used the best balance between bias and vari- ment of Gaussian mixture models and the
throughout the book and the forthcoming ance. In the case of classification, the lin- use of the expectation-maximization 共EM兲
companion book that deals with practical ear discriminant functions and the well- algorithm to find their parameters is de-
and software implementations of the key known Fisher’s linear discriminant are scribed. Chapters 10 and 11 deal with sto-
algorithms. A strong point of this book is provided in Chapter 4. The concepts of chastic or deterministic approximations of
that the mathematical expressions or algo- probabilistic generative and discriminative posterior distribution used in probabilistic
rithms are usually accompanied with col- models are also reviewed in this chapter. graphical models that were previously cov-
orful graphs and figures. This definitely Chapter 4 concludes with Bayesian treat- ered in Chapter 8. Although, these approxi-
helps to communicate the concepts much ment of logistic regression and also intro- mation techniques are well known in the
better to the students or the interested re- duces the concept of Laplace approxima- stochastic literature,2,3 Bishop has been
searchers than pure description of the algo- tion which fits a Gaussian distribution able to give an excellent review of these
rithms. The book also provides an interest- centered at the mode of a given posterior topics with examples on linear and logistic
ing short biography of the key scientists distribution in order to provide a Bayesian regressions. Also, the author’s personal re-
and mathematicians who have contributed inference for the logistic regression. An al- search experience on variational inference
historically to the basic mathematical con- terative method to perform classification is in graphical models is reflected in this
cepts and methods in each chapter. to use neural networks motivated by bio- chapter. Chapter 12 covers topics such as
This book consists of 14 chapters cov- logical systems. Feed-forward neural net- principal component analysis 共PCA兲,
ering the basic concepts of the probability works and their training procedures are probabilistic PCA, factor analysis, kernel
theory, classical linear regression, binary discussed in Chapter 5. Use of regulariza- PCA, independent component analysis,
discrimination or classification, neural net- tion in neural networks to reduce the num- and a short summary of the more recent
works, and advanced topics such as kernel ber of weights as well as Bayesian neural techniques on modeling nonlinear mani-
methods, Bayesian graphical models, networks are described in this chapter. folds. An introduction to Markov models
variational inference, Monte Carlo sam- Most of the engineers in signal and im- and hidden Markov models 共HMM兲, which
pling methods, hidden Markov models, age processing are more or less familiar are well known in signal and image pro-
and fusion of classifiers. Chapter 1 intro- with the topics covered in the first four cessing literature, is given in Chapter 13.
duces the basics of machine learning and chapters of this book. In the following Linear dynamic systems 共LDS兲, such as
classical pattern recognition by introducing chapters more advanced concepts in ma- Kalman filtering and inference/learning in
two examples—recognition of handwritten chine learning and statistics are introduced. LDS, are also covered in this chapter. Fi-
digits and polynomial curve fitting. Using Chapter 6 introduces kernel functions, the nally, Chapter 14 deals with combining

Journal of Electronic Imaging 049901-1 Oct–Dec 2007/Vol. 16(4)


Downloaded From: https://www.spiedigitallibrary.org/journals/Journal-of-Electronic-Imaging on 3/29/2018 Terms of Use: https://www.spiedigitallibrary.org/terms-of-use
different models or experts. Classical tech- Bishop are the books by Hastie et al.4 and 4. T. Hastie, R. Tibshirani, and J. Friedman,
niques such as committees, boosting, and Mackay.5 The Elements of Statistical Learning,
Springer Verlag, New York 共2002兲.
mixture of experts are reviewed in Chapter 5. D. J. C. Mackay, Information Theory, In-
14. References ference, and Learning Algorithms, Cam-
A comparison of this book with other 1. J. A. K. Suykens, T. V. Gestel, J. D. bridge University Press, Cambridge, UK
texts in this area is instructive. Almost all
Brabanter, B. D. Moor, and J. Vandewalle, 共2003兲.
Least Squares Support Vector Machines,
the topics covered in this book can be World Scientific, New Jersey 共2002兲.
2. A. Gelman, J. B. Carlin, H. S. Stern, and D. Nasser M. Nasrabadi is a senior research
found in more detail in other references. In
B. Rubin, Bayesian Data Analysis, Chap- scientist at the U.S. Army Research Labo-
fact, in this book each chapter cites several man & Hall/CRC, Boca Raton, Florida ratory in the field of EO/IR image process-
excellent references for the source or for 共2003兲. ing. His current research interests are ma-
more detailed information. The closest 3. C. P. Robert and G. Casella, Monte Carlo
Statistical Methods, Springer Verlag, New chine learning kernel-based signal and
books in terms of topics covered by York 共2004兲. image processing.

Journal of Electronic Imaging 049901-2 Oct–Dec 2007/Vol. 16(4)


Downloaded From: https://www.spiedigitallibrary.org/journals/Journal-of-Electronic-Imaging on 3/29/2018 Terms of Use: https://www.spiedigitallibrary.org/terms-of-use

Vous aimerez peut-être aussi