Académique Documents
Professionnel Documents
Culture Documents
• Reading Assignments
S. Gong et al., Dynamic Vision: From Images to Face Recognition, Imperial College
Press, 2001 (pp. 168-173 and Appendix C: Mathematical Details, hard copy).
K. Kastleman, Digital Image Processing, Prentice Hall, (Appendix 3: Mathematical
Background, hard copy).
F. Ham and I. Kostanic. Principles of Neurocomputing for Science and Engineering,
Prentice Hall, (Appendix A: Mathematical Foundation for Neurocomputing,
hard copy).
A. Jain, R. Duin, and J. Mao, "Statistical Pattern Recognition: A Review", IEEE
Transactions on Pattern Analysis and Machine Intelligenve, vol. 22, no. 1, pp.
4-37, 2000 (read pp. 11-13, on-line)
• Case Studies
-2-
a1 b1
a2 b2
x= − − > reduce dimensionality − − > y = ( K << N )
... ...
aN bK
- The goal of PCA is to reduce the dimensionality of the data while retaining as
much as possible of the variation present in the original dataset.
• Dimensionality reduction
- PCA allows us to compute a linear transformation that maps data from a high
dimensional space to a lower dimensional space.
b1 = t 11 a1 + t 12 a2 +. . . +t 1n a N
b2 = t 21 a1 + t 22 a2 +. . . +t 2n a N
...
b K = t K 1 a1 + t K 2 a2 +. . . +t KN a N
t 11 t 12 ... t 1N
t 21 t 22 ... t 2N
or y = Tx where T =
... ... ... ...
tK1 tK2 ... t KN
-- --
-3-
x = a1 v1 + a2 v2 + . . . + a N v N
v 1 , v 2 , ..., v N is a basis of the N -dimensional space
(2) Lower-dimensional space representation:
x̂ = b1 u1 + b2 u2 + . . . + b K u K
Example
-4-
• Information loss
The best low-dimensional space can be determined by the "best" eigenvectors of the
covariance matrix of x (i.e., the eigenvectors corresponding to the "largest" eigen-
values -- also called "principal components").
• Methodology
1 M
Step 1: x =
M
Σ
i=1
xi
N
x − x = b1 u1 + b2 u2 + . . . + b N u N = Σ bi ui
i=1
-- --
-5-
b1
b2
...
K
b
• An example
-6-
• Geometrical interpretation
- PCA projects the data along the directions where the data varies the most.
1 0 0
0 2 0
. . .
the covariance of bi ’s is: U T CU =
. . .
. . .
0 0 K
- The covariance matrix represents only second order statistics among the vector
values.
- Since the new variables are linear combinations of the original variables, it is
usally difficult to interpret their meaning.
-- --
-7-
K
Σ
i=1
i
N
> Threshold (e.g., 0.9 or 0.95)
Σ
i=1
i
- We saw above that an original vector x can be reconstructed using its the prin-
cipla components:
K K
x̂ − x = Σ bi ui or x̂ =
i=1
Σ
i=1
bi ui + x
e = ||x − x̂||
N
e = 1/2 Σ
i=K +1
i
• Standardization
- The principal components are dependent on the units used to measure the orig-
inal variables as well as on the range of values they assume.
-8-
• Other problems