Académique Documents
Professionnel Documents
Culture Documents
DOI: http://dx.doi.org/10.13005/bpj/587
ABSTRACT
between inputs, generating observations that lie linear mapping is done so that the principal
on or close to a smooth low-dimensional manifold. components can be computed efficiently in high –
To compare and classify such observations- in dimensional feature spaces. Kernel PCA finds the
effect, to reason about the world- depend crucially principal components where the latest structure is
on modelling the nonlinear geometry of these low- low dimensional and is easier to discover.
dimensional manifolds.
Sammon Map
Suppose the data lie on or near a smooth It is a typical example of a perfect non-
non-liner manifold of lower dimensionality d<<D. linear feature extraction technique. It is widely used
To a good approximation then, there exists a linear for the 2-D visualization of data which has quite a
mapping- consisting of a translation, rotation, and high dimension. In order to minimize the error
rescaling-that maps the high-dimensional function, Sammon’s algorithm uses gradient
coordinates of each neighbourhood to global descent technique.
internal coordinates on the manifold3. By design,
the reconstruction weights W ij reflect intrinsic Multidimensional Scaling (MDS)
geometric properties of the data that are invariant For the easy analysis of proximity data on
to exactly such transformations. Therefore expect a moderate set of stimuli, Multidimensional scaling
their characterization of local geometry in the can be used. It is used to express or reveal the
original data space is expected to be equally valid hidden structure which underlies in the data. Here
for local patches on the manifold. In particular, the the points are mostly placed in a very low
same weights Wij that reconstruct the ith data point dimensional space and it is widely considered as
in D dimensions should also reconstruct its an exploratory data analysis technique. The
embedded manifold coordinates in d dimensions. visualization of the structures in the data can be
easily learnt by the researchers if the representation
Suitable Dimensionality Reduction Techniques of the patterns of proximity is done in two or three
Principal Component Analysis dimensions. Multidimensional scaling is mostly
It is a second order method and it is the classified into two types namely MDS analysis of
best in the mean-square error sense. In different interval of ratio data and MDS analysis of ordinal
domains, it is called as Singular Value data.
Decomposition (SVD), empirical orthogonal
function (EOF), Hotelling transform and the Locally Linear Embedding
Karhunen-Loeve transform. By finding a few One of the widely used approaches for
orthogonal and linear combinations of the original the problems addressing to the non-linear
variables, PCA finds to reduce the data dimension. dimensionality reduction is Locally Liner
In the field of engineering mathematics and Embedding. It is done by the computation of low-
statistics, PCA has been successfully applied in a dimensional data which preserves the
variety of domains. For high dimensional data, the neighborhood embedding of high dimensional
computation of the Eigen vectors might be infeasible data 4. The main aim here is to map the high-
since the covariance matrix is proportional to the dimensionality regional data points to a single
data point’s dimensionality. coordinate system which is assigned to a global
position.
Kernel PCA
In high dimensional data, PCA is used for Isomap
the design and modeling of the linear variabilities. It is a non-linear dimensionality reduction
But the characteristic nature of the high technique based on the Multi-Dimensional
dimensional data set is that it has a non-linear technique. The classical MDS can be represented
nature. In such cases PCA cannot determine the in a generalized non-linear manner and that is
variability of data accurately. To address the problem called as an Isomap. In the input space, MDS is not
of non-linear dimensionality reduction kernel PCA performed, but it is performed at the non linear data
can be used. With the usage of kernels, some non manifold’s geodesic space. Isomap does not apply
HARIKUMAR & KUMAR, Biomed. & Pharmacol. J., Vol. 8(1), 103-106 (2015) 105
the classic MDS to the straight line distances. It Independent Component Analysis (ICA)
always applies to the geodesic distances in order ICA is a variant of principal component
to find a low-dimensional mapping that preserves analysis (PCA) in which the components are
those pair wise distances. assumed to be mutually statistically independent
instead of merely uncorrelated. It is a computational
Factor Analysis method for separating a multivariate signal into
It is a statistical technique which is closely additive subcomponents supposing the mutual
related to the Principal Component Analysis (PCA) statistical independence of the non-Gaussian
and dimensionality reduction. It is basically a source signals. Independent Component Analysis
generative model. From a set of latents, the observed describes a model for multivariate data describing
data has been produced and through the analysis large database of samples. The variables in the
of mathematical equations the unobserved model are assumed non-Gaussian and mutually
variables has been produced. The factors always independent and they are called the independent
follow a multivariate normal distribution and it is components of the observed data.
assumed to be more uncorrelated to noise.
Projection Pursuit
Generalized SVD Unlike the PCA and FA, it is a complete
It is a very versatile tool because it can be linear method and it can easily incorporate higher
particularized to correspondence analysis under than second order information. Projection Pursuit
the appropriate choices of the constrain matrices. It is used mostly for the analysis of non-Gaussian
can also be particularized to discriminant analysis datasets. More than second order methods it is
and canonical correlation analysis. Generalized more computationally intensive. Any aspect of non-
SVD performs a similar decomposition and it Gaussianity is usually measured through projection
modifies or relaxes the conditions of orthogonality pursuits only. A projection index always defines the
according the two constrain metrices. interestingness of a direction and looks for the
directions that optimize that index.
Maximum Variance Unfolding
It is more similar to Isomap in a respect Multi-unit Objective functions
when upon the data , a neighbourhood graph is To specify the objective functions there are
defined and the pairwise distances in the resulting many different ways. Inspite of their different
graph is retained. Here the Euclidean distances formulations they all are closely related and they
are maximized between the data points under the even become equivalent in some conditions. Under
condition that the corresponding distances in the certain conditions the non-linear PCA method is
neighbourhood graph is left unchanged. Then as a applicable where the mutual information method is
result of it, the optimization problem comes into most equivalent to maximum likelihood principle.
existences and then it can be solved efficiently Some of the most important Multi-unit Objective
using semi definite programming. functions are Maximum likelihood and network
entropy, Mutual information and Kullback-Leibler
Diffusion Maps divergence, Non-linear cross-correlations, Non-
It is originating from the field of system linear PCA and Higher-order cumulant tensors.
dynamics. On the graph of a particular data, a
Markov random walk is defined upon it. By Multi-layer autoencoders
performing the random walk for a number of times They are feed-forward neural networks
the data points proximity is obtained. Using this type with hidden layers usually having an odd number.
of means only diffusion maps are defined. If the The network is trained well for the minimization
data is represented in a low dimension then the process of the mean squared error between the
pair wise diffusion distances are retained there input and the output. If the activation functions are
itself. linear and it is used in the neural network, then the
perforamance of autoencoder will be more similar
106 HARIKUMAR & KUMAR, Biomed. & Pharmacol. J., Vol. 8(1), 103-106 (2015)
REFERENCES