Académique Documents
Professionnel Documents
Culture Documents
Near East University/Electrical & Electronic Engineering, Lefkosa, via Mersin-10, TurkeyMember, Centre of
Innovation for Artificial Intelligence, CiAi,
Email: oyebade.oyedotun@yahoo.com, obalolu117@yahoo.com
Adnan Khashman3
3
Founding Director, Centre of Innovation for Artificial Intelligence (CiAi), British University of Nicosia, Girne, Mersin
10, Turkey
Email: adnan.khashman@bun.edu.tr
I. INTRODUCTION
Pattern recognition has been successfully applied in
different fields, and the benefits therein are somewhat
obvious; in particular, character recognition has sufficed
in applications such as reading bank checks, postal codes,
document analysis and retrieval etc.
In view of this, several recognition systems exist today,
which employs different approaches and algorithms for
achieving such tasks. Some common recognition systems
approach that have been in use include template matching,
syntactic analysis, statistical analysis, and artificial neural
networks. The above mentioned approaches to the
problem of pattern recognition, and therefore character
can be classified into non-intelligent and intelligent
systems. If intelligence is the ability to adapt to the
environment and to learn from experience [1]; then we
can say that approaches to pattern recognition such as
Copyright 2015 MECS
Fig.4. (a), (b), and (c) shows the images with 20%,
30%, and 40% levels of salt & pepper noise added
respectively.
It is the aim to design machine learning algorithms and
novel neural network architectures that are more robust to
pattern invariance and noise etc.
( y)
( L1)
( L1)
W
x)
encoder
( L1)
W
L1( x))
decoder
(1)
(2)
C ( x, y ) ( y k x k ) 2
1
(3)
C ( x, y) ( xk log( yk ) (1 xk ) log(1 yk ))
(4)
p(h j 1 | v; ) (Wij xi b j )
(8)
p( xi 1 | h; ) (Wij h j bi )
(9)
(10)
E ( x, h; ) Wij xi h j bi xi b j h j (5)
i
e ( E ( x ,h; ))
p( x, h; )
Z
(6)
Z i h e ( E ( x,h; ))
(7)
B. Database A3
For the purpose of evaluating the tolerance of the
trained networks to translation a separate database was
collected with the same characters and other feature
characteristics in databases A1 and A2 save that the
characters in the images have now been translated
horizontally and vertically. The figure below describes
these translations.
C. Database A4
This database contains the rotated characters contained
in database A1 and A2. Its sole purpose is to further
evaluate the performance of trained networks on pattern
rotation. See fig.12 for samples.
D.Database A5
This database is essentially databases A1 and A2
except that the scales of characters in the images have
now been purposely made different in order to evaluate
the performance of the networks on scale mismatch. It
will be seen that some characters are now bigger or
smaller as compared to the training and validation
characters earlier shown in Fig.10.
Fig.13 show samples contained in this database.
E. Database A6
In order to assess the performance of the networks on
noisy data, sub-databases with added salt & pepper noise
of different densities were collected as described below.
Database A6_1:
Database A6_2:
Database A6_3:
Database A6_4:
Database A6_5:
Translation
Rotation
Scale
BPNN 1 layer
0.8571
0.3140
0.3571
BPNN 2 layers
0.8286
0.2729
0.3658
DAE 1 layer
0.8000
0.2486
0.3057
SDAE 2 layers
0.7429
0.2214
0.2743
DBN - 2 layers
0.8143
0.1986
0.2329
2.5%
5%
10%
BPNN 1 layer
0.3214
0.3486
0.3471
BPNN 2 layers
0.2943
0.3158
0.3414
DAE 1 layer
0.2843
0.3243
0.4357
SDAE 2 layers
0.2586
0.3043
0.3857
DBN - 2 layers
0.2371
0.2771
0.3843
1st layer
2nd layer
BPNN 1 layer
65
BPNN 2 layers
95
20%
30%
65
BPNN 1 layer
0.4129
0.4829
0.4486
0.5586
DAE 1 layer
100
BPNN 2 layers
SDAE 2 layers
95
65
DAE 1 layer
0.5986
0.6843
DBN - 2 layers
200
150
SDAE 2 layers
0.5386
0.6371
DBN - 2 layers
0.6414
0.7729
Train
Validate
BPNN 1 layer
0.0433
0.0734
BPNN 2 layers
0.0277
0.0639
DAE 1 layer
0.0047
0.0679
SDAE 2 layers
0.0028
0.0567
DBN - 2 layers
0.0023
0.0377
VII. CONCLUSION
This research is meant to explore and investigate some
common problems that occur in recognition systems that
are neural network based. It will be seen that the deep
belief networks on the average, performed best compared
to the other networks on variances like translation,
rotation and scale mismatch; while its tolerance to noise
decreased noticeably as the level of noise was increased
as shown in table 4, table 5, and fig.15.
A noteworthy attribute of the patterns (Yoruba vowel
characters) used in validating this research is that they
contain diacritical marks which increases the achievable
variations of each pattern, and as such, recognition
systems designed and described in this work have been
tasked with a harder classification problem.
The performances of the denoising auto encoder (DAE1 hidden layer) and stacked denoising auto encoder
(SDAE-2 hidden layer), on the average with respect to the
variances in character images seems to second the deep
belief network.
The performance of the denoising auto encoder is
lower than that of the stacked denoising auto encoder, it
can be conjured that the stacked denoising auto encoder is
less sensitive to the randomness of the input; of course the
training and validation errors for the SDAE are also lower
to the DAE, and the tolerance to variances introduced into
the input significantly higher. i.e. a kind of higher
hierarchical knowledge of the training data achieved.
It is the hope that machine learning algorithms and
neural network architectures, which when trained once,
perform better on invariances that can occur in the
patterns that they have been trained with can be explored
for more robust applications. This also obviously saves
time and expenses in contrast to training many different
networks for such situations.
Furthermore, building invariances by the inclusion of
all possible pattern invariances which can occur when
deployed in applications during the training phase is one
solution that has been exploited; unfortunately this is not
always feasible as the capacity of the network is
concerned. i.e. considering number of training samples
enough to guarantee that proper learning has been
achieved.
It can be seen that the major problem in deep learning is
not in obtaining low error rates on the training and
validation sets (i.e. optimization) but on the other
databases which contain variant constraints of interest (i.e.
regularization). These variances are common constraints
that occur in real life recognition systems for handwritten
characters, and some of the solutions have been
constraining the users (writers) to some particular possible
domains of writing spaces or earmarked pattern of writing
in order for low error rates to be achieved.
Copyright 2015 MECS
REFERENCES
[1] Sternberg, R. J., & Detterman, D. K. (Eds.), What is
Intelligence?, Norwood, USA: Ablex, 1986, pp.1
[2] Morgan McGuire, "An image registration technique for
recovering rotation, scale and translation parameters",
Massachusetts Institute of Technology, Cambridge MA,
1998, pp.3
[3] Gorge, D., and Hawkins, J., Invariant Pattern Recognition
using Bayesian Inference on Hierarchical Sequences, In
Proceedings of the International Joint Conference on
Neural Networks. IEEE, 2005. pp.1-7
[4] Esa Rahtu, Mikko Salo and Janne Heikkil, Affine
invariant
pattern
recognition
using
Multiscale
Autoconvolution, IEEE Transactions on Pattern Analysis
and Machine Intelligence, 2005, 27(6): pp.908-18
[5] Joarder Kamruzzaman and S. M. Aziz, A Neural Network
Based Character Recognition System Using Double
Backpropagation, Malaysian Journal of Computer Science,
Vol. 11 No. 1, 1998, pp. 58-64
10
Authors Profiles
Oyebade K. Oyedotun is a member of
Centre of Innovation for Artificial
Intelligence, British University of Nicosia,
Girne, via Mersin-10, Turkey and currently
pursuing a masters degree program in
Electrical/Electronic Engineering at Near
East University, Lefkosa, via Mersin-10,
Turkey.
Research interests include artificial neural networks, pattern
recognition, machine learning, image processing, fuzzy systems
and robotics.
PH-+905428892591. E-mail: oyebade.oyedotun@yahoo.com
Ebenezer O. Olaniyi is a member of
Centre of Innovation for Artificial
Intelligence, British University of Nicosia,
Girne, via Mersin-10, Turkey and currently
pursuing a masters degree program in
Electrical/Electronic Engineering at Near
East University, Lefkosa, via Mersin-10,
Turkey.
His research interest areas are Artificial neural network,
Pattern Recognition, Image processing, Machine Learning,
Speech Processing and Robotics.
PH- +905428827442. E-mail: obalolu117@yahoo.com
Adnan Khashman received the B.Eng.
degree in electronic and communication
engineering
from
University
of
Birmingham, England, UK, in 1991, and
the M.S and Ph.D. degrees in electronic
engineering
from
University
of
Nottingham, England, UK, in 1992 and
1997.
During 1998-2001 he was an Assistant
Professor and the Chairman of Computer Engineering
Department, Near East University, Lefkosa, Turkey. During
2001-2009 he was an Associate Professor and Chairman of
Electrical and Electronic Engineering Department, and in 2009
he received the professorship. In 2001 he established the
Intelligent Systems Research Group (ISRG) at the same
university, and has been chairing ISRG until present. From
2007 until 2008 he was also the Vice-Dean of Engineering
Faculty, later on from 2013 until 2014 the Dean of the faculty at
the same university.
Since 2014, Professor Khashman was appointed as the
Founding Dean of Engineering Faculty at the British University
of Nicosia in N. Cyprus, where he also established the Centre of
Innovation for Artificial Intelligence His current research
interests include image processing, pattern recognition,
emotional modeling and intelligent systems.
E-mail: adnan.khashman@bun.edu.tr