Académique Documents
Professionnel Documents
Culture Documents
OF
NEURAL NETWORKS
AND
FUZZY SYSTEMS
PRESENTED BY
K. NEELIMA
B TECH 2 ND YEAR
CVR COLLEGE,
IBP, HYDERABAD
Abstract
This paper deals with the concepts of Neural Networks and Fuzzy systems. At
First an overview of the Artificial Intelligence is given. The scope of this
presentation to make a brief induction to Artificial Neural Networks (ANNs) for
people who have no previous knowledge of them. We first make a brief
introduction to models of networks, for then describing in general terms ANNs.
As an application, we explain the backpropagation algorithm, since it is widely
used and many other algorithms are derived from it. Lastly a brief introduction to
Fuzzy systems is presented.
CONTENTS
Artificial intelligence 1
Networks 2
Perceptrons 3
Fuzzy systems 8
Bibliography 10
1. Artificial Intelligence
There are many definitions of given for Artificial Intelligence(1). Here are some
of them.
“The exciting new effort to make computers think . . .machines with minds, in the
full and literal sense” (Haugeland, 1985).
“The automation of activities that we associate with human thinking” (Bellman,
1978).
“The art of creating machines that perform functions that require intelligence
when performed by people” (Kurzweil, 1990).
“The study of how to make computers do things which, at the moment, people
do better” (Rich and Knight, 1991).
“A field of study that seeks to explain and emulate intelligent behaviour in terms
of computational processes” (Schalkoff, 1992).
“The branch of computer science that is concerned with the automation of
intelligent behaviour” (Luger and Stubblefi eld, 1992).
We can see that many definitions exist. But the question is that which one to
follow? Infact the answer is none. We have to build our own definition.
2. Networks
One efficient way of solving complex problems is following the lemma
“divide and conquer”. A complex system may be decomposed into simpler
elements, in order to be able to understand it. Also simple elements may be
gathered to produce a complex system (Bar Yam, 1997). Networks(2) are one
approach for achieving this. There are a large number of different types of
networks, but they all are characterized by the following components: a set of
nodes, and connections between nodes.
The nodes can be seen as computational units. They receive inputs, and
process them to obtain an output. This processing might be very simple (such
as summing the inputs), or quite complex (a node might contain another
network...). The connections determine the information flow between nodes.
They can be unidirectional, when the information flows only in one sense, and
bi-directional, when the information flows in either sense. The interactions of
nodes though the connections lead to a global behavior of the network, which
cannot be observed in the elements of the network. This global behavior is
said to be emergent. This means that the abilities of the network supercede the
ones of its elements, making networks a very powerful tool.
One type of network sees the nodes as ‘artificial neurons’. These are called
artificial neural networks(2) (ANNs). An artificial neuron is a computational
model inspired in the natural neurons. Natural neurons receive signals through
synapses located on the dendrites or membrane of the neuron. When the
signals received are strong enough (surpass a certain threshold), the neuron is
activated and emits a signal though the axon. This signal might be sent to
another synapse, and might activate other neurons.
Figure 1. Neuron architecture(3)
2.2 Perceptrons
Σi wixi > T
φ(Σi wixi)
• Typically also, each node in a layer (other than the output layer) is connected
to every node in the next layer by a trainable weight.
• The overall layout is illustrated in Figure 5.
Figure 4. A multi-layer network
1. If node j is an output node, then δj is the product of φ'(vj) and the error
signal ej, where φ(*) is the logistic function and vj is the total input to node j
(i.e. Σi wjiyi), and ej is the error signal for node j (i.e. the difference between
the desired output and the actual output);
2. If node j is a hidden node, then δj is the product of φ'(vj) and the weighted
sum of the δ's computed for the nodes in the next hidden or output layer that
are connected to node j.[The actual formula is δj = φ'(vj) &Sigmak δkwkj where
k ranges over those nodes for which wkj is non-zero (i.e. nodes k that actually
have connections from node j. The δk values have already been computed as
they are in the output layer (or a layer closer to the output layer than node j).]
FORWARD PASS: weights fixed, input signals propagated through network and
outputs calculated. Outputs oj are compared with desired outputs dj; the error
signal ej = dj - oj is computed.
BACKWARD PASS: starts with output layer and recursively computes the local
gradient δj for each node. Then the weights are updated using the equation above
for Δwji, and back to another forward pass.
2.8 Sigmoidal Nonlinearity
With the sigmoidal function φ(x) defined above, it is the case that φ'(vj) = yj(1 -
yj), a fact that simplifies the computations.
• If the learning rate η is very small, then the algorithm proceeds slowly, but
accurately follows the path of steepest descent in weight space.
• If η is largish, the algorithm may oscillate ("bounce off the canyon walls").
• A simple method of effectively increasing the rate of learning is to modify the
delta rule by including a momentum term:
Δwji(n) = α Δwji(n-1) + η δj(n)yi(n)
where α is a positive constant termed the momentum constant. This is called the
generalized delta rule.
The effect is that if the basic delta rule is consistently pushing a weight in the
same direction, then it gradually gathers "momentum" in that direction.
• stop after a certain number of runs through all the training data (each run
through all the training data is called an epoch);
• stop when the total sum-squared error reaches some low level. By total
sum-squared error we mean Σ p Σi ei2 where p ranges over all of the
training patterns and i ranges over all of the output units.
2.11 Initialization
Backpropagation tends to work well in some situations where human experts are
unable to articulate a rule for what they are doing - e.g. in areas depending on raw
perception, and where it is difficult to determine the attributes (in the ID3 sense)
that are relevant to the problem at hand.
For example, there is a proprietary system, which includes a backpropagation
component, for assisting in classifying Pap smears.
The system picks out from the image the most suspicious-looking
cells.
A human expert then inspects these cells.
This reduces the problem from looking at maybe 10,000 cells to
looking at maybe 100 cells - this reduces the boredom-induced
error rate.
Other successful systems have been built for tasks like reading
handwritten postcodes.
3. FUZZY SYSTEMS
3.1 Introduction
Fuzzy logic(4) is concerned with the theory of fuzzy sets, which describe
vagueness. It is a mathematical framework describing human thinking
using linguistic variables.
Fuzzy sets are defined as sets of ordered pairs of objects and their
membership functions.
1.ArtificialIntelligence
<http://www2.du.se/kurser/resurser.asp?iKursId=2097698285&kategoriID
2086658928>
2. Artificial Neural Networks for Beginners
Author Carlos
Gershenson(cgershen@vub.ac.bemailto:carlos@jlagunez.iquimica.unam.mx)
<http://www.cogs.susx.ac.uk/users/carlos/doc/FCS-ANN-tutorial.htm>
4. Fuzzy logic
<http://www2.du.se/kurser/resurser.asp?iKursId=2097698467&kategoriID=
2086658216>