Académique Documents
Professionnel Documents
Culture Documents
History of NN
Pitts & McCulloch (1943)
First mathematical model of biological neurons All Boolean operations can be implemented by these neuronlike nodes (with different threshold and excitatory/inhibitory connections). Competitor to Von Neumann model for general purpose computing device Origin of automata theory.
Hebb (1949)
Hebbian rule of learning: increase the connection strength between neurons i and j whenever both i and j are activated. Or increase the connection strength between nodes i and j whenever both nodes are simultaneously ON or OFF.
History of NN
Early booming (50s early 60s) Rosenblatt (1958) Perceptron: network of threshold nodes for pattern classification Perceptron learning rule Percenptron convergence theorem: everything that can be represented by a perceptron can be learned Widow and Hoff (1960, 19062) Learning rule based on gradient descent (with differentiable unit) Minskys attempt to build a general purpose machine with Pitts/McCullock units
History of NN
Renewed enthusiasm and flourish (80s present) New techniques Backpropagation learning for multi-layer feed forward nets (with non-linear, differentiable node functions) Thermodynamic models (Hopfield net, Boltzmann machine, etc.) Unsupervised learning Impressive application (character recognition, speech recognition, text-to-speech transformation, process control, associative memory, etc.) Traditional approaches face difficult challenges Caution:
Dont underestimate difficulties and limitations Poses more problems than solutions
neural networks are adjusted, or trained, so that a particular input leads to a specific target output. The next figure illustrates such a situation.There, the network is adjusted, based on a comparison of the output and thetarget, until the network output matches the target. Typically, many suchinput/target pairs are needed to train a network.Neural networks have been trained to perform complex functions in variousfields, including pattern recognition,identification,classification, speech,vision, and control systems. Target
Neural network including connections (called weights) between neurons
input
compare output
Adjust weigths
Network layers
The commonest type of artificial neural network consists of three groups, or layers, of units: a layer of "input" units is connected to a layer of "hidden" units, which is connected to a layer of "output" units. The activity of the input units represents the raw information that is fed into the network. The activity of each hidden unit is determined by the activities of the input units and the weights on the connections between the input and the hidden units. The behaviour of the output units depends on the activity of the hidden units and the weights between the hidden and output units. This simple type of network is interesting because the hidden units are free to construct their own representations of the input. The weights between the input and hidden units determine when each hidden unit is active, and so by modifying these weights, a hidden unit can choose what it represents. We also distinguish single-layer and multi-layer architectures. The single-layer organisation, in which all units are connected to one another, constitutes the most general case and is of more potential computational power than hierarchically structured multi-layer organisations. In multi-layer networks, units are often numbered by layer, instead of following a global numbering.
Credit Evaluation
The HNC company, founded by Robert Hecht-Nielsen, has developed several neural network applications. One of them is the Credit Scoring system which increase the profitability of the existing model up to 27%. The HNC neural systems were also applied to mortgage screening. A neural network automated mortgage insurance underwritting system was developed by the Nestor Company. This system was trained with 5048 applications of which 2597 were certified. The data related to property and borrower qualifications. In a conservative mode the system agreed on the underwritters on 97% of the cases. In the liberal model the system agreed 84% of the cases. This is system run on an Apollo DN3000 and used 250K memory while processing a case file in approximately 1 sec.