Artificial Neural Networks
Artificial Neural Networks
Artificial Neural Networks
Networks
Overview
1. Biological inspiration
2. Artificial neurons and neural networks
3. Learning processes
4. Learning with artificial neural networks
Biological inspiration
Animals are able to react adaptively to changes in their
external and internal environment, and they use their nervous
system to perform these behaviours.
An appropriate model/simulation of the nervous system
should be able to produce similar responses and behaviours in
artificial systems.
The nervous system is build by relatively simple units, the
neurons, so copying their behavior and functionality should be
the solution.
Biological inspiration
Dendrites
Axon
Biological inspiration
dendrites
axon
synapses
xm ......... x2 x1
Input ...
wm w2
weights ... w1
..
Processing ∑
Transfer Function
(Activation Function) f(vk)
Output y
The output is a function of the input, that is
affected by the weights, and the transfer
functions
Artificial Neural Networks
An ANN can:
1. compute any computable function, by the appropriate
selection of the network topology and weights values.
2. learn from experience!
1. Specifically, by trial‐and‐error
Multilayer Perceptron
• Input Layer
• Hidden Layers
• Output Layer
x1
xn
MPL
• Weights Updates
• Back propagation
• Two phases of computation:
– Forward pass: run the NN and compute the error for each
neuron of the output layer.
– Backward pass: start at the output layer, and pass the
errors backwards through the network, layer by layer, by
recursively computing the local gradient of each neuron.
Learning by trial‐and‐error
Continuous process of:
Trial:
Processing an input to produce an output (In terms of ANN:
Compute the output function of a given input)
Evaluate:
Evaluating this output by comparing the actual output
with the expected output.
Adjust:
Adjust the weights.
Design Issues
Initial weights (small random values
[‐1,1])
Transfer function (How the inputs and
the weights are combined to produce
output?)
Error estimation
Weights adjusting
Number of neurons
Data representation
Size of training set
Transfer Functions – Activation
Function
Linear: The output is proportional to the total
weighted input.
Threshold: The output is set at one of two values,
depending on whether the total weighted input is
greater than or less than some threshold value.
Non‐linear: The output varies continuously but not
linearly as the input changes.
Error Estimation
The root mean square error (RMSE) is a frequently-
used measure of the differences between values
predicted by a model or an estimator and the values
actually observed from the thing being modeled or
estimated
Weights Adjusting
After each iteration, weights should be adjusted to
minimize the error.
– All possible weights
– Back propagation
Back Propagation
Back-propagation is an example of supervised learning is
used at each layer to minimize the error between the
layer’s response and the actual data
The error at each hidden layer is an average of the
evaluated error
Hidden layer networks are trained this way
Back Propagation
N is a neuron.
Nw is one of N’s inputs weights
Nout is N’s output.
Nw = Nw +Δ Nw
Δ Nw = Nout * (1‐ Nout)* NErrorFactor
NErrorFactor = NExpectedOutput – NActualOutput
This works only for the last layer, as we can know
the actual output, and the expected output.
Number of neurons
Many neurons:
Higher accuracy
Slower
Risk of over‐fitting
Memorizing, rather than understanding
The network will be useless with new problems.
Few neurons:
Lower accuracy
Inability to learn at all
Optimal number.
Data representation
Usually input/output data needs pre‐processing
Pictures
Pixel intensity
Text:
A pattern
Size of training set
No one‐fits‐all formula
Over fitting can occur if a “good” training set
is not chosen
What constitutes a “good” training set?
Samples must represent the general population.
Samples must contain members of each class.
Samples in each class must contain a wide range of
variations or noise effect.
The size of the training set is related to the
number of hidden neurons
Applications Areas
Function approximation
including time series prediction and modeling.
Classification
including patterns and sequences recognition, novelty
detection and sequential decision making.
(radar systems, face identification, handwritten text recognition)
Data processing
including filtering, clustering blinds source separation and
compression.
(data mining, e-mail Spam filtering)
Advantages / Disadvantages
Advantages
Adapt to unknown situations
Powerful, it can model complex functions.
Ease of use, learns by example, and very little user
domain‐specific expertise needed
Disadvantages
Forgets
Not exact
Large complexity of the network structure
Conclusion
Artificial Neural Networks are an imitation of the biological
neural networks, but much simpler ones.
The computing would have a lot to gain from neural networks.
Their ability to learn by example makes them very flexible and
powerful furthermore there is need to device an algorithm in
order to perform a specific task.
Conclusion
Neural networks also contributes to area of research
such a neurology and psychology. They are regularly
used to model parts of living organizations and to
investigate the internal mechanisms of the brain.
Many factors affect the performance of ANNs, such
as the transfer functions, size of training sample,
network topology, weights adjusting algorithm, …
Neural network mathematics
Output
Inputs
y 14 f ( x4 , w14 ) 4
Summary