E-Note 15956 Content Document 20240219104507AM
E-Note 15956 Content Document 20240219104507AM
E-Note 15956 Content Document 20240219104507AM
In the above example it has two layers, out if which one is input layer and other one is output
layer and the minimum condition to construct any neural network is it should have minimum
of two layers.
For given example In input layers there are two neurons (X 1, X2) and one output neuron (Y)
thus making total of three neurons. x1 and x2 are inputs to X1 and X2 and y is out of Y neuron.
Each of neurons are connected by separate link and each links are associated with some
weights. Initially some random weight is assigned to links and after performing computation
and out put is obtained once the termination condition is satisfied (say error should be less
than 0.5) whatever weight considered that will be valid if termination condition is not
satisfied then weights need to be modified. To modify the weights there are different
algorithm like gradient, decent, back propagation etc which will be applied to update weights.
Usyally every neuron performs to types of computational operation
Summation
Apply activation function (Activation functions are not fixed one based on
requirements it can be changed like sigmoid and other functions)
i.e for the above simple neuron net architect, the net input has to be calculated in the
following way
Summation, yin = x1w1 + x2w2
Apply activation function, y = f(yin)
STATE THE MAJOR DIFFERENCES BETWEEN BIOLOGICAL AND
ARTIFICIANEURAL NETWORKS
1. Size: Our brain contains about 86 billion neurons and more than a 100 synapses
(connections). The number of “neurons” in artificial networks is much less than that.
2. Signal transport and processing: The human brain works asynchronously, ANNs work
synchronously.
3. Processing speed: Single biological neurons are slow, while standard neurons in ANNs are
fast.
4. Topology: Biological neural networks have complicated topologies, while ANNs are often
in a tree structure.
5. Speed: certain biological neurons can fire around 200 times a second on average. Signals
travel at different speeds depending on the type of the nerve impulse, ranging from 0.61 m/s
up to 119 m/s. Signal travel speeds also vary from person to person depending on their sex,
age, height, temperature, medical condition, lack of sleep etc. Information in artificial
neurons is carried over by the continuous, floating point number values of synaptic weights.
There are no refractory periods for artificial neural networks (periods while it is impossible to
send another action potential, due to the sodium channels being lock shut) and artificial
neurons do not experience “fatigue”: they are functions that can be calculated as many times
and as fast as the computer architecture would allow.
6. Fault-tolerance: biological neuron networks due to their topology are also fault-tolerant.
Artificial neural networks are not modeled for fault tolerance or self regeneration (similarly
to fatigue, these ideas are not applicable to matrix operations), though recovery is possible by
saving the current state (weight values) of the model and continuing the training from that
save state.
7. Power consumption: the brain consumes about 20% of all the human body’s energy —
despite it’s large cut, an adult brain operates on about 20 watts (barely enough to dimly light a
bulb) being extremely efficient. Taking into account how humans can still operate for a while,
when only given some c-vitamin rich lemon juice and beef tallow, this is quite remarkable.
For benchmark: a single Nvidia GeForce Titan X GPU runs on 250 watts alone, and requires
a power supply. Our machines are way less efficient than biological systems. Computers also
generate a lot of heat when used, with consumer GPUs operating safely between 50–
80°Celsius instead of 36.5–37.5 °C.
8. Learning: we still do not understand how brains learn, or how redundant connections store
and recall information. By learning, we are building on information that is already stored in
the brain. Our knowledge deepens by repetition and during sleep, and tasks that once required
a focus can be executed automatically once mastered. Artificial neural networks in the other
hand, have a predefined model, where no further neurons or connections can be added or
removed. Only the weights of the connections (and biases representing thresholds) can
change during training. The networks start with random weight values and will slowly try to
reach a point where further changes in the weights would no longer improve performance.
Biological networks usually don't stop / start learning. ANNs have different fitting (train) and
prediction (evaluate) phases.
9. Field of application: ANNs are specialized. They can perform one task. They might be
perfect at playing chess, but they fail at playing go (or vice versa). Biological neural networks
can learn completely new tasks. 10. Training algorithm: ANNs use Gradient Descent for
learning. Human brains use something different
APPLICATIONS OF ANN
1. Data Mining: Discovery of meaningful patterns (knowledge) from large volumes of data.
2. Expert Systems: A computer program for decision making that simulates thought process
of a human expert.
3. Fuzzy Logic: Theory of approximate reasoning.
4. Artificial Life: Evolutionary Computation, Swarm Intelligence.
5. Artificial Immune System: A computer program based on the biological immune system.
6. Medical: At the moment, the research is mostly on modelling parts of the human body and
recognizing diseases from various scans (e.g. cardiograms, CAT scans, ultrasonic scans,
etc.).Neural networks are ideal in recognizing diseases using scans since there is no need to
provide a specific algorithm on how to identify the disease. Neural networks learn by
example so the details of how to recognize the disease are not needed. What is needed is a set
of examples that are representative of all the variations of the disease. The quantity of
examples is not as important as the 'quantity'. The examples need to be selected very carefully
if the system is to perform reliably and efficiently.
7. Computer Science: Researchers in quest of artificial intelligence have created spin offs like
dynamic programming, object oriented programming, symbolic programming, intelligent
storage management systems and many more such tools. The primary goal of creating an
artificial intelligence still remains a distant dream but people are getting an idea of the
ultimate path, which could lead to it.
8. Aviation: Airlines use expert systems in planes to monitor atmospheric conditions and
system status. The plane can be put on autopilot once a course is set for the destination.
9. Weather Forecast: Neural networks are used for predicting weather conditions. Previous
data is fed to a neural network, which learns the pattern and uses that knowledge to predict
weather patterns.
10. Neural Networks in business: Business is a diverted field with several general areas of
specialization such as accounting or financial analysis. Almost any neural network
application would fit into one business area or financial analysis.
11. There is some potential for using neural networks for business purposes, including
resource allocation and scheduling.
12. There is also a strong potential for using neural networks for database mining, which is,
searching for patterns implicit within the explicitly stored information in databases. Most of
the funded work in this area is classified as proprietary. Thus, it is not possible to report on
the full extent of the work going on. Most work is applying neural networks, such as the
Hopfield-Tank network for optimization and scheduling.
13. Marketing: There is a marketing application which has been integrated with a neural
network system. The Airline Marketing Tactician (a trademark abbreviated as AMT) is a
computer system made of various intelligent technologies including expert systems. A feed
forward neural network is integrated with the AMT and was trained using back-propagation
to assist the marketing control of airline seat allocations. The adaptive neural approach was
amenable to rule expression. Additionally, the application's environment changed rapidly and
constantly, which required a continuously adaptive solution.
14. Credit Evaluation: The HNC company, founded by Robert Hecht-Nielsen, has developed
several neural network applications. One of them is the Credit Scoring system which
increases the profitability of the existing model up to 27%. The HNC neural systems were
also applied to mortgage screening. A neural network automated mortgage insurance under
writing system was developed by the Nestor Company. This system was trained with 5048
applications of which 2597 were certified. The data related to property and borrower
qualifications. In a conservative mode the system agreed on the under writers on 97% of the
cases. In the liberal model the system agreed 84% of the cases. This is system run on an
Apollo DN3000 and used 250K memory while processing a case file in approximately 1 sec.
ADVANTAGES OF ANN
1. Adaptive learning: An ability to learn how to do tasks based on the data given for training
or initial experience.
2. Self-Organisation: An ANN can create its own organisation or representation of the
information it receives during learning time.
3. Real Time Operation: ANN computations may be carried out in parallel, and special
hardware devices are being designed and manufactured which take advantage of this
capability.
4. Pattern recognition: is a powerful technique for harnessing the information in the data and
generalizing about it. Neural nets learn to recognize the patterns which exist in the data set.
5. The system is developed through learning rather than programming. Neural nets teach
themselves the patterns in the data freeing the analyst for more interesting work.
6. Neural networks are flexible in a changing environment. Although neural networks may
take some time to learn a sudden drastic change they are excellent at adapting to constantly
changing information.
7. Neural networks can build informative models whenever conventional approaches fail.
Because neural networks can handle very complex interactions they can easily model data
which is too difficult to model with traditional approaches such as inferential statistics or
programming logic.
8. Performance of neural networks is at least as good as classical statistical modelling, and
better on most problems. The neural networks build models that are more reflective of the
structure of the data in significantly less time.
LIMITATIONS OF ANN
In this technological era everything has Merits and some Demerits in others words there is a
Limitation with every system which makes this ANN technology weak in some points. The
various Limitations of ANN are:-
1) ANN is not a daily life general purpose problem solver.
2) There is no structured methodology available in ANN.
3) There is no single standardized paradigm for ANN development.
4) The Output Quality of an ANN may be unpredictable.
5) Many ANN Systems does not describe how they solve problems.
6) Black box Nature
7) Greater computational burden.
8) Proneness to over fitting.
9) Empirical nature of model development
ARTIFICIAL NEURAL NETWORK CONCEPTS/TERMINOLOGY
Here is a glossary of basic terms you should be familiar with before learning the details of
neural networks.
Inputs: Source data fed into the neural network, with the goal of making a decision or
prediction about the data. Inputs to a neural network are typically a set of real values; each
value is fed into one of the neurons in the input layer.
Training Set: A set of inputs for which the correct outputs are known, used to train the neural
network.
Outputs : Neural networks generate their predictions in the form of a set of real values or
boolean decisions. Each output value is generated by one of the neurons in the output layer.
Neuron/perceptron: The basic unit of the neural network. Accepts an input and generates a
prediction. Each neuron accepts part of the input and passes it through the activation function.
Common activation functions are sigmoid, TanH and ReLu. Activation functions help
generate output values within an acceptable range, and their non-linear form is crucial for
training the network.
Weight Space: Each neuron is given a numeric weight. The weights, together with the
activation function, define each neuron’s output. Neural networks are trained by fine-tuning
weights, to discover the optimal set of weights that generates the most accurate prediction.
Forward Pass: The forward pass takes the inputs, passes them through the network and
allows each neuron to react to a fraction of the input. Neurons generate their outputs and pass
them on to the next layer, until eventually the network generates an output.
Error Function: Defines how far the actual output of the current model is from the correct
output. When training the model, the objective is to minimize the error function and bring
output as close as possible to the correct value.
Backpropagation: In order to discover the optimal weights for the neurons, we perform a
backward pass, moving back from the network’s prediction to the neurons that generated that
prediction. This is called backpropagation. Backpropagation tracks the derivatives of the
activation functions in each successive neuron, to find weights that bring the loss function to
a minimum, which will generate the best prediction. This is a mathematical process called
gradient descent.
Bias and Variance: When training neural networks, like in other machine learning
techniques, we try to balance between bias and variance. Bias measures how well the model
fits the training set—able to correctly predict the known outputs of the training examples.
Variance measures how well the model works with unknown inputs that were not available
during training. Another meaning of bias is a “bias neuron” which is used in every layer of
the neural network. The bias neuron holds the number 1, and makes it possible to move the
activation function up, down, left and right on the number graph.
Hyperparameters: A hyper parameter is a setting that affects the structure or operation of
the neural network. In real deep learning projects, tuning hyper parameters is the primary way
to build a network that provides accurate predictions for a certain problem. Common hyper
parameters include the number of hidden layers, the activation function, and how many times
(epochs) training should be repeated.
In Multilayer feed-forward network along with input and output layers there will be hidden
layers. Each layers can have multiple neurons, input layers are conneted to hidden layers,
hidden layer connected to output layer.
Single node with its own feedback
In Single node with its own feedback there will be only one layer that is input layer and
output of input neuron will be the output and out put is validated and if the validation is
correct then model is accepted if validation is not correct feedback and modification is done
on computation part of neuron.
Single-layer recurrent network
Single-layer recurrent network has both input and output layers while modifying weights
instead of considering only input and previous weights output of previous iteration is
considered as well
Multilayer recurrent network.
Multilayer recurrent network is same as Single-layer recurrent network except that it has
hidden layer in addition to input and output layers.
2. Learning
The main property of an ANN is its capability to learn.
Learning or training is a process by means of which a neural network adapts itself to a
stimulus by making proper parameter adjustments, resulting in the production of desired
response.
Broadly, there are two kinds of learning in ANNs:
1. Parameter learning: It updates the connecting weights in a neural net.
2. Structure learning: It focuses on the change in network structure (which includes the
number of processing elements as well as their connection types).
WHAT IS A PERCEPTRON?
A perceptron is a binary classification algorithm modelled after the functioning of the human
brain it was intended to emulate the neuron. The perceptron, while it has a simple structure,
has the ability to learn input and output