Unit 2
Unit 2
Unit 2
wij (l ) ErrjOi
A multilayer perceptron is a neural network connecting multiple layers in a directed graph, which
means that the signal path through the nodes only goes one way. Each node, apart from the
input nodes, has a nonlinear activation function.
BACKPROPAGATION IS USED TO TRAIN THE MLP:
An MLP uses backpropagation as a supervised learning technique. Since there are
multiple layers of neurons, MLP is a deep learning technique.
MLP is widely used for solving problems that require supervised learning as well as
research into computational neuroscience and parallel distributed processing.
Desired outputs are compared to achieved system outputs, and then the
systems are tuned by adjusting connection weights to narrow the difference
between the two as much as possible.
LINEAR SEPARABILITY:
Linear separability refers to the fact that classes of patterns with -dimensional
vector can be separated with a single decision surface. In the case
above, the line represents the decision surface.
Figure 2.9: Linearly Separable Pattern
So each hidden layer applies a non-linearity to its input, and the more hidden layers you
stack together, the more complex functions you will be able to model. This is why neural
networks are said to be universal approximators, because with enough hidden layers,
the network is able to approximate any mapping from the network’s input to the
expected output.
You need to remember a thumb rule to identify the right hyper-plane: “Select
the hyper-plane which segregates the two classes better”. In this scenario, hyper-
plane “B” has excellently performed this job.
Above, you can see that the margin for hyper-plane C is high as compared to
both A and B. Hence, we name the right hyper-plane as C. Another lightning reason
for selecting the hyper-plane with higher margin is robustness. If we select a hyper-
plane having low margin then there is high chance of miss-classification.
Identify the right hyper-plane (Scenario-3):Hint: Use the rules as discussed in
previous section to identify the right hyper-plane.
Some of you may have selected the hyper- plane B as it has higher margin
compared to A. But, here is the catch, SVM selects the hyper-plane which
classifies the classes accurately prior to maximizing margin. Here, hyper-
plane B has a classification error and A has classified all correctly.
Therefore, the right hyper-plane is A.
As I have already mentioned, one star at other end is like an outlier for star
class. SVM has a feature to ignore outliers and find the hyper-plane that has
maximum margin. Hence, we can say, SVM is robust to outliers.
Find the hyper-plane to segregate to classes (Scenario-5): In the
scenario below, we can’t have linear hyper-plane between the two
classes, so how does SVM classify these two classes? Till now, we have
only looked at the linear hyper-plane.