Concepts in Deep Learning
Concepts in Deep Learning
Concepts in Deep Learning
LEARNING
VAC 3 1 0 4 2019
Preamble:
This course aims to introduce the learner to an overview of the concepts and algorithms involved in
deep learning. Deep learning is a subfield of machine learning, a subfield of artificial intelligence.
Basic concepts and application areas of machine learning, deep networks, convolutional neural
network and recurrent neural network are covered here. This is a foundational program that will
help students understand the capabilities, challenges, and consequences of deep learning and
prepare them to participate in the development of leading-edge AI technology. They will be able to
gain the knowledge needed to take a definitive step in the world of AI.
CO2 Illustrate the validation process of machine learning models using hyper-parameters
and validation sets. (Cognitive Knowledge Level: Understand)
CO3 Demonstrate the concept of the feed forward neural network and its training process.
(Cognitive Knowledge Level: Apply)
CO4 Build CNN and Recurrent Neural Network (RNN) models for different use cases.
(Cognitive Knowledge Level: Apply)
CO5 Use different neural network/deep learning models for practical applications.
(Cognitive Knowledge Level: Apply)
295
Downloaded from Ktunotes.in
COMPUTER SCIENCE AND ENGINEERING
PO1 PO2 PO3 PO4 PO5 PO6 PO7 PO8 PO9 PO10 PO11 PO12
CO1
CO2
CO3
CO4
CO5
296
Downloaded from Ktunotes.in
COMPUTER SCIENCE AND ENGINEERING
Assessment Pattern
Remember 30 30 30
Understand 40 40 40
Apply 30 30 30
Analyse
Evaluate
Create
Mark Distribution
Attendance : 10 marks
Each of the two internal examinations has to be conducted out of 50 marks. First Internal
Examination shall be preferably conducted after completing the first half of the syllabus and the
Second Internal Examination shall be preferably conducted after completing remaining part of the
syllabus.
297
Downloaded from Ktunotes.in
COMPUTER SCIENCE AND ENGINEERING
There will be two parts: Part A and Part B. Part A contains 5 questions (preferably, 2 questions
each from the completed modules and 1 question from the partly covered module), having 3 marks
for each question adding up to 15 marks for part A. Students should answer all questions from Part
A. Part B contains 7 questions (preferably, 3 questions each from the completed modules and 1
question from the partly covered module), each with 7 marks. Out of the 7 questions in Part B, a
student should answer any 5.
There will be two parts; Part A and Part B. Part A contains 10 questions with 2 questions from each
module, having 3 marks for each question. Students should answer all questions. Part B contains 2
questions from each module of which a student should answer any one. Each question can have
maximum 2 sub-divisions and carry 14 marks.
Syllabus
(General Instructions: Instructors are to introduce students to any one software platform and
demonstrate the working of the algorithms in the syllabus using suitable use cases and public
datasets to give a better understanding of the concepts discussed. Tutorial hour may be used for this
purpose)
Module-1 (Introduction)
Key components - Data, models, objective functions, optimization algorithms, Learning algorithm.
Supervised learning- regression, classification, tagging, web search, page ranking, recommender
systems, sequence learning, Unsupervised learning, Reinforcement learning, Historical Trends in
Deep Learning. Other Concepts - overfitting, underfitting, hyperparameters and validation sets,
estimators, bias and variance.
Neural Networks –Perceptron, Gradient Descent solution for Perceptron, Multilayer perceptron,
activation functions, architecture design, chain rule, back propagation, gradient based learning.
Introduction to optimization– Gradient based optimization, linear least squares. Stochastic gradient
descent, Building ML algorithms and challenges.
298
Downloaded from Ktunotes.in
COMPUTER SCIENCE AND ENGINEERING
Recurrent neural networks – Computational graphs, RNN design, encoder – decoder sequence to
sequence architectures, deep recurrent networks, recursive neural networks, modern RNNs LSTM
and GRU, Practical use cases for RNNs.
Applications – computer vision, speech recognition, natural language processing, common word
embedding: continuous Bag-of-Words, Word2Vec, global vectors for word representation (GloVe).
Research Areas – autoencoders, representation learning, boltzmann machines, deep belief
networks.
Text Book
1. Ian Goodfellow, YoshuaBengio, Aaron Courville, Deep Learning, MIT Press 2015 ed.
2. Aston Zhang, Zachary C. Lipton, Mu Li, and Alexander J. Smola, Dive into Deep Learning,
August 2019.
3. Neural Networks and Deep Learning, Aggarwal, Charu C., c Springer International
Publishing AG, part of Springer Nature 2018
Reference Books
1. Neural Smithing: Supervised Learning in Feedforward Artificial Neural Networks by
Russell Reed, Robert J MarksII, A Bradford Book,2014
2. Practical Convolutional Neural Networks by MohitSewak, Md. Rezaul Karim,
PradeepPujari,Packt Publishing 2018
3. Hands-On Deep Learning Algorithms with Python by SudharsanRavichandran,Packt
Publishing 2019
4. Deep Learning with Python by Francois Chollet,Manning Publications Co.,2018
299
Downloaded from Ktunotes.in
COMPUTER SCIENCE AND ENGINEERING
0.9
0.8
300
Downloaded from Ktunotes.in
COMPUTER SCIENCE AND ENGINEERING
QP CODE: PAGES:4
Reg No:_______________
Name:_________________
PART A
Answer all Questions. Each question carries 3 Marks
1. Distinguish between supervised learning and Reinforcement learning. Illustrate
with an example.
301
Downloaded from Ktunotes.in
COMPUTER SCIENCE AND ENGINEERING
4. Why does a single perceptron cannot simulate simple XOR function? Explain
how this limitation is overcome?
7. How many parameters are there in AlexNet? Why the dataset size (1.2 million) is
important for the success of AlexNet?
Part B
(Answer any one question from each module. Each question carries 14
Marks)
11. (a) “A computer program is said to learn from experience E with respect to some
class of
(10)
tasks T and performance measure P, if its performance at tasks in T, as
measured by P, improves with experience E.” What is your understanding of
the terms task, performance and experience. Explain with two example
(b) “How does bias and variance trade-off affect machine learning algorithms?
(4)
OR
12. (a) Illustrate the concepts of Web search, Page Ranking, Recommender systems
with suitable examples.
(10)
(b) List and discuss the different hyper parameters used in fine tuning the (4)
302
Downloaded from Ktunotes.in
COMPUTER SCIENCE AND ENGINEERING
13. (a) How multilayer neural networks learn and encode higher level features from
input features. (7)
(b) Explain gradient decent and delta rule? Why stochastic approximation to
gradient descent is needed? (7)
OR
14. (a) Find the new weights for the network using backpropogation algorithm, the (7)
network is given with a input pattern[-1,1] and target output as +1, Use
learning rate of alpha=0.3 and bipolar sigmoid function.
(b) Write an algorithm for backpropgation which uses stochastic gradient descent (7)
method. Comment on the effect of adding momentum to the network.
15. (a) Input to CNN architecture is a color image of size 112x112x3. The first (5)
convolution layer comprises of 64 kernels of size 5x5 applied with a stride
of 2 and padding 0. What will be the number of parameters?
(b) Let X=[-1, 0, 3, 5] W=[.3 ,.5 .2,.1] be the the input of ith layer of a neural (4)
network and to apply softmax function. What should be the output of it?
OR
16. (a) Explain the concept behind i) Early stopping ii) dropout iii) weight decay (9)
303
Downloaded from Ktunotes.in
COMPUTER SCIENCE AND ENGINEERING
17. (a) Explain the working of RNN and discuss how backpropagation through time
is used in recurrent networks. (8)
(b) Describe the working of a long short term memory in RNNs. (6)
OR
18. (a) What is the vanishing gradient problem and exploding gradient problem? (8)
(b) Why do RNNs have a tendency to suffer from exploding/vanishing gradient? (6)
How to overcome this challenge?
(b) Explain the merits and demerits of using Auto encoders in Computer Vision. (6)
OR
20. (a) Illustrate the use of representation learning in object classification. (7)
Teaching Plan
1.1 Key components - Data, models, objective functions, optimization algorithms. 1 hour
(TB2: Section 1.1-1.2)
304
Downloaded from Ktunotes.in
COMPUTER SCIENCE AND ENGINEERING
1.2 Learning algorithm (TB1: Section 5.1), Supervised learning- regression, 1 hour
classification (TB2: Section 1.3.1)
1.3 tagging, web search, page ranking (TB2: Section 1.3.1) 1 hour
1.7 Concepts: Estimators, bias and variance. (TB1: Section 5.4) 1 hour
1.8 Demonstrate the concepts of supervised learning algorithms using a suitable 1 hour
platform.
2.1 Perceptron, Stochastic Gradient descent, Gradient descent solution for 1 hour
perceptron (TB3: Section 1.1 - 1.2.1)
2.2 Multilayer perceptron (TB3: Section 1.2.2), (TB1: Section 6.1,6.3) 1 hour
2.3 Activation functions- Sigmoid, tanh, Softmax, ReLU, leaky ReLU (TB3: 1 hour
Section 1.2.1.3 - 1.2.1.5)
2.4 Architecture design (TB1: Section 6.4, TB3: Section 1.6) 1 hour
305
Downloaded from Ktunotes.in
COMPUTER SCIENCE AND ENGINEERING
2.8 Linear least squares using a suitable platform. (TB1: Section 4.5) 1 hour
2.9 Building ML Algorithms and Challenges (TB3: 1.4, TB1: 5.10-5.11) 1 hour
3.3 Convolution and Pooling as an infinitely strong prior (TB1: Section 9.4) 1 hour
3.8 Practical challenges of common deep learning architectures- early Stopping 1 hour
(TB3: 4.6)
306
Downloaded from Ktunotes.in
COMPUTER SCIENCE AND ENGINEERING
4.3 Encoder – decoder sequence to sequence architectures. (TB1: Section 10.4) 1 hour
4.5 Recursive neural networks , Modern RNNs, (TB1: Section 10.6, 10.10) 1 hour
4.6 LSTM and GRU (TB1: Section 10.10, TB3: Section 7.5-7.6) 1 hour
4.7 Practical use cases for RNNs. (TB1: Section 11.1-11.4) 1 hour
307
Downloaded from Ktunotes.in
COMPUTER SCIENCE AND ENGINEERING
5.7 Brief introduction on current research areas- representation learning. (TB3: 1 hour
Section 9.3)
5.8 Brief introduction on current research areas- Boltzmann Machines, Deep belief 1 hour
networks. (TB1: Section 20.1, TB3 Section 6.3)
5.9 Brief introduction on current research areas- Deep belief networks. (TB1: 1 hour
Section 20.3)
308
Downloaded from Ktunotes.in