Deep Learning and Reinforcement Learning Week 5

Course Name: Deep Learning and Reinforcement Learning

Course Link: Deep Learning and Reinforcement Learning

These are answers of Deep Learning and Reinforcement Learning Week 5 Quiz


Practice: Transfer Learning

Q1. The main idea of transfer learning of a neural network is:
To keep the early layers of a pre-trained network and re-train the later layers for a specific application.
To use the early layers to capture features that are more particular to the specific data you are trying to classify.
To train the early layers such that their weights have a higher impact on the final result.
To re-train the early layers for a specific application and transfer it to a different data set

Answer: To keep the early layers of a pre-trained network and re-train the later layers for a specific application.


Q2. In the context of transfer learning, which is a guiding principle of fine tuning?
Fine tuning the hyperparameters of the CNNs
Using data that is similar to the pre-trained network
Adjust the weights of the neural network
Increase the number of later layers iteratively

Answer: Using data that is similar to the pre-trained network


Q3. In the context of transfer learning, what do we call the process in which you only train the last or a few layers instead of all layers of a neural network?
Frozen layers
Frozen weights
Updated learning
Updated layers

Answer: Frozen layers


These are answers of Deep Learning and Reinforcement Learning Week 5 Quiz


Practice: Convolutional Neural Network Architectures

Q1. This concept came as a solution to CNNs in which each layer is turned into branches of convolutions:
Inception
Workload portion
Hebbian Principle
Network Concatenation

See also  Deep Learning and Reinforcement Learning Week 6

Answer: Inception


Q2. Which CNN Architecture is considered the flash point for modern Deep Learning?
AlexNet
VGG
Inception
ResNet
LeNet

Answer: AlexNet


Q3. Which CNN Architecture can be described as a “simplified, deeper LeNet” in which the more layers, the better?
Deep Lenet
AlexNet
VGG
Inception
ResNet

Answer: VGG


These are answers of Deep Learning and Reinforcement Learning Week 5 Quiz


Q4. Which CNN Architecture is the precursor of using convolutions to obtain better features and was first used to solve the MNIST data set?
AlexNet
VGG
Inception
ResNet
LeNet

Answer: LeNet


Q5. The motivation behind this CNN Architecture was to solve the inability of deep neural networks to fit or overfit the training data better when adding layers.
LeNet
AlexNet
VGG
Inception
ResNet

Answer: ResNet


Q6. This CNN Architecture keeps passing both the initial unchanged information and the transformed information to the next layer.
LeNet
AlexNet
VGG
Inception
ResNet

Answer: ResNet


These are answers of Deep Learning and Reinforcement Learning Week 5 Quiz


Q7. Which activation function was notably used in AlexNet and contributed to its success?
ReLU (Rectified Linear Unit)
Sigmoid
Tanh
Leaky ReLU

Answer: ReLU (Rectified Linear Unit)


These are answers of Deep Learning and Reinforcement Learning Week 5 Quiz


Practice: Regularization

Q1. Which regularization technique can shrink the coefficients of the less important features to zero?
L1
Dropout
L2
Batch Normalization 

Answer: L1


Q2. (True/False) Batch Normalization tackles the internal covariate shift issue by always normalizing the input signals, thus accelerating the training of deep neural nets and increasing the generalization power of the networks.
True
False

See also  Deep Learning and Reinforcement Learning Week 2

Answer: True


Q3. Regularization is used to mitigate which issue in model training?
Both underfitting and overfitting 
High bias and low variance 
Overfitting
Underfitting

Answer: Overfitting


These are answers of Deep Learning and Reinforcement Learning Week 5 Quiz


Final Quiz

Q1. (True/False) In Keras, the Dropout layer has an argument called rate, which is a probability that represents how often we want to invoke the layer in the training.
True
False

Answer: False


Q2. What is a benefit of applying transfer learning to neural networks? 
Train early layers for specific applications and generalize that with later pre-trained layers.
Save early layers for generalization before re-training later layers for specific applications. 
Easily adjust weights of early layers to reduce training time. 
Place heavy focus on training layers that generalize the model. 

Answer: Save early layers for generalization before re-training later layers for specific applications. 


Q3. By setting ` layer.trainable = False` for certain layers in a neural network, we____ 
exclude the layers during training because they should be discarded 
freeze the layers such thattheir weights change synchronously during training. 
set the layers’ weights to zero
freeze the layers such that their weights don’t update during training. 

Answer: freeze the layers such that their weights don’t update during training. 


These are answers of Deep Learning and Reinforcement Learning Week 5 Quiz


Q4. Which option correctly orders the steps of implementing transfer learning? 
1. Freeze the early layers of the pre-trained model.
2. Improve the model by fine-tuning.
3. Train the model with a new output layer in place.
4. Select a pre-trained model as the base of our training.
3, 2, 4, 1
4, 2, 3, 1
3, 1, 2, 4
4, 1, 3, 2

See also  Deep Learning and Reinforcement Learning Week 8

Answer: 4, 1, 3, 2


These are answers of Deep Learning and Reinforcement Learning Week 5 Quiz


Q5. Given a 100×100 pixels RGB image, there are _____ features. 
300
100
10000
30000

Answer: 30000


Q6. Before a CNN is ready for classifying images, what layer must we add as the last? 
Dense layer with the number of units corresponding to (number of classes*input size)
Dense layer with the number of units corresponding to the number of classes 
Flattening layer with the number of units corresponding to the number of classes 
Flattening layer with the number of units corresponding to (number of classes*input size)

Answer: Dense layer with the number of units corresponding to the number of classes 


These are answers of Deep Learning and Reinforcement Learning Week 5 Quiz


Q7. In a CNN, the depth of a layer corresponds to the number of: 
color channels
input layers
filters applied
channel-filter combinations

Answer: filters applied


These are answers of Deep Learning and Reinforcement Learning Week 5 Quiz


More Weeks of this course: Click Here

More Coursera Courses: http://progiez.com/coursera


These are answers of Deep Learning and Reinforcement Learning Week 5 Quiz