Introduction to Machine Learning Nptel Week 3 Answers

Are you looking for Introduction to Machine Learning Nptel Week 3 Answers? You’ve come to the right place! Access the latest and most accurate solutions for your Week 3 assignment in the Introduction to Machine Learning course.

Course Link: Click Here


Introduction to Machine Learning Nptel Week 3 Answers
Introduction to Machine Learning Nptel Week 3 Answers

Introduction to Machine Learning Nptel Week 3 Answers (July-Dec 2024)


Q1.For a two-class problem using discriminant functions (δ discriminant function for class k
), where is the separating hyperplane located?

Where δ1>δ2

Where δ1<δ2

Where δ1=δ2

Where δ1+δ2=1

Answer: Where δ1=δ2


Q2. Given the following dataset consisting of two classes, A and B ,calculate the prior probability of each class.

What are the prior probabilities of class A and class B ?

P(A)=0.5,P(B)=0.5

P(A)=0.625,P(B)=0.375

P(A)=0.375,P(B)=0.625

P(A)=0.6,P(B)=0.4

Answer: P(A)=0.625,P(B)=0.375


For answers or latest updates join our telegram channel: Click here to join

These are Introduction to Machine Learning Nptel Week 3 Answers


Q3.In a 3-class classification problem using linear regression, the output vectors for three data points are [0.8, 0.3, -0.1], [0.2, 0.6, 0.2], and [-0.1, 0.4, 0.7]. To which classes would these points be assigned?
1, 2, 1
1, 2, 2
1, 3, 2
1, 2, 3

Answer: 1, 2, 3


Q4.If you have a 5-class classification problem and want to avoid masking using polynomial regression, what is the minimum degree of the polynomial you should use?
3
4
5
6

Answer: 4


For answers or latest updates join our telegram channel: Click here to join

These are Introduction to Machine Learning Nptel Week 3 Answers


Q5. Consider a logistic regression model where the predicted probability for a given data point is 0.4. If the actual label for this data point is 1, what is the contribution of this data point to the log-likelihood?
-1.3219
-0.9163
+1.3219
+0.9163

Answer: -0.9163


Q6.What additional assumption does LDA make about the covariance matrix in comparison to the basic assumption of Gaussian class conditional density?
The covariance matrix is diagonal
The covariance matrix is identity
The covariance matrix is the same for all classes
The covariance matrix is different for each class

Answer: The covariance matrix is the same for all classes


For answers or latest updates join our telegram channel: Click here to join

These are Introduction to Machine Learning Nptel Week 3 Answers


Q7.What is the shape of the decision boundary in LDA?
Quadratic
Linear
Circular
Can not be determined

Answer: Linear


Q8.For two classes C1 and C2 with within-class variances σ2w1=1 and σ2w2=4 respectively, if the projected means are µ′1=1 and µ′2=3 , what is the Fisher criterion J(w) ?
0.5
0.8
1.25
1.5

Answer: 0.8


Q9.Given two classes C1 and C2 with means µ1=[23] and µ2=[57] respectively, what is the direction vector w for LDA when the within-class covariance matrix Sw is the identity matrix I ?

[43]

[57]

[0.70.7]

[0.60.8]

Answer: [43]

For answers or latest updates join our telegram channel: Click here to join

These are Introduction to Machine Learning Nptel Week 3 Answers


All weeks of Introduction to Machine Learning: Click Here

For answers to additional Nptel courses, please refer to this link: NPTEL Assignment Answers


Introduction to Machine Learning Nptel Week 3 Answers (Jan-Apr 2024)

Course name: Introduction to Machine Learning

Course Link: Click Here

For answers or latest updates join our telegram channel: Click here to join

These are Introduction to Machine Learning Nptel Week 3 Answers


Q1. Which of the following statement(s) about decision boundaries and discriminant functions of classifiers is/are true?
In a binary classification problem, all points x on the decision boundary satisfy δ1(x)=δ2(x)
In a three-class classification problem, all points on the decision boundary satisfy δ1(x) = δ2(x) = δ3(x)
In a three-class classification problem, all points on the decision boundary satisfy at least one of δ1(x) = δ2(x), δ2(x) = δ3(x) or δ3(x) = δ1(x).
Let the input space be Rn. If x does not lie on the decision boundary, there exists an ϵ>0 such that all inputs y satisfying ||y−x||<ϵ belong to the same class.

Answer: A, B, D


Q2. The following table gives the binary ground truth labels yi for four input points xi (not given). We have a logistic regression model with some parameter values that computes the probability p(xi) that the label is 1. Compute the likelihood of observing the data given these model parameters.
0.346
0.230
0.058
0.086

Answer: 0.230


For answers or latest updates join our telegram channel: Click here to join

These are Introduction to Machine Learning Nptel Week 3 Answers


Q3. Which of the following statement(s) about logistic regression is/are true?
It learns a model for the probability distribution of the data points in each class.
The output of a linear model is transformed to the range (0, 1) by a sigmoid function.
The parameters are learned by optimizing the mean-squared loss.
The loss function is optimized by using an iterative numerical algorithm.

See also  Introduction to Machine Learning Nptel Week 4 Answers

Answer: b, d


Q4. Consider a modified form of logistic regression given below where k is a positive constant and β0 and β1 are parameters.
log(1−p(x)/kp(x))=β0−β1x
Then find p(x).

eβ0/keβ0+eβ1x
eβ1x/eβ0+keβ1x
eβ1x/keβ0+eβ1x
eβ1x/keβ0+e−β1x

Answer: c. eβ1x/keβ0+eβ1x


For answers or latest updates join our telegram channel: Click here to join

These are Introduction to Machine Learning Nptel Week 3 Answers


Q5. Consider a Bayesian classifier for a 3-class classification problem. The following tables give the class-conditioned density fk(x) for three classes k=1,2,3 at some point x in the input space.
Note that πk denotes the prior probability of class k. Which of the following statement(s) about the predicted label at x is/are true?

If the three classes have equal priors, the prediction must be class 2
If π3<π2 and π1<π2, the prediction may not necessarily be class 2
If π1>2π2, the prediction could be class 1 or class 3
If π1>π2>π3, the prediction must be class 1

Answer: a, c


Q6. The following table gives the binary labels (y(i)) for four points (x(i)1,x(i)2) where i = 1,2,3,4. Among the given options, which set of parameter values β0,β1,β2 of a standard logistic regression model p(xi)=1/1+e−(β0+β1x+β2x) results in the highest likelihood for this data?
β0=0.5,β1=1.0,β2=2.0
β0=−0.5,β1=−1.0,β2=2.0
β0=0.5,β1=1.0,β2=−2.0
β0=−0.5,β1=1.0,β2=2.0

Answer: c. β0=0.5,β1=1.0,β2=−2.0


For answers or latest updates join our telegram channel: Click here to join

These are Introduction to Machine Learning Nptel Week 3 Answers


Q7. Which of the following statement(s) about a two-class LDA model is/are true?
It is assumed that the class-conditioned probability density of each class is a Gaussian
A different covariance matrix is estimated for each class
At a given point on the decision boundary, the class-conditioned probability densities corresponding to both classes must be equal
At a given point on the decision boundary, the class-conditioned probability densities corresponding to both classes may or may not be equal

Answer: A, D


Q8. Consider the following two datasets and two LDA models trained respectively on these datasets.
Dataset A: 100 samples of class 0; 50 samples of class 1
Dataset B: 100 samples of class 0 (same as Dataset A); 100 samples of class 1 created by repeating twice the class 1 samples from Dataset A
The classifier is defined as follows in terms of the decision boundary wTx+b=0. Here, w is called the slope and b is called the intercept.
x={0 if wTx+b<0 1 if wTx+b≥0

Which of the given statement is true?
The learned decision boundary will be the same for both models
The two models will have the same slope but different intercepts
The two models will have different slopes but the same intercept
The two models may have different slopes and different intercepts

Answer: The two models will have the same slope but different intercepts


For answers or latest updates join our telegram channel: Click here to join

These are Introduction to Machine Learning Nptel Week 3 Answers


Q9. Which of the following statement(s) about LDA is/are true?
It minimizes the between-class variance relative to the within-class variance
It maximizes the between-class variance relative to the within-class variance
Maximizing the Fisher information results in the same direction of the separating hyperplane as the one obtained by equating the posterior probabilities of classes
Maximizing the Fisher information results in a different direction of the separating hyperplane from the one obtained by equating the posterior probabilities of classes

Answer: b, c


Q10. Which of the following statement(s) regarding logistic regression and LDA is/are true for a binary classification problem?
For any classification dataset, both algorithms learn the same decision boundary
Adding a few outliers to the dataset is likely to cause a larger change in the decision boundary of LDA compared to that of logistic regression
Adding a few outliers to the dataset is likely to cause a similar change in the decision boundaries of both classifiers
If the within-class distributions deviate significantly from the Gaussian distribution, logistic regression is likely to perform better than LDA

Answer: b, c


For answers or latest updates join our telegram channel: Click here to join

These are Introduction to Machine Learning Nptel Week 3 Answers

More Weeks of Introduction to Machine Learning: Click here

More Nptel Courses: https://progiez.com/nptel-assignment-answers


Introduction to Machine Learning Nptel Week 3 Answers (July-Dec 2023)

Course Name: Introduction to Machine Learning

Course Link: Click Here

These are Introduction to Machine Learning Nptel Week 3 Answers


Q1. Which of the following are differences between LDA and Logistic Regression?
Logistic Regression is typically suited for binary classification, whereas LDA is directly applicable to multi-class problems
Logistic Regression is robust to outliers whereas LDA is sensitive to outliers
both (a) and (b)
None of these

Answer: both (a) and (b)


Q2. We have two classes in our dataset. The two classes have the same mean but different variance.
LDA can classify them perfectly.
LDA can NOT classify them perfectly.
LDA is not applicable in data with these properties
Insufficient information

See also  NPTEL Introduction To Machine Learning IITKGP ASSIGNMENT 7

Answer: LDA can NOT classify them perfectly.


These are Introduction to Machine Learning Nptel Week 3 Answers


Q3. We have two classes in our dataset. The two classes have the same variance but different mean.
LDA can classify them perfectly.
LDA can NOT classify them perfectly.
LDA is not applicable in data with these properties
Insufficient information

Answer: Insufficient information


These are Introduction to Machine Learning Nptel Week 3 Answers


Q4. Given the following distribution of data points:
What method would you choose to perform Dimensionality Reduction?

Linear Discriminant Analysis
Principal Component Analysis
Both LDA and/or PCA.
None of the above.

Answer: Linear Discriminant Analysis


These are Introduction to Machine Learning Nptel Week 3 Answers


Q5. If log(1−p(x)1+p(x))=β0+βx
What is p(x)?

a. p(x)=1+eβ0+βxeβ0+βx
b. p(x)=1+eβ0+βx1−eβ0+βx
c. p(x)=eβ0+βx1+eβ0+βx
d. p(x)=1−eβ0+βx1+eβ0+βx

Answer: d. p(x)=1−eβ0+βx1+eβ0+βx


These are Introduction to Machine Learning Nptel Week 3 Answers


Q6. For the two classes ’+’ and ’-’ shown below.
While performing LDA on it, which line is the most appropriate for projecting data points?

Red
Orange
Blue
Green

Answer: Blue


These are Introduction to Machine Learning Nptel Week 3 Answers


Q7. Which of these techniques do we use to optimise Logistic Regression:
Least Square Error
Maximum Likelihood
(a) or (b) are equally good
(a) and (b) perform very poorly, so we generally avoid using Logistic Regression
None of these

Answer: Maximum Likelihood


These are Introduction to Machine Learning Nptel Week 3 Answers


Q8. LDA assumes that the class data is distributed as:
Poisson
Uniform
Gaussian
LDA makes no such assumption.

Answer: Gaussian


These are Introduction to Machine Learning Nptel Week 3 Answers


Q9. Suppose we have two variables, X and Y (the dependent variable), and we wish to find their relation. An expert tells us that relation between the two has the form Y=meX+c. Suppose the samples of the variables X and Y are available to us. Is it possible to apply linear regression to this data to estimate the values of m and c?
No.
Yes.
Insufficient information.
None of the above.

Answer: Yes.


These are Introduction to Machine Learning Nptel Week 3 Answers


Q10. What might happen to our logistic regression model if the number of features is more than the number of samples in our dataset?
It will remain unaffected
It will not find a hyperplane as the decision boundary
It will over fit
None of the above

Answer: It will over fit


These are Introduction to Machine Learning Nptel Week 3 Answers

More Weeks of INTRODUCTION TO MACHINE LEARNING: Click here

More Nptel Courses: Click here


Introduction to Machine Learning Nptel Week 3 Answers (Jan-Apr 2023)

Course Name: Introduction to Machine Learning

Course Link: Click Here

These are Introduction to Machine Learning Nptel Week 3 Answers


Q1. Which of the following is false about a logistic regression based classifier?
a. The logistic function is non-linear in the weights
b. The logistic function is linear in the weights
c. The decision boundary is non-linear in the weights
d. The decision boundary is linear in the weights

Answer: b, c


Q2. Consider the case where two classes follow Gaussian distribution which are cen- tered at (3, 9) and (−3, 3) and have identity covariance matrix. Which of the following is the separating decision boundary using LDA assuming the priors to be equal?
a. y−x=3
b. x+y=3
c. x+y=6
d. both (b) and (c)
e. None of the above
f. Can not be found from the given information

Answer: c. x+y=6


These are Introduction to Machine Learning Nptel Week 3 Answers


Q3. Consider the following relation between a dependent variable and an independent variable identified by doing simple linear regression. Which among the following relations between the two variables does the graph indicate?

image 11

a. as the independent variable increases, so does the dependent variable
b. as the independent variable increases, the dependent variable decreases
c. if an increase in the value of the dependent variable is observed, then the independent variable will show a corresponding increase
d. if an increase in the value of the dependent variable is observed, then the independent variable will show a corresponding decrease
e. the dependent variable in this graph does not actually depend on the independent variable
f. none of the above

Answer: e. the dependent variable in this graph does not actually depend on the independent variable


Q4. Given the following distribution of data points:

image 12

What method would you choose to perform Dimensionality Reduction?
a. Linear Discriminant Analysis
b. Principal Component Analysis

Answer: a. Linear Discriminant Analysis


These are Introduction to Machine Learning Nptel Week 3 Answers


Q5. In general, which of the following classification methods is the most resistant to gross outliers?
a. Quadratic Discriminant Analysis (QDA)
b. Linear Regression
c. Logistic regression
d. Linear Discriminant Analysis (LDA)

See also  Introduction to Machine Learning | Week 6

Answer: c. Logistic regression


These are Introduction to Machine Learning Nptel Week 3 Answers


Q6. Suppose that we have two variables, X and Y (the dependent variable). We wish to find the relation between them. An expert tells us that relation between the two has the form Y=m+X2+c. Available to us are samples of the variables X and Y. Is it possible to apply linear regression to this data to estimate the values of m and c?
a. no
b. yes
c. insufficient information

Answer: a. no


These are Introduction to Machine Learning Nptel Week 3 Answers


Q7. In a binary classification scenario where x is the independent variable and y is the dependent variable, logistic regression assumes that the conditional distribution y|x follows a
a. Bernoulli distribution
b. binomial distribution
c. normal distribution
d. exponential distribution

Answer: a. Bernoulli distribution


These are Introduction to Machine Learning Nptel Week 3 Answers


Q8. Consider the following data:

image 13

Assuming that you apply LDA to this data, what is the estimated covariance matrix?
a. [1.8750.31250.31250.9375]
b. [2.50.41670.41671.25]
c. [1.8750.31250.31251.2188]
d. [2.50.41670.41671.625]
e. [3.251.16671.16672.375]
f. [2.43750.8750.8751.7812]
g. None of these

Answer: g. None of these


Q9. Given the following 3D input data, identify the principal component.

image 14

(Steps: center the data, calculate the sample covariance matrix, calculate the eigenvectors and eigenvalues, identify the principal component)
a. [−0.1022 0.0018 0.9948]
b. [0.5742 −0.8164 0.0605]
c. [0.5742 0.8164 0.0605]
d. [−0.5742 0.8164 0.0605]
e. [0.8123 0.5774 0.0824]
f. None of the above

Answer: e. [0.8123 0.5774 0.0824]


Q10. For the data given in the previous question, find the transformed input along the first two principal components.

a.

image 15

b.

image 16

c.

image 17

d.

image 18

e. None of the above

Answer: c


These are Introduction to Machine Learning Nptel Week 3 Answers

More Weeks of Introduction to Machine Learning: Click Here

More Nptel courses: https://progiez.com/nptel


Introduction to Machine Learning Nptel Week 3 Answers (Jul-Dec 2022)

These are Introduction to Machine Learning Nptel Week 3 Answers

Course Name: INTRODUCTION TO MACHINE LEARNING

Link to Enroll: Click Here


Q1. For linear classification we use:
a. A linear function to separate the classes.
b. A linear function to model the data.
c. A linear loss.
d. Non-linear function to fit the data.

Answer: b. A linear function to model the data.


Q2. Logit transformation for Pr(X=1) for given data is S=[0,1,1,0,1,0,1]
a. 3/4
b. 4/3
c. 4/7
d. 3/7

Answer: c. 4/7


These are Introduction to Machine Learning Nptel Week 3 Answers


Q3. The output of binary class logistic regression lies in this range.
a. [−∞,∞]
b. [−1,1]
c. [0,1]
d. [−∞,0]

Answer: d. [−∞,0]


These are Introduction to Machine Learning Nptel Week 3 Answers


Q4. If log(1−p(x)1+p(x))=β0+βxlog What is p(x)p(x)?

Answer: d.


Q5. Logistic regression is robust to outliers. Why?
a. The squashing of output values between [0, 1] dampens the affect of outliers.
b. Linear models are robust to outliers.
c. The parameters in logistic regression tend to take small values due to the nature of the problem setting and hence outliers get translated to the same range as other samples.
d. The given statement is false.

Answer: d. The given statement is false.


These are Introduction to Machine Learning Nptel Week 3 Answers


Q6. Aim of LDA is (multiple options may apply)
a. Minimize intra-class variability.
b. Maximize intra-class variability.
c. Minimize the distance between the mean of classes
d. Maximize the distance between the mean of classes

Answer: b. Maximize intra-class variability.


Q7. We have two classes in our dataset with mean 0 and 1, and variance 2 and 3.
a. LDA may be able to classify them perfectly.
b. LDA will definitely be able to classify them perfectly.
c. LDA will definitely NOT be able to classify them perfectly.
d. None of the above.

Answer: a. LDA may be able to classify them perfectly.


These are Introduction to Machine Learning Nptel Week 3 Answers


Q8. We have two classes in our dataset with mean 0 and 5, and variance 1 and 2.
a. LDA may be able to classify them perfectly.
b. LDA will definitely be able to classify them perfectly.
c. LDA will definitely NOT be able to classify them perfectly.
d. None of the above.

Answer: b. LDA will definitely be able to classify them perfectly.


Q9. For the two classes ’+’ and ’-’ shown below.
While performing LDA on it, which line is the most appropriate for projecting data points?

a. Red
b. Orange
c. Blue
d. Green

Answer: b. Orange


These are Introduction to Machine Learning Nptel Week 3 Answers


Q10. LDA assumes that the class data is distributed as:
a. Poisson
b. Uniform
c. Gaussian
d. LDA makes no such assumption.

Answer: d. LDA makes no such assumption.


These are Introduction to Machine Learning Nptel Week 3 Answers