ML Assignment 7
ML Assignment 7
1. Which of the following option is / are correct regarding the benefits of ensemble model?
1. Better performance
2. More generalized model
3. Better interpretability
A) 1 and 3
B) 2 and 3
C) 1 and 2
D) 1, 2 and 3
Answer: C
(1 and 2 are the benefits of ensemble modelling. Option 3 is incorrect because when we
ensemble multiple models, we lose interpretability of the models).
Answer: B, C
Outliers tend to get misclassified. As the number of iterations increase, the weight
corresponding to outlier points can become very large resulting in subsequent classifier
models trying to classify the outlier points correctly. This generally has an adverse effect on
the overall classifier. Restricting the weights is one way of mitigating this problem.
However, this can also lower the performance of the classifier.
3. Which among the following are some of the differences between bagging and boosting?
NPTEL Online Certification Courses
Indian Institute of Technology Kharagpur
A) In bagging we use the same classification algorithm for training on each sample
of the data, whereas in boosting, we use different classification algorithms on the
different training data samples.
B) Bagging is easy to parallelize whereas boosting is inherently a sequential
process.
C) In bagging we typically use sampling with replacement whereas in boosting,
we typically use weighted sampling techniques.
D) In comparison with the performance of a base classifier on a particular
dataset, bagging will generally not increase the error whereas as boosting may
leadto an increase in the error.
A) 3
B) 4
C) 5
D) 6
5. Considering the AdaBoost algorithm, which among the following statements is true?
Answer: B, D
The classifier chosen at each stage is the one that minimizes the weighted error at that
stage. The weight of a point is high if it has been misclassified more number of times in
the previous iterations. Thus, maximum error minimization is performed by trying to
correctly predict the points which were misclassified in earlier iterations. Also, weights
are assigned to the classifiers depending upon their accuracy which again depends upon
the weighted error (for that classifier).
NPTEL Online Certification Courses
Indian Institute of Technology Kharagpur
6. Suppose the VC dimension of a hypothesis space is 6. Which of the following are true?
7. Ensembles will yield bad results when there is a significant diversity among the models.
Write True or False.
A) True
B) False
Answer: B
Ensemble is a collection of diverse set of learners to improve the stability and the
performance of the algorithm. So, more diverse the models are, the better will be the
performance of ensemble.
Answer: D.
Decision trees do not aggregate the results of multiple trees, so it is not an ensemble
algorithm.
9. Which of the following can be true for selecting base learners for an ensemble?
A) Different learners can come from same algorithm with different hyper parameters
B) Different learners can come from different algorithms
C) Different learners can come from different training spaces
D) All of the above.
Answer: D
We can create an ensemble by following any / all of the options mentioned above.
10. Generally, an ensemble method works better, if the individual base models have
____________?
Note: Individual models have accuracy greater than 50%
A) Less correlation among predictions
B) High correlation among predictions