Here are 1000 MCQs on Machine Learning (Chapterwise).
1. What is Machine learning?
a) The selective acquisition of knowledge through the use of computer programs
b) The selective acquisition of knowledge through the use of manual programs
c) The autonomous acquisition of knowledge through the use of computer programs
d) The autonomous acquisition of knowledge through the use of manual programs
View Answer
Explanation: Machine learning is the autonomous acquisition of knowledge through the use of computer programs.
2. K-Nearest Neighbors (KNN) is classified as what type of machine learning algorithm?
a) Instance-based learning
b) Parametric learning
c) Non-parametric learning
d) Model-based learning
View Answer
Explanation: KNN doesn’t build a parametric model of the data. Instead, it directly classifies new data points based on the k nearest points in the training data.
3. Which of the following is not a supervised machine learning algorithm?
a) K-means
b) Naïve Bayes
c) SVM for classification problems
d) Decision tree
View Answer
Explanation: Decision tree, SVM (Support vector machines) for classification problems and Naïve Bayes are the examples of supervised machine learning algorithm. K-means is an example of unsupervised machine learning algorithm.
4. What’s the key benefit of using deep learning for tasks like recognizing images?
a) They need less training data than other methods.
b) They’re easier to explain and understand than other models.
c) They can learn complex details from the data on their own.
d) They work faster and are more efficient computationally.
View Answer
Explanation: Deep learning is great at figuring out intricate details from data, especially in tasks like recognizing images.
5. Which algorithm is best suited for a binary classification problem?
a) K-nearest Neighbors
b) Decision Trees
c) Random Forest
d) Linear Regression
View Answer
Explanation: Decision Trees are versatile and can be used for classification problems, particularly for binary classification, where the output is divided into two classes.
6. What is the key difference between supervised and unsupervised learning?
a) Supervised learning requires labeled data, while unsupervised learning does not.
b) Supervised learning predicts labels, while unsupervised learning discovers patterns.
c) Supervised learning is used for classification, while unsupervised learning is used for regression.
d) Supervised learning is always more accurate than unsupervised learning.
View Answer
Explanation: The presence or absence of labeled data in the training set distinguishes supervised and unsupervised learning approaches.
7. Which type of machine learning algorithm falls under the category of “unsupervised learning”?
a) Linear Regression
b) K-means Clustering
c) Decision Trees
d) Random Forest
View Answer
Explanation: K-means Clustering is an example of unsupervised learning used for clustering unlabeled data based on similarities.
8. Which of the following statements is true about AdaBoost?
a) It is particularly prone to overfitting on noisy datasets
b) Complexity of the weak learner is important in AdaBoost
c) It is generally more prone to overfitting
d) It improves classification accuracy
View Answer
Explanation: AdaBoost is generally not more prone to overfitting but is less prone to overfitting. And it is prone to overfitting on noisy datasets. If you use very simple weak learners, then the algorithms are much less prone to overfitting and it improves classification accuracy. So Complexity of the weak learner is important in AdaBoost.
9. Which one of the following models is a generative model used in machine learning?
a) Support vector machines
b) Naïve Bayes
c) Logistic Regression
d) Linear Regression
View Answer
Explanation: Naïve Bayes is a type of generative model which is used in machine learning. Linear Regression, Logistic Regression and Support vector machines are the types of discriminative models which are used in machine learning.
10. An artificially intelligent car decreases its speed based on its distance from the car in front of it. Which algorithm is used?
a) Naïve-Bayes
b) Decision Tree
c) Linear Regression
d) Logistic Regression
View Answer
Explanation: The output is numerical. It determines the speed of the car. Hence it is not a classification problem. All the three, decision tree, naïve-Bayes, and logistic regression are classification algorithms. Linear regression, on the other hand, outputs numerical values based on input. So, this can be used.
11. Which of the following statements is false about Ensemble learning?
a) It is a supervised learning algorithm
b) It is an unsupervised learning algorithm
c) More random algorithms can be used to produce a stronger ensemble
d) Ensembles can be shown to have more flexibility in the functions they can represent
View Answer
Explanation: Ensemble learning is not an unsupervised learning algorithm. It is a supervised learning algorithm that combines several machine learning techniques into one predictive model to decrease variance and bias. It can be trained and then used to make predictions. And this ensemble can be shown to have more flexibility in the functions they can represent.
12. Which of the following statements is true about stochastic gradient descent?
a) It processes one training example per iteration
b) It is not preferred, if the number of training examples is large
c) It processes all the training examples for each iteration of gradient descent
d) It is computationally very expensive, if the number of training examples is large
View Answer
Explanation: Stochastic gradient descent processes one training example per iteration. That is it updates the weight vector based on one data point at a time. All other three are the features of Batch Gradient Descent.
13. Decision tree uses the inductive learning machine learning approach.
a) False
b) True
View Answer
Explanation: Decision tree uses the inductive learning machine learning approach. Inductive learning enables the system to recognize patterns and regularities in previous knowledge or training data and extract the general rules from them. A decision tree is considered to be an inductive learning task as it uses particular facts to make more generalized conclusions.
14. What elements describe the Candidate-Elimination algorithm?
a) depends on the dataset
b) just a set of candidate hypotheses
c) just a set of instances
d) set of instances, set of candidate hypotheses
View Answer
Explanation: A set of instances is required. A set of candidate hypotheses are given. These are applied to the training data and the list of accurate hypotheses is output in accordance with the candidate-elimination algorithm.
15. Which of the following statements is not true about boosting?
a) It mainly increases the bias and the variance
b) It tries to generate complementary base-learners by training the next learner on the mistakes of the previous learners
c) It is a technique for solving two-class classification problems
d) It uses the mechanism of increasing the weights of misclassified data in preceding classifiers
View Answer
Explanation: Boosting does not increase the bias and variance but it mainly reduces the bias and the variance. It is a technique for solving two-class classification problems. And it tries to generate complementary base-learners by training the next learner (by increasing the weights) on the mistakes (misclassified data) of the previous learners.
Chapterwise Multiple Choice Questions on Machine Learning
- Formal Learning Model
- Version Spaces
- VC-Dimension
- Linear Regression
- Multivariate Linear Regression
- Logistic Regression
- Ensemble Learning
- Stochastic Gradient Descent
- Kernels
- Support Vector Machines
- Decision Trees
- Nearest Neighbor
- Naive-Bayes Algorithm
- Machine Learning – Neural Networks
1. MCQ on Formal Learning Model
The section contains multiple choice questions and answers on statistical learning framework, empirical minimization framework and PAC learning.
|
|
2. Machine Learning MCQ on Version Spaces
The section contains questions and answers on version spaces, find-s algorithm and candidate elimination algorithm.
|
|
3. Machine Learning Multiple Choice Questions on VC-Dimension
The section contains Machine Learning MCQs on VC-dimension and the Fundamental Theorem of PAC Learning.
|
|
4. MCQ on Linear Regression
The section contains Machine Learning multiple choice questions and answers on linear regression in machine learning, linear regression cost functions, and gradient descent.
|
|
5. Machine Learning MCQ on Multivariate Linear Regression
The section contains Machine Learning questions and answers on multivariate linear regression, gradient descent for multiple variables, and polynomial regression.
|
|
6. Machine Learning Multiple Choice Questions on Logistic Regression
This section features MCQs on logistic regression, hypothesis representation, decision boundary, cost function and gradient descent, logistic regression for multiple classification, and advanced optimization.
7. MCQ on Ensemble Learning
The section contains multiple choice questions and answers on ensemble learning, covering error-correcting output codes, model combination schemes, boosting weak learnability, the AdaBoost algorithm, and stacking.
|
|
8. Machine Learning MCQ on Stochastic Gradient Descent
The section contains questions and answers on optimization algorithms, specifically focusing on Stochastic Gradient Descent (SGD), its variants, the standard Gradient Descent Algorithm, and Subgradient Descent.
|
|
9. Machine Learning Multiple Choice Questions on Kernels
The section contains Machine Learning MCQs on kernels and kernel trick.
|
|
10. MCQ on Support Vector Machines
The section contains multiple choice questions and answers on support vector machines (SVMs), covering key concepts like the large margin intuition, margins and hard/soft SVMs, norm regularization, optimality conditions and support vectors, and finally, implementing soft SVMs using Stochastic Gradient Descent (SGD).
|
|
11. Machine Learning MCQ on Decision Trees
The section contains questions and answers on decision trees, covering core concepts such as decision tree pruning, inductive bias, classification trees, regression trees, and the powerful Random Forest algorithm.
12. Machine Learning Multiple Choice Questions on Nearest Neighbor
The section contains MCQs on K-Nearest Neighbor Algorithm and Nearest Neighbor Analysis.
|
|
13. MCQ on Naive-Bayes Algorithm
The section contains multiple choice questions and answers on Naive-Bayes Algorithm.
|
14. Neural Networks in Machine Learning
The section contains multiple choice questions and answers on nonlinear hypothesis, neurons and the brain, model representation, multiclass classification, cost function, gradient checking, and random initialization.
Wish you the best in your endeavor to learn and master Machine Learning!
If you find a mistake in question / option / answer, kindly take a screenshot and email to [email protected]