On This Page

This set of Machine Learning (ML) Multiple Choice Questions & Answers (MCQs) focuses on Machine Learning Set 17

Q1 | What are the two methods used for the calibration in Supervised Learning?
  • platt calibration and isotonic regression
  • statistics and informal retrieval
Q2 | Which of the following are several models for feature extraction
  • regression
  • classification
  • none of the above
Q3 | Lets say, a Linear regression model perfectly fits the training data (train error
  • a. you will always have test error zero
  • b. you can not have test error zero
  • c. none of the above
Q4 | Which of the one is true about Heteroskedasticity?
  • a. linear regression with varying error terms
  • b. linear regression with constant error terms
  • c. linear regression with zero error terms
  • d. none of these
Q5 | Which of the following assumptions do we make while deriving linear regression parameters?1. The true relationship between dependent y and predictor x is linear2. The model errors are statistically independent3. The errors are normally distributed with a 0 mean and constant standard deviation4. The predictor x is non-stochastic and is measured error-free
  • a. 1,2 and 3.
  • b. 1,3 and 4.
  • c. 1 and 3.
  • d. all of above.
Q6 | To test linear relationship of y(dependent) and x(independent) continuous variables, which of the following plot best suited?
  • a. scatter plot
  • b. barchart
  • c. histograms
  • d. none of these
Q7 | Generally, which of the following method(s) is used for predicting continuous dependent variable?1. Linear Regression2. Logistic Regression
  • a. 1 and 2
  • b. only 1
  • c. only 2
  • d. none of these.
Q8 | Suppose you are training a linear regression model. Now consider these points.1. Overfitting is more likely if we have less data2. Overfitting is more likely when the hypothesis space is small.Which of the above statement(s) are correct?
  • a. both are false
  • b. 1 is false and 2 is true
  • c. 1 is true and 2 is false
  • d. both are true
Q9 | Suppose we fit Lasso Regression to a data set, which has 100 features (X1,X2X100). Now, we rescale one of these feature by multiplying with 10 (say that feature is X1), and then refit Lasso regression with the same regularization parameter.Now, which of the following option will be correct?
  • a. it is more likely for x1 to be excluded from the model
  • b. it is more likely for x1 to be included in the model
  • c. cant say
  • d. none of these
Q10 | Which of the following is true aboutRidge or Lasso regression methods in case of feature selection?
  • a. ridge regression uses subset selection of features
  • b. lasso regression uses subset selection of features
  • c. both use subset selection of features
  • d. none of above
Q11 | Which of the following statement(s) can
  • a. 1 and 2
  • b. 1 and 3
  • c. 2 and 4
  • d. none of the above
Q12 | We can also compute the coefficient of linear regression with the help of an analytical method called Normal Equation. Which of the following is/are true about Normal Equation?1. We dont have to choose the learning rate2. It becomes slow when number of features is very large3. No need to iterate
  • a. 1 and 2
  • b. 1 and 3.
  • c. 2 and 3.
  • d. 1,2 and 3.
Q13 | If two variables are correlated, is it necessary that they have a linear relationship?
  • a. yes
  • b. no
Q14 | Correlated variables can have zero correlation coeffficient. True or False?
  • a. true
  • b. false
Q15 | Which of the following option is true regarding Regression andCorrelation ?Note: y is dependent variable and x is independent variable.
  • a. the relationship is symmetric between x and y in both.
  • b. the relationship is not symmetric between x and y in both.
  • c. the relationship is not symmetric between x and y in case of correlation but in case of regression it is symmetric.
  • d. the relationship is symmetric between x and y in case of correlation but in case of regression it is not symmetric.
Q16 | Suppose you are using a Linear SVM classifier with 2 class classification
  • yes
  • no
Q17 | If you remove the non-red circled points from the data, the decision boundary will change?
  • true
  • false
Q18 | When the C parameter is set to infinite, which of the following holds true?
  • the optimal hyperplane if exists, will be the one that completely separates the data
  • the soft-margin classifier will separate the data
  • none of the above
Q19 | Suppose you are building a SVM model on data X. The data X can be error prone which means that you should not trust any specific data point too much. Now think that you want to build a SVM model which has quadratic kernel function of polynomial degree 2 that uses Slack variable C as one of its hyper parameter.What would happen when you use very large value of C(C->infinity)?
  • we can still classify data correctly for given setting of hyper parameter c
  • we can not classify data correctly for given setting of hyper parameter c
  • cant say
  • none of these
Q20 | SVM can solvelinearand non- linearproblems
  • true
  • false
Q21 | The objective of the support vector machine algorithm is to find a hyperplane in an N-dimensional space(N the number of features) that distinctly classifies the data points.
  • true
  • false
Q22 | Hyperplanes are                        boundaries that help classify the data points.
  • usual
  • decision
  • parallel
Q23 | The          of the hyperplane depends upon the number of features.
  • dimension
  • classification
  • reduction
Q24 | Hyperplanes are decision boundaries that help classify the data points.
  • true
  • false