Home
MCQS
Machine Learning (ML) MCQ Quiz Hub
Machine Learning (ML) MCQ Set 10
Choose a topic to test your knowledge and improve your Machine Learning (ML) skills
1. A supervised scenario is characterized by the concept of a _____.
Programmer
Teacher
Author
Farmer
2. overlearning causes due to an excessive ______.
Capacity
Regression
Reinforcement
Accuracy
3. Which of the following are several models for feature extraction
classification
regression
Both A and B
None of the above
4. _ provides some built-in datasets that can be used for testing purposes.
scikit-learn
classification
regression
None of the above
5. While using _____ all labels areturned into sequential numbers.
LabelEncoder class
LabelBinarizer class
DictVectorizer
FeatureHasher
6. ___produce sparse matrices of real numbers that can be fed into any machine learning model.
DictVectorizer
FeatureHasher
Both A & B
None of the Mentioned
7. scikit-learn offers the class______, which is responsible for filling the holes using a strategy based on the mean, median, or frequency
LabelEncoder
LabelBinarizer
DictVectorizer
Imputer
8. scikit-learn also provides a class for per-sample normalization,_____
Normalizer
Imputer
Classifier
All of the above
9. _____dataset with many features contains information proportional to the independence of all features and their variance.
normalized
unnormalized
Both A and B
None of the Mentioned
10. In order to assess how much information is brought by each component, and the correlation among them, a useful tool is the_____.
Concuttent matrix
Convergance matrix
Supportive matrix
Covariance matrix
11. The_____ parameter can assume different values which determine how the data matrix is initially processed.
run
start
init
stop
12. Which of the following statement is true about outliers in Linear regression?
Linear regression is sensitive to outliers
Linear regression is not sensitive to outliers
Can’t say
None of these
13. Let’s say, a “Linear regression” model perfectly fits the training data (train error is zero). Now, Which of the following statement is true?
You will always have test error zero
You can not have test error zero
both (a) and (b)
None of the above
14. In a linear regression problem, we are using “R-squared” to measure goodness-of-fit. We add a feature in linear regression model and retrain the same model.Which of the following option is true?
If R Squared increases, this variable is significant.
If R Squared decreases, this variable is not significant.
Individually R squared cannot tell about variable importance. We can’t say anything about it right now.
None of the above
15. To test linear relationship of y(dependent) and x(independent) continuous variables, which of the following plot best suited?
Scatter plot B. C.
Barchart
Histograms
None of these
16. which of the following step / assumption in regression modeling impacts the trade-off between under-fitting and over-fitting the most.
The polynomial degree
Whether we learn the weights by matrix inversion or gradient descent
The use of a constant-term
None of these
17. Which of the following is true about “Ridge” or “Lasso” regression methods in case of feature selection?
Ridge regression uses subset selection of features B.
Lasso regression uses subset selection of features
Both use subset selection of features
None of above
18. Which of the following statement(s) can be true post adding a variable in a linear regression model?1. R-Squared and Adjusted R-squared both increase2. R-Squared increases and Adjusted R-squared decreases3. R-Squared decreases and Adjusted R-squared decreases4. R-Squared decreases and Adjusted R-squared increases
1 and 2
1 and 3
2 and 4
None of the above
19. What is/are true about kernel in SVM?1. Kernel function map low dimensional data to high dimensional space2. It’s a similarity function
1
2
1 and 2
None of these
20. Suppose you are building a SVM model on data X. The data X can be error prone which means that you should not trust any specific data point too much. Now think that you want to build a SVM model which has quadratic kernel function of polynomial degree 2 that uses Slack variable C as one of it’s hyper parameter.What would happen when you use very small C (C~0)?
Misclassification would happen
Data will be correctly classified
Can’t say
None of these
21. The cost parameter in the SVM means:
The number of cross-validations to be made
The kernel to be used
The tradeoff between misclassification and simplicity of the model
None of the above
22. How do you handle missing or corrupted data in a dataset?
Drop missing rows or columns
Replace missing values with mean/median/mode
Assign a unique category to missing values
All of the above
23. Which of the following statements about Naive Bayes is incorrect?
Attributes are equally important.
Attributes are statistically dependent of one another given the class value. C. C.
Attributes are statistically independent of one another given the class value.
Attributes can be nominal or numeric
24. The SVM’s are less effective when:
The data is linearly separable
The data is clean and ready to use
The data is noisy and contains overlapping points
None of the above
25. If there is only a discrete number of possible outcomes called _____.
Modelfree
Categories
Prediction
None of the above
26. Some people are using the term ___ instead of prediction only to avoid the weird idea that machine learning is a sort of modern magic.
Inference
Interference
Accuracy
None of the above
27. The term _____ can be freely used, but with the same meaning adopted in physics or system theory.
Accuracy
Cluster
Regression
Prediction
28. Common deep learning applications / problems can also be solved using____
Real-time visual object identification
Classic approaches
Automatic labeling
Bio-inspired adaptive systems
29. what is the function of ‘Unsupervised Learning’?
Find clusters of the data and find low-dimensional representations of the data
Find interesting directions in data and find novel observations/ database cleaning
Interesting coordinates and correlations
All
30. In a linear regression problem, we are using “R-squared” to measure goodness-of-fit. We add a feature in linear regression model and retrain the same model.Which of the following option is true?
If R Squared increases, this variable is significant.
If R Squared decreases, this variable is not significant.
Individually R squared cannot tell about variable importance. We can’t say anything about it right now.
None of These
31. Suppose we fit “Lasso Regression” to a data set, which has 100 features (X1,X2…X100). Now, we rescale one of these feature by multiplying with 10 (say that feature is X1), and then refit Lasso regression with the same regularization parameter.Now, which of the following option will be correct?
It is more likely for X1 to be excluded from the model
It is more likely for X1 to be included in the model
Can’t say
None of These
32. Which of the following is true about “Ridge” or “Lasso” regression methods in case of feature selection?
Ridge regression uses subset selection of features
Lasso regression uses subset selection of features
Both use subset selection of features
None of these
33. We can also compute the coefficient of linear regression with the help of an analytical method called “Normal Equation”. Which of the following is/are true about “Normal Equation”?1. We don’t have to choose the learning rate2. It becomes slow when number of features is very large3. No need to iterate
1 and 2
1 and 3.
2 and 3.
1,2 and 3.
34. Which of the following option is true regarding “Regression” and “Correlation” ?Note: y is dependent variable and x is independent variable.
The relationship is symmetric between x and y in both.
The relationship is not symmetric between x and y in both.
The relationship is not symmetric between x and y in case of correlation but in case of regression it is symmetric.
The relationship is symmetric between x and y in case of correlation but in case of regression it is not symmetric.
35. Suppose you are building a SVM model on data X. The data X can be error prone which means that you should not trust any specific data point too much. Now think that you want to build a SVM model which has quadratic kernel function of polynomial degree 2 that uses Slack variable C as one of it’s hyper parameter.What would happen when you use very large value of C(C->infinity)?
We can still classify data correctly for given setting of hyper parameter C
We can not classify data correctly for given setting of hyper parameter
Can’t Say
None of these
36. Hyperplanes are _____________boundaries that help classify the data points.
usual
decision
parallel
None of These
37. The _____of the hyperplane depends upon the number of features.
dimension
classification
reduction
None of these
38. What is the purpose of performing cross-validation?
To assess the predictive performance of the models
To judge how the trained model performs outside the sample on test data C. c.
Both A and B
None of These
39. Which of the following is true about Naive Bayes ?
Assumes that all the features in a dataset are equally important
Assumes that all the features in a dataset are independent
both (a) and (b)
None of the above option
40. Which of the following is not supervised learning?
PCA
Decision Tree
Naive Bayesian
Linerar regression
41. ___can be adopted when it's necessary to categorize a large amount of data with a few complete examples or when there's the need to impose some constraints to a clustering algorithm.
Supervised
Semi-supervised
Reinforcement
Clusters
42. In reinforcement learning, this feedback is usually called as___.
Overfitting
Overlearning
Reward
None of the above
43. In the last decade, many researchers started training bigger and bigger models, built with several different layers that's why this approach is called_____.
Deep learning
Machine learning
Reinforcement learning
Unsupervised learning
44. Suppose we fit “Lasso Regression” to a data set, which has 100 features (X1,X2…X100). Now, we rescale one of these feature by multiplying with 10 (say that feature is X1), and then refit Lasso regression with the same regularization parameter.Now, which of the following option will be correct?
It is more likely for X1 to be excluded from the model
It is more likely for X1 to be included in the model
Can’t say
None of these
45. . If Linear regression model perfectly first i.e., train error is zero, then ____________
Test error is also always zero
Test error is non zero
Couldn’t comment on Test error
Test error is equal to Train error
46. In syntax of linear model lm(formula,data,..), data refers to ______
Matrix
Vector
Array
List
47. Which of the following option is true regarding “Regression” and “Correlation” ?Note: y is dependent variable and x is independent variable.
The relationship is symmetric between x and y in both.
The relationship is not symmetric between x and y in both.
The relationship is not symmetric between x and y in case of correlation but in case of regression it is symmetric.
The relationship is symmetric between x and y in case of correlation but in case of regression it is not symmetric.
48. Let’s say, you are working with categorical feature(s) and you have not looked at the distribution of the categorical variable in the test data. You want to apply one hot encoding (OHE) on the categorical feature(s). What challenges you may face if you have applied OHE on a categorical variable of train dataset?
All categories of categorical variable are not present in the test dataset.
Frequency distribution of categories is different in train as compared to the test dataset.
Train and Test always have same distribution.
Both A and B
49. __which can accept a NumPy RandomState generator or an integer seed.
make_blobs
random_state
test_size
None of these
50. . In many classification problems, the target dataset is made up of categorical labels which cannot immediately be processed by any algorithm. An encoding is needed and scikit-learn offers at least_____valid options
1
2
3
4
51. _is the most drastic one and should be considered only when the dataset is quite large, the number of missing features is high, and any prediction could be risky.
Removing the whole line
Creating sub-model to predict those features
Using an automatic strategy to input them according to the other known values
All above
52. Suppose you have fitted a complex regression model on a dataset. Now, you are using Ridge regression with tuning parameter lambda to reduce its complexity. Choose the option(s) below which describes relationship of bias and variance with lambda.
In case of very large lambda; bias is low, variance is low
In case of very large lambda; bias is low, variance is high
In case of very large lambda; bias is high, variance is low
In case of very large lambda; bias is high, variance is high
53. Function used for linear regression in R is __________
lm(formula, data)
lr(formula, data)
lrm(formula, data)
regression.linear(formula, data)
54. Suppose, you got a situation where you find that your linear regression model is under fitting the data. In such situation which of the following options would you consider?1. I will add more variables2. I will start introducing polynomial degree variables3. I will remove some variables
1 and 2
2 and 3
1 and 3
1, 2 and 3
55. We usually use feature normalization before using the Gaussian kernel in SVM. What is true about feature normalization? 1. We do feature normalization so that new feature will dominate other 2. Some times, feature normalization is not feasible in case of categorical variables3. Feature normalization always helps when we use Gaussian kernel in SVM
1
1 and 2
1 and 3
2 and 3
Submit