machine-learning banner

Machine Learning Multiple Choice Questions (MCQs) and Answers

Master Machine Learning with Practice MCQs. Explore our curated collection of Multiple Choice Questions. Ideal for placement and interview preparation, our questions range from basic to advanced, ensuring comprehensive coverage of machine learning concepts. Begin your placement preparation journey now!

Q61

Q61 Which function in sklearn is used to implement logistic regression?

A

log_regression()

B

LogisticRegression()

C

reg_log()

D

log_reg()

Q62

Q62 How do you calculate the precision score in a classification task using sklearn?

A

precision_score()

B

precision()

C

calc_precision()

D

precision_calc()

Q63

Q63 Which library is used for implementing decision trees in Python?

A

numpy

B

pandas

C

sklearn

D

matplotlib

Q64

Q64 A classification model has high accuracy but low recall. What does this indicate?

A

Model is overfitting

B

Model has high false negatives

C

Model has high false positives

D

Data is imbalanced

Q65

Q65 A logistic regression model is overfitting on training data. What can be done to mitigate this?

A

Reduce the training data size

B

Increase the number of features

C

Apply regularization

D

Use higher learning rate

Q66

Q66 A decision tree model has a very high depth and performs poorly on test data. What could be the reason?

A

Overfitting

B

Underfitting

C

Insufficient training data

D

Wrong cost function

Q67

Q67 What is a decision tree in machine learning?

A

A model for regression tasks

B

A model for clustering

C

A tree-like structure for decision making

D

A method for dimensionality reduction

Q68

Q68 What is the purpose of pruning in decision trees?

A

To increase the depth of the tree

B

To reduce the size of the tree

C

To improve training speed

D

To prevent overfitting

Q69

Q69 How does a decision tree split data at each node?

A

By maximizing accuracy

B

By minimizing distance between data points

C

By maximizing information gain

D

By using random splits

Q70

Q70 What is the main advantage of using Random Forest over a single Decision Tree?

A

It reduces the complexity of the model

B

It reduces overfitting

C

It increases the depth of trees

D

It uses fewer features

Q71

Q71 What is the role of entropy in decision trees?

A

It measures the homogeneity of data

B

It measures model accuracy

C

It measures the distance between data points

D

It calculates the tree depth

Q72

Q72 Which function in sklearn is used to implement decision trees?

A

DecisionTree()

B

TreeDecision()

C

DecisionTreeClassifier()

D

ClassifierTree()

Q73

Q73 How can you visualize a decision tree in Python using sklearn?

A

visualize_tree()

B

plot_tree()

C

draw_tree()

D

show_tree()

Q74

Q74 Which sklearn function is used to implement a Random Forest Classifier?

A

RandomForest()

B

ForestClassifier()

C

RandomForestClassifier()

D

RandomClass()

Q75

Q75 How does the max_depth parameter in a decision tree model affect its performance?

A

Controls the number of features

B

Controls the depth of the tree

C

Controls the number of samples

D

Controls the number of trees

Q76

Q76 A decision tree is overfitting the training data. What can you do to resolve this?

A

Increase the tree depth

B

Use fewer features

C

Prune the tree

D

Increase the training data

Q77

Q77 A Random Forest model gives inconsistent predictions across different datasets. What could be the cause?

A

Not enough trees

B

Too many features

C

Overfitting in individual trees

D

No randomness in the dataset

Q78

Q78 A decision tree has poor performance on test data but excellent performance on training data. What is the issue?

A

High variance

B

Underfitting

C

Overfitting

D

Data imbalance

Q79

Q79 What is the primary goal of clustering algorithms?

A

To label data

B

To classify data

C

To group similar data points

D

To reduce dimensionality

Q80

Q80 Which of the following is an example of a density-based clustering algorithm?

A

K-Means

B

DBSCAN

C

Hierarchical clustering

D

Agglomerative clustering

Q81

Q81 What is the key difference between K-Means and Hierarchical clustering?

A

K-Means requires predefined number of clusters

B

Hierarchical clustering is faster

C

K-Means doesn't need distance metrics

D

Hierarchical clustering cannot be used for large datasets

Q82

Q82 Which clustering algorithm does not require specifying the number of clusters in advance?

A

K-Means

B

DBSCAN

C

K-Nearest Neighbors

D

PCA

Q83

Q83 Which function in sklearn is used to implement K-Means clustering?

A

kmeans_clustering()

B

KMeans()

C

cluster_means()

D

cluster_KMeans()

Q84

Q84 How do you specify the number of clusters in K-Means clustering using sklearn?

A

num_clusters()

B

cluster_count()

C

n_clusters

D

num_clust

Q85

Q85 Which function is used to implement the DBSCAN algorithm in sklearn?

A

dbscan_clustering()

B

DBSCAN()

C

density_cluster()

D

cluster_density

Q86

Q86 In K-Means, what is the effect of using a large number of clusters (n_clusters)?

A

Increases overfitting

B

Increases the number of iterations

C

Increases randomness

D

Improves accuracy

Q87

Q87 A K-Means model produces clusters of very different sizes. What could be the reason?

A

Wrong distance metric

B

Data is not scaled

C

Overfitting

D

High dimensionality

Q88

Q88 A DBSCAN model labels a large number of points as noise. What could be the cause?

A

High epsilon value

B

Low min_samples value

C

Low epsilon value

D

High min_samples value

Q89

Q89 A hierarchical clustering algorithm produces a large number of small clusters. What could resolve this issue?

A

Increase the number of clusters

B

Reduce the number of clusters

C

Increase the linkage criterion

D

Use a different distance metric

Q90

Q90 What is the purpose of dimensionality reduction in machine learning?

A

To increase model complexity

B

To remove redundant features

C

To reduce model accuracy

D

To create more features

ad verticalad vertical
ad