SVM Kernels
Which of the following is NOT a typical kernel used in Support Vector Machines?
- A. Linear kernel
- B. Polynomial kernel
- C. Radial basis kernel
- D. Sigmoid kernel
- E. Tangential kernel
Machine Learning Motivation
What is the primary reason for the increasing adoption of Machine Learning?
- A. To replace all human jobs with computers.
- B. To solve real-world problems by learning from data instead of hard-coded rules.
- C. To make computers think like humans.
- D. To build more complex algorithms.
- E. Because everyone else is doing it.
Classification vs. Regression
Which task is suitable for classification?
- A. Predicting stock prices.
- B. Determining the temperature for tomorrow.
- C. Categorizing emails as spam or non-spam.
- D. Forecasting sales figures.
- E. Estimating the height of a building.
Bias in Machine Learning
What does bias in data indicate in machine learning?
- A. Perfectly balanced dataset
- B. Complete lack of errors.
- C. Inconsistency in the data.
- D. High precision in predictions.
- E. The data is up to date.
Cross-Validation
What is the main purpose of cross-validation in machine learning?
- A. To increase the size of the training data.
- B. To reduce the training time of the model.
- C. To improve the model's performance on the training data.
- D. To assess how the results of a statistical analysis will generalize to an independent data set.
- E. To validate the code syntax.
Support Vectors
What are support vectors in SVM?
- A. All data points used for training the model.
- B. The vectors that define the axes of the data.
- C. Points on the edge of the dividing hyperplane that determine the margin.
- D. Features that are most relevant in determining the output.
- E. The lines used to graph the data.
PCA Purpose
What is the most common use for Principal Component Analysis (PCA)?
- A. Increasing the number of dimensions in a dataset.
- B. Improving the accuracy of regression models.
- C. Dimension reduction.
- D. Enhancing the visualization of high-dimensional data.
- E. Data encryption.
Naive Bayes Assumption
What is the 'naive' assumption in a Naive Bayes classifier?
- A. That the data is normally distributed.
- B. That all features are equally important.
- C. That all attributes are independent of each other.
- D. That there is no noise in the data.
- E. That the data does not have any missing values.
Unsupervised Learning
Which of the following tasks is an example of unsupervised learning?
- A. Predicting housing prices based on features like size and location.
- B. Classifying images of cats and dogs.
- C. Grouping customers into segments based on their purchasing behavior.
- D. Identifying the gender of a person based on height and weight.
- E. Sorting books by title.
Supervised Learning Example
Which of the following is an example of supervised learning?
- A. Clustering similar articles together.
- B. Reducing the dimensionality of a dataset.
- C. Training a model to predict whether an email is spam based on labeled data.
- D. Discovering hidden patterns in customer transactions.
- E. Generating a new sequence of text based on existing text.
F1 Score Calculation
The F1 score is calculated using which two metrics?
- A. Accuracy and Specificity.
- B. Sensitivity and Specificity.
- C. Precision and Recall.
- D. Error and Variance.
- E. Bias and Variance.
Precision Definition
What does precision measure in machine learning?
- A. The proportion of actual positives that are correctly identified.
- B. The proportion of predicted positives that are actually positive.
- C. The overall correctness of the model.
- D. The number of false negatives.
- E. The number of data points.
Tackling Overfitting
What is a common technique to tackle overfitting in machine learning models?
- A. Adding more features to the dataset.
- B. Using a more complex model.
- C. Resampling the data and using k-fold cross-validation.
- D. Removing all the outliers from the data.
- E. Decreasing the learning rate.
Ensemble Learning
What is the primary goal of ensemble learning?
- A. To simplify the model and reduce training time.
- B. To create more powerful models by combining multiple machine learning models.
- C. To remove bias from the dataset.
- D. To increase the variance of the model.
- E. To reduce the need for data preprocessing.
Loss vs Cost Function
What is the key difference between a Loss Function and a Cost Function?
- A. Loss function is used for regression, cost function is used for classification.
- B. Cost function is used for a single data point, loss function is used for multiple.
- C. Loss function is for single data point, cost function aggregates over the entire training data.
- D. They are the same thing.
- E. Loss function is used during training, cost function is used during testing.