Supervised Model Evaluation Flashcards

(36 cards)

1
Q

What is the key difference in how supervised and unsupervised learning models are evaluated?

A

Supervised models are evaluated using metrics that compare predictions to true labels, while unsupervised models are evaluated based on the task, as there is no ground truth.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Which classification metric is defined as the ratio of correctly predicted instances to the total number of instances?

A

Accuracy.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Why is Accuracy not a suitable metric for imbalanced datasets?

A

It can be misleadingly high if the model simply predicts the majority class for all instances.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the formula for Accuracy using TP, TN, FP, and FN?

A

Accuracy = $\frac{TP+TN}{TP+TN+FP+FN}$

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Term: Precision

A

Definition: The ratio of true positive predictions to the total number of positive predictions made by the model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

In what type of scenario is maximizing Precision particularly important?

A

In cases where the cost of a false positive is high.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the formula for Precision?

A

Precision = $\frac{TP}{TP+FP}$

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Term: Recall (Sensitivity)

A

Definition: The ratio of true positive predictions to the total number of actual positive instances in the dataset.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

In what type of scenario is maximizing Recall particularly important?

A

In scenarios where the cost of a false negative is high.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the formula for Recall?

A

Recall = $\frac{TP}{TP+FN}$

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

The _____ is the harmonic mean of precision and recall.

A

F1 Score.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the main purpose of using the F1 Score?

A

It provides a single metric that balances the trade-off between precision and recall.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the formula for the F1 Score?

A

F1 = $2 \times \frac{Precision \times Recall}{Precision+Recall}$

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How does Macro F1 calculate the final score for a multi-class problem?

A

It calculates the F1 score for each class independently and then takes the unweighted average, treating all classes equally.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How does Micro F1 calculate the final score for a multi-class problem?

A

It calculates the F1 score globally by considering the total number of true positives, false negatives, and false positives across all classes.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Which variant of the F1 score gives more weight to the performance on larger classes?

17
Q

What does AUC-ROC stand for?

A

Area Under the Receiver Operating Characteristic curve.

18
Q

The ROC curve plots the true positive rate (sensitivity) against which other rate?

A

The false positive rate (1 - specificity).

19
Q

What does the AUC-ROC metric quantify about a classification model?

A

The model’s overall discriminative power across all possible classification thresholds.

20
Q

An AUC-ROC score of 0.5 suggests the model’s performance is equivalent to ____.

A

random guessing

21
Q

What is a confusion matrix?

A

A table showing the counts of true positives, false positives, true negatives, and false negatives for a classification model.

22
Q

Which regression metric measures the average of the squared differences between predicted and actual values?

A

Mean Squared Error (MSE).

23
Q

What is the primary drawback of using MSE as an evaluation metric?

A

It is sensitive to outliers due to the squaring of errors.

24
Q

What is the formula for Mean Squared Error (MSE)?

A

MSE = $\frac{1}{m}\sum_{i=1}^{m}(y_{true}^{(i)} - y_{pred}^{(i)})^2$

25
Which regression metric measures the average of the absolute differences between predicted and actual values?
Mean Absolute Error (MAE).
26
What is the main advantage of using MAE over MSE?
MAE is more robust to outliers because it does not square the errors.
27
What is the formula for Mean Absolute Error (MAE)?
MAE = $\frac{1}{n}\sum_{i=1}^{n} |y_i - \hat{y}_i|$
28
What does the R-squared ($R^2$) metric, or coefficient of determination, represent?
The proportion of variance in the dependent variable that is explained by the model.
29
What does an R-squared ($R^2$) value close to 1 indicate?
It indicates that a large proportion of the variance in the dependent variable is predictable from the independent variables.
30
What is the formula for R-squared ($R^2$) in terms of MSE?
$R^2 = 1 - \frac{MSE}{Total \ variance}$
31
For senior candidates, what do AIC and BIC add to model evaluation compared to a metric like $R^2$?
They add a penalty for the number of parameters and samples, thus adjusting for model complexity.
32
When comparing two models, does a higher or lower AIC/BIC score indicate a better model?
A lower AIC/BIC score indicates a better model.
33
Intuitively, a model with more parameters would score _____ in AIC/BIC, suggesting a poorer fit after penalization.
higher
34
In a standard OLS regression output, what do the 'coefficients' represent?
The estimated relationship between each predictor variable and the outcome.
35
What do the 'p-values' in a regression output table assess?
The statistical significance of each coefficient.
36
What do the 'standard errors' in a regression output table measure?
The uncertainty or variability in the coefficient estimates.