Untitled Deck Flashcards

(44 cards)

1
Q

What is the definition of Labels (y ∈ Y) in supervised learning?

A

The target variable or output representing the quantity or category to be predicted

Y = R^k is referred to as regression. For classification, Y is a finite set.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the purpose of Labels (y ∈ Y) in supervised learning?

A

Provides the ‘ground truth’ for training models and evaluating predictions

Does not appear in unsupervised learning.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are Features (x ∈ X) in the context of machine learning?

A

The input data used to make predictions, represented as a feature vector x ∈ R^d

Each component represents a measurable attribute of the data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the purpose of Features (x ∈ X)?

A

Encodes the information from which the model learns patterns to predict labels.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Define Feature Map (ϕ : X → H).

A

A transformation that maps input data to a potentially alternative feature representation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is the purpose of a Feature Map?

A

Enables non-linear relationships or convenient features to be captured.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is a Predictor (fθ : X → Y)?

A

A function that maps x ∈ X to predicted labels f(x) = ˆy ∈ Y.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the purpose of a Predictor?

A

Learns the relationship between features and labels to make accurate predictions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Name a key example of a Predictor.

A
  • Support Vector Machine (SVM)
  • Logistic Regression
  • Discriminative Models
  • Generative Models
  • Likelihood Ratio Tests
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Define Loss (l(y, z)).

A

A function quantifying the penalty for predicting z when the true label is y.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the purpose of Loss?

A

Guides the optimisation process by measuring prediction errors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is Empirical Risk (Rˆ)?

A

The average loss computed over the training data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the purpose of Empirical Risk?

A

Measures how well the model fits the training data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Define Risk (R[fθ]).

A

The expected loss over the true data distribution.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the purpose of Risk?

A

Represents the model’s generalisation performance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the Generalisation Gap?

A

The difference between the true risk and the empirical risk.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What is the purpose of the Generalisation Gap?

A

Quantifies how much worse the model performs on unseen data compared to training data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Define Bayes Predictor.

A

The risk-minimising predictor that achieves the lowest possible risk.

19
Q

What is the purpose of the Bayes Predictor?

A

Serves as the theoretical optimal benchmark for any predictive model.

20
Q

What is Excess Risk?

A

The difference between the risk of the learned model and the risk of the Bayes predictor.

21
Q

What is the purpose of Excess Risk?

A

Measures how far the learned model is from the optimal predictor.

22
Q

Define Overfitting.

A

When a model learns overly complex patterns in the training data.

23
Q

What is the purpose of Overfitting?

A

Highlights the need for regularisation, cross-validation, or simpler models.

24
Q

What are Hyperparameters?

A

Parameters that specify the learning process and are tuned but not learned directly.

25
What is the **purpose** of Hyperparameters?
Optimises model performance and generalisation.
26
Define **k-Fold Cross-Validation**.
A technique where the data is split into k folds for training and validation.
27
What is the **purpose** of k-Fold Cross-Validation?
Provides a robust estimate of generalisation performance.
28
What is **Gradient Descent**?
An iterative optimisation algorithm that updates θ using the gradient of the loss.
29
What is the **purpose** of Gradient Descent?
Efficiently minimises the empirical risk.
30
Define **Learning Rate (η)**.
A hyperparameter controlling the step size at each iteration of gradient descent.
31
What is the **purpose** of Learning Rate?
Balances convergence speed and stability during optimisation.
32
What is **Regularisation**?
Techniques to reduce overfitting and improve generalisation.
33
What is the **purpose** of Regularisation?
Aims to bias the learning algorithm towards preferred solutions.
34
Define **Reproducing Kernel**.
A function that computes inner products in the feature space H.
35
What is the **purpose** of a Reproducing Kernel?
Enables efficient computation of similarities in high-dimensional spaces.
36
What are **Neural Networks**?
A class of models composed of layers that process input data.
37
What is the **purpose** of Neural Networks?
To learn complex patterns and representations from data.
38
What is a **Residual Layer**?
A layer that adds its input to its output.
39
What is the **purpose** of a Residual Layer?
Mitigates the vanishing gradient problem.
40
What is **Backpropagation**?
An algorithm for computing gradients of the loss with respect to the weights.
41
What is the **purpose** of Backpropagation?
Enables efficient gradient computation for updating weights.
42
Define **Universal Approximators**.
Neural networks that can approximate any continuous function on compact subsets.
43
What is the **purpose** of Universal Approximators?
To represent complex functions through neural networks.
44
What is the difference between **Underparameterised** and **Overparameterised Models**?
* Underparameterised: Lacks capacity to fit training data well * Overparameterised: Sufficient capacity to interpolate training data exactly