Parameter Estimation (needs practice (mixups)) Flashcards

(13 cards)

1
Q

Describe the parametric approach to building a classifier

A

Choose distribution for P(x|ωi) with parameters θ

For each class ωi, find the parameters that best fit the training data

Determine the prior probabilities P(wi), determine what share of the training data this class makes up

Compute P(ωi|x) for each class and make a classification

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What does p(x) means

A

a function of x, same as f(x)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the parameter (θ) likelihood function with respect to X

A

p(x1,x2, … ,xN:θ)

We can think of this as the probability that all these datapoints came from the distribution with parameters θ

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the maximum likelihood estimate

A

A function that chooses the parameter values 𝜃 that make the observed data 𝑋
as likely as possible under the model.

The function chooses parameters that maximises the likelihood

likelihood: probability of the training data given the class

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Why maximise the likelihood

A

By maximising the likelihood function, we find the parameter values that best explain the observed data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

In many cases the data is independent of each other (peoples heights are independent of one another), what do we do to ensure that Max likelihood function only has one input

A

Factorise the probabilities

P(A,B) = P(A)P(B)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

How is the product of all the max likelihood probabilities shown

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

How do you find the minimum and maximum of f(x)

A

Find the turning points

df(x)/dx = 0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Using a cunning trick, instead of maximising the likelihood, what should we maximise

A

the log likelihood

ln p(X; θ)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How do we find the best parameters using the log likelihood

A

Start with the likelihood

Take the log of the likelihood, this turns it into a sum

Take the derivative of the log likelihood and set it to zero

Calculate θ

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the advantage of taking the log likelihood

A

It turns the product of all the likelihoods into the sum of all the log likelihoods, which is easier to deal with

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the negative log likelihood

A

the minus of the negative log likelihood, you minimise it rather than maximise it

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the maximum of the gaussian function for the mean and variance and therefore the parameters

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly