Bayesian Basics for ML Flashcards

(40 cards)

1
Q

What is the core idea of Bayesian inference?

A

To treat unknown parameters as random variables and update a prior belief about them using observed data to obtain a posterior belief.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

In Bayesian terms, what is a prior distribution?

A

A probability distribution that represents our beliefs about a parameter before seeing the current data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is a likelihood in Bayesian inference?

A

The probability of the observed data as a function of the parameter, reflecting how plausible the data are under different parameter values.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is a posterior distribution?

A

The updated distribution of the parameter after combining the prior and the likelihood using Bayes’ rule.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is Bayes’ rule in the parameter-data form?

A

Posterior ∝ Likelihood × Prior, or p(θ|data) ∝ p(data|θ)p(θ).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What does the normalization constant in Bayes’ rule ensure?

A

That the posterior distribution integrates (or sums) to 1 over all possible parameter values.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is a point estimate in the Bayesian framework?

A

A single summary of the posterior such as the posterior mean, median, or maximum a posteriori (MAP) estimate.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the MAP (maximum a posteriori) estimate?

A

The parameter value that maximizes the posterior distribution p(θ|data).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How is MAP estimation related to regularized maximum likelihood?

A

MAP is equivalent to maximizing log-likelihood plus log-prior, which often looks like a regularized loss function.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What type of prior corresponds to L2 regularization in linear models?

A

A Normal (Gaussian) prior on coefficients, leading to ridge-like penalties.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What type of prior corresponds to L1 regularization in linear models?

A

A Laplace (double-exponential) prior on coefficients, leading to lasso-like penalties.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Why is it useful to view regularization as a Bayesian prior?

A

It provides an interpretation of regularization as encoding prior beliefs about parameter magnitude and supports probabilistic reasoning.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is a conjugate prior?

A

A prior distribution chosen so that the posterior is in the same family as the prior when combined with a given likelihood.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Why are conjugate priors convenient?

A

They allow closed-form posterior updates, simplifying analytic calculations and reducing computational cost.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is a conjugate prior for a Bernoulli or Binomial likelihood?

A

The Beta distribution is conjugate to Bernoulli and Binomial likelihoods for a probability parameter.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is a conjugate prior for a Poisson likelihood?

A

The Gamma distribution is conjugate for the rate parameter of a Poisson likelihood.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What is a conjugate prior for a Normal likelihood with known variance and unknown mean?

A

A Normal prior on the mean is conjugate, yielding a Normal posterior for the mean.

18
Q

What is a Bayesian credible interval?

A

An interval [a,b] such that the posterior probability that the parameter lies in [a,b] equals a chosen level (e.g., 95%).

19
Q

How does a credible interval differ conceptually from a frequentist confidence interval?

A

A credible interval directly expresses probability about the parameter given the data; a confidence interval concerns the long-run coverage of the procedure under repeated sampling.

20
Q

What is the posterior predictive distribution?

A

The distribution of a future observation, obtained by averaging the likelihood of new data over the posterior distribution of the parameters.

21
Q

Why is the posterior predictive distribution useful in ML?

A

It captures both parameter uncertainty and data noise, providing more realistic uncertainty about future predictions.

22
Q

What is hierarchical (multilevel) modeling in the Bayesian context?

A

Modeling parameters themselves as drawn from higher-level distributions, allowing sharing of information across groups or entities.

23
Q

Why are hierarchical models powerful?

A

They enable partial pooling across groups, improving estimates for small-sample groups and capturing structure in multi-level data.

24
Q

What is partial pooling?

A

An approach where group-specific estimates are shrunk towards a global mean based on data and prior, balancing between no pooling and complete pooling.

25
What is a posterior mean estimator for a probability parameter under a Beta–Binomial model?
Often a weighted average of the prior mean and observed proportion, with weights depending on prior strength and sample size.
26
Why can Bayesian updating be seen as a form of 'smart smoothing' of empirical estimates?
Priors temper extreme values from small samples, pulling estimates towards plausible ranges until more data reduces prior influence.
27
What role do priors play when we have lots of data?
Under regular conditions, the likelihood dominates and the influence of reasonable priors diminishes as sample size grows.
28
Why can prior choice matter a lot in small-sample settings?
With limited data, the prior contributes substantially to the posterior, affecting estimates and uncertainty.
29
What is a noninformative or weakly informative prior?
A prior chosen to have minimal influence or to rule out only extreme or implausible parameter values while remaining relatively diffuse.
30
Why are truly 'noninformative' priors often elusive?
Priors that appear flat in one parameterization can be highly informative under reparameterization, and boundaries/constraints complicate neutrality.
31
What is MCMC (Markov chain Monte Carlo) in Bayesian computation?
A class of algorithms that approximate posterior distributions by drawing samples via a Markov chain whose stationary distribution is the posterior.
32
Why is MCMC needed for many Bayesian models?
Because closed-form posteriors are unavailable and direct integration is intractable in high dimensions.
33
What are variational inference methods in Bayesian ML?
Optimization-based approximations that replace the true posterior with a simpler distribution chosen to minimize a divergence, often for scalability.
34
Why is approximate Bayesian inference important in modern ML?
Exact inference is often infeasible in complex models, so practical algorithms rely on approximations to capture key posterior structure.
35
What is Bayesian model averaging at a high level?
Combining predictions from multiple models weighted by their posterior probabilities instead of choosing a single best model.
36
Why is Bayesian model averaging conceptually attractive?
It accounts for model uncertainty, often improving predictive performance and calibration compared to relying on one model.
37
How does Bayesian thinking relate to ML hyperparameter tuning?
Bayesian optimization treats the performance as a function of hyperparameters and uses a probabilistic model to choose promising configurations.
38
Why is a full Bayesian treatment of deep nets often impractical?
The parameter space is huge and highly non-linear, making exact posterior inference computationally difficult.
39
What is a practical Bayesian-inspired idea used in ML without full inference?
Using priors as regularizers, ensembles as approximations to posterior predictive distributions, or uncertainty estimates from dropout or ensembles.
40
In one sentence, what is the key mental model for Bayesian stats in ML?
Bayesian methods update prior beliefs with data to produce posterior distributions, which in ML often appear as regularization, uncertainty estimates, and principled predictive distributions.