Exact Inference - 07 Flashcards

(33 cards)

1
Q

What is marginal inference?

A

We ask the question ‘What is the probability of a given variable in our model (possibly conditioned on evidence)?’.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is maximum a posteriori (MAP) inference?

A

We ask the question ‘What is the most likely assignment to the variables in the model (possibly conditioned on evidence)?’

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Inference is not challenging. True or False?

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Chains and trees are ______________. Loopy graphs are not.

A

tractable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

If a problem is intractable, we are still able to obtain useful answers via _______________ inference methods.

A

approximate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Which are the methods for inference in graphical models?

A
  1. Variable elimination
  2. Belief propagation (aka message passing algorithms)
  3. Approximate inference methods (aka Monte Carlo techniques)
  4. Variational inference
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is dynamic programming?

A

Dynamic programming ”inverts” the order of computation, performing it inside out instead of outside in.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are the 2 ideas that help us address the exponential blowup of the joint distribution in the variable elimination method?

A
  1. Because of the structure of the Bayesian network, some subexpressions in the joint depend only on a small number of variables.
  2. By computing these expressions once and caching the results, we can avoid generating them exponentially many times.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Why does the order of elimination matter?

A

!= elimination orders –> != intermediate factor sizes –> != complexity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the naïve approach in the variable elimination method?

A

First, evaluate the joint distribution and then perform the summation explicitly.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Define tree in an undirected graph.

A

In an undirected graph, a tree is defined as a graph in which there is one path between any pair of nodes.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Trees have loops. True or False?

A

False

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Define tree in a directed graph.

A

In directed graphs, a tree is defined such that there is a single node, called the root, which has no parents, and all other nodes have one parent.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

If there are nodes in a directed graph that have more than one parent, but there is still one path between any two nodes, then the graph is called a _____________.

Such a graph will have more than one node with the property of having ____ ___________, and the corresponding moralized undirected graph will have __________.

A

polytree; no parents; loops

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

______________ algorithm is used for marginal inference.

_______________ algorithm is used for MAP inference.

A

Sum-product; Max-product

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Marginal is proportional to the product of the incoming messages. True or False?

17
Q

In the sum-product algorithm, if the factor graph is derived from a directed graph, then the joint distribution is already correctly ________________, and so the marginals obtained will be normalized correctly.

If the factor graph is derived from an undirected graph, then we have the _______________ _________ Z and in this case we just need to marginalize over a single ____________ rather than the entire set of variables.

A

normalized; normalization factor; variable

18
Q

We can use the messages to compute the marginals of all variables in the graph. True or False?

19
Q

If each edge has a message in both directions, we can compute the marginals of all variables in the graph. True or False?

20
Q

The correspondence between messages and effective __________ allows us to find the joint distribution for variables connected to the same factor node (_____________).

A

factors; neighbors

21
Q

From a leaf variable node to a factor node, the message is _____.

22
Q

From a leaf factor node to a variable node, the message is the _____.

23
Q

In the sum-product algorithm, the factor graph must be a _______.

24
Q

The sum-product algorithm can be used to compute _______________.

This can be viewed as defining a new __________ ________ on the non-evidential variables.

Or, one can keep the original factor graph, but for factor-to-variable messages, the sum is only taken over _______________ variables; any evidential variables in the potential are set to their evidential states.

A

conditionals; factor graph; non-evidential

25
The sum-product algorithm may be used for ______________ random variables. All stays the same, but we replace sums with ___________. In special cases, integral can be computed in closed form (e.g., Gaussian family). If not, need for ________________. ________________ are also needed for discrete random variables when K is _________.
continuous; integrals; approximations; approximations; large
26
What do we need to do if our factor graph is not a tree?
If factor graph is NOT a tree: - Group variables together so that the factor becomes a tree (called junction tree or clique tree) - Pretend the factor graph is a tree and use message passing (loopy belief propagation)
27
What is the max-sum algorithm?
The max-sum algorithm is the message passing algorithm by replacing sum with max, products with sums, and factors with log-factors.
28
In the max-sum algorithm, from a leaf variable node to a factor node, the message is ____.
0
29
In the max-sum algorithm, from a leaf factor node to a variable node, the message is the ________________ of the _________.
logarithm; factor
30
The message passing framework can be generalized to arbitrary graph topologies, giving an exact inference procedure known as the _____________ ________ algorithm. If the starting point is a directed graph, it is first converted to an undirected graph by ______________, whereas if starting from an undirected graph, this step is not required. Next, the graph is ________________, which involves finding chordless cycles containing four or more nodes and adding extra links to eliminate such chordless cycles. Next, the triangulated graph is used to construct a new tree-structured undirected graph called a join tree (or clique tree), whose nodes correspond to the maximal _________ of the triangulated graph, and whose links connect pairs of cliques that have variables in common.
junction tree; moralization; triangulated; cliques
31
Understanding approximate inference gives a deeper context for why and how exact inference was needed. True or False?
False. It's the opposite.
32
Distinguish variable elimination and belief propragation.
Variable Elimination: Bayesian/Markov network; Expensive for many queries; Exact for all DAGs; Difficult parallelism Belief Propagation: Factor graph; Reuses computation via message; Exact only for trees, approximate for loopy graphs; Naturally parallelizable
33
A variable is conditionally independent of everything else given its Markov blanket. For a factor graph, the Markov blanket of a variable x is all ________ nodes adjacent to x , and all ___________ nodes adjacent to those _________.
factor; variable; factors