What is the amount of information yielded (entropy) by the outcome of a future event if you already know what the outcome will be?
0
The information content of an event E is what?
A function that increases as the probability p(E) of a function decreases
If p(E) is close to zero then what do we know about the information content of that event?
It is high. We learned a lot from it [“wow that was super surprising. We need to update our models about the world a LOT”]
The smaller the cross entropy the ______ two probability distributions are
more similar
what is a discrete probability distribution?
a vector whose elements lie in 0…1 and that sum to 1