Posterior Estimator#
With an assumption of the likelihood function (i.e., the distribution of the data), we can determine the distribution of the parameters which is the posterior distribution. Using Bayes rule we find that,
: the likelihood function : the normalization distribution called the evidence.
Expectation of Posterior#
If there must be one value instead of a distribution, the expectation is one natural choice
Maximum a Posterior#
If there must be one value, then the mode or maximum a posteriori (MAP) is a another natural choice.
Bayesian Update#
Motivating Example#
Consider a coin where we do not know the chance of heads
The chance that the coin lands head depends on the value of
In fact this looks like an expected value as a function of
The chance of head given
This is not so surprising since
However, what’s the chance of getting two heads
We know that
The answer is not
Where did the
Normalizing we find that,
Thus,
Beta and Binomial Update#
We are interested in
The Bayesian update rule is,
Let the prior for the parameter
the likelihood for
Hence, the posterior distribution is,
Thus, the Bayesian update adds the number of successes
Expectation
: $
MAP
: $
Transition Rule
: $
Evidence
: The chance of
$$
P(S_n = k) = {n \choose k}\frac{C(r,s)}{C(r+k, s+n-k)}
$$