Bayesian estimation is also a form of posterior estimation, but it is not maximum a posteriori estimation. Its goal is not to maximize the posterior, but to maximize the posterior expectation of the parameter, as shown below:
$$ \hat{\theta}_{\text{Bayes}} = E[\theta \mid \mathbf{x}] = \int \theta p(\theta \mid \mathbf{x}) d\theta $$
However, it is important to note that the goal of Bayesian estimation is often not just to obtain the final value of $\hat{\theta}_{\text{Bayes}}$, but to obtain the probability distribution of the parameter $\theta$ itself.
Let's demonstrate the process of Bayesian estimation through a simple example, such as estimating the probability of a coin landing heads up, denoted as $p$. This is a typical binomial distribution problem, and we can use Bayesian methods to update our belief about $p$.
Assumptions: we don't know whether the coin is fair. Therefore, our prior belief about $p$ is that it could be any value between 0 and 1, and all values are equally likely (i.e., the prior distribution is uniform). Mathematically, we can express this as $p \sim U(0, 1)$.
Experimental Observation: Suppose we flip the coin 10 times and observe 7 heads and 3 tails. We want to update our knowledge about $p$.
Bayesian Update Steps:
Based on the observed data, we have updated our knowledge about the probability of the coin landing heads up, $p$. Initially, we knew nothing about $p$ and assumed it was uniformly distributed between $[0, 1]$. The observed data has updated our belief to a $\text{Beta}(8, 4)$ distribution, with an expected value of approximately 0.667. This indicates that, based on the current data, the probability of the coin landing heads up is about 2/3.
It needs to extend to the bayes linear regression.