If the loss is squared error, the Bayes decision is found by minimizing or simplifying the notation, .
(5.3) | |||||
Differentiating wrt to find the minimum loss
And Since , the
posterior mean , is
the Bayes decision or rule.
Bayes risk can be found by substituting in (5.3) to get
It is symmetrical
It is easy to interpret. This means that, with such a loss, we can summarize a posterior with a mean (Bayes rule) and variance (Bayes risk).
Other losses also have a Bayes rule of the posterior mean (Lindley 1985)
The squared loss is often criticized for penalising large errors to heavily
The Bayes decision is found by minimizing
Differentiating wrt to find the minimum loss
The Bayes decision or rule is the mean of the weighted posterior over the posterior mean of the weights
Recall this loss function is given by
The Bayes decision can be found by minimising
The Bayes decision is
the fractile (quantile) of the posterior.
In particular for absolute loss the Bayes decision is the median.
It is usual that given positive error may be more serious than a given negative error of the same magnitude or vice-versa. Examples include the following
The plug is pulled out of a ventilator of a very sick hospital patient when the probability that a patient is dead exceeds a threshold say
A nuclear power plant is to be shut down if the probability of a meltdown is greater than a threshold
The safe concentration of CO2 in the atmosphere is thought to exceed a threshold, . When this level is exceeded, the risk of a runaway greenhouse gas effect is thought too high and expensive correctional procedures are carried out.
An interval of length ; say is said to be a modal interval of length
for the distribution of a random variable if
takes on its maximum value out of all such
intervals.
For the loss function
is maximized if is chosen to
be the midpoint of the modal interval of length .
For the limiting case of this as is the hit or miss loss:
where is the Kronecker function. If the posterior distribution is uni-modal the Bayes decision is
the mode of the posterior (The MAP).
Loss function can be thought of as depicting the truth of a model. When a model is either right or wrong this is the appropriate loss function.
This is mainly used in the classical formulation of hypothesis testing as formalized by Newman and Pearson.
It does not take into account shades of usefulness.