Consider a collection of m linear regression problems for n observations, related through a set of common predictor variables
, and a jointly normal errors
:



where the subscript c denotes a column vector of k observations for each measurement
(
).
The noise terms are jointly normal over each collection of k observations. That is, each row vector
represents an m length bvector of of correlated observations on each of the dependent variables:

where the noise
is i.i.d. and normally distributed for all rows
.

where B is an
matrix
![{\displaystyle B=[\beta _{1},\cdots ,\beta _{c},\cdots ,\beta _{m}]\,}](/media/api/rest_v1/media/math/render/svg/223cb43ae8419947ab44568b8d5879f6633e74c4)
We can write the entire regression problem in matrix form as:

where Y and E are
matrices.
The classical, frequentists least linear squares solution is to simply estimate the matirx of regression coeeficients
using the Moore-Penrose pseudoinverse:
.
To obtain the Bayesian solution, we need to specify the confitional likelihood and then find the appropriate conjugate prior. As with the univerate case of Linear Bayesian Regression, we will find that we can specify a natural conditional conjugate prior (which is scale dependent).
Let us write our conditional likelihood as

writing the error E in terms Y,X, and B yields

We seek a natural conjugate prior--a joint density
which is of the same functional form as the likelihood. Since the likelihood is quadratic in
, we re-write the likelihood so it is normal in
(the deviation from classical sample estimate)
Using the same technique as iwith Linear Bayesian Regression, we decompose the exponential term using a matrix-form of the sum-of-squares techique. Here, however, we will also need to use the Kronecker product and vectorization transformations.
First, let us apply sum-of-squares to obtain new expression for the likelihood:


We would like to develop a condition form for the priors:

where
is an inverse-Wishsart distribution
and
is some form of Normal distribution in the matrix
. This is accomplished using the vectorizaton transformation, which converts the likelihood from a function of the matrices
to a function of the vectors
:

With the prior now specified, we can express the posterior distribution as




With some re-arrangement, we can re-write the posterior so that the posterior mean
is weighted average of the least squares estimator and the prior mean:

where
comes from the LU decomposition of
(which is a positive-definite matrix by design)

This is the key result of the Empirical Bayes approach; it allows us to estimate the slope
for our original linear regression problem by combining estimates using the least squares estimate
for a single set of measurements with the empirical prior estimate
from a large collection of similar measurements. (Notice that the weighted average also depends on the empirical estimate of the prior covariance matrix
.)
To justify this, collect the quadratic terms in the exponential and
and now express this as a quadratic form in
:


where
![{\displaystyle ns^{2}=(v-W{\tilde {\beta }})^{T}(v-W{\tilde {\beta }}),v=[y,U{\bar {B}}],W=[X,U]}](/media/api/rest_v1/media/math/render/svg/1f82aa6592eae79be9dfd6441e10ef50c3aa21a1)
The posterior can now be expressed as a Normal distribution
times an inverse-gamma distribution:

A similar analysis can be performed for general case of multi-variate regression for a Bayesian Estimation of covariance matrices.
Example:
References
- Bradley P. Carlin and Thomas A. Louis, Bayes and Empirical Bayes Methods for Data Analysis, Chapman & Hall/CRC, Second edition 2000,
- Peter E. Rossi, Greg M. Allenby, and Robert McCulloch, Bayesian Statistics and Marketing, John Wiley & Sons, Ltd, 2006
External links