Jump to content

Bayesian multivariate linear regression

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Charlesmartin14 (talk | contribs) at 22:44, 19 October 2006. The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.
(diff) ← Previous revision | Latest revision (diff) | Newer revision → (diff)

Consider a collection of m linear regression problems for n observations, related through a set of common predictor variables , and a jointly normal errors  :

where the subscript c denotes a column vector of k observations for each measurement ().

The noise terms are jointly normal over each collection of k observations. That is, each row vector represents an m length bvector of of correlated observations on each of the dependent variables:

where the noise is i.i.d. and normally distributed for all rows .

where B is an matrix


We can write the entire regression problem in matrix form as:

where Y and E are matrices.

The classical, frequentists least linear squares solution is to simply estimate the matirx of regression coeeficients using the Moore-Penrose pseudoinverse:

.

To obtain the Bayesian solution, we need to specify the confitional likelihood and then find the appropriate conjugate prior. As with the univerate case of Linear Bayesian Regression, we will find that we can specify a natural conditional conjugate prior (which is scale dependent).

Let us write our conditional likelihood as

writing the error E in terms Y,X, and B yields


We seek a natural conjugate prior--a joint density which is of the same functional form as the likelihood. Since the likelihood is quadratic in , we re-write the likelihood so it is normal in (the deviation from classical sample estimate)

Using the same technique as iwith Linear Bayesian Regression, we decompose the exponential term using a matrix-form of the sum-of-squares techique. Here, however, we will also need to use the Kronecker product and vectorization transformations.

First, let us apply sum-of-squares to obtain new expression for the likelihood:


We would like to develop a condition form for the priors:

where is an inverse-Wishsart distribution and is some form of Normal distribution in the matrix . This is accomplished using the vectorizaton transformation, which converts the likelihood from a function of the matrices to a function of the vectors :


With the prior now specified, we can express the posterior distribution as

With some re-arrangement, we can re-write the posterior so that the posterior mean is weighted average of the least squares estimator and the prior mean:

where comes from the LU decomposition of (which is a positive-definite matrix by design)

This is the key result of the Empirical Bayes approach; it allows us to estimate the slope for our original linear regression problem by combining estimates using the least squares estimate for a single set of measurements with the empirical prior estimate from a large collection of similar measurements. (Notice that the weighted average also depends on the empirical estimate of the prior covariance matrix .)


To justify this, collect the quadratic terms in the exponential and and now express this as a quadratic form in :

where

The posterior can now be expressed as a Normal distribution times an inverse-gamma distribution:


A similar analysis can be performed for general case of multi-variate regression for a Bayesian Estimation of covariance matrices.


Example:


References

  • Bradley P. Carlin and Thomas A. Louis, Bayes and Empirical Bayes Methods for Data Analysis, Chapman & Hall/CRC, Second edition 2000,
  • Peter E. Rossi, Greg M. Allenby, and Robert McCulloch, Bayesian Statistics and Marketing, John Wiley & Sons, Ltd, 2006