Jump to content

Least-squares estimation of linear regression coefficients

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Deimos 28 (talk | contribs) at 10:35, 10 January 2006 (Lightening the Regression Analysis article). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.
(diff) ← Previous revision | Latest revision (diff) | Newer revision → (diff)

Given the hypothesis of linear regression and the Gauss-Markov hypothesis, we can find an explicit form for the function which lies the most clsely to the dependant variable .

As and are random variables, we only have a concrete realization and of them. Based on these numbers, we can only find an estimate of .

Therefore, we want an estimate of . Under the Gauss-Markov assumptions, there exists an optimal solution. We can see the unknown function as the projection of on the subspace of generated by . , where is the matrix whose columns are .

If we define the scalar product by and write for the induced norm, the metric d can be written . Minimizing this norm is equivalent to projecting orthogonally on the subspace induced by with the projection .

The projection being orthogonal, is orthogonal to the subspace generated by . Therefore, . As , this equation yields to .

If is of full rank,, then so is . In that case,

. Given the realizations and of and , we choose and .