Given the hypothesis of linear regression and the Gauss-Markov hypothesis, we can find an explicit form for the function which lies the most clsely to the dependant variable
.
As
and
are random variables, we only have a concrete realization
and
of them. Based on these numbers, we can only find an estimate
of
.
Therefore, we want an estimate of
. Under the Gauss-Markov assumptions, there exists an optimal solution. We can see the unknown function
as the projection of
on the subspace of
generated by
.
, where
is the matrix whose columns are
.
If we define the scalar product
by
and write
for the induced norm, the metric d can be written
. Minimizing this norm is equivalent to projecting orthogonally
on the subspace induced by
with the projection
.
The projection being orthogonal,
is orthogonal to the subspace generated by
. Therefore,
. As
, this equation yields to
.
If
is of full rank,, then so is
. In that case,
. Given the realizations
and
of
and
, we choose
and
.