Warmup
Let and be scalar constants, and be a scalar random variable.
Fill in the blanks
Goal
Recall that the least squares estimates are:
Our goal today is to learn about the statistical properties of these estimates, in particular their expectation and variance.
Random Vectors
is a vector-valued random variable, so we first need to cover a little more background.
Let be scalar random variables. Then the vector is a vector valued random variable, a.k.a. a random vector.
The expectation of is the vector of expectations,
And the variance-covariance matrix of is
For example, the errors in multiple linear regression, are independent with mean 0 and variance .
Then,
Properties of Expectation and Variance for random vectors
Let:
- be a random vector
- be a constant matrix
- be a constant vector
Then:
These are the vector analogs of the properties you wrote down in the warmup.
Find and where satisfies the multiple linear regression equation: and ,
Expectation of the least squares estimates
Assume the regression set up (with the usual dimensions): where is fixed with rank , , and .
Fill in the blanks to show the least squares estimates are unbiased
Variance-covariance matrix of the least square estimates
Fill in the blanks to find the variance covariance matrix of least squares estimates
We can pull out the variance of a particular parameter estimate, say , from the diagonal of the matrix: where indicates the element in the i’th row and j’th column of the matrix .
Why ?
The off diagonal terms tell us about the covariance between parameter estimates.
Estimating
To make use of the variance-covariance results we need to be able to estimate .
An unbiased estimate is:
The denominator is known as the model degrees of freedom.
Standard errors of particular parameters
The standard error of a particular parameter is then the squareroot of the variance replacing with its estimate:
Gauss Markov Theorem
You might wonder if we can find estimates with better properties. The Guass-Markov theorem says the least squares estimates are BLUE (Best Linear Unbiased Estimator).
Of all linear, unbiased estimates, the least squares estimates have the smallest variance.
Of course if you are willing to have a non-linear estimate and/or, a biased estimate you might be able to find an estimate with smaller variance.
Proof, see Section 2.8 in Faraday
Summary
For the linear regression model where , , and the matrix is fixed with rank .
The least squares estimates are
Furthermore, the least squares estimates are BLUE, and
We have not used any Normality assumptions to show these properties.