In: Statistics and Probability
(i)
Standard error of the OLS estimate is
given as,
where
is
the standard error of the regression
is the total sample
variation in
, and
is the
R-squared from regressing
on all other
independent variables (and including an intercept).
(ii)
is the
R-squared from regressing
on all other
independent variables (and including an
intercept). Because the R-squared measures
goodness-of-fit, a value of
close to one
indicates that
explains much of the
variation in all other independent variables in the sample. This
means that
are highly correlated
with all other independent variables.
where n are number of observations and k are number of independent variables
and are the OLS
residuals.
(iii)
Gauss Markov Theorem - Under Assumptions given below,
are the best linear unbiased estimators (BLUEs) of
respectively.
Assumptions -
Assumption MLR.1 (Linear in Parameters) - The model in the population can be written as
where
are the unknown parameters (constants) of interest and u is an
unobservable random error or disturbance term.
Assumption MLR.2 (Random Sampling) - We have a random sample of
n observations,
, following the population model in Assumption
MLR.1.
Assumption MLR.3 (No Perfect Collinearity) - In the sample (and therefore in the population), none of the independent variables is constant, and there are no exact linear relationships among the independent variables.
Assumption MLR.4 (Zero Conditional Mean) - The error u has an expected value of zero given any values of the independent variables. In other words,
Assumption MLR.5 (Homoskedasticity)- The error u has
the same variance given any values of the explanatory variables. In
other words,
.