BLUE
BLUE refers to a property of estimators in linear regression models, stating that under certain conditions, the estimator of the regression coefficients is the Best Linear Unbiased Estimator. The term “best” implies the estimator has the smallest variance among all linear unbiased estimators.
Key Properties of BLUE
- Best:
 The estimator has the smallest variance among all unbiased linear estimators, ensuring efficiency.
- Linear:
 The estimator is a linear combination of the observed data points.
- Unbiased:
 The expected value of the estimator equals the true parameter value, ensuring there is no systematic error:
 E(β^)=β
- Estimator:
 A formula or method used to estimate population parameters (e.g., mean, variance).
Conditions for BLUE (Gauss-Markov Theorem):
The Gauss-Markov theorem states that under certain assumptions, the OLS estimator is BLUE for the coefficients of a linear regression model. These assumptions include:
- Linearity:
 The relationship between the dependent (Y) and independent (X) variables is linear:
 Y=Xβ+ϵ
 Where:- Y: Dependent variable.
- X: Independent variable(s).
- β: Coefficients to be estimated.
- ϵ: Error term.
 
- No Endogeneity:
 The expected value of the error term is zero:
 E(ϵ)=0
- Homoscedasticity:
 The variance of the error term is constant:
 Var(ϵ)=σ^2
- No Autocorrelation:
 Errors are uncorrelated with each other:
 Cov(ϵi,ϵj)=0 for i≠j
- Full Rank:
 The independent variables in X are linearly independent.
Why is BLUE Important?
- Efficiency:
 Ensures that the estimator is precise by minimizing the variance.
- Interpretability:
 Provides reliable estimates for predictive and inferential analysis.
- Applicability:
 Foundational in econometrics, machine learning, and other data-driven fields.
Limitations
- If any of the Gauss-Markov assumptions are violated, the OLS estimator may no longer be BLUE.
- In the presence of heteroscedasticity or autocorrelation, alternative estimation methods like Generalized Least Squares (GLS) or robust standard errors are needed.
