Lesson 3.12: The Generalized LRT and Wilks' Theorem
The Neyman-Pearson Lemma gave us the 'Most Powerful' test, but only for simple hypotheses. In this final theoretical lesson, we generalize this idea to handle real-world composite hypotheses (e.g., H₀: β₁=β₂=0). We'll build the Generalized Likelihood Ratio Test (LRT) and introduce Wilks' Theorem, which provides a universal test statistic.
Part 1: The Real-World Challenge
The Neyman-Pearson Lemma is elegant, but it requires a simple alternative like . Real-world alternatives are composite, like , which contains an infinite number of possible values. Which one should we use in our likelihood ratio?
The 'Handcuffs' Analogy
Imagine comparing two models:
- The Restricted Model (H₀): This is our model with "handcuffs" on. We force the null hypothesis to be true (e.g., we force ).
- The Unrestricted Model (H₁): This is our model with the handcuffs off. The parameters are free to be whatever the data demands.
The Core Question: Does removing the handcuffs *significantly* improve the model's fit to the data?
Part 2: The Generalized Likelihood Ratio (GLR) Test
The solution is to compare the best possible likelihood of the restricted model to the best possible likelihood of the unrestricted model.
Definition: The GLR Statistic (Λ)
The GLR statistic is the ratio of the maximized likelihood under the null (Restricted) to the maximized likelihood under the alternative (Unrestricted).
Interpreting the Ratio:
Since the unrestricted model has more freedom, , which means .
- If : The handcuffs didn't matter. The restricted model fits almost as well as the unrestricted one. This supports H₀.
- If : The handcuffs were a major problem. The unrestricted model fits the data vastly better. This provides strong evidence against H₀.
Our decision rule is: Reject H₀ if is "too small."
Part 3: The Magic Bullet: Wilks' Theorem
We have a test statistic (), but finding its exact distribution is nearly impossible. This is where a beautiful asymptotic result, **Wilks' Theorem**, comes to the rescue.
For large sample sizes (), under the null hypothesis, the statistic converges in distribution to a **Chi-squared () distribution**.
The degrees of freedom, , is the **number of independent restrictions** imposed by H₀. (e.g., for , ).
Why this specific transformation?
The form is clever for two reasons:
- It turns the difficult ratio of likelihoods into an easy subtraction of log-likelihoods.
- It conveniently flips our decision rule. A small (evidence against H₀) corresponds to a **large** . This means our decision rule is now the standard "Reject H₀ if the test statistic is large," which is much more intuitive.
The Connection: For OLS models with the assumption of Normal errors, the F-statistic is just a simple mathematical transformation of the Likelihood Ratio statistic, . They are two different ways of measuring the exact same thing: the loss of fit from imposing a restriction.
Why It Matters: This proves that the F-test, which we motivated by comparing sums of squares, is also the most powerful test from the perspective of likelihood theory. More importantly, Wilks' Theorem shows that the LRT is **more general**. While the F-test is specific to linear models, the LRT can be used to compare *any* nested models (linear, logistic, GARCH, etc.) for which a likelihood can be written, making it the workhorse of modern statistical model comparison.
This reveals a profound connection between the tools we've learned.
What's Next? Module 3 Complete!
Congratulations! You have now completed the entire theoretical arc of **Statistical Inference & Estimation Theory**.
You've learned the properties of good estimators, the methods to build them (MoM, MLE), and the complete framework for using them to make decisions (CIs, Hypothesis Testing, p-values, and the theory of optimal tests).
You now possess the foundational knowledge to understand every statistical test and model that follows. In **Module 4**, we will put this entire framework into practice as we build and rigorously test the most important model in all of quantitative analysis: the **Linear Regression Model**.