Lesson 3.4: Method of Moments (MoM) Estimation
We now learn our first 'recipe' for creating estimators. The Method of Moments (MoM) is the oldest and most intuitive estimation technique. Its principle is simple: what we observe in our sample should mirror what is true in the population. We'll use this idea to derive estimators for the mean and variance.
Part 1: The Principle of Matching Moments
The entire philosophy of the Method of Moments, developed by Karl Pearson, can be summarized in one sentence:
The Core Idea: Match the sample moments (which you can calculate from your data) to the population moments (which are functions of the parameters you want to find), and then solve for the parameters.
This principle is built directly on the Law of Large Numbers, which guarantees that our sample moments are consistent estimators of the population moments.
The Method of Moments (MoM) Recipe
To find unknown parameters, you follow three steps:
- Step 1: Express Population Moments: Write the first population moments () as functions of the unknown parameters .
- Step 2: Calculate Sample Moments: Calculate the first sample moments () from your data.
- Step 3: Equate and Solve: Set the corresponding moments equal to each other () to create a system of equations, and solve for your parameters.
Part 2: MoM in Action - Deriving Estimators
We need to find one parameter (), so we need one moment equation.
- Population Moment 1:
- Sample Moment 1:
- Equate and Solve:
Result: The MoM estimator for is .
No-Skip Derivation for μ and σ²
Equation 1 (First Moments):
This is the same as above, giving us our first result: .
Equation 2 (Second Moments):
- Population Moment 2: From the variance identity, .
- Sample Moment 2: .
- Equate: .
Solve the System:
We substitute our first result () into the second equation:
Solving for gives:
The Critical Insight
The MoM estimator for the variance is .
This is the **biased** estimator we identified in Lesson 3.1! This is our first concrete proof that while MoM is intuitive and easy, it doesn't always produce the "best" (unbiased) estimators.
- Consistent: YES. Because sample moments converge to population moments (by the WLLN), MoM estimators are generally consistent.
- Simple to Calculate: YES. The logic is straightforward and usually only involves basic algebra.
- Unbiased: NOT ALWAYS. As we saw with the variance, MoM estimators can be biased in finite samples.
- Efficient: RARELY. MoM only uses the first few moments of the data, potentially ignoring valuable information contained in the full shape of the distribution. This often leads to estimators with higher variance than other methods.
What's Next? A More Powerful Engine
The Method of Moments is a great starting point, but its potential for bias and inefficiency means it's not the workhorse of modern statistics. We need a more powerful, more principled method.
In the next lesson, we will introduce the undisputed champion of estimation techniques: **Maximum Likelihood Estimation (MLE)**. MLE uses the entire probability distribution to find the parameter values that make our observed data "most likely," and it produces estimators with outstanding properties.
Up Next: Let's Learn the Champion: Maximum Likelihood Estimation