Statistical Estimation - Maximum Likelihood Estimation (MLE)
📘 Statistical Estimation
When dealing with statistics, we usually have:
-
A population with unknown parameters (e.g., mean , variance , probability , etc.).
-
A sample of observations drawn from that population.
Since population parameters are unknown constants, we need to estimate them from sample data.
1. Point Estimation
A point estimator is a single statistic (function of sample observations) that provides a “best guess” of the parameter.
-
Example: Sample mean is an estimator of population mean .
-
The obtained numerical value is called the point estimate.
2. Interval Estimation
Instead of one value, we provide an interval of plausible values with a given level of confidence.
-
Example:
is a 95% confidence interval for .
3. Properties of Good Estimators
the following properties are essential:
-
Unbiasedness:
The expected value of the estimator equals the true parameter.
-
Consistency:
As , in probability. -
Efficiency:
Among unbiased estimators, the one with minimum variance is preferred. -
Sufficiency:
An estimator is sufficient if it uses all available information in the sample about the parameter.
📘 Maximum Likelihood Estimation (MLE)
Idea
Step-by-Step Procedure
Suppose is a random sample from a distribution with pdf/pmf , where is an unknown parameter.
-
Likelihood Function:
This is the joint probability of the sample, considered as a function of .
-
Log-Likelihood:
For easier calculations, take logs: First Derivative (Likelihood Equation):
Solving this gives the MLE, .
-
Second Derivative Test:
Ensureto confirm a maximum.
Example 1: MLE for Bernoulli / Binomial
Let successes are observed.
-
Likelihood:
-
Log-likelihood:
-
Differentiate:
-
Solve:
✅ Thus, the MLE of is the sample proportion.
Example 2: MLE for Normal Mean ()
Suppose , with known.
-
Likelihood:
-
Log-likelihood:
-
Differentiate:
-
Solve:
✅ Hence, the MLE of the population mean is the sample mean.
Advantages of MLE
-
Consistency: As , .
-
Asymptotic normality: For large samples, the distribution of tends to normal.
-
Efficiency: Attains the Cramér–Rao lower bound asymptotically.
-
General applicability: Works for discrete, continuous, and complex models.
Limitations of MLE
-
Can be algebraically complicated (often requires iterative methods).
-
Sensitive to outliers.
-
For small samples, MLE may be biased.
Comments
Post a Comment