I. Parameter Estimation
1 Statistical Modeling and Quality Criteria
Statistical Estimation
The goal is to determine the probability distribution of the random variables based on avaliable samples. / The stochastic model is a set of probability spaces and the task of statistical estimation is to select the most appropriate candidate based on the observed outcomes of a random experiment.
Statistical Model
In a statistical model, data is assumed to be generated by some underlying probability distribution, and the goal is to estimate the parameters of this distribution.
Statistics
refers to the probability of an event according to the probability distribution selected by . The same holds for and .
1.1 Introductory Example: Estimating upper bound
Given
Solution 1: Use
Solution 2: Intuition
1.2 Consistency and Unbiasedness
Consistent estimator :
Using Chebyshev inequality, we derive the law of large numbers,
Using law of large numbers, we get
Unbiased estimator :
is unbiased:
is asymptorically unbiased:
Proof. For
is unbiased:
1.3 Variance:
A further quality measure for an estimator is its variance.
1.4 Mean Squared Error(MSE):
An extension of the variance is the MSE(mean squared error), where
1.5 Bias/Variance Trade-Off
MSE of an estimator
and can be decomposed into its bias and variance
Choose
to get optimal
Therefore, an unbiased estimator is not necessarily the optimal estimator, but for large
2 Maximum Likelihood Estimation
2.1 Maximum Likelihood Principle
The maximum likelihood principle suggests to select a candidate probability measure such that the observed outcomes of the experiment become most probable. A maximum likelihood estimator
The likelihood function depends on the statistical model, assuming all observations are iid, we obtain,
Normally, we use log-likelihood function,
In the slides,
is used rather than .
2.2 Parameter Estimation
Channel Estimation
Consider an AWGN channel
Given
The ML estimator is obviously identical with the least squares estimator, which changes drastically when the statistics
are correlated or when is non-Gaussian distributed.
Introductory Example: Estimating upper bound
Suppose the distribution of observations is uniform, the likelihood function of
Bernoulli Experiments
Given
and
The ML-estimator is obtained by
In the following, we analyze the quality of
Since the estimator is unbiased, the MSE is equal to the variance of the estimator,
However, biased estimator can have less MSE and thus provide better estimates than unbiased estimator.
Alternative Solution:
2.3 Best Unbiased Estimator
ML estimators are not necessarily the best estimators. However, a wide class of estimators is defined by minimizing the MSE under an unbiasedness constraint.
We call an estimator
for any alternative unbiased estimator
Best unbiased estimators are also referred to as UMVU(Uniformly Minimum Variance Unbiased) estimators.
3. Fisher's Information Inequality
An universal lower bound for the variance of an estimator can be introduced, if the following consition is fulfilled:
We define the score function as the slope of
3.1 Cramer-Rao Lower Bound
With
which can be interpreted as the negative mean curvature of the log-likelihood function at
The variance of an estimator can be lower bounded by the Cramer-Rao lower bound
If
Properties of the Fisher Information:
depends on given observations and the unknown parameterA large value of
corresponds to a strong curvature and more information in . A small value of corresponds to a weak curvature and little information in . is monotonically increasing with the number of independent observation statistics.
3.2 Exponential Models
A exponential model is a statistical model with