Jump to content

Introduction to Likelihood Theory/Maximum Likelihood Estimation

From Wikiversity

A simple idea: Since we have seen in previous section (Intuitive Meaning?), we may regard likelihood as the probability of the sample showing up in the space of all the possibles samples of size for a given value of . But the sample has actually occurred and the sample space is infinite, so this sample might have something special - perhaps it's probability of showing up is bigger than the others. Why don't we maximize it?

Maximizing the Likelihood

[edit | edit source]

If we try to maximize likelihoods of variables with densities like the Gaussian

directly, things can be rather complicated. But, we know from calculus that maximizing (or minimizing) when is strictly crescent is equivalent to maximizing . So, we take the logarithm of the likelihood and call the function

log-likelihood. So, we will maximize the log-likelihood instead of maximizing the proper likelihood.

Likelihoods Without Explicit Points of Maximum

[edit | edit source]

Statistics, Estimators, Unbiased Estimators, Variance and Consistency of an Estimator

[edit | edit source]

A statistic is a function from a sample in any space we are interested, and suppose that there a log-likelihood function ; A estimator for is a statistic used to estimate . These definitions don't add much, but since these words are commonplace for staticians worldwide, they are worth defining. Note that any statistic or estimator is itself a random variable.

In this section we will use our previous notation for samples. A estimator for is called unbiased if , and asymptotically unbiased if

where is a sequence of samples(as defined in the section about sampling). In classical views of statistics, only estimators that possessed at least one of these properties where admissible, but the Maximum Likehood Estimators (MLE) are not necessarily unbiased nor asymptotically unbiased.

TODO: Examples of biased estimatives (the easier one is the variance of a normal distribution).

The variance of a estimator is another important thing; We expect that, as our data grows in size, the variance of the estimator goes to zero, for more data means more information, more information means more certanity. So, we say that the estimator is consistent if

but such a strong condition is rarely satisfied; Usually, we have to contempt ourselves with weakly consistent estimators, estimators where

Reparametrization and Invariance

[edit | edit source]

TODO: Show how to reparametrize a gaussian using the coefficient of variation instead of the variance. Show how to reparametrize to .

A estimator of is called invariant under the transformation if the estimator of the parameter is . One important property of maximum likelihood estimators is that they are invariant under strictly monotone transformations.

TODO: Show estimators that aren't invariant under monotone functions.

Properties of Maximum Likelihood Estimates

[edit | edit source]

TODO: Show that the MLE are consistent.

Examples of finding Maximum Likelihood Estimators

[edit | edit source]

TODO: Example with a single parameter, continuous case.

TODO: Example of discrete case. Mark-Recapture Sampling.

A wildlife biologist captures and tags n1=300 ducks, and then releases them. After allowing time for mixing of the tagged birds with the population, a second sample of n2=200 birds is taken. It is found that m2=10 of these birds are found to be tagged.

Write down a binomial likelihood for m2.

Plot the likelihood of m2 as a function of the population abundance, N.

What is a maximum likelihood estimator of N?

TODO: Example with a two or more parameters, continuous case. that is nothing

TODO: Example Finding MLE's with a numerical method.

Often in applied problems the likelihood is so complicated that a closed form solution for the MLE cannot be found. In such cases, one must use numerical methods to find an approximate solution. The Metropolis Algorithm is a commonly used approach.