[Home]Likelihood principle

HomePage | Recent Changes | Preferences

The likelihood principle asserts that the information in any sample can be found, if at all, from the likelihood function, that function of [unknown parameters]? which specifies the probability of the sample observed.

Suppose, for example, that we have observed N independent flips of a coin which we regard as having a constant probabilty, p, of falling heads up. The likelihood function is then the product of N factors, each of which is either p or 1-p. If we observe X heads and N-X tails, then the likelihood function is

No multiplicative constant of C(N,X) is included because only the part of the probability which involves the parameter, p, is relevant. In particular, this principle suggests that it does not matter whether you started out planning to observe N trials or you just decided to stop on a whim. The issue of the likelihood principle is still controversial. A deeper discussion of the topic is available in the article about the maximum likelihood principle.

back to statistical theory


HomePage | Recent Changes | Preferences
This page is read-only | View other revisions
Last edited July 3, 2001 2:30 pm by Larry Sanger (diff)
Search: