Main Page: Difference between revisions

From formulasearchengine
Jump to navigation Jump to search
No edit summary
No edit summary
 
(495 intermediate revisions by more than 100 users not shown)
Line 1: Line 1:
{{ distinguish|a priori probability}}
This is a preview for the new '''MathML rendering mode''' (with SVG fallback), which is availble in production for registered users.


{{No footnotes|article|date=February 2008}}
If you would like use the '''MathML''' rendering mode, you need a wikipedia user account that can be registered here [[https://en.wikipedia.org/wiki/Special:UserLogin/signup]]
* Only registered users will be able to execute this rendering mode.
* Note: you need not enter a email address (nor any other private information). Please do not use a password that you use elsewhere.


{{Bayesian statistics}}
Registered users will be able to choose between the following three rendering modes:


In [[Bayesian probability|Bayesian]] [[statistical inference]], a '''prior probability distribution''', often called simply the '''prior''', of an uncertain quantity ''p'' is the [[probability distribution]] that would express one's uncertainty about ''p'' before some evidence is taken into account. For example, ''p'' could be the proportion of voters who will vote for a particular politician in a future election. It is meant to attribute uncertainty rather than randomness to the uncertain quantity. The unknown quantity may be a [[parameter]] or [[latent variable]].
'''MathML'''
:<math forcemathmode="mathml">E=mc^2</math>


One applies [[Bayes' theorem]], multiplying the prior by the [[likelihood function]] and then normalizing, to get the ''[[posterior probability distribution]]'', which is the conditional distribution of the uncertain quantity given the data.
<!--'''PNG'''  (currently default in production)
:<math forcemathmode="png">E=mc^2</math>


A prior is often the purely subjective assessment of an experienced expert. Some will choose a ''[[conjugate prior]]'' when they can, to make calculation of the posterior distribution easier.
'''source'''
:<math forcemathmode="source">E=mc^2</math> -->


Parameters of prior distributions are called ''[[hyperparameter]]s,'' to distinguish them from parameters of the model of the underlying data. For instance, if one is using a [[beta distribution]] to model the distribution of the parameter ''p'' of a [[Bernoulli distribution]], then:
<span style="color: red">Follow this [https://en.wikipedia.org/wiki/Special:Preferences#mw-prefsection-rendering link] to change your Math rendering settings.</span> You can also add a [https://en.wikipedia.org/wiki/Special:Preferences#mw-prefsection-rendering-skin Custom CSS] to force the MathML/SVG rendering or select different font families. See [https://www.mediawiki.org/wiki/Extension:Math#CSS_for_the_MathML_with_SVG_fallback_mode these examples].
* ''p'' is a parameter of the underlying system (Bernoulli distribution), and
* ''α'' and ''β'' are parameters of the prior distribution (beta distribution), hence ''hyper''parameters.


== Informative priors ==
==Demos==


An ''informative prior'' expresses specific, definite information about a variable.
Here are some [https://commons.wikimedia.org/w/index.php?title=Special:ListFiles/Frederic.wang demos]:
An example is a prior distribution for the temperature at noon tomorrow.
A reasonable approach is to make the prior a [[normal distribution]] with [[expected value]] equal to today's noontime temperature, with [[variance]] equal to the day-to-day variance of atmospheric temperature,
or a distribution of the temperature for that day of the year.


This example has a property in common with many priors,
namely, that the posterior from one problem (today's temperature) becomes the prior for another problem (tomorrow's temperature); pre-existing evidence which has already been taken into account is part of the prior and as more evidence accumulates the prior is determined largely by the evidence rather than any original assumption, provided that the original assumption admitted the possibility of what the evidence is suggesting. The terms "prior" and "posterior" are generally relative to a specific datum or observation.


== Uninformative priors ==<!-- This section is linked from [[Non-informative prior]] -->
* accessibility:
** Safari + VoiceOver: [https://commons.wikimedia.org/wiki/File:VoiceOver-Mac-Safari.ogv video only], [[File:Voiceover-mathml-example-1.wav|thumb|Voiceover-mathml-example-1]], [[File:Voiceover-mathml-example-2.wav|thumb|Voiceover-mathml-example-2]], [[File:Voiceover-mathml-example-3.wav|thumb|Voiceover-mathml-example-3]], [[File:Voiceover-mathml-example-4.wav|thumb|Voiceover-mathml-example-4]], [[File:Voiceover-mathml-example-5.wav|thumb|Voiceover-mathml-example-5]], [[File:Voiceover-mathml-example-6.wav|thumb|Voiceover-mathml-example-6]], [[File:Voiceover-mathml-example-7.wav|thumb|Voiceover-mathml-example-7]]
** [https://commons.wikimedia.org/wiki/File:MathPlayer-Audio-Windows7-InternetExplorer.ogg Internet Explorer + MathPlayer (audio)]
** [https://commons.wikimedia.org/wiki/File:MathPlayer-SynchronizedHighlighting-WIndows7-InternetExplorer.png Internet Explorer + MathPlayer (synchronized highlighting)]
** [https://commons.wikimedia.org/wiki/File:MathPlayer-Braille-Windows7-InternetExplorer.png Internet Explorer + MathPlayer (braille)]
** NVDA+MathPlayer: [[File:Nvda-mathml-example-1.wav|thumb|Nvda-mathml-example-1]], [[File:Nvda-mathml-example-2.wav|thumb|Nvda-mathml-example-2]], [[File:Nvda-mathml-example-3.wav|thumb|Nvda-mathml-example-3]], [[File:Nvda-mathml-example-4.wav|thumb|Nvda-mathml-example-4]], [[File:Nvda-mathml-example-5.wav|thumb|Nvda-mathml-example-5]], [[File:Nvda-mathml-example-6.wav|thumb|Nvda-mathml-example-6]], [[File:Nvda-mathml-example-7.wav|thumb|Nvda-mathml-example-7]].
** Orca: There is ongoing work, but no support at all at the moment [[File:Orca-mathml-example-1.wav|thumb|Orca-mathml-example-1]], [[File:Orca-mathml-example-2.wav|thumb|Orca-mathml-example-2]], [[File:Orca-mathml-example-3.wav|thumb|Orca-mathml-example-3]], [[File:Orca-mathml-example-4.wav|thumb|Orca-mathml-example-4]], [[File:Orca-mathml-example-5.wav|thumb|Orca-mathml-example-5]], [[File:Orca-mathml-example-6.wav|thumb|Orca-mathml-example-6]], [[File:Orca-mathml-example-7.wav|thumb|Orca-mathml-example-7]].
** From our testing, ChromeVox and JAWS are not able to read the formulas generated by the MathML mode.


An ''uninformative prior'' expresses vague or general information about a variable.
==Test pages ==
The term "uninformative prior" may be somewhat of a misnomer; often, such a prior might be called a ''not very informative prior'', or an ''objective prior'', i.e. one that's not subjectively elicited.
Uninformative priors can express "objective" information such as "the variable is positive" or "the variable is less than some limit".


The simplest and oldest rule for determining a non-informative prior is the [[principle of indifference]], which assigns equal probabilities to all possibilities.
To test the '''MathML''', '''PNG''', and '''source''' rendering modes, please go to one of the following test pages:
*[[Displaystyle]]
*[[MathAxisAlignment]]
*[[Styling]]
*[[Linebreaking]]
*[[Unique Ids]]
*[[Help:Formula]]


In parameter estimation problems, the use of an uninformative prior typically yields results which are not too different from conventional statistical analysis, as the likelihood function often yields more information than the uninformative prior.
*[[Inputtypes|Inputtypes (private Wikis only)]]
 
*[[Url2Image|Url2Image (private Wikis only)]]
Some attempts have been made at finding [[a priori probability|a priori probabilities]], i.e. probability distributions in some sense logically required by the nature of one's state of uncertainty; these are a subject of philosophical controversy, with Bayesians being roughly divided into two schools: "objective Bayesians", who believe such priors exist in many useful situations, and "subjective Bayesians" who believe that in practice priors usually represent subjective judgements of opinion that cannot be rigorously justified (Williamson 2010).  Perhaps the strongest arguments for objective Bayesianism were given by [[Edwin T. Jaynes]], based mainly on the consequences of symmetries and on the principle of maximum entropy.
==Bug reporting==
 
If you find any bugs, please report them at [https://bugzilla.wikimedia.org/enter_bug.cgi?product=MediaWiki%20extensions&component=Math&version=master&short_desc=Math-preview%20rendering%20problem Bugzilla], or write an email to math_bugs (at) ckurs (dot) de .
As an example of an a priori prior, due to Jaynes (2003), consider a situation in which one knows a ball has been hidden under one of three cups, A, B or C, but no other information is available about its location.  In this case a ''uniform prior'' of ''p''(''A'')&nbsp;= ''p''(''B'')&nbsp;= ''p''(''C'')&nbsp;= 1/3 seems intuitively like the only reasonable choice.  More formally, we can see that the problem remains the same if we swap around the labels ("A", "B" and "C") of the cups.  It would therefore be odd to choose a prior for which a permutation of the labels would cause a change in our predictions about which cup the ball will be found under; the uniform prior is the only one which preserves this invariance.  If one accepts this invariance principle then one can see that the uniform prior is the logically correct prior to represent this state of knowledge.  It should be noted that this prior is "objective" in the sense of being the correct choice to represent a particular state of knowledge, but it is not objective in the sense of being an observer-independent feature of the world: in reality the ball exists under a particular cup, and it only makes sense to speak of probabilities in this situation if there is an observer with limited knowledge about the system.
 
As a more contentious example, Jaynes published an argument (Jaynes 1968) based on [[Lie group]]s that
suggests that the prior representing complete uncertainty about a probability should be the [[Haldane prior]] ''p''<sup>&minus;1</sup>(1&nbsp;&minus;&nbsp;''p'')<sup>&minus;1</sup>.  The example Jaynes gives is of finding a chemical in a lab and asking whether it will dissolve in water in repeated experiments.  The Haldane prior<ref>This prior was proposed by [[J.B.S. Haldane]] in "A note on inverse probability", Mathematical Proceedings of the Cambridge Philosophical Society 28, 55–61, 1932, available online at http://journals.cambridge.org/action/displayAbstract?aid=1733860. See also J. Haldane, "The precision of observed values of small frequencies", Biometrika, 35:297–300, 1948, available online at http://www.jstor.org/pss/2332350.</ref> gives by far the most weight to <math>p=0</math> and <math>p=1</math>, indicating that the sample will either dissolve every time or never dissolve, with equal probability.  However, if one has observed samples of the chemical to dissolve in one experiment and not to dissolve in another experiment then this prior is updated to the [[uniform distribution (continuous)|uniform distribution]] on the interval [0, 1].  This is obtained by applying [[Bayes' theorem]] to the data set consisting of one observation of dissolving and one of not dissolving, using the above prior.  The Haldane prior has been criticized{{By whom|date=July 2010}} on the grounds that it yields an improper posterior distribution that puts 100% of the probability content at either ''p'' = 0 or at ''p'' = 1 if a finite number of observations have given the same result.  The [[Jeffreys prior]] ''p''<sup>&minus;1/2</sup>(1&nbsp;&minus;&nbsp;''p'')<sup>&minus;1/2</sup> is therefore preferred{{By whom|date=July 2010}} (see below).
 
Priors can be constructed which are proportional to the [[Haar measure]] if the parameter space ''X'' carries a [[transformation group|natural group structure]] which leaves invariant our Bayesian state of knowledge (Jaynes, 1968). This can be seen as a generalisation of the invariance principle used to justify the uniform prior over the three cups in the example above.  For example, in physics we might expect that an experiment will give the same results regardless of our choice of the origin of a coordinate system. This induces the group structure of the [[translation group]] on ''X'', which determines the prior probability as a constant [[improper prior]]. Similarly, some measurements are naturally invariant to the choice of an arbitrary scale (i.e., it doesn't matter if we use centimeters or inches, we should get results that are physically the same). In such a case, the scale group is the natural group structure, and the corresponding prior on ''X'' is proportional to 1/''x''. It sometimes matters whether we use the left-invariant or right-invariant Haar measure. For example, the left and right invariant Haar measures on the [[affine group]] are not equal. Berger (1985, p.&nbsp;413) argues that the right-invariant Haar measure is the correct choice.
 
Another idea, championed by [[Edwin T. Jaynes]], is to use the [[principle of maximum entropy]] (MAXENT). The motivation is that the [[Shannon entropy]] of a probability distribution measures the amount of information contained in the distribution. The larger the entropy, the less information is provided by the distribution. Thus, by maximizing the entropy over a suitable set of probability distributions on ''X'', one finds the distribution that is least informative in the sense that it contains the least amount of information consistent with the constraints that define the set. For example, the maximum entropy prior on a discrete space, given only that the probability is normalized to 1, is the prior that assigns equal probability to each state. And in the continuous case, the maximum entropy prior given that the density is normalized with mean zero and variance unity is the standard [[normal distribution]].  The principle of ''[[minxent|minimum cross-entropy]]'' generalizes MAXENT to the case of "updating" an arbitrary prior distribution with suitability constraints in the maximum-entropy sense.
 
A related idea, [[reference prior]]s, was introduced by [[José-Miguel Bernardo]]. Here, the idea is to maximize the expected [[Kullback–Leibler divergence]] of the posterior distribution relative to the prior. This maximizes the expected posterior information about ''X'' when the prior density is ''p''(''x''); thus, in some sense, ''p''(''x'') is the "least informative" prior about X. The reference prior is defined in the asymptotic limit, i.e., one considers the limit of the priors so obtained as the number of data points goes to infinity. Reference priors are often the objective prior of choice in multivariate problems, since other rules (e.g., [[Jeffreys prior|Jeffreys' rule]]) may result in priors with problematic behavior.
 
Objective prior distributions may also be derived from other principles, such as [[information theory|information]] or [[coding theory]] (see e.g. [[minimum description length]]) or [[frequentist statistics]] (see [[frequentist matching]]).
 
Philosophical problems associated with uninformative priors are associated with the choice of an appropriate metric, or measurement scale. Suppose we want a prior for the running speed of a runner who is unknown to us. We could specify, say, a normal distribution as the prior for his speed, but alternatively we could specify a normal prior for the time he takes to complete 100 metres, which is proportional to the reciprocal of the first prior. These are very different priors, but it is not clear which is to be preferred.  Jaynes' often-overlooked method of transformation groups can answer this question in some situations.<ref>Jaynes (1968), pp. 17, see also Jaynes (2003), chapter 12.  Note that chapter 12 is not available in the online preprint but can be previewed via Google Books.</ref>
 
Similarly, if asked to estimate an unknown proportion between 0 and 1, we might say that all proportions are equally likely and use a uniform prior. Alternatively, we might say that all orders of magnitude for the proportion are equally likely, the '''{{visible anchor|logarithmic prior}}''', which is the uniform prior on the logarithm of proportion. The [[Jeffreys prior]] attempts to solve this problem by computing a prior which expresses the same belief no matter which metric is used. The Jeffreys prior for an unknown proportion ''p'' is ''p''<sup>&minus;1/2</sup>(1&nbsp;&minus;&nbsp;''p'')<sup>&minus;1/2</sup>, which differs from Jaynes' recommendation.
 
Priors based on notions of [[algorithmic probability]] are used in [[inductive inference]] as a basis for induction in very general settings.
 
Practical problems associated with uninformative priors include the requirement that the posterior distribution be proper. The usual uninformative priors on continuous, unbounded variables are improper. This need not be a problem if the posterior distribution is proper. Another issue of importance is that if an uninformative prior is to be used ''routinely'', i.e., with many different data sets, it should have good [[frequentist]] properties. Normally a [[Bayesian probability|Bayesian]] would not be concerned with such issues, but it can be important in this situation. For example, one would want any [[decision theory|decision rule]] based on the posterior distribution to be [[admissible decision rule|admissible]] under the adopted loss function. Unfortunately, admissibility is often difficult to check, although some results are known (e.g., Berger and Strawderman 1996). The issue is particularly acute with [[hierarchical Bayes model]]s; the usual priors (e.g., Jeffreys' prior) may give badly inadmissible decision rules if employed at the higher levels of the hierarchy.
 
==Improper priors==
 
If Bayes' theorem is written as
:<math>P(A_i|B) = \frac{P(B | A_i) P(A_i)}{\sum_j P(B|A_j)P(A_j)}\, ,</math>
then it is clear that the same result would be obtained if all the prior probabilities ''P''(''A''<sub>''i''</sub>) and ''P''(''A''<sub>''j''</sub>) were multiplied by a given constant; the same would be true for a [[continuous random variable]].  If the summation in the denominator converges, the posterior probabilities will still sum (or integrate) to 1 even if the prior values do not, and so the priors may only need to be specified in the correct proportion. Taking this idea further, in many cases the sum or integral of the prior values may not even need to be finite to get sensible answers for the posterior probabilities. When this is the case, the prior is called an '''improper prior'''.  However, the posterior distribution need not be a proper distribution if the prior is improper. This is clear from the case where event ''B'' is independent of all of the ''A''<sub>''j''</sub>.
 
Some statisticians{{Citation needed|date=December 2008}} use improper priors as [[uninformative prior]]s.  For example, if they need a prior distribution for the mean and variance of a random variable, they may assume ''p''(''m'',&nbsp;''v'')&nbsp;~&nbsp;1/''v'' (for ''v''&nbsp;>&nbsp;0) which would suggest that any value for the mean is "equally likely" and that a value for the positive variance becomes "less likely" in inverse proportion to its value.  Many authors (Lindley, 1973; De Groot, 1937; Kass and Wasserman, 1996){{Citation needed|date=December 2008}} warn against the danger of over-interpreting those priors since they are not probability densities. The only relevance they have is found in the corresponding posterior, as long as it is well-defined for all observations. (The [[Beta distribution#Haldane.27s prior probability .28 Beta.280.2C0.29 .29|Haldane prior]] is a typical counterexample.{{Clarify|reason=counterexample of what?|date=May 2011}}{{Citation needed|date=May 2011}})
 
=== Examples ===
Examples of improper priors include:
* Beta(0,0), the [[beta distribution]] for α=0, β=0.
* The [[uniform distribution (continuous)|uniform distribution]] on an infinite interval (i.e., a half-line or the entire real line).
* The logarithmic prior on the positive reals.{{Citation needed|date=October 2010}}
 
==Other priors==
The concept of [[algorithmic probability]] provides a route to specifying prior probabilities based on the relative complexity of the alternative models being considered.
 
== Notes ==
 
<references/>
 
== References ==
 
* {{cite book |author=Rubin, Donald B.; [[Andrew Gelman|Gelman, Andrew]]; John B. Carlin; Stern, Hal |title=Bayesian Data Analysis |edition=2nd |publisher=Chapman & Hall/CRC |location=Boca Raton |year=2003 |pages= |isbn=1-58488-388-X |doi= |accessdate= |mr=2027492 }}
* {{cite book |last=Berger |first=James O. |title=Statistical decision theory and Bayesian analysis |publisher=Springer-Verlag |location=Berlin |year=1985 |pages= |isbn=0-387-96098-8 |doi= |accessdate= |mr=0804611 }}
* {{cite journal
|first1=James O. |last1=Berger
|first2=William E. |last2=Strawderman
|title=Choice of hierarchical priors: admissibility in estimation of normal means
|journal=[[Annals of Statistics]]
|volume=24 |issue=3 |pages=931&ndash;951 |year=1996
|doi=10.1214/aos/1032526950
|mr=1401831 | zbl = 0865.62004
}}
* {{cite journal |first=Jose M. |last=Bernardo
|title=Reference Posterior Distributions for Bayesian Inference
|journal=[[Journal of the Royal Statistical Society]], Series B
|volume=41 |issue= 2 |pages=113&ndash;147 |year=1979
|mr=0547240 | jstor = 2985028
}}
* {{cite journal | title=The formal definition of reference priors
|author1=James O. Berger
|authorlink1=James Berger (statistician)
|author2=José M. Bernardo
|authorlink2=José-Miguel Bernardo
|author3=Dongchu Sun
|journal=Annals of Statistics
|year=2009
|volume=37
|issue=2
|pages=905–938
|arxiv=0904.0156 | doi=10.1214/07-AOS587
}}
* {{cite journal|last=Jaynes |first=Edwin T. |authorlink= Edwin T. Jaynes |title=Prior Probabilities |journal=IEEE Transactions on Systems Science and Cybernetics |volume=4 |issue=3 |pages=227&ndash;241 |date=Sep 1968 |doi=10.1109/TSSC.1968.300117 |url=http://bayes.wustl.edu/etj/articles/prior.pdf |accessdate=2009-03-27}}
** Reprinted in {{cite book |author=Rosenkrantz, Roger D. |title=E. T. Jaynes: papers on probability, statistics, and statistical physics |publisher=Kluwer Academic Publishers |location=Boston |year=1989 |isbn=90-277-1448-7 |doi= |accessdate= |pages=116&ndash;130}}
* {{cite book |last=Jaynes |first=Edwin T. |authorlink= Edwin T. Jaynes |title= Probability Theory: The Logic of Science |publisher=Cambridge University Press |year=2003 |pages= |isbn=0-521-59271-2 |doi= |accessdate= |url=http://www-biba.inrialpes.fr/Jaynes/prob.html }}
* {{cite journal|last=Williamson |first=Jon |title=review of Bruno di Finetti. Philosophical Lectures on Probability |journal=Philosophia Mathematica |volume=18 |issue=1 |pages=130&ndash;135 |year=2010 |doi=10.1093/philmat/nkp019  |url=http://www.kent.ac.uk/secl/philosophy/jw/2009/deFinetti.pdf |accessdate=2010-07-02}}
 
{{DEFAULTSORT:Prior Probability}}
[[Category:Bayesian statistics]]
[[Category:Probability assessment]]

Latest revision as of 22:52, 15 September 2019

This is a preview for the new MathML rendering mode (with SVG fallback), which is availble in production for registered users.

If you would like use the MathML rendering mode, you need a wikipedia user account that can be registered here [[1]]

  • Only registered users will be able to execute this rendering mode.
  • Note: you need not enter a email address (nor any other private information). Please do not use a password that you use elsewhere.

Registered users will be able to choose between the following three rendering modes:

MathML

E=mc2


Follow this link to change your Math rendering settings. You can also add a Custom CSS to force the MathML/SVG rendering or select different font families. See these examples.

Demos

Here are some demos:


Test pages

To test the MathML, PNG, and source rendering modes, please go to one of the following test pages:

Bug reporting

If you find any bugs, please report them at Bugzilla, or write an email to math_bugs (at) ckurs (dot) de .