A prior probability distribution of an uncertain quantity, often simply called the prior, is its assumed probability distribution before some evidence is taken into account. For example, the prior could be the probability distribution representing the relative proportions of voters who will vote for a particular politician in a future election. The unknown quantity may be a parameter of the model or a latent variable rather than an observable variable.
In Bayesian statistics, Bayes' rule prescribes how to update the prior with new information to obtain the posterior probability distribution, which is the conditional distribution of the uncertain quantity given new data. Historically, the choice of priors was often constrained to a conjugate family of a given likelihood function, for that it would result in a tractable posterior of the same family. The widespread availability of Markov chain Monte Carlo methods, however, has made this less of a concern.
There are many ways to construct a prior distribution.[1] In some cases, a prior may be determined from past information, such as previous experiments. A prior can also be elicited from the purely subjective assessment of an experienced expert.[2][3] When no information is available, an uninformative prior may be adopted as justified by the principle of indifference.[4][5] In modern applications, priors are also often chosen for their mechanical properties, such as regularization and feature selection.[6][7][8]
The prior distributions of model parameters will often depend on parameters of their own. Uncertainty about these hyperparameters can, in turn, be expressed as hyperprior probability distributions. For example, if one uses a beta distribution to model the distribution of the parameter p of a Bernoulli distribution, then:
p is a parameter of the underlying system (Bernoulli distribution), and
α and β are parameters of the prior distribution (beta distribution); hence hyperparameters.
In principle, priors can be decomposed into many conditional levels of distributions, so-called hierarchical priors.[9]
^Robert, Christian (1994). "From Prior Information to Prior Distributions". The Bayesian Choice. New York: Springer. pp. 89–136. ISBN 0-387-94296-3.
^Chaloner, Kathryn (1996). "Elicitation of Prior Distributions". In Berry, Donald A.; Stangl, Dalene (eds.). Bayesian Biostatistics. New York: Marcel Dekker. pp. 141–156. ISBN 0-8247-9334-X.
^Mikkola, Petrus; et al. (2023). "Prior Knowledge Elicitation: The Past, Present, and Future". Bayesian Analysis. Forthcoming. doi:10.1214/23-BA1381. hdl:11336/183197. S2CID 244798734.
^Zellner, Arnold (1971). "Prior Distributions to Represent 'Knowing Little'". An Introduction to Bayesian Inference in Econometrics. New York: John Wiley & Sons. pp. 41–53. ISBN 0-471-98165-6.
^Price, Harold J.; Manson, Allison R. (2001). "Uninformative priors for Bayes' theorem". AIP Conf. Proc. 617: 379–391. doi:10.1063/1.1477060.
^Piironen, Juho; Vehtari, Aki (2017). "Sparsity information and regularization in the horseshoe and other shrinkage priors". Electronic Journal of Statistics. 11 (2): 5018–5051. arXiv:1707.01694. doi:10.1214/17-EJS1337SI.
^Simpson, Daniel; et al. (2017). "Penalising Model Component Complexity: A Principled, Practical Approach to Constructing Priors". Statistical Science. 32 (1): 1–28. arXiv:1403.4630. doi:10.1214/16-STS576. S2CID 88513041.
^Fortuin, Vincent (2022). "Priors in Bayesian Deep Learning: A Review". International Statistical Review. 90 (3): 563–591. doi:10.1111/insr.12502. hdl:20.500.11850/547969. S2CID 234681651.
^Congdon, Peter D. (2020). "Regression Techniques using Hierarchical Priors". Bayesian Hierarchical Models (2nd ed.). Boca Raton: CRC Press. pp. 253–315. ISBN 978-1-03-217715-1.
A priorprobability distribution of an uncertain quantity, often simply called the prior, is its assumed probability distribution before some evidence...
the Bayesian probabilist specifies a priorprobability. This, in turn, is then updated to a posterior probability in the light of new, relevant data (evidence)...
The posterior probability is a type of conditional probability that results from updating the priorprobability with information summarized by the likelihood...
proportions. In Bayesian inference, the beta distribution is the conjugate priorprobability distribution for the Bernoulli, binomial, negative binomial, and geometric...
defendant in a criminal case Priorprobability, in Bayesian statistics Prior knowledge for pattern recognition Saint Prior (4th century), an Egyptian hermit...
x)} is in the same probability distribution family as the priorprobability distribution p ( θ ) {\displaystyle p(\theta )} , the prior and posterior are...
Probability is the branch of mathematics concerning events and numerical descriptions of how likely they are to occur. The probability of an event is a...
In probability theory, conditional probability is a measure of the probability of an event occurring, given that another event (by assumption, presumption...
interpretation of probability, where probability expresses a degree of belief in an event. The degree of belief may be based on prior knowledge about the...
The word probability has been used in a variety of ways since it was first applied to the mathematical study of games of chance. Does probability measure...
information about a probability distribution function. Consider the set of all trial probability distributions that would encode the prior data. According...
theory, algorithmic probability, also known as Solomonoff probability, is a mathematical method of assigning a priorprobability to a given observation...
In Bayesian probability, the Jeffreys prior, named after Sir Harold Jeffreys, is a non-informative prior distribution for a parameter space; its density...
to update the probability for a hypothesis as more evidence or information becomes available. Fundamentally, Bayesian inference uses prior knowledge, in...
In probability theory and statistics, the binomial distribution with parameters n and p is the discrete probability distribution of the number of successes...
the priorprobability distribution is estimated from the data. This approach stands in contrast to standard Bayesian methods, for which the prior distribution...
Frequentist probability or frequentism is an interpretation of probability; it defines an event's probability as the limit of its relative frequency in...
assumptions are made for the prior distribution of the probability. If a trial yields more information, the empirical probability can be improved on by adopting...
catalog of articles in probability theory. For distributions, see List of probability distributions. For journals, see list of probability journals. For contributors...
The probability assigned to the hypothesis before the event is called priorprobability. The probability afterward is called posterior probability. According...
possible outcomes under consideration. In Bayesian probability, this is the simplest non-informative prior. The principle of indifference is meaningless under...
trillion. Note that as remarked above, this argument assumes that the priorprobability for N is flat, or 50% for N1 and 50% for N2 in the absence of any...
sorts due to the general interest in Bayesian probability, because Bayesian methods require a priorprobability distribution and the principle of indifference...
where g(r) represents the priorprobability density distribution of r, which lies in the range 0 to 1. The priorprobability density distribution summarizes...
} In plain English, using Bayesian probability terminology, the above equation can be written as posterior = prior × likelihood evidence {\displaystyle...
the null is close to 100, if the hypothesis was implausible, with a priorprobability of a real effect being 0.1, even the observation of p = 0.001 would...
of priorprobabilities for the parameters—so-called conjugate priors—the posterior can be found analytically. With more arbitrarily chosen priors, the...