Precision (statistics)

[4] Other summary statistics of statistical dispersion also called precision (or imprecision[5][6]) include the reciprocal of the standard deviation,

[9] One particular use of the precision matrix is in the context of Bayesian analysis of the multivariate normal distribution: for example, Bernardo & Smith prefer to parameterise the multivariate normal distribution in terms of the precision matrix, rather than the covariance matrix, because of certain simplifications that then arise.

[10] For instance, if both the prior and the likelihood have Gaussian form, and the precision matrix of both of these exist (because their covariance matrix is full rank and thus invertible), then the precision matrix of the posterior will simply be the sum of the precision matrices of the prior and the likelihood.

As the inverse of a Hermitian matrix, the precision matrix of real-valued random variables, if it exists, is positive definite and symmetrical.

This means that precision matrices tend to be sparse when many of the dimensions are conditionally independent, which can lead to computational efficiencies when working with them.

It also means that precision matrices are closely related to the idea of partial correlation.

The precision matrix plays a central role in generalized least squares, compared to ordinary least squares, where

The term precision in this sense ("mensura praecisionis observationum") first appeared in the works of Gauss (1809) "Theoria motus corporum coelestium in sectionibus conicis solem ambientium" (page 212).

Gauss's definition differs from the modern one by a factor of

He writes, for the density function of a normal distribution with precision

Later Whittaker & Robinson (1924) "Calculus of observations" called this quantity the modulus (of precision), but this term has dropped out of use.