词条 | Wishart distribution |
释义 |
| name =Wishart | type =density | pdf_image = | cdf_image = | notation ={{math|X ~ Wp(V, n)}} | parameters ={{math|n > p − 1}} degrees of freedom (real) {{math|V > 0}} scale matrix ({{math|p × p}} pos. def) | support ={{math|X(p × p)}} positive definite matrix | pdf =
| cdf = | mean =| | median = | mode ={{math|(n − p − 1)V}} for {{math|n ≥ p + 1}} | variance = | skewness = | kurtosis = | entropy =see below | mgf = | char = }} In statistics, the Wishart distribution is a generalization to multiple dimensions of the gamma distribution. It is named in honor of John Wishart, who first formulated the distribution in 1928.[1] It is a family of probability distributions defined over symmetric, nonnegative-definite matrix-valued random variables (“random matrices”). These distributions are of great importance in the estimation of covariance matrices in multivariate statistics. In Bayesian statistics, the Wishart distribution is the conjugate prior of the inverse covariance-matrix of a multivariate-normal random-vector. DefinitionSuppose {{mvar|X}} is an {{math|p × n}} matrix, each column of which is independently drawn from a {{mvar|p}}-variate normal distribution with zero mean: Then the Wishart distribution is the probability distribution of the {{math|p × p}} random matrix known as the scatter matrix. One indicates that {{mvar|S}} has that probability distribution by writing The positive integer {{mvar|n}} is the number of degrees of freedom. Sometimes this is written {{math|W(V, p, n)}}. For {{math|n ≥ p}} the matrix {{mvar|S}} is invertible with probability {{math|1}} if {{mvar|V}} is invertible. If {{math|p {{=}} V {{=}} 1}} then this distribution is a chi-squared distribution with {{mvar|n}} degrees of freedom. OccurrenceThe Wishart distribution arises as the distribution of the sample covariance matrix for a sample from a multivariate normal distribution. It occurs frequently in likelihood-ratio tests in multivariate statistical analysis. It also arises in the spectral theory of random matrices{{Citation needed|date=October 2010}} and in multidimensional Bayesian analysis.[2] It is also encountered in wireless communications, while analyzing the performance of Rayleigh fading MIMO wireless channels .[3] Probability density functionThe Wishart distribution can be characterized by its probability density function as follows: Let {{math|X}} be a {{math|p × p}} symmetric matrix of random variables that is positive definite. Let {{math|V}} be a (fixed) symmetric positive definite matrix of size {{math|p × p}}. Then, if {{math|n ≥ p}}, {{math|X}} has a Wishart distribution with {{mvar|n}} degrees of freedom if it has the probability density function where is the determinant of and {{math|Γp}} is the multivariate gamma function defined as The density above is not the joint density of all the elements of the random matrix {{math|X}} (such {{nowrap|-dimensional}} density does not exists because of the symmetry constrains ), it is rather the joint density of elements for ([1], page 38). Also, the density formula above applies only to positive definite matrices for other matrices the density is equal to zero. The joint-eigenvalue density for eigenvalues is where is a constant. In fact the above definition can be extended to any real {{math|n > p − 1}}. If {{math|n ≤ p − 1}}, then the Wishart no longer has a density—instead it represents a singular distribution that takes values in a lower-dimension subspace of the space of {{math|p × p}} matrices.[4] Use in Bayesian statisticsIn Bayesian statistics, in the context of the multivariate normal distribution, the Wishart distribution is the conjugate prior to the precision matrix {{math|Ω {{=}} Σ−1}}, where {{math|Σ}} is the covariance matrix.[6]{{rp|135}} Choice of parametersThe least informative, proper Wishart prior is obtained by setting {{math|n {{=}} p}}.{{Citation needed|date=June 2014}} The prior mean of {{math|Wp(V, n)}} is {{math|nV}}, suggesting that a reasonable choice for {{math|V}} would be {{math|n−1Σ0}}, where {{math|Σ0}} is some prior guess for the covariance matrix. PropertiesLog-expectationThe following formula plays a role in variational Bayes derivations for Bayes networks involving the Wishart distribution: [6]{{rp|693}} where is the multivariate digamma function (the derivative of the log of the multivariate gamma function). Log-varianceThe following variance computation could be of help in Bayesian statistics: where is the trigamma function. This comes up when computing the Fisher information of the Wishart random variable. EntropyThe information entropy of the distribution has the following formula:[6]{{rp|693}} where {{math|B(V, n)}} is the normalizing constant of the distribution: This can be expanded as follows: Cross-entropyThe cross entropy of two Wishart distributions with parameters and with parameters is Note that when we recover the entropy. KL-divergenceThe Kullback–Leibler divergence of from is Characteristic functionThe characteristic function of the Wishart distribution is In other words, where {{math|E[⋅]}} denotes expectation. (Here {{math|Θ}} and {{math|I}} are matrices the same size as {{math|V}}({{math|I}} is the identity matrix); and {{mvar|i}} is the square root of −1).[5] Since the determinant's range contains a closed line through the origin for matrix dimensions greater than two, the above formula is only correct for small values of the Fourier variable. (see https://arxiv.org/pdf/1901.09347.pdf) TheoremIf a {{math|p × p}} random matrix {{math|X}} has a Wishart distribution with {{mvar|m}} degrees of freedom and variance matrix {{math|V}} — write — and {{math|C}} is a {{math|q × p}} matrix of rank {{mvar|q}}, then [6] Corollary 1If {{math|z}} is a nonzero {{math|p × 1}} constant vector, then:[6] In this case, is the chi-squared distribution and (note that is a constant; it is positive because {{math|V}} is positive definite). Corollary 2Consider the case where {{math|zT {{=}} (0, ..., 0, 1, 0, ..., 0)}} (that is, the {{mvar|j}}-th element is one and all others zero). Then corollary 1 above shows that gives the marginal distribution of each of the elements on the matrix's diagonal. George Seber points out that the Wishart distribution is not called the “multivariate chi-squared distribution” because the marginal distribution of the off-diagonal elements is not chi-squared. Seber prefers to reserve the term multivariate for the case when all univariate marginals belong to the same family.[7]Estimator of the multivariate normal distributionThe Wishart distribution is the sampling distribution of the maximum-likelihood estimator (MLE) of the covariance matrix of a multivariate normal distribution.[8] A derivation of the MLE uses the spectral theorem. Bartlett decompositionThe Bartlett decomposition of a matrix {{math|X}} from a {{mvar|p}}-variate Wishart distribution with scale matrix {{math|V}} and {{mvar|n}} degrees of freedom is the factorization: where {{math|L}} is the Cholesky factor of {{math|V}}, and: where and {{math|nij ~ N(0, 1)}} independently.[9] This provides a useful method for obtaining random samples from a Wishart distribution.[10] Marginal distribution of matrix elementsLet {{math|V}} be a {{math|2 × 2}} variance matrix characterized by correlation coefficient {{math|−1 < ρ < 1}} and {{math|L}} its lower Cholesky factor: Multiplying through the Bartlett decomposition above, we find that a random sample from the {{math|2 × 2}} Wishart distribution is The diagonal elements, most evidently in the first element, follow the {{math|χ2}} distribution with {{mvar|n}} degrees of freedom (scaled by {{math|σ2}}) as expected. The off-diagonal element is less familiar but can be identified as a normal variance-mean mixture where the mixing density is a {{math|χ2}} distribution. The corresponding marginal probability density for the off-diagonal element is therefore the variance-gamma distribution where {{math|Kν(z)}} is the modified Bessel function of the second kind.[11] Similar results may be found for higher dimensions, but the interdependence of the off-diagonal correlations becomes increasingly complicated. It is also possible to write down the moment-generating function even in the noncentral case (essentially the nth power of Craig (1936)[12] equation 10) although the probability density becomes an infinite sum of Bessel functions. The range of the shape parameterIt can be shown [13] that the Wishart distribution can be defined if and only if the shape parameter {{math|n}} belongs to the set This set is named after Gindikin, who introduced it[14] in the seventies in the context of gamma distributions on homogeneous cones. However, for the new parameters in the discrete spectrum of the Gindikin ensemble, namely, the corresponding Wishart distribution has no Lebesgue density. Relationships to other distributions
See also{{Colbegin}}
References1. ^1 {{cite journal |first=J. |last=Wishart |authorlink=John Wishart (statistician) |title=The generalised product moment distribution in samples from a normal multivariate population |journal=Biometrika |volume=20A |issue=1–2 |pages=32–52 |year=1928 |doi=10.1093/biomet/20A.1-2.32 |jfm=54.0565.02 |jstor=2331939}} 2. ^{{cite book |last=Gelman |first=Andrew |date=2003 |title=Bayesian Data Analysis |url=http://www.stat.columbia.edu/~gelman/book/ |publisher=Chapman & Hall |page=582 |isbn=158488388X |access-date=3 June 2015 |location=Boca Raton, Fla. |edition=2nd}} 3. ^{{cite journal| last=Zanella| first=A.|author2=Chiani, M. |author3=Win, M.Z. |title=On the marginal distribution of the eigenvalues of wishart matrices| journal=IEEE Transactions on Communications|date=April 2009| volume=57| issue=4| pages=1050–1060 | doi=10.1109/TCOMM.2009.04.070143}} 4. ^{{Cite journal | doi = 10.1214/aos/1176325375| title = On Singular Wishart and Singular Multivariate Beta Distributions| journal = The Annals of Statistics| volume = 22| pages = 395–405| year = 1994| last1 = Uhlig | first1 = H. }} 5. ^{{cite book | last = Anderson | first = T. W. | authorlink = T. W. Anderson | title = An Introduction to Multivariate Statistical Analysis | publisher = Wiley Interscience | edition = 3rd | location = Hoboken, N. J. | year = 2003 | page = 259 | isbn = 0-471-36091-0 }} 6. ^1 {{cite book |last=Rao |first=C. R. |title=Linear Statistical Inference and its Applications |location= |publisher=Wiley |year=1965 |page=535 }} 7. ^{{cite book | last = Seber | first = George A. F. | title = Multivariate Observations | publisher = Wiley | year = 2004 | isbn = 978-0471691211 }} 8. ^{{cite book |first=C. |last=Chatfield |first2=A. J. |last2=Collins |year=1980 |title=Introduction to Multivariate Analysis |location=London |publisher=Chapman and Hall |pages=103–108 |isbn=0-412-16030-7 }} 9. ^{{cite book | last = Anderson | first = T. W. | authorlink = T. W. Anderson | title = An Introduction to Multivariate Statistical Analysis | publisher = Wiley Interscience | edition = 3rd | location = Hoboken, N. J. | year = 2003 | page = 257 | isbn = 0-471-36091-0 }} 10. ^{{cite journal |title=Algorithm AS 53: Wishart Variate Generator |first1= W. B. |last1=Smith |first2= R. R. |last2=Hocking |journal=Journal of the Royal Statistical Society, Series C |volume=21 |issue=3 |year=1972 |pages=341–345 |jstor=2346290}} 11. ^{{cite journal | last1 = Pearson | first1 = Karl | author1-link = Karl Pearson | last2 = Jeffery | first2 = G. B. | author2-link = George Barker Jeffery | last3 = Elderton | first3 = Ethel M. | author3-link = Ethel M. Elderton | title = On the Distribution of the First Product Moment-Coefficient, in Samples Drawn from an Indefinitely Large Normal Population | journal = Biometrika | volume = 21 | issue = | pages = 164–201 | publisher = Biometrika Trust | date = December 1929 | jstor = 2332556 | doi = 10.2307/2332556}} 12. ^{{cite journal | last = Craig | first = Cecil C. | title = On the Frequency Function of xy | journal = Ann. Math. Statist. | volume = 7 | issue = | pages = 1–15 | year = 1936 | url = http://projecteuclid.org/euclid.aoms/1177732541 | doi = 10.1214/aoms/1177732541}} 13. ^{{cite journal |doi=10.1214/aop/1176990455 |last=Peddada and Richards |first1=Shyamal Das |last2=Richards |first2=Donald St. P. |title=Proof of a Conjecture of M. L. Eaton on the Characteristic Function of the Wishart Distribution, |journal=Annals of Probability |volume=19 |issue=2 |pages=868–874 |year=1991 }} 14. ^{{cite journal |doi=10.1007/BF01078179 |first=S.G. |last=Gindikin |title=Invariant generalized functions in homogeneous domains, |journal=Funct. Anal. Appl. |volume=9 |issue=1 |pages=50–52 |year=1975}} 15. ^{{cite journal |first=Paul S. |last=Dwyer |title=Some Applications of Matrix Derivatives in Multivariate Analysis |journal=J. Amer. Statist. Assoc. |year=1967 |volume=62 |issue=318 |pages=607–625 |jstor=2283988 }} 16. ^1 2 3 {{cite book |first=C. M. |last=Bishop |title=Pattern Recognition and Machine Learning |location= |publisher=Springer |year=2006 }} External links
6 : Continuous distributions|Multivariate continuous distributions|Covariance and correlation|Random matrices|Conjugate prior distributions|Exponential family distributions |
随便看 |
|
开放百科全书收录14589846条英语、德语、日语等多语种百科知识,基本涵盖了大多数领域的百科知识,是一部内容自由、开放的电子版国际百科全书。