词条 | Complex random vector |
释义 |
In probability theory and statistics, a complex random vector is typically a tuple of complex-valued random variables, and generally is a random variable taking values in a vector space over the field of complex numbers. If are complex-valued random variables, then the n-tuple is a complex random vector. Complex random variables can always be considered as pairs of real random vectors: their real and imaginary parts. Some concepts of real random vectors have a straightforward generalization to complex random vectors. For example, the definition of the mean of a complex random vector. Other concepts are unique to complex random vectors. Applications of complex random vectors are found in digital signal processing. {{Probability fundamentals}}DefinitionA complex random vector on the probability space is a function such that the vector is a real real random vector on where denotes the real part of and denotes the imaginary part of .[1]{{rp|p. 292}} Cumulative distribution functionThe generalization of the cumulative distribution function from real to complex random variables is not obvious because expressions of the form make no sense. However expessions of the form make sense. Therefore, the cumulative distribution function of a random vector is defined as {{Equation box 1|indent = |title= |equation = {{NumBlk|||{{EquationRef|Eq.1}}}} |cellpadding= 6 |border |border colour = #0073CF |background colour=#F5FFFA}} where . ExpectationAs in the real case the expectation (also called expected value) of a complex random vector is taken component-wise.[1]{{rp|p. 293}} {{Equation box 1|indent = |title= |equation = {{NumBlk|||{{EquationRef|Eq.2}}}} |cellpadding= 6 |border |border colour = #0073CF |background colour=#F5FFFA}} Covariance matrix and pseudo-covariance matrixDefinitionsThe covariance matrix (also called second central moment) contains the covariances between all pairs of components. The covariance matrix of an random vector is an matrix whose th element is the covariance between the i th and the j th random variables.[2]{{rp|p.372}} Unlike in the case of real random variables, the covariance between two random variables involves the complex conjugate of one of the two. Thus the covariance matrix is a Hermitian matrix.[1]{{rp|p. 293}} {{Equation box 1|indent = |title= |equation = {{NumBlk|| |{{EquationRef|Eq.3}}}} |cellpadding= 6 |border |border colour = #0073CF |background colour=#F5FFFA}} The pseudo-covariance matrix (also called relation matrix) is defined as follows. In contrast to the covariance matrix defined above Hermitian transposition gets replaced by transposition in the definition. {{Equation box 1|indent = |title= |equation = {{NumBlk|| |{{EquationRef|Eq.4}}}} |cellpadding= 6 |border |border colour = #0073CF |background colour=#F5FFFA}} PropertiesThe covariance matrix is a hermitian matrix, i.e.[1]{{rp|p. 293}} . The pseudo-covariance matrix is a symmetric matrix, i.e. . The covariance matrix is a positive semidefinite matrix, i.e. . Covariance matrices of real and imaginary partsBy decomposing the random vector into its real part and imaginary part (i.e. ), the matrices and can be related to the covariance matrices of and via the following expressions: and conversely Cross-covariance matrix and pseudo-cross-covariance matrixDefinitionsThe cross-covariance matrix between two complex random vectors is defined as: {{Equation box 1|indent = |title= |equation = {{NumBlk|||{{EquationRef|Eq.5}}}} |cellpadding= 6 |border |border colour = #0073CF |background colour=#F5FFFA}} And the pseudo-cross-covariance matrix is defined as: {{Equation box 1|indent = |title= |equation = {{NumBlk|||{{EquationRef|Eq.6}}}} |cellpadding= 6 |border |border colour = #0073CF |background colour=#F5FFFA}} UncorrelatednessTwo complex random vectors and are called uncorrelated if . Independence{{main|Independence (probability theory)}}Two complex random vectors and are called independent if {{Equation box 1|indent = |title= |equation = {{NumBlk|||{{EquationRef|Eq.7}}}} |cellpadding= 6 |border |border colour = #0073CF |background colour=#F5FFFA}} where and denote the cumulative distribution functions of and as defined in {{EquationNote|Eq.1}} and denotes their joint cumulative distribution function. Independence of and is often denoted by . Written component-wise, and are called independent if . Circular symmetryDefinitionA complex random vector is called circularly symmetric if for every deterministic the distribution of equals the distribution of .[3]{{rp|pp. 500–501}} Properties
Proper complex random vectorsDefinitionA complex random vector is called proper if the following three conditions are all satisfied:[1]{{rp|p. 293}}
Two complex random vectors are called jointly proper is the composite random vector is proper. Properties
Cauchy-Schwarz inequalityThe Cauchy-Schwarz inequality for complex random vectors is . Characteristic functionThe characteristic function of a complex random vector with components is a function defined by:[1]{{rp|p. 295}} See also
References1. ^1 2 3 4 5 6 7 8 9 {{cite book |first=Amos |last=Lapidoth |year=2009 |title=A Foundation in Digital Communication |publisher=Cambridge University Press |isbn=978-0-521-19395-5}} 2. ^{{cite book |first=John A. |last=Gubner |year=2006 |title=Probability and Random Processes for Electrical and Computer Engineers |publisher=Cambridge University Press |isbn=978-0-521-86470-1}} 3. ^1 2 {{cite book |first=David |last=Tse |year=2005 |title=Fundamentals of Wireless Communication |publisher=Cambridge University Press}} 3 : Probability theory|Randomness|Algebra of random variables |
随便看 |
|
开放百科全书收录14589846条英语、德语、日语等多语种百科知识,基本涵盖了大多数领域的百科知识,是一部内容自由、开放的电子版国际百科全书。