Jump to content
 







Main menu
   


Navigation  



Main page
Contents
Current events
Random article
About Wikipedia
Contact us
Donate
 




Contribute  



Help
Learn to edit
Community portal
Recent changes
Upload file
 








Search  

































Create account

Log in
 









Create account
 Log in
 




Pages for logged out editors learn more  



Contributions
Talk
 



















Contents

   



(Top)
 


1 Definition  





2 Computing the pdf/cdf/inverse cdf/random numbers  





3 Applications  



3.1  In model fitting and selection  





3.2  Classifying normal vectors using Gaussian discriminant analysis  





3.3  In signal processing  







4 See also  





5 References  





6 External links  














Generalized chi-squared distribution: Difference between revisions







Add links
 









Article
Talk
 

















Read
Edit
View history
 








Tools
   


Actions  



Read
Edit
View history
 




General  



What links here
Related changes
Upload file
Special pages
Permanent link
Page information
Cite this page
Get shortened URL
Download QR code
Wikidata item
 




Print/export  



Download as PDF
Printable version
 
















Appearance
   

 





Help
 

From Wikipedia, the free encyclopedia
 


Browse history interactively
 Previous edit
Content deleted Content added
No edit summary
→‎See also: no longer exists
 
(47 intermediate revisions by 12 users not shown)
Line 5: Line 5:

| pdf_image = [[File:Generalized chi-square PDF.svg|340px|Generalized chi-square probability density function]]

| pdf_image = [[File:Generalized chi-square PDF.svg|340px|Generalized chi-square probability density function]]

| cdf_image = [[File:Generalized chi-square cumulative distribution function.svg|340px|Generalized chi-square cumulative distribution function]]

| cdf_image = [[File:Generalized chi-square cumulative distribution function.svg|340px|Generalized chi-square cumulative distribution function]]

| parameters = <math>\boldsymbol{\lambda}</math>, vector of weights of chi-square components<br /><math>\boldsymbol{m}</math>, vector of degrees of freedom of chi-square components<br /><math>\boldsymbol{\delta}</math>, vector of non-centrality parameters of chi-square components<br /><math>\sigma</math>, scale of normal term

| notation = <math>\tilde{\chi}(\boldsymbol{w}, \boldsymbol{k}, \boldsymbol{\lambda},s,m)</math>

| parameters = <math>\boldsymbol{w}</math>, vector of weights of noncentral chi-square components<br /><math>\boldsymbol{k}</math>, vector of degrees of freedom of noncentral chi-square components<br /><math>\boldsymbol{\lambda}</math>, vector of non-centrality parameters of chi-square components<br /><math>s</math>, scale of normal term<br /><math>m</math>, offset

| support = <math>x\in\R</math>

| support = <math> x \in \begin{cases} [m, +\infty) \text{ if } w_i \geq 0, s=0,\\ (-\infty,m] \text{ if } w_i \leq 0, s=0,\\ \R \text{ otherwise.} \end{cases}</math> <br />

| mean = <math>\sum_j \lambda_j (m_j+\delta_j^2)</math>

| mean = <math>\sum_j w_j (k_j+\lambda_j)+m</math>

| variance = <math>2 \sum_j \lambda_j^2 (m_j+2 \delta_j^2) + \sigma^2</math>

| variance = <math>2 \sum_j w_j^2 (k_j+2 \lambda_j) + s^2</math>

| char = <math>\frac{\exp\left(it \sum_j \frac{\lambda_j \delta_j^2}{1-2i \lambda_j t}-\frac{\sigma^2 t^2}{2}\right)}{\prod_j \left(1-2i \lambda_j t \right)^{m_j/2}}</math>

| char = <math>\frac{\exp\left[it \left( m + \sum_j \frac{w_j \lambda_j}{1-2i w_j t} \right)-\frac{s^2 t^2}{2}\right]}{\prod_j \left(1-2i w_j t \right)^{k_j/2}}</math>

}}

}}



In [[probability theory]] and [[statistics]], the '''generalized chi-squared distribution''' (or '''generalized chi-square distribution''') is the distribution of a linear sum of independent [[noncentral chi-squared distribution|noncentral chi-square variables]] and a [[normal distribution|normal variable]], or equivalently, the distribution of a [[quadratic form (statistics)|quadratic form]] of a [[multivariate normal distribution]]. There are several other such generalizations for which the same term is sometimes used; some of them are special cases of the family discussed here, for example the [[gamma distribution]].

In [[probability theory]] and [[statistics]], the '''generalized chi-squared distribution''' (or '''generalized chi-square distribution''') is the distribution of a [[quadratic form (statistics)|quadratic form]] of a [[multivariate normal distribution|multinormal variable (normal vector)]], or a linear combinationof different normal variables and squares of normal variables. Equivalently, it is alsoa linear sum of independent [[noncentral chi-squared distribution|noncentral chi-square variables]] and a [[normal distribution|normal variable]]. There are several other such generalizations for which the same term is sometimes used; some of them are special cases of the family discussed here, for example the [[gamma distribution]].



==Definition==

==Definition==

The generalized chi-squared variable may be described in multiple ways. One is to write it as a linear sum of independent noncentral chi-square variables and a normal variable <ref name=Davies1>Davies, R.B. (1973) Numerical inversion of a characteristic function. [[Biometrika]], 60 (2), 415&ndash;417</ref><ref name=Davies2>Davies, R,B. (1980) "Algorithm AS155: The distribution of a linear combination of ''&chi;''<sup>2</sup> random variables", ''Applied Statistics'', 29, 323&ndash;333</ref>:

The generalized chi-squared variable may be described in multiple ways. One is to write it as a weighted sum of independent [[Noncentral chi-squared distribution|noncentral chi-square]] variables <math>{{\chi}'}^2</math> and a standard normal variable <math>z</math>:<ref name=Davies1>Davies, R.B. (1973) Numerical inversion of a characteristic function. [[Biometrika]], 60 (2), 415&ndash;417</ref><ref name=Davies2>Davies, R.B. (1980) "Algorithm AS155: The distribution of a linear combination of ''&chi;''<sup>2</sup> random variables", ''Applied Statistics'', 29, 323&ndash;333</ref>



:<math>\xi=\sum_i \lambda_i y_i + \sigma z, \quad y_i \sim \chi'^2(m_i,\delta_i^2), \quad z \sim N(0,1).</math>

:<math>\tilde{\chi}(\boldsymbol{w}, \boldsymbol{k}, \boldsymbol{\lambda},s,m)=\sum_i w_i {{\chi}'}^2 (k_i,\lambda_i) + sz+m.</math>



Here the parameters are the weights <math>\lambda_i</math> and <math>\sigma</math>, and the degrees of freedom <math>m_i</math> and non-centralities <math>\delta_i^2</math> of the constituent chi-squares. Some important special cases of this have all coefficients the same sign, omit the normal term or have central chi-squared components.

Here the parameters are the weights <math>w_i</math>, the degrees of freedom <math>k_i</math> and non-centralities <math>\lambda_i</math> of the constituent non-central chi-squares, and the coefficients <math>s</math> and <math>m</math> of the normal. Some important special cases of this have all weights <math>w_i</math>of the same sign, or have central chi-squared components, or omit the normal term.



Since a non-central chi-squared variable is a sum of squares of normal variables with different means, the generalized chi-square variable is also defined as a sum of squares of independent normal variables, plus an independent normal variable: that is, a quadratic in normal variables.

Another is to formulate it as a quadratic form of a normal vector <math>\boldsymbol{x}</math> <ref name=Jones1 />:



Another equivalent way is to formulate it as a quadratic form of a normal vector <math>\boldsymbol{x}</math>:<ref name=Jones1 /><ref name="Das" />

:<math>\xi=q(\boldsymbol{x}) = \boldsymbol{x}' \mathbf{Q_2} \boldsymbol{x} + \boldsymbol{q_1}' \boldsymbol{x} + q_0</math>.



:<math>\tilde{\chi}=q(\boldsymbol{x}) = \boldsymbol{x}' \mathbf{Q_2} \boldsymbol{x} + \boldsymbol{q_1}' \boldsymbol{x} + q_0</math>.

Here <math>\mathbf{Q_2}</math> is a matrix, <math>\boldsymbol{q_1}</math> is a vector, and <math>q_0</math> is a scalar. These, together with the mean <math>\boldsymbol{\mu}</math> and covariance matrix <math>\mathbf{\Sigma}</math> of the normal vector <math>\boldsymbol{x}</math>, parameterize the distribution. If (and only if) <math>\mathbf{Q_2}</math> in this formulation is positive-definite, all the <math>\lambda_i</math> in the other formulation will have the same sign.


Here <math>\mathbf{Q_2}</math> is a matrix, <math>\boldsymbol{q_1}</math> is a vector, and <math>q_0</math> is a scalar. These, together with the mean <math>\boldsymbol{\mu}</math> and covariance matrix <math>\mathbf{\Sigma}</math> of the normal vector <math>\boldsymbol{x}</math>, parameterize the distribution. The parameters of the former expression (in terms of non-central chi-squares, a normal and a constant) can be calculated in terms of the parameters of the latter expression (quadratic form of a normal vector).<ref name="Das">{{cite arXiv |eprint=2012.14331 |last1=Das |first1=Abhranil |author2=Wilson S Geisler |title=Methods to integrate multinormals and compute classification measures |date=2020 |class=stat.ML }}</ref> If (and only if) <math>\mathbf{Q_2}</math> in this formulation is [[positive-definite matrix|positive-definite]], then all the <math>w_i</math> in the first formulation will have the same sign.



For the most general case, a reduction towards a common standard form can be made by using a representation of the following form:<ref name=Sheil>Sheil, J., O'Muircheartaigh, I. (1977) "Algorithm AS106: The distribution of non-negative quadratic forms in normal variables",''Applied Statistics'', 26, 92&ndash;98</ref>

For the most general case, a reduction towards a common standard form can be made by using a representation of the following form:<ref name=Sheil>Sheil, J., O'Muircheartaigh, I. (1977) "Algorithm AS106: The distribution of non-negative quadratic forms in normal variables",''Applied Statistics'', 26, 92&ndash;98</ref>

Line 33: Line 36:

where ''D'' is a diagonal matrix and where ''x'' represents a vector of uncorrelated [[standard normal]] random variables.

where ''D'' is a diagonal matrix and where ''x'' represents a vector of uncorrelated [[standard normal]] random variables.



==Computing the pdf/cdf/inverse cdf/random numbers==

==Probability density and cumulative distribution functions==



The probability density and cumulative distribution functions of a generalized chi-squared variable do not have simple closed-form expressions. However, numerical algorithms <ref name=Sheil /><ref name=Davies2 /><ref name="Imhof">{{cite journal |last1=Imhof |first1=J. P. |title=Computing the Distribution of Quadratic Forms in Normal Variables |journal=Biometrika |pages=419–426 |doi=10.2307/2332763 |date=1961|volume=48 |issue=3/4 |jstor=2332763 |url=http://doc.rero.ch/record/294852/files/48-3-4-419.pdf }}</ref> and computer code ([http://www.statsresearch.co.nz/robert/QF.htm Fortran and C], [https://www.mathworks.com/matlabcentral/fileexchange/74663-generalized-chi-squared-pdf-and-cdf Matlab], [https://cran.r-project.org/web/packages/CompQuadForm/index.html R]) for evaluating them have been published.

The probability density, cumulative distribution, and inverse cumulative distribution functions of a generalized chi-squared variable do not have simple closed-form expressions. However, numerical algorithms <ref name=Sheil /><ref name=Davies2 /><ref name="Imhof">{{cite journal |last1=Imhof |first1=J. P. |title=Computing the Distribution of Quadratic Forms in Normal Variables |journal=Biometrika |pages=419–426 |doi=10.2307/2332763 |date=1961|volume=48 |issue=3/4 |jstor=2332763 |url=http://doc.rero.ch/record/294852/files/48-3-4-419.pdf }}</ref><ref name=Das /> and computer code ([http://www.statsresearch.co.nz/robert/QF.htm Fortran and C], [https://www.mathworks.com/matlabcentral/fileexchange/85028-generalized-chi-square-distribution Matlab], [https://cran.r-project.org/web/packages/CompQuadForm/index.html R], [https://pypi.org/project/chi2comb/ Python], [https://github.com/heliosdrm/Distributions.jl/tree/GChisq/ Julia]) have been published to evaluate some of these, and to generate random samples.



==Applications==

==Applications==



The generalized chi-squared is the distribution of [[estimation|statistical estimates]] in cases where the usual [[statistical theory]] does not hold. For example, if a [[predictive modelling|predictive model]] is fitted by [[least squares]], but the model errors have either [[autocorrelation]] or [[heteroscedasticity]], then alternative models can be compared by relating changes in the [[Explained sum of squares|sum of squares]] to an [[asymptotic distribution|asymptotically valid]] generalized chi-squared distribution.<ref name=Jones1>Jones, D.A. (1983) "Statistical analysis of empirical models fitted by optimisation", [[Biometrika]], 70 (1), 67&ndash;88</ref>

The generalized chi-squared is the distribution of [[estimation|statistical estimates]] in cases where the usual [[statistical theory]] does not hold, as in the examples below.

===In model fitting and selection===

If a [[predictive modelling|predictive model]] is fitted by [[least squares]], but the [[residuals (statistics)|residuals]] have either [[autocorrelation]] or [[heteroscedasticity]], then alternative models can be compared (in [[model selection]]) by relating changes in the [[Explained sum of squares|sum of squares]] to an [[asymptotic distribution|asymptotically valid]] generalized chi-squared distribution.<ref name=Jones1>Jones, D.A. (1983) "Statistical analysis of empirical models fitted by optimisation", [[Biometrika]], 70 (1), 67&ndash;88</ref>



===Classifying normal samples using Gaussian discriminant analysis===

===Classifying normal vectors using Gaussian discriminant analysis===

If <math>\boldsymbol{x}</math> is a normal variable, its log likelihood is a [[quadratic form]] of <math>\boldsymbol{x}</math>, and is hence distributed as a generalized chi-squared. The log likelihood ratio that <math>\boldsymbol{x}</math> arises from one normal distribution versus another is also a [[quadratic form]], so distributed as a generalized chi-squared.

If <math>\boldsymbol{x}</math> is a normal vector, its log likelihood is a [[quadratic form]] of <math>\boldsymbol{x}</math>, and is hence distributed as a generalized chi-squared. The log likelihood ratio that <math>\boldsymbol{x}</math> arises from one normal distribution versus another is also a [[quadratic form]], so distributed as a generalized chi-squared.<ref name="Das" />



In Gaussian discriminant analysis, samples from normal distributions are optimally separated by using a [[quadratic classifier]], a boundary that is a quadratic function (e.g. the curve defined by setting the likelihood ratio between two Gaussians to 1). The classification error rates of different types (false positives and false negatives) are integrals of the normal distributions within the quadratic regions defined by this classifier. Since this is mathematically equivalent to integrating a quadratic form of a normal variable, the result is an integral of a generalized-chi-squared variable.

In Gaussian discriminant analysis, samples from multinormal distributions are optimally separated by using a [[quadratic classifier]], a boundary that is a quadratic function (e.g. the curve defined by setting the likelihood ratio between two Gaussians to 1). The classification error rates of different types (false positives and false negatives) are integrals of the normal distributions within the quadratic regions defined by this classifier. Since this is mathematically equivalent to integrating a quadratic form of a normal vector, the result is an integral of a generalized-chi-squared variable.<ref name="Das" />



===In signal processing===

===In signal processing===

The following application arises in the context of [[Fourier analysis]] in [[signal processing]], [[renewal theory]] in [[probability theory]], and [[MIMO|multi-antenna systems]] in [[Wireless|wireless communication]]. The common factor of these areas is that the sum of exponentially distributed variables is of importance (or identically, the sum of squared magnitudes [[complex normal distribution|circular symmetric complex Gaussian]] variables).

The following application arises in the context of [[Fourier analysis]] in [[signal processing]], [[renewal theory]] in [[probability theory]], and [[MIMO|multi-antenna systems]] in [[Wireless|wireless communication]]. The common factor of these areas is that the sum of exponentially distributed variables is of importance (or identically, the sum of squared magnitudesof [[complex normal distribution|circularly-symmetric centered complex Gaussian]] variables).



If <math>Z_i</math> are ''k'' [[statistical independence|independent]], [[complex normal distribution|circular symmetric complex Gaussian]] random variables with [[mean]] 0 and [[variance]] <math>\sigma_i^2</math>, then the random variable

If <math>Z_i</math> are ''k'' [[statistical independence|independent]], [[complex normal distribution|circularly-symmetric centered complex Gaussian]] random variables with [[mean]] 0 and [[variance]] <math>\sigma_i^2</math>, then the random variable



:<math>\tilde{Q} = \sum_{i=1}^k |Z_i|^2</math>

:<math>\tilde{Q} = \sum_{i=1}^k |Z_i|^2</math>

Line 82: Line 88:


==See also==

==See also==

* [[Degrees of freedom (statistics)#Alternative]]

* [[Noncentral chi-squared distribution]]

* [[Noncentral chi-squared distribution]]

* [[Chi-squared distribution]]

* [[Chi-squared distribution]]

Line 91: Line 96:

==External links==

==External links==

* [http://www.statsresearch.co.nz/robert/QF.htm Davies, R.B.: Fortran and C source code for "Linear combination of chi-squared random variables"]

* [http://www.statsresearch.co.nz/robert/QF.htm Davies, R.B.: Fortran and C source code for "Linear combination of chi-squared random variables"]

* [https://www.mathworks.com/matlabcentral/fileexchange/74663-generalized-chi-squared-pdf-and-cdf Das, A: MATLAB code to compute PDF and CDFofa generalized chi-squared distribution.]

* [https://www.mathworks.com/matlabcentral/fileexchange/85028-generalized-chi-square-distribution Das, A: MATLAB code to compute the statistics, pdf, cdf, inverse cdf and random numbersofthe generalized chi-square distribution.]



{{ProbDistributions}}

{{ProbDistributions}}


Latest revision as of 23:16, 22 June 2024

Generalized chi-squared distribution
Probability density function
Generalized chi-square probability density function
Cumulative distribution function
Generalized chi-square cumulative distribution function
Notation
Parameters , vector of weights of noncentral chi-square components
, vector of degrees of freedom of noncentral chi-square components
, vector of non-centrality parameters of chi-square components
, scale of normal term
, offset
Support
Mean
Variance
CF

Inprobability theory and statistics, the generalized chi-squared distribution (orgeneralized chi-square distribution) is the distribution of a quadratic form of a multinormal variable (normal vector), or a linear combination of different normal variables and squares of normal variables. Equivalently, it is also a linear sum of independent noncentral chi-square variables and a normal variable. There are several other such generalizations for which the same term is sometimes used; some of them are special cases of the family discussed here, for example the gamma distribution.

Definition[edit]

The generalized chi-squared variable may be described in multiple ways. One is to write it as a weighted sum of independent noncentral chi-square variables and a standard normal variable :[1][2]

Here the parameters are the weights , the degrees of freedom and non-centralities of the constituent non-central chi-squares, and the coefficients and of the normal. Some important special cases of this have all weights of the same sign, or have central chi-squared components, or omit the normal term.

Since a non-central chi-squared variable is a sum of squares of normal variables with different means, the generalized chi-square variable is also defined as a sum of squares of independent normal variables, plus an independent normal variable: that is, a quadratic in normal variables.

Another equivalent way is to formulate it as a quadratic form of a normal vector :[3][4]

.

Here is a matrix, is a vector, and is a scalar. These, together with the mean and covariance matrix of the normal vector , parameterize the distribution. The parameters of the former expression (in terms of non-central chi-squares, a normal and a constant) can be calculated in terms of the parameters of the latter expression (quadratic form of a normal vector).[4] If (and only if) in this formulation is positive-definite, then all the in the first formulation will have the same sign.

For the most general case, a reduction towards a common standard form can be made by using a representation of the following form:[5]

where D is a diagonal matrix and where x represents a vector of uncorrelated standard normal random variables.

Computing the pdf/cdf/inverse cdf/random numbers[edit]

The probability density, cumulative distribution, and inverse cumulative distribution functions of a generalized chi-squared variable do not have simple closed-form expressions. However, numerical algorithms [5][2][6][4] and computer code (Fortran and C, Matlab, R, Python, Julia) have been published to evaluate some of these, and to generate random samples.

Applications[edit]

The generalized chi-squared is the distribution of statistical estimates in cases where the usual statistical theory does not hold, as in the examples below.

In model fitting and selection[edit]

If a predictive model is fitted by least squares, but the residuals have either autocorrelationorheteroscedasticity, then alternative models can be compared (inmodel selection) by relating changes in the sum of squares to an asymptotically valid generalized chi-squared distribution.[3]

Classifying normal vectors using Gaussian discriminant analysis[edit]

If is a normal vector, its log likelihood is a quadratic formof, and is hence distributed as a generalized chi-squared. The log likelihood ratio that arises from one normal distribution versus another is also a quadratic form, so distributed as a generalized chi-squared.[4]

In Gaussian discriminant analysis, samples from multinormal distributions are optimally separated by using a quadratic classifier, a boundary that is a quadratic function (e.g. the curve defined by setting the likelihood ratio between two Gaussians to 1). The classification error rates of different types (false positives and false negatives) are integrals of the normal distributions within the quadratic regions defined by this classifier. Since this is mathematically equivalent to integrating a quadratic form of a normal vector, the result is an integral of a generalized-chi-squared variable.[4]

In signal processing[edit]

The following application arises in the context of Fourier analysisinsignal processing, renewal theoryinprobability theory, and multi-antenna systemsinwireless communication. The common factor of these areas is that the sum of exponentially distributed variables is of importance (or identically, the sum of squared magnitudes of circularly-symmetric centered complex Gaussian variables).

If are k independent, circularly-symmetric centered complex Gaussian random variables with mean 0 and variance , then the random variable

has a generalized chi-squared distribution of a particular form. The difference from the standard chi-squared distribution is that are complex and can have different variances, and the difference from the more general generalized chi-squared distribution is that the relevant scaling matrix A is diagonal. If for all i, then , scaled down by (i.e. multiplied by ), has a chi-squared distribution, , also known as an Erlang distribution. If have distinct values for all i, then has the pdf[7]

If there are sets of repeated variances among , assume that they are divided into M sets, each representing a certain variance value. Denote to be the number of repetitions in each group. That is, the mth set contains variables that have variance It represents an arbitrary linear combination of independent -distributed random variables with different degrees of freedom:

The pdf of is[8]

where

with from the set of all partitions of (with ) defined as

See also[edit]

References[edit]

  1. ^ Davies, R.B. (1973) Numerical inversion of a characteristic function. Biometrika, 60 (2), 415–417
  • ^ a b Davies, R.B. (1980) "Algorithm AS155: The distribution of a linear combination of χ2 random variables", Applied Statistics, 29, 323–333
  • ^ a b Jones, D.A. (1983) "Statistical analysis of empirical models fitted by optimisation", Biometrika, 70 (1), 67–88
  • ^ a b c d e Das, Abhranil; Wilson S Geisler (2020). "Methods to integrate multinormals and compute classification measures". arXiv:2012.14331 [stat.ML].
  • ^ a b Sheil, J., O'Muircheartaigh, I. (1977) "Algorithm AS106: The distribution of non-negative quadratic forms in normal variables",Applied Statistics, 26, 92–98
  • ^ Imhof, J. P. (1961). "Computing the Distribution of Quadratic Forms in Normal Variables" (PDF). Biometrika. 48 (3/4): 419–426. doi:10.2307/2332763. JSTOR 2332763.
  • ^ D. Hammarwall, M. Bengtsson, B. Ottersten (2008) "Acquiring Partial CSI for Spatially Selective Transmission by Instantaneous Channel Norm Feedback", IEEE Transactions on Signal Processing, 56, 1188–1204
  • ^ E. Björnson, D. Hammarwall, B. Ottersten (2009) "Exploiting Quantized Channel Norm Feedback through Conditional Statistics in Arbitrarily Correlated MIMO Systems", IEEE Transactions on Signal Processing, 57, 4027–4041
  • External links[edit]


    Retrieved from "https://en.wikipedia.org/w/index.php?title=Generalized_chi-squared_distribution&oldid=1230476684"

    Category: 
    Continuous distributions
     



    This page was last edited on 22 June 2024, at 23:16 (UTC).

    Text is available under the Creative Commons Attribution-ShareAlike License 4.0; additional terms may apply. By using this site, you agree to the Terms of Use and Privacy Policy. Wikipedia® is a registered trademark of the Wikimedia Foundation, Inc., a non-profit organization.



    Privacy policy

    About Wikipedia

    Disclaimers

    Contact Wikipedia

    Code of Conduct

    Developers

    Statistics

    Cookie statement

    Mobile view



    Wikimedia Foundation
    Powered by MediaWiki