Some Expectations of a Non-Central Chi-Square Distribution With an Even Number of Degrees of Freedom Stefan M. Moser April 7, 007 Abstract The non-central chi-square distribution plays an important role in communications, for example in the analysis of mobile and wireless communication systems. It not only includes the important cases of a squared Rayleigh distribution and a squared Rice distribution, but also the generalizations to a sum of independent squared Gaussian random variables of identical variance with or without mean, i.e., a squared MIMO Rayleigh and squared MIMO Rice distribution. In this paper closed-form expressions are derived for the expectation of the logarithm and for the expectation of the n-th power of the reciprocal value of a non-central chi-square random variable. It is shown that these expectations can be expressed by a family of continuous functions g m ( ) and that these families have nice properties (monotonicity, convexity, etc.). Moreover, some tight upper and lower bounds are derived that are helpful in situations where the closed-form expression of g m ( ) is too complex for further analysis. Index Terms: Non-central chi-square distribution, Rayleigh, Rice, expected logarithm, expected reciprocal value. Introduction It is well known that adding several independent squared Gaussian random variables of identical variance yields a random variable that is non-central chi-square distributed. This distribution often shows up in information theory and communications. As an example we mention the situation of a non-coherent multiple-input multiple-output (MIMO) fading channel Y = Hx + Z () with an additive white Gaussian noise vector Z and with a fading matrix H that consists of independent and unit-variance Gaussian distributed components with or without mean. Here conditional on the input x the squared magnitude of the output Y is non-central chi-square distributed. The author is with the Department of Communication Engineering, National Chiao Tung University (NCTU), Hsinchu, Taiwan. This work was supported by the Industrial Technology Research Institute (ITRI), Zhudong, Taiwan, under JRC NCTU-ITRI and by the National Science Council under NSC 95--E-009-046. Part of this work has been submitted to TENCON 007 in Taipei, Taiwan. Stefan M. Moser, April 7, 007, submitted
While the special case of a squared Rayleigh distribution is well understood in the sense that there exist closed-form expressions for more or less all interesting expected values, the more general situation of a non-central chi-square distribution is far more complex. Here, standard integration tools (e.g., Maple) or integration lookup tables (e.g., []) will very quickly cease to provide closed-form expressions. In this paper we will state closed-form expressions for some of these situations: we will give closed-form solutions to E[lnV ] and E [ ] V for a non-central chi-square n random variable V with an even number of degrees of freedom. Note that in practice we often have an even number of degrees of freedom because we usually consider complex Gaussian random variables consisting of two real Gaussian components. We will see that these expectations are all related to a family of functions g m ( ) that is defined in Definition in the following section. There we will also state the main results. In Section 3 we will then derive some properties of the functions g m ( ) and in Section 4 we will state tight upper and lower bounds. In Section 5 another property of g m ( ) is derived as an example of how the bounds from Section 4 could be applied. We conclude in Section 6. Definitions and Main Results A non-negative real random variable is said to have a non-central chi-square distribution with n degrees of freedom and non-centrality parameter s if it is distributed like n (X j + µ j ), () j= where {X j } n j= are IID N R (0, ) and the real constants {µ j } n j= satisfy s = n µ j. (3) j= (The distribution of () depends on the constants {µ j } only via the sum of their squares.) The probability density function of such a distribution is given by [3, Chapter 9] ( x ) n 4 s e s +x ( ) I n/ s x, x 0. (4) Here I ν ( ) denotes the modified Bessel function of the first kind of order ν R, i.e., I ν (x) ( x ) ν+k, x 0 (5) k! Γ(ν + k + ) (see [, Equation 8.445]). If the number of degrees of freedom n is even, i.e., if n = m for some positive integer m, then the non-central chi-square distribution can also be expressed as a sum of the squared norms of complex Gaussian random variables: Definition. Let the random variable V have a non-central chi-square distribution with an even number m of degrees of freedom, i.e., m V U j + µ j (6) A squared Rayleigh random variable is actually exponentially distributed. j= Stefan M. Moser, April 7, 007, submitted
where {U j } m j= are IID N C(0, ), and {µ j } m j= are complex constants. Let further the non-centrality parameter s be defined as s m µ j. (7) j= Next we define the following continuous functions: Definition. The functions g m ( ) are defined as follows: m ln() Ei( ) + ( ) [e j (j )! j= ] ( ) g m () (m )! j, > 0 j(m j)! ψ(m), = 0 (8) for m N, where Ei( ) denotes the exponential integral function defined as and ψ( ) is Euler s psi function given by e t Ei( x) dt, x > 0 (9) x t m ψ(m) γ + j j= (0) with γ 0.577 denoting Euler s constant. lim 0 Note that g m () is continuous for all 0, i.e., in particular { m [ ln() Ei( ) + ( ) j e (m )! (j )! j(m j)! j= ]( ) } j = ψ(m) () for all m N. Therefore its first derivative is defined for all 0 and can be evaluated to g m() g m() = ( )m m Γ(m) e ( ) j m j () j! (see [4, Eq. (47)], [5, Eq. (A.39)]). Note that g m( ) is also continuous, i.e., in particular ( ) m m Γ(m) lim e ( ) j 0 m j j! = m = g m(0). (3) j=0 Now we give a closed-form expression for a first expectation of a non-central chi-square random variable: j=0 Stefan M. Moser, April 7, 007, submitted 3
Theorem 3. The expected value of the logarithm of a non-central chi-square random variable with an even number m of degrees of freedom is given as E[lnV ] = g m (s ), (4) where V and s are defined in (6) and (7). Hence, we have the solution to the following integral: for any m N and s 0. 0 ( v ) m lnv s e v s I m (s v)dv = g m (s ) (5) Proof. A proof can be found in [4, Lem. 0.], [5, Lem. A.6] Next we derive the following expectations: Theorem 4. Let n N with n < m. The expected value of the n-th power reciprocal value of a non-central chi-square random variable with an even number m of degrees of freedom is given as where [ ] E V n = ( )n (n!) g(n) m n (s ), m > n (6) g (l) m () = l g m () l (7) denotes the l-th derivative of g m ( ) and where V and s are defined in (6) and (7). In particular, for m > [ ] E = g V m (s ). (8) Hence, we have the solution to the following integral: 0 ( v ) m v n s e v s I m (s v)dv = ( )n (n!) g(n) m n (s ) (9) for any m, n N, m > n, and any real s 0. Note that in the cases where m n, the expectation is unbounded. Proof. A proof can be found in Appendix A. 3 Properties of g m ( ) and g m( ) In this section we will show that the family of functions g m ( ) and g m( ) are wellbehaved. Corollary 5. The functions g m ( ) are monotonically strictly increasing and strictly concave in the interval [0, ) for all m N. Stefan M. Moser, April 7, 007, submitted 4
Proof. From [4, Eqs. (45), (46)], [5, Eqs. (A.37), (A.38)] we know that g m() = g m() g m() = g m () = e k! = e k + m k, (0) k! (k + m)(k + m + ) k, () i.e., the first derivative of g m ( ) is positive and the second derivative is negative. Corollary 6. The function g m () is monotonically strictly increasing in m for all 0. Proof. Fix two arbitrary natural numbers m, m N such that m < m. Choose µ = s, µ =... = µ m = 0, with an arbitrary s 0. Let {U j } m j= be IID N C(0, ). Then m g m (s ) = E ln U j + µ j () j= m = E ln U j + µ j + j= m j=m + U j + µ j (3) m > E ln U j + µ j (4) j= = g m (s ), (5) where the first equality follows from (4); the subsequent equality from splitting the sum into two parts; the subsequent inequality from dropping some positive terms; and the final equality again from (4). Corollary 7. The functions g m( ) are positive, monotonically strictly decreasing, and strictly convex functions for all m N. Proof. The positivity and the monotonicity follow from (0) and (). To see the convexity, compute from () g m() = 3 g m () 3 = e which is positive. k! (k + m)(k + m + )(k + m + ) k (6) Corollary 8. The function g m() is monotonically strictly decreasing in m for all 0. Proof. Fix two arbitrary natural numbers m, m N such that m > m >. Choose µ = s, µ =... = µ m = 0, with an arbitrary s 0. Let {U j } m j= be IID Stefan M. Moser, April 7, 007, submitted 5
N C (0, ). Then [ ] g m (s ) = E m j= U j + µ j [ ] = E m j= U j + µ j + m j=m + U j + µ j [ ] < E m j= U j + µ j (7) (8) (9) = g m (s ), (30) where the first equality follows from (8); the subsequent equality from splitting the sum into two parts; the subsequent inequality from dropping some positive terms in the denominator; and the final equality again from (8). Theorem 9. We have the following relation: for all m N and all 0. Proof. A proof is given in Appendix B. Theorem 0. We have the following relation: for all m N and all 0. Proof. A proof is given in Appendix C. 4 Bounds on g m ( ) and g m( ) g m+ () = g m () + g m() (3) g m+() = m g m() (3) In this section we derive some tight bounds on the functions g m ( ) and g m( ). Theorem. The function g m( ) can be bounded as follows: { + m g m() min m + m( + m + ), + m }. (33) Note that for < m+ the first of the two upper bounds is tighter than second, while for > m + the second is tighter. Moreover, the first upper bound coincides with g m () for = 0, and the second upper bound is asymptotically tight when tends to infinity. Proof. A proof is given in Appendix D. The bounds (33) are depicted in Figure for the cases of m =, m = 3, and m = 8. Stefan M. Moser, April 7, 007, submitted 6
second upper bound (for > m + ) first upper bound (for < m + ) g m ( ) lower bound 0.8 g m() 0.6 0.4 0. 0 0 3 4 5 6 7 8 9 0 Figure : Upper and lower bounds on g m( ) according to Theorem. The top four curves correspond to m =, the middle four to m = 3, and the lowest group of four curves to m = 8. 3.5 3.5 gm().5 0.5 0 second upper bound first upper bound 0.5 upper bound (35) g m( ) lower bound (34) 0 3 4 5 6 7 8 9 0 Figure : Upper and lower bounds on g m ( ) according to (34) and (35) in Theorem. The lowest curve corresponds to m = (in this case all bounds coincides with g ( )), the next five curves correspond to m = 3, and the top five to m = 8. Stefan M. Moser, April 7, 007, submitted 7
3.5 3.5 gm().5 0.5 0 0.5 upper bound g m( ) lower bound 0 3 4 5 6 7 8 9 0 Figure 3: Upper and lower bounds on g m ( ) according to (36) in Theorem. The lowest three curves correspond to m =, the next three to m = 3, and the top three to m = 8. Theorem. For the functions g m ( ) we state two sets of bounds. The first set is tighter for smaller values of m: m g m () ln Ei( ) + + j, (34) j= m g m () ln Ei( ) + j= { min j + j( + j + ), + j Secondly, we give a set of bounds that is tight for large values of m: }. (35) ln( + m ) g m () ln( + m). (36) Note that this second set of bounds is very simple, i.e., it is particularly interesting for further analysis. Proof. A proof is given in Appendix E. The bounds (34) and (35) are depicted in Figure and the bounds (36) in Figure 3, both times for the cases of m =, m = 3, and m = 8. 5 Additional Properties As an example of how the properties given in Section 3 and the bounds given in Section 4 can be used to derive further results we state the following corollary: Stefan M. Moser, April 7, 007, submitted 8
( ) Corollary 3. The functions g m are monotonically strictly decreasing and convex in for all m N. Proof. We start as follows: g m ( ) = g m ( ) = ( g m+ ( (37) ) ( )) g m, (38) where the last equality follows from Theorem 9. From ( ) Corollary 7 we know that g m( ) > 0, so that we can conclude from (37) that g m is monotonically strictly decreasing. To check convexity, we continue as follows: ( ) g m = ( ) 3g m+ + ( ) 4g m+ ( ) 3g m ( ) 4g m (39) ( ) ( ) = ( ) mg m g m 3g m + 4 (40) ( ) = g m m 4 + 3 (4) m + m 4 + 3 (4) = (m + ). (43) Here, in the second equality we use Theorems 9 and 0; and the inequality follows from the lower bound of Theorem. Hence, the second derivative is positive and the statement proved. 6 Conclusions We have derived closed-form expressions for some important expectations in the field of information theory and communications. We have shown that the resulting functions behave nicely and we have given tight upper and lower bounds to them. For brevity we will not include a detailed example of how these results can be used, but only mention that the derivation of the fading number of a non-coherent MIMO Gaussian fading channel [6] strongly depends on them. There are many more examples. A Proof of Theorem 4 Let n N be arbitrary and assume that m > n. Then the required expectation can be written as [ ] ( v ) m E V n = 0 v n s e v s I m (s v)dv. (44) Expressing I m ( ) as a power series (5) ( z ) m +k I m (z) = (45) k! Γ(m + k) Stefan M. Moser, April 7, 007, submitted 9
we obtain from [, Eq. 3.35-3] (using that m > n) [ ] E V n = s m e s = e s = e s k! Γ(m + k) sk+m 0 v k+m n e v dv (46) k! (m + k )! sk (k + m n)! (47) k! (k + m ) (k + m n) sk. (48) Generalizing (0), (), and (6) to the l-th derivative, we have g m (l) () = l g m () l = ( ) l e k! (l )! (k + m) (k + m + l ) k. (49) Comparing this with (48) we see that [ ] E V n = ( )n (n!) g(n) m n (s ). (50) B Proof of Theorem 9 Using (8) and () we get g m () + g m() m ( ) j = ln() Ei( ) + ( ) j e (j )! m j= j= ( ) j (m )! j(m j)! ( ) j + ( )m (m )! m e m ( ) i+m(m )! i m (5) i! i=0 m ( ) j = ln() Ei( ) + ( ) j e (j )! m j= j= ( ) j (m )! j(m j)! = ln() Ei( ) + m ( ) j m j= m ( ) j e (j )! j= ( ) j+m }{{} ( =( ) j ) j ( ) ( ) j (m )! j(m j)! + (m j)! j= }{{} j= = m j(m j)! (m )! (m j)! j (5) ( ) j ( ) ( ) m(m )! m (53) 0! m ( ) j = ln() Ei( ) + ( ) j e (j )! Stefan M. Moser, April 7, 007, submitted 0
m ( ) j m (m )! j(m j)! j= m = ln() Ei( ) + ( ) j e (j )! m j= j= ( ) j m! j(m j)! ( ) j ( ) m m! m 0! ( ) j ( ) m (54) ( ) j (55) = g m+ (). (56) Here, the first equality follows from the definitions given in (8) and (); in the subsequent equality we combine the second last term with the first sum and reorder the last summation by introducing a new counter-variable j = m i; the subsequent three equalities follows from arithmetic rearrangements; and the final equality follows again from definition (8). C Proof of Theorem 0 Using (0) we get m g m() = e e m e = e k! k e = e = e = e k= = e k! ( m k + m k! k + m k (57) k! m k + m k (58) ) k (59) k! k k + m k (60) (k )! k + m k (6) k! k + m + k (6) = g m+(). (63) Here, the first equality follows from (0); in the subsequent equality we use the series expansion of e which is valid for all 0; the subsequent two equalities follow from algebraic rearrangements; in the next equality we note that for k = 0 the terms in the sum are equal to zero; the second last equality then follows from renumbering the terms; and the last equality follows again from (0). D Proof of Theorem We start with the lower bound and note that for = 0 the bound is tight: g m(0) = g m+ (0) g m (0) = ψ(m + ) ψ(m) = m (64) Stefan M. Moser, April 7, 007, submitted
where the first equality follows from Theorem 9, the second from (8) and the final from (0); and + m = =0 m. (65) Moreover we notice that the bound is asymptotically tight, too: lim g m() = 0; (66) = 0. + m (67) lim Hence, since additionally both functions g m( ) and +m are monotonically strictly decreasing and strictly convex, they cannot cross. Moreover, it is shown in Theorem that g m () ln( + m), so that we must have g m() ln( + m) = + m. (68) Next we turn to the first upper bound which will follow from the lower bound (33) derived above together with Theorem 0: g m () = g m() m +m m = m (m )( + m). (69) To derive the second upper bound we use once again Theorem 0 and the lower bound (33) derived above: + m g m() = + m + m g m () (70) + m + m + m (7) = 0, (7) where the first equality follows from Theorem 0 and the inequality from the lower bound in (33). E Proof of Theorem The first set of bounds (34) and (35) follow directly from Theorems 9 and and from the fact that g () = ln Ei( ). (73) The upper bound in the second set of bounds (36) has been proven before in [7, App. B]. The proof is based on Jensen s inequality. The lower bound in (36) follows from a slightly more complicated argument: Note that both g m ( ) and ln( + m ) are monotonically strictly increasing and strictly concave (see Corollary 5). Hence, they can cross at most twice. Now asymptotically as the two functions coincide which corresponds to one of these crossings. So they can only cross at most once more for finite. For = 0, we have g m (0) = ψ(m) > ln(m ) (74) Note that this part of the proof of Theorem does not rely in any way on the theorem under consideration. Stefan M. Moser, April 7, 007, submitted
for all m N (where for m = we take ln0 = ). Let s assume for the moment that there is another crossing for a finite value 0. Then, for > 0, ln( + m ) is strictly larger than g m (). However, since asymptotically they will coincide, the slope of ln( + m ) must then be strictly smaller than the slope of g m ( ). But we know from Theorem that ln( + m ) = + m g m(). (75) This is a contradiction which leads to the conclusion that there cannot be another crossing and ln( + m ) must be a strict lower bound to g m ( ). Acknowledgment This work was supported by the Industrial Technology Research Institute (ITRI), Zhudong, Taiwan, under JRC NCTU-ITRI and by the National Science Council under NSC 95--E-009-046. References [] A. Lapidoth and S. M. Moser, The expected logarithm of a non-central chi-square random variable, website. [Online]. Available: http://moser.cm.nctu.edu.tw/explog.html [] I. S. Gradshteyn and I. M. Ryzhik, Table of Integrals, Series, and Products, 5th ed., A. Jeffrey, Ed. Academic Press, San Diego, 996. [3] N. L. Johnson, S. Kotz, and N. Balakrishnan, Continuous Univariate Distributions, nd ed. John Wiley & Sons, 995, vol.. [4] A. Lapidoth and S. M. Moser, Capacity bounds via duality with applications to multiple-antenna systems on flat fading channels, IEEE Transactions on Information Theory, vol. 49, no. 0, pp. 46 467, October 003. [5] S. M. Moser, Duality-based bounds on channel capacity, Ph.D. dissertation, Swiss Federal Institute of Technology, Zurich, October 004, Diss. ETH No. 5769. [Online]. Available: http://moser.cm.nctu.edu.tw/ [6], Fading number and degrees of freedom, March 007, in preparation. [7], The fading number of memoryless multiple-input multipleoutput fading channels, July 007, to appear in IEEE Transactions on Information Theory, vol. 53, no. 7. [Online]. Available: http://moser.cm.nctu.edu.tw/publications.html Stefan M. Moser, April 7, 007, submitted 3