Académique Documents
Professionnel Documents
Culture Documents
j=1
(X
j
+
j
)
2
, (1)
where {X
j
}
r
j=1
are independent and identically dis-
tributed (IID) N
R
(0, 1) and the real constants
{
j
}
r
j=1
satisfy =
r
j=1
2
j
. (The distribution of
(1) depends on the constants {
j
} only via the sum
of their squares.) The probability density function of
such a distribution is given by [1, Chapter 29]
f
W
(r),
(x) =
1
2
_
x
_
r2
4
e
+x
2
I
r/21
_
_
x
_
, x 0.
(2)
Here I
j=1
U
j
+
j
2
(4)
where {U
j
}
m
j=1
are IID N
C
(0, 1), {
j
}
m
j=1
are com-
plex constants, and where the noncentrality parameter
is dened as
m
j=1
|
j
|
2
. (5)
Note the following relation to the other denition
of a noncentral chi-square distribution (1):
V
(m),
=
1
2
W
(2m),2
. (6)
Next we dene the following family of continuous
functions.
Denition 2. The functions g
m
() are dened as fol-
lows:
g
m
()
_
_
ln() Ei () +
m1
j=1
(1)
j
_
e
(j 1)!
(m1)!
j(m1 j)!
_ _
1
_
j
, > 0
(m), = 0
(7)
for m N, where Ei () denotes the exponential integral
function dened as
Ei ()
_
e
t
t
dt, > 0 (8)
and () is Eulers psi function given by
(m) +
m1
j=1
1
j
, m N (9)
with 0.577 denoting Eulers constant.
Note that g
m
() is continuous over [0, ) for all m
N, i.e., in particular
lim
0
_
ln() Ei () +
m1
j=1
(1)
j
_
e
(j 1)!
(m1)!
j(m1 j)!
_ _
1
_
j
_
= (m) (10)
for all m N. Therefore its rst derivative is dened
2
for all 0 and can be evaluated to
g
m
()
g
m
()
=
(1)
m
(m)
m
_
_
e
m1
j=0
(1)
j
j!
j
_
_
(11)
(see [4, Eq. (417)], [8, Eq. (A.39)]). Note that g
m
() is
also continuous, i.e., in particular
lim
0
_
_
_
(1)
m
(m)
m
_
_
e
m1
j=0
(1)
j
j!
j
_
_
_
_
_
=
1
m
= g
m
(0). (12)
Now we will give closed-form expressions for some
expectations of a noncentral chi-square random vari-
able. We start with the logarithm.
Theorem 3. The expected value of the logarithm of
a noncentral chi-square random variable with an even
number 2m of degrees of freedom is given as
E
_
ln V
(m),
= g
m
() (13)
where V
(m),
and are dened in (4) and (5).
Next we look at the inverse moments.
Theorem 4. Let n N with n < m. The expected
value of the n-th power reciprocal value of a noncentral
chi-square random variable with an even number 2m of
degrees of freedom is given as
E
__
1
V
(m),
_
n
_
=
(1)
n1
(n 1)!
g
(n)
mn
(), m > n
(14)
where g
(n)
m
()
n
g
m
()
n
denotes the n-th derivative of
g
m
() and where V
(m),
and are dened in (4) and
(5). In particular, for m > 1
E
_
1
V
(m),
_
= g
m1
(). (15)
Note that in the cases where m n the expectation
is unbounded.
2
As a matter of fact all derivative of g
m
() are dened and
continuous.
3. PROPERTIES AND BOUNDS OF g
m
()
AND g
m
()
Next we will present some properties and bounds of
g
m
() and g
m
().
Theorem 5. The functions g
m
() are monotonically
strictly increasing and strictly concave in the interval
[0, ) for all m N. The functions g
m
() are mono-
tonically strictly increasing in m for all 0.
The functions g
m
_
1
_
are monotonically strictly de-
creasing and convex in for all m N.
The functions g
m
() are positive, monotonically
strictly decreasing, and strictly convex functions for all
m N. The functions g
m
() are monotonically strictly
decreasing in m for all 0.
We have the following relations:
g
m+1
() = g
m
() +g
m
() (16)
for all m N and all 0, and
g
m+1
() =
1
m
(), > 0, (17)
and g
m
() =
1
m
m
g
m+1
(), 0, (18)
for all m N.
Note that relation (16) nicely reects a property
discovered by Cohen [9] [1, Eq. (29.23a)].
Next we state some bounds.
Theorem 6. The function g
m
() can be bounded as
follows:
1
+m
g
m
() min
_
m+ 1
m( +m+ 1)
,
1
+m1
_
.
(19)
Note that for < m+1 the rst of the two upper bounds
is tighter than second, while for > m + 1 the second
is tighter. Moreover, the rst upper bound coincides
with g
m
() for = 0, and the second upper bound is
asymptotically tight when tends to innity.
For the functions g
m
() we state two sets of bounds.
The rst set is tighter:
g
m
() ln Ei () +
m1
j=1
1
+j
, (20)
g
m
() ln Ei ()
+
m1
j=1
min
_
j + 1
j( +j + 1)
,
1
+j 1
_
. (21)
Secondly, we give a set of bounds that is slightly less
tight, but that is very appealing because the expressions
are simple and easy to use for further analysis:
ln( +m1) g
m
() ln( +m). (22)
The bounds (19) are depicted in Figure 1, the
bounds (20) and (21) in Figure 2 and the bounds (22)
are depicted in Figure 3. In all three cases the bounds
are shown for the values m = 1, m = 3, and m = 8.
0 1 2 3 4 5 6 7 8 9 10
0
0.2
0.4
0.6
0.8
1
g
m
(
)
upper bound (for > m + 1)
upper bound (for < m + 1)
g
m
()
lower bound
Figure 1: Upper and lower bounds on g
m
() according
to (19) in Theorem 6. The top four curves correspond
to m = 1, the middle four to m = 3, and the lowest
group of four curves to m = 8.
Note that some of the asymptotic results given in
[3] can be directly derived from (15) and (19).
We also remark that the known expression [2] [1,
Eq. (29.32a)]
E
__
1
V
(m),
_
n
_
=
(1)
n+m
(n 1)!
n1
=0
_
n 1
+1m
(m 1)
_
_
e
m2
j=0
(1)
j
j!
j
_
_
(23)
can be used as an alternative way of nding the n-th
derivative of g
m
():
g
(n)
m
() = (1)
m+n1
n1
=0
_
n 1
(m)(n1)
_
m (n 1)
_
_
_
e
m+n2
j=0
(1)
j
j!
j
_
_
.
(24)
Finally, we also state another expression for the de-
rivatives of g
m
() that uses an innite sum [4] [1,
0 1 2 3 4 5 6 7 8 9 10
1
0.5
0
0.5
1
1.5
2
2.5
3
3.5
g
m
(
)
second loosened bound
first loosened bound
upper bound (21)
g
m
()
lower bound (20)
Figure 2: Upper and lower bounds on g
m
() according
to (20) and (21) in Theorem 6. We also depict the two
additional upper bounds that are gained by loosening
the minimum in (21) to always take on one of the two
values. The lowest curve corresponds to m = 1 (in
this case all bounds coincides with g
1
()), the next ve
curves correspond to m = 3, and the top ve to m = 8.
0 1 2 3 4 5 6 7 8 9 10
1
0.5
0
0.5
1
1.5
2
2.5
3
3.5
g
m
(
)
upper bound
g
m
()
lower bound
Figure 3: Upper and lower bounds on g
m
() according
to (22) in Theorem 6. The lowest three curves corre-
spond to m = 1, the next three to m = 3, and the top
three to m = 8.
Eq. (29.32c)]:
g
(n)
m
() = e
k=0
(1)
n1
(n 1)!
(k +m) (k +m+n 1)
k
k!
.
(25)
4. FADING NUMBER OF AN IID GAUS-
SIAN FADING CHANNEL WITH A SCALAR
LINE-OF-SIGHT COMPONENT
In this section a small example is shown where some
of the results from Sections 2 and 3 can be applied. To
that goal we consider the following Gaussian MIMO
fading channel:
Y = Hx +Z. (26)
Here the n
R
-vector Y denotes the signal at the receiver,
the n
T
-vector x denotes the transmitted signal, Z is
additive white Gaussian noise, and the n
R
n
T
fad-
ing matrix H corresponds to multiplicative noise. It is
assumed that H can be written as
H = D +
H (27)
where all components of the n
R
n
T
random matrix
H are IID N
C
(0, 1) and where the constant n
R
n
T
line-of-sight matrix D is scalar in the sense that for
some constant d C we have in the case n
R
n
T
:
D = d
_
I
n
R
0
n
R
(n
T
n
R
)
_
(28)
and in the case n
R
> n
T
:
D = d
_
I
n
T
0
(n
R
n
T
)n
T
_
, (29)
where I
n
denotes the n n identity matrix and 0
nm
is the n m zero-matrix.
Note that conditional on the input x the squared
magnitude of the output Y
2
is noncentral chi-square
distributed.
For this fading channel we would like to derive the
asymptotic capacity in the case when the available SNR
tends to innity. In particular we are interested in the
fading number which is the second term in the high-
SNR expansion of capacity and has been introduced in
[4]:
(H) lim
snr
_
C(snr) log(1 +log(1 +snr))
_
. (30)
An expression for the general memoryless MIMO fading
number has already been derived in [10]:
(H) = sup
Q
X
_
h
_
H
X
H
X
_
+n
R
E
_
log H
X
2
_
log 2 h
_
H
X
_
_
, (31)
however, this expression still contains an optimization.
Hence, we would like to evaluate (31) in the situation of
the IID MIMO Gaussian fading channel with a scalar
line-of-sight component (27)(29).
Again we omit all proofs.
Theorem 7. Assume n
R
n
T
and a Gaussian fading
matrix as given in (27) and (28). Then
(H) = n
R
g
n
R
_
|d|
2
_
n
R
log (n
R
) (32)
where g
m
() is dened in (7).
For the case n
R
> n
T
we rstly need to introduce
some notation. We note that
H
X = D
X+
H
X
L
=
_
d
X
0
_
+
H, (33)
where
H N
C
(0, I
n
R
), and we split the vector
H into
two parts:
H =
_
H
1
H
2
_
(34)
where
H
1
N
C
(0, I
n
T
) and
H
2
N
C
(0, I
n
R
n
T
)
are two independent white Gaussian random vectors
in C
n
T
and C
n
R
n
T
, respectively. Then we write
H
X
L
=
_
d
X+
H
1
H
2
_
(35)
and dene
S
1
_
_
d
X+
H
1
_
_
2
, (36)
S
2
_
_
H
2
_
_
2
. (37)
Note that S
1
is noncentral chi-square distributed with
2n
T
degrees of freedom and noncentrality parameter
d
X
2
= |d|
2
independently of the distribution of
X, and that S
2
is central chi-square distributed with
2(n
R
n
T
) degrees of freedom. Moreover, S
1
and S
2
are independent of each other.
Theorem 8. Assume n
R
> n
T
and a Gaussian fading
matrix as given in (27) and (29). Then
(H) = n
T
g
n
T
_
|d|
2
_
n
T
log (n
T
) +I
_
S
1
;
S
2
S
1
_
.
(38)
Unfortunately, we have not succeeded in deriving
the term I(S
1
; S
2
/S
1
) precisely. However, we can state
the asymptotic behavior when d gets large.
Corollary 9. The fading number of the IID MIMO
Gaussian fading channel as dened in (27)(29) is
given by
(H) = n
m
g
n
m
_
|d|
2
_
n
m
log
_
n
m
_
+f(n
R
, n
T
; d)
(39)
where
n
m
min{n
R
, n
T
} (40)
is called degree of freedom of a MIMO fading channel
and where f(n
R
, n
T
; d) depends primarily on n
T
and
n
R
and is bounded in d:
0 f(n
R
, n
T
; d) (n
R
n
T
1)(n
R
n
T
)
(n
R
n
T
1) log
(n
R
n
T
)
n
R
n
T
+log
_
min
_
n
T
n
T
1
,
e
2
__
. (41)
Using (22) we therefore have for |d| 1
n
m
log
_
|d|
2
+n
m
_
. (42)
Acknowledgment
This work was supported by the NCTU/ITRI Joint
Research Center under contract G1-97006.
References
[1] N. L. Johnson, S. Kotz, and N. Balakrishnan, Con-
tinuous Univariate Distributions, 2nd ed. John
Wiley & Sons, 1995, vol. 2.
[2] M. E. Bock, G. G. Judge, and T. A. Yancey, A
simple form for the inverse moments of noncentral
2
and F random variables and certain conu-
ent hypergeometric functions, J. Econometrics,
vol. 25, no. 12, pp. 217234, MayJun. 1984.
[3] T. Fujioka, Asymptotic approximations of the in-
verse moment of the noncentral chi-squared vari-
able, J. Japan Statist. Soc., vol. 31, no. 1, pp.
99109, 2001.
[4] A. Lapidoth and S. M. Moser, Capacity bounds
via duality with applications to multiple-antenna
systems on at fading channels, IEEE Trans. Inf.
Theory, vol. 49, no. 10, pp. 24262467, Oct. 2003.
[5] S. M. Moser, The fading number of multiple-
inputmultiple-output fading channels with mem-
ory, May 2008, to app. in IEEE Trans. Inf. The-
ory. [Online]. Available: http://moser.cm.nctu.ed
u.tw/publications.html
[6] S. M. Moser, Some expectations of a noncentral
chi-square distribution with an even number of
degrees of freedom, Apr. 2007, subm. [Online].
Available: http://moser.cm.nctu.edu.tw/publicat
ions.html
[7] I. S. Gradshteyn and I. M. Ryzhik, Table of In-
tegrals, Series, and Products, 5th ed., A. Jerey,
Ed. Academic Press, San Diego, 1996.
[8] S. M. Moser, Duality-based bounds on channel
capacity, Ph.D. dissertation, Swiss Fed. Inst. of
Technol. (ETH), Zurich, Oct. 2004, Diss. ETH
No. 15769. [Online]. Available: http://moser.cm
.nctu.edu.tw/
[9] J. P. Cohen, Noncentral chi-square: Some ob-
servations on recurrence, The American Statisti-
cian, vol. 42, no. 2, pp. 120122, 1988.
[10] S. M. Moser, The fading number of memoryless
multiple-input multiple-output fading channels,
IEEE Trans. Inf. Theory, vol. 53, no. 7, pp. 2652
2666, Jul. 2007.