Image

Figure 8.5:  ROC curve for single-source detection, K = 1, N = 4, M = 8, SNR = −3 dB, FAR range of practical interest.

The integral in the numerator is then extended on its first column, the remainder of the matrix having a Vandermonde determinant. What remains is then an integral form of the g1 parameter. Hence the result.

The scenario where K ≥ 1 unfolds similarly. The final theorem can be found in [2]. The receiver operating characteristics (ROC) curve of the Neyman-Pearson test against that of the energy detector is provided in Figure 8.5 for N = 4, M = 8 and σ2 =3 dBm. We observed a significant performance gain in terms of detection rate incurred by the Neyman-Pearson test compared to the classical energy detector.

This completes this section on hypothesis testing. In the following section, we go beyond the hypothesis test and move to the question of parameter inference in a slightly more complex data plus noise model than above.

8.5.2      Parameter Estimation

We consider a similar scenario as in the previous section, where now the K sources use different transmit powers P1,…,PK, which the receiver is entitled to infer from successive observations.

Consider K data sources which are transmitting data simultaneously. Transmitter k ∈ {1,…, K} has power Pk and has space dimension nk, e.g., is composed of nk antennas. We denote nΔ__Kk = 1nknΔKk = 1nk the total number of transmit dimensions. Consider also a sink or receiver with space dimension N, N > n. Denote HkN × nkHkCN × nk the multidimensional filter matrix between transmitter k and the receiver. We assume that the entries of NHkNHk are independent and identically distributed with zero mean, unit variance and finite fourth order moment. At time instant m, transmitter k emits the signal x(m)knkx(m)kCnk, with entries assumed to be independent, independent along m, k, identically distributed along m, and have all zero mean, unit variance and finite fourth order moment (the x(m)kx(m)k need not be identically distributed along k). Assume further that at time instant m the receive signal is impaired by additive white Gaussian noise with entries of zero mean and variance σ2, denoted σw(m)∈ ℂN. At time m, the receiver therefore senses the signal y(m) ∈ ℂN defined as

y(m) = Kk = 1PkHkx(m)k + σw(m).y(m) = k = 1KPkHkx(m)k + σw(m).

Assuming the filter coefficients are constant over at least M consecutive sampling periods, by concatenating M successive signal realizations into Y = [y(1),…, y(M)] ∈ ℂN×M, we have

Y = Kk = 1PkHkXk + σW,Y = k = 1KPkHkXk + σW,

where Xk = [x(1)k,x(M)k]nk × MXk = [x(1)k,x(M)k]Cnk × M, for every k, and W = [w(1),…, w(M)] ∈ ℂN×M. This can be further rewritten as

Y = HP12X + σW,Y = HP12X + σW,

(8.31)

where P ∈ ℝn×n is diagonal with first n1 entries P1, subsequent n2 entries P2, etc. and last nK entries PK, H = [H1,…, HK] ∈ ℂN×n and X = [XT1,XTK]Tn × MX = [XT1,XTK]TCn × M. By convention, we assume P1 ≤ …≤ PK.

Our objective is to infer the values of the powers P1,…, PK from the realization of a single random matrix Y. This is successively performed from different approaches in the following sections. We first consider the conventional approach that assumes n small, N much larger than n, and M much larger than N. This will lead to a simple although largely biased estimation algorithm. This algorithm will be improved using Stieltjes transform approaches in the same spirit as in Section 8.4.

Conventional Approach

The first approach assumes numerous sensors in order to have much diversity in the observation vectors, as well as an even larger number of observations, in order to create an averaging effect on the incoming random data. In this situation, let us rewrite (8.31) under the form

Y = (HP12σIN)(XW).Y = (HP12σIN)(XW).

(8.32)

We shall denote λ1 ≤ … ≤ λN the ordered eigenvalues of 1MYY1MYY (the non-zero eigenvalues of which are almost surely different).

Appending Y ∈ ℂN×M into the larger matrix ∈ ℂ(N+nM

Y_ = (HP12σIN00)(XW),Y = (HP120σIN0)(XW),

we recognize that, conditioned on H,1MYY_H,1MYY is a sample covariance matrix, for which the population covariance matrix is

T(HPH + σ2IN000)T(HPH + σ2IN000)

and the random matrix

(XW)(XW)

has independent (non-necessarily identically distributed) entries with zero mean and unit variance. The population covariance matrix T, whose upper left entries also form a matrix unitarily equivalent to a sample covariance matrix, clearly has an almost sure limit spectral distribution as N grows large for fixed or slowly growing n. Extending Theorem 8.2.9 and Theorem 8.3.3 to c = 0 and applying them twice (once for the population covariance matrix T and once for 1MYY_1MYY, we finally have that, as M, N, n → ∞ with M/N → ∞ and N/n, the distribution of the largest n eigenvalues of 1MYY_1MYY is asymptotically almost surely composed of a mass σ2 + P1 of weight lim n1/n, a mass σ2 + P2 of weight lim n2/n, etc. and a mass σ2 + PK of weight lim nK/n. As for the distribution of the smallest Nn eigenvalues of 1MYY1MYY, it converges to a single mass in σ2.

If σ2 is a priori known, a rather trivial estimator of Pk is then given by

1nkiNk(λi  σ2),1nkiNk(λi  σ2),

where Nk{k  1j = 1nj + 1,,kj = 1nj}Nk{k  1j = 1nj + 1,,kj = 1nj} and we recall that λ1 ≤ … ≤ λN are the ordered eigenvalues of 1MYY1MYY.

This means in practice that PK is asymptotically well approximated by the averaged value of the nK largest eigenvalues of 1MYY1MYY, PK − 1 is well approximated by the averaged value of the nK−1 eigenvalues before that, etc. This also assumes that σ2 is perfectly known at the receiver. If it were not, observe that the averaged value of the Nn smallest eigenvalues of 1MYY1MYY is a consistent estimate for σ2. This therefore leads to the second estimator ˆPkPˆk for Pk, that will constitute our reference estimator,

ˆPk = 1nkiNk(λi  ˆσ2),Pˆk = 1nkiNk(λi  σˆ2),

where

ˆσ2 = 1N  nN  ni = 1λi.σˆ2 = 1N  ni = 1N  nλi.

Incidentally, although not derived on purpose, the refined (n, N, M)-consistent estimator of Section 8.5.2 will appear not to depend on a prior knowledge of σ2. Note that the estimation of Pk only relies on nk contiguous eigenvalues of 1MYY1MYY, which suggests that the other eigenvalues are asymptotically uncorrelated from these. It will turn out that the improved (n, N, M)-consistent estimator does take into account all eigenvalues for each k, in a certain manner.

The Stieltjes Transform Method

The Stieltjes transform approach relies heavily on the techniques from Mestre, established in [24] and introduced in Section 8.4. We provide hereafter only the main steps of the method. The details can be found in [23].

Limiting spectrum of BN. In this section, we prove the following result.

Theorem 8.5.3. Let BN = 1MYYBN = 1MYY, with Y defined as in (8.31). Then, for M, N, n growing large with limit ratios M/N → c, N/nk → ck , 0 < c, c1, …, cK < ∞, the empirical spectral distribution FBNFBN of BN converges almost surely to the distribution function F, whose Stieltjes transform mF(z) satisfies, for z ∈ ℂ+,

mF(z) = cmF_(z) + (c  1)1z,mF(z) = cmF(z) + (c  1)1z,

(8.33)

where m(z) is the unique solution with positive imaginary part of the implicit equation in m ,

1mF_ =   σ2 + 1f  Kk = 11ckPkPkf1mF =   σ2 + 1f  k = 1K1ckPkPkf

(8.34)

in which we denoted f the value

f(1  c)mF_  czm2F_.f(1  c)mF  czm2F.

Proof. First remember that the matrix Y in (8.31) can be extended into the larger sample covariance matrix ∈ ℂ(N+nM

Y_ = (HP12σIN00)(XW).Y = (HP120σIN0)(XW).

From Theorem 8.2.9, since H has independent entries with finite fourth order moments, we have that the e.s.d. of HPH† converges weakly and almost surely to a limit distribution G as N, n1,…, nK → ∞ with N/nkck > 0. For z ∈ ℂ+, the Stieltjes transform mG(z) of G is the unique solution with positive imaginary part of the equation in mG,

z =   1mG + kk = 11ckPk1 + PkmG.z =   1mG + k = 1k1ckPk1 + PkmG.

(8.35)

The almost sure convergence of the e.s.d. of HPH† ensures the almost sure convergence of the e.s.d. of the matrix (HPH0 + σ2IN00)(HPH0 + σ2IN00). Since mG(z) evaluated at z ∈ ℂ+ is the Stieltjes transform of the l.s.d. of HPH† + σ2IN evaluated at z + σ2, adding n zero eigenvalues, we finally have that the e.s.d. of (HPH0 + σ2IN00)(HPH0 + σ2IN00) tends almost surely to a distribution H whose Stieltjes transform mH(z) satisfies

mH(z) = c01 + c0mG(z  σ2)  11 + c01z,mH(z) = c01 + c0mG(z  σ2)  11 + c01z,

(8.36)

for z ∈ ℂ+, where we denoted by c0 the limit of the ratio N/n, i.e., c0 = (c  11 +  + c  1K)  1c0 = (c  11 +  + c  1K)  1.

As a consequence, the sample covariance matrix 1MYY_1MYY has a population covariance matrix which is not deterministic but whose e.s.d. has an almost sure limit H for increasing dimensions. Since X and W have entries with finite fourth order moment, we can again apply Theorem 8.2.9 and we have that the e.s.d. of B_NΔ__1MY_Y_BNΔ1MYY converges almost surely to the limit whose Stieltjes transform m(z) is the unique solution in ℂ+ of the equation in mF̲

z =   1mF_ + 1c(1 + 1c0)t1 + tmF_dH(t) =   1mF_ + 1 + 1c0cmF_(1  1mF_mH(  1mF_))z =  =   1mF + 1c(1 + 1c0)t1 + tmFdH(t)  1mF + 1 + 1c0cmF(1  1mFmH(  1mF))

(8.37)

for all z ∈ ℂ+.

For z ∈ ℂ+, m(z) ∈ ℂ+. Therefore −1/m(z) ∈ ℂ+ and one can evaluate (8.36) at −1/m(z). Combining (8.36) and (8.37), we then have

z =   1c1cmF_(z)2mG(  1mF_(z)  σ2) + (1c  1)1mF_(z),z =   1c1cmF(z)2mG(  1mF(z)  σ2) + (1c  1)1mF(z),

(8.38)

where, according to (8.35), mG(−1/mF(z) − σ2) satisfies

1mF_(z) =   σ2 + 1mG(  1mF_(z)  σ2)  Kk = 11ckPk1 + PkmG(  1mF_(z)  σ2).1mF(z) =   σ2 + 1mG(  1mF(z)  σ2)  k = 1K1ckPk1 + PkmG(  1mF(z)  σ2).

(8.39)

Together with (8.38), this is exactly (8.34), with f(z) = mG(  1mF_(z)  σ2) = (1  c)mF_(z)  czmF_(z)2f(z) = mG(  1mF(z)  σ2) = (1  c)mF(z)  czmF(z)2.

Since the eigenvalues of the matrices BN and N only differ by MN zeros, we also have that the Stieltjes transform mF(z) of the l.s.d. of BN satisfies

mF(z) = cmF_(z) + (c  1)1z.mF(z) = cmF(z) + (c  1)1z.

(8.40)

This completes the proof of Theorem 8.5.3.

For further usage, notice here that (8.40) provides a simplified expression for mG (−1/m(z) − σ2). Indeed we have,

mG(  1/mF_(z)  σ2) =   zmF(z)mF_(z).mG(  1/mF(z)  σ2) =   zmF(z)mF(z).

(8.41)

Therefore, the support of the (almost sure) l.s.d. F of BN can be evaluated as follows: for any z ∈ ℂ+, mF(z) is given by (8.33), in which m(z) is solution of (8.34); the inverse Stieltjes transform formula (8.4) allows then to evaluate F from mF(z), for values of z spanning over the set {z = x + iy, x > 0} and y small.

Multisource power inference. In the following, we finally prove the main result of this section, which provides the G-estimator 1,…, K of the transmit powers P1,…, PK.

Theorem 8.5.4. Let BN ∈ ℂN×N be defined as BN = 1MYYBN = 1MYY with Y defined as in (8.31), and λ = (λ1,…, λN), λ1 ≤ … ≤ λN, be the vector of the ordered eigenvalues of BN. Further assume that the limiting ratios c0, C1,…,CK, c and P are such that the cluster mapped to Pk in BN does not map another Pi, ik. Then, as N, n, M grow large, we have

ˆPk  Pka.s.0,Pˆk  Pka.s.0,

where the estimate P̂k is given by

•  if MN,

ˆPk = NMnk(M  N)iNk(ηi  μi),Pˆk = NMnk(M  N)iNk(ηi  μi),

•  if M = N,

ˆPk = Nnk(N  n)iNk(Nj = 1ηi(λj  ηi)2)  1,Pˆk = Nnk(N  n)iNk(j = 1Nηi(λj  ηi)2)  1,

in which Nk{k  1j = 1nj + 1,,kj = 1nj},η1ηNNk{k  1j = 1nj + 1,,kj = 1nj},η1ηN are the ordered eigenvalues of the matrix diag (λ)  1Nλλ(λ)  1Nλλ and μ1 ≤ … ≤ are the ordered eigenvalues of the matrix diag (λ)  1Mλλ(λ)  1Mλλ.

Remark 8.5.1. We immediately notice that, if N < n, the powers P1,…,Pl , with l the largest integer such that N  Ki = lni<0N  Ki = lni<0, cannot be estimated since clusters may be empty. The case Nn turns out to be of no practical interest as clusters always merge and no consistent estimate of either Pi can be described.

Proof. The approach pursued to prove Theorem 8.5.4 relies strongly on the original idea of [22], which was detailed for the case of sample covariance matrices in Section 8.4. From Cauchy’s integration formula,

Pk = ck12πick1ckωPk  ωdω = ck12πick1crωPr  ωdωPk = ck12πick1ckωPk  ωdω = ck12πick1crωPr  ωdω

(8.42)

for any negatively oriented contour Ck ⊂ ℂ, such that Pk is contained in the surface described by the contour, while for every ik, Pi is outside this surface. The strategy is very similar to that used for the sample covariance matrix case in Section 8.4. It comes as follows: we first propose a convenient integration contour Ck which is parametrized by a functional of the Stieltjes transform mF (z) of the l.s.d. of BN. We proceed to a variable change in (8.42) to express Pk as a function of mF(z). We then evaluate the complex integral resulting from replacing the limiting mF(z) in (8.42) by its empirical counterpart ˆmF(z) = 1Ntr(BN  zIN)  1mˆF(z) = 1Ntr(BN  zIN)  1. This new integral, whose value we name k, is shown to be almost surely equal to Pk in the large N limit. It then suffices to evaluate k, which is just a matter of residue calculus.

Similar to Section 8.4, it turns out that the clusters generated in the spectrum of BN can be mapped to one or many power values Pk. In what follows, we assume that the clusters are disjoint so that no holomorphicity problem arises. We can prove the following (given in detail in [23] and [28]). There exist x(l)kFx(l)kF and x(r)kFx(r)kF outside the support of F, on either side of cluster kF (i.e., the cluster in F that is uniquely mapped to Pk) such that m(z) has limits m(l)F_,kGΔ__mF(x(l)kF)m(l)F,kGΔmF(x(l)kF) and m(r)F_,kGΔ__mF(x(r)kF)m(r)F,kGΔmF(x(r)kF), as zx(l)kFzx(l)kF and zx(r)kFzx(r)kF, respectively, with mFmF the analytic extension of m in the points x(l)kFx(l)kFR and x(r)kFx(r)kFR. These limits m(l)F_,kGm(l)F,kG and m(r)F_,kGm(r)F,kG are on either side of cluster kG (i.e., the cluster in G mapped uniquely to Pk) in the support of 1/H, and therefore   1/m(l)F_,kG  σ2  1/m(l)F,kG  σ2 and   1/m(l)F_,kG  σ2  1/m(l)F,kG  σ2 are on either side of cluster kG in the support of G.

Consider any continuously differentiable complex path ΓF ,k with endpoints x(l)kFx(l)kF and x(r)kFx(r)kF, and interior points of positive imaginary part. We define the contour CF,k as the union of ΓF ,k oriented from x(l)kFx(l)kF to x(r)kFx(r)kF and its complex conjugateΓ*F,kΓF,k oriented backwards from x(r)kFx(r)kF to x(l)kFx(l)kF. The contour CF ,k is clearly continuous and piecewise continuously differentiable. Also, the support of cluster kF in is completely inside CF,k, while the support of the neighboring clusters is away from CF,k. The support of cluster kG in H is then inside −1/m(CF,k), and therefore the support of cluster kG in G is inside CGk ≜ −1/mF(CGk) − σ2. Since m is continuously differentiable on CC ℝ (it is in fact holomorphic [21]) and has limits in x(l)kFx(l)kF and x(r)kFx(r)kF is also continuous and piecewise continuously differentiable. Going one last step in this process, we finally have that Pk is inside the contour Ck ≜ −1/mG(CG,k), while Pi, for all ik, is outside Ck. Since mG is also holomorphic on CC ℝ and has limits in   1/mF(x(l)kF)  σ2  1/mF(x(l)kF)  σ2 and   1/mF(x(r)kF)  σ2  1/mF(x(r)kF)  σ2, Ck is a continuous and piecewise continuously differentiable complex path, which is sufficient to perform complex integration [30].

Recall now that Pk was defined as

Pk = ck12πickKr = 11crωPr  ωdω.Pk = ck12πickr = 1K1crωPr  ωdω.

With the variable change ω = − 1/mG (t), this becomes

Pk = ck12πiCG,kKr = 11cr  11 + PrmG(t)mG(t)mG(t)2dt = ck12πiCG,k(mG(t)[  1mG(t)Kr = 11cr  11 + PrmG(t)] + c0  1c0)mG(t)mG(t)2dt.Pk =  = ck12πiCG,kr = 1K1cr  11 + PrmG(t)mG(t)mG(t)2dtck12πiCG,k(mG(t)[  1mG(t)r = 1K1cr  11 + PrmG(t)] + c0  1c0)mG(t)mG(t)2dt.

From Equation (8.35), this simplifies into

Pk = ckc012πiCG,k(c0tmG(t) + c0  1)mG(t)mG(t)2dt.Pk = ckc012πiCG,k(c0tmG(t) + c0  1)mG(t)mG(t)2dt.

(8.43)

Using (8.38) and proceeding with the further change of variable t = −1/m(z) − σ2, (8.43) becomes

Pk = ckc012πiCG,k(1 + σ2mF_(z))[  1zmF_(z)  mF_(z)mF_(z)2  mF(z)mF(z)mF_(z)]dz.Pk = ckc012πiCG,k(1 + σ2mF(z))[  1zmF(z)  mF(z)mF(z)2  mF(z)mF(z)mF(z)]dz.

(8.44)

This whole process of variable changes allows us to describe Pk as a function of mF (z), the Stieltjes transform of the almost sure limiting spectral distribution of BN, as N → ∞. It then remains to exhibit a relation between Pk and the empirical spectral distribution of BN for finite N. This is what the subsequent section is dedicated to.

Let us now define F (z) and (z) as the Stieltjes transforms of the empirical eigenvalue distributions of BN and N, respectively, i.e.,

ˆmF(z) = 1NNi = 11λi  zmˆF(z) = 1Ni = 1N1λi  z

(8.45)

and

ˆmF_(z) = NMˆmF(z)  M  NM1z.mˆF(z) = NMmˆF(z)  M  NM1z.

Instead of going further with (8.44), define Pk, the “empirical counterpart” of Pk , as

ˆPk = nnk12πiCF,kNn(1 + σ2ˆmF_(z))[  1zˆmF_(z)  ˆmF_(z)ˆmF_(z)2  ˆmF_(z)ˆmF(z)ˆmF_(z)]dz.Pˆk = nnk12πiCF,kNn(1 + σ2mˆF(z))[  1zmˆF(z)  mˆF(z)mˆF(z)2  mˆF(z)mˆF(z)mˆF(z)]dz.

(8.46)

The integrand can then be expanded into nine terms, for which residue calculus can easily be performed. Denote first η1,…, ηN the N real roots of F(z) = 0 and (μ1, …, μN the N real roots of (z) = 0. We identify three sets of possible poles for the nine aforementioned terms: (i) the set {λ1,λN}[x(l)kF,x(r)kF]{λ1,λN}[x(l)kF,x(r)kF], (ii) the set {η1,ηN}[x(l)kF,x(r)kF]{η1,ηN}[x(l)kF,x(r)kF], and (iii) the set {μ1,μN}[x(l)kF,x(r)kF]{μ1,μN}[x(l)kF,x(r)kF]. For MN, the full calculus leads to

ˆPk=NMnk(MN)[1iNx(l)kFηix(r)kFηi1iNx(l)kFηix(r)kFμi]+Nnk[1iNx(l)kFηix(r)kFσ21iNx(l)kFλix(r)kFσ2]+Nnk[1iNx(l)kFμix(r)kFσ21iNx(l)kFλix(r)kFσ2].Pˆk=NMnk(MN)1iNx(l)kFηix(r)kFηi1iNx(l)kFηix(r)kFμi+Nnk1iNx(l)kFηix(r)kFσ21iNx(l)kFλix(r)kFσ2+Nnk1iNx(l)kFμix(r)kFσ21iNx(l)kFλix(r)kFσ2.

(8.47)

Now, we know from Theorem 8.5.3 that ˆmF(z) = a.s.mF(z) and ˆmF_(z) = a.s.mF_(z) as N → ∞. Observing that the integrand in (8.46) is uniformly bounded on the compact CF,k, the dominated convergence theorem, Theorem 16.4 of [16], ensures ˆmF(z) = a.s.mF(z)1Ntr(BN  zIN)  1ˆPka.s.Pk.

To go further, we now need to determine which of λ1, …, λN, η1, …, ηN and (η1, …, ηN lie inside CF ,k. It can be proved, by extending Theorem 8.3.1 and Theorem 8.3.4 to the current model, that there will be no eigenvalue of BN (or N) outside the support of F, and the number of eigenvalues inside cluster kF is exactly nk. Since CF,k encloses cluster kF and is away from the other clusters, {λ1,λN}[x(l)kF,x(r)kF] = {λiiNk} almost surely, for all N large. Also, for any i ∈ {1, …, N}, it is easy to see from (8.45) that F(z) → ∞ when zλi and F(z) → −∞ when zλi. Therefore F(z) → ∞ has at least one solution in each interval (λi−1, λi), with λ0 = 0, hence μ1 < λ1 < μ2 … < μN < λN. This implies that, if k0 is the index such that CF,k contains exactly λk0,,λk0 + (nk  1), then CF,k also contains {μk0 + 1,,μk0 + (nk  1)}. The same result holds for ηk0 + 1,,ηk0 + (nk  1). When the indexes exist, due to cluster separability, ηk0  1 and μk0  1 belong, for N large, to cluster kF − 1. We are then left with determining whether μk0 and ηk0 are asymptotically found inside CF,k.

For this, we use the same approach as in [22] by noticing that, since 0 is not included in Ck , one has

12πiCk1ωdω = 0.

Performing the same changes of variables as previously, we have

CF,k  mF_(z)mF(z)  zmF_(z)mF(z)  zmF_(z)mF_(z)z2mF_(z)2mF(z)2dz = 0.

(8.48)

For N large, the dominated convergence theorem ensures again that the left-hand side of the (8.48) is close to

CF,k  ˆmF_(z)(z)  z^mF_(z)(z)  zˆmF_(z)(z)z2ˆmF_(z)2(z)2dz.

(8.49)

Residue calculus of (8.49) then leads to

1iNλi[x(l)kF,x(r)kF]2  1iNηi[x(l)kF,x(r)kF]1  1iNμi[x(l)kF,x(r)kF]1a.s.0.

(8.50)

Since the cardinalities of {i,ηi[x(l)kF,x(r)kF]} and i,μi[x(l)kF,x(r)kF] are at most nk, (8.50) is satisfied only if both cardinalities equal nk in the limit. As a consequence, μk0[x(l)kF,x(r)kF] and ηk0[x(l)kF,x(r)kF]. For N large, NM, this allows us to simplify (8.47) into

ˆPk = NMnk(M  N)1iNλiNk(ηi  μi)

(8.51)

with probability one. The same reasoning holds for M = N. This is our final relation. It now remains to show that the ηi and the μi are the eigenvalues of diag (λ)  1NλλT and diag (λ)  1MλλT, respectively. But this is merely a consequence of [23, Lemma 1].

This concludes the proof of Theorem 8.5.4.

Image

Figure 8.6:  Distribution function of the estimators ˆPk and k for k ∈ {1, 2, 3}, P1 = 1/16, P2 = 1/4, P3 = 1, n1 = n2 = n3 = 4 antennas per user, N = 24 sensors, M = 128 samples, and SNR = 20 dB. Optimum estimator is shown in dashed lines.

We now evaluate the performance difference between the traditional and the Stieltjes transform inference methods, for K = 3 sources, P1 = 1/16, P2 = 1/4, N = 24 sensors, M = 128 samples, and n1 = n2 = n3 = 4. This is provided in Figure 8.6 which compares the distribution function of the estimates engendered by both methods. As anticipated, we observe a significant gain in terms of bias reduction for the Stieltjes transform method.

8.6      Conclusion

Random matrix theory for signal processing is a fast growing field of research whose interest is mainly motivated by the increase of the dimensionality and complexity of today’s systems. While the first years of random matrix theory were mainly focusing on Gaussian and invariant matrix distributions, the last ten years of research were mainly targeting large dimensional matrices with independent entries. This provided interesting results in particular on the limiting spectrum of sample covariance matrices, which led to new results on inverse problems for large dimensional systems. These results are often surprisingly simple and efficient as they perform well against exact maximum likelihood solutions, even for systems of not too large dimensions. Much more is however needed from a mathematical viewpoint relative in particular to second order statistics, see e.g., [35, 36], in order to evaluate theoretically the performance of these methods as well as a generalization to more intricate random matrix structures, such as Vandermonde matrices for array processing, see e.g., [37], or unitary random matrices, see e.g., [38]. A more exhaustive account of random matrix methods as well as more details on the methods presented here can be found in [10, 17, 28].

8.7      Exercises

Exercise 8.7.1 (Sampling and Signal Energy). Based on Theorem 8.2.9, prove the Marc̆enko-Pastur law, Theorem 8.2.5.

Hint 8.7.1. Observe that the fixed-point equation in mFB reduces now to a second order polynomial from which mFB(z) takes an explicit form. The inverse Stieltjes transform formula 8.4 gives the expression of FB.

Exercise 8.7.2. Let XN ∈ ℂN×n be a random matrix with i.i.d. Gaussian entries of zero mean and variance 1/n. For RN ∈ ℂN×N and TN ∈ ℂn×n deterministic and of uniformly bounded spectral norm such that FRNFR and FTNFT, as N,n → ∞, determine an expression of the Stieltjes transform of the limiting eigenvalue distribution of BN=R12NXNTNXNR12N as N/nc.

Hint 8.7.2. Follow the proof of Theorem 8.2.9 by looking for a deterministic equivalent of 1Ntr(A(BN  zIN)  1) for some deterministic A, taken to be successively RN and IN. A good choice of the matrix DN is DN = aNRN.

Exercise 8.7.3. Based on the definition of the Shannon-transform and on the G-estimator for the Stieltjes transform, determine a G-estimator for

VTN(x) = 1Nlogdet(xTN + IN)

based on the observations

yk = T12Nxk

with xk ∈ ℂN with i.i.d. entries of zero mean and variance 1/N, independent across k, for k ∈ {1,…, n}.

Hint 8.7.3. Write the expression of VTN(x) as a function of the Stieltjes transform of TN and operate a variable change in the resulting integral using Theorem 8.2.9.

Exercise 8.7.4. From the result of Theorem 8.3.3, propose a hypothesis test for the presence of a signal transmitted by a signal source and observed by a large array of sensors, assuming that the additive noise variance is either perfectly known or not.

Hint 8.7.4. Observe that the ratio of the extreme eigenvalues in both H0 and H1 hypotheses is asymptotically independent of the noise variance.

Exercise 8.7.5. For W ∈ ℂN×n, n < N, the n columns of a random unitarily invariant unitary matrix, w a column vector of W, prove that, if BN is a random matrix with bounded spectral norm, function of all columns of W but w, then, as N, n → ∞ with n/Nc < 1,

wBNw  1N  ntr(()IN  WW)BNa.s.0.

Hint 8.7.5. write w as the normalized projection of a Gaussian vector x on the subspace orthogonal to the space spanned by the columns of W but w, i.e., w = Πx, with Π = INWW + WW.

References

[1]  J. Wishart, “The generalized product moment distribution in samples from a normal multivariate population,” Biometrika, vol. 20, no. 1–2, pp. 32–52, Dec. 1928.

[2]  R. Couillet and M. Debbah, “A Bayesian framework for collaborative multisource signal detection,” IEEE Transactions on Signal Processing, vol. 58, no. 10, pp. 5186–5195, Oct. 2010.

[3]  P. Bianchi, J. Najim, M. Maida, and M. Debbah, “Performance of Some Eigenbased Hypothesis Tests for Collaborative Sensing,” Proceedings of the IEEE Workshop on Statistical Signal Processing, SSP’09, Sep., 2010.

[4]  R. A. Fisher, “The sampling distribution of some statistics obtained from non-linear equations,” The Annals of Eugenics, vol. 9, pp. 238–249, 1939.

[5]  M. A. Girshick, “On the sampling theory of roots of determinantal equations,” The Annals of Math. Statistics, vol. 10, pp. 203–204, 1939.

[6]  P. L. Hsu, “On the distribution of roots of certain determinantal equations,” The Annals of Eugenics, vol. 9, pp. 250–258, 1939.

[7]  S. Roy, “p-statistics or some generalizations in the analysis of variance appropriate to multi-variate problems,” Sankhya: The Indian Journal of Statistics, vol. 4, pp. 381–396, 1939.

[8]  Harish–Chandra, “Differential operators on a semi-simple Lie algebra,” American Journal of Mathematics, vol. 79, pp. 87–120, 1957.

[9]  C. Itzykson and J. B. Zuber, Quantum Field Theory. McGraw-Hill, 1980; Dover Publications, 2005, p. 705.

[10]  Z. Bai and J. W. Silverstein, Spectral Analysis of Large Dimensional Random Matrices, Springer Series in Statistics, 2009.

[11]  P. Billingsley, Convergence of Probability Measures. Hoboken, NJ: John Wiley & Sons, Inc., 1968.

[12]  S. Wagner, R. Couillet, M. Debbah, and D. T. M. Slock, “Large System Analysis of Linear Precoding in MISO Broadcast Channels with Limited Feedback,” 2010. [Online]. Available:http://arxiv.org/abs/0906.3682

[13]  V. A. Marc̆enko and L. A. Pastur, “Distributions of eigenvalues for some sets of random matrices,” Math USSR-Sbornik, vol. 1, no. 4, pp. 457–483, Apr. 1967.

[14]  J. W. Silverstein and Z. D. Bai, “On the empirical distribution of eigenvalues of a class of large dimensional random matrices,” Journal of Multivariate Analysis, vol. 54, no. 2, pp. 175–192, 1995.

[15]  Z. D. Bai and J. W. Silverstein, “No Eigenvalues Outside the Support of the Limiting Spectral Distribution of Large Dimensional Sample Covariance Matrices,” Annals of Probability, vol. 26, no. 1, pp. 316–345, Jan. 1998.

[16]  P. Billingsley, Probability and Measure, 3rd ed. Hoboken, NJ: John Wiley & Sons, Inc., 1995.

[17]  A. M. Tulino and S. Verdú, “Random matrix theory and wireless communications,” Foundations and Trends in Communications and Information Theory, vol. 1, no. 1, 2004.

[18]  D. N. C. Tse and O. Zeitouni, “Linear multiuser receivers in random environments,” IEEE Transactions on Information Theory, vol. 46, no. 1, pp. 171–188, 2000.

[19]  J. Baik and J. W. Silverstein, “Eigenvalues of large sample covariance matrices of spiked population models,” Journal of Multivariate Analysis, vol. 97, no. 6, pp. 1382–1408, 2006.

[20]  Z. D. Bai and J. W. Silverstein, “Exact Separation of Eigenvalues of Large Dimensional Sample Covariance Matrices,” The Annals of Probability, vol. 27, no. 3, pp. 1536–1555, 1999.

[21]  J. W. Silverstein and S. Choi, “Analysis of the limiting spectral distribution of large dimensional random matrices,” Journal of Multivariate Analysis, vol. 54, no. 2, pp. 295–309, 1995.

[22]  X. Mestre, “On the asymptotic behavior of the sample estimates of eigenvalues and eigenvectors of covariance matrices,” IEEE Transactions on Signal Processing, vol. 56, no. 11, pp. 5353–5368, Nov. 2008.

[23]  R. Couillet, J. W. Silverstein, and M. Debbah, “Eigen-Inference for Energy Estimation of Multiple Sources,” 2011. [Online]. Available:http://arxiv.org/abs/1001.3934

[24]  X. Mestre, “Improved estimation of eigenvalues of covariance matrices and their associated subspaces using their sample estimates,” IEEE Transactions on Information Theory, vol. 54, no. 11, pp. 5113–5129, Nov. 2008.

[25]  F. Hiai and D. Petz, The semicircle law, free random variables and entropy - Mathematical Surveys and Monographs No. 77. Providence, RI, USA: American Mathematical Society, 2006.

[26]  Ø. Ryan and M. Debbah, “Free deconvolution for signal processing applications,” in (ISIT’07), Nice, France, June 2007, pp. 1846–1850.

[27]  N. R. Rao and A. Edelman, “The polynomial method for random matrices,” Foundations of Computational Mathematics, vol. 8, no. 6, pp. 649–702, Dec. 2008.

[28]  R. Couillet and M. Debbah, Random Matrix Methods for Wireless Communication., 1st ed. New York, NY: Cambridge University Press, 2011.

[29]  P. Vallet, P. Loubaton, and X. Mestre, “Improved subspace DoA estimation methods with large arrays: The deterministic signals case,” in (ICASSP’09), 2009, pp. 2137–2140.

[30]  W. Rudin, Real and Complex Analysis, 3rd ed. McGraw-Hill Series in Higher Mathematics, May 1986.

[31]  D. Gregoratti and X. Mestre, “Random DS/CDMA for the amplify and forward relay channel,” IEEE Transactions on Wireless Communications, vol. 8, no. 2, pp. 1017–1027, 2009.

[32]  R. Couillet and M. Debbah, “Free deconvolution for OFDM multicell SNR detection,” in (PIMRC’08), Cannes, France, 2008, pp. 1–5.

[33]  I. S. Gradshteyn and I. M. Ryzhik, “Table of Integrals, Series and Products,” Academic Press, 6th edition, 2000.

[34]  S. H. Simon, A. L. Moustakas, and L. Marinelli, “Capacity and character expansions: Moment generating function and other exact results for MIMO correlated channels,” IEEE Transactions on Information Theory, vol. 52, no. 12, pp. 5336–5351, 2006.

[35]  Z. D. Bai and J. W. Silverstein, “CLT of linear spectral statistics of large dimensional sample covariance matrices,” Annals of Probability, vol. 32, no. 1A, pp. 553–605, 2004.

[36]  J. Yao, R. Couillet, J. Najim, E. Mouline, and M. Debbah, “CLT for eigeninference methods in cognitive radios,” in (ICASSP’11), Prague, Czech Republic, 2011, pp. 2980–2983.

[37]  Ø. Ryan and M. Debbah, “Asymptotic behavior of random Vandermonde matrices with entries on the unit circle,” IEEE Transactions on Information Theory, vol. 55, no. 7, pp. 3115–3148, July 2009.

[38]  R. Couillet, J. Hoydis, and M. Debbah, “Deterministic equivalents for the analysis of unitary precoded systems,” IEEE Transactions on Information Theory, 2011.

1We remind that a unitary matrix U ∈ ℂN×N is such that UU† = UU = IN.

2Throughout this work, we will respect the convention that x (be it a scalar or an Hermitian matrix) is non-negative if x ≥ 0, while x is positive if x ≥ 0.

3The Hermitian property is fundamental to ensure that all eigenvalues of XN belong to the real line. However, the extension of the empirical spectral distribution (e.s.d.) to non-Hermitian matrices is sometimes required; for a definition, see (1.2.2) of [10].

4We borrow here the notation m due to a large number of contributions from Bai, Silverstein et al. [14, 15]. In other works, the notation s or S for the Stieltjes transform is used.

5We recall that the support Supp (F) of a real function F is the set {x ∈ℝ, ∣F(x)∣ > 0}.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
3.16.69.143