11.6.2    UMP Detection with Both Composite Hypotheses

We now consider the more general case where both the hypotheses are composite. The UMP optimization problem can be stated as:

MaximizePD(˜δ;θ),forallθΛ1,subjecttosupθΛ0PF(˜δ;θ)α.MaximizePD(δ˜;θ),forallθΛ1,subjecttosupθΛ0PF(δ˜;θ)α.

(11.21)

If a UMP test δ̃UMP exists, then it must satisfy the following conditions. First,

supθ0Λ0PF(˜δUMP;θ0)α.supθ0Λ0PF(δ˜UMP;θ0)α.

(11.22)

Second, for any δ̃ ∈ Δ̃ that satisfies supθ0Λ0PF(˜δ;θ0)αsupθ0Λ0PF(δ˜;θ0)α we must have

PD(˜δ;θ1)PD(˜δUMP;θ1)forallθ1Λ1.PD(δ˜;θ1)PD(δ˜UMP;θ1)forallθ1Λ1.

(11.23)

The following example illustrates a case where a UMP solution can be found. Also see Exercises 11.9.12 and 11.9.13.

Example 11.6.4. Testing Between Two One-Sided Composite Signals in Gaussian Noise. This is an extension of Example 11.6.1 in which the observation is Y = θ+Z, with Z ~ NN(0, σ2), and we are testing

H0:θΛ0 = [0,1]versusH1:θΛ1 = (1,).H0:θΛ0 = [0,1]versusH1:θΛ1 = (1,).

For fixed θ0 ∈ Λ0 and θ1 ∈ Λ1, Lθ0,θ1(y)Lθ0,θ1(y) has no point masses under Pθ0Pθ0 or Pθ1Pθ1, and therefore δ̃NP(y; θ0, θ1) is a deterministic LRT:

δNP(y:θ0,θ1) = {1ifLθ0,θ1(y)η(θ0,θ1)0ifLθ0,θ1(y)<η(θ0,θ1) = {1ifyη(θ0,θ1)0ify<η(θ0,θ1)δNP(y:θ0,θ1) = {10ififLθ0,θ1(y)η(θ0,θ1)Lθ0,θ1(y)<η(θ0,θ1) = {10ififyη(θ0,θ1)y<η(θ0,θ1)

where η′(θ0, θ1) is given by

η(θ0,θ1) = σ2logη(θ0,θ1)θ1  θ0 + θ0 + θ12.η(θ0,θ1) = σ2logη(θ0,θ1)θ1  θ0 + θ0 + θ12.

Now in order to set the threshold η′ to meet the constraint on PF given in (11.21), we first compute:

PF(δη;θ0) = Pθ0{Yη} = Q(η  θ0σ)PF(δη;θ0) = Pθ0{Yη} = Q(η  θ0σ)

and note that this probability is an increasing function in θ0. Therefore

supθ0[0,1] PF(δη;θ0) = Q(η  1σ)supθ0[0,1] PF(δη;θ0) = Q(η  1σ)

and we can meet the PF constraint with equality by setting η′ such that:

Q(η  1σ) = αηα = σQ  1(α) + 1.Q(η  1σ) = αηα = σQ  1(α) + 1.

Note that η′α is independent of θ0 and θ1. Define the test

δηα(y) = {1ifyηα0ify<ηα.δηα(y) = {10ififyηαy<ηα.

We will now establish that δηαδηα is a UMP test, by showing that conditions (11.22) and (11.23) hold. By construction,

supθ0[0,1] PF(δηα;θ0) = PF(δηα;1) = αsupθ0[0,1] PF(δηα;θ0) = PF(δηα;1) = α

and so (11.22) holds. Also, δηαδηα is an a-level N-P test between the simple hypotheses H0 : θ = 1 and H1 : θ = θ1, and being independent of θ1, it is an α-level N-P test between these hypotheses for all θ1 ∈ (1, ∞). Now, consider any test δ̃ ∈ Δ̃ that satisfies supθ∈[0, 1] PF(δ̃; θ) ≤ α. Then clearly it is also true that PF(δ̃; 1) ≤ α. This means that δ̃ is an α-level test for testing the simple hypotheses H0 : θ = 1 versus H1 : θ = θ1, and it cannot be more powerful than δηαδηα, i.e.,

PD(˜δ;θ1)PD(δηα;θ1)forallθ1(1,).PD(δ˜;θ1)PD(δηα;θ1)forallθ1(1,).

Therefore (11.23) holds and we have:

δUMP(y) = δηα(y) = {1ifyσQ  1(α) + 10ify<σQ  1(α) + 1.δUMP(y) = δηα(y) = {10ififyσQ  1(α) + 1y<σQ  1(α) + 1.

Again, while the test δUMP is independent of the θ1, the performance of the test in terms of the PD depends on θ1. In particular

PD(δUMP;θ1) = Pθ{YσQ  1(α) + 1} = Q(Q  1(α)  θ1  1σ).PD(δUMP;θ1) = Pθ{YσQ  1(α) + 1} = Q(Q  1(α)  θ1  1σ).

11.6.3    Generalized Likelihood Ratio (GLR) Detection

While it is always desirable to have a UMP solution to the composite hypothesis testing problem, such solutions rarely exist in practice, especially in situations where both hypotheses are composite. One approach to generating a good test when UMP solutions do not exist is through the use of a “GLR” defined by

TGLR(y) = supθ1Λ1pθ1(y)supθ0Λ0pθ0(y).TGLR(y) = supθ1Λ1pθ1(y)supθ0Λ0pθ0(y).

It is important to note that the maximization over θ0 and θ1 has to be performed for each realization of the observation y, and so this test statistic is considerably more complex that the LRT. Also the result of the maximization may not produce a PDF (or PMF) in the numerator and denominator. We can use the statistic TGLR(y) to produce a test, which is called the “generalized likelihood ratio test (GLRT)”:

˜δGLRT(y) = {1ifTGLRT(y)>η1w.pifTGLRT(y) = η.0ifTGLRT(y)<ηδ˜GLRT(y) = 11w.p.γ0ifififTGLRT(y)>ηTGLRT(y) = η.TGLRT(y)<η

The use of the GLRT can be justified via an asymptotic analysis with a sequence of independent and identically distributed (i.i.d.) observations under each hypothesis, where it can be shown to have certain optimality properties. The maximization in numerator and denominator in TGLR(y) can also be justified from the viewpoint of maximum likelihood parameter estimation [2].

Example 11.6.5. Detection of One-Sided Composite Signal in Cauchy Noise (continued). This problem was introduced in Example 11.6.3. The conditional PDF is given by

pθ(y) = 1π[1 + (y  θ)2]pθ(y) = 1π[1 + (y  θ)2]

and we are testing H0 : θ = 0 against the one-sided composite hypothesis H1 : θ > 0. As we saw in Example 11.6.3, there is no UMP solution to this problem. The GLR statistic is given by

TGLR(y) = supθ>0pθ(y)p0(y)TGLR(y) = supθ>0pθ(y)p0(y)

with

supθ>0pθ(y) = supθ>01π[1 + (y  θ)2] = {1πify01π(1 + y2)ify<0.supθ>0pθ(y) = supθ>01π[1 + (y  θ)2] = {1π1π(1 + y2)ifify0y<0.

Thus

TGLR(y) = {1 + y2ify01ify<0.TGLR(y) = {1 + y21ifify0y<0.

To find an α-level test we need to evaluate P0{TGLR(Y) ≥ η}. Clearly

P0{TGLR(y)η} = 1for0η<1.P0{TGLR(y)η} = 1for0η<1.

For η ≥ 1

P0{TGLR(Y)η} = η  11π11 + y2dy = 0.5  tan  1(η  1)π.P0{TGLR(Y)η} = η  11π11 + y2dy = 0.5  tan  1(η  1)π.

There is a point of discontinuity in P0{TGLR(Y) ≥ η} at η = 1 as the value drops from 1 to the left to 0.5 to the right. For α ∈ (0.5, 1], we would need to randomize to meet the PF constraint with equality. For α ∈ (0, 0.5], which would be more relevant in practice, the GLRT is a deterministic test:

δGLRT(y) = {1ifTGLR(y)ηα0ifTGLR(y)<ηαδGLRT(y) = {10ififTGLR(y)ηαTGLR(y)<ηα

where

ηα = [tan(π(0.5  α))]2 + 1.ηα = [tan(π(0.5  α))]2 + 1.

11.6.4    Locally Most Powerful (LMP) Detection

Another approach to finding good detectors in cases where UMP tests do not exist is via a local optimization approach, which works when only one of the hypotheses is composite. Consider the scenario where Y ~ Pθ, we are interested in testing H0 : θ = θ0 versus H1 : θ > θ0, and there is no UMP solution. Also, suppose that θ takes values close to θ0 under H1; this might occur in practice in the detection of weak signals with unknown amplitude in noise.

Fix θ > θ0 and let δ̃θ be an α-level N-P test between θ and θ0. Then assuming that PD(δ̃θ; θ) is differentiable with respect to θ, we can write the Taylor series approximation:

PD(˜δθ;θ) = PD(˜δθ0;θ0) + (θ  θ0)θPD(˜δθ;θ)|θ = θ0 + (θ  θ0)α + (θ  θ0)θPD(˜δθ;θ)|θ = θ0.PD(δ˜θ;θ) = PD(δ˜θ0;θ0) + (θ  θ0)θPD(δ˜θ;θ)|θ = θ0 + (θ  θ0)α + (θ  θ0)θPD(δ˜θ;θ)|θ = θ0.

The locally optimal criterion can described as:

MaximizeθPD(˜δθ;θ)|θ = θ0subjecttoPF(˜δ;θ0)αMaximizeθPD(δ˜θ;θ)|θ = θ0subjecttoPF(δ˜;θ0)α

(11.24)

the idea being that maximizing PD should be approximately the same as maximizing the slope of PD at θ = θ0 for values of θ close to θ0. Now

PD(˜δθ;θ) = yI{˜δ(y) = 1}pθ(y)μ(dy).PD(δ˜θ;θ) = yI{δ˜(y) = 1}pθ(y)μ(dy).

Assuming that pθ (y) is differentiable in θ

θPD(˜δθ;θ)|θ = θ0 = yI{˜δ(y) = 1}θpθ(y)|θ = θ0μ(dy).θPD(δ˜θ;θ)|θ = θ0 = yI{δ˜(y) = 1}θpθ(y)|θ = θ0μ(dy).

Therefore, the solution to the locally optimal detection problem of (11.24) can be seen as being equivalent to N-P testing between pθ0(y)pθ0(y) and

θpθ(y)|θ = θ0.θpθ(y)|θ = θ0.

Even though the latter quantity is not necessarily a PDF (or PMF), the steps that we followed in deriving the N-P solution in Section 11.4 can be repeated to show that the solution to (11.24) has the form:

˜δLMP(y) = {1ifTlo(y)>η1w.pγifTlo(y) = η0ifTlo(y)<ηδ˜LMP(y) = 11w.pγ0ifififTlo(y)>ηTlo(y) = ηTlo(y)<η

where

Tlo(y) = θpθ(y)|θ = θ0pθ0(y).Tlo(y) = θpθ(y)θ = θ0pθ0(y).

Example 11.6.6. Detection of One-Sided Composite Signal in Cauchy Noise (continued). This problem was introduced in Example 11.6.3, and we saw that there was no UMP solution. We studied the GLRT in Example 11.6.5, and now we examine the LMP solution.

pθ(y) = 1π[1 + (y  θ)2]θpθ(y)|θ = 0 = 2yπ(1 + y2)2pθ(y) = 1π[1 + (y  θ)2]θpθ(y)|θ = 0 = 2yπ(1 + y2)2

Thus

Tlo(y) = 2y1 + y2Tlo(y) = 2y1 + y2

and

˜δLMP(y) = {1ifTlo(y)η0ifTlo(y)<η.δ˜LMP(y) = {10ififTlo(y)ηTlo(y)<η.

Randomization is not needed since Tlo(y) does not have point masses under P0.

11.7    Binary Detection with Vector Observations

In the detection problems we have studied so far, we did not make any explicit assumptions about the observation space, although the examples were restricted to scalar observations. The theory that we have developed applies equally to scalar and vector observations. Nevertheless, it is useful to study the case of vector observations in more detail as such a study reveals aspects of detector structures that are useful in applications.

Consider the detection problem:

H0:Yp0(y)versusH1:Yp1(y)H0:Yp0(y)versusH1:Yp1(y)

where Y = [Y1 Y2Yn] and y = [y1 y2yn]. The optimum detector for this problem, no matter which criterion (Bayes, Neyman-Pearson, minimax) we choose, is of the form

˜δOPT(y) = {1iflogL(y)>η1w.piflogL(y) = η0iflogL(y)<ηδ˜OPT(y) = 11w.p.γ0ifififlogL(y)>ηlogL(y) = ηlogL(y)<η

(11.25)

where L(y) = p1(y)/p0(y) is the likelihood ratio, and taking the log of L(y) does not affect the structure of the test since log is a monotonic function. The threshold η and randomization parameter γ are chosen based on the criterion used for detection. Of course, in the Bayesian setting, η = log τ, with τ given in (11.11), and γ = 0.

11.7.1    Conditionally Independent Observations

Consider the special case where the observations are (conditionally) independent under each hypothesis. In this case

pj(y) = nΠk = 1pj,k(yk)pj(y) = Πk = 1npj,k(yk)

and the log likelihood ratio in (11.25) can be written as

log L(y) = nk = 1logLk(yk)log L(y) = k = 1nlogLk(yk)

where Lk(yk) = p1,k(yk)/p0,k(yk).

Example 11.7.1. Deterministic signals in i.i.d. noise. Here, the hypotheses are given by:

H0:Y = s0 + ZversusH1:Y = s1 + ZH0:Y = s0 + ZversusH1:Y = s1 + Z

where s0 and s1 are deterministic vectors (signals) and Z1, Z2, …, Zn are i.i.d. random variables with zero mean and density given by pZ. Hence, the log likelihood ratio in (11.25) can be written as:

logL (y) = nk = 1logpZ(yk  s1,k)pZ(yk  s0,k).logL (y) = k = 1nlogpZ(yk  s1,k)pZ(yk  s0,k).

A special case of this example is one where Z is a vector of i.i.d. NN(0, σ2) random variables, in which case (based on the more general result derived in the following section), we can show that the optimum detector structure is of the form:

δOPT(y) = {1if(s1  s0)yη0if(s1  s0)y<η.δOPT(y) = {10ifif(s1  s0)yη(s1  s0)y<η.

11.7.2    Deterministic Signals in Correlated Gaussian Noise

In general, the detection problem with vector observations that are conditionally dependent, given the hypothesis, does not admit any special structure beyond what is described in (11.25). However, in some special cases, we can simplify the expression for the log likelihood ratio to obtain some more insight into the detector structure. In this section, we consider the example of detecting deterministic signals in correlated Gaussian noise, for which the hypotheses are described by:

H0:Y = s0 + ZversusH1:Y = s1 + ZH0:Y = s0 + ZversusH1:Y = s1 + Z

with s0 and s1 being deterministic signals as in Example 11.7.1, and Z is a Gaussian vector with zero mean and covariance matrix Σ, denoted by Z ~ NN(0, Σ). In this case

pj(y) = 1(2π)n|Σ|exp{  12(y  sj)Σ  1(y  sj)}.pj(y) = 1(2π)n|Σ|exp{  12(y  sj)Σ  1(y  sj)}.

where ∣Σ∣ is the absolute value of the determinant of Σ. Therefore

log L(y)  logp1(y)p0(y) = (s1  s0)Σ  1(y  s1  s02).log L(y)  logp1(y)p0(y) = (s1  s0)Σ  1(y  s1  s02).

Since log L(y) does not have any point masses under either hypothesis, the optimum detector is deterministic and has the form:

δOPT(y) = {1ifT(y)η0ifT(y)<ηδOPT(y) = {10ififT(y)ηT(y)<η

where T(y) = (s1s0)Σ−1y and the η is chosen based on the detection criterion. In the special case of Bayesian detection,

η = logτ + 12(s1  s0)Σ  1(s1 + s0)η = logτ + 12(s1  s0)Σ  1(s1 + s0)

with τ given in (11.11).

If we define the pseudosignal by

˜sΣ  1(s1  s0)s˜Σ  1(s1  s0)

then the test statistic T(y) can be written as:

T(y) = ˜sy = nk = 1˜skyk.T(y) = s˜y = k = 1ns˜kyk.

We see that the optimum detector is a correlation detector or matched filter [2].

Note that T(y) is linear in Y and hence has a Gaussian PDF under both H0 and H1. In particular,

Ej[T(Y)] = ˜s˜sj˜μjEj[T(Y)] = s˜s˜jμ˜j

and

Varj[T(Y)] = Var(˜sZ) = ˜sΣ˜s = ˜μ1  ˜μ0d2Varj[T(Y)] = Var(s˜Z) = s˜Σs˜ = μ˜1  μ˜0d2

where d2 is called the Mahalanobis distance between the signals s1 and s0.

Based on the above characterization of T(y), we can conclude that the problem of deterministic signal detection in correlated Gaussian noise is equivalent to the following detection problem involving the scalar observation T(y):

H0:T(y)~N(˜μ0,d2)versusH1:T(y)~N(˜μ1,d2).H0:T(y)~N(μ˜0,d2)versusH1:T(y)~N(μ˜1,d2).

11.7.3    Gaussian Signals in Gaussian Noise

In this section we consider another important example involving dependent observations, that of detecting Gaussian signals in Gaussian noise. The hypotheses are described by:

H0:Y = S0 + ZversusH1:Y = S1 + ZH0:Y = S0 + ZversusH1:Y = S1 + Z

where S0, S1, and Z are jointly Gaussian random vectors. It is easy to see that this problem is equivalent to the following detection problem:

H0:Y~N(μ0,0)versusH1:Y~N(μ1,1)H0:Y~N(μ0,0)versusH1:Y~N(μ1,1)

(11.26)

for some vectors μ0, μ1, and covariance matrices Σ0 and Σ1. Note that

pj(y) = 1(2π)n|j|exp{  12(y  μj)  1j(y  μj)}pj(y) = 1(2π)njexp{  12(y  μj)  1j(y  μj)}

and therefore the log likelihood ratio is given by:

logL(y) = 12y(  10    10)y + (μ1  11  μ0  10)y + 12[log|0||1| + μ01  11μ1].logL(y) = 12y(  10    10)y + (μ1  11  μ0  10)y + 12[log01 + μ01  11μ1].

Thus, the optimum detector in general involves both a quadratic term as well as a linear term in y. If Σ0 = Σ1 and μ0μ1, then the quadratic term vanishes and we have the detector structure we saw earlier for the detection of deterministic signals in Gaussian noise. If μ0 = μ1 = 0 and Σ1 ≠ Σ0, then the linear term vanishes and we have a purely quadratic detector.

Example 11.7.2. Signaling over Rayleigh Fading Channel with Random Phase. The following detection problem arises in the context of wireless communication systems, when the carrier phase is not known at the receiver:

H0:Y = ZversusH1:Y = [AcosϕAsinϕ] + ZH0:Y = ZversusH1:Y = [AcosϕAsinϕ] + Z

(11.27)

where Z ~ NN(0, σ2I), A is the fading amplitude that is Rayleigh distributed, and ϕ is the random phase that is uniformly distributed on [0, 2π]. The PDF of A is given by:

pA(a) = av2exp[  a22v2]I{a0}pA(a) = av2exp[  a22v2]I{a0}

If we define the fading signal vector S to have components S1 = A cos ϕ and S2 = A sin ϕ, then it is not difficult to show that S1 and S2 are independent NN(0, ν2) random variables. Thus the hypothesis test of (11.27) reduces to:

H0:Y~N(0,σ2I)versusH1:Y~N(0,(σ2 + v2)I).H0:Y~N(0,σ2I)versusH1:Y~N(0,(σ2 + v2)I).

This is a special case of (11.26) with μ0 = μ1 = 0, and Σ0 = σ2I, Σ1 = (σ2 + ν2)I. Thus the log likelihood ratio has the form:

logL(y) = (constant)yy + (constant)logL(y) = (constant)yy + (constant)

from which we can conclude that the optimum detector is of the form:

δOPT(y) = {1ifyyη0ifyyη.δOPT(y) = {10ifyyηifyyη.

The test statistic YY=Y21+Y22YY=Y21+Y22 has an exponential distribution with mean 2σ2 under H0, and an exponential distribution with mean 2(σ2 + ν2) under H1. Thus, if we are interested in N-P detection, for an α-level test, we can set P0{YYηα] = α by setting

exp[ηα2σ2] = αηα =   2σ2logα.exp[ηα2σ2] = αηα =   2σ2logα.

The corresponding power of the test is given by:

PD(δOPT) = P1{YTYηα} = exp[  ηα2(σ2 + v2)] = ασ2σ2 + v2.PD(δOPT) = P1{YTYηα} = exp[  ηα2(σ2 + v2)] = ασ2σ2 + v2.

11.8    Summary and Further Reading

This chapter covered the fundamentals of detection theory, with an emphasis on binary detection problems. In Section 11.1, we provided a general statistical decision theory framework for detection problems. In Section 11.2, Section 11.4, we introduced the three basic formulations for the binary detection problem: Bayesian, minimax, and Neyman-Pearson. We saw that in all cases the optimum detection rule is a LRT with possible randomization. In Section 11.5, Section 11.6, we studied composite detection problems where the distributions of the observations are not completely specified. In particular, we saw that Bayesian composite detection can be reduced to an equivalent simple detection problem. The Neyman-Pearson version of the composite detection problem is more interesting, and we studied various approaches to this problem, including UMP detection, GLR detection, and LMP detection. Finally, we examined the detection problem with vector observations in more detail, and discussed optimum detector structures for both the cases where the observations are conditionally independent and dependent, under each hypothesis.

This chapter was inspired by the textbook on detection and estimation theory by Poor [2]. While we focused almost exclusively on binary detection problems, extension to M-ary detection is straightforward at least in the Bayesian setting (see Exercise 11.9.6). More details on M-ary detection can be found in the books by Van Trees [3], Levy [4] and Kay [5]. An alternative formulation to the detection problem with incompletely specified distributions is the robust formulation of Huber [6]. Other extensions of detection theory include sequential [7] and quickest change detection [8], where observations are taken sequentially in time and decisions about the hypothesis need to be made online. Asymptotic performance analysis and design of detection procedures for large number of observations using tools from large deviations theory has been an active area of research (see, e.g., [9]). Finally, distributed sensor networks have generated interesting new directions for research in detection theory [10].

Acknowledgments

The writing of this chapter was supported in part by the U.S. National Science Foundation, under grant CCF-0830169, through the University of Illinois at Urbana-Champaign. The author would also like to thank Taposh Banerjee for help with the figures.

11.9    Exercises

Exercise 11.9.1. Consider the binary statistical decision theory problem for which S=D={0,1}S=D={0,1}. Suppose the cost function is given by

C(i,j) = {0ifi = j1ifj = 0,i = 110ifj = 1,i = 0C(i,j) = 0110ifi = jifj = 0,i = 1ifj = 1,i = 0

The observation Y takes values in the set Γ = {a, b, c} and the conditional p.m.f.’s of Y are:

p0(a) = p0(b) = 0.5p1(a) = p1(b) = 0.25,p1(c) = 0.5p0(a) = p0(b) = 0.5p1(a) = p1(b) = 0.25,p1(c) = 0.5

1.  Is there a best decision rule based on conditional risks?

2.  Find Bayes (for equal priors) and minimax rules within the set of deterministic decision rules.

3.  Now consider the set of randomized decision rules. Find a Bayes rule (for equal priors). Also construct a randomized rule whose maximum risk is smaller than that of the minimax rule of part (b).

Exercise 11.9.2. For the binary hypothesis testing problem, with C0,0 < C1,0 and C1,1 < C0,1, show there is no “best” rule based on conditional risks, except in the trivial case case where p0(y) and p1(y) have disjoint supports.

Exercise 11.9.3. Let SS = {0, 1}, and DD = {0, 1, e}. This would correspond to binary communication with erasures. Now suppose

pj(y) = 12πσexp[  (y  (  1)j + 1)22σ2],j = 0,1,  <y<.pj(y) = 12πσexp[  (y  (  1)j + 1)22σ2],j = 0,1,  <y<.

That is, Y has distribution NN(−1, σ2) when the state is 0, and Y has distribution NN(1, σ2) when the state is 1. Assume a cost structure

Ci,j = {0ifi = 0,j = 0ori = 1,j = 11ifi = 1,j = 0ori = 0,j = 1cifi = eCi,j = 0ifi = 0,j = 0ori = 1,j = 11ifi = 1,j = 0ori = 0,j = 1cifi = e

Furthermore, assume that the two states are equally likely.

1.  First assume that c < 0.5. Show that the Bayes rule for this problem has the form:

δB(y) = {0y  te  t<y<t1ytδB(y) = 0y  te  t<y<t1yt

Also give an expression for t in terms of the parameters of the problem.

2.  Now find δB(y) when c ≥ 0.5.

Exercise 11.9.4. Consider the binary detection problem with

p1(y) = {1/4ify[0,4]00otherwisep1(y) = {1/400ify[0,4]otherwise

and

p0(y) = {(y + 3)/18ify[  3,3]0otherwisep0(y) = {(y + 3)/180ify[  3,3]otherwise

1.  Find a Bayes rule for uniform costs and equal priors and the corresponding minimum Bayes risk.

2.  Find a minimax rule for uniform costs, and the corresponding minimax risk.

Exercise 11.9.5. For Exercise 11.9.2 above, find the minimum Bayes risk function V0), and then find a minimax rule in the set of randomized decision rules using V0).

Exercise 11.9.6. In this chapter, we formulated and solved the general Bayesian binary detection problem. We may generalize this formulation to M-ary detection (M > 2) as follows:

•  SS = {0, …, M − 1}, with a priori probability of state j being πj.

•  DD = {0, …, M − 1}

•  C(i, j) = Cij ≥ 0, for i, j = 0, …, M − 1.

•  YY, the observation space being continuous/discrete with conditional density (PDF/PMF) pj(y), j = 0, …, M − 1.

•  δ ∈ Δ, δ partitions YY into M regions YY0, …, YYM−1, where δ(y) = i when yYYi.

Find δB(y) by specifying the Bayes decision regions YYi, i = 0, …, M − 1. Simplify as much as possible.

Exercise 11.9.7. Consider the 5-ary detection problem in which the hypotheses are given by

Hj:Y = (j  2) + Z,j = 0,1,2,3,4,Hj:Y = (j  2) + Z,j = 0,1,2,3,4,

where Z ~ NN(0, 1). Assume that the hypotheses are equally likely.

1.  Find the decision rule with minimum probability of error (i.e., Bayes rule with uniform costs).

2.  Also find the corresponding minimum Bayes risk.

Hint: Find the probability of correct decision making first.

Exercise 11.9.8. Consider the binary detection problem with

p0(y) = 12e  |y|andp1(y) = e  2|y|,yp0(y) = 12e  |y|andp1(y) = e  2|y|,yR

1.  Find the Bayes rule for equal priors and a cost structure of the form C00 = C11 = 0, C10 = 1, and C01 = 2.

2.  Find the Bayes risk for the Bayes rule of part (a). (Note that the costs are not uniform.)

3.  Find a Neyman-Pearson rule for α = 1/4.

4.  Find the probability of detection for the rule of part (c).

Exercise 11.9.9. Consider the detection problem for which Γ = {0, 1, 2, …} and the PMF’s of the observations under the two hypotheses are:

p0(y) = (1  β0)βy0,y = 0,1,2,p0(y) = (1  β0)βy0,y = 0,1,2,

and

p1(y) = (1  β0)βy1,y = 0,1,2,p1(y) = (1  β0)βy1,y = 0,1,2,

Assume that 0 < β0 < β1 < 1.

1.  Find the Bayes rule for uniform costs and equal priors.

2.  Find the Neyman-Pearson rule with false-alarm probability α ∈ (0, 1). Also find the corresponding probability of detection as a function of α.

Exercise 11.9.10. Consider a binary detection problem, where the goal is to minimize the following risk measure

ρ(˜δ) = [PF(˜δ)]2 + PM(˜δ)ρ(δ˜) = [PF(δ˜)]2 + PM(δ˜)

1.  Show that the optimal solution is a (possibly randomized) likelihood-ratio test.

2.  Find the optimal solution for the observation model

p0(y){1ify[0,1]0otherwisep0(y){10ify[0,1]otherwise

and

p1(y){2yify[0,1]0otherwisep1(y){2y0ify[0,1]otherwise

Exercise 11.9.11. Consider the detection problem where L(y) has no point masses under either hypothesis. Let δη denote the likelihood ratio test:

δη(y) = {1ifL(y)η0ifL(y)<η.δη(y) = {10ifL(y)ηifL(y)<η.

As discussed in Section 11.4.2, a plot of PD(δη) versus PF(δη) for various values of η is called the ROC. This plot is a concave function with the point (0, 0) corresponding to η = ∞, and the point (1, 1) corresponding to η = 0. Prove the following properties of ROC’s:

1.  PD(δη) ≥ PF(δη) for all η. (Hint: consider cases η ≤ 1 and η > 1 separately.)

2.  The slope of the ROC at a particular point is equal to the value of the threshold η required to acheive the PD and PF at that point, i.e.,

dPDdPF = η.dPDdPF = η.

(Hint: Use the fact that L(Y) has a density under each hypothesis.)

Exercise 11.9.12. Consider the following composite detection problem with Λ = ℝ:

H0:θ˜θversusH1:θ>˜θH0:θθ˜versusH1:θ>θ˜

where θ̃ is a fixed real number. Now suppose that for each fixed θ0θ̃ and each fixed θ1 > θ̃, we have

pθ1(y)pθ0(y) = gθ0,θ1(T(y))pθ1(y)pθ0(y) = gθ0,θ1(T(y))

where the function T does not depend on θ1 or θ0, and the function gθ0,θ1gθ0,θ1 is strictly increasing in its argument.

Show that for any level α, a UMP test between H0 and H1 exists.

Exercise 11.9.13. Consider the composite binary detection problem in which

pθ(y) = {θe  θyify00ify<0pθ(y) = {θe  θy0ify0ify<0

1.  For α ∈ (0, 1), show that a UMP test of level α exists for testing the hypotheses

H0:Λ0 = [1,2]versusH1:Λ1 = (2,).H0:Λ0 = [1,2]versusH1:Λ1 = (2,).

Find this UMP test as a function of α.

2.  Find the structure of the generalized likelihood ratio test.

Exercise 11.9.14. (UMP testing with Laplacian Observations) Consider the composite binary detection problem in which

pθ(y) = 12e  |y  θ|,y.pθ(y) = 12e  |y  θ|,yR.

and we are testing:

H0:θ = 0versusH1:θ>0H0:θ = 0versusH1:θ>0

1.  Does a UMP test exist? If so, find it for level α and derive its power PD. If not, find the generalized likelihood ratio test for level α.

2.  Find a locally most powerful α-level test and derive its power PD.

Exercise 11.9.15. Consider the detection problem:

H0:Y = [  a0] + ZversusH1:Y = [a0] + ZH0:Y = [  a0] + ZversusH1:Y = [a0] + Z

where Z ~ NN(0, Σ) with

 = [1ρρ1 + ρ2]. = [1ρρ1 + ρ2].

Assume that a > 0 and ρ ∈ (0, 1).

1.  For equal priors show that the minimum-probability-of-error detector is given by

δB(y) = {1ify1  by2τ0ify1  by2<τδB(y) = {10ify1  by2τify1  by2<τ

where b = ρ/(1 + ρ2) and τ = 0.

2.  Determine the minimum probability of error.

3.  Consider the test of part (a) in the limit as ρ → 0. Explain why the dependence on y2 goes away in this limit.

4.  Now suppose the observations Y ~ NN([a 0], Σ), with ρ = 1 but a being an unknown parameter, and we wish to test between the hypotheses:

H0:0<a<1versusH1:a>1.H0:0<a<1versusH1:a>1.

Show that a UMP test exists for this problem, and find the UMP test of level α ∈ (0, 1).

Exercise 11.9.16. Consider the detection problem with n-dimensional observations:

H0:Y = ZversusH1:Y = s + ZH0:Y = ZversusH1:Y = s + Z

where the components of Z are zero mean correlated random variables with

E[ZkZ] = σ2ρ|k  |,forall1k,nE[ZkZ] = σ2ρ|k  |,forall1k,n

where ∣ρ∣ < 1.

1.  Show that the N-P test for this problem has the form:

δη(y) = {1ifnk = 1bkxkη0ifnk = 1bkxk<η

where b1 = s1/σ, x1 = y1/σ, and

bk = sk  ρsk  1σ1  ρ2,xk = yk  ρyk  1σ1  ρ2,k = 2,,n.

Hint: Note that   1Z = A/(σ2(1  ρ2)), where A is a tridiagonal matrix with main diagonal (1 1 + ρ2 1 + ρ2 … 1 + ρ2 1) and superdiagonal and subdiagonal entries all being −ρ.

2.  Find the α-level N-P test, δηα.

3.  Find the ROC for the above detector, i.e., find PD(δηα) as a function of α.

Exercise 11.9.17. Consider the composite detection problem with twodimensional observations:

H0:Y = ZversusH1:Y = θs + Z

where Z1 and Z2 are independent N(0, 1) random variables, and s1 = 1 and s2 = −1.

The parameter θ is a deterministic but unknown parameter that takes one of two possible values +1 or −1.

1.  Is there a UMP test for this problem? If so, find it for level a. If not, explain why not.

2.  Show that an α-level GLRT for this problem is given by:

δGLRT(y) = {1if|y1  y2|ηα0othwerwise

with ηα = 2Q  1(α2).

3.  Give a clear argument to establish that the probability of detection for the GLRT of part (b) is independent of θ.

4.  Now find the probability of detection for the GLRT as a function of ηα.

References

[1]  Ferguson, T.S., Mathematical Statistics: A Decision Theoretic Approach. Academic Press, 1967.

[2]  Poor, H.V., An Introduction to Signal Detection and Estimation, second edition. Springer-Verlag, 1994.

[3]  Van Trees, H.L., Detection, Estimation and Modulation Theory, Part 1. Wiley, 1968.

[4]  Levy, B.C., Principles of Signal Detection and Parameter Estimation. Springer-Verlag, 2008.

[5]  Kay, S.M., Fundamentals of Statistical Signal Processing: Detection Theory. Prentice Hall, 1998.

[6]  Huber, P.J., Robust Statistics. Wiley, 1981.

[7]  Wald, A., Sequential Analysis. Wiley, 1947.

[8]  Poor, H.V. and Hadjiliadis, O., Quickest Detection. Cambridge University Press, 2009.

[9]  Dembo, A. and Zeitouni, O., Large Deviations Techniques and Applications, Second Edition. Springer-Verlag, 1998.

[10]  Varshney, P.K., Distributed Detection and Data Fusion. Springer-Verlag, 1997.

1As will be the convention in the rest of the chapter, we denote random variables by uppercase letters and their corresponding realizations by lowercase letters. In particular, a realization of Y is denoted by y.

2This condition typically holds for continuous observations when p0(y) and p1(y) are PDF’s with the same support, but not necessarily even in this case.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
3.140.198.43