Lesson 9
Best Linear Unbiased Estimation

Summary

The main purpose of this lesson is to develop our second estimator. It is both unbiased and efficient by design and is a linear function of the measurements Z(k). It is called a best linear unbiased estimator (BLUE).

As in the derivation of the WLSE, we begin with our generic linear model; but now we make two assumptions about this model: (1) H(k) must be deterministic, and (2) V(k) must be zero mean with positive definite known covariance matrix R(k). The derivation of the BLUE is more complicated than the derivation of the WLSE because of the design constraints; however, its performance analysis is much easier because we build good performance into its design.

A very remarkable connection exists between the BLUE and WLSE: the BLUE of θ is the special case of the WLSE of θ when W(k) = R–1 (k). Consequently, all results obtained in Lessons 3, 4, and 5 for Image can be applied to Image by setting W(k) = R–1(k).

In recursive WLSs, matrix P(k) has no special meaning. In recursive BLUE, matrix P(k) turns out to be the covariance matrix for the error between θ and Image. Recall that in Lesson 3 we showed that weighted least-squares estimates may change in numerical value under changes in scale. BLUEs are invariant under changes in scale.

The fact that H(k) must be deterministic limits the applicability of BLUEs.

When you complete this lesson you will be able to (1) explain and demonstrate how to design an estimator that processes the measurements linearly and has built into it the desirable properties of unbiasedness and efficiency; (2) connect this linear estimator (i.e., the BLUE) with the WLSE; (3) explain why the BLUE is insensitive to scale change; and (4) derive and use recursive BLUE algorithms.

Introduction

Least-squares estimation, as described in Lessons 3, 4, and 5, is for the linear model

(9-1)

Image

where θ is a deterministic, but unknown vector of parameters, H(k) can be deterministic or random, and we do not know anything about V(k) ahead of time. By minimizing Image, where Image, we determined that Image is a linear transformation of Z(k), i.e., Image. After establishing the structure of Image, we studied its small and large-sample properties. Unfortunately, Image is not always unbiased or efficient. These properties were not built into Image during its design.

In this lesson we develop our second estimator. It will be both unbiased and efficient, by design. In addition, we want the estimator to be a linear function of the measurements Z(k). This estimator is called a best linear unbiased estimator (BLUE) or an unbiased minimum-variance estimator (UMVE). To keep notation relatively simple, we will use Image to denote the BLUE of θ.

As in least squares, we begin with the linear model in (9-1), where θ is deterministic. Now, however, H(k) must be deterministic and V(k) is assumed to be zero mean with positive definite known covariance matrix R(k). An example of such a covariance matrix occurs for white noise. In the case of scalar measurements, z(k), this means that scalar noise v(k) is white, i.e.,

(9-2)

Image

where δkj is the Kronecker δ (i.e., δkj = 0 for kj and δkj = 1 for k = j). Thus,

(9-3)

Image

In the case of vector measurements, z(k), this means that vector noise, v(k), is white, i.e.,

(9-4)

Image

Thus,

(9-5)

Image

i.e., R(k) is block diagonal.

Problem Statement and Objective Function

We begin by assuming the following linear structure for Image:

(9-6)

Image

where, for notational simplicity, we have omitted subscripting F(k) as FBLU(k). We shall design F(k) such that

a. Image is an unbiased estimator of θ, and

b. the error variance for each of the n parameters is minimized. In this way, Image will be unbiased and efficient, by design.

Recall, from Theorem 6-1, that unbiasedness constrains design matrix F(k), such that

(9-7)

Image

Our objective now is to choose the elements of F(k), subject to the constraint of (9-7), in such a way that the error variance for each of the n parameters is minimized.

In solving for FBLU(k), it will be convenient to partition matrix F(k) as

(9-8)

Image

Equation (9-7) can now be expressed in terms of the vector components of F(k). For our purposes, it is easier to work with the transpose of (9-7), H′F′ = I, which can be expressed as

(9-9)

Image

where ei, is the ith unit vector,

(9-10)

Image

in which the nonzero element occurs in the ith position. Equating respective elements on both sides of (9-9), we find that

(9-11)

Image

Our single unbiasedness constraint on matrix F(k) is now a set of n constraints on the rows of F(k).

Next, we express Image in terms of fi(i = 1, 2,…, N). We shall make use of (9-11), (9-1), and the following equivalent representation of (9-6):

(9-12)

Image

Proceeding, we find that

(9-13)

Image

Observe that the error variance for the ith parameter depends only on the ith row of design matrix F(k). We, therefore, establish the following objective function:

(9-14)

Image

where λi, is the ith vector of Lagrange multipliers (see the Supplementary Material at the end of this lesson for an explanation of Lagrange’s method), which is associated with the ith unbiasedness constraint. Observe that the first term in (9-14) is associated with mean-squared error, whereas the second term is associated with unbiasedness. Our objective now is to minimize Ji, with respect to fi and λi (i = 1, 2,…, n).

Why is it permissible to determine fi and λi, for i = 1, 2,…, n by solving n independent optimization problems? Ideally, we want to minimize all Ji,(fi, λi,)’s simultaneously. Consider the following objective function:

(9-15)

Image

Observe from (9-14) that

(9-16)

Image

Clearly, the minimum of J (f1, f2,…, fn, λ1, λ2,…, λn) with respect to all f1, f2,…, fn, λ1, λ2,…,λn is obtained by minimizing each Ji,(fi, λi) with respect to each fi, λi. This is because the Ji’s are uncoupled; i.e., J1 depends only on f1 and λ1, J2 depends only on f2 and λ2,…, and Jn depends only on fn and λn.

Derivation of Estimator

A necessary condition for minimizing Ji,(fi, λi,) is ∂, Ji (fi, λi,)/∂fi = 0(i = 1, 2,…, n); hence,

(9-17)

Image

from which we determine fi as

(9-18)

Image

For (9-18) to be valid, R–1 must exist. Any noise V(k) whose covariance matrix R is positive definite qualifies. Of course, if V(k) is white, then R is diagonal (or block diagonal) and R–1 exists. This may also be true if V(k) is not white. A second necessary condition for minimizing Ji,(fi, λi) is ∂ Ji(fi, λi)/∂λi, = 0(i = 1, 2,…, n), which gives us the unbiasedness constraints

(9-19)

Image

To determine λi, substitute (9-18) into (9-19). Doing this, we find that

(9-20)

Image

whereupon

(9-21)

Image

(i = 1, 2,…, n). Matrix F(k) is reconstructed from fi(k), as follows:

(9-22)

Image

Hence,

(9-23)

Image

which means that

(9-24)

Image

Comparison of Image and Image

We are struck by the close similarity between Image and Image.

Theorem 9-1. The BLUE of θ is the special case of the WLSE of θ when

(9-25)

W(k) = R-1(k)

If W(k) is diagonal, then (9-25) requires V(k) to be white.

Proof. Compare the formulas for Image in (9-24) and Image in (3-10). If W(k) is a diagonal matrix (which is required in the Lesson 5 derivation of the recursive WLSE), then R(k) is a diagonal matrix only if V(k) is white.Image

Matrix R–1(k) weights the contributions of precise measurements heavily and de-emphasizes the contributions of imprecise measurements. The best linear unbiased estimation design technique has led to a weighting matrix that is quite sensible.

Corollary 9-1. All results obtained in Lessons 3, 4, and 5 for Image can be applied to Image by setting W(k) = R–1 (k).Image

We leave it to the reader to explore the full implications of this important corollary by reexamining the wide range of topics that was discussed in Lessons 3, 4, and 5.

Theorem 9-2 (Gauss-Markov Theorem). If H(k) is deterministic and R(k) = ImageI, then Image = Image.

Proof. Using (9-22) and the fact that R(k) = ImageI, we find that

Image

Comparison of Image and Image

Why is this a very important result? We have connected two seemingly different estimators, one of which, Image has the properties of unbiased and minimum variance by design; hence, in this case Image inherits these properties. Remember though that the derivation of Image required H(k) to be deterministic.

Some Properties of Image

To begin, we direct our attention at the covariance matrix of parameter estimation error Image.

Theorem 9-3. If V(k) is zero mean, then

(9-26)

Image

Proof. We apply Corollary 9-1 to cov[Image] [given in (8-18)] for the case when H(k) is deterministic, to see that

Image

Observe the great simplification of the expression for covImage, when Image. Note, also, that the error variance of Image given by the ith diagonal element of covImage.

Corollary 9-2. When W(k) = R–1(k), then matrix P(k), which appears in the recursive WLSE of θ equals covImage, i.e.,

(9-27)

Image

Proof. Recall Equation (4-13), that

(9-28)

Image

When W(k) = R–1(k), then

(9-29)

Image

Hence, P(k) = covImage because of (9-26). Image

Soon we will examine a recursive BLUE. Matrix P(k) will have to be calculated, just as it has to be calculated for the recursive WLSE. Every time P(k) is calculated in our recursive BLUE, we obtain a quantitative measure of how well we are estimating θ. Just look at the diagonal elements of P(k), k = 1, 2,…. The same statement cannot be made for the meaning of P(k) in the recursive WLSE. In the recursive WLSE, P(k) has no special meaning.

Next, we examine covImage in more detail.

Theorem 9-4. Image is a most efficient estimator of θ within the class of all unbiased estimators that are linearly related to the measurements Z(k).

In the econometric’s literature (e.g., Fomby et al., 1984) this theorem is known as Aitken’s Theorem (Aitken, 1935).

Proof (Mendel, 1973, pp. 155–156). According to Definition 6-3, we must show that

(9-30)

Image

is positive semidefinite. In (9-30), Imagea(k) is the error associated with an arbitrary linear unbiased estimate of θ. For convenience, we write Σ as

(9-31)

Image

To compute Σa we use the facts that

(9-32)

Image

and

(9-33)

Image

Thus,

(9-34)

Image

Because Image = F(k)Z(k) and F(k)H(k) = I,

(9-35)

Image

Substituting (9-34) and (9-35) into (9-31) and making repeated use of the unbiasedness constraints H′F′ = H′F′a = FaHa = I, we find that

(9-36)

Image

Making use of the structure of F(k), given in (9-23), we see that Σ can also be written as

(9-37)

Image

Some Properties of Image

To investigate the definiteness of Σ, consider the definiteness of a′Σa, where a is an arbitrary nonzero vector:

(9-38)

Image

Matrix F (i.e., FBLU) is unique; therefore, (Fa – F)′a is a nonzero vector, unless Fa – F and a are orthogonal, which is a possibility that cannot be excluded. Because matrix R is positive definite, a′Σa ≥ 0, which means that Σ is positive semidefinite.Image

These results serve as further confirmation that designing F(k) as we have done, by minimizing only the diagonal elements of covImage, is sound.

Theorem 9-4 proves that Image is most efficient within the class of linear estimators. The cov [Image] is given in (9-26) as [H′(k)R–1(k)H(k)]–1; hence, it must be true that the Fisher information matrix [see (6-45)] for this situation is J(θ) = H′(k)R–1 (k)H(k). We have been able to obtain the Fisher information matrix in this lesson without knowledge of the probability density function for Z(k), because our BLU estimator has been designed to be of minimum variance within the class of linear estimators.

Next, let us compare our just determined Fisher information matrix with the results in Example 6-4. In that example, J(θ) was obtained for the generic linear model in (9-1) under the assumptions that H(A:) and θ are deterministic and that generic noise V(k) is zero mean multivariate Gaussian with covariance matrix equal to R(k). The latter assumption then lets us determine a probability density function for Z(k), after which we were able to compute the Fisher information matrix, in (6-56), as J(θ) = H′(k)R–1(k)H(k).

How is it possible that we have arrived at exactly the same Fisher information matrix for these two different situations? The answer is simple. The Fisher information matrix in (6-56) is not dependent on the specific structure of an estimator. It merely provides a lower bound for cov[Image(k)]. We have shown, in this lesson, that a linear estimator achieves this lower bound. It is conceivable that a nonlinear estimator might also achieve this lower bound; but, because linear processing is computationally simpler than nonlinear processing, there is no need to search for such a nonlinear estimator.

Corollary 9-3. If R(k) = ImageI, then Image is a most efficient estimator of θ.

The proof of this result is a direct consequence of Theorems 9-2 and 9-4.Image

Note that we also proved the truth of Corollary 9-3 in Lesson 8 in Theorem 8-3. The proof of Theorem 8-3 is totally within the context of least squares, whereas the proof of Corollary 9-3 is totally within the context of BLUE, which, of course, is linked to least squares by the Gauss-Markov Theorem 9-2.

At the end of Lesson 3 we noted that Image may not be invariant under scale changes. We demonstrate next that Image is invariant to such changes.

Theorem 9-5. Image is invariant under changes of scale.

Proof (Mendel, 1973, pp. 156–157). Assume that observers A and B are observing a process; but observer A reads the measurements in one set of units and B in another. Let M be a symmetric matrix of scale factors relating A to B (e.g., 5280 ft/mile, 454 g/lb, etc.), and ZA(k) and ZB(k) denote the total measurement vectors of A and B, respectively. Then

(9-39)

Image

which means that

(9-40)

Image

(9-41)

Image

and

(9-42)

Image

Let Image, and Image denote the BLUEs associated with observers A and B, respectively; then,

(9-43)

Image

Here is an interesting interpretation of Theorem 9-5. Suppose we “whiten” the data by multiplying Z(k) by R–1/2(k). Doing this, our generic linear model can be written as

(9-44)

Image

where Z1(k) = R–1/2(k)Z(k), H1(k) = R–1/2(k)H(k), and V1(k) = R–1/2(k)V(k). Note that, when the data are whitened, cov V1(k) = R1 = I. If (9-44) is the starting point for BLUE, then

(9-45)

Image

which is precisely the correct formula for Image. Creating R–1/2(k)Z(k) properly normalizes the data. This normalization is not actually performed on the data in BLUE. It is performed by the BLUE algorithm automatically.

This suggests that we should choose the normalization matrix N in Lesson 3 as R–1/2(k); but this can only be done if we know the covariance matrix of V(k), something that was not assumed known for the method of least squares. When R(k) is known and H(k) is deterministic, then use Image. When R(k) is known and H(k) is not deterministic, use Image with W(k) = R–1 (k). We cannot use Image in the latter case because the BLUE was derived under the assumption that H(k) is deterministic.

Recursive BLUEs

Because of Corollary 9-1, we obtain recursive formulas for the BLUE of θ by setting 1/w(k + 1) = r(k + 1) in the recursive formulas for the WLSEs of θ, which are given in Lesson 5. In the case of a vector of measurements, we set (see Table 5-1) w–1 (k + 1) = R(k + 1).

Theorem 9-6 (Information form of recursive BLUE). A recursive structure for Image is

(9-46)

Image

where

(9-47)

Image

and

(9-48)

Image

These equations are initialized by Image and P–1 (n) [where P(k) is covImage, given in (9-26)] and are used for k = n, n + 1,…, N – 1. These equations can also be used for k = 0, 1,…, N – 1 as long as Image and P–1(0) are chosen using Equations (5-21) and (5-20) in Lesson 5, respectively, in which w(0) is replaced by r–1(0). Image

Theorem 9-7 (Covariance form of recursive BLUE). Another recursive structure/or Image is (9-46) in which

(9-49)

Image

and

(9-50)

Image

These equations are initialized by Image and P(n) and are used for k = n, n + 1,…, N – 1. They can also be used for k = 0, 1,…, N – 1 as long as Image and P(0) are chosen using Equations (5-21) and (5-20), respectively, in which w(0) is replaced by r–1(0).Image

Recall that, in best linear unbiased estimation, Image. Observe, in Theorem 9-7, that we compute P(k) recursively, and not P–1(k). This is why the results in Theorem 9-7 (and, subsequently, Theorem 5-2) are referred to as the covariance form of recursive BLUE.

See Problem 9-10 for a generalization of the recursive BLUE to a nondiagonal symmetric covariance matrix.

Computation

Because BLUE is a special case of WLSE, refer to Lessons 4 and 5 for discussions on computation. Of course, we must now set W(k) = R–1(k) in batch algorithms, or w(k) = 1/r(k) in recursive algorithms.

Supplementary Material: Lagrange’s Method

Lagrange’s method for handling optimization problems with constraints was devised by the great eighteenth century mathematician Joseph Louis Lagrange. In the case of optimizing F(x) subject to the scalar constraint G(x) = g, Lagrange’s method tells us to proceed as follows: form the function

(9-51)

Image

where λ. is a constant, as yet undetermined in value. Treat the elements of x as independent variables, and write down the conditions

(9-52)

Image

Solve these n equations along with the equation of the constraint

(9-53)

Image

to find the values of the n + 1 quantities x1, x2,…, xn, λ. More than one point (x1, x2,…, xn, λ) may be found in this way, but among the points so found will be the points of optimal values of F(x).

Equivalently, form the function

(9-54)

Image

where λ is treated as before. In addition to the n equations ∂J(x, λ)/∂xi = 0, write down the condition

(9-55)

Image

Solve these n + 1 equations for x and λ.

The parameter λ is called a Lagrange multiplier. Multiple constraints are handled by introducing one Lagrange multiplier for each constraint. So, for example, the vector of nc constraints G(x) = g is handled in (9-54) by replacing the term λ[G(x) – g] by the term λ′[G(x) – g].

Summary Questions

1. For a BLUE:

(a) H(k) and V(k) must be known

(b) H(k) may be random

(c) H(k) must be deterministic and V(k) may be zero mean

2. Image = Image if:

(a) R(k) = W(k)

(b) R(k) = wI

(c) R(k) = ImageI

3. By design, a BLUE is:

(a) consistent and unbiased

(b) efficient and unbiased

(c) a least-squares estimator

4. Image is invariant under scale changes because:

(a) it was designed to be so

(b) W(k) = R–1 (k)

(c) RB(k) = MRA(k)M

5. The name covariance form of the recursive BLUE (and, subsequently, recursive WLSE) arises, because:

(a) P(k) = covImage and we compute P(k) recursively

(b) P(k) = covImage and we compute P–1(k) recursively

(c) P(k) = covImage

6. The BLUE of θ is the special case of the WLSE of θ when W(k) equals:

(a) R(k)

(b) R–1(k)

(c) wI

7. Check off the two design constraints imposed on the BLUE:

(a) Image is an unbiased estimator of θ

(b) Image is a linear function of the measurements

(c) the error variance of each of the n parameters is minimized

8. Recursive formulas for the BLUE of θ are obtained from the recursive formulas in Lesson 5 for the WLSE of θ by setting:

(a) w(k + 1) = constant

(b) 1/w(k + 1) = r(k + 1)

(c) w(k + 1) = r(k + 1)

9. When H(k) is random, computations in the derivation of Image break down in:

(a) the unbiasedness constraint

(b) computing the Lagrange multipliers

(c) computing the error variance

10. When W(k) = R–1(k), then matrix P(k), which appears in the recursive WLSE of θ, equals:

(a) covImage

(b) a diagonal matrix

(c) W(k)

Problems

9-1. (Mendel, 1973, Exercise 3-2, p. 175). Assume H(k) is random and that Image.

(a) Show that unbiasedness of the estimate is attained when E{F(k)H(k)} = I.

(b) At what point in the derivation of Image do the computations break down because H(k) is random?

9-2. Here we examine the situation when V(k) is colored noise and how to use a model to compute R(k). Now our linear model is

Z(k + 1) = H(k + 1)θ + v(k + 1)

where v(k) is colored noise modeled as

v(k + 1) = Avv(k) + ξ(k)

We assume that deterministic matrix Av is known and that ξ(k) is zero-mean white noise with covariance Rξ,(k). Working with the measurement difference z*(k + 1) = z(k + 1) – Avz(k), write down the formula for Image in batch form. Be sure to define all concatenated quantities.

9-3. (Sorenson, 1980, Exercise 3-15, p. 130). Suppose Image and Image are unbiased estimators of θ with var Image and var Image. Let Image.

(a) Prove that Image is unbiased.

(b) Assume that Image and Image are statistically independent, and find the mean-squared error of Image.

(c) What choice of α minimizes the mean-squared error?

9-4. (Mendel, 1973, Exercise 3-12, pp. 176–177). A series of measurements z(k) is made, where z(k) = Hθ + v(k), H is an m × n constant matrix, E{v(k)} = 0, and cov[v(k)] = R is a constant matrix.

(a) Using the two formulations of the recursive BLUE, show that (Ho, 1963, pp. 152–154):

(i) P(k + 1)H′ = P(k)H′[HP(k)H′ + R]–1 R, and

(ii) HP(k) = R[HP(k – 1)H′ + R]–1HP(k – 1).

(b) Next, show that

(i) P(k)H′ = P(k – 2)H′[2HP(k – 2)H′ + R]–1R;

(ii) P(k)H′ = P(k – 3)H′[3HP(k – 3)H′ + R]–1R; and

(iii)P(k)H′ = P(0)H′[kHP(0)H′ + R]–1R.

(c) Finally, show that the asymptotic form (k → ∞) for the BLUE of θ is (Ho, 1963, pp. 152–154)

Image

This equation, with its 1/(k + 1) weighting function, represents a form of multidimensional stochastic approximation.

9-5. (Iraj Manocheri, Spring 1992) We are given the BLUE Image of θ that is based on measurements Z1(k), where Z1(k) = H1(k)θ + V1(k), E{V1(k)} = 0, and Image. Additional measurements Z2 (k) = H2(k)θ + V2 (k), E{V2 (k)} = 0, and Image are obtained, where V2(k) is independent of V1,(k).

(a) What are the formulas for Image and Image?

(b) Find the batch BLUE of θ given both Z1,(k) and Z2 (k).

(c) Find the recursive BLUE of θ given Z2(k) and Image.

9-6. (Bruce Rebhan, Spring 1992) One of your colleagues at Major Defense Corporation has come to you with an estimation problem.

(a) Two sets of sensors are used to provide an altitude measurement for the fighter radar you are developing. These sensors have the following covariance (R) matrices associated with their noises:

Image

Your colleague tells you that he plans to take a BLUE estimate of the altitude from each set of sensors. You tell him that you don’t think that’s a good idea. Why not?

(b) Your colleague returns to you, again seeking your sage advice regarding BLUE estimates. This time he has two different sets of sensors with noise covariance matrices

Image

He tells you that neither set of sensors is conducive to a BLUE estimate, since neither covariance matrix is “white.” Do you agree with him?

(c) Your colleague returns once more, this time with covariance matrices associated with another two sets of sensors:

Image

He tells you that since both of the noise covariance matrices are of the form R = Image he may take BLUEs from both set of sensors. You agree with this much. He then says that, furthermore, since both BLUEs will reduce to the LSE (by the Gauss-Markov theorem), it makes no difference which BLUE is used by the radar processor. You disagree with his conclusion, and tell him to use the BLUE from the sensors associated with R1. What is your reasoning?

9-7. (Richard S. Lee, Jr., Spring 1992) This computer-oriented problem is designed to explore the differences between the LS and BLU estimators. (a) Generate a set of measurement data, {z(k), k = 0, 1,…, 309}, where

z(k) = a + bk + v(k)

in which a = 1.0 and b = 0.5, v(k) is Gaussian zero-mean white noise with variance r(k), where r(k) = 1.0 for k = 0, 1,…,9, and r(k) is uniformly distributed between 1 and 100 for k = 10, 11,…, 309.

(b) Using the generated measurements [z(k), k = 0, 1,…, 9}, determine Image and Image using the batch LS estimator.

(c) For each value of k = 10, 11,…, 309, determine Image and Image using the recursive LS algorithm. Use the values of Image(9) and Image(9) determined in part (b) to initialize the recursive algorithm. Plot Image(k) and Image(k) versus k.

(d) Repeat part (b) using the BLUE. Compare your results to those obtained in part (b).

(e) Repeat part (c) using the BLUE. Compare your results to those obtained in part (c).

9-8. Suppose covariance matrix R is unknown and we decide to estimate it using least squares so that we can then use this estimate in the formula for the BLUE of θ; i.e., we estimate R as

Image

Show that Image is rank 1 and is therefore singular; consequently, Image cannot be computed. Note that if R can be suitably parameterized (e.g., see Problem 9-9) it may be possible to estimate its parameters. Note, also, that if we know that H(k) is deterministic and V(k) is multivariate Gaussian (see Lesson 12) we can determine a maximum-likelihood estimator for R.

9-9. Consider the generic linear model Z(k) = H(k)θ + V(k), where the elements of V(k) are not white but rather are “colored,” in the sense that they satisfy the following first-order difference equation (for more extensive discussions about colored noise, see Lesson 22):

v(k) = ρv(k – 1)+ n(k), k = 1, 2,…, N

where ρ is a parameter that may or may not be known ahead of time, |ρ| < 1, and n(k) is zero mean, independent, and identically distributed with known variance Image (Fomby et al, 1984, pp. 206–208).

(a) Show that

Image

(b) Show that Ω–1 is a tridiagonal matrix with diagonal elements equal to (1, (1 + ρ2), (1 + ρ2),…, (1 + ρ2), 1), superdiagonal elements all equal to –ρ, and subdiagonal elements all equal to –ρ.

(c) Assume that ρ is known. What is Image?

9-10. (Prof. Tom Hebert) In this lesson we derived the recursive BLUE under the assumption that R(k) is a diagonal matrix. In this problem we generalize the results to a nondiagonal symmetric covariance matrix (e.g., see Problem 9-9). This could be useful if the generic noise vector is colored where the coloring is known.

(a) Prove that:

Image

where

Image

and

P–1(k + 1) = P–1(k) + ak+1β(k + 1)β′(k + 1)

(b) Prepare a flow chart that shows the correct order of computations for this recursive algorithm.

(c) Demonstrate that the algorithm reduces to the usual recursive BLUE when R(k) is diagonal.

(d) What is the recursive WLSE for which W(k + 1) = R–1(k + 1), when W(k + 1) is nondiagonal but symmetric?

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
18.188.10.246