11 Main Statements of Statistical Estimation Theory

11.1 Main Definitions and Problem Statement

In a general form, the problem of estimates of stochastic process parameters can be formulated in the following manner. Let the incoming realization x(t) or a set of realizations xi(t), i = 1,…, ν of the random process ξ(t) be observed within the limits of the fixed time interval [0, T]. In general, the multidimensional (one-dimensional or n-dimensional) probability density function (pdf) of the random process ξ(t) contains μ unknown parameters l = {l1, l2, …, l} to be estimated. We assume that the estimated multidimensional parameter vector l = {l1, l2, …, l} is the continuous function within the limits of some range of possible values . Based on observation and analysis of the incoming realization x(t) or realizations xi(t), i = 1,…, ν, we need to decide what magnitudes from the given domain of possible values possessing the parameters l = {l1, l2, …, lμ} are of interest for the user. Furthermore, only the realization x(t) of the random process ξ(t) is considered if the special conditions are not discussed. In other words, we need to define the estimate of the required multidimensional parameter l = {l1, l2, …, lμ} on processing the observed realization x(t) within the limits of the fixed time interval [0, T].

Estimate of the parameter l = {l1, l2,…, l} of the random process ξ(t) is a set of functions or a single function of the observed data x(t). The values of these functions for the fixed realization x(t) estimate, that is, defined by a given way, are the unknown parameters of stochastic process. Depending on the requirements of estimation process and estimates of parameters, various estimation procedures are possible. In doing so, each estimate is characterized by its own quality performance that, in majority of cases, indicates a measure of the estimate’s closeness to the true value of estimated random process parameter. The quality performance, in turn, is defined by a choice of estimation criterion. Because of this, before estimate definition, we need to define a criterion of estimation. Selection of estimate criterion depends on the end problem, for which this random process parameter estimate is used. Because these problems can differ substantially, it is impossible to define and use the integrated criterion and one common estimate for the given random process parameter. This circumstance makes it difficult to compare various estimates.

In many practical applications, the estimate criteria are selected based on assumption about the purpose of estimate. At the same time, a design of process to choose definite estimate criteria is of great interest because such approach allows us to understand more clearly the essence and characteristic property of the problem to estimate the random process parameter and, that is very important, gives us a possibility to define the problem of valid selection of the estimate criterion more fairly. Owing to the finite time of observation and presence of noise and interference concurrent to observation, specific errors in the course of estimate definition arise. These errors are defined both by the quality performance and by conditions under which an estimation process is carried out. Because of this, the problem of optimal estimation of the parameter l = {l1, l2,…,lμ} is to define a procedure that allows minimizing the errors in estimation of the parameter l = {l1, l2,…,lμ}. In general, the requirement of minimizing estimation error is not based on one particular aspect. However, if the criterion of estimation is given, the quality performance is measured using that criterion. Then the problem of obtaining the optimal estimation is reduced to a definition of solution procedure that minimizes or maximizes the quality performance. In doing so, the parameter estimate must be close, in a certain sense, to the true value of estimated parameter, and the optimal estimate must minimize this measure of closeness in accordance with the chosen criterion.

To simplify the written form and discussion in future, we assume, if it is not particularly fixed, that the unknown parameter of the random process ξ(t) is the only parameter l = {l1, l2,…, lμ}. Nevertheless, all conclusions made based on our analysis of estimation process of the parameter l = {l1, l2,…, lμ} of the random process ξ(t) are correct with respect to joint estimation process of several parameters of the same random process ξ(t). Thus, it is natural to obtain one function from the observed realization x(t) to estimate a single parameter l = {l1, l2,…, lμ} of the random process ξ(t). Evidently, the more knowledge we have about the characteristics of the analyzed random process ξ(t) and noise and interference in the received realization x(t), the more accurate will be our estimation of possible values of the parameters of the random process ξ(t), and thus more accurate will be the solution based on synthesis of the devices designed using the chosen criterion with the minimal errors in the estimation of random process parameters of interest.

More specifically, the estimated parameter is a random variable. Under this condition, the most complete data about the possible values of the parameter l = {l1, l2,…, lμ} of the random process ξ(t) are given by the a posteriori pdf ppost(l) = p{l|x(t)}, which is the conditional pdf if the given realization x(t) is received. The formula of a posteriori pdf can be obtained based on the theorem about conditional probabilities of two random variables l and X, where X{x1, x2, …, xn} is the multidimensional (n-dimensional) sample of the realization x(t) within the limits of the interval [0, T]. According to the theorem about the conditional probabilities [1]

p(l,X)=p(l)p(l|X)=p(X)p(l|X),(11.1)

we can write

ppost(l)=p(l|X)=p(l)p(X|l)p(X).(11.2)

In (11.1) and (11.2) p(l) = pprior (l) is the a priori pdf of the estimated parameter l; p(X) is the pdf of multidimensional sample X of the realization x(t). The pdf p(X) does not depend on the current value of the estimated parameter l and can be determined based on the condition of normalization of ppost(l):

p(X)=p(X|l)pprior(l)dl.(11.3)

Integration is carried out by the a priori region (interval) ℒ of all possible values of the estimated parameter l. Taking into consideration (11.3), we can rewrite (11.2) in the following form:

ppost(l)=pprior(l)p(X|l)p(X|l)pprior(l)dl.(11.4)

The conditional pdf of the observed data sample X, under the condition that the estimated parameter takes a value l, has the following form:

p(X|l)=p(x1,x1,,xn|l),(11.5)

and can be considered as the function of l and is called the likelihood function. For the fixed sample X, this function shows that one possible value of parameter l is more likely in comparison with other value.

The likelihood function plays a very important role in the course of solution of signal detection problems, especially in radar systems. However, in a set of applications, it is worthwhile to consider the likelihood ratio instead of the likelihood function:

Λ(l)=p(x1,x2,,xn|l)p(x1,x2,,xn|lfix),(11.6)

where p(x1, x2, …, xn|lfix) is the pdf of observed data sample at some fixed value of the estimated random process parameter lfix. As applied to analysis of continuous realization x(t) within the limits of the interval [0, T], we introduce the likelihood functional in the following form:

Λ^(l)=limnp(x1,x2,,xn|l)p(x1,x2,,xn|lfix),(11.7)

where the interval between samples is defined as

Δ=Tn.(11.8)

Using the introduced notation, the a posteriori pdf takes the following form:

ppost(l)=κpprior(l)Λ(l),(11.9)

where κ is the normalized coefficient independent of the current value of the estimated parameter l;

κ=1pprior(l)Λ(l)dl.(11.10)

We need to note that the a posteriori pdf ppost(l) of the estimated parameter l and the likelihood ratio Λ(l) are the random functions depending on the received realization x(t).

In theory, for statistical parameter estimation, two types of estimates are used:

  • Interval estimations based on the definition of confidence interval

  • Point estimations, that is, the estimate defined at the point

Employing the interval estimations, we need to indicate the interval, within the limits of which there exists the true value of unknown random process parameter with the probability that is not less than the value given before. This earlier-given probability is called the confidence factor, and the..indicated interval of possible values of the estimated random process parameter is called the confidence interval. The upper and lower bounds of the confidence interval, which are called the confidence limits, and the confidence interval are the functions to be considered for both digital signal processing (a dis-cretization) and analog signal processing (continuous function) of the received realization x (t). In the point estimation case, we assign one parameter value to the unknown parameter from the interval of possible parameter values; that is, some value is obtained based on the analysis of the received realization x (t) and we use this value as the true value of the evaluated parameters.

In addition to the procedure of analysis of the random process parameter based on the value of received realization x (t), there is a sequential estimation method. This method essentially involves the sequential statistical analysis that estimates the random process parameter [2, 3]. The basic idea of sequential estimation is to define the time of analysis of the received realization x(t), within the limits of which we are able to obtain the estimate of parameter with the earlier-given reliability. In the case of point estimate, the root-mean-square deviation of estimate or other convenience function characterizing a deviation of estimate from the true value of estimated random process parameter can be considered as the measure of reliability. From the viewpoint of interval sequential estimation, the estimate reliability can be defined using the length of the confidence interval with a given confidence coefficient.

11.2 Point Estimate and its Properties

To make the point estimation means that some number γ = γ[x(t)] from the interval ℒ of possible values of the estimated random process parameter l must correspond to each possible received realization x(t). This number γ = γ[x(t)] is called the point estimate. Owing to the random nature of the point estimate of random process parameter, it is characterized by the conditional pdf p(γ|l). This is a general and total characteristic of the point estimate. The shape of this pdf defines the quality of point estimate definition and, consequently, all properties of the point estimate. At the given estimation rule γ = γ[x (t)], the conditional pdf p(γ|l) can be obtained from the pdf of received realization x(t) based on the well-known transformations of pdf [4]. We need to note that a direct determination of the pdf p(γ| l) is very difficult for many application problems. Because of this, if there are reasons to suppose that this pdf is a unimodal function and is very close to symmetrical function, then the bias, dispersion, and variance of estimate that can be determined without direct definition of the p(γ|l) are widely used as characteristics of the estimate γ.

In accordance with definitions, the bias, dispersion, and variance of estimate are defined as follows:

b(γ|l)=(γl)=x[γ(X)l]p(X|l)dX;(11.11)

D(γ|l)=(γl)2=x[γ(X)l]2p(X|l)dX;(11.12)

Var(γ|l)=[γγ]2=x[γ(X)γ]2p(X|l)dX.(11.13)

Here and further 〈…〉 means averaging by realizations. The estimate obtained taking into consideration of a priori pdf is called the unconditional estimate. The unconditional estimates are obtained as a result of averaging (11.11) through (11.13) on possible values of the variable l with a priori pdf pprior (l); that is, the unconditional bias, dispersion, and variance of estimate are determined in the following form:

b(γ)=xb(γ|l)pprior(l)dl;(11.14)

D(γ)=xD(γ|l)pprior(l)dl;(11.15)

Var(γ|l)=xVar(γ|l)pprior(l)dl.(11.16)

Since the conditional and unconditional estimate characteristics have different notations, we will drop the term “conditional” when discussing a single type of characteristics.

The estimate of random process parameters, for which the conditional bias is equal to zero, is called the conditionally unbiased estimate; that is, in this case, the mathematical expectation of the estimate coincides with the true value of estimated parameter: 〈γ〉 = l. If the unconditional bias is equal to zero, then the estimate is unconditionally unbiased estimate; that is, 〈γ〉 = lprior, where lprior is the a priori mathematical expectation of the estimated parameter. Evidently, if the estimate is conditionally unbiased, then we can be sure that the estimate is unconditionally unbiased. Inverse proposition, generally speaking, is not correct. In practice, the conditional unbiasedness often plays a very important role. During simultaneous estimation of several random process parameters, for example, estimation of the vector parameter l = {l1, l2, …, lμ}, we need to know the statistical relationship between estimates in addition to introduced conditional and unconditional bias, dispersion, and variance of estimate. For this purpose, we can use the mutual correlation function of estimates.

If estimations of the random process parameters l1, l2, …, l are denoted by γ1, γ2,…, γμ, then the conditional mutual correlation function of estimations of the parameters li and lj is defined in the following form:

Rij(v|l)=[(vivi)(vjvj)].(11.17)

The correlation matrix is formed based on these elements Rij(v|l); moreover, the matrix diagonal elements are the conditional variances of estimations. By averaging the conditional mutual correlation function using possible a priori values of the estimated random process parameters, we obtain the unconditional mutual correlation function of estimations.

There are several approaches to define the properties of the point estimations. We consider the following requirements to properties of the point estimations in terminology of conditional characteristics:

  • It is natural to try to define such point estimate γ so that the conditional pdf p(γ|l) stays very close to the value l.

  • It is desirable that while increasing the observation interval, that is, T → ∞, the estimation would coincide with or approach stochastically the true value of estimated random process parameter. In this case, we can say that the estimate is the consistent estimate.

  • The estimate must be unbiased 〈γ〉 = l or, in extreme cases, asymptotically unbiased, that is, limTγ=l

  • The estimate must be the best by some criterion; for example, it must be characterized by minimal values of dispersion or variance at zero or constant bias.

  • The estimate must be statistically sufficient.

The statistics, that is, in the considered case, the function or functions of the observed data, is sufficient if all statements about the estimated random process parameter can be defined based on the considered statistical data without any additional observation of received realization data. Evidently, the a posteriori pdf is always a sufficient statistic. Conditions of estimation sufficiency can be formulated in terms of the likelihood function: The necessary and sufficient condition of such estimation means the possibility to present the likelihood function in the form of product between two functions [5, 6]:

p(X|l)=h[x(t)]g(γ|l).(11.18)

Here, h[x(t)] is the arbitrary function of the received realization x(t) independent of the current value of the estimated random process parameter l. Since the parameter l does not enter into the function h[x(t)], we cannot use this function to obtain any information about the parameter l. The factor g(γ|l) depends on the received realization x(t) over the estimation γ = γ[x (t)] only. For this reason, all information about the estimated random process parameter l must be contained into γ[x (t)].

11.3 Effective Estimations

One of the main requirements is to obtain an estimate with minimal variance or minimal dispersion. Accordingly, a statement of effective estimations was introduced in the mathematical statistics. As applied to the bias estimations of the random process parameter, the estimation lef is considered effective if the mathematical expectation of its squared deviation from the true value of the estimated random process parameter l does not exceed the mathematical expectation of quadratic deviation of any other estimation γ; in other words, the following condition

Def(l)=(lefl)2(γl)2.(11.19)

must be satisfied. Dispersion of the unbiased estimate coincides with its variance and, consequently, the effective unbiased estimate is defined as the estimation with the minimal variance.

Cramer–Rao lower bound [5] was defined for the conditional variance and dispersion of estimations that are the variance and dispersion of effective estimations under the condition that they exist for the given random process parameters. Thus, in particular, the biased estimate variance is defined as

Var(γ|l)>[1+db(γ|l)dl]2[ddllnΛ(l)]2.(11.20)

The variance of unbiased estimations and estimations with the constant bias is simplified and takes the following form:

Var(γ|l)>1[ddllnΛ(l)]2.(11.21)

We need to note that in the case of analog signal processing of all possible realizations x (t) the averaging is carried out using a multidimensional sample of the observed data X and the derivatives are taken at the point where the estimated random process parameter has a true value. Equality in (11.20) and (11.21). takes place only in the case of effective estimations and if the two conditions are satisfied. The first condition is the condition that estimation remains sufficient (11.18). The second condition is the following: The likelihood function or likelihood ratio logarithm derivative should satisfy the equality [5]

ddllnΛ(l)=q(l)(γγ),(11.22)

where the function q(l) does not depend on the estimate δ and sample of observed data but depends on the current value of the estimated random process parameter l. At the same time, the condition (11.22) exists if and only if the estimate is sufficient; that is, the condition (11.18) is satisfied and the condition of sufficiency can exist when (11.22) is not satisfied. Analogous limitations are applied to the effective unbiased estimations, at which point the sign of inequality in (11.21) becomes the sign of equality.

11.4 Loss Function and Average Risk

There are two ways to make a decision in the theory of statistical estimations: nonrandom and random. In the case of nonrandom decision making (estimation), the definite decision is made by each specific realization of the received data x(t); that is, there is a deterministic dependence between the received realization and the decision made. However, owing to the random nature of the observed data, the values of estimation are viewed as random variables. The probability of definite decision making is assigned in the case of random decision making using each specific realization x(t); that is, the relationship between the received realization and the decision made has a probabilistic character. Furthermore, we consider only the nonrandom decision-making rules.

Based on the random character of the observed realization, there are errors for any decisionmaking rules; that is, the decision made γ does not coincide with the true value of parameter l. Evidently, by applying various decision-making rules different errors will appear with the various probabilities. Since the nonzero probability of error exists forever, we need to characterize the quality of different estimations in one way or another. For this purpose, the loss function is introduced in the theory of decisions. This function defines a definite loss ℒ(γ, l) for each combination from the decision γ and parameter l. As a rule, the losses are selected as positive values and the correct decisions are assigned zero or negative losses. The physical sense of the loss function is described as follows. A definite nonnegative weight is assigned to each incorrect decision. In doing so, depending on targets, for which the estimate is defined, the most undesirable decisions are assigned the greatest weights. A choice of definite loss function is made depending on a specific problem of estimation of the random process parameter l. Unfortunately, there is no general decision-making rule to select the loss function. Each decision-making rule is selected based on a subjective principle. A definite arbitrariness in selecting losses leads to definite difficulties with applying the theory of statistical decisions. The following types of loss functions are widely used (see Figures 11.1 through 11.4):

  • Simple loss function (see Figure 11.1)

    (γ,l)=1δ(γl);(11.23)

    where δ(z) is the Dirac delta function;

  • Linear modulo loss function (see Figure 11.2)

    (γ,l)=|γl|;(11.24)

    Images

    FIGURE 11.1 Simple loss function.

    Images

    FIGURE 11.2 Linear modulo loss function.

    Images

    FIGURE 11.3 Quadratic loss function.

    Images

    FIGURE 11.4 Rectangle loss function.

  • Quadratic loss function (see Figure 11.3)

    (γ,l)=(γl)2;(11.25)

  • Rectangle loss function (see Figure 11.4)

    (γ,l)={0if|γl|<η,η>0,1if|γl|>η,η>0.(11.26)

In general, there may be a factor on the right side of loss functions given by (11.23) through (11.24). These functions are the symmetric functions of difference |γ − l|. In doing so, deviations of the parameter estimate with respect to the true value of estimated random process parameter are undesirable. In addition, there are some application problems when the observer ignores the sign of estimate. In this case, the loss function is not symmetric.

Based on the random nature of estimations y and random process parameter l, the losses are random for any decision-making rules and cannot be used to characterize an estimate quality. To characterize an estimate quality, we can apply the mathematical expectation of the loss function that takes into consideration all incorrect solutions and relative frequency of their appearance. Choice of the mathematical expectation to characterize a quality of estimate, not another statistical characteristic, is rational but a little bit arbitrary. The mathematical expectation (conditional or unconditional) of the loss function is called the risk (conditional or unconditional). Conditional risk is obtained by averaging the loss function over all possible values of multidimensional sample of the observed data that are characterized by the conditional pdf p(X|l)

(γ|l)=X(γ,l)p(X|l)dX.(11.27)

As we can see from (11.27), the preferable estimations are the estimations with minimal conditional risk. However, at various values of the estimated random process parameter l, the conditional risk will have different values. For this reason, the preferable decision-making rules can be various, too. Thus, if we know the a priori pdf of the estimated parameter, then it is worthwhile to define the best decision-making rule for the definition of estimation based on the condition of the unconditional average minimum risk, which can be written in the following form:

(γ)=Xp(X)[(γ,l)ppost(l)dl]dX,(11.28)

where p(X) is the pdf of the observed data sample.

Estimations obtained by the criterion of minimum conditional and unconditional average risk are called the conditional and unconditional Bayes estimates. The unconditional Bayes estimate is often called the Bayes estimate. Furthermore, we will understand that for the Bayes estimate γm of the parameter l the estimate ensuring the minimum average risk at the given loss function is ℒ(γ, l). The minimal value of average risk corresponding to the Bayes estimate is called the Bayes risk

m=(γ,l)fpost(l)dl.(11.29)

Here, the averaging is carried out by samples (digital signal processing) of the observed data X or by realizations x(t) (analog signal processing). The average risk can be determined for any given decision-making rule and because of the definition of Bayes estimate the following condition is satisfied forever:

m(γ).(11.30)

Computing the average risk by (11.28) for different estimations and comparing each of these risks with Bayes risk, we can evaluate and conclude how one estimate can be better compared to another estimate and that the best estimate is the one close to the optimal Bayes estimate. Since the conditional or unconditional risk has a varying physical sense depending on the shape and physical interpretation of the loss function ££(y, l), the sense of criterion of optimality depends on the shape of the loss function, too.

The pdf p(X) is the nonnegative function. For this reason, a minimization of (11.28) by y at the fixed sample of observed data is reduced to minimization of the function

post(γ)=(γ,l)ppost(l)dl,(11.31)

called the a posteriori risk. If the a posteriori risk ℛpost(γ) is differentiated by γ, then the Bayes estimate γm can be defined as a solution of the following equation:

[dpost(γ)dγ]γm=0.(11.32)

To ensure a global minimum (minimum minimorum) of the a posteriori risk ℛpost(γ), we need to use the root square.

The criterion of minimum average risk is based on the knowledge of the total a priori information about the estimated parameter, which gives the answer to the question about how we can use all a priori information to obtain the best estimate. However, the absence of the total a priori information about the estimated parameter that takes place in the majority of applications leads us to definite problems (a priori problems) with applying the methods of the theory of statistical estimations. Several approaches are found to solve the problems of definition of the optimal estimations at the unknown a priori distribution of estimated parameters. One of them is based on definition of the Bayes estimations that are invariant with respect to the sufficiently wide class of a priori distributions. In other problems, we are limited by a choice of estimation based on the minimization of the conditional risk or we make some assumptions with respect to a priori distribution of estimated parameters. The least favorable distribution, with the Bayes risk at the maximum, is considered as the a priori distribution of estimated parameters; the obtained estimation of the random process parameter is called the minimax estimate. The minimax estimate defines the upper bound of the Bayes risk that is called the minimax risk. In spite of the fact that minimax estimate can generate heavy losses compared to other possible estimate, it can be useful if losses under the most unfavorable a priori conditions are avoided.

In accordance with the definition, the minimax estimate can be defined in the following way. In the case of a priori distribution of estimated parameter in accordance with the given loss function, we need to define the Bayes estimate γm = γm[x(t)]. After that we select such a priori distribution of estimated parameter, at which the minimal value of the average risk (Bayes risk) reaches the maximum. The Bayes estimate will be minimax at such a priori distribution. We need to note that a rigorous definition of the least favorable a priori distribution of estimated parameter is related to great mathematical problems. However, in the majority of applications, including the problems of random parameter estimations, the least favorable pdf is the uniform distribution within the limits of the given interval.

11.5 Bayesian Estimates for Various Loss Functions

Now, we discuss the properties of Bayes estimations for some loss functions mentioned previously.

11.5.1 SIMPLE LOSS FUNCTION

Substituting the simple loss function given by (11.23) into (11.31) and using the filtering property of the delta function

φ(z)δ(zz0)dz=φ(z0).(11.33)

we obtain

post(γ)=1ppost(γ).(11.34)

The a posteriori risk ℛpost(γ) and, consequently, the average risk ℛ(γ) are minimum if the a posteriori pdf ppost(γ) is the maximum for the given estimate. In other words, the a posteriori pdf ppost(γ) is maximum, if there is a single maximum, or maximum maximorum, if there are several local maxima. This means that the probable value γm, under which the following condition

ppost(γm)fpost(l),  γm,  l(11.35)

is satisfied, can be considered as the Bayes estimate. If the a posteriori pdf ppost(γ) is differentiable with respect to the parameter l, then the estimate γm can be defined based on the following equation:

[dppost(l)dl]γm=0 at [d2ppost(l)dl2]γm<0,(11.36)

and we take into consideration only the solution of this equation satisfying (11.35). Substituting the simple loss function in (11.29), we obtain the Bayes risk

m=1maxppost(l)=1ppost(γm).(11.37)

The second term on the right-hand side of (11.37) is the average probability of correct decision making that is accurate within the constant factor. Consequently, in the case of the simple loss function, the probability of correct decision-making is maximum at the Bayes estimate. It is evident that, in the case of the simple loss function, the probability of incorrect decision-making is minimal when applying the Bayes estimate to random process parameter. At the same time, all errors have the same weight that means that all errors are undesirable independent of their values. In the case of simple loss function, the Bayes estimate is well known in literature as the estimate by maximum maximorum of the a posteriori pdf.

If the a priori pdf ppr(l) is constant within the limits of interval of possible values of the estimated random process parameter, then, according to (11.9), the a posteriori pdf is matched accurately within the constant factor with the likelihood ratio Λ(l). The estimate by maximum of the a posteriori pdf γm transforms into the estimate of the maximum likelihood lm. The maximum likelihood estimate lm is defined as a maximum maximorum position of the likelihood ratio. As a rule, the maximum likelihood ratio estimations are applied in the following practical cases:

  • The a priori pdf of estimated random process parameter is unknown.

  • It is complicated to obtain the a posteriori pdf in comparison with the likelihood ratio or function.

The method of maximum likelihood has the following advantages compared to other methods.

  1. In practice, the maximum likelihood estimation is very close to the estimation by maximum a posteriori pdf for a wide class of a priori pdfs of the estimated random process parameter. In other words, the maximum likelihood estimation is the Bayes estimate of the simple loss function. Indeed, for highly accurate measurement of random process parameters, the a priori pdf about the estimate can be considered very often as the constant value and the a posteriori pdf about the estimate is matched by shape with the likelihood function or likelihood ratio. This property is very important when the a priori pdf of estimated random process parameter is unknown and a definition of the least favorable a priori pdf is associated with mathematical difficulties.

  2. Estimation of random process parameter by the maximum likelihood is independent of the mutual one-valued no inertial transformation by estimated parameter of the signal at the receiver output in the form F[A(t)], since the maximum likelihood point remains invariant during these transformations. This property is very important for practical realizations of optimal measurers.

  3. Analytical definition of random process parameter estimate quality using the maximum likelihood ratio requires less mathematical problems and difficulties compared to other methods of estimation. As a rule, the other methods of estimation require a mathematical or physical simulation to define the characteristics of estimate quality that, in turn, complicates the design and construction of the corresponding random process parameter measurers.

  4. It is defined in the mathematical statistics [5] that if there is an effective estimate, then the maximum likelihood estimate is effective too.

  5. If the random process observation interval tends to approach infinity, the maximum likelihood estimate is asymptotically effective and unbiased and, in this case, it is the limiting form of Bayes estimations for a wide class of a priori pdfs and loss functions.

These and other particular advantages of the maximum likelihood estimate method are the reasons for its wide application. We need to note that the estimation of random process parameters by the maximum likelihood method has a major disadvantage if there is noise, because the noise will be analyzed jointly with the considered random signal parameters. In this case, the maximum maximorum caused by the presence of noise can appear.

11.5.2 LINEAR MODULE LOSS FUNCTION

According to (11.31), the a posteriori risk for the loss function given by (11.24) can be determined based on the following equation:

post(γ)=|γl|ppost(l)dl.(11.38)

We can divide the integral on two integrals with the following limits of integration −∞ < l ≤ γ and γ < l ≤ ∞ . Based on the definition of the extremum of ℛpost(γ), we obtain

[dpost(γ)dγ]γm=γmppost(l)dlγmppost(l)dl=0.(11.39)

It follows that the Bayes estimate is the median of the a posteriori pdf.

11.5.3 QUADRATIC LOSS FUNCTION

According to (11.31), in the case of the quadratic loss function, we can write

post(γ)=(γl)2ppost(l)dl.(11.40)

From the condition of extremum of the function ℛpost(γ), we obtain the following estimate:

γm=lppost(l)dl=lpost.(11.41)

Thus, the average of the a posteriori pdf of lpost, that is, the main point of the a posteriori pdf, is considered as the estimate of the random process parameter l at the quadratic loss function. The value of ℛpost(γ) characterizes the minimum dispersion of random process parameter estimate. Since ℛpost(γ) depends on a specific form of the received realization x(t), the conditional dispersion is a random variable. In the case of the quadratic loss function, the Bayes risk coincides with the unconditional dispersion of estimate given by (11.15):

m=Var(γm)=Var(lpost)=X(llpost)2ppost(l)p(X)dldX.(11.42)

The estimate (11.41) is defined based on average risk minimum condition. For this reason, we can state that the Bayes estimate for the quadratic loss function makes minimum the unconditional dispersion of estimate. In other words, in the case of quadratic loss function, the Bayes estimate ensures the minimum estimate variance value with respect to the true value of estimated random process parameter among all possible estimations. We need to note one more property of the Bayes estimate when the loss function is quadratic. Substituting the a posteriori pdf (11.2) into (11.41) and averaging by the sample of observed data X, we can write

lpost=Xlpprior(l)p(X|l)dldX.(11.43)

Changing the order of integration and taking into consideration the condition of normalization, we obtain

lpost=lpprior(l)dl=lprior,(11.44)

that is, in the case of the quadratic loss function, the Bayes estimate is unconditional and unbiased, forever.

The quadratic loss function is proportional to the deviation square of the estimate from the true value of estimated random process parameter; that is, the weight is assigned to deviations and this weight increases as the square function of deviation values. These losses take place very often in various applications of mathematical statistics and theory of estimations. However, although the quadratic loss function has a set of advantages, that is, it is very convenient from the mathematical viewpoint and takes into consideration efficiently the highest value of big deviations compared to small ones, it is seldom used for the solution of optimal estimation problems compared to the simplest loss function. This is associated with the fact that the device that generates the estimate given by (11.41) has very complex construction and functioning. In the quadratic loss function case, definite difficulties can be met during computation of variance of the Bayes estimations.

11.5.4 RECTANGLE LOSS FUNCTION

In the case of rectangle loss function (11.26), all estimate deviations from the true value of estimated random process parameter, which are less than the given modulo value η, are not equally dangerous and do not affect the quality of random process parameter estimation. Deviations, the instantaneous values of which exceed the modulo value η, are not desirable and are assigned the same weight. Substituting (11.26) into (11.31) and dividing the limits of integration on three intervals, namely, −∞ < l < γ − η, γ − η < l< γ + η, and γ + η < l < ∞, we obtain

post(γ)=1γηγ+ηppost(l)dl.(11.45)

As we can see from (11.45), in the case of rectangular loss function, the value γ = γm is the Bayes estimate, for which the probability to satisfy the condition |γ − l| < η is maximum. Based on the condition of the a posteriori risk ℛpost(γ) extremum, we obtain the equation for estimation

ppost=(γmη)=ppost(γm+η).(11.46)

This equation shows that we need to use that value of the parameter γm as the Bayes estimate, under which the values of a posteriori pdf corresponding to random process parameters deviating from the estimate on the value η both at right and left sides are the same.

If the a posteriori pdf ppost(l) is differentiable, then, at small values η, there can be a limitation by the first three terms of Taylor series expansion with respect to the point γm of the a posteriori pdfs given by (11.46). As a result, we obtain the equation coinciding with (11.36). Consequently, at small insensibility areas η, the Bayes estimations for the simplest and rectangular loss functions coincide. The Bayes risk given by (11.29) for the rectangular loss function takes the following form:

m=1X[γmηγm+ηppost(l)p(X)dl]dX.(11.47)

The random variable given by (11.45) is the a posteriori probability of the event that the true value of estimated random process parameter of the given realization is not included within the limits of the interval −η < γm < η. In doing so, in the case of the Bayes estimate γm, this probability is less compared to that of other types of estimations. The Bayes risk given by (11.47) defines the minimum average probability of incorrect decision making by the given criterion.

11.6 Summary and Discussion

In many practical applications, the estimate criteria are selected based on assumptions about the purpose of estimate. At the same time, during the process design, choosing a definite estimate criteria is of great interest because such approach allows us to understand more clearly the essence and characteristic property of the problem to estimate the random process parameter, which is very important, giving us a possibility to define the problem of valid selection of the estimate criterion more fairly. Owing to the finite time of observation and presence of noise and interference concurrent to observation, specific errors in the course of estimate definition arise. These errors are defined both by the quality performance and by conditions under which an estimation process is carried out. In general, the requirement of minimization of estimation error by value does not depend on just one assumption. However, if the criterion of estimation is given, the quality performance is measured based on this criterion. Then the problem of obtaining the optimal estimation is reduced to a definition of solution procedure that minimizes or maximizes the quality performance. In doing so, the parameter estimate must be close, in a certain sense, to the true value of estimated parameter and the optimal estimate must minimize this measure of closeness in accordance with the chosen criterion.

In theory, for statistical parameter estimation, two types of estimates are used: the interval estimations based on the definition of confidence interval, and the point estimation, that is, the estimate defined at the point. Employing the interval estimations, we need to indicate the interval, within the limits of which there is the true value of unknown random process parameter with the probability that is not less than the predetermined value. This predetermined probability is called the confidence factor and the indicated interval of possible values of estimated random process parameter is called the confidence interval. The upper and lower bounds of the confidence interval, which are called the confidence limits, and the confidence interval are the functions to be considered during digital signal processing (a discretization) or during analog signal processing (continuous function). of the received realization x (t). In the case of point estimation, we assign one parameter value to the unknown parameter from the interval of possible parameter values; that is, some value is obtained based on the analysis of the received realization x (t) and we use this value as the true value of the evaluated parameters.

In addition to the procedure of analysis of the estimation random process parameter based on an analysis of the received realization x (t), a sequential estimation method is used. Essentially, this method is used for the sequential statistical analysis to estimate the random process parameter. The basic idea of sequential estimation is to define the time of analysis of the received realization x (t), within the limits of which we are able to obtain the estimate of parameter with the preset criteria for reliability. In the case of the point estimate, the root-mean-square deviation of estimate or other convenience function characterizing a deviation of estimate from the true value of estimated random process parameter can be considered as the measure of reliability. From a viewpoint of interval sequential estimation, the estimate reliability can be defined using a length of the confidence interval with the given confidence coefficient.

To make the point estimation means that some number from the interval of possible values of the estimated random process parameter must correspond to each possible received realization. This number is called the point estimate. Owing to the random character of the point estimate of random process parameter, it is characterized by the conditional pdf. This is a general and total characteristic of the point estimate. A shape of this pdf defines a quality of point estimate definition and, consequently, all properties of the point estimate.

There are several approaches to define the properties of the point estimations: (a) It is natural to try to define such point estimate that the conditional pdf would be grouped very close to the estimate value; (b) it is desirable that while increasing the observation interval, the estimation would coincide with or approach stochastically the true value of the estimated random process parameter (in this case, we can say that the estimate is the consistent estimate); (c) the estimate must be unbiased or, in an extreme case, asymptotically unbiased; (d) the estimate must be the best by some criterion; for example, it must be characterized by minimal values of dispersion or variance at zero or constant bias; and (e) the estimate must be a statistically sufficient measure.

Based on the random character of the observed realization, we can expect errors in any decision-making rules; that is, such a decision does not coincide with the true value of a parameter By applying various decision-making rules, different errors will appear with different levels of probability. Since the nonzero probability of error exists forever, we need to characterize the quality of different estimations in one way or another. For this purpose, the loss function is introduced in the theory of decisions. This function defines a definite loss for each combination from the decision and parameter. The physical sense of the loss function is the following. A definite nonnegative weight is assigned to each incorrect decision. In doing so, depending on targets, for which the estimate is defined, the most undesirable decisions are assigned the greatest weights. A choice of definite loss function is made depending on a specific problem of estimation of the random process parameter. There is no general decision-making rule to select the loss function. Each decision-making rule is selected based on a subjective principle. Any arbitrariness with selecting losses leads to definite difficulties with applying the theory of statistical decisions. The following types of loss functions are widely used in practice: the simple loss function, the linear modulo loss function, the quadratic loss function, and the rectangle loss function.

References

1. Kay, S.M. 2006. Intuitive Probability and Random Processes Using MATLAB. New York: Springer + Business Media, Inc.

2. Govindarajulu, Z. 1987. The Sequential Statistical Analysis of Hypothesis Testing Point and Interval Estimation, and Decision Theory. (American Series in Mathematical and Management Science). New York: American Sciences Press, Inc.

3. Sieqmund, D. 2010. Sequential Analysis: Test and Confidence Intervals. (Springer Series in Statistics). New York: Springer + Business Media, Inc.

4. Kay, S.M. 1993. Fundamentals of Statistical Signal Processing: Estimation Theory. Upper Saddle River, NJ: Prentice Hall, Inc.

5. Cramer, H. 1946. Mathematical Methods of Statistics. Princeton, NJ: Princeton University Press.

6. Cramer, H. and M.R. Leadbetter. 2004. Stationary and Related Stochastic Processes: Sample Function Properties and Their Applications. Mineola, NY: Dover Publications.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
3.142.173.227