11
CONFIDENCE ESTIMATION

11.1 INTRODUCTION

In many problems of statistical inference the experimenter is interested in constructing a family of sets that contain the true (unknown) parameter value with a specified (high) probability. If X, for example, represents the length of life of a piece of equipment, the experimenter is interested in a lower bound images for the mean θ of X. Since images will be a function of the observations, one cannot ensure with probability 1 that images . All that one can do is to choose a number 1 – α that is close to 1 so that images for all θ. Problems of this type are called problems of confidence estimation. In this chapter we restrict ourselves mostly to the case where images and consider the problem of setting confidence limits for the parameter θ.

In Section 11.2 we introduce the basic ideas of confidence estimation. Section 11.3 deals with various methods of finding confidence intervals, and Section 11.4 deals with shortest-length confidence intervals. In Section 11.5 we study unbiased and equivariant confidence intervals.

11.2 SOME FUNDAMENTAL NOTIONS OF CONFIDENCE ESTIMATION

So far we have considered a random variable or some function of it as the basic observable quantity. Let X be an RV, and a, b, be two given positive real numbers. Then

images

and if we know the distribution of X and a, b, we can determine the probability images . Consider the interval images . This is an interval with end points that are functions of the RV X, and hence it takes the value (x,bx/a) when X takes the value x. In other words, I(X) assumes the value I(x) whenever X assumes the value x. Thus I(X) is a random quantity and is an example of a random interval. Note that I(X) includes the value b with a certain fixed probability. For example, if b = 1, images , and X is U(0,1), the interval (X, 2X) includes point 1 with probability images . We note that I(X) is a family of intervals with associated coverage probability images . It has (random) length images . In general the larger the length of the interval the larger the coverage probability. Let us formalize these notions.

In a wide variety of inference problems one is not interested in estimating the parameter or testing some hypothesis concerning it. Rather, one wishes to establish a lower or an upper bound, or both, for the real-valued parameter. For example, if X is the time to failure of a piece of equipment, one may be interested in a lower bound for the mean of X. If the RV X measures the toxicity of a drug, the concern is to find an upper bound for the mean. Similarly, if the RV X measures the nicotine content of a certain brand of cigarettes, one may be interested in determining an upper and a lower bound for the average nicotine content of these cigarettes.

In this chapter we are interested in the problem of confidence estimation, namely, that of finding a family of random sets S(x) for a parameter θ such that, for a given α, images (usually small),

(1)images

We restrict our attention mainly to the case where images .

Remark 1. Suppose X ~ Pθ and (2) holds. Then the smallest probability of true coverage, images is 1 – α. Then the probability of false (or incorrect) coverage is images for images . According to Definition 3 among the class of all lower confidence bounds satisfying (2), a UMA lower confidence bound has the smallest probability of false coverage.

Similar definitions are given for an upper confidence bound for θ and a UMA upper confidence bound.

Remark 2. We write S(X)∋ θ to indicate that X, and hence S(X), is random here and not θ so the probability distribution referred to is that of X.

Remark 3. When images is the realization the confidence interval (set) S(x) is a fixed subset of images . No probability is attached to S(x) itself since neither θ nor S(x) has a probability distribution. In fact either S(x) covers θ or it does not and we will never know which since θ is unknown. One can give a relative frequency interpretation. If (1 – α) -level confidence sets for θ were computed a large number of times, then a fraction (approximately) 1 – α of these would contain the true (but unknown) parameter value.

If σ is not known, we have from

images

that

images

and once again we can choose pairs of values (c1 , c2) using a t-distribution with n – 1 d.f. such that

images

In particular, if we take images , say, then

images

and images , is a (1 – α) -level confidence interval for μ. The length of this interval is images , which is no longer constant. Therefore we cannot choose n to get a fixed-width confidence interval of level 1 – α. Indeed, the length of this interval can be quite large if σ is large. Its expected length is

images

which can be made as small as we please by choosing n large enough.

Next suppose that both μ and σ2 are unknown and that we want a confidence set for (μ, σ2). We have from Boole’s inequality

images

so that the Cartesian product,

images

is a (1–α1α2) -level confidence set for (μ, σ2).

11.3 METHODS OF FINDING CONFIDENCE INTERVALS

We now consider some common methods of constructing confidence sets. The most common of these is the method of pivots.

In many problems, especially in location and scale problems, pivots are easily found. For example, in sampling from images is a pivot and so is images . In sampling from (1/σ)f(x/σ) , a scale family, images is a pivot and so is X(1)/σ, and in sampling from (1/σ)f((x–θ)/σ), a location-scale family, images is a pivot, and so is images .

If the DF Fθ of Xi is continuous, then images and, in case of random sampling, we can take

images

or,

images

as a pivot. Since images and images . It follows that images is a pivot.

The following result gives a simple sufficient condition for a pivot to yield a confidence interval for a real-valued parameter θ.

Remark 1. The condition that images be solvable will be satisfied if, for example, T is continuous and strictly increasing or decreasing as a function of θ in Θ.

Note that in the continuous case (that is, when the DF of T is continuous) we can find a confidence interval with equality on the right side of (1). In the discrete case, however, this is usually not possible.

Remark 2. Relation (1) is valid even when the assumption of monotonicity of T in the theorem is dropped. In that case inversion of the inequalities may yield a set of intervals (random set) S(X) in Θ instead of a confidence interval.

Remark 3. The argument used in Theorem 1 can be extended to cover the multiparameter case, and the method will determine a confidence set for all the parameters of a distribution.

We next consider the method of test inversion and explore the relationship between a test of hypothesis for a parameter θ and confidence interval for θ. Consider the following example.

Conversely, a family of α-level tests for the hypothesis images generates a family of confidence intervals for μ by simply taking, as the confidence interval for μ0, the set of those μ for which one cannot reject images .

Similarly, we can generate a family of α-level tests from a (1 – α) -level lower (or upper) confidence bound. Suppose that we start with the (1 – α) -level lower confidence bound images for μ. Then, by defining a test φ (Χ) that rejects images if and only if images , we get an α-level test for a hypothesis of the form images .

Example 1 is a special case of the duality principle proved in Theorem 2 below. In the following we restrict attention to the case in which the rejection (acceptance) region of the test is the indicator function of a (Borel-measurable) set, that is, we consider only nonrandomized tests (and confidence intervals). For notational convenience we write Η0(θ0) for the hypothesis images and Η1(θ0) for the alternative hypothesis, which may be one- or two-sided.

In particular, if X1, X2, … ,Xn is a sample from

images

then images and for testing images against images , the UMP acceptance region is of the form

images

where c(θ0) is the unique solution of

images

The UMA family of (1 – α) -level confidence sets is of the form

images

In the case images and images .

The third method we consider is based on Bayesian analysis where we take into account any prior knowledge that the experimenter has about θ. This is reflected in the specification of the prior distribution π(θ) on Θ. Under this setup the claims of probability of coverage are based not on the distribution of X but on the conditional distribution of θ given images , the posterior distribution of θ.

Let Θ be the parameter set, and let the observable RV X have PDF (PMF) fθ(x). Suppose that we consider θ as an RV with distribution π(θ) on Θ. Then fθ(x) can be considered as the conditional PDF (PMF) of X, given that the RV θ takes the value θ. Note that we are using the same symbol for the RV θ and the value that it assumes. We can determine the joint distribution of X and θ, the marginal distribution of X, and also the conditional distribution of θ, given images as usual. Thus the joint distribution is given by

(7)images

and the marginal distribution of X by

(8)images

The conditional distribution of θ, given that x is observed, is given by

(9)images

Given h(θ | x), it is easy to find functions l(x), u(x) such that

images

where

(10)images

depending on whether h is a PDF or a PMF.

One can similarly define one-sided Bayes intervals or (1 – α) -level lower and upper Bayes limits.

Remark 5. We note that, under the Bayesian set-up, we can speak of the probability that θ lies in the interval (l(x), u(x)) with probability 1–α because l and u are computed based on the posterior distribution of θ given x. In order to emphasize this distinction between Bayesian and classical analysis, some authors prefer the term credible sets for Bayesian confidence sets.

Finally, we consider some large sample methods of constructing confidence intervals. Suppose images . Then

images

where Z ~ N(0,1). Suppose further that there is a statistic S(X) such that images . Then, by Slutsky’s theorem

images

and we can obtain an (approximate) images -level confidence interval for θ by inverting the inequality

images

In particular, we can choose –c1 = c2 = zα/2 to give

images

as an approximate images -level confidence interval for μ.

Recall that if images is the MLE of θ and the conditions of Theorem 8.7.4 or 8.7.5 are satisfied (caution: See Remark 8.7.4), then

images

where

images

Then we can invert the statement

images

to give an approximate images -level confidence interval for θ.

Yet another possible procedure has universal applicability and hence can be used for large or small samples. Unfortunately, however, this procedure usually yields confidence intervals that are much too large in length. The method employs the well-known Chebychev inequality (see Section 3.4):

images

If images is an estimate of θ (not necessarily unbiased) with finite variance σ2(θ), then by Chebychev’s inequality

images

It follows that

images

is a images -level confidence interval for θ. Under some mild consistency conditions one can replace the normalizing constant images , which will be some function λ(θ) of θ, by images .

Note that the estimator images need not have a limiting normal law.

Actually the confidence interval obtained above can be improved somewhat. We note that

images

so that

images

Now

images

if and only if

images

This last inequality holds if and only if p lies between the two roots of the quadratic equation

images

The two roots are

images

and

images

It follows that

images

Note that when n is large

images

as one should expect in view of the fact that images with probability 1 and images estimates images . Alternatively, we could have used the CLT (or large-sample property of the MLE) to arrive at the same result but with ε replaced by zα/2.

In the examples given above we see that, for a given confidence interval images , a wide choice of confidence intervals is available. Clearly, the larger the interval, the better the chance of trapping a true parameter value. Thus the interval images , which ignores the data completely will include the real-valued parameter θ with confidence level 1. However, the larger the confidence interval, the less meaningful it is. Therefore, for a given confidence level 1–α, it is desirable to choose the shortest possible confidence interval. Since the length images , in general, is a random variable, one can show that a confidence interval of level 1–α with uniformly minimum length among all such intervals does not exist in most cases. The alternative, to minimize images , is also quite unsatisfactory. In the next section we consider the problem of finding shortest-length confidence interval based on some suitable statistic.

PROBLEMS 11.3

  1. A sample of size 25 from a normal population with variance 81 produced a mean of 81.2. Find a 0.95 level confidence interval for the mean μ.
  2. Let images be the mean of a random sample of size n from images . Find the smallest sample size n such that images is a 0.90 level confidence interval for μ.
  3. Let X1, X2, … ,Xm and Y1, Y2, … ,Yn be independent random samples from images and images , respectively. Find a confidence interval for images at confidence level 1–α when (a) σ is known and (b) σ is unknown.
  4. Two independent samples, each of size 7, from normal populations with common unknown variance σ2 produced sample means 4.8 and 5.4 and sample variances 8.38 and 7.62, respectively. Find a 0.95 level confidence interval for μ1μ2, the difference between the means of samples 1 and 2.
  5. In Problem 3 suppose that the first population has variance images and the second population has variance images , where both images , and images are known. Find a images -level confidence interval for μ1μ2. What happens if both images and images are unknown and unequal?
  6. In Problem 5 find a confidence interval for the ratio images , both when μ1, μ2 are known and when μ1, μ2 are unknown. What happens if either μ1 or μ2 is unknown but the other is known?
  7. Let X1, X2, … ,Xn be a sample from a G(1) distribution. Find a confidence interval for the parameter β with confidence level 1–α.
    1. Use the large-sample properties of the MLE to construct a (1–α) -level confidence interval for the parameter θ in each of the following cases: (i) X1. X2, … ,Xn is a sample from G(1,1/θ) and (ii) X1,X2, … ,Xn is a sample from Ρ(θ).
    2. In part (a) use Chebychev’s inequality to do the same.
  8. For a sample of size 1 from the population
    images

    find a images -level confidence interval for θ.

  9. Let X1,X2, … ,Xn be a sample from the uniform distribution on N points. Find an upper images -level confidence bound for N, based on max(X1, X2, … ,Xn).
  10. In Example 10 find the smallest n such that the length of the images -level confidence interval images , provided it is known that images , where a is a known constant.
  11. Let X and Y be independent RVs with PDFs images and images , respectively. Find a images -level confidence region for (λ, μ) of the form images .
  12. Let X1,X2, … ,Xn be a sample from images , where σ2 is known. Find a UMA images -level upper confidence bound for μ.
  13. Let X1,X2, … ,Xn be a sample from a Poisson distribution with unknown parameter λ. Assuming that λ is a value assumed by a G(α,ß) RV, find a Bayesian confidence interval for λ.
  14. Let X1,X2, … ,Xn be a sample from a geometric distribution with parameter θ. Assuming that θ has a priori PDF that is given by the density of a B(a,ß) RV, find a Bayesian confidence interval for θ.
  15. Let X1, X2, … ,Xn be a sample from images , and suppose that the a priori PDF for μ is U(– 1, 1). Find a Bayesian confidence interval for μ.

11.4 SHORTEST-LENGTH CONFIDENCE INTERVALS

We have already remarked that we can increase the confidence level by simply taking a larger-length confidence interval. Indeed, the worthless interval images , which simply says that θ is a point on the real line, has confidence level 1. In practice, one would like to set the level at a given fixed number images and, if possible, construct an interval as short in length as possible among all confidence intervals with the same level. Such an interval is desirable since it is more informative. We have already remarked that shortest-length confidence intervals do not always exist. In this section we will investigate the possibility of constructing shortest-length confidence intervals based on simple RVs. The discussion here is based on Guenther [37]. Theorem 11.3.1 is really the key to the following discussion.

Let X1,X2, … ,Xn be a sample from a PDF fθ(x), and images be a pivot for θ. Also, let images be chosen so that

(1)images

and suppose that (1) can be rewritten as

(2)images

For every Tθ, λ1 and λ2 can be chosen in many ways. We would like to choose λ1 and λ2 so that images is minimum. Such an interval is a images -level shortest-length confidence interval based on Tθ. It may be possible, however, to find another RV images that may yield an even shorter interval. Therefore we are not asserting that the procedure, if it succeeds, will lead to a images -level confidence interval that has shortest length among all intervals of this level. For Tθ we use the simplest RV that is a function of a sufficient statistic and θ.

Remark 1. An alternative to minimizing the length of the confidence interval is to minimize the expected length images . Unfortunately, this also is quite unsatisfactory since, in general, there does not exist a member of the class of all images -level confidence intervals that minimizes images for all θ. The procedures applied in finding the shortest-length confidence interval based on a pivot are also applicable in finding an interval that minimizes the expected length. We remark here that the restriction to unbiased confidence intervals is natural if we wish to minimize images . See Section 11.5 for definitions and further details.

The length of this interval is images . In this case we can plan our experiment to give a prescribed confidence level and a prescribed length for the interval. To have level images and length images , we choose the smallest n such that

images

This can also be interpreted as follows. If we estimate μ by images , taking a sample of size images , we are 100 images percent confident that the error in our estimate is at most d.

It follows that the minimum occurs at images (the other solution, images , is not admissible). The shortest-length confidence interval based on Tμ is the equal-tails interval,

images

The length of this interval is images , which, being random, may be arbitrarily large. Note that the same confidence interval minimizes the expected length of the interval, namely, images , where cn is a constant determined from images and the minimum expected length is images .

Numerical results giving values of a and b to four significant places of decimals are available (see Tate and Klett [112]). In practice, the simpler equal-tails interval,

images

may be used.

If μ is unknown, we use

images

as a pivot. images has a χ2(n–1) distribution. Proceeding as above, we can show that the shortest-length confidence interval based on images is images ; here a and b are a solution of

images

and

images

where fn—1 is the PDF of a images RV. Numerical solutions due to Tate and Klett [112]may be used, but, in practice, the simpler equal-tails confidence interval,

images

is employed.

Using Tθ as pivot, we see that the confidence interval is (X(n)/b, X(n)/a) with length images . We minimize L subject to

images

Now

images

and

images

so that the minimum occurs at images . The shortest interval is therefore images .

Note that

images

which is minimized subject to

images

where images and images . The expected length of the interval that minimizes EL is images , which is also the expected length of the shortest confidence interval based on X(n).

Note that the length of the interval images goes to 0 as images .

For some results on asymptotically shortest-length confidence intervals, we refer the reader to Wilks [118, pp. 374–376].

PROBLEMS 11.4

  1. Let X1,X2, … ,Xn be a sample from
    images

    Find the shortest-length confidence interval for θ at level images based on a sufficient statistic for θ.

  2. Let X1,X2, … ,Xn be a sample from G(1). Find the shortest-length confidence interval for θ at level images , based on a sufficient statistic for θ.
  3. In Problem 11.3.9 how will you find the shortest-length confidence interval for θ at level images , based on the statistic X/θ?
  4. Let T (Χ, θ) be a pivot of the form images . Show how one can construct a confidence interval for θ with fixed width d and maximum possible confidence coefficient. In particular, construct a confidence interval that has fixed width d and maximum possible confidence coefficient for the mean μ of a normal population with variance 1. Find the smallest size n for which this confidence interval has a confidence coefficient images . Repeat the above in sampling from an exponential PDF
    images

    (Desu [21])

  5. Let X1, X2, … ,Xn be a random sample from
    images

    Find the shortest-length (1–α) -level confidence interval for θ, based on the sufficient statistic images .

  6. In Example 4, let images . Find a (1–α) -level confidence interval for θ of the form (R, R/c). Compare the expected length of this interval to the one computed in Example 4.
  7. Let X1, X2, … ,Xn be a random sample from a Pareto PDF images , and = 0 for images . Show that the shortest-length confidence interval for θ based on X(1) is (X(1)α1/n, X(1)). (Use θ/X(1) as a pivot.)
  8. Let X1, X2, … ,Xn be a sample from PDF images and = 0 otherwise. Let images . Using images as a pivot for estimating images , show that the shortest-length confidence interval is of the form (R, R/c), where c is determined from the level as a solution of cn—1{(n – 1)c – n} + a = 0 (Ferentinos [27]).

11.5 UNBIASED AND EQUIVARIANT CONFIDENCE INTERVALS

In Section 11.3 we studied test inversion as one of the methods of constructing confidence intervals. We showed that UMP tests lead to UMA confidence intervals. In Chapter 9 we saw that UMP tests generally do not exist. In such situations we either restrict consideration to smaller subclasses of tests by requiring that the test functions have some desirable properties, or we restrict the class of alternatives to those near the null parameter values.

In this section will follow a similar approach in constructing confidence intervals.

Remark 1. Definition 1 says that a family S(X) of confidence sets for a parameter θ is unbiased at level images if the probability of true coverage is at least images and that of false coverage is at most images . In other words, S(X) traps a true parameter value more often than it does a false one.

We next show that S(X) is UMA. Let S*(x) be any other unbiased images -level family of confidence sets, and write images contains θ}. Then images , and it follows that Α*(θ) is the acceptance region of an unbiased size α test. Hence

images

The inequality follows since Α(θ) is the acceptance region of a UMP unbiased test. This completes the proof.

If the measure of precision of a confidence interval is its expected length, one is naturally led to a consideration of unbiased confidence intervals. Pratt [81]has shown that the expected length of a confidence interval is the average of false coverage probabilities.

Remark 2. If S(X) is a family of UMAU images -level confidence intervals, the expected length of S(X) is minimal. This follows since the left-hand side of (3) is the expected length, if θ is the true value, of S(X) and Ρθ{S(X) contains θ'} is minimal [because S(X) is UMAU], by Theorem 1, with respect to all families of images unbiased confidence intervals uniformly in images .

Since a reasonably complete discussion of UMP unbiased tests (see Section 9.5) is beyond the scope of this text, the following procedure for determining unbiased confidence intervals is sometimes quite useful (see Guenther [38]). Let X1,X2, … ,Xn be a sample from an absolutely continuous DF with PDF fθ (x) and suppose that we seek an unbiased confidence interval for θ. Following the discussion in Section 11.4, suppose that

images

is a pivot, and suppose that the statement

images

can be converted to

images

In order for images to be unbiased, we must have

(4)images

and

(5)images

If Ρ(θ, θ') depends only on a function γ of θ, θ', we may write

(6)images

and it follows that Ρ(γ) has a maximum at images .

Finally, let us briefly investigate how invariance considerations apply to confidence estimation. Let images . Let images be a group of transformations on images which leaves images invariant. Let S(X) be a images -level confidence set for θ.

The most useful method of constructing invariant confidence intervals is test inversion. Inverting the acceptance region of invariant tests often leads to equivariant confidence intervals under certain conditions. Recall that a group images of transformations leaves a hypothesis testing problem invariant if images leaves both Θ0 and Θ1 invariant. For each images , we have a different group of transformations, images , which leaves the problem of testing images invariant. The equivariant confidence interval, on the other hand, must be equivariant with respect to images , which is a much larger group since images for all θ0. The relationship between an equivariant confidence set and invariant tests is more complicated when the family images has a nuisance parameter τ.

Under certain conditions there is a relationship between equivariant confidence sets and associated invariant tests. Rather than pursue this relationship, we refer the reader to Ferguson [28, p. 262]; it is generally easy to check that (10) holds for a given confidence interval S to show that S is invariant. The following example illustrates this point.

PROBLEMS 11.5

  1. Let X1, X2, … ,Xn be a sample from U(0). Show that the unbiased confidence intervals for θ based on the pivot max Xi coincides with the shortest-length confidence interval based on the same pivot.
  2. Let X1,X2, … ,Xn be a sample from G(1). Find the unbiased confidence interval for θ based on the pivot 2 images .
  3. Let X1, X2, … ,Xn be a sample from PDF
    images

    Find the unbiased confidence interval based on the pivot 2n[min Xiθ].

  4. Let X1,X2, … ,Xn be iid images (μ,σ2) RVs where both μ and σ2 are unknown. Using the pivot images show that the shortest-length unbiased (1–α) -level confidence interval for μ is the equal-tails interval images .
  5. Let X1, X2, … ,Xn be iid with PDF images , and images otherwise. Find the shortest-length images -level unbiased confidence interval for θ based on the pivot θ/X(1) .
  6. Let X1,X2, … ,Xn be a random sample from a location family images . Show that a confidence interval of the form images where T(x) is an equivariant estimate under location group is an equivariant confidence interval.
  7. Let X1,X2, … ,Xn be iid RVs with common scale PDF images . Consider the scale group images . If T(x) is an equivariant estimate of σ, show that a confidence interval of the form
    images

    is equivariant.

  8. Let X1,X2, … ,Xn be iid RVs with PDF images and, images , otherwise. For testing images against images , consider the (UMP) test
    images

    Is the acceptance region of this α-level test an equivariant (1 – α) -level confidence interval (lower bound) for θ with respect to the location group?

    Yes.

11.6 RESAMPLING: BOOTSTRAP METHOD

In many applications of statistical inference the investigator has a random sample from a population distribution DF F which may or may not be completely specified. Indeed the empirical data may not even fit any known distribution. The inference is typically based on some statistic such as images , S2, a percentile or some much more complicated statistic such as sample correlation coefficient or odds ratio. For this purpose we need to know the distribution of the statistic being used and/or its moments. Except for the simple situations such as those described in Chapter 6 this is not easy. And even if we are able to get a handle on it, it may be inconvenient to deal with it. Often, when the sample is large enough, one can resort to asymptotic approximations considered in Chapter 7. Alternatively, one can use computer-intensive techniques which have become quite popular in the last 25 years due to the availability of fast home or office laptops and desktops.

Suppose x1, x2, … ,xn is a random sample from a distribution F with unknown parameter θ(F), and let images be an estimate of θ(F). What is the bias of images and its SE? Resampling refers to sampling from x1, x2, … ,xn and using these samples to estimate the statistical properties of images . Jackknife is one such method where one uses subsets of the sample by excluding one or more observations at a time. For each of these subsamples an estimate θj of θ is computed, and these estimates are then used to investigate the statistical properties of images .

The most commonly used resampling method is the bootstrap , introduced by Efron [22], where one draws random samples of size n, with replacement, from x1, x2, … ,xn. This allows us to generate a large number of bootstrap samples and hence bootstrap estimates images of θ. This bootstrap distribution of images is then used to study the statistical properties of images .

Let images , be iid RVs with common DF images , the empirical DF corresponding to the sample x1, x2, … ,xn. Then images is called a bootstrap sample . Let θ be the parameter of interest associated with DF F and suppose we have chosen images to be an estimate of θ based on the sample x1, x2, … ,xn. For each bootstrap sample let images , be the corresponding bootstrap estimate of θ. We can now study the statistical properties of images based on the distribution of the images , values. Let images Then the variance of images is estimated by the bootstrap variance.

(1)images

Similarly the bias of images , is estimated by

(2)images

Arranging the values of images , in increasing order of magnitude and then excluding 100α/2 percent smallest and largest values we get a (1–α) -level confidence interval for θ. This is the so-called percentile confidence interval. One can also use this confidence interval to test hypotheses concerning θ.

We took 1000 random samples, with replacement, of size 20 each from this sample with the following distribution of images .

IntervalFrequency
0.49–0.566
0.53–0.5729
0.57–0.61109
0.61–0.65200
0.65–0.69234
0.69–0.73229
0.73–0.77123
0.77–0.8159
0.81–0.8510
0.85–0.892

The bootstrap estimate of θ is images and that of the variance is 0.061. By excluding the smallest and the largest twenty-five vales of images a 95 percent confidence interval for θ is given by (0.564, 0.793). (We note that images and images so the images .)

Figure 1 show the frequency distribution of the bootstrap statistic images .

c11-fig-0001

Fig. 1

It is natural to ask how well does the distribution of the bootstrap statistic images approximate the distribution of images ? The bootstrap approximation is often better when applied to the appropriately centered images . Thus to estimate population mean θ bootstrap is applied to the centered sample mean images . The corresponding bootstrapped version will then be images , where images is the sample mean of the bth bootstrap sample. Similarly if images then the bootstrap approximation will be applied to the centered sample median images . The bootstrap version will be then be images . Similarly , in estimation of the distribution of sample variance S2, the bootstrap version will be applied to the ratio S22, where σ2 is the variance of the DF F.

We have already considered the percentile method of constructing confidence intervals. Let us denote the αth percentile of the distribution of images ,by Bα. Suppose that the sampling distribution of images is approximated by the bootstrap distribution of images . Then the probability that images is covered by the interval images is approximately images . This is called a images -level centered bootstrap percentile confidence interval for θ.

Recall that in sampling from a normal distribution when both mean and the variance are unknown, a images -level confidence interval for the mean θ is based on t-statistic and is given by images . For nonnormal distributions the bootstrap analog of the Student’s t-statistic is the statistic images . The bootstrap version is the statistic images , where SEb is the SE computed from the bootstrap sample distribution. A images -level confidence interval is now easily constructed.

In our discussion above we have assumed that F(θ) is completely unspecified. What if we know F except for the parameter θ? In that case we take bootstrap samples from the distribution images .

We refer the reader to Efron and Tibshirani [23]for further details.

PROBLEMS 11.6

    1. Show that there are images distinct bootstrap samples of size n. [Hint: Problem 1.4.17.]
    2. What is the probability that a bootstrap sample is identical to the original samples?
    3. What is the most likely bootstrap sample to be drawn?
    4. What is the mean number of times that xi appears in the bootstrap samples?
  1. Let x1, x2,…xn be a random sample. Then images is an estimate of the unknown mean μ. Consider the leave-one-out Jackknife sample. Let images be the mean of the remaining (n – 1) observations when xi is excluded:
    1. Show that images .
    2. Now suppose we need to estimate a parameter θ and choose images to be an estimate from the sample. Imitating the Jackknife procedure for estimating μ we note that images . What is the Jackknife estimate of θ? What is the Jackknife estimate of the bias of images and its variance?
  2. Let x1, x2,… ,xn be a random sample from images and suppose that images is an estimate of θ. Let images be a bootstrap sample from images . Show that both images and images have the same images distribution.
  3. Consider the data set
    images

    Let images , be a bootstrap sample from this data set:

    1. Find the probability that the bootstrap mean equals 2.
    2. Find the probability that the maximum value of the bootstrap mean is 9.
    3. Find the probability that the bootstrap sample mean is 4.
..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
18.191.233.205