Appendix

A.1 Notation

Notation

Explanation

Chapter

≻, ≺

preference relations

2

~

equivalence relation

2

Z

set of outcomes, rewards

3

z

outcome, generic element of Z

3

z0

best outcome

3

z0

worst outcome

3

Zi1i2...is0k

outcome from actions Image when the state of nature is Image

12

Zi1i2...iSk

outcome from actions Image when the state of nature is Image

12

Θ

set of states of nature, parameter space

3

θ

event, state of nature, parameter, generic element of Θ, scalar

2

θ

state of nature, parameter, generic element of Θ, vector

7

Image

generic state of nature upon taking stopping action Image at stage s

12

Image

generic state of nature at last stage S upon taking action Image

12

π

subjective probability

2

πθ

price, relative to stake S, for a bet on θ

2

πθ1|θ2

price of a bet on θ1 called off if θ2 does not occur

2

Image

current probability assessment of θ

2

Image

future probability assessment of θ

2

πM

least favorable prior

7

Sθ

stake associated with event θ

2

gθ

net gains if event θ occurs

2

A

action space

3

a

action, generic element of A

3

a(θ)

VNM action, function from states to outcomes

3

Image

expected value of the rewards for a lottery a

4

z*(a)

certainty equivalent of lottery a

4

λ(z)

Arrow–Pratt (local) measure of risk aversion at z

4

P

set of probability functions on Z

6

a

AA action, horse lottery, list of VNM lotteries a(θ) in P for θ ∊ Θ

6

a(θ, z)

probability that lottery a(θ) assigns to outcome z

6

aS,θ

action of betting stake S on event θ

2

aθ

action that would maximize decision maker’s utility if θ was known

13

a

action, vector

7

Image

generic action, indexed by is, at stage s of a multistage decision problem

12

Image

stopping action at stage s of a multistage decision problem

12

aM

minimax action

7

a*

Bayes action

3

p(z)

probability of outcome z

3

p

lottery or gamble, probability distribution over Z

3

χz

degenerate action with mass 1 at reward z

3

u

utility

3

u(z)

utility of outcome z

3

u(a(θ))

utility of outcome a(θ)

3

uθ(z)

state-dependent utility of outcome z

6

S(q)

expected score of the forecast probability q

10

s(θ, q)

scoring rule for the distribution q and event θ

10

RP(a)

risk premium associated with action a

4

X

sample space

7

x

random variable or observed outcome

7

xn

random sample (x1, ..., xn)

14

x

multivariate random sample

7

Image

random variable with possible values xi1, ..., xiJs observed at stage s, upon taking continuation action Image

12

f (x|θ)

probability density function of x or likelihood function

7

m(x)

marginal density function of x

7

F(x|θ)

distribution function of x

7

π(θ)

prior probability of θ; may indicate a density if θ is continuous

7

π(θ|x), πx

posterior density of θ given x

7

Ex[g(x)]

expectation of the function g(x) with respect to m(x)

7

Ex|θ[g(x)], E[g(x)|θ]

expectation of the function g(x) with respect to f (x|θ)

7

uπ(a)

expected utility of action a, using prior π

3

uπ(d)

expected utility of the sequential decision procedure d

15

δ

decision rule, function with domain X and range A

7

δ

multivariate decision rule

7

δM

minimax decision rule

7

δ*

Bayes decision rule

7

δR(x, ·)

randomized decision rule for a given x

7

δn

terminal decision rule after n observations xn

15

δ

sequence of decision rules δ1(x1), δ2(x2), ...

15

ζn

stopping rule after n observations xn

15

ζ

sequence of stopping rules ζ0, ζ1(x1), ...

15

d

sequential decision rule: pair (δ, ζ)

15

L(θ, a)

loss function (in regret form)

7

Lu(θ, a)

loss function as the negative of the utility function

7

L(θ, a, n)

loss function for n observations

14

u(θ, a, n)

utility function for n observations

14

L(θ, δR(x))

loss function for randomized decision rule δR

7

Lπ(a)

prior expected loss for action a

7

Lπx(a)

posterior expected loss for action a

7

U(θ, d)

average utility of the sequential procedure d for given θ

15

R(θ, δ)

risk function of decision rule δ

7

V(π)

maximum expected utility with respect to prior π

13

W0(πxn, n)

posterior expected utility, at time n, of colleting 0 additional observations

15

Wkn(πxn, n)

posterior expected utility, at time n, of continuing for at most an additional kn steps

15

r(π, δ)

Bayes risk associated with prior π and decision δ

7

r(π, n)

Bayes risk adopting the optimal terminal decision with n observations

14

D

class of decision rules

7

DR

class of all (including randomized) decision rules

7

ɛ

perfect experiment

13

ɛ

generic statistical experiment

13

ɛ12

statistical experiment consisting of observing both variables x1 and x2

13

ɛ(n)

statistical experiment consisting of experiments ɛ1, ɛ2, ..., ɛn where ɛ2, ..., ɛn are conditionally i.i.d. repetitions of ɛ1

13

Vθ(ɛ∞)

conditional value of perfect information for a given θ

13

V(ɛ)

expected value of perfect information

13

Vx(ɛ)

observed value of information for a given x in experiment ɛ

13

V(ɛ)

expected value of information in the experiment ɛ

13

V(ɛ12

expected value of information in the experiment ɛ12

13

V(ɛ2|ɛ1)

expected information of x2 conditional on observed x1

13

Ix(ɛ)

observed (Lindley) information provided by observing x in experiment ɛ

13

I(ɛ)

expected (Lindley) information provided by the experiment ɛ

13

Ix(ɛ12)

expected (Lindley) information provided by the experiment ɛ12

13

I(ɛ2|ɛ1)

expected (Lindley) information provided by E2 conditional on E1

13

c

cost per observation

14

C(n)

cost function

14

Image

Bayes rule based on n observations

14

Uπ (n)

expected utility of making n observations and adopting the optimal terminal decision

14

n*

Bayesian optimal sample size

14

nM

Minimax optimal sample size

14

η

nuisance parameter

7

Φ(.)

cumulative distribution function of the N(0,1)

8

M

class of parametric models in a decision problem

11

M

generic model within M

11

Hs

history set at stage s of a multistage decision problem

12

A.2 Relations

p(z)

=

Image

(3.3)

Image

=

Image

(4.4)

z*(a)

=

Image

(4.9)

RP(a)

=

Image

(4.10)

λ(z)

=

u″ (z)/u′ (z) = –(d/dz)(log u′ (z))

(4.13)

Lu(θ, a)

=

u(a(θ))

(7.1)

L(θ, a)

=

Lu(θ, a) – infaA Lu(θ, a)

(7.2)

 

=

supa′(θ) u(a′ (θ)) – u(a(θ))

(7.3)

Lπ(a)

=

Image

(7.6)

Lπx(a)

=

Image

(7.13)

Uπ(a)

=

Image

(3.4)

 

=

Image

(3.5)

S(q)

=

Image

(10.1)

V(π)

=

supa Uπ(a)

(13.2)

aM

=

argmin maxθ L(θ, a)

(7.4)

a*

=

argmax Uπ(a)

(3.6)

 

=

argmin ∫Θ L(θ, a)π(θ)

(7.5)

aθ

=

arg supaA u(a(θ))

(13.4)

m(x)

=

Θ π(θ)f (x|θ)

(7.12)

π(θ|x)

=

π(θ)f (x|θ)/m(x)

(7.11)

L(θ, δR(x))

=

Image

(7.16)

R(θ, δ)

=

Image

(7.7)

r(π, δ)

=

Image

(7.9)

δM

s.t. supθ R(θ, δM) = infδ supθ R(θ, δ)

(7.8)

δ*

s.t. r(π, δ*) = infδ r(π, δ)

(7.10)

Vθ(ɛ∞)

=

u(aθ(θ)) – u(a*(θ))

(13.5)

V(ɛ)

=

Eθ[Vθ(ɛ)]

(13.6)

 

=

Eθ [supa u(a(θ))] – V(π)

(13.7)

Vx(ɛ)

=

V(πx) – Uπx(a*)

(13.9)

V(ɛ)

=

Ex[Vx(ɛ)] = Ex[V(πx) – Uπx(a*)]

(13.10)

 

=

E[Vx(ɛ)] = Ex[V(πx)] – V(π)

(13.12)

 

=

Ex[V(πx)] – V(Ex[πx])

(13.13)

V(ɛ2|ɛ1)

=

Ex1x2[V(πx1x2)] – Ex1[V(πx1)]

(13.17)

V(ɛ12

=

Ex1x2[V(πx1x2)] – V(π)

(13.16)

 

=

V(ɛ1) + V(ɛ2|ɛ1)

(13.18)

I(ɛ)

=

XΣ log (πx(θ)/π(θ)) πx(θ)m(x)dθdx

(13.27)

 

=

ΣX log (f (x|θ)/m(x)) f (x|θ)π(θ)dxdθ

(13.28)

 

=

Ex [Eθ|x [log (f (x|θ)/m(x))]]

(13.29)

 

=

Ex [Eθ|x[log (fx(θ)/m(x))]]

(13.30)

I(ɛ2|ɛ1)

=

Ex1Ex2|x1Eθ|x1,x2 [log (f (x2|θ, x1)/m(x2|x1))]

(13.33)

I(ɛ12

=

I(ɛ1) + I(ɛ2|ɛ1)

(13.31)

Ix(ɛ)

=

Σ πx(θ)log x(θ)(θ)) dθ

(13.26)

u(a, θ, n)

=

u(a(θ)) – C(n)

(14.1)

L(θ, a, n)

=

L(θ, a) + C(n)

(14.2)

r(π, n)

=

Image

(14.3)

 

=

Image

(14.4)

Uπ(n)

=

Image

(14.5)

A.3 Probability (density) functions of some distributions

  1. If x ~ Bin(n, θ) then

    Image
  2. If x ~ N(θ, σ2) then

    Image
  3. If x ~ Gamma(α, β) then

    Image
  4. If x ~ Exp(θ) then

    f (x|θ) = θeθx.

  5. If x ~ BetaBinomial(n, α, β) then

    Image
  6. If x ~ Beta(α, β) then

    Image

A.4 Conjugate updating

In the following description, prior hyperparameters have subscript 0, while parameters of the posterior distributions have subscript x.

  1. Sampling model: binomial

    1. Data: x|θ ~ Bin(n, θ).

    2. Prior: θ ~ Beta(α0, β0).

    3. Posterior: θ|x ~ Beta(αx, βx) where

      αx = α0 + x and βx = β0 + nx.

    4. Marginal: x ~ BetaBinomial(n, α0, β0).

  2. Sampling model: normal

    1. Data: x ~ N(θ, σ2)(σ2 known).

    2. Prior: Image.

    3. Posterior: Image where

      Image
    4. Marginal: Image.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
18.216.55.20