Top Banner

of 30

Good Conditional Expectation Notes

Jun 04, 2018

Download

Documents

ingramk
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
  • 8/13/2019 Good Conditional Expectation Notes

    1/30

    Stochastic Processes

    (Lecture #4)

    Davy Paindaveine

    Universite Libre de Bruxelles

    Stochastic Processes (Lecture #4) p. 1/

  • 8/13/2019 Good Conditional Expectation Notes

    2/30

    Today

    Today, our goal will be

    to finish with conditional expectations,

    to have a quick look at conditional variances,

    and

    to discuss limits of sequences of r.v.s, and

    to study famous limiting results.

    Stochastic Processes (Lecture #4) p. 2/

  • 8/13/2019 Good Conditional Expectation Notes

    3/30

    Conditional expectation

    Conditional expectationallows for exploiting some information(e.g., the information that some event occurred) in order to

    improve the (unconditional) best guessE[X].

    This available information can take various forms:

    (the occurrence of) an event.

    (the value of )a r.v.

    (the occurrence of some event in) a sigma-algebra.

    Stochastic Processes (Lecture #4) p. 3/

  • 8/13/2019 Good Conditional Expectation Notes

    4/30

    Conditional expectation (w.r.t. a r.v.)

    LetXbe an integrable r.v. on(,A, P).LetY be adiscreter.v. on(,A, P), say with distribution

    distribution ofY

    values y1 y2 . . .

    probabilities p1 p2 . . .

    Then we defineE[X|Y]as the r.v.

    E[X|Y] : R

    E[X|Y =Y()].The last conditional expectation is w.r.t. an event, and hence is

    well defined.

    Stochastic Processes (Lecture #4) p. 4/

  • 8/13/2019 Good Conditional Expectation Notes

    5/30

  • 8/13/2019 Good Conditional Expectation Notes

    6/30

    Conditional expectation (w.r.t. a r.v.)

    LetXbe an integrable r.v. on(,A, P).LetYbe anabsolutely continuousr.v. on(,A, P), say with pdff.

    Then we defineE[X

    |Y]as the r.v. such that

    (i)E[X|Y]is(Y)-measurable.(ii)

    A

    E[X|Y]() dP() = A

    X() dP(), for allA (Y).

    In practice:

    IfX is discrete,E[X|Y =y] =i xiP[X=xi|Y =y].

    IfXis absolutely continuous,

    E[X|Y =y] = 1fY(y)

    R

    x f(X,Y)(x, y) dx=

    R

    x f(X|Y)(x, y) dx,

    wheref(X

    |Y)

    (x, y)denotes the pdf ofXconditionally onY =y.

    Stochastic Processes (Lecture #4) p. 6/

  • 8/13/2019 Good Conditional Expectation Notes

    7/30

    Conditional expectation (w.r.t. a r.v.)

    Same properties as in the discrete case:

    E[X|Y]is(Y)-measurable E[X

    |Y] = E[X]iffX

    Y.

    E[X|Y] =X iffX is(Y)-measurable. E

    E[X|Y]

    = E[X].

    Remark: E[X]= E

    E[X|Y]

    =R

    E[X|Y =y] fY(y) dy.(//total probability formula).

    We still define: P[A|Y] = E[IA|Y]]. P[A] = E[IA] = E

    E[IA|Y]

    =

    R

    E[IA|Y =y] fY(y) dy=

    RP[A|Y =y] fY(y) dy.

    Stochastic Processes (Lecture #4) p. 7/

  • 8/13/2019 Good Conditional Expectation Notes

    8/30

    Conditional expectation (w.r.t. a r.v.)

    Example: letX Unif(0, 1). IfX=x, you flipm times a coin suchthatP[Head] =x. LetNbe the number of heads.

    E[N] =? Distribution ofN?

    E[N|X=x] =mn=0

    n P[N=n|X=x] =mn=0

    n

    m

    n

    xn(1 x)mn =

    =

    mn=1

    m

    m 1n 1

    xn(1x)mn =mx

    mn=1

    m 1n 1

    xn1(1x)(m1)(n1) =mx.

    E[N|X] =mX ((X)-measurable).

    Consequently, E[N] =E

    E[N|X]

    = E[mX] =m E[X] = m2.

    Stochastic Processes (Lecture #4) p. 8/

  • 8/13/2019 Good Conditional Expectation Notes

    9/30

    Conditional expectation (w.r.t. a r.v.)

    Distribution ofN?

    P[N=n] = R

    P[N=n

    |X=x] fX(x) dx=

    1

    0

    P[N=n

    |X=x] dx

    =

    10

    m

    n

    xn(1x)mn dx=

    m

    n

    10

    xn(1x)mn dx=. . .= 1m + 1

    ,

    for alln= 0, 1, . . . , m.

    Hence,Nis uniformly distributed on {0, 1, . . . , m}.(which now makes clear whyE[N] = m2).

    Stochastic Processes (Lecture #4) p. 9/

  • 8/13/2019 Good Conditional Expectation Notes

    10/30

    Conditional expectation

    Conditional expectationallows for exploiting some information(e.g., the information that some event occurred) in order to

    improve the (unconditional) best guessE[X].

    This available information can take various forms:

    (the occurrence of) an event.

    (the value of ) a r.v.

    (the occurrence of some event in)a sigma-algebra.

    Stochastic Processes (Lecture #4) p. 10/

  • 8/13/2019 Good Conditional Expectation Notes

    11/30

    Conditional expectation (w.r.t. a-algebra)

    LetXbe an integrable r.v. on(,A, P).Let Fbe a sigma-algebra A.

    Then we defineE[X

    |F]as the r.v. such that

    (i)E[X|F]is F-measurable.(ii)

    A

    E[X|F]() dP() = A X() dP(), for allA F.

    Same kind of properties as forE[X|Y]: E[X|F]is F-measurable E[X

    |F] = E[X]iffX

    F.

    E[X|F] =X iffX is F-measurable. E

    E[X|F]

    = E[X].

    Stochastic Processes (Lecture #4) p. 11/

  • 8/13/2019 Good Conditional Expectation Notes

    12/30

    Conditional expectation (w.r.t. a-algebra)

    Let F1 F2be two-algebras in A.Extra properties:

    EE[X|F2

    ]F1= E[X

    |F1].

    E

    E[X|F1]F2

    = E[X|F1].

    Conditional expectation w.r.t. to a-algebra also allows for

    defining conditional expectationw.r.t. a collection of r.v.s.

    More specifically, we define

    E[X|Y1, . . . , Y n] := E[X|(Y1, . . . , Y n)],

    where(Y1, . . . , Y n)is the smallest-algebra containing

    {Y1i (B)

    |B

    B, i= 1, . . . , n

    }.

    Stochastic Processes (Lecture #4) p. 12/

  • 8/13/2019 Good Conditional Expectation Notes

    13/30

    Conditional expectation (w.r.t. a-algebra)

    To understand how one can computeE[X|F]in practice,see exercise 6, in homework 1.

    Stochastic Processes (Lecture #4) p. 13/

  • 8/13/2019 Good Conditional Expectation Notes

    14/30

    Conditional variances

    LetXandYbe r.v.s on(,A, P).As we have seen, one can exploit the information sitting inY

    in order to improve the (unconditional) best guessE[X].

    the improved guess is the conditional expectationE[X|Y].Similarly, onceYis observed, the knowledge of the

    (unconditional) dispersion ofX, namely

    Var[X] = E

    (X E[X])2

    ,

    can be improved into theconditional variance

    Var[X|Y] = E

    (X E[X|Y])2Y.

    Stochastic Processes (Lecture #4) p. 14/

  • 8/13/2019 Good Conditional Expectation Notes

    15/30

    Conditional variances

    In time series analysis, many models are designed to explain thedynamics of the conditional varianceVar[Xt|Xt1, Xt2, . . .](e.g., the so-called stochastic volatility models).

    Remarks:

    Similarly as forVar[X],

    Var[X|Y] = E[X2

    |Y] E[X|Y]2

    (exercise; better for computations, not for the interpretation).

    Var[X]

    = E[Var[X

    |Y]], but we have

    Var[X] = E

    Var[X|Y]+ VarE[X|Y].Indeed, E[Var[X|Y]] = E[X2] E[

    E[X|Y]

    2

    ] = Var[X] +

    E[X]

    2

    E[E[X|Y]2

    ] = Var[X] +

    E

    E[X|Y]2

    E[E[X|Y]2

    ].

    Stochastic Processes (Lecture #4) p. 15/

  • 8/13/2019 Good Conditional Expectation Notes

    16/30

    Today

    Today, our goal will be

    to finish with conditional expectations,

    to have a quick look at conditional variances,

    and

    to discuss limits of sequences of r.v.s, and

    to study famous limiting results.

    Stochastic Processes (Lecture #4) p. 16/

  • 8/13/2019 Good Conditional Expectation Notes

    17/30

    Convergence of sequences of r.v.s

    LetX1, X2, . . .bei.i.d.r.v.s, that is, r.v.s that areindependent andidenticallydistributed. AssumeX1 is square-integrable, and

    write:= E[X1]and2 := Var[X1].

    Let Xn := 1nn

    i=1 Xi. Then

    E[Xn] = 1nn

    i=1E[Xi] = E[X1] =, and

    Var[Xn] =

    1n2 Var[

    ni=1 Xi] =

    1n2

    ni=1Var[Xi] =

    1nVar[X1] =

    2

    n,

    which converges to0asn .

    Consequently, we feel intuitively that Xn X, whereX=a.s.

    How to make this convergence precise?

    Stochastic Processes (Lecture #4) p. 17/

  • 8/13/2019 Good Conditional Expectation Notes

    18/30

    Convergence of sequences of r.v.s

    Consider a sequence of r.v.s(Xn)and a r.v. X, defined on(,A, P).How to defineXn X(asn )?

    Xna.s.

    X(almost surely) P[{ |Xn() X()}] = 1.Xn

    P X(in probability) For all >0,P[|Xn X| > ] 0.

    Xn

    Lr

    X(inLr

    ,r >0) E[|Xn X|r

    ] 0.Xn

    D Xin distribution (or in law) FXn(x) FX(x)for allxatwhichFX is continuous.

    Stochastic Processes (Lecture #4) p. 18/

  • 8/13/2019 Good Conditional Expectation Notes

    19/30

    Convergence of sequences of r.v.s

    Example:

    LetY1, Y2, . . .be i.i.d. r.v.s, with common distribution

    distribution ofYi

    values 0 2

    probabilities 12

    1

    2

    DefineXn=ni=1 Yi.

    The distribution ofXn is

    distribution ofXn

    values 0 2n

    probabilities 1 12n

    1

    2n

    We feel thatXn X, whereX= 0a.s... But in what sense?Stochastic Processes (Lecture #4) p. 19/

  • 8/13/2019 Good Conditional Expectation Notes

    20/30

    Convergence of sequences of r.v.s

    We feel thatXn X, whereX= 0a.s...But in what sense?

    In probability?

    For all >0,P[|Xn X| > ] =P[Xn> ] P[Xn >0] = 12n 0,asn .

    Xn P X, asn .

    InL1?

    E[

    |Xn

    X

    |] = E[Xn] = 0

    1

    12n + 2n

    12n = 1,

    which does not go zero, asn . the convergence does not hold in theL1 sense.

    Stochastic Processes (Lecture #4) p. 20/

  • 8/13/2019 Good Conditional Expectation Notes

    21/30

    Convergence of sequences of r.v.s

    It can be shown that:

    Xn a.s. X Xn P X Xn Lr

    X

    Xn D X

    Stochastic Processes (Lecture #4) p. 21/

    C f f

  • 8/13/2019 Good Conditional Expectation Notes

    22/30

    Convergence of sequences of r.v.s

    Some arrows can sometimes be reverted:

    Xn P X there exists a subsequence(Xnk)for whichXnk

    a.s. X. Xn P Xand theXns are uniformly integrable Xn L

    r X. Xn D a Xn P a(a is some constant).

    Definition: theXns are uniformly integrable (i)supnE[|Xn|]< .(ii)P[A]

    0

    supn A |

    Xn()

    |d

    0.

    Recall: ifX is integrable,P[A] 0 A |X()| d 0(hence,only thesupnbrings some extra condition in (ii) above).

    Stochastic Processes (Lecture #4) p. 22/

    Li iti th

  • 8/13/2019 Good Conditional Expectation Notes

    23/30

    Limiting theorems

    Here are the two most famouslimiting resultsin probability and

    statistics...

    Stochastic Processes (Lecture #4) p. 23/

    Li iti th

  • 8/13/2019 Good Conditional Expectation Notes

    24/30

    Limiting theorems

    Thelaw of large numbers(LLN):

    LetX1, X2, . . .be i.i.d. integrable r.v.s.

    Write:= E[X1].

    Then

    Xn := 1

    n

    n

    i=1

    Xia.s. .

    Remarks:

    Basically no assumption. Interpretation for favourable/fair/defavourable games of

    chance.

    Stochastic Processes (Lecture #4) p. 24/

    Li iti th

  • 8/13/2019 Good Conditional Expectation Notes

    25/30

    Limiting theorems

    An example...

    LetX1, X2, . . .be i.i.d., withXi Bern(p).Then= E[Xi] =p, so that

    Xn := 1

    n

    ni=1

    Xia.s. p.

    Interpretation:

    The empirical proportion of successes converges to the

    theoretical proportion of successes (that is, the probability of

    success).

    Remark: this also shows that, if(Yn)is a sequence of r.v.s with

    Yn Bin(n, p), thenYn/na.s.

    p.Stochastic Processes (Lecture #4) p. 25/

    Limiting theorems

  • 8/13/2019 Good Conditional Expectation Notes

    26/30

    Limiting theorems

    ... and some applets:

    http://hadm.sph.sc.edu/COURSES/J716/a01/stat.html

    http://www.stat.berkeley.edu/stark/Java/Html/lln.htm ...

    Stochastic Processes (Lecture #4) p. 26/

    Limiting theorems

  • 8/13/2019 Good Conditional Expectation Notes

    27/30

    Limiting theorems

    Thecentral limit theorem(CLT):

    LetX1, X2, . . .be i.i.d. square-integrable r.v.s.

    Write:= E[X1]and2 := Var[X1].

    ThenXn /

    n

    D Z, withZ N(0, 1).

    Remarks:

    Xn/n

    =XnE[Xn]

    Var[Xn].

    It says sth about the speed of convergence in Xn a.s. . It allows for computingP[Xn B]for largen...

    It is valid whatever the distribution of theXis!Stochastic Processes (Lecture #4) p. 27/

  • 8/13/2019 Good Conditional Expectation Notes

    28/30

    Limiting theorems

  • 8/13/2019 Good Conditional Expectation Notes

    29/30

    Limiting theorems

    (B) Throw 9 dices and denote by Xthe sum of the results.Repetition of this yields i.i.d. r.v.sX1, X2, . . .

    Note thatXi =Yi1+ . . . + Yi9, whereYij is the result of the jth

    dice in theith roll. Then we have

    = E[Xi] = E[Yi1] + . . . + E[Yi9] = 9 3.5 = 31.5

    and

    2 = Var[Xi] = Var[Yi1] + . . . + Var[Yi9] = 9 2.91666...= 26.25,

    so that

    n(Xn 31.5)

    26.25

    D Z, withZ N(0, 1).

    Stochastic Processes (Lecture #4) p. 29/

    Limiting theorems

  • 8/13/2019 Good Conditional Expectation Notes

    30/30

    Limiting theorems

    ... and some applets:

    http://www.math.csusb.edu/faculty/stanton/probstat/clt.html

    http://www.stat.vt.edu/sundar/java/applets/(then go to Statistical Theory Central Limit Theorem

    Main page).

    http://www.jcu.edu/math/isep/Quincunx/Quincunx.html http://bcs.whfreeman.com/ips4e/cat 010/applets/

    CLT-Binomial.html

    ...

    Stochastic Processes (Lecture #4) p. 30/