stat420c3

Embed Size (px)

Citation preview

  • 8/6/2019 stat420c3

    1/33

    Chapter 3: Some Time-Series Models

    1

  • 8/6/2019 stat420c3

    2/33

    Section 3.1 Stochastic Processes and Their

    Properties

    There are several terminologies in this sections:

    Stochastic processes (random processes): xtfor t R, where xt is a random variable whent is given.

    continuous: t (,).

    discrete: t = 0,1,2, . Ensemble: the set of xt for all possible t.

    Realization: the element of the ensemble.

    We write X(t) or Xt if we treat time series

    as random, and x(t) or xt if we treat it as

    observations.

    2

  • 8/6/2019 stat420c3

    3/33

    LetXt

    be the time series. Then,

    the mean function is

    (t) = E[X(t)];

    the variance function is

    2(t) = V ar[X(t)];

    the autocovariance is(t1, t2) =Cov(X(t1), X(t2))

    =E{[X(t1) (t1)][X(t2) (t2)]} the autocorrelation is

    (t1, t2) =Cor(X(t1), X(t2))

    =E{[X(t1) (t1)][X(t2) (t2)]}

    (t1)(t2)

    = (t1, t2)(t1)(t2)

    .

    Clearly, there is (t, t) = 2(t) and (t, t) = 1.

    3

  • 8/6/2019 stat420c3

    4/33

    3.2. Stationary Processes

    Strictly stationary.

    A time series is said strictly stationary if the joint

    distribution ofX

    (t1

    ), , X

    (tk

    ) is the same as the

    joint distribution of X(t1 + ), , X(tk + ) forany possible values of and k

    If X(t) is strictly stationary, then (t) i s aconstant and (t1, t2) only depends on t1 t2.

    Then, we can write

    (t1, t2) = (|t1 t2|)

    and

    (t1, t2) = (|t1 t2|).

    4

  • 8/6/2019 stat420c3

    5/33

    Second-order stationary.

    A time series is said second-order stationary if

    (t) is independent oft, and (t1, t2) only depends

    on |t1 t2|.

    If the joint distribution of X(t1), , X(tk) isalways normal, then we call X(t) is a normal

    time series (or just normal).

    If X(t) is normal, then second-order stationaryand strictly stationary are equivalent.

    Mostly, we simply call second-order stationary asstationary.

    5

  • 8/6/2019 stat420c3

    6/33

    Let X(t) be a time series. If the limit distribution

    of X(t) exist, then this distribution is equilibrium

    distribution.

    If the conditional distribution of X(t + 1) givenX(t) is invariant, then there exists an equilibrium

    distribution. This is also called a Markov Chain.

    6

  • 8/6/2019 stat420c3

    7/33

    3.3 Some Properties of Autocorrelation Function

    Let X(t) be a (second-order) stationary time se-

    ries. Then we can define the following: Autocorrelation function:

    () =()

    (0).

    The correlation matrix of X(t1), , X(tn) is

    1 (t1 t2) (t1 tn)(t2 t1) 1 (t2 tn)

    ... ... . . . ...

    (tn t1) (tn t2) 1

    ,

    which should be nonnegative-definite.

    Therefore, an autocorrelation function ()satisfies

    (i) |()| 1;

    (ii) (0) = 1;

    (iii) () is nonnegative-definite.

    Nugget effect: if(0+) = lim

    0 () < 1,

    there there exist a nugget effect.7

  • 8/6/2019 stat420c3

    8/33

    3.4. Some Useful Models

    3.4.1 Purely random processes:

    Let Zt be a discrete stationary time series. If Zt1and Zt2 are independent when t1

    = t2, then Zt is

    call a pure random processes or a white noise. It

    has:

    an autocovariance function as(k) =Cov(Zt, Zt+k)

    =

    2Z k = 00, k = 0.

    an autocorrelation function as(k) =Corr(Zt, Zt+k)

    =1 k = 00, k = 0.

    Sometimes, we make the assumption weaker from

    independent to uncorrelated. This is enough for

    any inference of linear operations.

    8

  • 8/6/2019 stat420c3

    9/33

    3.4.2. Random walks:

    Let Xt be a discrete time series and Zt be the

    white noise. Then, Xt is called a random walk if

    Xt = Xt1 + Zt and X0 = 0.

    For a random walk, we have

    Xt =t

    i=1

    Zi.

    Suppose E(Zt) = and V(Zt) = 2Z. Then, we

    have E(Xt) = t and V(Xt) = t2Z. By CLT, we

    have

    Xt tt

    L N(0, 2Z).

    9

  • 8/6/2019 stat420c3

    10/33

    An example:

    Assume Tom and Jerry are gambling. Tom has

    m units of money and Jerry has n units of money.

    Each time, either Tom or Jerry wins 1 unit of

    money. Suppose the probability of Tom win is p.

    Compute the probability that Tom wins all Jerrysmoney.

    Solution: Let Zt be defined by

    P(Zt = 1) = p; P(Zt = 1) = 1 p.Define X0 = 0 and

    Xt =t

    i=1

    Zi.

    Then, if Xt attains n first, Tom wins the game;

    otherwise Jerry wins the game.

    Let f(a) be the probability that Tom wins the

    game if Tom has a units of money and Jerry has

    m + n a units of money.

    10

  • 8/6/2019 stat420c3

    11/33

    Then, we have

    (i) f(n) = 1;

    (ii) f(

    m) = 0;

    (iii) ifm < a < n,

    f(a) = pf(a + 1 ) + qf(a 1).

    Thus, we have

    f(a + 1) f(a) =qp

    [f(a) f(a 1)]

    =(q

    p)2[f(a 1) f(a 2)]

    =(q

    p)a+m[f(m + 1) f(m)]

    =(qp

    )a+mf(m + 1).

    11

  • 8/6/2019 stat420c3

    12/33

    Assume p = q. Thus, we haven

    a=m[f(a + 1) f(a)] =

    nm

    (q

    p)a+mf(m + 1)

    f(m + 1)(1 (q/p)m+n

    1

    (q/p)

    ) = 1

    f(m + 1) = 1 (q/p)1 (q/p)m+n.

    Therefore,

    1

    a=m

    [f(a + 1) f(a)] =1

    a=m

    (q

    p)a+mf(m + 1)

    f(0) = f(m + 1)(1 (q/p)m

    1 (q/p) )

    f(0) = 1 (q/p)m

    1 (q/p)m+n.

    When p = q, we have

    f(0) =m

    m + n

    by taking the limit pq 1.

    12

  • 8/6/2019 stat420c3

    13/33

    Thus, we have

    f(m) =

    1(q/p)m1(q/p)m+n when p = q

    mm+n when p = q = 1/2

    Clearly, we have when

    limn f(m) =

    1 (q/p)m when p > 1/20 when p 1/2

    13

  • 8/6/2019 stat420c3

    14/33

    3.4.3. Moving average (MA) processes

    A moving average process xT has the form of

    Xt = 0Zt + 1Zt1 + + qZtq, (1)where Zt is a white noise (purely random process)

    with E(Zt) = 0 and V(Zt) = 2Z.

    Usually, 0 is scaled to 0 = 1.

    We write

    Zt W N(0, 2Z)for such Zt and

    Xt M A(q)for such Xt.

    14

  • 8/6/2019 stat420c3

    15/33

    Clearly, we have

    E(Xt) = 0

    V(Xt) = 2Z

    qi=1

    2i

    (k) = (k)

    =2Zqki=0 ii+k, when k = 0,

    , q

    0 when k > q.and

    (k) = (k)

    =

    qk

    i=0 ii+k

    qi=0

    2i

    when k = 0, , q

    0 when k > q.

    15

  • 8/6/2019 stat420c3

    16/33

    A MA process Xt expressed by above is invertible

    if the innovation expression converges that is

    Zt =

    j=0

    jXtj

    with absolute convergence coefficient as

    j=0

    |j| < .

    16

  • 8/6/2019 stat420c3

    17/33

    The backward shift operator B is defined by

    BXt = Xt1; B2Xt = Xt2; ,and in general there is

    Bj

    Xt = Xtj.

    Then, equation (1) can be expressed as

    Xt = (0 + 1B + + qBq)Zt = (B)Zt

    where

    (B) = 0 + 1B + + qBq.

    If an MA(q) process is invertible if the roots of

    the equation

    (B) = 0

    are all outside the unit circle in the complex plane.

    17

  • 8/6/2019 stat420c3

    18/33

    Example: Consider the following M A(1) model

    (a) : Xt = Zt 0.5Zt1 = (1 0.5B)Zt.Then,

    Zt =1

    1 0.5BXt

    =

    k=0

    0.5kBkXt

    =Xt +1

    2Xt1 +

    1

    4Xt2 +

    1

    8Xt3 + .

    Then, based on Xt, we can define Zt according to

    the above formula.

    (b) : Xt = Zt 2Zt1 = (1 2B)Zt.Then,

    Zt =1

    1 2BXt

    =Xt + 2Xt1 + 4Xt2 + 8Xt3 + .Then, based on Xt, we cannot define Zt according

    to the above formula.

    18

  • 8/6/2019 stat420c3

    19/33

    The autocorrelation for (a) is

    (1) =0.5

    1 + 0.52= 0.4

    The autocorrelation for (b) is

    (1) = 2

    1 + 22 = 0.4.Thus, MA(1) model in (a) and (b) are equivalent,

    which implies (a) is more approrpiate.

    19

  • 8/6/2019 stat420c3

    20/33

    Example: check whether the following MA pro-

    cesses are invertible

    (a) Xt = (1 + B)Zt.

    (b) Xt = (1 + 0.7B + 0.1B2)Zt.

    20

  • 8/6/2019 stat420c3

    21/33

    3.4.4. Autoregressive processes

    Suppose Zt W N(0, 2Z). A process {Xt} is saidan autoregressive of order p (AR(p)) if

    Xt = 1Xt1 + + pXtp + Zt. (2)This can also be expressed as

    Zt = Xt (1Xt1 + + pXtp). (3)

    Since we only observed Xt, equation (3) can be

    used to derived the white noise Zt.

    21

  • 8/6/2019 stat420c3

    22/33

    First-order process

    Xt = Xt1 + Zt.

    This is equivalent to an infinite MA process

    Xt =

    j=0

    jZt

    j,

    which is well defined when || < 1. For this pro-cess, we have

    E(Xt) = 0

    V(Xt) = 2Z

    j=0

    2j

    (k) = 2Z

    j=0

    k+2j,

    and

    (k) = k

    for k = 0, 1, .

    22

  • 8/6/2019 stat420c3

    23/33

    General-order process

    Zt = (1 1B pBp)Xtor equivalently as

    Xt =Zt

    1

    1

    B

    p

    Bp= f(B)Zt,

    where

    f(B) =(1 1B pBp)1=(1 + 1B + 2B

    2 + ).If

    j=1

    |j| <

    then Xt is well defined which is also stationary.

    It is also equivalent that if the roots of

    (B) = (1 1B pBp) = 0are all outside of the unit circle on complex plane,

    then Xt is well defined.

    23

  • 8/6/2019 stat420c3

    24/33

    Since finding 1, 2, usually is not easy, wesometimes use stationarity and derived the fol-

    lowing formulae (Yule-Walker equations)

    (k) = 1(k 1) + + p(k p)for k > 0. Note that (0) = 1 and (k) = (k).We can solve those by linear equations.

    This has a general expression

    (k) = A1|k| + + App|k|

    where i are roots of the auxiliary equations

    yp 1yp1 p = 0,

    where A1, , Ap can be solved by the first p linearequations.

    24

  • 8/6/2019 stat420c3

    25/33

    Example: Consider AR(2) processXt = 1Xt1 + 2Xt2 + Zt.

    Then, Xt is stationary if

    |

    1 21 + 42

    2 |< 1.

    This requires

    1 + 2 < 1; 1 2 > 1; 2 > 1.

    The roots are real if

    1 + 42 0.

    Suppose conditions are satisfied. Then, the Yule-

    Walker equations are

    (0) =1

    (1) =1(0) + 2(1) = 1 + 2(1)(k) =1(k 1) + 2(k 2).

    25

  • 8/6/2019 stat420c3

    26/33

    Example 3.1. Consider the AR(2) process

    Xt = Xt1 1

    2Xt2 + Zt.

    Then,

    (B) = 1 B + 12

    B2.

    The roots are

    1

    1

    2

    2=

    1

    1

    2=

    1

    2 1

    2i.

    Therefore, it is stationary. By Yule-Walker equa-

    tions, we have

    (1) =(0) 12

    (1)

    (1) =11

    2(1)

    (1) =23

    .

    For other (k), we can use

    (k) = (k

    1)

    1

    2

    (k

    2).

    Another expression

    (k) =A1(1

    2+

    1

    2i)|k| + A2(

    1

    2 1

    2i)|k|

    =(1

    2)|k|(cos k

    4

    +1

    3

    sink

    4

    ).

    26

  • 8/6/2019 stat420c3

    27/33

    3.4.5. Mixed ARMA models

    An ARMA(p,q) processes is

    Xt = 1Xt1+ +pXtp+1Xt1+ +qZtq.It can also write as

    (B)Xt = (B)Zt

    where

    (B) = 1 1B pBp

    and

    (B) = 1 + 1B + + qBq.The process is stationary if

    (B) = 0

    and

    (B) = 0

    are outside of the unit disc on complex plane.

    27

  • 8/6/2019 stat420c3

    28/33

    Let (B) = (B)/(B). Then, we have

    Xt = (B)Ztwhich is a pure MA process.

    Alternative, let (B) = 1/(B). Then, we have

    (B)Xt = Zt

    which is a pure AR process.

    In general, the above expressions can be used in

    theoretical inference and are rarely used in appli-

    cations.

    28

  • 8/6/2019 stat420c3

    29/33

    Example 3.2: Find the and i weights for

    ARMA(1,1) model given by

    Xt = 0.5Xt1 + Zt 0.3Zt1.Solution: Let (B) = 1 0.5B and (B) = (1 0.3B). Then,

    (B) =B

    (B)

    =(1 0.3B)(1 0.5B)

    =(1

    0.3B)

    i=0

    0.5iBi

    =1 +

    i=0

    0.2 0.5i1Bi.

    Thus,

    i = 0.2

    0.5i1

    for i = 1, 2, . Similarly, we have

    i = 0.2 0.3i1

    for i = 1, 2, . We always have 0 = 0 = 1.

    29

  • 8/6/2019 stat420c3

    30/33

    3.4.6. Integrated ARMA (ARIMA) models

    An ARIMA(p,d,q) model is defined by

    (B)(1 B)dXt = (B)Ztwhere Zt W N(0, 2Z).

    If we write Wt = (1 B)d, then we have(B)Wt = (B)Zt

    and Wt is stationary under some conditions. How-

    ever, Xt is not stationary.

    30

  • 8/6/2019 stat420c3

    31/33

    3.4.7. The general linear process

    A general linear process is

    Xt =

    i=0

    iZti.

    If

    i=0

    |i| <

    then Xt is stationary.

    Clearly, MA(q), AR(p), ARMA(p,q)

    and AMIMA(p,d,q) are all linear process.

    31

  • 8/6/2019 stat420c3

    32/33

    3.4.8. Continuous processes

    Suppose xt is a continuous time series. Then,

    () = Corr(Xt, Xt+)

    is a function defined in (,).A continuous time series can be approximated by

    a discrete time series.

    32

  • 8/6/2019 stat420c3

    33/33

    3.5. The Wold Decomposition Theorem

    Consider the regression Xt on (Xtq, Xtq1, )and denote the residual variance by 2q .

    If

    limq q = V(Xt)

    then we call Xt purely indeterministic.

    If

    limq q = 0

    then we call Xt purely deterministic.

    The Wold Decomposition Theorem says: any dis-crete time stationary series can be expressed as

    the sum of two uncorrelated processes, one purely

    deterministic and another purely indeterministic.