+ All Categories
Home > Documents > M343 Unit 15

M343 Unit 15

Date post: 02-Jun-2018
Category:
Upload: robbah112
View: 233 times
Download: 3 times
Share this document with a friend

of 52

Transcript
  • 8/10/2019 M343 Unit 15

    1/52

    FOR

    REFEBENCE

    ONLY

    Mothemotics:

    A Third

    Level

    Course

    The

    Open

    University

    Unit

    l5

    Time

    series

    ffiffiffiffiffiffiffiffiffi

    W

    W

    n

    ffiffi

    fi

    ffi

    W

    W

    W

    r

    *Wffi

    WW

    ffieffiffi

    ffiffiws@wt

    ffiffi*ffimgws

    ffifffiffiffireffi

    wsE*&wtff

    rsffi*@ffi

    @re

    WW

    ffiEs ' GF6 ry

    WW

    WWffiffiW

    ffimffiffiw

    nnmffim

    ffi

    W

    W

    W

    n

    ffi

    W

    W

    ffi

    ffi

  • 8/10/2019 M343 Unit 15

    2/52

    M343

    APPLICATIONS

    OF PROBABITIW

    Mothemotics:

    A Third

    Level

    Course

    The

    Open

    Universifu

    Unit l5

    Time

    series

    Prepored by the

    Course Teom

    CONTENTS

    Introduction

    1

    What

    is

    a

    time

    series?

    1.1 Time

    series data

    1.2 Requirements

    for

    modelling

    time series

    2 Definitions

    of stationarity

    2.1 Strictly

    stationary

    processes

    2.2

    The

    autocorrelation

    function

    2.3

    Weakly

    stationary

    processes

    (Video)

    3

    Autoregressive processes

    of

    order I

    3.1

    Definition

    of

    an

    AR(1)

    process

    3.2

    Stationary

    AR(1)

    processes

    3.3

    Aspects

    of

    stationarity

    for

    autoregressive

    processes

    4 More on

    autoregressive

    processes

    4.1 Autoregressive

    processes

    of

    order 2

    (Audio)

    4.2

    General autoregressive

    processes

    5 Moving aYerage

    processes

    5.1 Moving

    average

    processes

    of order

    |

    (Video)

    5.2 Moving

    average

    processes

    of

    general

    order

    5.3

    The

    duality between AR

    and

    MA

    processes

    Objectives

    Appendix:

    Solutions to

    questions

    ?

    A

    +

    7

    9

    13

    IJ

    t4

    15

    t7

    20

    2l

    2l

    31

    31

    31

    34

    35

    37

    38

    The

    Open

    Universitv Press

  • 8/10/2019 M343 Unit 15

    3/52

    Statistics tables

    The

    recommended

    book

    of

    statistics

    tables

    for

    this course

    is

    H.

    R. Neave,

    Elementary

    statistics

    Tables

    (George

    Allen

    &

    Unwin, 1981).

    In

    this

    unit,

    these

    tables

    are referred

    to

    as Neaue.

    Unit

    titles

    1

    Random

    Processes

    2 Events

    in

    Time

    3 Patterns

    in

    Space

    4 Branching

    Processes

    5 Random

    Walks

    6

    Markov

    Chains

    7

    Birth

    Processes

    8

    Birth

    and

    Death

    Processes

    9

    Queues

    10

    Epidemics

    11

    More

    Population

    Models

    12

    Genetics

    13 Renewal

    Models

    14

    Diffusion

    Processes

    15

    Time

    Series

    16

    Problems, Problems,

    Problems,

    ...

    The

    Open

    University

    Press,

    Walton Hall,

    Milton

    Keynes.

    First

    published

    1988.

    Copyright

    O

    1988

    The Open

    University

    A11 rights reserved.

    No

    part

    of

    this

    publication

    may

    be

    reproduced,

    stored in a retrieval

    system

    or transmitted

    in

    any

    form or

    by any means,

    without written

    permission

    from

    the

    publisher.

    Designed

    by the

    Graphic

    Design

    Group

    of the Open

    University.

    Produced

    in

    Great

    Britain

    by

    Speedlith Photo

    Litho

    Limited,

    Longford

    Trading

    Estate,

    Manchester

    M32

    0JT.

    ISBN

    0 335

    14299

    0

    This

    text forms

    part

    of the

    correspondence

    element

    of

    an Open

    University

    Third

    Level

    Course.

    Ior

    general

    availqbility

    of supporting

    material

    referred

    to in

    this

    text,

    please

    write

    to:

    o_p_en

    U-niirersity

    Pducational

    Enterprises

    Limited,

    12

    cofferidge

    close,

    Stony

    stratford,

    Milton

    Keynes

    MK11lBY,

    Great Britain.

    Further

    information

    on

    Open University

    courses may

    be obtained from:

    The Admissions

    Office,

    The

    Open University,

    P.O.

    Box

    48,

    Milton

    Keynes MK7

    648.

    1.1

  • 8/10/2019 M343 Unit 15

    4/52

    ntroduction

    -

    .

    ::.:

    ::.irrst

    commoniy

    occurring

    forms

    of

    statistical

    data is the

    time

    series.

    slp1r.

    a

    time

    series

    is

    a sequence

    of

    observations

    made

    at

    regular

    time

    ri: :-

    \'tru have

    met

    such

    sequences

    before in

    the

    course,

    and elsewhere.

    For

    :,

    ::-,:.

    ::e

    ueekly

    chicago

    wheat

    prices

    discussed in[]nits

    1

    and 14

    form

    a time

    \ld

    the

    monthly

    UK

    unemployment

    figures

    form

    a

    time

    series,

    which is

    :-

    ::3:irred

    graphically

    in

    the news media.

    -

    --:

    :-':rstical

    anaiysis

    of data

    from

    time

    series is

    a

    topic

    of

    importance.

    The

    main

    --

    such

    an

    analysis is

    often

    to

    fit

    a model

    to

    the data

    that

    will allow

    forecasts

    ,

    :'. :-ade

    of future

    values

    in

    the

    series.

    In

    this

    unit

    some

    of the most

    important

    -,:rirstic

    models

    used

    for

    this

    purpose

    are

    introduced

    and

    developed;

    but

    it

    is

    .

    :,rssib1e

    in

    one

    unit to

    describe

    also

    how

    to

    fit the models

    to data,

    or how

    to

    ,:-m

    lor

    forecasting.

    However,

    after

    studying this

    unit,

    you

    should have

    a

    ';nderstanding

    of

    how

    the

    models

    work

    that

    will

    stand

    you

    in

    good

    stead

    if

    .

    -

    Jlr

    on

    to

    study

    time

    series model

    fitting

    and forecasting

    elsewhere.

    And

    the

    -::ls.

    as well

    as

    being useful,

    have some

    fascinating theoretical

    properties.

    S:;tion

    1 several

    examples

    of

    time series data are

    presented,

    and the

    kinds

    of

    that

    they show are

    discussed.

    The idea

    of a stationary time

    series-

    -

    -_llilv

    speaking, one

    that 'looks

    the

    same

    all

    the way

    along'-

    is

    introduced.

    of

    stationarity

    are developed in

    Section

    2.

    One

    of

    the key features

    of

    -:

    series data is

    that successive

    observations

    are usually correlated,

    and

    in

    3 we consider

    one

    kind

    of

    stochastic

    process,

    the autoregressive

    process

    of

    '::r

    1.

    which represents

    this

    dependence

    between successive observations

    in

    a

    simple

    way.

    In

    Section 4,

    the models

    of

    Section 3 are extended,

    and

    processes

    of

    order 2 and

    higher orders

    are defined. Another

    kind

    of

    used to model

    time

    series data

    -

    the moving

    average

    process

    -

    is

    in

    Section

    5, and

    the relationship

    between

    autoregressive processes

    and

    average

    processes

    is discussed.

    r

    carry

    out

    some

    of

    the calculations

    in

    this unit,

    you

    need

    to

    be

    able to

    solve

    first-

    and

    second-order

    recurrence

    relations

    with

    constant

    coefficients.

    You

    be

    familiar

    with

    the

    techniques

    required

    from

    (Jnits

    5

    and

    6;

    if

    you want

    to

    yourself

    of

    them

    you

    may

    find

    it

    helpful

    to

    read the

    Handbook

    entry

    on

    relations.

    Throughout

    this unit

    the ideas

    of stationarity

    and

    of

    stochastic

    processes

    are used;

    you

    met these

    first

    in

    Unit

    2,

    in

    context

    of

    point

    processes.

    And

    Section 2 contains

    references

    to

    an

    example

    unit

    6

    on rainfall

    in Melbourne,

    though

    you

    do not

    need

    to remember

    all

    the

    of this

    example

    in

    order

    to

    study this

    unit.

    1 is

    relatively

    short,

    while

    the

    other

    four

    sections

    are all

    about

    average

    in

    Band

    G

    of

    the video-cassette

    is

    associated

    with

    this unit.

    you

    will

    gain

    benefit

    from

    it

    if

    you

    watch it

    at

    the

    end

    of Section

    2,

    before

    the various

    time

    models

    are introduced,

    and again

    in

    Section

    5, after they

    have

    been

    studied.

    is

    also an

    audio-tape

    session;

    this is

    associated

    with Subsection

    4.1.

    What

    is

    a time

    series?

    time

    series is

    a sequence

    of observations

    made

    at

    regular intervals

    of time

    -

    erhaps

    every minute,

    or

    every

    week,

    or

    every

    year.

    The

    kind

    of

    stochastic process

    can

    be used

    to

    model

    time

    series

    data

    thus appears

    to

    be

    a

    process

    1,...) in

    discrete

    time,

    where

    the

    time

    unit

    is

    one minute,

    or

    one

    week,

    whatever

    length

    is

    appropriate

    for

    the data.

    what

    about the

    state

    space?

    It

    be reasonable

    for

    the

    state space

    of the

    process

    to

    match

    the

    data

    -

    if the

    consist

    of, say,

    temperatures

    taken

    hourly in

    a cold

    store, then

    a continuous

    space

    would

    seem

    appropriate,

    while

    if

    the data

    were counts

    of

    animals

    livinga certain place,

    a discrete

    state

    space

    might

    be

    better.

    In

    fact, in

    this

    unit

    (as

    in

    time

    series

    analysis)

    only

    a continuous

    state

    space is considered.

    If

    the

    data

    really

    discrete,

    the model

    will

    be

    only

    an approximation

    -

    but

    then,

    models

    always

    approximations

    Unit l,Example 4.10;

    Unit 14, Subsection

    1.1.

    In

    earlier units

    n

    has

    been used

    for

    the time variable in

    discrete-time

    stochastic

    processes.

    However,

    it is

    usual

    to

    use

    t for the time variable

    in time

    series models, so r will

    be

    used

    in

    this

    unit.

  • 8/10/2019 M343 Unit 15

    5/52

    units

    4,

    5, 6,

    13

    Unit

    15

    Continuous

    units

    2, 7, 8, 9, 10, 13

    Unit

    14

    To

    summarize,

    therefore,

    in

    this

    unit

    we shall

    investigate discrete-time,

    continuous

    state space stochastic

    processes

    that

    are

    useful

    for modelling time

    series

    data.

    And

    so the course

    has

    covered

    all four

    possibilities

    shown

    in

    the table below.

    1.1

    Time

    series data

    To

    progress

    further

    with

    the

    development

    of

    models,

    we

    must

    look at some

    time

    series

    data

    to

    investigate

    the sort

    of

    properties that

    the

    processes must model.

    This

    subsection

    consists

    largely

    of

    a number

    of

    plots

    of

    time series

    data which,

    taken

    together,

    illustrate

    most of

    the

    important

    features

    arising

    in such data.

    Figure

    1.1 shows

    the

    annual

    production

    of lumber

    in

    the

    USA

    ftom

    1947 to

    1976'

    Figure

    1.2 shows

    the annual

    population of the

    yellow-eyed

    penguin in Dunedin,

    New

    Zealand,

    from

    1937

    to

    1952.

    Lumber

    (millions

    of board

    feet)

    Pairs of

    penguins

    34000

    32000

    30000

    Though the

    values of

    the series

    in

    Figures

    1.1 and

    1.2 are defined

    onl;'

    for whole

    years

    -

    1947,

    1948, ar.d

    so on,

    in

    Figure

    1.1

    -

    in the

    graPhs

    the

    points corresponding

    to

    these

    years

    are

    joined

    with straight

    lines.

    This is a common

    convention

    for

    time

    series

    plots,

    and

    it

    will

    be

    used

    throughout

    this

    unit.

    i935

    1910 1945

    1950

    1955

    Year

    Figure 1.2

    Population of

    the

    yellow-eyed

    penguin

    (Megadyptes

    antipodes)

    in

    Dunedin,

    New

    Zealand

    (Source:

    MST204 Unit 3)

    100

    t946

    1956

    1966

    1976 Year

    Figure 1.1

    Annual US

    production

    of

    lumber

    (Source:

    B. Abraham and J.

    Ledolter, Statistical

    Methods

    for

    Forecasting

    (Wiley,

    1983)

    p.83)

    In both

    cases, the

    graphs

    appear

    to

    be

    moving

    irregularly up and down within a

    fairly

    restricted

    range.

    On

    the

    evidence of Figure

    1..2,

    for example, one would

    probably

    be

    prepared

    to

    predict

    that

    the

    penguin

    population

    in 1954

    would

    be

    somewhere

    near

    90

    pairs.

    Both

    these

    plots

    have the

    property

    that,

    roughly

    speaking,

    they have

    similar

    characteristics

    ('look

    the

    same') all

    the way along.

    Other

    timeseries

    do

    not look

    like

    these

    two.

    Figure

    1.3 shows

    the

    production of

    sulphuric

    acid

    in

    the

    UK,

    measured

    monthly,

    from

    July

    1983

    to November

    1987.

    Question

    1.1 Describe

    the

    pattern

    in

    the

    data

    of

    Figure

    1.3.

    Slow changes

    in

    the

    overall

    level

    of

    the

    process,

    like

    those

    in Figure

    1.3, are

    called

    trends.

    Another

    common

    pattern in

    time

    series

    is

    seasonal

    uariation.

    Figure

    1.4

    shows

    the

    monthly mean daily sunshine

    in

    Northern

    lreland, from

    January

    1976

    to

    June

    1986.

    The series

    moves

    up

    and down,

    but not

    entirely

    irregularly:

    there

    is

    a

    pattern

    related

    to

    the time

    of

    year,

    with

    sunshine

    hours

    not

    surprisingiy

    being

    high in

    summer

    and

    low

    in

    winter.

    However, this

    seasonal

    variation is

    not

    completely

    regular;

    there

    is random

    variability

    superimposed

    on

    the seasonal

    pattern. Other

    time series

    show seasonal

    variation

    in addition

    to

    overall trends

    in

    the

    level of the

    process

    -

    see,

    for

    example,

    Figure

    1.5

    which

    shows

    monthly

    car

    sales

    in

    Quebec

    from

    January

    1960

    to

    December

    1967.In

    Figure

    1.5

    the trend

    in

    the

    process

    appears

    to

    be a

    linear

    increase.

    70

  • 8/10/2019 M343 Unit 15

    6/52

    iuutrr"t.-:

    =trd

    r

    T '[lrilU[-jr1]

    ilm uLk-,:

    n mntr\

    l:,

    ;

    -

    llll

    ,

    1976

    1978

    198.1

    1986

    Year

    Figure

    1.4

    Monthly

    mean

    daily

    hours

    of sunshine

    in Northern

    Ireland

    (Source:

    Annual

    Abstract

    of Statistics,

    1987

    {HMSO)

    p.

    a

    )

    Price

    (f)

    4

    0

    f-.-,

    0

    1987

    Year

    100% acid)

    '

    lg15 lq76

    1t)77

    1978

    1979

    Year

    Figure

    1.6

    Monthly average

    wholesale

    price

    for

    10

    England

    and Wales

    produced Baccara

    roses

    (Source:

    Agriculture

    Statistics,

    England,

    1976-77 and

    1978

    79,

    Ministry

    of

    Agriculture.

    Fisheries

    and

    Food)

    1983

    1984

    1985

    1986

    Fgure

    1.3

    UK

    production of

    sulphuric

    acid

    (thousand

    tonnes,

    as

    S".urse:

    tr{onthly

    D,igest

    of Statis/ics,

    January

    1988

    (HMSO) p'

    57)

    Hours of

    f

    :u

    ns

    n

    il'lc

    I

    i0l

    I

    l

    8l

    I

    j

    6l

    1

    950

    1962

    t964

    1966

    Year

    Figure

    1.5

    N{orrthly car

    sales in

    Quebec

    (Source:

    Abraham

    and

    Lcdoiter,

    Statistic(il

    Methods

    for

    Forecasting,

    p.420).

    Questier*

    n.2 tr)escribe

    the

    pattcrn of

    the time

    series

    in

    Figure

    1.6, which

    shows

    the

    rnr-rrrflrly

    averilge

    wholesale

    price

    for

    10 red

    roses

    in English and

    Welsh

    whuJ;r;:ii+

    milrket

    irorrr

    January

    1975

    to

    December

    1979.

    tr

  • 8/10/2019 M343 Unit 15

    7/52

    $tofornnr,,-c

    r.-rd

    r

    Ilnrudlu-d'n

    mni{ltffi}i}:

    XJEITIb:\

    ll

    I

    _

    llr

    )

    .l

    i

    lfil

    i

    i

    l8f)-1

    I

    I

    t

    I

    i60

    FgEe

    13

    UK

    production of

    sulphuric

    acid

    (thousand

    tonnes,

    as

    100%

    acid)

    ,S..urce: Monthly

    D,igest

    of

    Stati.sfics,

    January

    1988

    (HMSO) p'

    57)

    1 983

    Hours of

    \unshine

    i{l

    1984

    1986

    Year

    Figure

    1.4

    Monthly

    mean

    daily

    hours of

    sunshine

    in

    Northern

    Ireland

    (Source:

    Annual

    Abstract of Statistics,

    1987

    (HMSO)

    p.

    a

    )

    Price

    ()

    4

    198,i

    1985

    1986

    1987

    Year

    "

    1975 1976

    197'1

    1978

    1919

    Year

    Figure

    1.6 Monthly

    average

    wholesale

    price for 10 England

    and Wales

    produced Baccara

    roses

    (Source:

    Agriculture

    Statistics,

    England,

    1976-77

    and

    1978-79,

    Ministry

    of

    Agriculture,

    Fisheries

    and

    Food)

    (t

    4

    Car

    sales

    I

    ,

    rhousands)

    |

    -301

    l

    j

    l

    2aI

    1964

    1966

    Figure 1.5 Monthly

    car sales

    in

    Quebec

    (Source: Abraham

    and

    Ledcriter. Statistical

    Methods

    for

    Forecasting,

    p.420).

    Questig*

    1.2 Describe

    the

    pattern of the

    time

    series

    in Figure

    1.6,

    which shows

    the

    motthlir

    average

    wholesale

    price

    for

    10 red

    roses

    in

    English and

    Welsh

    whcjl";rlai,j

    niarkf,ts

    from

    January

    1975

    to

    December

    1979.

    tr

  • 8/10/2019 M343 Unit 15

    8/52

    The

    patterns

    shown

    by

    the time series

    in Figures

    1.7,1.8

    and

    1.9 are

    particularly

    interestine.

    Number

    of lynx

    pelts

    (thousands)

    182()

    1840 i860 1880

    1900 1970

    1940

    Ycar

    t'igure

    1.7

    Variations in the

    population

    of lynx

    in

    the

    Hudson Bay

    area

    of

    Canada,

    as reflected

    by the number of

    pelts

    sold

    to a company each

    year

    (Source:

    MST204 Unit

    3)

    Number of

    lemmings

    (per

    hectrrc)

    1931

    1933

    1935

    1931

    [943

    Year

    Figure 1.8 Population density

    of lemmings

    in

    the area

    near

    Churchill, Canada

    (Source:

    MST204 Unit 3)

    1770 l7q0 l8l0 1830

    l8-5{

    Figure 1.9 Annual sunspot

    numbers,

    1770-1869

    (Originally pubiished by Wcilfer)

    1870

    Year

    Figures

    1.7 and

    1.8 show estimates

    of the

    populations

    of

    two

    mammals,

    lynx

    and

    iemrning,

    in

    parts

    of

    Canada,

    while

    Figure

    1.9 shows

    the

    sunspot

    numbers

    counted

    annually,

    according

    to

    a standard

    method,

    over

    a

    period

    of 100

    years.

    All these

    series

    exhibit

    an

    oscillatory

    variation

    over

    time,

    but

    the

    period

    of

    the

    oscillation

    does

    not

    appear

    to

    relate to the

    seasons

    of

    the

    year,

    for

    instance. Furthermore,

    the

    'peaks'

    of

    the oscillation are

    not

    always exactly

    the

    same number of

    time

    periods

    apart.

    For

    instance,

    in

    Figure

    1.9 the times between

    peak

    sunspot

    years

    range from

    7

    to

    l1

    years.

    Thus

    these series appear

    to

    show

    a different type of

    periodic

    behaviour

    from that

    in

    the series

    in

    Figures 1.4,

    1.5

    and

    1.6. In those series,

    the

    periodic

    variation

    is

    clearly

    tied

    to

    the seasons,

    and

    the

    peaks

    and troughs

    occur at

    ioughly

    the same

    times every

    year.

    Therefore,

    different

    kinds

    of

    modei

    are

    needed

    to

    deal

    with

    the

    two

    different

    kinds of

    situation.

  • 8/10/2019 M343 Unit 15

    9/52

    fi

    .

    "';irements

    for

    rnodelling tirne

    series

    -

    -

    ::

    --,c;ls

    for

    time

    series

    data,

    as

    in

    most rnodelling,

    it

    pays

    to

    start

    :.:i..rrv modeis

    can

    be developed

    for

    data

    iike

    those

    in Figure 1.1,

    '

    r.-

    ,:end and

    no

    seasonal

    variation,

    tlien

    it may be

    possible

    to

    extend

    ,

    -

    -:

    -:

    :.,

    take account of trend and seasonal

    variation.

    '

    -

    -

    :

    :' e

    similar

    characteristics

    aiong the entire time

    axis, like those

    in

    :

    :-3

    iometimes

    described as

    stationary"

    They

    can

    be modeiled by

    :

    :.--,-esses

    that

    are also,

    on

    average,'the

    same

    ali

    the way along the time

    -

    ,.

    -:-l.tic

    process

    defined

    in

    such

    a way

    that

    its

    properties

    do not

    vary

    *

    .

    s

    s:id.

    ioosely,

    to

    be stationary;

    definitions of what

    is meant

    by a

    *:r-

    :ltrCh&Stic

    process

    are developed

    in Section

    2.

    -

    :,-

    -i

    of one very simple

    kind of

    stationary

    process is illustrated in

    : ,.;

    In

    this simuiation, successive

    values

    of X,

    are

    independent

    -..

    :s iiom the

    same distribution

    in

    this

    case. a

    normal

    distribution.

    0501001

    Figure

    1.10

    A

    series of

    (simulated)

    independent

    observations

    from

    a normal distribution

    with

    mean 0 and

    variance 1

    ::'rre

    time series data can be

    modelled

    by

    a

    process

    as

    simple as this; but

    most

    are

    :r.rie

    complicated.

    Look

    againat Figure

    1.1.

    In

    the

    early

    1950s,

    the

    data'stuck'for

    . :eu'

    years

    around

    the 38 000

    level.

    In the

    late

    1960s, the

    data again

    remained

    ,:irl,v-

    constant for a

    few

    years,

    this time

    around 35000.

    In

    these data,

    as

    in

    many

    :iher

    time

    series, there

    is a tendency

    for

    successive

    values

    of the

    process

    to

    be

    ::iativeiy

    close

    to

    one another.

    Such a series

    will

    tend

    to

    look

    'smoother'

    than

    a

    .uccession of

    independent values:

    the differences

    between

    successive observations

    *il1

    be smaller

    in

    comparison

    with

    the

    variability

    of

    the

    series as a whole than

    *'ould

    be

    the case for independent observations.

    Therefore,

    if

    a

    process

    {X,}

    is to

    ce

    a

    suitable

    model

    for

    these data, then

    it must have the

    property

    that

    successive

    ralues

    of

    X,

    are correlated

    rather than independent.

    This

    type of correlation

    is

    :alled

    autocorrelation because

    it

    involves the

    correlation

    of

    part

    of the

    process

    with

    another

    part

    of

    itself.

    A

    sequence

    of independent

    random

    variables does

    not

    have

    any autocorrelation.

    Therefore

    a

    more complicated

    model is needed

    for

    a series

    like that

    in

    Figure

    1.1.

    Suitable

    processes

    that

    involve autocorrelation

    will

    be

    discussed

    in

    the remaining

    sections

    of

    this

    unit.

    What

    about

    models

    for

    time

    series that

    are

    not

    stationary,

    in that

    they exhibit

    trends or seasonal

    variation

    (or

    both)?

    It

    is beyond

    the scope

    of

    this

    unit

    to

    describe

    in

    any detail

    how to

    set

    up

    such

    models,

    but two

    basic

    approaches

    to

    the

    problem

    will

    be outlined.

    In

    the

    first,

    a

    series

    that

    exhibits

    trend and

    seasonal

    variation can be thought

    of as

    consisting

    of a

    slowly changing

    level,

    with

    perhaps

    a

    seasonal

    pattern

    superimposed

    on

    it,

    and

    with

    irregular variability

    superimposed

    on

    these

    two.

    It may

    be

    reasonable to

    model the values

    in

    such a series as

    a

    (deterministic)

    trend

    plus

    a

    (deterministic)

    seasonal

    pattern

    plus

    a

    stationaty

    irregular

    (stochastic)

    component.

    Once the

    trend and

    seasonal components

    have

    been

    removed, the

    irregular

    component

    can

    be modelled

    using one

    of

    the

    models

    Autocorrelation was

    introduced in

    Subsection

    4.3 of Unit

    2.

  • 8/10/2019 M343 Unit 15

    10/52

    for

    stationary

    series.

    This

    approach

    is

    illustrated

    in

    Figure

    1.11,

    which

    shows

    estimates

    of

    the

    deterministic

    trend

    and

    seasonal

    variaiion

    for

    the

    euebec

    car

    sales

    {1i1

    from

    Figure

    1.5,

    together

    with

    the

    remaining

    stationary

    irregular

    component.

    with

    other

    series

    it

    may

    be more

    appropriate

    to

    multiply

    the

    co#ponents

    of

    the

    series

    together.

    Car

    sales

    (thousands)

    0

    Seasonal

    vanatlon

    10

    0

    -l0

    Stationary

    lrregular

    componcnl

    10

    1060

    1962

    1964

    1906

    Year

    Figure

    1.I1

    Trend,

    seasonal

    and

    stationary

    irregular

    components

    of the

    euebec

    car sales

    series

    from

    Figure

    1.5

    Another

    approach

    to

    modelling

    non-stationary

    time

    series

    data

    involves

    looking

    at

    the

    differences

    between

    successive

    observations

    and

    fitting

    a stationary

    model

    to'

    those.

    Both

    these

    approaches

    to

    modelling

    non-stationa

    ry

    data

    involve

    doing

    something

    to

    the data

    -

    removing trend and

    seasonal

    components,

    or taking

    differences

    -

    nd

    then

    fitting

    a

    stationary

    model

    to

    the

    new

    series.

    So, even

    if

    tiuly

    stationary

    time

    series

    never

    occurred,

    it

    would

    still

    be

    important

    to

    develop

    models

    for

    them,

    and it is

    to

    such

    models

    that

    we

    now

    turn

    our

    attention.

    8

  • 8/10/2019 M343 Unit 15

    11/52

    I

    Definitions

    of stationaritv

    -

    :':,,'

    : ,

    :he

    basic

    idea

    of a

    stationary

    stochastic

    process

    was introduced;

    in

    "

    ".'"'

    ::, :,:rmai

    definitions

    of

    stationarity

    are developed.

    A

    stationary

    point

    rri

    "r'

    :

    ;*::

    defined

    inUnit

    2.

    In

    Subsection2.1

    this

    definition

    is

    extended

    to

    s:-r:--

    ,,:,-nastic

    processes

    in

    discrete

    time;

    the type

    of

    statiorrarity

    defined

    there

    .",:,r

    :r-"rr

    stationarity.

    Another

    type

    of

    stationarity,

    known

    asweak

    ;i

    ';'

    -..:s

    of more

    use

    in

    modelling

    time series;

    this type

    of

    stationarity

    ::i

    ::

    .:e

    .'oncept

    of

    the

    autocorrelation

    function

    of

    a

    process,

    which

    was

    iT

    -

    ,:-:.i

    lor point

    processes

    in

    unit

    2.

    Autocorrelation

    is

    discussed

    in

    more

    *:r

    i:--

    :-:rns

    in

    Subsection

    2.2,

    and

    weak

    stationarity

    is studied

    in

    Subsection

    2.3.

    I I

    Strictll'

    stationary

    processes

    :

    -

    :

    --.

    a

    point

    process

    was defined

    to

    be stationary if

    the

    distribution

    of

    the

    "r

    "-:.::

    -ri

    erents

    occurring

    in

    the

    interval

    (0,r]

    is

    the same as

    the

    distribution

    of

    ':

    -:r'rer

    of events

    in

    the interval(r,t

    *

    r],

    for all

    r

    > 0 and r > 0.

    That

    is,

    a

    r

    ::.rcess

    is stationary

    if its distributional

    properties

    are unaffected

    by

    shifting

    '

    : 1;

    aris by

    an amount

    r.

    An

    extension

    of this idea can be used

    to

    define what

    ,

    :

    .i

    1r

    as

    s/rlcf

    stationarity

    for

    general

    stochastic

    processes.

    A

    stochastic

    process

    . >.rid

    to be

    strictly

    stationary if

    the

    probability

    laws describing

    the shifted

    --

    :; >

    -Y,*,]

    are the same

    as those describing

    the

    original

    process

    {X,}.

    To

    be

    -

    -=

    :recise. consider

    the

    original

    process

    at

    a number

    of

    time

    points

    ty t2, ...,

    tn.

    '-:

    trocess is

    to

    be

    strictly

    stationary,

    then the

    joint

    distribution

    of

    the original

    r'

    -;i:

    ai these

    time

    points

    should

    be the same

    as the

    joint

    distribution

    of

    the

    -

    ..:i

    process

    at

    the

    corresponding

    shifted

    time

    points

    11

    +

    r,

    t2

    +

    r,

    ...,

    tn+

    r.

    -

    r .s stated

    formally

    for

    a discrete-time

    stochastic

    process

    in

    the

    followine

    : .::-:;tOn.

    Unlt

    2,

    Subsection

    1.2

    A similar

    definition can be

    given

    for

    proccsscs

    in

    continuous

    time.

    \

    stochastic process

    {Xr

    t

    :0,1,

    ...i

    is

    strictly

    stationary

    if

    the

    joint

    ::.trrbution

    of X,,,

    Xlrr,...,

    X,"

    is

    the same

    as ihe

    joint

    distribution

    of

    \,

    -,.X,r*,,

    ..",

    Xt,+,,

    for all

    positive

    integers

    n, aii

    (non-negative

    integer-

    .:lued)

    times

    rr,

    t2,

    ...,

    t,,

    and

    all

    positive

    integers

    r.

    :::..re

    investigating

    how

    this

    definition

    applies

    to

    the

    kind

    of

    stochastic

    process

    -.:i

    to

    model

    time

    series,

    we shall investigate

    how

    it

    applies

    to

    a simple process

    ,:rch

    you

    met

    earlier

    in

    the

    course.

    Erample

    2.1

    Melbourne

    weather

    '-:.-

    L'nit

    6 the sequence

    of

    dry

    and

    wet days

    in Melbourne

    was

    modelled

    by a two-

    ::.:te Markov

    chain,

    with

    states

    0 (wet)

    and

    1

    (dry), and

    transition matrix

    01

    o

    fo.azo

    o. r7l'l

    P: I I

    I

    l0.600

    i.400_l

    1t

    was

    shown

    that this Markov

    chain

    had a limiting

    distribution

    with

    probabilities

    -.o

    :0.175

    and

    n1

    :

    0.225

    (correct

    to

    three

    decimal

    places).

    It

    was

    pointed

    out

    ihat

    a finite,

    irreducible,

    aperiodic

    Markov

    chain

    (such

    as

    this

    one) has

    a

    unique

    stationary

    distribution,

    and

    that this

    is the

    same as the

    limiting

    distribution.

    The

    stationary

    distribution

    has

    the following property:

    if the

    initial

    state

    of

    the

    Markov

    chain

    has

    this

    stationary

    distribution,

    then the

    state

    distribution

    at

    every

    subsequent

    time is

    also the

    stationary

    distribution.

    In

    symbols, if

    the

    process

    is

    denoted

    by

    {X,; t

    :

    0,

    7,...},

    and

    if

    P(Xo:

    0)

    :

    no

    and

    P(Xo

    :

    1)

    :

    nt,then,

    for

    all

    integers

    t> 0, P(4:

    0)

    :

    z6 and P(X,: 1)

    :

    nr.Now, if

    n

    :

    1 is

    put

    in

    the

    definition

    of

    a strictly

    stationary

    process,

    then

    this

    definition

    implies

    that

    the

    distribution

    of

    X,

    must

    be the

    same

    for

    all

    integers

    r

    )

    0, so if the

    distribution

    of

    Xo is

    the stationary

    distribution,

    then

    this Markov

    chain looks

    as

    if

    it

    misht

    possibly

    be strictly

    staiionary.

    Unit 6, Example 1.4

    In

    this

    example

    X,

    is

    discrete,

    taking

    only the

    values

    0 and 1.

  • 8/10/2019 M343 Unit 15

    12/52

    I

    Definitions

    of

    stationaritv

    : :i,:-'

    :

    "

    :he

    basic idea

    of

    a stationary

    stochastic

    process

    was introduced;

    in

    '

    .

    d

    -

    : :rrma1

    definitions

    of stationarity

    are developed.

    A

    stationary

    point

    n

    -:

    ','

    defined

    in

    Unit

    2.

    In

    Subsection

    2.1

    this

    definition

    is

    extended

    to

    *i:-r:*r

    ':-;hastic

    processes

    in

    discrete

    time;

    the type

    of statiorrarity

    defined

    there

    , :'-

    r-:;i

    sttttionarity.

    Another

    type of

    stationarity, known

    as weak

    r

    .;

    'r-."...

    i.

    of

    more

    use

    in

    modelling

    time

    series;

    this

    type

    of stationarity

    lr

    :: .:-

    concept

    of

    thc

    autocorrelation

    function

    of

    a

    process,

    which

    was

    :

    -

    ,:,:=i

    1.rr

    point

    processes

    inunit

    2.

    Autocorrelation

    is

    discussed

    in

    more

    -::

    i

    :--

    r::rls

    rn

    Subsection

    2.2,

    and

    weak

    stationarity

    is

    studied

    in

    Subsection

    2.3.

    : I

    Strictll.'

    stationary

    processes

    r

    -

    '

    -'.

    a

    point

    process

    was

    defined

    to

    be

    stationary

    if

    the

    distribution

    of

    the

    r

    "-

    r,:.

    .ri

    events

    occurring

    in

    the

    interval

    (0,r]

    is the

    same as the

    distribution

    of

    ':

    --:-:er

    ofevents

    in

    the interval(t,t

    *

    r], for allt> 0 and

    r >0.

    That

    is,

    a

    r

    .:

    ::.rcess is

    stationary

    if its

    distributional

    properties

    are unaffected

    by

    shifting

    '

    :

    r---i

    aris by

    an amount

    r.

    An

    extension

    of this idea can be used

    to

    define

    what

    ,

    , .

    ,i

    n as

    stricf

    stotionarity

    for

    general

    stochastic

    processes.

    A

    stochastic

    process

    .

    .iid

    to

    be

    strictly st(itionary

    if the

    probability

    laws

    describing

    the shifted

    --

    -;s. -Y,,.].

    are

    the same

    as

    those

    describing the

    originai

    process

    {Xr].

    To

    be

    -

    -;

    rrecise.

    consider the

    original

    process

    at

    a number of time

    points

    tt,

    tz,

    ..., tn.

    ::

    process

    is

    to

    be

    strictly

    stationary,

    then

    the

    joint

    distribution

    of

    the

    original

    :- ,'s at

    these

    time

    points

    should be the

    same as the

    joint

    distribution

    of

    the

    ',:.1

    process

    at

    the corresponding

    shifted time

    points

    tl

    +

    r,

    t2

    +

    x,

    ...,

    t,

    +.t.

    :.:

    .s

    stated

    formally

    for

    a

    discrete-time

    stochastic

    process

    in

    the

    following

    -

    r

    .:-litLrn.

    Unir 2, Subsection 1.2

    A

    similar definition

    can be

    given

    for

    flrocesses

    in

    cont.inuous time.

    \

    stochastic process

    t

    X,; t

    :

    0,

    1,

    "..-|

    is

    strictly

    stationary

    if

    the

    joint

    :r.tribution

    of Xrr,

    Xr.,

    ...,

    X,,

    is

    the

    same

    zrs

    the

    joint

    distribution

    of

    \,

    -..X,r*,,

    ..., Xtn+,,

    for

    all

    positive

    integers

    n,

    aii

    (non-negative

    integer-

    .;lued)

    times

    /1,

    t2,

    ..

    .,

    tn, and

    all

    positive

    integers

    r.

    3:rt1re

    investigating

    how

    this

    definition

    applies

    to

    the

    kind

    of stochastic process

    ,:r'i

    to

    model

    time

    series,

    we shall

    investigate

    how

    it

    applies

    to a

    simple

    process

    i.

    :rr-h

    you

    met

    earlier in

    the

    course.

    Erample

    2.1

    Melbourne

    weather

    '-r-

    L-nit

    6 the

    sequence

    of

    dry

    and wet

    days

    in

    Melbourne

    was modelled

    by a two-

    ,--:te

    Markov

    chain,

    with

    states

    0 (wet)

    and

    1

    (dry), and

    transition matrix

    01

    o

    lo.sze

    o.

    r74l

    P:

    I

    t.

    /

    |

    0.600

    1.400

    ]

    Ii

    was

    shown tfrat

    tnls Markov

    chain had

    a limiting

    distribution

    with

    probabilities

    -o

    :

    0.775

    and zr

    :

    0.225

    (correct

    to

    three decimai

    places).

    It

    was

    pointed

    out

    that

    a finite,

    irreducible,

    aperiodic

    Markov

    chain

    (such

    as this one)

    has

    a unique

    stationary

    distribution,

    and

    that

    this is

    the same

    as the

    limiting

    distribution.

    The

    stationary

    distribution

    has

    the

    following property:

    if

    the

    initial

    state

    of the

    Markov

    chain has

    this

    stationary

    distribution,

    then

    the state

    distribution

    at

    every

    subsequent

    time

    is also

    the stationary

    distribution.

    In symbols,

    if the

    process

    is

    denoted

    by

    lX,;

    t

    :

    0,

    1,

    ...],

    and

    if

    P(Xo

    :

    0)

    :

    z6 and

    P(X.

    :

    l)

    :

    rr,rhen,

    for

    all

    integers

    t

    >

    0, P(Xt:0):

    no

    and

    P(.Xt: l):

    nr. Now, if

    n:

    1 is

    put

    in

    the

    definition

    of a

    strictly stationary

    process,

    then this

    definition

    implies

    that

    the

    distribution

    of

    x,

    must

    be the same

    for all

    integers

    /

    )

    0, so if

    the

    distribution

    of

    Xo

    is

    the stationary

    distribution,

    then

    this Markov

    chain looks

    as

    if

    it mieht

    possiblv

    be

    strictlv

    staiionarv.

    Unit 6,

    Example

    1.4

    In

    this

    example

    X,

    is

    discrete,

    taking

    only the values

    0 and

    1.

  • 8/10/2019 M343 Unit 15

    13/52

    Question

    2.1

    Could this Markov

    chain be

    a

    strictly

    stationary

    process

    if the

    distribution of

    the

    initial

    state were

    not

    given

    by no

    and

    nr?

    n

    Showing

    that

    the

    definition

    is

    satisfied when n

    :

    1 is

    not

    sufficient

    for

    the

    process

    to be strictly

    stationary; the

    definition

    must

    be satisfied

    for all

    n.

    A

    complete

    proof

    of this will not

    be

    given

    here; instead,

    the

    principles

    underlying the

    proof

    will

    be

    sketched.

    Let

    t, t

    t

    I

    and

    t

    +

    i

    +7

    be

    three

    integer

    times

    in

    increasing

    order

    (so

    that

    I and

    j

    are

    positive

    integers). Then,

    if this

    process

    really is

    strictiy

    stationary,

    it

    must

    be

    the case that,

    for

    example,

    P(Xt+t+:

    :

    l, Xt+i:

    0,

    X,

    :

    1)

    :

    P(.Xt+i+

    j+,:

    I,Xt+i+,:0,Xra":

    I),

    (2.1)

    for

    any

    positive

    integer

    r.

    Are these

    two

    probabilities

    really equal?

    Consider

    first

    the

    probability

    on

    the left-

    hand side. It is

    the

    probability

    of the

    following

    event:

    that

    the

    process

    is

    in

    state 1

    on

    day

    /,

    then

    in

    state 0 I days later, and

    finally

    in

    state

    1

    after another

    j

    days.

    Since the

    process

    is a

    Markov

    chain, this

    probability

    is equal to

    P(X,:

    I)

    x

    P(X,*t:01X,:

    1)

    x

    P(Xt+i+i:

    llX,*;:0)

    :

    P(X,:

    t)pf).p8)'.

    Similarly, the

    right-hand

    side of

    Equation

    (2.1)

    can

    be

    written

    as

    P(X,

    -,

    :

    ll

    P'ito

    P'd\

    But,

    as

    you

    have

    already

    seen,

    X,

    has the same

    distribution

    for

    all

    r, so

    P(X,*,: 1)

    :

    P(X,: 11

    for

    any

    positive integer

    t.

    Therefore,

    the

    two

    probabilities

    in

    Equation

    (2.1)

    arc equal, and we have

    P(Xt+i+:: l,Xt+i:

    0,Xr: 1)

    :

    P(Xt+i+

    j+t:

    l,Xt+i+t:0,Xra,:

    l),

    for

    any

    positive

    integer

    r.

    Now

    the sequence

    of

    states

    101

    in

    Equation

    (2.1)

    could be replaced

    by any other

    sequence

    100,

    for

    example.

    So a

    condition

    similar to

    Equation

    (2.1)

    holds for

    any sequence

    of three states. In

    order

    to

    prove

    that

    the

    process

    is strictly

    stationary,

    it

    is necessary

    to

    extend

    the argument

    to

    cover sequences of any

    number

    of states,

    not

    just

    three; and this

    can be

    done

    by

    a

    similar

    argument

    to

    that

    used above.

    So this

    Markov

    chain is strictly

    stationary.

    tr

    If a

    stochastic

    process

    {X,}

    is

    strictly stationary, then the distribution

    of

    X,

    is the

    same

    for all

    integers

    /,

    so the

    mean

    of

    X,

    is

    the same for all r;

    that

    is,

    E(X):p

    fort:0,1,....

    Question

    2.2

    Calculate

    p

    for

    the Melbourne

    weather

    process,

    given

    that

    the

    initial

    distribution

    is the

    stationary distribution.

    tr

    In statistical

    terms, as

    you

    saw

    in

    Section 1,

    what

    makes a time

    series

    different

    from

    a

    random

    sample

    is

    that

    successive

    observations

    are

    correlated.

    How

    can

    this

    correlation be

    measured? IfX and Yarc

    any

    two

    random

    variables, the

    degree of

    relationship

    between

    them can

    be

    quantified

    by their covariance:

    C(x,Y): El(X

    -

    p)(Y- p)l:

    E(XY)

    -

    Hxtty,

    where

    p"

    and

    trt,

    are the expected

    values of

    X and Y

    Therefore

    the relationship

    between successive

    observations

    Xrand

    X,*,

    in

    a time series

    (or

    any other

    stochastic

    process)

    can be measured

    by

    the covariance

    of these random

    variables:

    C(XbXt+):

    El(X,

    -

    H*,)(Xr*r

    -

    ltx,*,)1.

    Now, in

    general,

    the covariance

    of, say, Xrand

    Xrneed

    not

    be the

    same as

    that

    of, say,

    Xro

    and Xrr.

    But

    suppose

    that

    the

    stochastic

    process

    {X,}

    is

    strictly

    stationary. Then

    the

    joint

    distribution

    of

    X,

    and X, is

    the same

    as

    that

    of

    Xro

    and

    Xrr,

    so

    that

    the

    covariance

    of

    X,

    and

    X,

    is the

    same as

    that

    of

    Xro

    and

    X,

    1in

    The sequence of states 1 0 1

    is

    being

    used merely as an example to

    demonstrate the ideas involved in

    ;

    proof

    of stationarity: there is

    nothing special about it. Another

    sequence, such as 00

    1,

    would

    have

    done

    just

    as well.

    10

  • 8/10/2019 M343 Unit 15

    14/52

    ffi

    m" Thus

    the

    covariance

    of

    X,

    and Xr*,

    is the

    same

    for

    all integer

    varues

    of

    iff}:fllh

    ir

    n usual

    to

    write

    this covariance

    as

    C|X"X'*t)

    :7r.

    ft*crip

    I represents

    the fact

    that

    the random

    variables

    on

    the

    left

    are

    one

    ddti'.e apart;

    this

    difference

    is

    called

    the tag. The

    covariance y,

    is

    called

    an

    ffiry[ri'r4ce

    because

    it involves

    the

    covariance

    between

    randcm

    variables

    from

    ,k

    parts

    of

    the

    same

    process,

    and

    hence

    y,

    is

    called

    the autocovariance

    at

    h

    f

    of

    the process

    {X,}.

    h$

    Zl

    Melbourne

    weather,

    continued

    mm

    ft

    Melbourne

    weather process,

    made

    station

    ary by

    choosing

    the appropriate

    'rkrfrution

    for

    Xs,

    the

    value

    of

    y,

    is

    given

    by

    "y1:

    C(XyX2):

    E(X:X)

    -

    E(X)E(XI).

    fu

    frxr)

    and

    E(Xr)

    are both

    equal

    to

    ,u,

    the mean

    of

    the

    process,

    which

    you

    ffi

    to

    &

    0.225 in

    Question

    2.2,

    so

    i'r:E(XtX)-0.2252.

    ,lr

    I,

    and X,

    can each

    take

    the

    values 0 and 1,

    the

    product

    XrX,

    can also

    take

    fr

    ralues

    0

    and

    1,

    and

    it

    takes

    the value

    1 only

    when X,

    :

    Xz

    :1.

    Therefore

    E(X$): I

    x

    P(X1:

    1,Xz:

    1)

    +

    0

    x P(X,

    and

    X2

    not

    both

    1)

    :

    P(Xr:

    l,Xz: l)

    :

    P(Xz:

    llXr:

    1)P(x1

    :

    1).

    Frm

    the

    transition matrix

    for

    the

    Markov

    chain, we have

    P(X2

    :

    llxr:

    1)

    :

    p,.,

    :0.4,

    ul since

    the chain is

    stationary,

    P(X1

    :1):nr:0.225.

    llence

    and

    E(Xd):

    0.4

    x

    0.22s

    :

    0.09

    7r

    :

    0.09

    -

    0.2252

    :

    0.039.

    tr

    The covariance

    that

    has

    just

    been calculated

    is called

    the

    autocouariance

    at

    tag 1

    of

    the process

    {x,}.

    other

    autocovariances

    can be

    defined

    in

    a

    similar

    way.

    Suppose

    that

    {X,;

    t:0,1,...}

    is

    a

    strictly

    stationary process.

    Then

    the

    joint

    distribution

    of

    -\

    and

    X,*.

    is

    the same

    as

    the

    joint

    distribution

    of

    Xo

    and

    X" for

    any

    positive

    integers

    t and.c,

    so

    that

    C(Xt,Xt+"):

    C(Xs,X")

    for

    all

    positive

    integers

    t

    and r.

    In

    other

    words,

    if

    r,

    and

    t2areintegers

    with

    tz) tt

    >

    0,

    then

    the

    covariance

    between

    x,,

    and

    Xr,

    does

    not

    depend

    directly

    on

    the

    values

    of

    r,

    and /r,

    but

    only

    on

    the

    difference

    tz

    -

    tt

    between

    them.

    This

    difference

    is

    called

    the lag.

    Thus

    if

    we

    write

    C(Xe,

    X,)

    :

    y"

    for

    all

    positive

    integers

    z, then

    C{X^,

    Xe)

    :

    ltz-tt.

    (2.2)

    The

    covarianca

    "-

    c(xo,x") is

    called the autocovariance

    function

    at

    lag r

    of the

    process

    {Xr}.

    (The

    quantity

    y"

    is called

    the autocovariance

    fuinction

    because it can

    be

    thought

    of as a

    function

    of

    the lag

    z.)

    So

    far,7"

    has

    been

    defined

    only

    for

    .c:1,2,...,

    but

    it is

    useful

    to extend

    the

    deflnition

    to non-positive

    integers.

    If

    r

    :

    0 is

    put

    in

    the definition

    above,

    then

    we

    obtain

    c(xo,xd: ys,

    but

    the

    covariance

    of

    a

    random

    variable

    with

    itself is

    just

    its

    vadance,

    so that

    lo:

    v(xd.

    Since

    {x,}

    is

    a

    strictly

    stationary process,

    we

    have

    V(X,)

    :

    V(Xo)

    :

    yo

    for r

    :

    0, 1,

    ...

    .

    What

    about

    negative

    values

    of c?

    Consider

    the covariance

    C(Xn,Xr). Attempting

    to

    use Equation (2.2)

    to find

    this

    covariance

    in

    terms

    of

    the

    autocovariance

    function

    leads

    to

    C(X4,X):Tz-+:t-2.

    But

    C(Xo,Xr):

    C(X2,Xa):

    lz,

    so

    ?-z

    :

    y2.

    And,

    in

    fact,

    |

    -":

    l,

    for

    all

    integers

    r.

    The

    prefix'auto'has

    Greek

    origin

    and means

    'self'.

    p

    _

    o[0.826

    0.t741

    I

    10.600

    0.4001

  • 8/10/2019 M343 Unit 15

    15/52

  • 8/10/2019 M343 Unit 15

    16/52

    1r

    ..,.

    {

    *--i

    :1e

    covariance

    of

    X,

    and X,*, is

    the same

    for

    all

    integer

    values

    of

    -:-ri

    :'.

    n'rite

    this covariance

    as

    -

    i 1 \-.,

    'l'

    w r";*::

    -

    represents

    the fact

    that

    the random

    variables

    on

    the

    left

    are one

    lr:

    *:

    ::art:

    this

    difference

    is called

    thelag. The

    covariance y1

    is called

    an

    ,iitrru,r

    -,r"

    j:-i'oecause

    it

    involves

    the

    covariance

    between

    random

    variables

    from

    ri

    iri-r:'-

    r:rli trf

    the

    same

    process,

    and

    hence

    7,

    is called

    the autocovariance

    at

    lilm

    '::;crcess{Xr}.

    il,rrurnmrnul

    1l

    \Ielbourne

    weather,

    continued

    '

    -

    1;

    \f

    li'rourne

    weather

    process,

    made

    stationary

    by

    choosing

    the appropriate

    ".

    ,

    -

    -

    -:

    :

    ior Xo,

    the

    value

    of

    7,

    is

    given

    by

    :

    C(XrXz):

    E(XrXz)

    -

    E(X)E(Xr).

    '*

    u

    :

    f

    _

    '

    and

    E(X)

    are both

    equal

    to

    p,

    the

    mean

    of

    the

    process,

    which

    you

    r-n: .,

    :e

    0.225

    in

    Question

    2.2, so

    .:E(X6z\-0.22s2.

    ,

    .

    .:.1

    ,Y:

    can

    each take the values 0 and 1, the

    product

    XrXrcan

    also take

    ,r

    :

    -is

    0 and

    1,

    and it

    takes the value

    1

    only when

    Xr:

    Xz:1. Therefore

    Et-Y,Xr):

    1

    x P(X,

    : l,Xz:

    1)

    +

    0 x

    P(X, and X,

    not

    both

    1)

    :

    P(Xr

    :

    l,Xz: 1)

    :

    P(Xz: llXt

    :

    1)P(X1

    :

    1).

    ;:

    -

    :

    ::e transition

    matrix

    for

    the

    Markov

    chain.

    we

    have

    prX,

    :11X, :

    t):

    prr:

    e.4,,

    *ri r :tlJe

    the

    chain is

    stationary,

    P(X1: l):

    nt:0.225.

    -

    :

    ':-

    E(XJ):0.4

    x

    0.225

    :0.09

    ...:

    I

    r

    :

    0.09

    -

    0.2252:

    0.039.

    n

    --.r

    "Lrvariance

    that

    has

    just

    been calcuiated

    is

    called the autocouariance

    at

    lag 1

    of

    -:3rocess

    {X,}.

    Other

    autocovariances

    can be defined

    in

    a

    similar

    way.

    Suppose

    '-

    :i .Y,:

    r

    :

    0,

    1, . . .

    ]

    is

    a strictly

    stationary

    process.

    Then the

    joint

    distribution

    of

    i .nd,Y,*,

    is

    the

    same as the

    joint

    distribution

    of

    Xo

    and

    X, for

    any positive

    :..::ers

    t and

    r,

    so

    that

    C(Xt,Xt+"):

    C(Xs,X")

    for all

    positive

    integers

    t and r.

    :.

    ..ther

    words,

    if

    r,

    and

    t,

    are integers

    with

    tz) tt

    >

    0, then

    the

    covariance

    :':.'''\een

    X,,

    and

    X,,

    does

    not

    depend

    directly

    on

    the

    values

    of /,

    and

    rr,

    but

    only

    :. the

    difference

    t2

    -

    r,

    between

    them. This

    difference is called

    the lag.

    Thus

    if

    we

    ,,:te

    C(Xo, X,):

    y, for

    all

    positive

    integers

    r,

    then

    C(Xt,,

    X,)

    :

    ltz_,t.

    (2.2)

    The

    prefix'auto'has

    Greek origin

    and means 'self'.

    p

    _

    o[0.826

    o.t74l

    I

    10.600

    0.400

    1

    The

    covarianc

    )"

    -

    C(Xo,X,)

    is

    called the autocovariance

    function

    at

    lag

    r

    of

    the

    rrt'rcSS

    {X,}.

    (The

    quantity

    7.

    is

    called

    the autocovariance

    function

    because

    it

    can

    r:

    thought

    of as

    a

    function

    of the lag

    r.)

    So

    far,

    y,

    has

    been

    defined

    only for

    r:

    ,2,...,

    but

    it

    is useful

    to extend

    the

    lefinition

    to

    non-positive

    integers.

    If r

    :

    0

    is

    put

    in

    the

    definition

    above,

    then

    we

    ,rbtain

    c(xo,xo):

    yo,

    but

    the covariance

    of

    a

    random variable

    with

    itself is

    just

    its

    iariance,

    so

    that

    lo:V(Xo).

    Since

    {X,}

    is a

    strictly

    stationary

    process,

    we have

    V(.X)

    :

    V(Xd

    :

    yo

    for /

    :

    0, 1,

    ... .

    What

    about

    negative

    values

    of

    r?

    Consider the

    covariance

    C(Xa,Xr).

    Attempting

    to

    use Equation

    (2.2)

    to

    find

    this

    covariance

    in

    terms

    of

    the autocovariance

    function

    leads

    to

    C(X4,X)

    :

    niz-+:

    I,z.

    But

    C(Xo,X)

    --

    C(X2,X+):

    lz,

    so

    I-z

    :7r.

    And,

    in

    fact,

    .i

    _":

    l,

    for

    all

    integers

    r.

  • 8/10/2019 M343 Unit 15

    17/52

    Question

    2.3 Find

    the

    values

    of

    ys

    and

    y-1for

    the

    (stationary)

    Meibourne

    weather

    process.

    n

    Example

    2.1

    Melbourne

    weathero

    continued

    To find

    the

    values

    of

    7"

    for ali

    integers

    r

    for

    the Melbourne

    weather process,

    we

    begin

    by using

    the

    same

    approach

    that

    was

    used to flnd

    yr^

    For

    positive

    integers

    r,

    y,:

    C(X6,X")

    :

    t'(X.

    X")

    *

    E(X}E(X"),

    and

    E(xo)

    and

    E(x,)

    are

    both

    equal

    to

    the

    process

    mean,e.225.

    Just

    as before,

    E(X,X"):

    I

    x

    P{XoX,:1)

    *0

    x

    p(Xo&:0)

    :

    p(Xo

    X,:

    I)

    :

    P(X": 1'Xo:

    1)

    :

    P(X,:

    llxo:

    1)P(Xo:

    1).

    Now

    P(X,:

    llxc

    :

    1)

    is

    the

    probability

    that

    the

    process

    will return

    to

    state

    1

    at

    time

    r

    given

    that

    it

    started there

    at

    time 0; this

    return

    probabiiity

    is

    written

    pf)r.

    Denoting

    the

    n-step

    transition

    matrix

    for

    the

    process

    by

    P('),

    and

    using

    the fact

    that

    P('+1)

    -

    P(n)P,

    it

    can be

    shown that,

    for

    n:0,1,.".,

    pli')

    :0.226pP'

    +

    a.174.

    This is

    a

    first-order

    recurrence

    relation

    of the form

    ilnrt

    :

    cun

    * d

    for n

    -0,

    1,...,

    which has the

    solution

    d(l

    -

    c")

    u":

    1_ c

    4-

    t"Llo,

    prcrvided

    c

    11.

    Comparing

    the recurrence

    relations

    (2.3)and (2.4),

    we have

    u":

    pY1,

    c

    :

    Q.226 and

    d

    :

    0.114.

    To

    use

    Formula

    (2.5)

    to

    write

    down

    the

    solutron

    of

    the

    recurrence

    reiation

    (2.3),

    we need

    uo; this is

    p ol,

    which is

    equal

    to

    1 by

    See Unit

    6,

    page

    l0

    definition. Therefore

    0.t74(1_0

    :?61

    _L

    o

    ))^n

    P.i,i:u,

    _0.226

    4.174

    :

    ffi(1

    -

    0.226\

    +

    0.226'

    :

    0.225

    (l

    -

    0.226\

    +

    0.226

    :0.775

    x

    0.226'

    +

    0.225.

    Question

    2.4

    Show thar,

    for

    positive

    lags

    z. the value

    of the

    autocovariance

    function

    l,

    for the

    Meibourne

    weather process

    is

    given

    b;r

    T,:

    4.775

    x

    0.225

    x

    0.226'.

    Check that

    this

    formula

    gives

    1,

    :

    0.039,

    as already

    caiculated.

    fl

    From

    Solution

    2.3,

    I'o

    :

    0.225

    -

    A.2252

    :

    A.225(l

    -

    0.225):

    A.225 x

    A.775,

    so the

    formula

    for

    y, given

    in

    Question

    2.4 also

    works

    for

    t

    :

    0.

    And

    since

    1t":l

    ,,

    this

    formula

    can be extended

    to

    cover

    all

    (integer) lags,

    including

    negative

    lags,

    by

    writing

    'it":

    0.775

    x

    0.225

    x

    0.2261"t,

    where

    lrl

    denotes the

    absolute

    value of r.

    n

    {2.3)

    See Unit 6, Solution 2.5;

    a stmllar

    argument is required

    to obtain

    this

    recurrence relation.

    (2.4)

    (2

    5)

    See

    Unit 6,

    page

    17.

    This

    method

    was used

    tn Unit

    6 to

    derive

    the stationary probabilities

    zo

    and z, for

    a two-state

    Markov

    chain. As

    n+

    a), the return

    probability

    p(tl

    4

    0,225,

    the

    value

    of n1.

    (.2.6)

  • 8/10/2019 M343 Unit 15

    18/52

    Question

    2.3

    Find

    the

    values

    of

    yo

    and

    y_,

    for

    the

    (stationary)

    Melbourne

    weather

    process.

    tr

    Example

    2.1

    Melbourne

    weather,

    continued

    To

    find

    the

    values

    of

    7"

    for

    ali

    integers

    r

    for

    the Melbourne

    weather process,

    we

    begin

    by

    using

    the

    same

    approach

    that

    was

    used to find

    yr.

    For

    positive

    integers

    r,

    1t,:

    C{Xs,X")

    :

    E(x.

    X,)

    -

    E(X}E(X,),

    and

    E(xo)

    and

    E(x,)

    are

    both

    equal

    to

    the

    process

    mean,0.225.

    Just

    as before,

    E(XyX,)

    :

    I

    x

    P(Xo

    X,

    :

    1)+

    0

    x

    P(X,X,

    :

    0)

    :

    F(Xo X.

    :

    1)

    :

    P(X,:

    1,Xo:

    1)

    :

    P(X,:

    liXo:

    1)P(Xo:

    1).

    Now

    P(x.

    :

    llxc

    :

    1)

    is

    the

    probability

    that

    the

    process

    will return

    to

    state

    1

    at

    time

    r

    given

    that

    it started

    there

    at time

    0; this

    return probabiiity

    is

    written

    pf)r.

    Denoting

    the

    n-step

    transition

    matrix

    for

    the

    process

    by

    P('),

    and

    using

    the fact

    that

    P('+1):

    P(nrP,

    it

    can be

    shown that,

    for

    n:0,1,...,

    Pfrnt):

    8-226pPt

    +

    a.174.

    This is

    a

    first-order

    recurrence relation

    of the form

    un+1

    :

    cu^

    I tl

    for

    n:0,

    1,...,

    which has the

    solution

    d(I

    -

    c')

    un:-,

    1

    C,uo.

    I

    -('

    provided

    c

    11.

    Comparing

    the recurrence

    relations

    (2.3)and

    (2.4),

    we have

    un:

    pYt,

    c

    :

    Q.226 and

    d

    :

    0.114.

    To

    use

    Formula

    (2.5)

    to

    write

    down

    the

    soluiron

    of

    the

    recurrence

    reiation

    (2.3),

    we need

    ao; this is

    pt0J,

    which is

    equal

    to

    1 by

    definition.

    Therefore

    0.t74(1

    -

    0.216')

    p'{'i

    :u,:-

    I

    _

    02N'

    +0.226"

    4.174

    :

    afr(I

    -

    0.226")

    +

    0.226'

    :

    0.225

    (l

    *

    0.226)

    +

    0.226

    :0.775

    x

    0.226,

    +

    0.225.

    Question

    2"4

    Show that,

    for

    positive

    lags

    r,

    the value

    of the autocovariance

    function

    y"

    for the Metbourne

    weather process

    is

    given

    bv

    "1":

    4.775

    x

    0.225

    x

    0.226'.

    Check that

    this

    formula

    gives

    i,,

    :

    0.039,

    as alreacly

    caiculated"

    n

    From

    Solution

    2.3,

    ta

    :0.225

    -

    A.2252

    :

    A.225(.1

    -

    0.225\

    :

    A.225 x

    0.775,

    so

    the

    formuia

    for

    ir,

    given

    in

    Question

    2.4 also

    works

    for

    r

    :

    0.

    And

    since

    )t":1,,,

    this

    formula

    can be extended

    to

    cover

    all

    (integer)

    lags,

    including

    negative

    lags,

    by writing

    "t,

    :

    A.l7

    5

    x

    0.225

    x

    0.2251"t,

    where

    lri

    denotes the

    absolute value

    of t.

    n

    (.23)

    See Unit 6, Solution 2.5;

    a similar

    argument is required

    to obtain

    this

    recurrence relation.

    (2

    4)

    See

    Unil 6,

    page

    17

    See Unit

    6,

    page

    10.

    This

    method

    was used in

    Unit

    6

    to

    derive

    the stationary probabilities

    no

    and z, for

    a two-state

    Markov

    chain. As

    n4

    u:),

    the return

    probability

    pY)t

    r

    0.225,

    the value

    of nr.

    (2

    6)

  • 8/10/2019 M343 Unit 15

    19/52

    T'

    The

    autocorrelation

    function

    :.:rgating

    the

    relationship

    between

    random

    varrables'

    it

    is sometimes

    -itoco,rsiaertheircorrelationthantheircovariance.Thisisthecasefor

    'iiabies

    associated

    with

    stochastic

    processes,

    particularly

    for

    those-used

    '':me

    series.

    The

    correlation

    p(X,Y)

    of

    two

    random

    variables

    X and

    Yis

    C8.Y\

    .\

    Yt:

    u/lvWvrvt)

    : i

    :

    0,

    1,

    . . .)

    is

    a stochastic

    process

    in

    discrete

    time'

    then

    the

    correlation

    ,In

    and

    X"

    is

    given bY

    c(xo.

    x,)

    .,t

    .\

    ^.

    X,)

    u/1v1xol

    vtx,t)

    -

    .

    rrLrcess

    is strictly

    stationary,

    then

    C(Xo,

    X")

    :

    y",

    and

    V(Xl

    :

    V(X")

    :

    )o'

    so

    :

    -

    -:ielation

    between

    Xn and

    {

    is equal

    to

    y",fy6'And

    then'

    since

    the

    process

    is

    --:.-.

    stationary,

    the

    corielation

    between

    X,aid

    X,*"

    is also

    given-by

    7'/y6'

    for

    l::-ger,>0.Thrscorrelationisknownastheautocorrelationfunctionofthe

    -: s

    at

    lag

    t,

    and

    is

    denoted

    bY

    P,,

    so

    (21)

    ['t

    -

    ltl

    lO'

    .,

    :,-,rrelationJunctionis

    often

    abbreviated

    to

    acf'

    but

    not

    in

    this

    course')

    1,1,rt:don

    2.5

    Find

    the

    autocorrelation

    function

    p"

    of

    the

    Melbourne

    weather

    --:;utocorrelationfunction,andestimatesofitfromdata'areofkeyimportance

    -

    .:e

    modelling

    and

    analysis

    of

    time

    series'

    :"3

    Weakly

    stationarY

    Processes

    S:

    tar

    in

    this

    section we have

    concentrated

    on

    one

    particular definition of

    .."lionarity

    -

    strict

    stationarity'

    Not

    surprisingly'

    there

    is-another

    type

    of

    .::rionarity

    called

    weak

    stationarity.

    A stochastic

    process

    {xt}

    it

    weakly

    stationary

    .-

    rhe

    mean

    of

    X,

    does

    noi

    a"p.nO

    t"

    t

    u"d

    the

    covariance

    between

    X"

    and

    X"

    :.p.nd,

    only

    onthe

    lag

    tz-'tt'

    This

    definition

    is

    stated

    formally

    below'

    A

    stochastic

    process

    {Xr;

    t

    :0,

    1,

    . .

    .}

    is weakly

    stationary

    if

    it satisfies

    the

    followine

    two

    conditions.

    E(X,)

    takes

    the

    same

    value,

    p,

    for

    all

    r'

    For

    all

    non-negative

    integers

    t1

    and

    t2,the

    value

    of

    C(Xt1,Xt2)

    depends

    only

    on

    the

    lag

    tz-

    tr

    The

    autocorrelation

    function

    for

    a

    stationary

    point

    process

    was

    defined

    rn

    Unit

    2,

    Subsection

    2.3'

    The

    definition

    given

    here

    for

    general

    orocesses

    does

    not

    correspono

    ixactlv

    to

    the

    deflnitionin

    Unit

    2,

    becauie

    that

    deflnition

    had

    to

    take

    account

    of

    the

    time

    interval

    over

    which

    the

    Process

    was

    observed

    on

    each

    occasron.

    There

    are

    several

    important

    points

    about

    this

    definition'

    First,

    Condition

    2 applies

    t o

    the

    case

    t

    r

    :

    t

    z

    :

    t,

    in

    which

    case

    the

    covariance

    is simply

    the

    variance

    of

    X"

    and

    the

    definition

    says

    that

    this

    variance

    is

    the

    same

    for

    all

    t'

    That

    is,

    the

    variance

    of

    the

    process

    is

    constant

    as

    well

    as

    its

    mean.

    This

    is clearly

    a

    basic

    requirement

    for

    a

    definition

    of

    stationurity'

    S""o"aly,

    if the

    covariance

    C(Xo'X')

    is

    denoted

    by

    1'.,

    then

    it

    must

    be

    true

    for weakly

    stati'onary

    processes

    (as

    for

    strictly

    stationary

    processes)

    that

    y-.:

    y,

    for

    all

    t'

    Thirdly,-it

    is

    important

    to

    realize

    what

    the

    definition

    does

    not

    ,uy.

    tt says

    nothing

    about

    any

    features

    of

    the

    joint

    distribution

    of

    the

    X,

    other

    than

    their

    firit-order

    moments

    (i.e.

    their

    means)

    and

    second-order

    moments

    (i.e.

    variances

    and

    covariances).

    Thus,

    for

    instance,

    it

    is

    quite

    possible

    to

    have a

    weakly stationary

    process

    in

    which

    the

    shapes

    of

    the

    distributions

    of,

    say,

    X, and

    Xo

    aie

    quite ditrerent,

    though

    the

    means

    and

    variances

    of the distributions

    13

  • 8/10/2019 M343 Unit 15

    20/52

    must

    be

    the same.

    Thus

    a

    process

    can be

    weakly stationary

    without

    being

    strictly

    stationary.

    on

    the

    other hand,

    for

    practical

    purposes,

    strictly

    stationary processes

    may

    be

    assumed

    to

    be

    weakly

    stationary

    as well.

    It is

    the

    weak

    definition

    of stationarity

    that

    is

    used most

    in

    modelling

    time

    series.

    There

    are

    basically

    two

    reasons

    for

    this.

    First,

    many

    of the

    properties

    of stationary

    processes

    that

    are

    useful in

    modelling

    time series

    data depend

    only

    on first-

    and

    second-order

    moments,

    so

    it

    is

    unnecessary

    to

    use the

    strict

    definition

    of

    stationarity

    for

    these

    properties

    to

    hold.

    Secondly,

    when

    faced

    with

    time

    series

    data, it

    is

    not

    usually possible

    in

    practice

    to

    estimate much

    more

    than

    the

    first

    and

    second

    moments

    of the

    underlying

    random

    variables. Thus

    it is

    just

    not

    practicable

    to

    check

    whether

    the

    process

    is

    strictly

    stationary.

    Question

    2.6

    Is

    the Melbourne

    weather

    process

    weakly stationary?

    n

    A Markov

    chain

    is

    not

    the sort

    of

    process

    that

    is commonly

    used

    to

    model

    stationary

    time

    series.

    In

    fact,

    the most commonly

    used

    models

    do

    not,

    in

    general,

    possess

    the

    Markov property;

    they are

    built

    up

    from sequences

    of independent

    identically

    distributed

    random

    variables.

    Now

    one does want

    successive

    observations

    in

    a stationary

    time

    series

    model

    to

    be

    identically

    distributed

    (or

    at

    any rate

    to

    have

    constant mean

    and constant

    variance),

    but

    a model in

    which all

    the observations

    are

    independent

    will

    not usually do, because,

    as

    you

    saw

    in

    Section 1,

    time series

    very often

    exhibit

    a

    kind

    of

    'smoothness'in

    which

    successive

    observations

    are relatively

    close to one another,

    compared

    to

    the

    overall level

    of

    variability

    in

    the

    process.

    We shall consider

    two

    different ways

    of

    getting

    the

    required dependence

    into

    the model.

    These

    two

    ways lead

    to

    mouing

    aDerage

    processes

    and

    to

    autoregressiue

    processes.

    Band

    G

    of the video-cassette

    demonstrates

    how

    simuiations

    of

    a

    simple moving

    average process

    and

    two

    simple

    autoregressive processes

    can be

    built up from

    sequences

    of independent

    normal

    random

    variables

    with

    zero

    mean

    and constant

    variance,

    which

    in

    the

    context

    of

    time

    series modelling

    are sometimes

    called white

    noise.

    (The

    term 'white

    noise'

    can

    mean

    different

    things

    in

    other

    contexts.

    In

    particular,

    a continuous-time

    white

    noise process

    can

    be defined

    in

    terms

    of

    Brownian

    motion

    (unit

    14),

    and

    this

    continuous-time

    process can be

    approximated

    by

    sequences

    of independent normal

    random

    variables.)

    This is

    a suitable

    point

    at which

    to

    watch Band

    G of the video-cassette.

    you

    may

    find

    it

    helpfui

    to

    watch it again

    after reading

    Subsection

    5.1,

    by

    which

    time

    you

    will

    have

    studied

    the

    definitions

    and

    properties

    of the

    processes

    simulated

    on

    the

    video-cassette.

    3

    Autoregressive

    processes

    of

    order

    1

    In

    this

    section

    and the next

    we

    study

    a class

    of

    processes

    known

    as autoregressiue

    processes,

    which are

    widely used

    in

    modelling

    time series

    data. In an

    autoregressive

    process,

    each successive

    observation

    is a linear

    function

    of

    previous

    observations,

    plus

    a

    further

    random

    term.

    In this

    section we discuss

    autoregressive processes

    of

    order

    1,

    in

    which the

    observation

    at time

    r is

    given

    by

    a

    multiple

    of the

    observation

    at time

    r

    -

    1,

    added to

    a random

    term. Autoregressive processes

    of

    higher

    order,

    in

    which

    the observation

    at

    time

    r

    is

    defined

    in

    terms

    of more

    than

    one

    previous

    observation,

    are left

    until

    Section

    4.

    Under

    some rather

    special

    clrcumstances,

    a

    process

    can

    in fa;:

    be

    strictly

    stationary

    without

    bein,i

    weakly stationary;

    but this is

    of

    1i::,:

    or no

    practical

    importance.

    Any

    strictly stationary process

    that

    has

    finite

    variances

    and autocovariane:.

    is

    weakly stationarv.

    HT

    R"T?T'|

  • 8/10/2019 M343 Unit 15

    21/52

    ;Jdrliromr

    j.l

    -{

    foraging

    animal

    -

    ::-

    movement

    of

    an animal

    that

    spent

    its

    waking time foraging

    for

    food

    unit

    14, Example

    2.2

    i

    'r

    -'"::

    bank

    was

    modelled. The

    path

    of

    the animal

    was approximated by

    "

    -

    -.-

    :rr)tion,

    and a simulation of the animal's

    path

    was obtained as a

    'r"

    '

    :

    :

    ::iacement

    of the animal

    from

    a reference

    point

    at time f

    -

    1. Then

    '

    --

    ,-

    time

    interval

    from

    time

    t

    -

    1

    to

    time

    r

    the animal

    moves

    by

    a random

    ..

    :

    -

    -::r.ni.

    Z, say, so

    that its

    displacement

    at

    time r is

    'f,:Xr-r*Zr.

    (3

    1)

    ":

    -

    "i.e

    assumptions

    of ordinary

    Brownian

    motion, Z, has

    a

    normal distribution

    :"n

    0 and fixed

    variance o2"(say).

    Moreover, displacements in

    different

    unit

    a

    :

    :-:i:\.als

    are independent. Thus

    the

    animal's movement is being modelled

    by

    .:

    : :r u'alk

    with normally distributed

    steps.

    ::-r.-ess demonstrates

    how

    a stochastic

    process,

    {X,},

    can be

    built

    up

    from a

    ,

    "

    -::-r- ',2,.,

    2r,...

    l

    of

    independent

    identically distributed

    random

    variables.

    -

    rurcsfion

    3.1

    S:orv

    that

    the sequence

    {Zt,2r,...}

    for

    the

    foraging animal, treated

    as a

    ,.-.;hastic

    process,

    is

    weakly

    stationary.

    r,\':ite

    down

    its

    autocorrelation function.

    r

    -.

    the

    process

    strictly

    stationary?

    -

    ::-.

    context

    of

    time series

    models, a sequence

    of

    independent

    identically

    : ,::r:uted

    random variables

    hke

    {Zr}

    is called a

    purely random

    process

    or

    white

    :xr,:,ne.

    Such a

    process

    is

    strictly

    stationary,

    whatever

    the

    distribution

    of Zr,

    and

    is

    i,::{i}'

    stationary

    as long as

    the

    variance

    is finite.

    But

    clearly

    it

    is of

    little

    use

    for

    :-,

    Je11ing

    time

    series

    that exhibit

    dependence.

    ',i

    :.rt

    about the

    process

    {X,}

    defined

    by

    Equation

    (3.1)?

    There is

    clearly

    -.:endence

    between successive

    observations

    where the animal

    is

    this

    minute

    -::ends

    on where

    it

    was

    a minute ago

    -

    but

    a

    major difficulty is that

    this

    process

    .i-. is

    not

    stationary.

    In Section 1

    of Unit

    14,

    you

    saw

    that

    the

    variance of X, is

    ::rportional to

    r,

    rather than being constant.

    So

    this

    process

    may

    be

    a

    use


Recommended