0% found this document useful (0 votes)
17 views

Stationary Random Series

Uploaded by

kipropnathan15
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
17 views

Stationary Random Series

Uploaded by

kipropnathan15
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 14

Page 6 CS2-13: Time series 1

2 Stationary random series

2.1 Stationary time series processes


The concept of stationarity was introduced in Chapter 1, along with the ideas of strict and
weak stationarity.

Question

Explain what it means for a process X to be:

(i) strictly stationary

(ii) weakly stationary.

Solution

(i) A process X is strictly stationary if the joint distributions of Xt1 , Xt2 , ..., Xtn and
X k  t1 , X k  t2 , ..., X k  tn are identical for all t1 , t2 , , tn and k  t1 , k  t2 , , k  tn in J and
all integers n. This means that the statistical properties of the process remain unchanged
as time elapses.

(ii) A process X is weakly stationary if E ( Xt ) is independent of t and cov( Xt , Xt  s ) depends


only on the lag, s .

A weakly stationary process has constant variance since, for such a process,
var( Xt )  cov( Xt , Xt ) is independent of t .

In the study of time series it is a convention that the word ‘stationary’ on its own is a
shorthand notation for ‘weakly stationary’, though in the case of a multivariate normal
process, strict and weak stationarity are equivalent.

This is because a distribution of a multivariate normal random variable is completely determined


by its mean vector and covariance matrix. We will consider this further in Chapter 14.

But we do need to be careful in our definition, as there are some processes which we wish
to exclude from consideration but which satisfy the definition of weak stationarity.

Purely indeterministic processes


A process X is called purely indeterministic if knowledge of the values of X 1, , X n is
progressively less useful at predicting the value of X N as N   . When we talk of a
‘stationary time series process’ we shall mean a weakly stationary purely indeterministic
process.

© IFE: 2019 Examinations The Actuarial Education Company


CS2-13: Time series 1 Page 7

Question

Let Yt be a sequence of independent standard normal random variables. Determine which of the
following processes are stationary time series (given the definition above).

(i) Xt  sin t  U  , where U is uniformly distributed on the interval  0,2 

(ii) Xt  sin t  Yt 

(iii) Xt  Xt 1  Yt

(iv) Xt  Yt 1  Yt

(v) Xt  2  3t  0.5 Xt 1  Yt  0.3Yt 1

Solution

(i) For this process, X 0  sin U , X1  sin(  U) , X2  sin(2  U) , …. Given the value of X 0 ,
future values of the process are fully determined. So this is not purely indeterministic,
and is not therefore a stationary time series in the sense defined in the Core Reading.

(ii) For this process, X 0  sinY0 , X1  sin(  Y1 ) , X2  sin(2  Y2 ) , …. Since E ( Xt ) varies


over time, this process is not stationary.

(iii) Here we have:

E ( Xt )  E ( Xt 1  Yt )  E ( Xt 1 )  E (Yt )  E ( Xt 1 )

So the process has a constant mean. However:

var( Xt )  var( Xt 1  Yt )  var( Xt 1 )  var(Yt )  var( Xt 1 )  1

Here we are using the fact that Yt is a sequence of independent standard normal random
variables. Since the variance is not constant, the process is not stationary.

(iv) This process is weakly stationary:

E ( Xt )  E (Yt 1 )  E (Yt )  0

and:

2 k0

cov  Xt , Xt  k   1 k 1
0 k 2

The Actuarial Education Company © IFE: 2019 Examinations


Page 8 CS2-13: Time series 1

For example:

cov( Xt , Xt )  cov(Yt 1  Yt ,Yt 1  Yt )

 cov(Yt 1 ,Yt 1 )  2cov(Yt ,Yt 1 )  cov(Yt ,Yt )

 var(Yt 1 )  var(Yt )  2

In addition, it is purely indeterministic. From the defining equation, we have:

X1  Y0  Y1

So Y1  X1  Y0 and:

X2  Y1  Y2  ( X1  Y0 )  Y2

Rearranging gives:

Y2  X2  X1  Y0

and hence:

X 3  Y2  Y3  ( X2  X1  Y0 )  Y3

Continuing in this way, we see that:

X n  X n 1  X n 2  X n  3    (1)n 2 X1  (1)n 1Y0  Yn

From this formula, we see that knowledge of the values of X1 , X2 and X 3 , say, becomes
progressively less useful in predicting the value of X n as n   .

(v) This process has a deterministic trend via the ‘3t’ term, ie its mean varies over time. So it
is not stationary.

A particular form of notation is used for time series: X is said to be I (0) (read ‘integrated
of order 0’) if it is a stationary time series process, X is I (1) if X itself is not stationary but
the increments Yt  X t  X t 1 form a stationary process, X is I (2) if it is non-stationary but
the process Y is I (1) , and so on.

We will see plenty of examples of integrated processes when we study the ARIMA class of
processes in Section 3.8.

The theory of stationary random processes plays an important role in the theory of time
series because the calibration of time series models (that is, estimation of the values of the
model’s parameters using historical data) can be performed efficiently only in the case of
stationary random processes. A non-stationary random process has to be transformed into
a stationary one before the calibration can be performed. (See Chapter 14.)

© IFE: 2019 Examinations The Actuarial Education Company


CS2-13: Time series 1 Page 9

Question

Suppose that we have a sample set of data that looks to be a realisation of an integrated process
of order 2. Explain what can we do to the data set in order to model it.

Solution

We can difference the data twice, ie look at the increments of the increments.

2.2 Autocovariance function


The mean function (or trend) of the process is t  E [ X t ] , the covariance function
cov( X s , X t )  E [ X s X t ]  E [ X s ]E [ X t ] . Both of these functions take a simpler form in the
case where X is stationary:

 The mean of a stationary time series process is constant, ie t   for all t.

 The covariance of any pair of elements X r and X s of a stationary sequence X


depends only on the difference r – s.

Autocovariance function

We can therefore define the autocovariance function  k : k  Z  of a stationary time series


process X as follows:

 k  cov  X t , X t  k   E  X t X t  k   E  X t  E  X t  k 

The common variance of the elements of a stationary process is given by:

 0  var( X t )

If a process is not stationary, then the autocovariance function depends on two variables, namely
the time t and the lag k. This could be denoted, for example,   t , k   cov  Xt , Xt  k  . However,
one of the main uses of the autocovariance function is to determine the type of process that will
be used to model a given set of data. Since this will be done only for stationary series, as
mentioned above, it is the autocovariance function for stationary series that is most important.

Because of the importance of the autocovariance function, we will have to calculate it for various
processes. This naturally involves calculating covariances and so we need to be familiar with all of
the properties of the covariance of two random variables. The following question is included as a
revision exercise.

The Actuarial Education Company © IFE: 2019 Examinations


Page 10 CS2-13: Time series 1

Question

Let X and Y denote any random variables.

(i) Express cov( X ,Y ) in terms of E ( X ) , E (Y ) and E ( XY ) .

(ii) Express each of the following in terms of cov( X ,Y ) :

(a) cov(Y , X )

(b) cov( X , c) , where c is a constant

(c) cov(2 X ,3Y )

(iii) Give an equivalent expression for cov( X , X ) .

(iv) Prove, using your formula in (i), that:

cov( X  Y ,W )  cov( X ,W )  cov(Y ,W )

(v) Simplify each of the following expressions assuming that { Xt } denotes a stationary time
series defined at integer times and { Zt } are independent N(0,  2 ) random variables.

(a) cov(Z2 , Z3 )

(b) cov(Z3 , Z3 )

(c) cov( X2 , Z3 )

(d) cov( X2 , X 3 )

(e) cov( X2 , X2 )

Solution

(i) cov( X ,Y )  E ( XY )  E ( X )E (Y )

(ii) (a) cov(Y , X )  cov( X ,Y )

(b) cov( X , c)  0

(c) cov(2 X ,3Y )  6cov( X ,Y )

(iii) cov(X , X)  E(X 2 )  [E(X)]2  var(X )

(iv) cov( X  Y ,W )  E[( X  Y )W ]  E ( X  Y )E (W )


 E ( XW  YW )  [E ( X )  E (Y )]E (W )
 E ( XW )  E (YW )  E ( X )E (W )  E (Y )E (W )

 cov( X ,W )  cov(Y ,W )

© IFE: 2019 Examinations The Actuarial Education Company


CS2-13: Time series 1 Page 11

(v) (a) cov( Z2 , Z3 )  0 , since they are independent.

(b) cov(Z3 , Z3 )  var(Z3 )   2

(c) cov( X2 , Z3 )  0

(d) and (e) will depend on the actual process. If it is stationary, then cov( X2 , X 3 )   1 , and
cov( X2 , X2 )   0 .

2.3 Autocorrelation function


The autocovariance function is measured in squared units, so that the values obtained depend on
the absolute size of the measurements. We can make this quantity independent of the absolute
sizes of X n by defining a dimensionless quantity, the autocorrelation function.

Autocorrelation function
The autocorrelation function (ACF) of a stationary process is defined by:

k
 k  corr( X t , X t  k ) 
0

The ACF of a purely indeterministic process satisfies k  0 as k   .

This statement is intuitive. We do not expect two values of a (purely indeterministic) time series
to be correlated if they are a long way apart.

Question

Write down the formula for the correlation coefficient between the random variables X and Y .

Hence deduce the formula for the autocorrelation function given above.

Solution

The formula for the correlation coefficient is:

cov  X ,Y 
corr  X ,Y  
var( X )var(Y )

So:

cov( Xt , Xt  k ) k 
k    k
var( Xt )var( Xt  k ) 0 0 0

The Actuarial Education Company © IFE: 2019 Examinations


Page 12 CS2-13: Time series 1

For a non-stationary process we could define an autocorrelation function by:

cov  Xt , Xt k   t , k 
 t , k   
var( Xt ) var( Xt k )   t ,0    t  k ,0 

However, as with the autocovariance function, it is the stationary case that is of most use in
practice.

A simple class of weakly stationary random processes is the white noise processes. A
random process et : t  Z  is a white noise if E et   0 for any t , and:

 2 if k  0
 k  cov(et , et  k )  
0 otherwise

An important representative of the white noise processes is a sequence of independent


normal random variables with common mean 0 and variance  2 .

Strictly speaking a white noise process only has to be a sequence of uncorrelated random
variables, ie not necessarily a sequence of independent random variables. We can also have white
noise processes without zero mean.

Result 13.1
The autocovariance function  and autocorrelation function  of a stationary random
process are even functions of k , that is,  k    k and  k    k .

Proof
Since the autocovariance function  k  cov  X t , X t  k  does not depend on t, we have:

 k  cov  X t  k , X t  k  k   cov  X t  k , X t   cov  X t , X t  k     k

Thus  is an even function, which in turn implies that  is even.

This result allows us to concentrate on positive lags when finding the autocorrelation functions of
stationary processes.

2.4 Correlograms
Autocorrelation functions are the most commonly used statistic in time series analysis. A lot of
information about a time series can be deduced from a plot of the sample autocorrelation
function (as a function of the lag). Such a plot is called a correlogram.

Typical stationary series


A typical sample autocorrelation function for a stationary series looks like the one shown below.
The lag is shown on the horizontal axis, and the autocorrelation on the vertical.

© IFE: 2019 Examinations The Actuarial Education Company


CS2-13: Time series 1 Page 13

0.5

1 2 10

-0.5

-1


At lag 0 the autocorrelation function takes the value 1, since 0  0  1 . Often the function
0
starts out at 1 but decays fairly quickly, which is indicative of the time series being stationary. The
above correlation function tells us that at lags 0, 1 and 2 there is some positive correlation so that
a value on one side of the mean will tend to have a couple of values following that are on the
same side of the mean. However, beyond lag 2 there is little correlation.

In fact, the above function comes from a sample path of a stationary AR(1) process, namely
X n  0.5 X n1  en . (We look in more detail at such processes in the next section.)

The data used for the first 50 values is plotted below. (The actual data used to produce the
autocorrelation function used the first 1,000 values.)

0 10 20 30 40 50

The ‘gap’ in the axes here is deliberate; the vertical axis does not start at zero. The horizontal axis
on this and the next graph shows time, and the vertical axis shows the value of the time series X .

This form of presentation is difficult to interpret. It’s easier to see if we ‘join the dots’.

The Actuarial Education Company © IFE: 2019 Examinations


Page 14 CS2-13: Time series 1

0 10 20 30 40 50

By inspection of this graph we can indeed see that one value tends to be followed by another
similar value. This is also true at lag 2, though slightly less clear. Once the lag is 3 or more, there
is little correlation.

The previous data set is in stark contrast to the following one.

Alternating series

0 10 20 30 40 50

The average of this data is obviously roughly in the middle of the extreme values. Given a
particular value, the following one tends to be on the other side of the mean. The series is
alternating. This is reflected in the autocorrelation function shown below. At lag 1 there is a
negative correlation. Conversely, at lag 2, the two points will generally be on the same side of the
mean and therefore will have positive correlation, and so on. The autocorrelation therefore also
alternates as shown.

© IFE: 2019 Examinations The Actuarial Education Company


CS2-13: Time series 1 Page 15

0.5

1 2 3 4 5 6 7 8 9 10

–0.5

–1

The data in this case actually came from a stationary autoregressive process, this time
X n  0.85 X n1  en . This is stationary, but because the coefficient of X n1 is larger in
magnitude, ie 0.85 vs 0.5, the decay of the autocorrelation function is slower. This is because the
X n1 term is not swamped by the random factors en as quickly. It is the fact that the coefficient
is negative that makes the series alternate.

Series with a trend


A final example comes from the following data generated from X n  0.1n  0.5 X n1  en .

0 10 20 30 40 50

In this time series, a strong trend is clearly visible. The effect of this is that any given value is
followed, in general, by terms that are greater. This gives positive correlation at all lags. The
decay of the autocorrelation function will be very slow, if it occurs at all.

The Actuarial Education Company © IFE: 2019 Examinations


Page 16 CS2-13: Time series 1

0.5

1 2 3 4 5 6 7 8 9 10

–0.5

–1

If the trend is weaker, for example X n  0.001n  0.5 X n1  en , then there may be some decay at
first as the trend is swamped by the other factors, but there will still be some residual correlation
at larger lags.

0 10 20 30 40 50

The trend is difficult to see from this small sample of the data but shows up in the autocorrelation
function as the residual correlation at higher lags.

0.5

1 2 3 4 5 6 7 8 9 10

–0.5

–1

© IFE: 2019 Examinations The Actuarial Education Company


CS2-13: Time series 1 Page 17

Question

Describe the associations we would expect to find in a time series representing the average
daytime temperature in successive months in a particular town, and hence sketch a diagram of
the autocorrelation function of this series.

Solution

We expect the temperature in different years to be roughly the same at the same time of year,
and hence there should be very strong positive correlation at lags of 12 months, 24 months and
so on.

Within each year we would also expect a positive correlation between nearby times, for example
with lags of 1 or 2 months, with decreasing correlation as the lag increases. On the other hand,
once we reach a lag of 6 months there should be strong negative correlation since one
temperature will be above the mean, the other below it. For example comparing June with
December.

The autocorrelation function will therefore oscillate with period of 12 months.

0.5

5 10 15 20 25

–0.5

–1

2.5 Partial autocorrelation function


Another important characteristic of a stationary random process is the partial
autocorrelation function (PACF), k : k  1,2, , defined as the conditional correlation of
X t  k with X t given X t 1, , X t  k 1 .

Unlike the autocovariance and autocorrelation functions, the PACF is defined for positive lags
only.

This may be derived as the coefficient k ,k in the problem to minimise:


 
2
E  X t  k ,1X t 1  k ,2 X t 2    k ,k X t k 

The Actuarial Education Company © IFE: 2019 Examinations


Page 18 CS2-13: Time series 1

We can explain the last expression as follows. Suppose that at time t  1 we are trying to
estimate Xt , but we are going to limit our choice of estimator to linear functions of the k
previous values Xt k , , Xt 1 . The most general linear estimator will be of the form:

k ,1 Xt 1  k ,2 Xt 2    k ,k Xt k

where k ,i are constants. We can choose the coefficients to minimise the mean square error,
which is the expression given above in Core Reading. The partial autocorrelation for lag k is then
the weight that we assign to the Xt k term.

Question

Consider the process Xt  0.5 Xt 2  et , where et forms a white noise process.

Determine the partial autocorrelation function for this process.

Solution

For k  1 we just have the correlation itself. However, in this case it is clear that the Xt for even
values of t are independent of those for odd values. It follows that the correlation at lag 1 is 0.

For k  2 the partial autocorrelation is the coefficient of Xt 2 in the best linear estimator:

2,1 Xt 1  2,2 Xt 2

Comparing this to the defining equation suggests that 2  0.5 .

Similarly, the defining equation suggests that the best linear estimator will not involve
Xt 3 , Xt 4 , . It follows that for k  3 , we have k  0 .

For the time series in the previous question, we have 4  0 . This is in contrast to the actual
correlation at lag four, since Xt depends on Xt 2 , which in turn depends on Xt 4 . Xt and Xt 4
will therefore be correlated. The partial autocorrelation is zero, however, because it effectively
removes the impact of the correlation at smaller lags.

In general it is difficult to calculate the PACF by hand.

The formula for calculating k involves a ratio of determinants of large matrices whose
entries are determined by 1, ,  k ; it may be found in standard works on time series
analysis, and is readily available in common computer packages like R.

© IFE: 2019 Examinations The Actuarial Education Company


CS2-13: Time series 1 Page 19

The diagrams below show the autocorrelation function and partial autocorrelation of an
ARMA(1,1) series. ARMA processes are discussed in detail in Section 3.7.

Figure 13.1: ACF and PACF values of some stationary time series model.

In particular the formulae for 1 and 2 are as follows:

Partial autocorrelation function at lags 1 and 2

 1 1 
det   2
1  1, 2   1 2   2  1
 1 1  1  12
det  
 1 1 

Note that for each k , k depends on only 1, 2 ,...,  k .

These formulae can be found on page 40 of the Tables. Their derivations are not required.

It is important to realise that the PACF is determined by the ACF, as the above expressions
suggest. The PACF does not therefore contain any extra information; it just gives an alternative
presentation of the same information. However, as we will see, this can be used to identify
certain types of process.

The Actuarial Education Company © IFE: 2019 Examinations

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy