Top Banner
Chapter 6 Autocorrelation
143

Chapter 6

Mar 23, 2016

Download

Documents

chesmu

Chapter 6. Autocorrelation. What is in this Chapter?. How do we detect this problem? What are the consequences? What are the solutions?. What is in this Chapter?. - PowerPoint PPT Presentation
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Chapter 6

Chapter 6

Autocorrelation

Page 2: Chapter 6

What is in this Chapter?

• How do we detect this problem?• What are the consequences?• What are the solutions?

Page 3: Chapter 6

What is in this Chapter?

• Regarding the problem of detection, we start with the Durbin-Watson (DW) statistic, and discuss its several limitations and extensions. We discuss Durbin's h-test for models with lagged dependent variables and tests for higher-order serial correlation.

• We discuss (in Section 6.5) the consequences of serially correlated errors and OLS estimators.

Page 4: Chapter 6

What is in this Chapter?

• The solutions to the problem of serial correlation are discussed in Section 6.3 (estimation in levels versus first differences), Section 6.9 (strategies when the DW test statistic is significant), and Section 6.10 (trends and random walks).

• This chapter is very important and the several ideas have to be understood thoroughly.

Page 5: Chapter 6

6.1 Introduction

• The order of autocorrelation• In the following sections we discuss how t

o:– 1. Test for the presence of serial correlation.– 2. Estimate the regression equation when the

errors are serially correlated.

Page 6: Chapter 6

6.2 Durbin-Watson Test

Page 7: Chapter 6

6.2 Durbin-Watson Test

Page 8: Chapter 6

6.2 Durbin-Watson Test

Page 9: Chapter 6

6.2 Durbin-Watson Test

Page 10: Chapter 6

6.2 Durbin-Watson Test

Page 11: Chapter 6

6.3 Estimation in Levels Versus First Differences

• Simple solutions to the serial correlation problem: First Difference

• If the DW test rejects the hypothesis of zero serial correlation, what is the next step?

• In such cases one estimates a regression by transforming all the variables by ρ-differencing (quasi-first difference) or first-difference

Page 12: Chapter 6

6.3 Estimation in Levels Versus First Differences

Page 13: Chapter 6

6.3 Estimation in Levels Versus First Differences

Page 14: Chapter 6

6.3 Estimation in Levels Versus First Differences

• When comparing equations in levels and first differences, one cannot compare the R2 because the explained variables are different.

• One can compare the residual sum of squares but only after making a rough adjustment. (Please refer to P.231)

Page 15: Chapter 6

6.3 Estimation in Levels Versus First Differences

Page 16: Chapter 6

6.3 Estimation in Levels Versus First Differences

Page 17: Chapter 6

6.3 Estimation in Levels Versus First Differences

• Since we have comparable residual sum of squares (RSS), we can get the comparable R2 as well, using the relationship RSS = Syy(l — R2)

Page 18: Chapter 6

6.3 Estimation in Levels Versus First Differences

Page 19: Chapter 6

6.3 Estimation in Levels Versus First Differences

• Illustrative Examples

Page 20: Chapter 6

6.3 Estimation in Levels Versus First Differences

Page 21: Chapter 6

6.3 Estimation in Levels Versus First Differences

Page 22: Chapter 6

6.3 Estimation in Levels Versus First Differences

Page 23: Chapter 6

6.3 Estimation in Levels Versus First Differences

Page 24: Chapter 6

6.3 Estimation in Levels Versus First Differences

• Usually, with time-series data, one gets high R2 values if the regressions are estimated with the levels yt and Xt but one gets low R2 values if the regressions are estimated in first differences (yt — yt-1) and (xt — xt-1)

• Since a high R2 is usually considered as proof of a strong relationship between the variables under investigation, there is a strong tendency to estimate the equations in levels rather than in first differences.

• This is sometimes called the “R2 syndrome."

Page 25: Chapter 6

6.3 Estimation in Levels Versus First Differences

• However, if the DW statistic is very low, it often implies a misspecified equation, no matter what the value of the R2 is

• In such cases one should estimate the regression equation in first differences and if the R2 is low, this merely indicates that the variables y and x are not related to each other.

Page 26: Chapter 6

6.3 Estimation in Levels Versus First Differences

• Granger and Newbold present some examples with artificially generated data where y, x, and the error u are each generated independently so that there is no relationship between y and x

• But the correlations between yt and yt-1,.Xt and Xt-1, and ut and ut-1 are very high

• Although there is no relationship between y and x the regression of y on x gives a high R2 but a low DW statistic

Page 27: Chapter 6

6.3 Estimation in Levels Versus First Differences

• When the regression is run in first differences, the R2 is close to zero and the DW statistic is close to 2

• Thus demonstrating that there is indeed no relationship between y and x and that the R2 obtained earlier is spurious

• Thus regressions in first differences might often reveal the true nature of the relationship between y and x.

• Further discussion of this problem is in Sections 6.10 and 14.7

Page 28: Chapter 6

Homework• Find the data

– Y is the Taiwan stock index– X is the U.S. stock index

• Run two equations– The equation in levels (log-based price)– The equation in the first differences

• A comparison between the two equations– The beta estimate and its significance– The R square– The value of DW statistic

• Q: Adopt the equation in levels or the first differences?

Page 29: Chapter 6

6.3 Estimation in Levels Versus First Differences

• For instance, suppose that we have quarterly data; then it is possible that the errors in any quarter this year are most highly correlated with the errors in the corresponding quarter last year rather than the errors in the preceding quarter

• That is, ut could be uncorrelated with ut-1 but it could be highly correlated with ut-4.

• If this is the case, the DW statistic will fail to detect it

• What we should be using is a modified statistic defined as

Page 30: Chapter 6

6.3 Estimation in Levels Versus First Differences

Page 31: Chapter 6

6.4 Estimation Procedures with Autocorrelated Errors

Page 32: Chapter 6

6.4 Estimation Procedures with Autocorrelated Errors

Page 33: Chapter 6

6.4 Estimation Procedures with Autocorrelated Errors

Page 34: Chapter 6

6.4 Estimation Procedures with Autocorrelated Errors

Page 35: Chapter 6

6.4 Estimation Procedures with Autocorrelated Errors

• GLS (Generalized least squares)

Page 36: Chapter 6

6.4 Estimation Procedures with Autocorrelated Errors

Page 37: Chapter 6

6.4 Estimation Procedures with Autocorrelated Errors

• In actual practice ρ is not known• There are two types of procedures for esti

mating – 1. Iterative procedures– 2. Grid-search procedures.

Page 38: Chapter 6

6.4 Estimation Procedures with Autocorrelated Errors

Page 39: Chapter 6

6.4 Estimation Procedures with Autocorrelated Errors

Page 40: Chapter 6

6.4 Estimation Procedures with Autocorrelated Errors

Page 41: Chapter 6

6.4 Estimation Procedures with Autocorrelated Errors

Page 42: Chapter 6

6.4 Estimation Procedures with Autocorrelated Errors

Page 43: Chapter 6

6.4 Estimation Procedures with Autocorrelated Errors

Page 44: Chapter 6

Homework

• Redo the example (see Table 3.11 for the data) in the Textbook– OLS – C-O procedure– H-L procedure with the interval of 0.01– Compare the R2 (Note: please calculate the

comparable R2 form the levels equation)

Page 45: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

• In Section 6.4 we described different procedures for the estimation of regression models with AR(1) errors

• We will now answer two questions that might arise with the use of these procedures:– 1. What do we gain from using these

procedures?– 2. When should we not use these procedures?

Page 46: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

• First, in the case we are considering (i.e., the case where the explanatory variable Xt is independent of the error ut), the OLS estimates are unbiased

• However, they will not be efficient• Further, the tests of significance we apply,

which will be based on the wrong covariance matrix, will be wrong.

Page 47: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

• In the case where the explanatory variables include lagged dependent variables, we will have some further problems, which we discuss in Section 6.7

• For the present, let us consider the simple regression model

Page 48: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

Page 49: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

Page 50: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

Page 51: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

Page 52: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

Page 53: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

Page 54: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

Page 55: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

Page 56: Chapter 6

An Alternative Method to Prove the Above Characteristics???

• Use “simulation method” as shown at Chapter 5

• Write your program by the Gauss program• Take the program at Chapter 5 and make

some modifications on it

Page 57: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

• Thus the consequences of autocorrelated errors are:– 1. The least squares estimators are unbiased but are

not efficient. Sometimes they are considerably less efficient than the procedures that take account of the autocorrelation

– 2. The sampling variances are biased and sometimes likely to be seriously understated. Thus R2 as well as t and F statistics tend to be exaggerated.

Page 58: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

Page 59: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

• 2. The discussion above assumes that the true errors are first-order autoregressive. If they have a more complicated structure (e.g., second-order autoregressive), it might be thought that it would still be better to proceed on the assumption that the errors are first-order autoregressive rather than ignore the problem completely and use the OLS method???– Engle shows that this is not necessarily true (i.e.,

sometimes one can be worse off making the assumption of first-order autocorrelation than ignoring the problem completely).

Page 60: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

Page 61: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

Page 62: Chapter 6

6.5 Effect of AR(1) Errors on OLS Estimates

Page 63: Chapter 6

6.7 Tests for Serial Correlation in Models with Lagged Dependent Variables

• In previous sections we considered explanatory variables that were uncorrelated with the error term

• This will not be the case if we have lagged dependent variables among the explanatory variables and we have serially correlated errors

• There are several situations under which we would be considering lagged dependent variables as explanatory variables

• These could arise through expectations, adjustment lags, and so on.

Page 64: Chapter 6

6.7 Tests for Serial Correlation in Models with Lagged Dependent Variables

• The various situations and models are explained in Chapter 10. For the present we will not be concerned with how the models arise. We will merely study the problem of testing for autocorrelation in these models

• Let us consider a simple model

Page 65: Chapter 6

6.7 Tests for Serial Correlation in Models with Lagged Dependent Variables

Page 66: Chapter 6

6.7 Tests for Serial Correlation in Models with Lagged Dependent Variables

Page 67: Chapter 6

• new;• format /m1 /rd 9,3;

• beta=2;

• T=30; @ sample number @ • u=Rndn(T,1); • x=Rndn(T,1)+0*u;• y=beta*x+u;

• @ OLS @• Beta_OLS=olsqr(y,x);

• print " OLS beta estimate ";• Beta_OLS;

Page 68: Chapter 6
Page 69: Chapter 6

• new;• format /m1 /rd 9,3;

• beta=2;

• T=50000; @ sample number @ • u=Rndn(T,1); • x=Rndn(T,1)+0*u;• y=beta*x+u;

• @ OLS @• Beta_OLS=olsqr(y,x);

• print " OLS beta estimate ";• Beta_OLS;

Page 70: Chapter 6
Page 71: Chapter 6

• new;• format /m1 /rd 9,3;

• beta=2;

• T=50000; @ sample number @ • u=Rndn(T,1); • x=Rndn(T,1)+0.5*u;• y=beta*x+u;

• @ OLS @• Beta_OLS=olsqr(y,x);

• print " OLS beta estimate ";• Beta_OLS;

Page 72: Chapter 6
Page 73: Chapter 6

6.7 Tests for Serial Correlation in Models with Lagged Dependent Variables

Page 74: Chapter 6

6.7 Tests for Serial Correlation in Models with Lagged Dependent Variables

Page 75: Chapter 6

6.7 Tests for Serial Correlation in Models with Lagged Dependent Variables

Page 76: Chapter 6

6.7 Tests for Serial Correlation in Models with Lagged Dependent Variables

Page 77: Chapter 6

6.7 Tests for Serial Correlation in Models with Lagged Dependent Variables

Page 78: Chapter 6

6.8 A General Test for Higher-Order Serial Correlation: The LM Test

• The h-test we have discussed is, like the Durbin-Watson test, a test for first-order autoregression.

• Breusch and Godfrey discuss some general tests that are easy to apply and are valid for very general hypotheses about the serial correlation in the errors

• These tests are derived from a general principle — called the Lagrange multiplier (LM) principle

• A discussion of this principle is beyond the scope of this book. For the present we will explain what the test is

• The test is similar to Durbin's second test that we have discussed

Page 79: Chapter 6

6.8 A General Test for Higher-Order Serial Correlation: The LM Test

Page 80: Chapter 6

6.8 A General Test for Higher-Order Serial Correlation: The LM Test

Page 81: Chapter 6

6.8 A General Test for Higher-Order Serial Correlation: The LM Test

Page 82: Chapter 6

6.8 A General Test for Higher-Order Serial Correlation: The LM Test

Page 83: Chapter 6

6.8 A General Test for Higher-Order Serial Correlation: The LM Test

Page 84: Chapter 6

6.9 Strategies When the DW Test Statistic is Significant

• The DW test is designed as a test for the hypothesis ρ = 0 if the errors follow a first-order autoregressive process

• However, the test has been found to be robust against other alternatives such as AR(2), MA(1), ARMA(1, 1), and so on.

• Further, and more disturbingly, it catches specification errors like omitted variables that are themselves autocorrelated, and misspecified dynamics (a term that we will explain). Thus the strategy to adopt, if the DW test statistic is significant, is not clear. We discuss three different strategies:

Page 85: Chapter 6

6.9 Strategies When the DW Test Statistic is Significant

• 1. Assume that the significant DW statistic is an indication of serial correlation but may not be due to AR(1) errors

• 2. Test whether serial correlation is due to omitted variables.

• 3. Test whether serial correlation is due to misspecified dynamics.

Page 86: Chapter 6

6.9 Strategies When the DW Test Statistic is Significant

Page 87: Chapter 6

6.9 Strategies When the DW Test Statistic is Significant

Page 88: Chapter 6

6.9 Strategies When the DW Test Statistic is Significant

Page 89: Chapter 6

6.9 Strategies When the DW Test Statistic is Significant

Page 90: Chapter 6

6.9 Strategies When the DW Test Statistic is Significant

Page 91: Chapter 6

6.9 Strategies When the DW Test Statistic is Significant

• Serial correlation due to misspecification dynamics

Page 92: Chapter 6

6.9 Strategies When the DW Test Statistic is Significant

Page 93: Chapter 6

6.9 Strategies When the DW Test Statistic is Significant

Page 94: Chapter 6

6.9 Strategies When the DW Test Statistic is Significant

Page 95: Chapter 6

6.9 Strategies When the DW Test Statistic is Significant

Page 96: Chapter 6

6.10 Trends and Random Walks

Page 97: Chapter 6

6.10 Trends and Random Walks

Page 98: Chapter 6

6.10 Trends and Random Walks

Page 99: Chapter 6

6.10 Trends and Random Walks

Page 100: Chapter 6

6.10 Trends and Random Walks

• Both the models exhibit a linear trend. But the appropriate method of eliminating the trend differs

• To test the hypothesis that a time series belongs to the TSP class against the alternative that it belongs to the DSP class, Nelson and Plosser use a test developed by Dickey and Fuller:

Page 101: Chapter 6

6.10 Trends and Random Walks

Page 102: Chapter 6

6.10 Trends and Random Walks

Page 103: Chapter 6

Three Types of RW• RW without drift: Yt=1*Yt-1+ut;

• RW with drift: Yt=alpha+1*Yt-1+ut;

• RW with drift and time trend: Yt=alpha+beta*t+1*Yt-1+ut

• ut~iid(0,sigma)

Page 104: Chapter 6
Page 105: Chapter 6
Page 106: Chapter 6
Page 107: Chapter 6
Page 108: Chapter 6

RW or Unit Root tests by E-view

• Additional Slides:• Augmented D-F tests

– Yt=a1*Yt-1+ut;– Yt-Yt-1=(a1-1)*Yt-1+ut– ΔYt=(a1-1)*Yt-1+ut– ΔYt=λ*Yt-1+ut– H0:a1=1 ≡ H0: λ=0

• ΔYt=λ*Yt-1+ΣΔYt-i+ut

Page 109: Chapter 6

6.10 Trends and Random Walks

Page 110: Chapter 6

6.10 Trends and Random Walks

• As an illustration consider the example given by Dickey and Fuller.36 For the logarithm of the quarterly Federal Reserve Board Production Index 1950-1 through 1977-4 they assume that the time series is adequately represented by the model

Page 111: Chapter 6

6.10 Trends and Random Walks

Page 112: Chapter 6

6.10 Trends and Random Walks

Page 113: Chapter 6

6.10 Trends and Random Walks

• 6. Regression of one random walk on another, with time included for trend, is strongly subject to the spurious regression phenomenon. That is, the conventional t-test will tend to indicate a relationship between the variables when none is present.

Page 114: Chapter 6

6.10 Trends and Random Walks• The main conclusion is that using a regression o

n time has serious consequences when, in fact, the time series is of the DSP type and, hence, differencing is the appropriate procedure for trend elimination

• Plosser and Schwert also argue that with most economic time series it is always best to work with differenced data rather than data in levels

• The reason is that if indeed the data series are of the DSP type, the errors in the levels equation will have variances increasing over time

Page 115: Chapter 6

6.10 Trends and Random Walks

• Under these circumstances many of the properties of least squares estimators as well as tests of significance are invalid

• On the other hand, suppose that the levels equation is correctly specified. Then all differencing will do is produce a moving average error and at worst ignoring it will give inefficient estimates

• For instance, suppose that we have the model

Page 116: Chapter 6

6.10 Trends and Random Walks

Page 117: Chapter 6

6.10 Trends and Random Walks

• Differencing and Long-Run Effects:The Concept of Cointegration– One drawback of the procedure of differencing is that

it results in a loss of valuable "long-run information" in the data

– Recently, the concept of cointegrated series has been suggested as one solution to this problem.39 First, we need to define the term "cointegration.“

– Although we do not need the assumption of normality and independence, we will define the terms under this assumption.

Page 118: Chapter 6

6.10 Trends and Random Walks

Page 119: Chapter 6

6.10 Trends and Random Walks

• Yt~I(1)– Yt is a random walk–△Yt is a white noise, or iid– No one could predict the future price change– The market is efficient– The impact of previous shock on the price will

remain and not approach to zero

Page 120: Chapter 6

6.10 Trends and Random Walks

Page 121: Chapter 6

6.10 Trends and Random Walks

Page 122: Chapter 6

6.10 Trends and Random Walks

Page 123: Chapter 6

Cointegration

ttt ZUNDADR 10

tU

q

jjtjUU

p

iitiUAtUUt

tA

q

jjtjAU

p

iitiAAtAAt

eUNDADRZUND

eUNDADRZADR

,1

,1

,1

,1

,1

,1

Page 124: Chapter 6

Cointegration

Table 1 Unit Root and Cointergration Tests U.K. VOD Brazil CVDO ADR UND ADR UND Log levels -1.4979 -1.5236 -0.6505 -0.2749 % Returns -16.8033* -16.7456* -16.5194* -15.8273* Error Correction Term -11.9252* -4.0032*

Page 125: Chapter 6

Cointegration

• Run the VECM (vector error correction model) by E-view

• Additional slides

Page 126: Chapter 6

CointegrationTable 2 Parameter Estimates of VECM

U.K. VOD Brazil CVDO ADR Equation

αA -0.0006 (0.0008) 0.0008 (0.0007) βA -0.0049 (0.0885) -0.0256 (0.0130)** γAA,1 -0.1518 (0.0871)* -0.0578 (0.0433) γAU,1 0.0829 (0.0852) 0.1248 (0.0453)*** γAA,2 -0.0767 (0.0812) -0.1534 (0.0465)*** γAU,2 -0.0158 (0.0778) 0.0560 (0.0479) γAA,3 -0.0766 (0.0710) -0.1452 (0.0464)*** γAU,3 -0.0204 (0.0660) 0.0952 (0.0472)** γAA,4 -0.0226 (0.0526) -0.0677 (0.0437) γAU,4 0.0408 (0.0439) 0.0287 (0.0429)

UND Equation αU -0.0006 (0.0007) 0.0012 (0.0007)* βU 0.5636 (0.0738)*** 0.0020 (0.0123) γUA,1 0.2844 (0.0727)*** 0.3833 (0.0411)*** γUU,1 -0.2748 (0.0711)*** -0.2970 (0.0430)*** γUA,2 0.1955 (0.0677)*** 0.1478 (0.0442)*** γUU,2 -0.2543 (0.0649)*** -0.1914 (0.0454)*** γUA,3 0.1262 (0.0592)** 0.0634 (0.0440) γUU,3 -0.1510 (0.0551)*** -0.0479 (0.0449) γUA,4 0.0487 (0.0439) -0.0309 (0.0415) γUU,4 -0.0152 (0.0366) -0.0123 (0.0408)

Page 127: Chapter 6

Lead-lag relation obtained with VECM model

• If beta_A is significant and beta_U is insignificant, – the price adjustment mainly depends on ADR

markets– ADR prices converge to UND prices– UND prices lead ADR prices in price discover

y process– UND prices provide an information advantage

Page 128: Chapter 6

• If beta_U is significant and beta_A is insignificant, – the price adjustment mainly depends on UND

markets– UND prices converge to ADR prices– ADR prices lead UND prices in price discover

y process– ADR prices provide an information advantage

Page 129: Chapter 6

• If both of beta_U and beta_A are significant– suggesting a bidirectional error correction – The equilibrium prices line within ADR and UN

D prices– Both ADR and UND prices converge to the eq

uilibrium prices

Page 130: Chapter 6

• If both of beta_U and beta_A are significant, but the beta_U is greater than beta_A in absolute velue– The finding denotes that it is the UND price th

at makes greater adjustment in order to reestablish the equilibrium

– That is, most of the price discovery takes place at the ADR market.

Page 131: Chapter 6

Homework

• Find the spot and futures prices• Daily and 5-year data at least• Run the cointegration test• Run the VECM

– Lead-lag relationship

Page 132: Chapter 6

6.11 ARCH Models and Serial Correlation

• We saw in Section 6.9 that a significant DW statistic can arise through a number of misspecifications.

• We will now discuss one other source. This is the ARCH model suggested by Engle which has, in recent years, been found useful in the analysis of speculative prices.

• ARCH stands for "autoregressive conditional heteroskedasticity."

Page 133: Chapter 6

6.11 ARCH Models and Serial Correlation

• GARCH (p,q) Model:

q

i jtjitp

i it

tt

t

m

lltlt

baa

yy

122

102

2

1

),0(~

Page 134: Chapter 6

6.11 ARCH Models and Serial Correlation

• The high level of persistence in GARCH models– the sum of the two GARCH parameter

estimates approximates unity in most cases– Li and Lin (2003): This finding provides some

support for the notion that GARCH models are handicapped by the inability to account for structural changes during the estimation period and thus suffers from a high persistence problem in variance settings.

Page 135: Chapter 6

6.11 ARCH Models and Serial Correlation

• Find the stock returns• Daily and 5-year data at least• Run the GARCH(1,1) model

– Check the sum of the two GARCH parameter estimates

– Parameter estimates – Graph the time-varying variance estimates

Page 136: Chapter 6

Could we identify RW? Low test power of the DF test

The Power of the test? The H0 is not true, but we accept the H0The data series is I(0), but we conclude it is

I(1)

Page 137: Chapter 6

Several Key Problems for Unit Root Tests

• Low test power • Structural change problem• Size distortion

• RW or non-stationary or I(1) : Yt=1*Yt-1+ut• Stationary Process or I(0): Yt=0.99*Yt-1+ut-1, T=1,000 Yt=0.98*Yt-1+ut-1, T=50 or 1000

Page 138: Chapter 6
Page 139: Chapter 6
Page 140: Chapter 6
Page 141: Chapter 6

Spurious Regression

• RW 1 : Yt=0.05+1*Yt-1+ut

• RW 2: Xt=0.03+1*Xt-1+vt

Page 142: Chapter 6

Spurious Regression• new;• format /m1 /rd 9,3;

• @ Data Gerneration Process @

• Y=zeros(1000,1); u=2*Rndn(1000,1);• X=zeros(1000,1); v=1*Rndn(1000,1);

• i=2;• do until i>1000;

• Y[i,1]=0.05+1*Y[i-1,1]+u[i,1];

• X[i,1]=0.03+1*X[i-1,1]+v[i,1];

• i=i+1;• endo;

• Output file=d:\Courses\Enclass\Unit\YX_Spur.out reset;

• Y~X;

• Output off;

Page 143: Chapter 6