Time-Series Econometrics
Time-Series Econometrics
Time-Series Econometrics
Stochastic Process:
Analysis of time-series is based on the modelling of stochastic process.
A stochastic process is a collection of random variables ordered in time.
The stochastic process evolves in time according to probabilistic laws.
An observed time-series is considered to be one realization of a stochastic process.
=>
E( Y t )=0 ;var( Y t )=σ
j =0
2
( )
∑ β 2j
Autoregressive (AR) Value of Y at time point t depends Autoregressive process of order r:
Process on its previous values and random Y t =α 1 Y t−1 +α 2 Y t−2 +.. .+α r Y t−r +ut
disturbance at that time point.
ARMA Process The variable Y has characteristics of ARMA process of order r and m, i.e.
both AR and MA ARMA (r, m):
Y t =α 1 Y t−1 +α 2 Y t−2 +.. .+α r Y t−r +ut +
β 1 ut−1 +β 2 ut−2 +. ..+β m u t−m
Here,
E(ut )=0 ∀ t ;var(u t )=σ 2 ∀ t ;
cov (ut ,u t−s )=0 ∀ s≠0
ARIMA Process Differencing non-stationary time- ARIMA (r, d, m) means the time-series
series to make it integrated and has to be differenced d times to make it
modelling the differenced series as stationary and the stationary time series
ARMA process can be modelled as ARMA(r, m)
Nature of Time-Series
Three possibilities
(1) When ρ>1 , the series is non-stationary and explosive. Past shocks have a greater
impact than current ones.
(2) When ρ=1 , the series is non-stationary shocks persist at full force, and the series is
not mean-reverting. This is the random walk model and where the variance increases
with t and we have the infinite variance problem.
(3) When ρ<1 , series is stationary and the effects of shocks die out exponentially
according to. The series reverts to its mean.
var ( Y t|Y 0 )=var
(∑ ) ∑
τ =0
ut−τ =
τ =0
var(ut−τ )=tσ 2
(conditional variance)
Thus, conditional variance varies with time.
Unconditional variance:
∞ ∞ ∞
var ( Y t ) =var
( )(
∑ uτ =
τ=0 τ=0
)
∑ var(u τ ) =∑ σ 2=∞
τ=0
2
Note: tσ →∞ as t →∞
Y 1 =ρY 0 +u1
When < 1 the time-series is stationary. In this case, the time-series looks jagged and it never
wanders too far from the mean. The effect of the errors decay and disappear over time. Impact
of recent events are relatively more important than what happened a long time ago.
With given initial value of Y at Y0,
t t−1
Y t =Y 0 + ∑ ρ u j=Y 0 + ∑ ρτ ut−τ
τ
j =1 τ =0
Unconditional variance:
∞ ∞
σ2
var ( Y t ) =var
(∑ ρ u ) (
τ =0
τ
t−τ
2
=σ ∑ ( ρ ) =
τ =0
2 τ
1−ρ2
<∞
)
var ( Y t )
What happens to for the following models?
(1)
Y t =α +ρY t−1 +ut with ρ=1
(2)
Y t =α +ρY t−1 +ut with ρ<1
Model 1:
t−1
Y t =Y 0 +tα + ∑ u t−τ
τ =0
E ( Y t |Y 0 ) =Y 0 +tα
and
var ( Y t|Y 0 )=tσ 2
∞
α
E(Y t )=α ∑ ( ρτ ) =
τ=0 1− ρ (Unconditional mean)
∞ ∞
2 τ σ2
var( Y t )=var
( ∑ ρτ (α +ut−τ ) =σ 2
τ=0
) ( ∑ (
τ=0
ρ ) =
) 2
1−ρ (Unconditional variance)
Both unconditional mean and variance are constant (independent of time)
Importance of Stationarity in Time-Series
The null hypothesis is generally defined as the presence of a unit root and the
alternative hypothesis is either stationarity, trend stationarity or explosive root
depending on the test used
Problem of spurious regressions - two variables trending over time can have a high R 2
even if the two are totally unrelated
Standard assumptions for asymptotic analysis will not be valid for non-stationary
variables - testing of hypothesis may not be valid
Phillips- Y t =θ2 Y t−1 +u t Y t =θ0 +θ2 Y t−1 +u t Y t =θ0 +θ1 t +θ2 Y t−1 +ut H 0 :γ 2 =0
Perron (PP)
ΔY t =θ0 +γ 2 Y t−1 +ut ; ΔY t =θ0 +θ1 t+γ 2 Y t−1 +ut ; H 1 :γ 2 <0
Test
γ 2=θ2 −1 γ 2=θ2 −1
ΔY t =θ2 Y t−1+u t ;
γ 2=θ2−1
Phillips-Perron Test Statistics
1 n2 σ^ ^ 2 2 γ^ 2 ( ^ρ−1 ) 1 n σ^ λ^ 2− γ^ 2
Z ( ρ )=n( ^ρ−1)− ( )( )
2 γ^ 2
( λ −^γ )
;
Z ( τ )= 2
^λ√ σ^
− ( ) ( )(
2 γ^ 2 ^λ2 )
Here, ρ^ = OLS estimate of; σ^ = Standard error of ρ^
n
∑ u^ 2t
γ^ 2= t=1
n−k (Sample variance of the least square residuals)
^λ2
= Newey-West long-run variance estimate of residuals
Choice of the Alternative Models:
Although the DF and the ADF tests are frequently used in testing for unit roots, there are problems of size distortions and low
power. In DF test, the problem of autocorrelation is not corrected for. There is problem of selection of lag length in ADF test.
The information criteria such as AIC or BIC often select a low value of the lag length.
The PP test is based on the similar equation as employed in the DF test (without the lagged differenced terms included in the
ADF test).
The PP test incorporates automatic non-parametric correction procedure for autocorrelated residuals, and usually gives the
same conclusions as the ADF tests
Monte Carlo studies suggest that the PP test has greater power than the ADF test
7. Making series stationary if there is unit root – (1) Detrending the series for the trend stationary process (TSP); or (2)
Differencing the series (DSP)
DF-GLS Test (Elliott, Rothenberg, and Stock, 1996)
While the ADF test corrects for higher-order serial correlation by adding lagged differenced terms as independent variables,
the PP test makes a nonparametric correction to account for the autocorrelation.
Monte Carlo studies suggest generally greater power of the PP test than the ADF test
The PP test is also robust to general forms of heteroscedasticity
The PP test does not require specification of lag length
However
The PP test is that it is based on asymptotic theory - works well only in large samples.
Also, like the ADF test, it is sensitive to structural breaks, poor small sample power often resulting in unit root conclusions.
The DF-GLS test is a modified version of the ADF test
The DF-GLS test has a higher power than the ADF tests
The DF-GLS tests the null hypothesis that the series is a random walk. There are two possible alternative hypotheses: (1) The series is
stationary about a linear time trend; or (2) It is stationary with a possibly non-zero mean but with no linear time trend.
The DF-GLS test proceeds by first de-trending the series as
d
Y t =Y t −α^ − β^ t
(a) Detrending Process using quasi difference series if there is a constant but no trend
Z t =Y t for t=1
7
( )
Z t =Y t − 1− Y for t= 2, 3, ..,T
T t−1
Similarly,
X t =1 for t=1
7
X t = for t= 2, 3, . .,T
T
Regressing Zt on Xt with no intercept
Z t =δX t +v t
Getting the de-trended series as
d
Y t =Y t −δ^
Subsequently, the DF-GLS test uses the following specification
p
ΔY dt =γ 2 Y dt−1 + ∑ λ j ΔY dt− j +ut
j=1
The DF-GLS test has the same null and alternative hypotheses as the ADF test, i.e.,
H 0 :γ 2 =0
H 1 :γ 2 <0
(b) Detrending Process using quasi difference series around deterministic trend with a constant
Z t =Y t for t=1
13. 5
(
Z t =Y t − 1−
T )Y t−1 for t= 2, 3, .. ,T
Similarly,
X 1 t =1 for t=1
13 .5
X 1 t= for t= 2, 3, . . ,T
T
X 2 t =1 for t=1
13 .5
(
X 2 t =t− 1−
T )
(t- 1) for t= 2, 3, . .,T
Z t =δ 1 X t +δX 2 t +v t
Getting the de-trended series as
d
Y t =Y t −δ^ 1 −δ^ 2 t
Subsequently, the DF-GLS test uses the following specification
p
ΔY dt =γ 2 Y dt−1 + ∑ λ j ΔY dt− j +ut
j=1
The DF-GLS test has the same null and alternative hypotheses as the traditional ADF test, i.e.,
H 0 :γ 2 =0
H 1 :γ 2 <0
Setting Maximum Lag Length
Information Criteria
1
T
Schwert Criteria (1989):
pmax =12×
100( ) 4
Autocorrelation Coefficient:
A partial correlation is a conditional correlation. It is the correlation between two variables controlling for influence of other
variables. It is defined as,
cov( y , x i|x −i )
r=
√ var( y|x−i )*var( x i|x−i )
Partial correlation (r) between y and x i (say x1) is the correlation between the two variables accounting for the influence of other
variables (say x2 and x3) on y and also on xi (say x1)
In regression, this partial correlation could be found by correlating the residuals from the following two regressions:
y=f ( x−i ) ; and (2) x i=f ( x−i )
(1)
For y=f ( x−i ) , the residuals give the part of y that is not predicted by other variables
For x i=f ( x−i ) , the residuals give the part of xi that is not predicted by other variables
Partial Autocorrelation Coefficient
It is defined as
k −1
^ρk − ∑ ( r^ k−1 )( j . ρ^ k− j)
j=1
r^ k = k −1
1−∑ ( r^ k−1 )( j. ρ^ j)
j=1
k
Y t =α + ∑ β j Y t− j +u t
For the regression equation j=1
Q Statistic
m
Q=n ∑ ρ 2k ~ χ 2m
Box-Pierce Statistic: k =1
(b) Unidirectional causality from Y to X: If at least one of the coefficients of lagged Y in the second equation are statistically
significant (the null hypothesis of the restricted F Test is rejected), whereas those of lagged X are not as a group in the first
equation ((the null hypothesis of the restricted F Test is not rejected)
(c) Feedback or bilateral causality: If the coefficients of both lagged X and lagged Y are statistically significant in both the
equations (when the coefficients are statistically significant in both the equations)
(d) Independence: If the coefficients of both lagged X and lagged Y are not statistically significant in both the equations (when
the coefficients are not statistically significant in either of the equations)
Decision Matrix on Granger Causality
If two time-series X and Y are cointegrated, there must exist Granger causality either from X to Y, or from Y to X or in both the
directions.
Presence of cointegration among the variables rules out the possibility of spurious regression. If the series are I(1) but not
cointegrated, Granger causality test may give misleading results unless the data are transformed to induce stationarity.
However, presence of Granger causality in either or both the directions between X and Y does not necessarily imply that the
series will be cointegrated.
Although cointegration indicates the presence or absence of Granger causality, it does not indicate in which direction causality
runs between the variables. The direction of Granger’s causality can be detected through the Vector Error Correction model of
long-run cointegrating vectors.
Following Oxley and Greasley (1998), a three-stage procedure can be used to test the direction of causality:
The first step tests for the order of integration (of natural logarithm) of the variables. If the variables are stationary, Granger
causality test can be carried out.
If the variables are not stationary, the second stage involves in investigating bivariate cointegration between the two
variables. If the variables have bivariate cointegration, Granger causality test can be carried out.
If bivariate cointegration is rejected, the variables are to be made stationary to carry out the Granger causality test.
Thus, there are three alternative specifications:
(a) When the variables are individually non-stationary i.e., I(1) but cointegrated
p q
Y t =α + ∑ ai Y t−i + ∑ b j X t− j +ut
i=1 j=1
r s
X t =β + ∑ ci X t−i + ∑ d j Y t− j +v t
i =1 j=1
(b) Cointegrated variables but use of I( 0) series with an error correction term (to capture short-run dynamics)
p q r s
ΔY t =α+ ∑ ai ΔY t−i + ∑ b j ΔX t− j +λ ECM t−1 +ut ΔX t =β+ ∑ c i ΔY t−i + ∑ d j ΔY t− j + θ ECM t−1 +v t
i=1 j=1 i=1 j=1
(c) When the variables are individually non-stationary, i.e., I(1) and not cointegrated
p q
ΔY t =α+ ∑ ai ΔY t−i + ∑ b j ΔX t− j +u t
i=1 j=1
r s
ΔX t =β+ ∑ c i ΔY t−i + ∑ d j ΔY t− j +v t
i=1 j=1
This is the structure of a simple ECM. The model relates change in one variable to change in another variable and the gap between
the two variables in the previous period.
It captures the short-run adjustments being guided by long-run theory. Here, the term ( x t−1 − y t−1 ) provides the short-run
disequilibrium adjustments. A test on is, therefore, the test for this disequilibrium component.
In a generalized form (4) can be rewritten as,
Δy t =α+β 2 Δx t +γ 1 x t−1 +γ 2 y t−1 +ut
Vector autoregression (VAR) (introduced by Sims, 1980) characterizes the joint dynamic behaviour of a collection of variables
without restrictions needed to identify the underlying structural parameters
A typical restriction takes the form of an assumption about the dynamic relationship between the pair of variables
A VAR system contains a set of m variables, each of which is expressed as a linear function of p lags of itself and of all of the
other m – 1 variables, plus an error term. However, one can also include exogenous variables such as seasonal dummies or
time trends in a VAR.
For two variables, X and Y, VAR of order p (without exogenous variable) is written as,
p p
Y t =α + ∑ ai Y t−i + ∑ b j X t− j +ut
i=1 j=1
p p
X t =β + ∑ ci X t−i + ∑ d j Y t− j +v t
i =1 j=1
If another variable Z is added to the system, there would be a third equation for Z and terms involving p lagged values of z would be
added to each of the three equations.
When the variables of a VAR are cointegrated, vector error-correction (VEC) model is estimated. A VEC model for two variables can
be expressed as
p p
ΔY t =α+ ∑ ai ΔY t−i + ∑ b j ΔX t− j −λ1 (Y t−1 −θ0 −θ 1 X t−1 )+u t
i=1 j=1
p p
ΔX t =β+ ∑ c i ΔY t−i + ∑ d j ΔY t− j −λ 2 (Y t−1−θ 0 −θ1 X t−1 )+v t
i=1 j=1
Here,
Y t =θ0 +θ1 X t
stands for long-run cointegrating relationship between X and Y and 1 and 2 are the error correction
parameters. These two parameters measure how the variables X and Y react to deviation from long run equilibrium.
In VEC model with more than two variables, there is possibility that more than one cointegrating relationship among the variables.
For example, if X, Y and Z tend to be equal in the long-run, Xt = Yt and Yt = Zt (or, Xt = Zt) would be the cointegrating relationships.
Impulse Response Function:
Large number of coefficients in the VAR model – difficulty in interpretation of coefficients
Impulse responses – tool of interpretation of VAR results
The impulse responses are the time‐paths of the variables response to shocks (u and v) - found by the recursion formula - functions
of the estimated VAR coefficients
Impulse variable - source of the shock; Response variable - the variable being affected
x t =a11 Lx t +a12 Ly t +u t
y t =a 21 Lx t + a22 Lyt + v t
1−a11 L −a12 L x t u
[ −a 21 L 1−a22 L y t ][ ] [ ]
= t
vt
−1
xt 1−a11 L −a12 L ut 1 1−a 22 L a12 L ut
[ ][yt
=
−a 21 L 1−a22 L ] [] [ =
v t Δ a21 L ][ ]
1−a 11 L v t
Or, Δ=(1−λ1 L)(1−λ2 L) with 1 and 2 being the two roots of the equation
When the stability condition is satisfied, x and y can be expressed as functions of the current and lagged values of u and v. These are
known as the impulse response functions. They show the current and lagged effects over time of changes in u and v on x and y.
Some Questions:
1. Let
y t =βx t +ut with ut =ρut−1 +v t and -1< ρ<1 Express this as ARDL model.
Ans.
y t =βx t + ρut−1 +v t
Again,
y t−1=βx t−1+ut−1 or y t−1−βx t−1=ut−1
y t =βx t + ρyt−1 −βρ x t−1 +v t
y t =β 0 x t + β1 Lx t +ρ Lyt +ut
(1−ρL) y t =( β 0 +β 1 L )x t +ut
( β 0 +β 1 L) ut
yt= xt +
(1−ρL) (1− ρL)
Ans.
y t =β 0 x t + β1 x t−1 +ρy t−1 +ut
Δxt =x t −x t−1
Δxt +x t−1=x t
−β 2
(
Δy t =β 0 Δxt +γ y t−1−
γ )
x t−1 +ut
−β 2 −( β 0 + β 1 )
Δy t =β 0 Δx t +γ ( y t−1−β 3 x t−1 ) +ut β 3= =
; γ γ
Δy t =β 0 Δxt +γ ( y t−1−β 3 x t−1 ) +ut
4. Write the ECM in lag operator form.
y t − y t−1= β0 xt − β0 x t−1 +γy t−1−γβ 3 xt−1 +u t
y t − y t−1−γy t−1 =β 0 x t −β 0 x t−1 −γβ 3 x t−1 +ut
y t −Ly t −γ Lyt =β 0 xt −β 0 Lxt −γβ 3 Lxt +ut
[1−(1+γ ) L] y t =( β 0− β0 L−γβ 3 L) x t +u t
γ ( β0 + β1)
[ {
[1−(1+γ ) L] y t = β 0− β 0 −
γ }]
L x t +u t
(1−ρL) y t =( β 0 +β 1 L )x t +ut
( β 0 +β 1 L) u
yt= xt + t
1−ρL 1−ρL
Δy t =θ0 +θ 1 Δx t +θ2 ( y t−1 −α−βx t−1 )+v t =(θ 0−αθ 2 )+θ1 Δx t +θ 2 y t−1 −θ2 βx t−1 +v t
In this setup, the coefficient of the lagged dependent variable is the coefficient on the error
correction mechanism and θ1 is the short-run effect of X on Y
t 0 t Δy =β Δx +γ y
( t−1 3 t−1 ) t −β x +u
6. Interpret the coefficients of the ECM
The model uses differences in both the dependent variables and the independent variables.
y −β x
Inclusion of the term ( t−1 3 t−1 ) reflects the assumption that X and Y have a long-term equilibrium relationship.
More specifically, any change in y is a sum of two effects: (i) the short-run impact of the
change in x on y, and (ii) the long-run impact of the deviation from the equilibrium in period t adjusted at each period (short-
run adjustments) at the rate γ.
Here, β0 captures the short-run relationship between X and Y. It indicates how Y and Y immediately change if X goes up one
period.
On the other hand, γ gives the rate at which the model re-equilibrates i.e. the speed at which it returns to its equilibrium
level. Formally, γ tells us the proportion of the disequilibrium which is corrected with each passing period.
This coefficient should be negative and less than the absolute value of one indicating its re-equilibrating properties. If γ = 0,
then the process never re-equilibrates and if γ = 1, re-equilibration occurs in one period.
7. Consider the following models:
1 β yt u
Now,
[ ][
1 α xt ][]
= t
vt
1 ut u β
| | |t |
1 v t v t −ut vt α αut −βv t
xt= = yt= =
1 β α−β 1 β α−β
| | | |
1 α and 1 α
Thus, both x and y are linear combinations of u and v.
y t =β 0 ut +β 1 Lut + β 2 Lu t
Invertibility requires
|λi|<1
−β 1±√ β 21−4 β 2
| |<1
2
β 1 +β 2 >−1 ; β 2−β 1 >−1 ; |β 2|<1
y t =α 1 Lyt +α 2 L2 y t +u t
(1−α 1 L−α 2 L2 ) y t =u t
1
yt= u
[(1−θ 1 L )(1−θ2 L )] t
2
Here, θ1 and θ2 are roots of the quadratic equation x −α 1 x−α 2 =0
Convergence requires
|θi|<1
α 1 ±√ α 12 +4 α 2
| |<1 y t =0.6 y t−1 +0.2 y t−2 +ut
2
α 1 +α 2 <1 ; α 1−α 2 >−1 ; |α 2|<1
Topics discussed in the class on 26/09/2020 (Saturday)
1. Application of unit root test: Example from the paper on instabilities in market concentration
2. Information criteria for lag length selection
3. Phillips-Perron test statistic (the mathematical details)
4. Autocorrelation function and partial autocorrelation function (PACF)
1. Assignment III for the Lab course (a) testing instabilities in index of industrial production (by use- based classification of industries); (b)
Examining impact of cropping intensity and irrigation facilities on use of chemical fertilizers in Indian agriculture sector
2. Focus Areas for (a): (i) Concept of instability, its implications and use of time-series econometrics; (ii) Splicing method (for converting
index number with different bases into a common base); (iii) interpretation of the results
3. Focus Areas for (b): (i) Measurement of the variables and their interpretation and implications; (ii) Different aspects of estimation (to be
discussed further)