SlideShare a Scribd company logo
1 of 64
TIME SERIES
ECONOMETRICS
ARIMA MODEL
Types of Data
• Time Series Data,
• Cross-Sectional Data and
• Panel Data.
Time Series Data
• Time series data, as the name suggests, are data that have been
collected over a period of time on one or more variables.
• Time series data have associated with them a particular frequency
of observation or frequency of collection of data points.
• The frequency is simply a measure of the interval over, or the
regularity with which, the data are collected or recorded.
Problems that could be tackled using time series data:
• How the value of a country’s stock index has varied with that
country’s macroeconomic fundamentals.
• How the value of a company’s stock price has varied when it
announced the value of its dividend payment.
• The effect on a country’s exchange rate of an increase in its trade
deficit.
In all of the above cases, it is clearly the time dimension which is the
most important, and the analysis will be conducted using the values
of the variables over time.
Cross-Sectional Data
Cross-sectional data are data on one or more variables collected at a
single point in time. For example, the data might be on:
• A poll of usage of internet stockbroking services.
• A cross-section of stock returns on the New York Stock Exchange
(NYSE)
• A sample of bond credit ratings for UK banks.
• Data on sales volume, sales revenue, number of customers and
expenses for the past month at each Starbucks location.
Problems that could be tackled using cross-sectional data:
• The relationship between company size and the return to investing
in its shares.
• The relationship between a country’s GDP level and the probability
that the government will default on its sovereign debt.
Panel, Longitudinal, or Micropanel Data
Panel data have the dimensions of both time series and cross-
sections, e.g. the daily prices of a number of blue chip stocks over
two years.
Cross sectional data is collected
at a particular point of time...For
instance, say your are studying the
GDP of 3 developing countries in
year 1999 only .... So u have data
like this :-
This is your cross sectional data since
you are studying the entities (Countries
here) at some point of time i.e. (year
1999 here) .
Country Time GDP
India 1999 ----
China 1999 ----
Brazil 1999 ----
Panel data is when you study again say GDP
of 3 developing countries over a period of time
(say 3 yrs. from 1999 to 2001)….So u have data
like this:
Here you are studying the same entities i.e.
(China, India and Brazil) over a period of time
i.e. 3 yrs. from 1999 to 2001. This is called panel
data.
Country Time GDP
India 1999 ----
India 2000 ----
India 2001 ----
China 1999 ----
China 2000 ----
China 2001 ----
Brazil 1999 ----
Brazil 2000 ----
Brazil 2001 ----
Balanced and Unbalanced Panel Data
• If all the companies/ person/
entities have the same number
of observations, we have what is
called a balanced panel.
• Balanced panel, as each person
is observed every year
• If the number of observations
is not the same for each
company, companies/ person/
entity it is called an
unbalanced panel.
• Unbalanced panel, since
person 1 is not observed in
year 2003 and person 3 is not
observed in 2003 or 2001.
Repeated Cross-Sections or Pooled Cross-Sections.
• There is also a type of data that's in between cross-sectional
data and panel data. It is typically called repeated cross-
sections or pooled cross-sections.
• For example, annual labour force surveys are repeated cross-
sections, because every year, a new random sample is taken from
the population. In this case, there is a time component, so
these are not cross-sectional data, but every year, new individuals
are surveyed, so these are also not panel data. That's why these
are called repeated cross-sections.
What is a time series?
A time series is any series of data that varies over time. For example
• Monthly Tourist Arrivals from Other countries.
• Quarterly GDP of USA.
• Hourly price of stocks and shares.
• Weekly quantity of beer sold in a pub
Because of widespread availability of time series databases most
empirical studies use time series data.
Definition of Time Series: An ordered sequence of values of a
variable at equally spaced time intervals.
Applications Of Time Series Analysis
Time Series Analysis is used for many applications such as:
• Economic Forecasting
• Sales Forecasting
• Budgetary Analysis
• Stock Market Analysis
• Yield Projections
• Process and Quality Control
• Inventory Studies
• Workload Projections
• Utility Studies
• Census Analysis.
Time Series Data
• One of the important and frequent types of data used in
empirical analysis.
• But it poses several challenges to
econometricians/practitioners. E.g.
1. Empirical work based on time series data assumes that the
underlying time series is stationary.
2. Autocorrelation: because the underlying time series data is
non-stationary.
3. Spurious/nonsense regression: a very high R2 and
significant regression coefficients (though there is no
meaningful relationship between the two variables)
Caveats in Using Time Series Data in
Applied Econometric Modeling
• Data Should be Stationary
• Presence of Autocorrelation
• Guard Against Spurious Regressions
• Establish Cointegration
• Reconcile SR with LR Behavior via ECM
• Implications to Forecasting
• Possibility of Volatility Clustering
Examples of Stationary Time Series
-6000
-4000
-2000
0
2000
4000
6000
8000
92 94 96 98 00 02 04
AUST
-6000
-4000
-2000
0
2000
4000
6000
92 94 96 98 00 02 04
CAN
-2000
-1000
0
1000
2000
3000
4000
92 94 96 98 00 02 04
CHI
-6000
-4000
-2000
0
2000
4000
6000
92 94 96 98 00 02 04
GERM
-12000
-8000
-4000
0
4000
8000
12000
92 94 96 98 00 02 04
HONG
-12000
-8000
-4000
0
4000
8000
12000
92 94 96 98 00 02 04
JAP
-12000
-8000
-4000
0
4000
8000
12000
92 94 96 98 00 02 04
KOR
-3000
-2000
-1000
0
1000
2000
92 94 96 98 00 02 04
MAL
-3000
-2000
-1000
0
1000
2000
3000
92 94 96 98 00 02 04
SING
-8000
-4000
0
4000
8000
12000
92 94 96 98 00 02 04
TWN
-4000
-3000
-2000
-1000
0
1000
2000
3000
4000
5000
92 94 96 98 00 02 04
UKK
-20000
-10000
0
10000
20000
30000
92 94 96 98 00 02 04
US
2000
4000
6000
8000
10000
12000
14000
16000
92 94 96 98 00 02 04
AUSTRALIA
0
2000
4000
6000
8000
10000
12000
92 94 96 98 00 02 04
CANADA
0
2000
4000
6000
8000
10000
92 94 96 98 00 02 04
CHINA
1000
2000
3000
4000
5000
6000
7000
8000
9000
92 94 96 98 00 02 04
GERMANY
0
4000
8000
12000
16000
20000
24000
92 94 96 98 00 02 04
HONGKONG
10000
15000
20000
25000
30000
35000
40000
45000
92 94 96 98 00 02 04
JAPAN
0
10000
20000
30000
40000
50000
92 94 96 98 00 02 04
KOREA
0
1000
2000
3000
4000
5000
6000
7000
92 94 96 98 00 02 04
MALAYSIA
0
1000
2000
3000
4000
5000
6000
7000
92 94 96 98 00 02 04
SINGAPORE
0
5000
10000
15000
20000
25000
30000
92 94 96 98 00 02 04
TAIWAN
0
2000
4000
6000
8000
10000
12000
92 94 96 98 00 02 04
UK
10000
20000
30000
40000
50000
60000
92 94 96 98 00 02 04
USA
Examples of Non-Stationary Time Series
Stationary Processes
• Stationary Processes: A stochastic process is said to be
stationary/ weakly /covariance/2nd-order stationary if:
o Its mean and variance are constant over time, and
o The value of the covariance between the two time periods depends only on
the distance/lag between the two time periods and not the actual time at
which the covariance is computed.
o E.g. let’s Yt be a stochastic process, then;
• Mean: E(Yt ) = µ …………………………………………..
(1)
• Variance: var (Yt ) = E(Yt − µ)2 = σ2 ………………………………..
(2)
• Covariance: γk = E[(Yt − µ)(Yt +k − µ)] ……………..…………
(3)
• Where γk, the covariance (or auto-covariance) at lag k,
• If k = 0, we obtain γ0, which is simply the variance of Y (= σ2); if k = 1,
γ1 is the covariance between two adjacent values of Y
Stationarity
A time series has stationarity if a shift in time doesn’t cause a
change in the shape of the distribution. Basic properties of the
distribution like the mean , variance and covariance are constant
over time.
Why are Stationary Time Series so
Important?
• Because if a time series is non-stationary, we can study its
behavior only for the time period under consideration, and as a
consequence, it is not possible to generalize it to other time
periods.
• Therefore, for the purpose of forecasting, such (non-stationary)
time series may be of little practical value.
• Non-stationary Stochastic Processes: Although our
interest is in stationary time series, one often encounters non-
stationary time series
• A non-stationary time series will have a time-varying mean
or a time-varying variance or both.
Transformations to Achieve Stationarity
If the time series is not stationary, we can often transform it to stationarity with one
of the following techniques.
1. We can difference the data. That is, given the series Yt, we create the
new series Yi = Yi − Yi−1. The differenced data will contain one less point
than the original data. Although you can difference the data more than
once, one difference is usually sufficient.
2. If the data contain a trend, we can fit some type of curve to the data
and then model the residuals from that fit. Since the purpose of the fit is
to simply remove long term trend, a simple fit, such as a straight line, is
typically used.
3. For non-constant variance, taking the logarithm or square root of the
series may stabilize the variance. For negative data, you can add a
suitable constant to make all the data positive before applying the
transformation. This constant can then be subtracted from the model to
obtain predicted (i.e., the fitted) values and forecasts for future points.
The above techniques are intended to generate series with constant location and
scale. Although seasonality also violates stationarity, this is usually explicitly
incorporated into the time series model
• Seasonality rules out series (d), (h) and (i).
• Trend rules out series (a), (c), (e), (f) and (i).
• Stationary series (b) and (g).
Differencing
In Figure below, the Google stock price was non-stationary in panel (a), but the
daily changes were stationary in panel (b). This shows one way to make a non-
stationary time series stationary — compute the differences between consecutive
observations. This is known as differencing.
1st Diff. 2nd Diff.
Year Quarter gdp loggdp d1loggdp d2loggdp
1947 1 1570.519 7.359161
1947 2 1568.653 7.357973 -0.001189
1947 3 1567.966 7.357535 -0.000438 0.000751
1947 4 1590.938 7.372079 0.014545 0.014983
1948 1 1616.069 7.387752 0.015673 0.001128
1948 2 1644.637 7.405275 0.017523 0.001850
1948 3 1654.061 7.410989 0.005714 -0.011809
1948 4 1657.988 7.413360 0.002371 -0.003342
1949 1 1633.249 7.398327 -0.015034 -0.017405
1949 2 1628.439 7.395377 -0.002949 0.012084
1949 3 1646.698 7.406527 0.011150 0.014100
1949 4 1629.911 7.396281 -0.010247 -0.021397
1950 1 1696.765 7.436479 0.040198 0.050445
1950 2 1747.322 7.465840 0.029361 -0.010837
1950 3 1815.845 7.504306 0.038467 0.009106
1950 4 1848.928 7.522361 0.018055 -0.020412
1951 1 1871.311 7.534395 0.012033 -0.006022
1951 2 1903.118 7.551249 0.016854 0.004821
1951 3 1941.109 7.571015 0.019766 0.002912
1951 4 1944.447 7.572733 0.001718 -0.018048
Differencing the data
EViews Commands
Log Loggdp = log(gdp)
1st Difference d1loggdp=d(loggdp)
1st Difference + Log d1loggdp=dlog(gdp)
2nd Difference
d2loggdp=d(d1loggdp)
or
d2loggdp=d(loggdp,2)
or
d2loggdp=dlog(gdp,2)
• First take log of series and check stationarity because standard
unit root tests assumes a linear structure.
• If log transformed series is not stationarity then do differencing of
log transformed series.
GDP Lag1 Lag2
1570.519
1568.653 1570.519
1567.966 1568.653 1570.519
1590.938 1567.966 1568.653
1616.069 1590.938 1567.966
1644.637 1616.069 1590.938
1654.061 1644.637 1616.069
1657.988 1654.061 1644.637
1633.249 1657.988 1654.061
1628.439 1633.249 1657.988
1646.698 1628.439 1633.249
1629.911 1646.698 1628.439
1696.765 1629.911 1646.698
1747.322 1696.765 1629.911
7.2
7.6
8.0
8.4
8.8
9.2
9.6
50 55 60 65 70 75 80 85 90 95 00 05
LOGGDP
0
2,000
4,000
6,000
8,000
10,000
12,000
50 55 60 65 70 75 80 85 90 95 00 05
GDP
Transformations to Achieve Stationarity
A seasonal difference is
the difference between
an observation and the
corresponding
observation from the
previous year. So,
Y′t = Yt−Yt−m
where m = number of
seasons.
These are also called
“lag-m differences” as
we subtract the
observation after a lag of
m periods.
• Sometimes it is
necessary to do both a
seasonal difference and
a first difference to
obtain stationary data,
as shown in Figure.
• Here, the data are first
transformed using
logarithms (second
panel).
• Then seasonal
differenced are
calculated (third panel).
• The data still seem a
little non-stationary,
and so a further lot of
first differences are
computed (bottom
panel).
US net electricity generation (billion kWh).
Types of Stationary
Models can show different types of stationarity:
• Strict stationarity means that the joint distribution of any moments of any degree
(e.g. expected values, variances, third order and higher moments) within the
process is never dependent on time. This definition is in practice too strict to be
used for any real-life model.
• First-order stationarity series have means that never changes with time. Any
other statistics (like variance) can change.
• Second-order stationarity (also called weak stationarity) time series have a
constant mean, variance and an autocovariance that doesn’t change with time.
Other statistics in the system are free to change over time. This constrained
version of strict stationarity is very common.
• Trend-stationary models fluctuate around a deterministic trend (the series
mean). These deterministic trends can be linear or quadratic, but the amplitude
(height of one oscillation) of the fluctuations neither increases nor decreases
across the series.
• Difference-stationary models are models that need one or more differencing's to
become stationary.
• We call a stochastic process(time series) purely random/white noise process
if it has zero mean, constant variance σ2, and is serially
uncorrelated i.e. [ut ∼ IIDN(0, σ2)].
• Note: Here onward, in all equations the assumption of “white noise” will be
applicable on ut .
White Noise Processes
White Noise Processes
Unit Root Testing: Formal Tests to
Establish Stationarity of Time Series
• Dickey-Fuller (DF) Test
• Augmented Dickey-Fuller (ADF) Test
• Phillips-Perron (PP) Unit Root Test
• Dickey-Pantula Unit Root Test
• GLS Transformed Dickey-Fuller Test
• ERS Point Optimal Test
• KPSS Test (run as a complement to the unit root tests)
• Ng and Perron Test
Some Useful Models for Time Series
1. A purely random process,
2. A random walk,
3. A moving average (MA) process,
4. An autoregressive (AR) process,
5. An autoregressive moving average (ARMA) process, and
6. An autoregressive integrated moving average (ARIMA)process.
Estimation of AR, MA, and ARMA Models
 Testing Goodness of Fit
• When an AR, MA, or ARMA model has been fitted to a given
time series, it is advisable to check that the model does really
give an adequate description of the data
• There are two criteria often used that reflect the closeness of fit
and the number of parameters estimated.
• One is the Akaike Information Criterion (AIC), and the other is
the Schwartz Bayesian Criterion (SBC)/ Bayesian information
criterion (BIC).
The Box-Jenkins Approach
• The Box-Jenkins approach is one of the most widely used
methodologies for the analysis of time-series data
• It is popular because of its generality; it can handle any series,
stationary or not, with or without seasonal elements, and it has
well-documented computer programs.
• Although Box and Jenkins have been neither the originators nor the
most important contributors in the field of ARMA models
• They have popularized these models and made them readily
accessible to everyone, so much that ARMA models are sometimes
referred to as Box-Jenkins models.
The Box-Jenkins Approach
 The basic steps in the Box-Jenkins methodology are
1. Differencing the series so as to achieve Stationarity,
2. Identification of a tentative model,
3. Estimation of the model,
4. Diagnostic checking (if the model is found inadequate, we go
back to step 2), and
5. Using the model for forecasting and control.
The Box-Jenkins Approach
1. Differencing to achieve Stationarity: How do we conclude
whether a time series is stationary or not?
• We can do this by studying the graph of the correlogram of the series.
• The correlogram of a stationary series drops off as k, the number of
lags, becomes large, but this is not usually the case for a non-
stationary series.
• Thus, the common procedure is to plot the correlogram of the given
series Yt and successive differences ΔY, ΔY, and so on, and look at the
correlograms at each stage.
• We keep differencing until the correlogram dampens.
The Box-Jenkins Approach
2. Once we have used the differencing procedure to get a
stationary time series, we examine the correlogram to decide on
the appropriate orders of the AR and MA components.
• The correlogram of a MA process is zero after a point.
• That of an AR process declines geometrically. The correlograms
of ARMA processes show different patterns (but all dampen
after a while).
• Based on these, one arrives at a tentative ARMA model.
• This step involves more of a judgmental procedure than the use
of any clear-cut rules.
The Box-Jenkins Approach
https://people.duke.edu/~rnau/411arim3.htm
http://rinterested.github.io/statistics/arima.html
http://qualityamerica.com/LSS-Knowledge-
Center/qualityimprovementtools/interpreting_an_autoc
orrelation_chart.php
The Box-Jenkins Approach
3. The next step is the estimation of the tentative ARMA model
identified in step 2. We have discussed in the preceding section
the estimation of ARMA models.
4. The next step is diagnostic checking to check the adequacy of the
tentative model. We discussed in the preceding section the Q and
Q* statistics commonly used in diagnostic checking. As argued
there, the (^-statistic is inappropriate in autoregressive models
and thus we need to replace it with some LM test statistic.
5. The final step is forecasting.
Differencing the series
to achieve Stationarity
Identify model to be
tentatively entertained
Estimate the parameters
of the tentative model
Diagnostic checking. Is
the model adequate?
No
Yes
Use the model for
forecasting and
control
Approaches to Economic Forecasting
The Box-Jenkins Approach
The defining characteristics of AR, MA and ARMA processes:
An autoregressive process has:
• a geometrically decaying acf
• a number of non-zero points of pacf = AR order.
A moving average process has:
• a geometrically decaying pacf.
• number of non-zero points of acf = MA order
A combination autoregressive moving average process has:
• a geometrically decaying acf
• a geometrically decaying pacf.
• Diagnostic Checking: Diagnostic checking consists of evaluating the
adequacy of the estimated model. Considerable skill is required to
choose the actual ARIMA (p,d,q) model so that the residuals
estimated from this model are white noise. So the autocorrelations
of the residuals are to be estimated for the diagnostic checking of
the model. These are also judged by Ljung-Box statistic under null
hypothesis that autocorrelation co-efficient is equal to zero.
MA(1) model
Sample autocorrelation and partial autocorrelation functions for an MA(1) model:
• The MA(1) has an acf that is significant for only lag 1, while the pacf declines
geometrically, and is significant until lag 7.
• The acf at lag 1 and all of the pacfs are negative as a result of the negative
coefficient in the MA generating process.
MA(2) model
Sample autocorrelation and partial autocorrelation functions for an MA(2) model:
• The first two autocorrelation coefficients only are significant, while the partial
autocorrelation coefficients are geometrically declining.
• Since, the second coefficient on the lagged error term in the MA is negative, the
acf and pacf alternate between positive and negative.
AR(1) model
Sample autocorrelation and partial autocorrelation functions for an AR(1) model:
• The AR(1) has an pacf that is significant for only lag 1, while the acf declines
geometrically.
• Only the first pacf coefficient is significant, while all others are virtually zero and
are not significant.
AR(1) model
Sample autocorrelation and partial autocorrelation functions for an AR(1) model:
• AR(1), which was generated using identical error terms, but a much smaller
autoregressive coefficient. In this case, the autocorrelation function dies away
much more quickly than in the previous example, and in fact becomes
insignificant after around five lags.
AR(1) model
• Sample autocorrelation and partial autocorrelation functions for a non-stationary
model (i.e. a unit coefficient):
• On some occasions, the acf does die away for a non-stationary process.
• The pacf, however, is significant only for lag 1, correctly suggesting that an
autoregressive model with no moving average term is most appropriate.
ARMA(1, 1) model
• Sample autocorrelation and partial autocorrelation functions for an ARMA(1, 1)
model:
• In such a process, both the acf and the pacf decline geometrically – the acf as a
result of the AR part and the pacf as a result of the MA part.
GDP
• Check through AIC and BIC criteria.
• Imp note: ARIMA model output differs in EViews 8 and later
versions. EViews 8 and former versions estimate ARIMA model on
the basis of conditional least squares method. Whereas version 9
and newer are based on ARIMA forecasting through maximum
likelihood method.
• Therefore AIC & BIC values will also differ in EViews 8 and later
versions.
SBIC & AIC code results with EViews 8
SBC MA(0) MA(1) MA(2) MA(3) MA(4) MA(5)
AR(0) --- -6.464488 -6.482979 -6.465608 -6.443753 -6.439384
AR(1) -6.490116 -6.471629 -6.462397 -6.440327 -6.448238 -6.425663
AR(2) -6.471464 -6.456046 -6.497155 -6.437803 -6.433839 -6.405617
AR(3) -6.462250 -6.453936 -6.520738 -6.505968 -6.488171 -6.446302
AR(4) -6.448194 -6.429773 -6.509408 -6.487179 -6.503980 -6.493222
AR(5) -6.428779 -6.407640 -6.384756 -6.389125 -6.518437 -6.500290
Best Model Is ARMA(3,2)
AIC MA(0) MA(1) MA(2) MA(3) MA(4) MA(5)
AR(0) --- -6.493237 -6.526104 -6.523107 -6.515626 -6.525633
AR(1) -6.518950 -6.514880 -6.520065 -6.512412 -6.534741 -6.526582
AR(2) -6.514843 -6.513885 -6.569454 -6.524561 -6.535057 -6.521295
AR(3) -6.520261 -6.526449 -6.607754 -6.607487 -6.604193 -6.576826
AR(4) -6.520924 -6.517048 -6.611230 -6.603546 -6.634893 -6.638681
AR(5) -6.516316 -6.509766 -6.501471 -6.520429 -6.664330 -6.660773
Best Model Is ARMA(5,4)
AIC Vs SBIC
• When large observations (data points) are available select model
as per AIC. This is because AIC will always select higher model than
SBIC. Higher model (eg. ARMA 5,5) means loss of 5 data points,
therefore more data should be available which can compensate
for loss of lags.
Computing Summary Statistics
• UKHP.xls
• Import into Eviews.
• Calculate simple
percentage changes in the
series:
dhp = 100*(hp-hp(-1))/hp(-1)
• To obtain descriptive
summary statistics of a
series select Quick/Series
Statistics/Histogram and
Stats and type the name of
the variable (DHP)
ARMA(p,q)
Model Equation
ARMA(1,1) dhp c ar(1) ma(1)
OBS DHP
1991M01
1991M02 0.83895
1991M03 -1.12892 0.83895
1991M04 1.483326 -1.12892
1991M05 1.319533 1.483326
1991M06 1.326908 1.319533
1991M07 -1.02755 1.326908
1991M08 -0.91749 -1.02755
1991M09 -1.44568 -0.91749
1991M10 0.388974 -1.44568
1991M11 -0.83902 0.388974
1991M12 0.162323 -0.83902
1992M01 -2.17099 0.162323
1992M02 -0.19444 -2.17099
1992M03 0.245605 -0.19444
1992M04 0.00058 0.245605
1992M05 1.413615 0.00058
1992M06 0.366004 1.413615
1992M07 -0.33771 0.366004
1992M08 -0.78737 -0.33771
1992M09 -2.32213 -0.78737
1992M10 -0.77807 -2.32213
1992M11 -2.16517 -0.77807
1992M12 0.288302 -2.16517
1993M01 0.665036 0.288302
1993M02 0.597487 0.665036
1993M03 -0.37186 0.597487
1993M04 2.968641 -0.37186
1993M05 -1.50191 2.968641
ARMA(p,q)
ARMA(5,5) dhp c ar(1) ar(2) ar(3) ar(4) ar(5) ma(1) ma(2) ma(3) ma(4) ma(5)
AIC selects an ARMA(4,5), while SBIC selects the smaller ARMA(2,0) model i.e. an
AR(2).
The values of all of the Akaike and Schwarz information criteria calculated using EViews are as
follows.
ARMA(5,5) Model
Why forecast?
Some examples in finance of where forecasts from econometric
models might be useful include:
• Forecasting tomorrow’s return on a particular share.
• Forecasting the price of a house given its characteristics.
• Forecasting the riskiness of a portfolio over the next year.
• Forecasting the volatility of bond returns
• Forecasting the correlation between US and UK stock market
movements tomorrow
• Forecasting the likely number of defaults on a portfolio of home
loans.
Checking Forecasting Accuracy
• Root Mean Squared Error: the smaller the error, the better the
forecasting ability of that model
• Theil Inequality Coefficient: always lies between zero and one,
where zero indicates a perfect fit.
http://www.eviews.com/help/helpintro.html#page/content/Forecast-Forecast_Basics.html
http://www.eviews.com/help/helpintro.html#page/content%2FForecast-
An_Illustration.html%23
Diagnostic Checking
• How do we know that the model is a reasonable fit to the data?
• One simple diagnostic is to obtain residuals from identified
Equation and obtain the ACF and PACF of these residuals, say, up
to lag 25 (or 1/3 or ¼ ) of total observations.
• In estimated AC and PACF Figure, none of the autocorrelations and
partial autocorrelations should be individually statistically
significant. Nor is the sum of the 25 squared autocorrelations, as
shown by the Box–Pierce Q and Ljung–Box (LB) statistics,
statistically significant.
• In other words, the correlograms of both autocorrelation and
partial autocorrelation give the impression that the residuals
estimated from Equation are purely random. Hence, there may
not be any need to look for another ARIMA model
Diagnostic Checking
Serial Correlation (Auto-Correlation)
• Normally serial correlation (auto-correlation) in regression model
(Y= a + bX) is detected through Durbin Watson Statistics (DW).
• However, when regression model is autoregressive (AR) in nature (Y
= a + bYt-1), Durbin Watson Statistics (DW) will give invalid results.
• Durbin Watson Statistics (DW) can be used for only for AR models
up to one lag.
• Testing serial correlation in Eviews:
1. Correlogram:
a. ACF and PACF.
b. the Ljung-Box (LB) Q-statistics and their p-values.
2. The Langrange Multiplier (LM test) (given by Breusch and
Godfrey)
• Model accuracy is generally assessed by using the root
mean squared error criterion.
• Calculate all the errors, square them, calculate the average, and
then take the square root of that average.
• The model with the lowest mean-squared error is judged the
most accurate
Diagnostic Checking Cont.……
Autoregressive (AR) Forecasting Equation
• To make forecasts j years into the future, using a third-order
autoregressive model (AR3), you need only the most recent p = 3
values (Yn, Yn-1 and Yn-2 ) and the regression estimates a0, a1, a2, and
a3.
To forecast one year ahead, Equation becomes:
To forecast two years ahead, Equation becomes:
To forecast two years ahead, Equation becomes:
Moving Average (MA) Forecasting Equation
• An autoregression of the residual error time series is called a
Moving Average (MA) model. This is confusing because it has
nothing to do with the moving average smoothing process. Think of
it as the sibling to the autoregressive (AR) process, except on
lagged residual error rather than laged raw observations.
MA order Regression Equation
MA(1) Yt = c + Ut-1
MA(2) Yt = c + Ut-1 + Ut-2
U = error
Seasonality
Time series that show regular patterns of movement within a
year across years.
• Seasonal lags are most often included as a lagged value
one year before the prior value.
• We detect such patterns through the autocorrelations in the
data.
• For quarterly data, the fourth autocorrelation will not be
statistically zero if there is quarterly seasonality.
 For monthly, the 12th, and so on.
• To correct for seasonality, we can include an additional lagged
term to capture the seasonality.
 For quarterly data, we would include a prior year quarterly seasonal lag
as
𝑥𝑡 = 𝑏0 + 𝑏1𝑥𝑡−1 + 𝑏2𝑥𝑡−4 + ε𝑡

More Related Content

What's hot

Time series analysis in Stata
Time series analysis in StataTime series analysis in Stata
Time series analysis in Statashahisec1
 
Econometric model ing
Econometric model ingEconometric model ing
Econometric model ingMatt Grant
 
Analysis of Time Series
Analysis of Time SeriesAnalysis of Time Series
Analysis of Time SeriesManu Antony
 
Time Series
Time SeriesTime Series
Time Seriesyush313
 
Time series and forecasting
Time series and forecastingTime series and forecasting
Time series and forecastingmvskrishna
 
Advanced Econometrics by Sajid Ali Khan Rawalakot: 0334-5439066
Advanced Econometrics by Sajid Ali Khan Rawalakot: 0334-5439066Advanced Econometrics by Sajid Ali Khan Rawalakot: 0334-5439066
Advanced Econometrics by Sajid Ali Khan Rawalakot: 0334-5439066Sajid Ali Khan
 
Arima model
Arima modelArima model
Arima modelJassika
 
Forecasting techniques, time series analysis
Forecasting techniques, time series analysisForecasting techniques, time series analysis
Forecasting techniques, time series analysisSATISH KUMAR
 
Time Series - Auto Regressive Models
Time Series - Auto Regressive ModelsTime Series - Auto Regressive Models
Time Series - Auto Regressive ModelsBhaskar T
 
Auto Correlation Presentation
Auto Correlation PresentationAuto Correlation Presentation
Auto Correlation PresentationIrfan Hussain
 

What's hot (20)

Autocorrelation
AutocorrelationAutocorrelation
Autocorrelation
 
Time series forecasting
Time series forecastingTime series forecasting
Time series forecasting
 
Time series analysis in Stata
Time series analysis in StataTime series analysis in Stata
Time series analysis in Stata
 
Panel Data Models
Panel Data ModelsPanel Data Models
Panel Data Models
 
Time series
Time seriesTime series
Time series
 
Econometric model ing
Econometric model ingEconometric model ing
Econometric model ing
 
Analysis of Time Series
Analysis of Time SeriesAnalysis of Time Series
Analysis of Time Series
 
Time Series
Time SeriesTime Series
Time Series
 
Difference-in-Difference Methods
Difference-in-Difference MethodsDifference-in-Difference Methods
Difference-in-Difference Methods
 
Regression models for panel data
Regression models for panel dataRegression models for panel data
Regression models for panel data
 
Time Series Decomposition
Time Series DecompositionTime Series Decomposition
Time Series Decomposition
 
Time series analysis
Time series analysisTime series analysis
Time series analysis
 
Time series and forecasting
Time series and forecastingTime series and forecasting
Time series and forecasting
 
Advanced Econometrics by Sajid Ali Khan Rawalakot: 0334-5439066
Advanced Econometrics by Sajid Ali Khan Rawalakot: 0334-5439066Advanced Econometrics by Sajid Ali Khan Rawalakot: 0334-5439066
Advanced Econometrics by Sajid Ali Khan Rawalakot: 0334-5439066
 
Arima model
Arima modelArima model
Arima model
 
Forecasting techniques, time series analysis
Forecasting techniques, time series analysisForecasting techniques, time series analysis
Forecasting techniques, time series analysis
 
ARIMA
ARIMA ARIMA
ARIMA
 
Time Series - Auto Regressive Models
Time Series - Auto Regressive ModelsTime Series - Auto Regressive Models
Time Series - Auto Regressive Models
 
time series analysis
time series analysistime series analysis
time series analysis
 
Auto Correlation Presentation
Auto Correlation PresentationAuto Correlation Presentation
Auto Correlation Presentation
 

Similar to TIME SERIES ECONOMETRICS ARIMA MODEL

Module 3 - Time Series.pptx
Module 3 - Time Series.pptxModule 3 - Time Series.pptx
Module 3 - Time Series.pptxnikshaikh786
 
R Packages for Time-Varying Networks and Extremal Dependence
R Packages for Time-Varying Networks and Extremal DependenceR Packages for Time-Varying Networks and Extremal Dependence
R Packages for Time-Varying Networks and Extremal DependenceWork-Bench
 
Application of panel data to the effect of five (5) world development indicat...
Application of panel data to the effect of five (5) world development indicat...Application of panel data to the effect of five (5) world development indicat...
Application of panel data to the effect of five (5) world development indicat...Alexander Decker
 
Application of panel data to the effect of five (5) world development indicat...
Application of panel data to the effect of five (5) world development indicat...Application of panel data to the effect of five (5) world development indicat...
Application of panel data to the effect of five (5) world development indicat...Alexander Decker
 
Applied Statistics Chapter 2 Time series (1).ppt
Applied Statistics Chapter 2 Time series (1).pptApplied Statistics Chapter 2 Time series (1).ppt
Applied Statistics Chapter 2 Time series (1).pptswamyvivekp
 
Time Series Analysis and Forecasting.ppt
Time Series Analysis and Forecasting.pptTime Series Analysis and Forecasting.ppt
Time Series Analysis and Forecasting.pptssuser220491
 
Basic Concepts of Statistics - Lecture Notes
Basic Concepts of Statistics - Lecture NotesBasic Concepts of Statistics - Lecture Notes
Basic Concepts of Statistics - Lecture NotesDr. Nirav Vyas
 
Enterprise_Planning_TimeSeries_And_Components
Enterprise_Planning_TimeSeries_And_ComponentsEnterprise_Planning_TimeSeries_And_Components
Enterprise_Planning_TimeSeries_And_Componentsnanfei
 
Demand forecasting by time series analysis
Demand forecasting by time series analysisDemand forecasting by time series analysis
Demand forecasting by time series analysisSunny Gandhi
 
Weather forecasting model.pptx
Weather forecasting model.pptxWeather forecasting model.pptx
Weather forecasting model.pptxVisheshYadav12
 
Holtwinters terakhir lengkap
Holtwinters terakhir lengkapHoltwinters terakhir lengkap
Holtwinters terakhir lengkapZulyy Astutik
 
Storm Prediction data analysis using R/SAS
Storm Prediction data analysis using R/SASStorm Prediction data analysis using R/SAS
Storm Prediction data analysis using R/SASGautam Sawant
 

Similar to TIME SERIES ECONOMETRICS ARIMA MODEL (20)

analysix.ppt
analysix.pptanalysix.ppt
analysix.ppt
 
Module 3 - Time Series.pptx
Module 3 - Time Series.pptxModule 3 - Time Series.pptx
Module 3 - Time Series.pptx
 
Panel data content
Panel data contentPanel data content
Panel data content
 
Time series
Time seriesTime series
Time series
 
R Packages for Time-Varying Networks and Extremal Dependence
R Packages for Time-Varying Networks and Extremal DependenceR Packages for Time-Varying Networks and Extremal Dependence
R Packages for Time-Varying Networks and Extremal Dependence
 
Application of panel data to the effect of five (5) world development indicat...
Application of panel data to the effect of five (5) world development indicat...Application of panel data to the effect of five (5) world development indicat...
Application of panel data to the effect of five (5) world development indicat...
 
Application of panel data to the effect of five (5) world development indicat...
Application of panel data to the effect of five (5) world development indicat...Application of panel data to the effect of five (5) world development indicat...
Application of panel data to the effect of five (5) world development indicat...
 
Forecasting
ForecastingForecasting
Forecasting
 
Applied Statistics Chapter 2 Time series (1).ppt
Applied Statistics Chapter 2 Time series (1).pptApplied Statistics Chapter 2 Time series (1).ppt
Applied Statistics Chapter 2 Time series (1).ppt
 
Time Series Analysis and Forecasting.ppt
Time Series Analysis and Forecasting.pptTime Series Analysis and Forecasting.ppt
Time Series Analysis and Forecasting.ppt
 
Basic Concepts of Statistics - Lecture Notes
Basic Concepts of Statistics - Lecture NotesBasic Concepts of Statistics - Lecture Notes
Basic Concepts of Statistics - Lecture Notes
 
forecast.ppt
forecast.pptforecast.ppt
forecast.ppt
 
Demand forecasting
Demand forecastingDemand forecasting
Demand forecasting
 
Enterprise_Planning_TimeSeries_And_Components
Enterprise_Planning_TimeSeries_And_ComponentsEnterprise_Planning_TimeSeries_And_Components
Enterprise_Planning_TimeSeries_And_Components
 
Demand forecasting by time series analysis
Demand forecasting by time series analysisDemand forecasting by time series analysis
Demand forecasting by time series analysis
 
Case study of s&p 500
Case study of s&p 500Case study of s&p 500
Case study of s&p 500
 
Weather forecasting model.pptx
Weather forecasting model.pptxWeather forecasting model.pptx
Weather forecasting model.pptx
 
Holtwinters terakhir lengkap
Holtwinters terakhir lengkapHoltwinters terakhir lengkap
Holtwinters terakhir lengkap
 
Storm Prediction data analysis using R/SAS
Storm Prediction data analysis using R/SASStorm Prediction data analysis using R/SAS
Storm Prediction data analysis using R/SAS
 
Cmt learning objective 36 case study of s&p 500
Cmt learning objective 36   case study of s&p 500Cmt learning objective 36   case study of s&p 500
Cmt learning objective 36 case study of s&p 500
 

Recently uploaded

PKS-TGC-1084-630 - Stage 1 Proposal.pptx
PKS-TGC-1084-630 - Stage 1 Proposal.pptxPKS-TGC-1084-630 - Stage 1 Proposal.pptx
PKS-TGC-1084-630 - Stage 1 Proposal.pptxPramod Kumar Srivastava
 
Customer Service Analytics - Make Sense of All Your Data.pptx
Customer Service Analytics - Make Sense of All Your Data.pptxCustomer Service Analytics - Make Sense of All Your Data.pptx
Customer Service Analytics - Make Sense of All Your Data.pptxEmmanuel Dauda
 
Dubai Call Girls Wifey O52&786472 Call Girls Dubai
Dubai Call Girls Wifey O52&786472 Call Girls DubaiDubai Call Girls Wifey O52&786472 Call Girls Dubai
Dubai Call Girls Wifey O52&786472 Call Girls Dubaihf8803863
 
9654467111 Call Girls In Munirka Hotel And Home Service
9654467111 Call Girls In Munirka Hotel And Home Service9654467111 Call Girls In Munirka Hotel And Home Service
9654467111 Call Girls In Munirka Hotel And Home ServiceSapana Sha
 
9711147426✨Call In girls Gurgaon Sector 31. SCO 25 escort service
9711147426✨Call In girls Gurgaon Sector 31. SCO 25 escort service9711147426✨Call In girls Gurgaon Sector 31. SCO 25 escort service
9711147426✨Call In girls Gurgaon Sector 31. SCO 25 escort servicejennyeacort
 
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.pptdokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.pptSonatrach
 
INTERNSHIP ON PURBASHA COMPOSITE TEX LTD
INTERNSHIP ON PURBASHA COMPOSITE TEX LTDINTERNSHIP ON PURBASHA COMPOSITE TEX LTD
INTERNSHIP ON PURBASHA COMPOSITE TEX LTDRafezzaman
 
办理学位证纽约大学毕业证(NYU毕业证书)原版一比一
办理学位证纽约大学毕业证(NYU毕业证书)原版一比一办理学位证纽约大学毕业证(NYU毕业证书)原版一比一
办理学位证纽约大学毕业证(NYU毕业证书)原版一比一fhwihughh
 
Data Science Jobs and Salaries Analysis.pptx
Data Science Jobs and Salaries Analysis.pptxData Science Jobs and Salaries Analysis.pptx
Data Science Jobs and Salaries Analysis.pptxFurkanTasci3
 
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改atducpo
 
How we prevented account sharing with MFA
How we prevented account sharing with MFAHow we prevented account sharing with MFA
How we prevented account sharing with MFAAndrei Kaleshka
 
From idea to production in a day – Leveraging Azure ML and Streamlit to build...
From idea to production in a day – Leveraging Azure ML and Streamlit to build...From idea to production in a day – Leveraging Azure ML and Streamlit to build...
From idea to production in a day – Leveraging Azure ML and Streamlit to build...Florian Roscheck
 
High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...
High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...
High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...soniya singh
 
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一F sss
 
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130Suhani Kapoor
 
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM TRACKING WITH GOOGLE ANALYTICS.pptx
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM  TRACKING WITH GOOGLE ANALYTICS.pptxEMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM  TRACKING WITH GOOGLE ANALYTICS.pptx
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM TRACKING WITH GOOGLE ANALYTICS.pptxthyngster
 

Recently uploaded (20)

Decoding Loan Approval: Predictive Modeling in Action
Decoding Loan Approval: Predictive Modeling in ActionDecoding Loan Approval: Predictive Modeling in Action
Decoding Loan Approval: Predictive Modeling in Action
 
PKS-TGC-1084-630 - Stage 1 Proposal.pptx
PKS-TGC-1084-630 - Stage 1 Proposal.pptxPKS-TGC-1084-630 - Stage 1 Proposal.pptx
PKS-TGC-1084-630 - Stage 1 Proposal.pptx
 
Call Girls in Saket 99530🔝 56974 Escort Service
Call Girls in Saket 99530🔝 56974 Escort ServiceCall Girls in Saket 99530🔝 56974 Escort Service
Call Girls in Saket 99530🔝 56974 Escort Service
 
Customer Service Analytics - Make Sense of All Your Data.pptx
Customer Service Analytics - Make Sense of All Your Data.pptxCustomer Service Analytics - Make Sense of All Your Data.pptx
Customer Service Analytics - Make Sense of All Your Data.pptx
 
꧁❤ Aerocity Call Girls Service Aerocity Delhi ❤꧂ 9999965857 ☎️ Hard And Sexy ...
꧁❤ Aerocity Call Girls Service Aerocity Delhi ❤꧂ 9999965857 ☎️ Hard And Sexy ...꧁❤ Aerocity Call Girls Service Aerocity Delhi ❤꧂ 9999965857 ☎️ Hard And Sexy ...
꧁❤ Aerocity Call Girls Service Aerocity Delhi ❤꧂ 9999965857 ☎️ Hard And Sexy ...
 
Dubai Call Girls Wifey O52&786472 Call Girls Dubai
Dubai Call Girls Wifey O52&786472 Call Girls DubaiDubai Call Girls Wifey O52&786472 Call Girls Dubai
Dubai Call Girls Wifey O52&786472 Call Girls Dubai
 
9654467111 Call Girls In Munirka Hotel And Home Service
9654467111 Call Girls In Munirka Hotel And Home Service9654467111 Call Girls In Munirka Hotel And Home Service
9654467111 Call Girls In Munirka Hotel And Home Service
 
9711147426✨Call In girls Gurgaon Sector 31. SCO 25 escort service
9711147426✨Call In girls Gurgaon Sector 31. SCO 25 escort service9711147426✨Call In girls Gurgaon Sector 31. SCO 25 escort service
9711147426✨Call In girls Gurgaon Sector 31. SCO 25 escort service
 
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.pptdokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
 
INTERNSHIP ON PURBASHA COMPOSITE TEX LTD
INTERNSHIP ON PURBASHA COMPOSITE TEX LTDINTERNSHIP ON PURBASHA COMPOSITE TEX LTD
INTERNSHIP ON PURBASHA COMPOSITE TEX LTD
 
办理学位证纽约大学毕业证(NYU毕业证书)原版一比一
办理学位证纽约大学毕业证(NYU毕业证书)原版一比一办理学位证纽约大学毕业证(NYU毕业证书)原版一比一
办理学位证纽约大学毕业证(NYU毕业证书)原版一比一
 
Data Science Jobs and Salaries Analysis.pptx
Data Science Jobs and Salaries Analysis.pptxData Science Jobs and Salaries Analysis.pptx
Data Science Jobs and Salaries Analysis.pptx
 
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
 
How we prevented account sharing with MFA
How we prevented account sharing with MFAHow we prevented account sharing with MFA
How we prevented account sharing with MFA
 
From idea to production in a day – Leveraging Azure ML and Streamlit to build...
From idea to production in a day – Leveraging Azure ML and Streamlit to build...From idea to production in a day – Leveraging Azure ML and Streamlit to build...
From idea to production in a day – Leveraging Azure ML and Streamlit to build...
 
High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...
High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...
High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...
 
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一
办理学位证中佛罗里达大学毕业证,UCF成绩单原版一比一
 
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130
 
VIP Call Girls Service Charbagh { Lucknow Call Girls Service 9548273370 } Boo...
VIP Call Girls Service Charbagh { Lucknow Call Girls Service 9548273370 } Boo...VIP Call Girls Service Charbagh { Lucknow Call Girls Service 9548273370 } Boo...
VIP Call Girls Service Charbagh { Lucknow Call Girls Service 9548273370 } Boo...
 
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM TRACKING WITH GOOGLE ANALYTICS.pptx
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM  TRACKING WITH GOOGLE ANALYTICS.pptxEMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM  TRACKING WITH GOOGLE ANALYTICS.pptx
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM TRACKING WITH GOOGLE ANALYTICS.pptx
 

TIME SERIES ECONOMETRICS ARIMA MODEL

  • 2. Types of Data • Time Series Data, • Cross-Sectional Data and • Panel Data.
  • 3. Time Series Data • Time series data, as the name suggests, are data that have been collected over a period of time on one or more variables. • Time series data have associated with them a particular frequency of observation or frequency of collection of data points. • The frequency is simply a measure of the interval over, or the regularity with which, the data are collected or recorded.
  • 4. Problems that could be tackled using time series data: • How the value of a country’s stock index has varied with that country’s macroeconomic fundamentals. • How the value of a company’s stock price has varied when it announced the value of its dividend payment. • The effect on a country’s exchange rate of an increase in its trade deficit. In all of the above cases, it is clearly the time dimension which is the most important, and the analysis will be conducted using the values of the variables over time.
  • 5. Cross-Sectional Data Cross-sectional data are data on one or more variables collected at a single point in time. For example, the data might be on: • A poll of usage of internet stockbroking services. • A cross-section of stock returns on the New York Stock Exchange (NYSE) • A sample of bond credit ratings for UK banks. • Data on sales volume, sales revenue, number of customers and expenses for the past month at each Starbucks location.
  • 6. Problems that could be tackled using cross-sectional data: • The relationship between company size and the return to investing in its shares. • The relationship between a country’s GDP level and the probability that the government will default on its sovereign debt.
  • 7. Panel, Longitudinal, or Micropanel Data Panel data have the dimensions of both time series and cross- sections, e.g. the daily prices of a number of blue chip stocks over two years. Cross sectional data is collected at a particular point of time...For instance, say your are studying the GDP of 3 developing countries in year 1999 only .... So u have data like this :- This is your cross sectional data since you are studying the entities (Countries here) at some point of time i.e. (year 1999 here) . Country Time GDP India 1999 ---- China 1999 ---- Brazil 1999 ---- Panel data is when you study again say GDP of 3 developing countries over a period of time (say 3 yrs. from 1999 to 2001)….So u have data like this: Here you are studying the same entities i.e. (China, India and Brazil) over a period of time i.e. 3 yrs. from 1999 to 2001. This is called panel data. Country Time GDP India 1999 ---- India 2000 ---- India 2001 ---- China 1999 ---- China 2000 ---- China 2001 ---- Brazil 1999 ---- Brazil 2000 ---- Brazil 2001 ----
  • 8. Balanced and Unbalanced Panel Data • If all the companies/ person/ entities have the same number of observations, we have what is called a balanced panel. • Balanced panel, as each person is observed every year • If the number of observations is not the same for each company, companies/ person/ entity it is called an unbalanced panel. • Unbalanced panel, since person 1 is not observed in year 2003 and person 3 is not observed in 2003 or 2001.
  • 9. Repeated Cross-Sections or Pooled Cross-Sections. • There is also a type of data that's in between cross-sectional data and panel data. It is typically called repeated cross- sections or pooled cross-sections. • For example, annual labour force surveys are repeated cross- sections, because every year, a new random sample is taken from the population. In this case, there is a time component, so these are not cross-sectional data, but every year, new individuals are surveyed, so these are also not panel data. That's why these are called repeated cross-sections.
  • 10. What is a time series? A time series is any series of data that varies over time. For example • Monthly Tourist Arrivals from Other countries. • Quarterly GDP of USA. • Hourly price of stocks and shares. • Weekly quantity of beer sold in a pub Because of widespread availability of time series databases most empirical studies use time series data. Definition of Time Series: An ordered sequence of values of a variable at equally spaced time intervals.
  • 11. Applications Of Time Series Analysis Time Series Analysis is used for many applications such as: • Economic Forecasting • Sales Forecasting • Budgetary Analysis • Stock Market Analysis • Yield Projections • Process and Quality Control • Inventory Studies • Workload Projections • Utility Studies • Census Analysis.
  • 12. Time Series Data • One of the important and frequent types of data used in empirical analysis. • But it poses several challenges to econometricians/practitioners. E.g. 1. Empirical work based on time series data assumes that the underlying time series is stationary. 2. Autocorrelation: because the underlying time series data is non-stationary. 3. Spurious/nonsense regression: a very high R2 and significant regression coefficients (though there is no meaningful relationship between the two variables)
  • 13. Caveats in Using Time Series Data in Applied Econometric Modeling • Data Should be Stationary • Presence of Autocorrelation • Guard Against Spurious Regressions • Establish Cointegration • Reconcile SR with LR Behavior via ECM • Implications to Forecasting • Possibility of Volatility Clustering
  • 14. Examples of Stationary Time Series -6000 -4000 -2000 0 2000 4000 6000 8000 92 94 96 98 00 02 04 AUST -6000 -4000 -2000 0 2000 4000 6000 92 94 96 98 00 02 04 CAN -2000 -1000 0 1000 2000 3000 4000 92 94 96 98 00 02 04 CHI -6000 -4000 -2000 0 2000 4000 6000 92 94 96 98 00 02 04 GERM -12000 -8000 -4000 0 4000 8000 12000 92 94 96 98 00 02 04 HONG -12000 -8000 -4000 0 4000 8000 12000 92 94 96 98 00 02 04 JAP -12000 -8000 -4000 0 4000 8000 12000 92 94 96 98 00 02 04 KOR -3000 -2000 -1000 0 1000 2000 92 94 96 98 00 02 04 MAL -3000 -2000 -1000 0 1000 2000 3000 92 94 96 98 00 02 04 SING -8000 -4000 0 4000 8000 12000 92 94 96 98 00 02 04 TWN -4000 -3000 -2000 -1000 0 1000 2000 3000 4000 5000 92 94 96 98 00 02 04 UKK -20000 -10000 0 10000 20000 30000 92 94 96 98 00 02 04 US
  • 15. 2000 4000 6000 8000 10000 12000 14000 16000 92 94 96 98 00 02 04 AUSTRALIA 0 2000 4000 6000 8000 10000 12000 92 94 96 98 00 02 04 CANADA 0 2000 4000 6000 8000 10000 92 94 96 98 00 02 04 CHINA 1000 2000 3000 4000 5000 6000 7000 8000 9000 92 94 96 98 00 02 04 GERMANY 0 4000 8000 12000 16000 20000 24000 92 94 96 98 00 02 04 HONGKONG 10000 15000 20000 25000 30000 35000 40000 45000 92 94 96 98 00 02 04 JAPAN 0 10000 20000 30000 40000 50000 92 94 96 98 00 02 04 KOREA 0 1000 2000 3000 4000 5000 6000 7000 92 94 96 98 00 02 04 MALAYSIA 0 1000 2000 3000 4000 5000 6000 7000 92 94 96 98 00 02 04 SINGAPORE 0 5000 10000 15000 20000 25000 30000 92 94 96 98 00 02 04 TAIWAN 0 2000 4000 6000 8000 10000 12000 92 94 96 98 00 02 04 UK 10000 20000 30000 40000 50000 60000 92 94 96 98 00 02 04 USA Examples of Non-Stationary Time Series
  • 16. Stationary Processes • Stationary Processes: A stochastic process is said to be stationary/ weakly /covariance/2nd-order stationary if: o Its mean and variance are constant over time, and o The value of the covariance between the two time periods depends only on the distance/lag between the two time periods and not the actual time at which the covariance is computed. o E.g. let’s Yt be a stochastic process, then; • Mean: E(Yt ) = µ ………………………………………….. (1) • Variance: var (Yt ) = E(Yt − µ)2 = σ2 ……………………………….. (2) • Covariance: γk = E[(Yt − µ)(Yt +k − µ)] ……………..………… (3) • Where γk, the covariance (or auto-covariance) at lag k, • If k = 0, we obtain γ0, which is simply the variance of Y (= σ2); if k = 1, γ1 is the covariance between two adjacent values of Y
  • 17. Stationarity A time series has stationarity if a shift in time doesn’t cause a change in the shape of the distribution. Basic properties of the distribution like the mean , variance and covariance are constant over time.
  • 18. Why are Stationary Time Series so Important? • Because if a time series is non-stationary, we can study its behavior only for the time period under consideration, and as a consequence, it is not possible to generalize it to other time periods. • Therefore, for the purpose of forecasting, such (non-stationary) time series may be of little practical value. • Non-stationary Stochastic Processes: Although our interest is in stationary time series, one often encounters non- stationary time series • A non-stationary time series will have a time-varying mean or a time-varying variance or both.
  • 19. Transformations to Achieve Stationarity If the time series is not stationary, we can often transform it to stationarity with one of the following techniques. 1. We can difference the data. That is, given the series Yt, we create the new series Yi = Yi − Yi−1. The differenced data will contain one less point than the original data. Although you can difference the data more than once, one difference is usually sufficient. 2. If the data contain a trend, we can fit some type of curve to the data and then model the residuals from that fit. Since the purpose of the fit is to simply remove long term trend, a simple fit, such as a straight line, is typically used. 3. For non-constant variance, taking the logarithm or square root of the series may stabilize the variance. For negative data, you can add a suitable constant to make all the data positive before applying the transformation. This constant can then be subtracted from the model to obtain predicted (i.e., the fitted) values and forecasts for future points. The above techniques are intended to generate series with constant location and scale. Although seasonality also violates stationarity, this is usually explicitly incorporated into the time series model
  • 20. • Seasonality rules out series (d), (h) and (i). • Trend rules out series (a), (c), (e), (f) and (i). • Stationary series (b) and (g).
  • 21. Differencing In Figure below, the Google stock price was non-stationary in panel (a), but the daily changes were stationary in panel (b). This shows one way to make a non- stationary time series stationary — compute the differences between consecutive observations. This is known as differencing.
  • 22. 1st Diff. 2nd Diff. Year Quarter gdp loggdp d1loggdp d2loggdp 1947 1 1570.519 7.359161 1947 2 1568.653 7.357973 -0.001189 1947 3 1567.966 7.357535 -0.000438 0.000751 1947 4 1590.938 7.372079 0.014545 0.014983 1948 1 1616.069 7.387752 0.015673 0.001128 1948 2 1644.637 7.405275 0.017523 0.001850 1948 3 1654.061 7.410989 0.005714 -0.011809 1948 4 1657.988 7.413360 0.002371 -0.003342 1949 1 1633.249 7.398327 -0.015034 -0.017405 1949 2 1628.439 7.395377 -0.002949 0.012084 1949 3 1646.698 7.406527 0.011150 0.014100 1949 4 1629.911 7.396281 -0.010247 -0.021397 1950 1 1696.765 7.436479 0.040198 0.050445 1950 2 1747.322 7.465840 0.029361 -0.010837 1950 3 1815.845 7.504306 0.038467 0.009106 1950 4 1848.928 7.522361 0.018055 -0.020412 1951 1 1871.311 7.534395 0.012033 -0.006022 1951 2 1903.118 7.551249 0.016854 0.004821 1951 3 1941.109 7.571015 0.019766 0.002912 1951 4 1944.447 7.572733 0.001718 -0.018048 Differencing the data
  • 23. EViews Commands Log Loggdp = log(gdp) 1st Difference d1loggdp=d(loggdp) 1st Difference + Log d1loggdp=dlog(gdp) 2nd Difference d2loggdp=d(d1loggdp) or d2loggdp=d(loggdp,2) or d2loggdp=dlog(gdp,2) • First take log of series and check stationarity because standard unit root tests assumes a linear structure. • If log transformed series is not stationarity then do differencing of log transformed series.
  • 24. GDP Lag1 Lag2 1570.519 1568.653 1570.519 1567.966 1568.653 1570.519 1590.938 1567.966 1568.653 1616.069 1590.938 1567.966 1644.637 1616.069 1590.938 1654.061 1644.637 1616.069 1657.988 1654.061 1644.637 1633.249 1657.988 1654.061 1628.439 1633.249 1657.988 1646.698 1628.439 1633.249 1629.911 1646.698 1628.439 1696.765 1629.911 1646.698 1747.322 1696.765 1629.911
  • 25. 7.2 7.6 8.0 8.4 8.8 9.2 9.6 50 55 60 65 70 75 80 85 90 95 00 05 LOGGDP 0 2,000 4,000 6,000 8,000 10,000 12,000 50 55 60 65 70 75 80 85 90 95 00 05 GDP
  • 26. Transformations to Achieve Stationarity A seasonal difference is the difference between an observation and the corresponding observation from the previous year. So, Y′t = Yt−Yt−m where m = number of seasons. These are also called “lag-m differences” as we subtract the observation after a lag of m periods.
  • 27. • Sometimes it is necessary to do both a seasonal difference and a first difference to obtain stationary data, as shown in Figure. • Here, the data are first transformed using logarithms (second panel). • Then seasonal differenced are calculated (third panel). • The data still seem a little non-stationary, and so a further lot of first differences are computed (bottom panel). US net electricity generation (billion kWh).
  • 28. Types of Stationary Models can show different types of stationarity: • Strict stationarity means that the joint distribution of any moments of any degree (e.g. expected values, variances, third order and higher moments) within the process is never dependent on time. This definition is in practice too strict to be used for any real-life model. • First-order stationarity series have means that never changes with time. Any other statistics (like variance) can change. • Second-order stationarity (also called weak stationarity) time series have a constant mean, variance and an autocovariance that doesn’t change with time. Other statistics in the system are free to change over time. This constrained version of strict stationarity is very common. • Trend-stationary models fluctuate around a deterministic trend (the series mean). These deterministic trends can be linear or quadratic, but the amplitude (height of one oscillation) of the fluctuations neither increases nor decreases across the series. • Difference-stationary models are models that need one or more differencing's to become stationary.
  • 29. • We call a stochastic process(time series) purely random/white noise process if it has zero mean, constant variance σ2, and is serially uncorrelated i.e. [ut ∼ IIDN(0, σ2)]. • Note: Here onward, in all equations the assumption of “white noise” will be applicable on ut . White Noise Processes
  • 31. Unit Root Testing: Formal Tests to Establish Stationarity of Time Series • Dickey-Fuller (DF) Test • Augmented Dickey-Fuller (ADF) Test • Phillips-Perron (PP) Unit Root Test • Dickey-Pantula Unit Root Test • GLS Transformed Dickey-Fuller Test • ERS Point Optimal Test • KPSS Test (run as a complement to the unit root tests) • Ng and Perron Test
  • 32. Some Useful Models for Time Series 1. A purely random process, 2. A random walk, 3. A moving average (MA) process, 4. An autoregressive (AR) process, 5. An autoregressive moving average (ARMA) process, and 6. An autoregressive integrated moving average (ARIMA)process.
  • 33. Estimation of AR, MA, and ARMA Models  Testing Goodness of Fit • When an AR, MA, or ARMA model has been fitted to a given time series, it is advisable to check that the model does really give an adequate description of the data • There are two criteria often used that reflect the closeness of fit and the number of parameters estimated. • One is the Akaike Information Criterion (AIC), and the other is the Schwartz Bayesian Criterion (SBC)/ Bayesian information criterion (BIC).
  • 34. The Box-Jenkins Approach • The Box-Jenkins approach is one of the most widely used methodologies for the analysis of time-series data • It is popular because of its generality; it can handle any series, stationary or not, with or without seasonal elements, and it has well-documented computer programs. • Although Box and Jenkins have been neither the originators nor the most important contributors in the field of ARMA models • They have popularized these models and made them readily accessible to everyone, so much that ARMA models are sometimes referred to as Box-Jenkins models.
  • 35. The Box-Jenkins Approach  The basic steps in the Box-Jenkins methodology are 1. Differencing the series so as to achieve Stationarity, 2. Identification of a tentative model, 3. Estimation of the model, 4. Diagnostic checking (if the model is found inadequate, we go back to step 2), and 5. Using the model for forecasting and control.
  • 36. The Box-Jenkins Approach 1. Differencing to achieve Stationarity: How do we conclude whether a time series is stationary or not? • We can do this by studying the graph of the correlogram of the series. • The correlogram of a stationary series drops off as k, the number of lags, becomes large, but this is not usually the case for a non- stationary series. • Thus, the common procedure is to plot the correlogram of the given series Yt and successive differences ΔY, ΔY, and so on, and look at the correlograms at each stage. • We keep differencing until the correlogram dampens.
  • 37. The Box-Jenkins Approach 2. Once we have used the differencing procedure to get a stationary time series, we examine the correlogram to decide on the appropriate orders of the AR and MA components. • The correlogram of a MA process is zero after a point. • That of an AR process declines geometrically. The correlograms of ARMA processes show different patterns (but all dampen after a while). • Based on these, one arrives at a tentative ARMA model. • This step involves more of a judgmental procedure than the use of any clear-cut rules.
  • 39. The Box-Jenkins Approach 3. The next step is the estimation of the tentative ARMA model identified in step 2. We have discussed in the preceding section the estimation of ARMA models. 4. The next step is diagnostic checking to check the adequacy of the tentative model. We discussed in the preceding section the Q and Q* statistics commonly used in diagnostic checking. As argued there, the (^-statistic is inappropriate in autoregressive models and thus we need to replace it with some LM test statistic. 5. The final step is forecasting.
  • 40. Differencing the series to achieve Stationarity Identify model to be tentatively entertained Estimate the parameters of the tentative model Diagnostic checking. Is the model adequate? No Yes Use the model for forecasting and control Approaches to Economic Forecasting The Box-Jenkins Approach
  • 41. The defining characteristics of AR, MA and ARMA processes: An autoregressive process has: • a geometrically decaying acf • a number of non-zero points of pacf = AR order. A moving average process has: • a geometrically decaying pacf. • number of non-zero points of acf = MA order A combination autoregressive moving average process has: • a geometrically decaying acf • a geometrically decaying pacf.
  • 42. • Diagnostic Checking: Diagnostic checking consists of evaluating the adequacy of the estimated model. Considerable skill is required to choose the actual ARIMA (p,d,q) model so that the residuals estimated from this model are white noise. So the autocorrelations of the residuals are to be estimated for the diagnostic checking of the model. These are also judged by Ljung-Box statistic under null hypothesis that autocorrelation co-efficient is equal to zero.
  • 43. MA(1) model Sample autocorrelation and partial autocorrelation functions for an MA(1) model: • The MA(1) has an acf that is significant for only lag 1, while the pacf declines geometrically, and is significant until lag 7. • The acf at lag 1 and all of the pacfs are negative as a result of the negative coefficient in the MA generating process.
  • 44. MA(2) model Sample autocorrelation and partial autocorrelation functions for an MA(2) model: • The first two autocorrelation coefficients only are significant, while the partial autocorrelation coefficients are geometrically declining. • Since, the second coefficient on the lagged error term in the MA is negative, the acf and pacf alternate between positive and negative.
  • 45. AR(1) model Sample autocorrelation and partial autocorrelation functions for an AR(1) model: • The AR(1) has an pacf that is significant for only lag 1, while the acf declines geometrically. • Only the first pacf coefficient is significant, while all others are virtually zero and are not significant.
  • 46. AR(1) model Sample autocorrelation and partial autocorrelation functions for an AR(1) model: • AR(1), which was generated using identical error terms, but a much smaller autoregressive coefficient. In this case, the autocorrelation function dies away much more quickly than in the previous example, and in fact becomes insignificant after around five lags.
  • 47. AR(1) model • Sample autocorrelation and partial autocorrelation functions for a non-stationary model (i.e. a unit coefficient): • On some occasions, the acf does die away for a non-stationary process. • The pacf, however, is significant only for lag 1, correctly suggesting that an autoregressive model with no moving average term is most appropriate.
  • 48. ARMA(1, 1) model • Sample autocorrelation and partial autocorrelation functions for an ARMA(1, 1) model: • In such a process, both the acf and the pacf decline geometrically – the acf as a result of the AR part and the pacf as a result of the MA part.
  • 49. GDP • Check through AIC and BIC criteria. • Imp note: ARIMA model output differs in EViews 8 and later versions. EViews 8 and former versions estimate ARIMA model on the basis of conditional least squares method. Whereas version 9 and newer are based on ARIMA forecasting through maximum likelihood method. • Therefore AIC & BIC values will also differ in EViews 8 and later versions.
  • 50. SBIC & AIC code results with EViews 8 SBC MA(0) MA(1) MA(2) MA(3) MA(4) MA(5) AR(0) --- -6.464488 -6.482979 -6.465608 -6.443753 -6.439384 AR(1) -6.490116 -6.471629 -6.462397 -6.440327 -6.448238 -6.425663 AR(2) -6.471464 -6.456046 -6.497155 -6.437803 -6.433839 -6.405617 AR(3) -6.462250 -6.453936 -6.520738 -6.505968 -6.488171 -6.446302 AR(4) -6.448194 -6.429773 -6.509408 -6.487179 -6.503980 -6.493222 AR(5) -6.428779 -6.407640 -6.384756 -6.389125 -6.518437 -6.500290 Best Model Is ARMA(3,2) AIC MA(0) MA(1) MA(2) MA(3) MA(4) MA(5) AR(0) --- -6.493237 -6.526104 -6.523107 -6.515626 -6.525633 AR(1) -6.518950 -6.514880 -6.520065 -6.512412 -6.534741 -6.526582 AR(2) -6.514843 -6.513885 -6.569454 -6.524561 -6.535057 -6.521295 AR(3) -6.520261 -6.526449 -6.607754 -6.607487 -6.604193 -6.576826 AR(4) -6.520924 -6.517048 -6.611230 -6.603546 -6.634893 -6.638681 AR(5) -6.516316 -6.509766 -6.501471 -6.520429 -6.664330 -6.660773 Best Model Is ARMA(5,4)
  • 51. AIC Vs SBIC • When large observations (data points) are available select model as per AIC. This is because AIC will always select higher model than SBIC. Higher model (eg. ARMA 5,5) means loss of 5 data points, therefore more data should be available which can compensate for loss of lags.
  • 52. Computing Summary Statistics • UKHP.xls • Import into Eviews. • Calculate simple percentage changes in the series: dhp = 100*(hp-hp(-1))/hp(-1) • To obtain descriptive summary statistics of a series select Quick/Series Statistics/Histogram and Stats and type the name of the variable (DHP)
  • 54. OBS DHP 1991M01 1991M02 0.83895 1991M03 -1.12892 0.83895 1991M04 1.483326 -1.12892 1991M05 1.319533 1.483326 1991M06 1.326908 1.319533 1991M07 -1.02755 1.326908 1991M08 -0.91749 -1.02755 1991M09 -1.44568 -0.91749 1991M10 0.388974 -1.44568 1991M11 -0.83902 0.388974 1991M12 0.162323 -0.83902 1992M01 -2.17099 0.162323 1992M02 -0.19444 -2.17099 1992M03 0.245605 -0.19444 1992M04 0.00058 0.245605 1992M05 1.413615 0.00058 1992M06 0.366004 1.413615 1992M07 -0.33771 0.366004 1992M08 -0.78737 -0.33771 1992M09 -2.32213 -0.78737 1992M10 -0.77807 -2.32213 1992M11 -2.16517 -0.77807 1992M12 0.288302 -2.16517 1993M01 0.665036 0.288302 1993M02 0.597487 0.665036 1993M03 -0.37186 0.597487 1993M04 2.968641 -0.37186 1993M05 -1.50191 2.968641
  • 55. ARMA(p,q) ARMA(5,5) dhp c ar(1) ar(2) ar(3) ar(4) ar(5) ma(1) ma(2) ma(3) ma(4) ma(5)
  • 56. AIC selects an ARMA(4,5), while SBIC selects the smaller ARMA(2,0) model i.e. an AR(2). The values of all of the Akaike and Schwarz information criteria calculated using EViews are as follows. ARMA(5,5) Model
  • 57. Why forecast? Some examples in finance of where forecasts from econometric models might be useful include: • Forecasting tomorrow’s return on a particular share. • Forecasting the price of a house given its characteristics. • Forecasting the riskiness of a portfolio over the next year. • Forecasting the volatility of bond returns • Forecasting the correlation between US and UK stock market movements tomorrow • Forecasting the likely number of defaults on a portfolio of home loans.
  • 58. Checking Forecasting Accuracy • Root Mean Squared Error: the smaller the error, the better the forecasting ability of that model • Theil Inequality Coefficient: always lies between zero and one, where zero indicates a perfect fit. http://www.eviews.com/help/helpintro.html#page/content/Forecast-Forecast_Basics.html http://www.eviews.com/help/helpintro.html#page/content%2FForecast- An_Illustration.html%23
  • 59. Diagnostic Checking • How do we know that the model is a reasonable fit to the data? • One simple diagnostic is to obtain residuals from identified Equation and obtain the ACF and PACF of these residuals, say, up to lag 25 (or 1/3 or ¼ ) of total observations. • In estimated AC and PACF Figure, none of the autocorrelations and partial autocorrelations should be individually statistically significant. Nor is the sum of the 25 squared autocorrelations, as shown by the Box–Pierce Q and Ljung–Box (LB) statistics, statistically significant. • In other words, the correlograms of both autocorrelation and partial autocorrelation give the impression that the residuals estimated from Equation are purely random. Hence, there may not be any need to look for another ARIMA model
  • 60. Diagnostic Checking Serial Correlation (Auto-Correlation) • Normally serial correlation (auto-correlation) in regression model (Y= a + bX) is detected through Durbin Watson Statistics (DW). • However, when regression model is autoregressive (AR) in nature (Y = a + bYt-1), Durbin Watson Statistics (DW) will give invalid results. • Durbin Watson Statistics (DW) can be used for only for AR models up to one lag. • Testing serial correlation in Eviews: 1. Correlogram: a. ACF and PACF. b. the Ljung-Box (LB) Q-statistics and their p-values. 2. The Langrange Multiplier (LM test) (given by Breusch and Godfrey)
  • 61. • Model accuracy is generally assessed by using the root mean squared error criterion. • Calculate all the errors, square them, calculate the average, and then take the square root of that average. • The model with the lowest mean-squared error is judged the most accurate Diagnostic Checking Cont.……
  • 62. Autoregressive (AR) Forecasting Equation • To make forecasts j years into the future, using a third-order autoregressive model (AR3), you need only the most recent p = 3 values (Yn, Yn-1 and Yn-2 ) and the regression estimates a0, a1, a2, and a3. To forecast one year ahead, Equation becomes: To forecast two years ahead, Equation becomes: To forecast two years ahead, Equation becomes:
  • 63. Moving Average (MA) Forecasting Equation • An autoregression of the residual error time series is called a Moving Average (MA) model. This is confusing because it has nothing to do with the moving average smoothing process. Think of it as the sibling to the autoregressive (AR) process, except on lagged residual error rather than laged raw observations. MA order Regression Equation MA(1) Yt = c + Ut-1 MA(2) Yt = c + Ut-1 + Ut-2 U = error
  • 64. Seasonality Time series that show regular patterns of movement within a year across years. • Seasonal lags are most often included as a lagged value one year before the prior value. • We detect such patterns through the autocorrelations in the data. • For quarterly data, the fourth autocorrelation will not be statistically zero if there is quarterly seasonality.  For monthly, the 12th, and so on. • To correct for seasonality, we can include an additional lagged term to capture the seasonality.  For quarterly data, we would include a prior year quarterly seasonal lag as 𝑥𝑡 = 𝑏0 + 𝑏1𝑥𝑡−1 + 𝑏2𝑥𝑡−4 + ε𝑡