This document proposes a new 3-way ANOVA test to help determine whether to use a direct, indirect, or mixed approach for seasonally adjusting time series data. The test examines seasonal patterns across time frequencies, years, and individual time series to see if they have common regular seasonal patterns. If seasonal patterns differ, an indirect or mixed approach should be used instead of a direct approach. The document provides an example application to construction industry data from France, Belgium, and Luxembourg, finding common seasonal patterns between Belgium and Luxembourg to allow their mixed adjustment. Future work is proposed to further develop the theoretical underpinnings and test additional applications of the approach.
A Study on Performance Analysis of Different Prediction Techniques in Predict...IJRES Journal
Time series data is a series of statistical data that is related to a specific instant or a specific time period. Here, the measurements are recorded on a regular basis such as monthly, quarterly and yearly. Most of the researchers have used one of the prediction techniques in prediction of time series data. But, they have not tested all prediction techniques on same data set. They have not even compared the performance of different prediction techniques on the same data set. In this research work, some well known prediction techniques have been applied in the same time series data set. The average error and residual analysis have been done for each and every applied technique. One technique has been selected based on the minimum average error and residual analysis among the all applied techniques. The residual analysis comprises of absolute residual, maximum residual, median of absolute residual, mean of absolute residual and standard deviation. To finalize the algorithm, same procedure has been applied on different time series data sets. Finally, one technique has been selected which has been given minimum error and minimum value of residual analysis in most cases.
Siegel-Tukey test named after Sidney Siegel and John Tukey, is a non-parametric test which may be applied to the data measured at least on an ordinal scale. It tests for the differences in scale between two groups.
The test is used to determine if one of two groups of data tends to have more widely dispersed values than the other.
The test was published in 1980 by Sidney Siegel and John Wilder Tukey in the journal of the American Statistical Association in the article “A Non-parametric Sum Of Ranks Procedure For Relative Spread in Unpaired Samples “.
A Study on Performance Analysis of Different Prediction Techniques in Predict...IJRES Journal
Time series data is a series of statistical data that is related to a specific instant or a specific time period. Here, the measurements are recorded on a regular basis such as monthly, quarterly and yearly. Most of the researchers have used one of the prediction techniques in prediction of time series data. But, they have not tested all prediction techniques on same data set. They have not even compared the performance of different prediction techniques on the same data set. In this research work, some well known prediction techniques have been applied in the same time series data set. The average error and residual analysis have been done for each and every applied technique. One technique has been selected based on the minimum average error and residual analysis among the all applied techniques. The residual analysis comprises of absolute residual, maximum residual, median of absolute residual, mean of absolute residual and standard deviation. To finalize the algorithm, same procedure has been applied on different time series data sets. Finally, one technique has been selected which has been given minimum error and minimum value of residual analysis in most cases.
Siegel-Tukey test named after Sidney Siegel and John Tukey, is a non-parametric test which may be applied to the data measured at least on an ordinal scale. It tests for the differences in scale between two groups.
The test is used to determine if one of two groups of data tends to have more widely dispersed values than the other.
The test was published in 1980 by Sidney Siegel and John Wilder Tukey in the journal of the American Statistical Association in the article “A Non-parametric Sum Of Ranks Procedure For Relative Spread in Unpaired Samples “.
This presentation explains the procedure involved in two-way repeated measures ANOVA(within-within design). An illustration has been discussed by using the functionality of SPSS.
Learning material on Measurement of Seasonal variations prepared in accordance to VTU I Sem MBA syllabus for the subject Business Statistics & Analytics
Hi semua, terima kasih sudah berkunjung kesini 😆 Semua file yang diupload adalah materi perkuliahan. Nah... materi ini dari dosen yang dikhususkan untuk teman-teman kelas #manabeve 💚
Biar gampang diakses, yah masukin sini aja kan😆 Sekalian membantu kalian yang mungkin butuh beberapa konten dalam file-file ini.
Jangan lupa di like yah 💙 Kalau mau dishare atau didownload PLEASE MINTA IZIN dulu oke??
Biar ngga salah paham cuy😆
ASK FOR PERMISSION ▶ itsmeroses@mail.ru
Kalau kesulitan untuk mendownload FEEL FREE untuk email ke aku🔝🔝🔝🔝
[DISCLAIMER] Mohon banget kalau udah didownload. Kemuadian ingin dijadikan materi atau referensi. Jangan lupa cantumkan sumbernya. Terima kasih atas pengertiannya💖
------------------------------------------------------------
Materi details :
Coming soon ")
------------------------------------------------------------
MEET CLASS FELLAS💚
Instagram ▶ https://www.instagram.com/manabeve
Blog ▶ https://manabeve.blogspot.com
Email ▶ manabeve@gmail.com
------------------------------------------------------------
LET'S BECOME FRIENDS WITH ME💜
Instagram ▶ https://www.instagram.com/ameldiana3
Twitter ▶ https://www.twitter.com/amlediana3
This presentation explains the procedure involved in two-way repeated measures ANOVA(within-within design). An illustration has been discussed by using the functionality of SPSS.
Learning material on Measurement of Seasonal variations prepared in accordance to VTU I Sem MBA syllabus for the subject Business Statistics & Analytics
Hi semua, terima kasih sudah berkunjung kesini 😆 Semua file yang diupload adalah materi perkuliahan. Nah... materi ini dari dosen yang dikhususkan untuk teman-teman kelas #manabeve 💚
Biar gampang diakses, yah masukin sini aja kan😆 Sekalian membantu kalian yang mungkin butuh beberapa konten dalam file-file ini.
Jangan lupa di like yah 💙 Kalau mau dishare atau didownload PLEASE MINTA IZIN dulu oke??
Biar ngga salah paham cuy😆
ASK FOR PERMISSION ▶ itsmeroses@mail.ru
Kalau kesulitan untuk mendownload FEEL FREE untuk email ke aku🔝🔝🔝🔝
[DISCLAIMER] Mohon banget kalau udah didownload. Kemuadian ingin dijadikan materi atau referensi. Jangan lupa cantumkan sumbernya. Terima kasih atas pengertiannya💖
------------------------------------------------------------
Materi details :
Coming soon ")
------------------------------------------------------------
MEET CLASS FELLAS💚
Instagram ▶ https://www.instagram.com/manabeve
Blog ▶ https://manabeve.blogspot.com
Email ▶ manabeve@gmail.com
------------------------------------------------------------
LET'S BECOME FRIENDS WITH ME💜
Instagram ▶ https://www.instagram.com/ameldiana3
Twitter ▶ https://www.twitter.com/amlediana3
Metastatistical Extreme Value distributionsRiccardo Rigon
Marco Marani and coworkers rethink the estreme value concepts, observinfg that Pearson's distributions are obtained as a limit of an infinite number of events. He proposed intermediate distribution, when the number of observations is limited. He, they, called these distribution metastistical. This is, I think new insight in old stuff. Pretty much necessary though.
Goodness–of–fit tests for regression models: the functional data caseNeuroMat
In this talk the topic of the goodness–of–fit for regression models with functional covariates is considered. Although several papers have been published in the last two decades for the checking of regression models, the case where the covariates are functional is quite recent and has became of interest in the last years. We will review the very recent advances in this area and we will propose a new goodness–of–fit test for the null hypothesis of a functional linear model with scalar response. Our test is based on a generalization to the functional framework of a previous one, designed for the goodness–of–fit of regression models with multivariate covariates using random projections. The test statistic is easy to compute using geometrical and matrix arguments, and simple to calibrate in its distribution by a wild bootstrap on the residuals. Some theoretical aspects are derived and the finite sample properties of the test are illustrated by a simulation study. Finally, the test is applied to real data for checking the assumption of the functional linear model and a graphical tool is introduced. Lecturer: Wenceslao González-Manteiga, Univ. de Santiago de Compostela, Spain.
Reporting uncertainties - too much information?Dario Buono
Ocial statistics are published by government agencies and other international institutes to provide infor-
mation on the economy, living conditions, social development etc. These metrics are evaluated using dierent
sources, primarily surveys and censuses and, in addition, data obtained from government administrations or
private sector information.
Several qualitative criteria are considered the basis for trustworthy ocial statistics, such as impartiality,
transparency, relevance and independency. However, as the published metrics are derived from statistical
analysis of imperfect and potentially incomplete data, errors and uncertainties are inevitable and, in some
cases, require revisions or corrections that can lead to reduced condence in the overall process. It is also
important to recognize that the uncertainties originate both from statistical errors, such as the use of limited
raw data, and from bias induced by incomplete information or modeling assumptions.
Understanding how dierent sources of errors lead to bias and variance helps us improve the overall process
resulting in more accurate predictions. Quantitative measures of the variance errors are commonplace and
easy to convey; among those the standard-error-in the mean is perhaps the most popular and results in the
symbol. Measures of the spread in the actual data are also easy to estimate and disseminated using the
variance, or more frequently the standard deviation. More complete representation of the statistical spread
in the row data leads to percentiles and, eventually, to reporting the complete probability distributions.
Measures of bias, on the other hand, are not well developed because in many cases are not directly computable.
In the engineering community rather than presenting the variance and bias error, the focus is to identify
and rank the sources of uncertainties that explain the imprecision in the estimates. In this work we will
discuss applications of two global sensitivity metrics, the Sobol indices and the active subspace variables as
tools to describe the variance errors. Furthermore, we will discuss the distance metric as a strategy to assess
bias errors derived from classical measures of discrepancies between probability distribution functions.
Skills for the new generation of statisticians Dario Buono
This presentation analyses the competence profile of official statisticians with a particular focus on new data science competences. Modernization of official statistics will depend on the capability to incorporate new data sources and benefit from “disruptive technologies”. This will require new capabilities, skills and competences that may not be part of the traditional skill set of official statisticians. The document was presented to the Conference of European Statisticians organised at the United Nation in Geneva
JDemetra+ Java Tool for Seasonal AdjustmentDario Buono
JDemetra+ is a tool for seasonal adjustment (SA) developed by the National Bank of Belgium (NBB) in cooperation with the Deutsche Bundesbank and Eurostat in accordance with the Guidelines of the European Statistical System (ESS). User support, training and methodological development is provided by the devoted Centre of Excellence on Seasonal Adjustment coordinated by INSEE, the French National Statistical Office.
JDemetra+ has been officially recommended, since February 2015, to the members of the ESS and the European System of Central Banks as software for seasonal and calendar adjustment of official statistics.
Big data and macroeconomic nowcasting from data access to modellingDario Buono
Parallel advances in IT and in the social use of Internet-related applications, provide the general public with access to a vast amount of information. The associated Big Data are potentially very useful for a variety of applications, ranging from marketing to tapering fiscal evasion.
From the point of view of official statistics, the main question is whether and to what extent Big Data are a field worth investing to expand, check and improve the data production process and which types of partnerships will have to be formed for this purpose. Nowcasting of macroeconomic indicators represents a well-identified field where Big Data has the potential to play a decisive role in the future.
In this paper we present the results and main recommendations from the Eurostat-funded project “Big Data and macroeconomic nowcasting”, implemented by GOPA Consultants, which benefits from the cooperation and work of the Eurostat task force on Big Data and a few external academic experts.
Big Data Analysis: The curse of dimensionality in official statisticsDario Buono
Statistical authorities need to produce accurate data faster and in a cost effective way, to become more responsive to users´ demands, while at the same time continuing to provide high quality output. One way to fulfil this is to make use of all new accessible data sources, as for example administrative data and big data. As a result, statistical offices will have to deal more and more with a "huge" number" of time series, in particular for producing model based statistics.
Using high dimensional datasets will most likely urge statistical authorities to follow a different approach, in particular to be conscious that the measurement of socio-economic variables will follow more and more non-linear processes that could not be described by probability distributions that could be easily described by few parameters.
It will thus imply to adapt the way to observe the world through data taking into account at a greater extent uncertainty and complexity, which will in turn impact dissemination and communication activities of statistical authorities.
Reliability of estimates in socio-demographic groups with small samplesDario Buono
The aim of this work is twofold: to investigate the possibilities of model-based approach implementation in the official statistics so to ensure reliability of data for social conditions by different breakdowns; and to discuss advantages, disadvantages, and the potentiality of use of small area estimation techniques and tools in production of the official statistics.
In order to try to analyse fitting of the models to different type of data, various run were conducted using several small area estimation techniques (such as empirical Bayesian, hierarchical Bayes, etc.) already built-in within the R software (packages sae, hbsae, etc.) to obtain area and unit level based at-risk-of-poverty estimates and the mean squared errors of the estimates
Enhanced Enterprise Intelligence with your personal AI Data Copilot.pdfGetInData
Recently we have observed the rise of open-source Large Language Models (LLMs) that are community-driven or developed by the AI market leaders, such as Meta (Llama3), Databricks (DBRX) and Snowflake (Arctic). On the other hand, there is a growth in interest in specialized, carefully fine-tuned yet relatively small models that can efficiently assist programmers in day-to-day tasks. Finally, Retrieval-Augmented Generation (RAG) architectures have gained a lot of traction as the preferred approach for LLMs context and prompt augmentation for building conversational SQL data copilots, code copilots and chatbots.
In this presentation, we will show how we built upon these three concepts a robust Data Copilot that can help to democratize access to company data assets and boost performance of everyone working with data platforms.
Why do we need yet another (open-source ) Copilot?
How can we build one?
Architecture and evaluation
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Data and AI
Round table discussion of vector databases, unstructured data, ai, big data, real-time, robots and Milvus.
A lively discussion with NJ Gen AI Meetup Lead, Prasad and Procure.FYI's Co-Found
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Adjusting OpenMP PageRank : SHORT REPORT / NOTESSubhajit Sahu
For massive graphs that fit in RAM, but not in GPU memory, it is possible to take
advantage of a shared memory system with multiple CPUs, each with multiple cores, to
accelerate pagerank computation. If the NUMA architecture of the system is properly taken
into account with good vertex partitioning, the speedup can be significant. To take steps in
this direction, experiments are conducted to implement pagerank in OpenMP using two
different approaches, uniform and hybrid. The uniform approach runs all primitives required
for pagerank in OpenMP mode (with multiple threads). On the other hand, the hybrid
approach runs certain primitives in sequential mode (i.e., sumAt, multiply).
Learn SQL from basic queries to Advance queriesmanishkhaire30
Dive into the world of data analysis with our comprehensive guide on mastering SQL! This presentation offers a practical approach to learning SQL, focusing on real-world applications and hands-on practice. Whether you're a beginner or looking to sharpen your skills, this guide provides the tools you need to extract, analyze, and interpret data effectively.
Key Highlights:
Foundations of SQL: Understand the basics of SQL, including data retrieval, filtering, and aggregation.
Advanced Queries: Learn to craft complex queries to uncover deep insights from your data.
Data Trends and Patterns: Discover how to identify and interpret trends and patterns in your datasets.
Practical Examples: Follow step-by-step examples to apply SQL techniques in real-world scenarios.
Actionable Insights: Gain the skills to derive actionable insights that drive informed decision-making.
Join us on this journey to enhance your data analysis capabilities and unlock the full potential of SQL. Perfect for data enthusiasts, analysts, and anyone eager to harness the power of data!
#DataAnalysis #SQL #LearningSQL #DataInsights #DataScience #Analytics
New innovative 3 way anova a-priori test for direct vs. indirect approach in seasonal adjustment
1. Enrico Infante*
University of Naples Federico II
Dario Buono*
EUROSTAT, Unit B1: Quality, Research and Methodology
Euroindicators PEEI WG – Luxembourg, 11-12 June 2012
*The views and the opinions expressed in this paper are solely of the authors
and do not necessarily reflect those of the institutions for which they work
New innovative 3-way ANOVA a-priori test for
direct vs. indirect approach in Seasonal Adjustment
2. 2
A generic time series Yt can be the result of an aggregation of p series:
( )pthttt XXXfY ,,,,1 =
We focus on the case of the additive function:
∑=
=++++=
p
h
hthptphthtt XXXXY
1
11 ϖϖϖϖ
Introduction
Enrico Infante, Dario Buono
3. 3
To Seasonally Adjust the aggregate, different approaches can be applied
Direct Approach
Indirect Approach
The Seasonally Adjusted data are
computed directly by Seasonally
Adjusting the aggregate
( )
= ∑=
p
h
htht XSAYSA
1
ϖ
The Seasonally Adjusted data are
computed indirectly by Seasonally
Adjusting data per each series
( ) ( )∑=
=
p
h
htht XSAYSA
1
ϖ
Introduction
Enrico Infante, Dario Buono
4. 4
If it is possible to divide the series into groups, then it is possible to
compute the Seasonally Adjusted figures by summing the Seasonally
Adjusted data of these groups
Mixed Approach
Example (two groups):
∑∑ ==
+=
r
u
utu
q
l
ltlt XXY
11
ϖϖ
Group A Group B
prq =+
( )
+
= ∑∑ ==
r
u
utu
q
l
ltlt XSAXSAYSA
11
ϖϖ
Introduction
Enrico Infante, Dario Buono
5. 5
To use the Mixed Approach, sub-aggregates must be defined
We would like to find a criterion to divide the series into groups
The series of each group must have
common regular seasonal patterns
How is it possible to decide that two or more series have common
seasonal patterns?
NEW TEST!!!
The basic idea
Enrico Infante, Dario Buono
6. 6
Direct and indirect: there is no consensus on which is the best approach
Direct Indirect
+
-
• Transparency
• Accuracy
• Accounting
Consistency
• No accounting
consistency
• Cancel-out
effect
• Residual
Seasonality
• Calculations
burden
It could be interesting to identify
which series can be aggregated in
groups and decide at which level
the SA procedure should be run
This test gives information about
the approach to follow before
SA of the series
Why a new test?
Enrico Infante, Dario Buono
The presence of residual seasonality should
always be checked in all of the Indirect and
Mixed Seasonally Adjusted aggregates
7. 7
The variable tested is the final estimation of the unmodified Seasonal-
Irregular ratios (or differences) absolute value
ijkSI
1−ijkSI
Additive model
Multiplicative model
It is considered that the decomposition model is the same on all the
series. The series is then considered already Calendar Adjusted
The classic test for moving seasonality is based on a 2-way ANOVA test,
where the two factors are the time frequency (usually months or
quarters) and the years. This test is based on a 3-way ANOVA model,
where the three factors are the time frequency, the years and the series
The test
Enrico Infante, Dario Buono
8. 8
The model is:
ijkkjiijk ecbaSI +++=
Where:
• ai, i=1,…,M, represents the numerical contribution due to the effect of
the i-th time frequency (usually M=12 or M=4)
• bj, j=1,…,N, represents the numerical contribution due to the effect of
the j-th year
• ck, k=1,…,S, represents the numerical contribution due to the effect of
the k-th series of the aggregate
• The residual component term eijk (assumed to be normally distributed
with zero mean, constant variance and zero covariance) represents the
effect on the values of the SI of the whole set of factors not explicitly
taken into account in the model
The test
Enrico Infante, Dario Buono
9. 9
The test is based on the decomposition of the variance of the
observations:
22222
RSNM SSSSS +++=
Sk ,,1 =
Nj ,,1 =
Between time frequencies variance
Between years variance
Between series variance
Residual variance
The test
Enrico Infante, Dario Buono
Mi ,,1 =
10. 10
VAR Mean df
2
MS
2
NS
2
SS
2
RS
∑∑= =
•• =
N
j
S
k
ijki SI
NS
x
1 1
1
∑∑= =
•• =
M
i
S
k
ijkj SI
MS
x
1 1
1
( )∑∑∑= = =
•••••• +−−−
M
i
N
j
S
k
kjiijk xxxxSI
1 1 1
2
2
( )∑=
•• −
M
i
i xxNS
1
2
( )∑=
•• −
N
j
j xxMS
1
2
( )∑=
•• −
S
k
k xxMN
1
2
∑∑= =
•• =
M
i
N
j
ijkk SI
MN
x
1 1
1
1−M
1−N
1−S
( )( )( )111 −−− SNM
The table for the ANOVA test
Sum of Squares
The test
Enrico Infante, Dario Buono
11. 11
The null hypothesis is made taking into consideration that there is no
change in seasonality over the series
( ) ( )( )( )111;12
2
~ −−−−= SNMS
R
S
T F
S
S
F
The test statistic is the ratio of the between series variance and the
residual variance, and follows a Fisher-Snedecor distribution with (S-1)
and (M-1)(N-1)(S-1) degrees of freedom
ScccH === 210 :
Rejecting the null hypothesis is to say that the pure Direct Approach
should be avoided, and an Indirect or a Mixed one should be considered
The test
Enrico Infante, Dario Buono
12. 12
ttt XXY 21 +=
The most simple case: the aggregate is formed of two series, using the
same decomposition model
Do X1t and X2t have the same seasonal patterns?
TEST
Rejecting H0: the two series
have different seasonal patterns
Not rejecting H0: the two series
have common regular seasonal
patterns
Direct Approach
Indirect Approach
Showing the procedure - Example
Enrico Infante, Dario Buono
13. 13
Let’s consider the Construction Production Index of the three French-
speaking European countries: France, Belgium and Luxembourg (data are
available on the EUROSTAT database). The time span is from January
2001 to December 2010
To take an example, a very simple aggregate could be the following:
tttt LUBEFRY ++=
VAR Mean Square df
Months 1.5003 11
Years 0.0226 9
Series 0.1356 2
Residual 0.0117 198
8122.5
0117.0
1356.0
==− ratioF 0035.0=− valueP
There is no evidence of common
seasonal patterns between the
series at 5 per cent level
The Direct Approach
should be avoided
Numerical example
Enrico Infante, Dario Buono
14. 14
If two of them have the same seasonal pattern, a Mixed Approach could
be used. So the test is now used for each couple of series
VAR Mean Square df
Months 2.0403 11
Years 0.0140 9
Series 0.1199 1
Residual 0.0016 99
7591.75=F 0000.0=− valueP 8313.4=F 0303.0=− valueP
VAR Mean Square df
Months 1.0464 11
Years 0.0172 9
Series 0.0793 1
Residual 0.0164 99
LU - FR BE - FR
There is no evidence of common
seasonal patterns between the
series at 5 per cent level
There is no evidence of common
seasonal patterns between the
series at 5 per cent level
Numerical example
Enrico Infante, Dario Buono
15. 15
An excel file with all the calculations is available on request
VAR Mean Square df
Months 0.9579 11
Years 0.0202 9
Series 0.0042 1
Residual 0.0181 99
2314.0=F 6315.0=− valueP
LU - BE
Common seasonal patterns
between the series present
at 5 per cent level
LU and BE have the same seasonal pattern, so it is possible to Seasonally
Adjust them together, using a Mixed Approach
( ) ( ) ( )tttt LUBESAFRSAYSA ++=
Numerical example
Enrico Infante, Dario Buono
16. 16
Theoretical review (F-ratio, trend, co-movements test)
Future research line
Enrico Infante, Dario Buono
• F-ratio: re-building the test upon the ratio of the between months
variance and the residual variance (comments by Kirchner)
Additive and multiplicative
decompositions
Moving Seasonality
+ -
• A-priori estimation of the trend
• Use of the co-movements test as benchmarking
17. 17
Case study (IPC using Demetra+) - ongoing
Simulations (R) - ongoing
Application with a Tukey’s range test
Future research line
Enrico Infante, Dario Buono
18. 18
[1] J. Higginson – An F Test for the Presence of Moving Seasonality When Using
Census Method II-X-11 Variant – Statistics Canada, 1975
[2] R. Astolfi, D. Ladiray, G. L. Mazzi – Seasonal Adjustment of European
Aggregates: Direct versus Indirect Approach – European Communities, 2001
[3] F. Busetti, A. Harvey – Seasonality Tests – Journal of Business and Economic
Statistics, Vol. 21, No. 3, pp. 420-436, Jul. 2003
[4] B. C. Surtradhar, E. B. Dagum – Bartlett-type modified test for moving
seasonality with applications – The Statistician, Vol. 47, Part 1, 1998
[5] M. Centoni, G. Cubbadda – Modelling Comovements of Economic Time Series:
A Selective Survey – CEIS, 2011
[7] A. Maravall – An application of the TRAMO-SEATS automatic procedure; direct
versus indirect approach – Computation Statistics & Data Analysis, 2005
[8] R. Cristadoro, R. Sabbatini - The Seasonal Adjustment of the Harmonised Index
of Consumer Prices for the Euro Area: a Comparison of Direct and Indirect
Method – Banca d’Italia, 2000
[9] B. Cohen – Explaning Psychological Statistics (3rd
ed.), Chapter 22: Three-way
ANOVA - New York: John Wiley & Sons, 2007
[10]I. Hindrayanto - Seasonal adjustment: direct, indirect or multivariate method?
– Aenorm, No. 43, 2004
References
Enrico Infante, Dario Buono
19. 19
Many Thanks!!!
Questions?
Enrico Infante, Dario Buono
We are really grateful for all the comments we already received
(in particular from R. Gatto, R. Kirchner, A. Maravall, G.L. Mazzi, J. Palate)