SlideShare a Scribd company logo
1 of 1
Download to read offline
Presentation:
This section aims to study the performance of a feed forward network when used with real data with a time step of 5
days, trying to approximate and predict the UT1-TAI from 1962 to 2013, in order to estimate the maximum prediction
horizon possible with ANN. Two kinds of networks have been simulated and summarized in table 1.
Table 1: ANN 1 & 2 simulated for long term predictions
Training data is used only for training, adjusting weights and bias
Validation set is used for training and to stop it if the errors increase 6 times in a row (validation stop)
Test data is only used to compare models and represents the answer of the network having no effect in training
Performance evaluation: using Mean Square Error & Regression Analysis
Procedure:
• Each training sample goes through N trials, defined by the user (set to 10 in this experimentation), where the
sample is used to train a new network and generate new predictions N times. All network parameters and results
are saved in an .mat file after each trial.
• Training sample is increased with one year of data and the procedure repeats this behavior until the training
sample size reaches its maximum.
Results
Figure 3 shows the long term prediction limit where |UT1-UTC| reaches 0,9s using Neural Network 1 & 2 and the
Least Squares Method.
Fig. 4 shows the prediction of the network 2 at cycle 21 (MJD=47519), and although it manages to begin to predict a
small sinusoidal signal, it fails to actually approximate the real data and diverges quickly soon after. The occasional
good and long prediction showed by these networks seem to be more a product of chance than a good result where
the prediction turns out to be a curve, without changes and tangent to the target data (see Fig.5 for cycle 27 of
Network 1).
Network 1 with validation checks Network 2 without validation checks
2 hidden layers with 6 neurons each
Network parameters:
• Multi-Layer Perceptron networks (MLP)
• Neural Network Structure: Feed Forward
• Tangent-Sigmoid transfer function
• Training algorithms Levenberg-Marquardt (LM) back
propagation
• Random Data Division: division splits the input data into
three sub datasets: 70% for Training, 15% for Validation and
15% for Testing
• Weights and bias initialization: random values
2 hidden layers with 12 neurons each
Network parameters:
• Multi-Layer Perceptron networks (MLP)
• Neural Network Structure: Feed Forward
• Tangent-Sigmoid transfer function
• Training Algorithm: Bayesian Regularization (BR)
• Random Data Division disabled: all the training sample will be
used for training and there will be no validation stop. This means
that each trial will take all 10000 iterations to finish or training will
be stopped when one of the other parameters converge
• Weights and bias initialization: random values
Pedro Lopes1, Jean-Yves Richard1, Christophe Barache1, Daniel Gambis1 , Christian Bizouard1
Prediction of EOPs using Artificial Neural Networks
(1) Observatoire de Paris SYRTE / GRGS, Paris, France (Pedro.Lopes@obspm.fr )
Session 4: Earth Rotation - Theory, modelling and observations
ABSTRACT
The monitoring of the Earth Orientation Parameters
(EOP) variations is the main task of the IERS Earth
Orientation Centre. EOP have applications on
navigation, precise orbit determination or leap seconds
announcements, meaning short and long term
predictions are required.
Currently, the method applied for predictions is based
on deterministic processes, namely Least Square fitting
and autoregressive filtering. Here, we present Artificial
Neural Networks (ANN) as an alternative. These have
successfully been applied for pattern recognition, and
have been tested by various authors for EOP
predictions. Though, no real improvement was shown
when compared with the current methods.
Howbeit, recent mechanisms allow reconsidering the
use of ANN for EOP predictions. Hence, a series of
simulations for short and long term predictions; as well
as statistics and comparisons with the current methods
are presented.
Short Term Prediction
Bibliography
Long Term Prediction
This project’s objective is to study the use of Neural Networks, namely Multi-Layer Perceptron
(MLP), to perform predictions of one of Earth’s rotation parameters, UT1 and to verify if it is
possible to use those networks to predict the Leap Second.
Artificial Neural Networks – ANNs
Like the human brain, the ANN learns by example and can perform tasks such as function
approximation, pattern recognition or prediction of future events. Fig. 1 shows a typical example.
Single Neuron
A single neuron follows a simple mathematical equation, where the output is
the result of the product between inputs and weights (Fig.2 and equation 1).
𝑜𝑢𝑡𝑝𝑢𝑡 = 𝑓 𝑾𝒑 + 𝒃 (1)
The input vector, p of dimensions R, is multiplied by the weight matrix, W. This
matrix is randomly generated by the toolbox every time a new network is
created and trained. The result of this dot product is called weighted input that
is added to the vector b, bias, to form the net input, n.
𝑛 = 𝑊1,1 𝑝1 + 𝑊1,2 𝑝2 + ⋯ 𝑊1,𝑅 𝑝 𝑅 + 𝑏 (2)
This net input is then passed through the transfer function, f, which will in turn
produce the network’s output. The most common and the one used for this
project is the tangent-sigmoid function.
Training the Neural Network
During training, the values of weights and bias are changed to increase the
network’s performance, which is measured by the network performance
function - mean squared error (MSE) between output and targets.
[1] SYRTE. (n.d.). Retrieved from http://hpiers.obspm.fr/eop-
pc/index.php?index=leapsecond&lang=en
[2] Mark Hudson Beale, M. T. (2013). Neural Network Toolbox™ User's Guide. Retrieved
from http://www.mathworks.co.uk/help/nnet/index.html
[3] Levenberg. (n.d.). Retrieved from Mathworks:
http://www.mathworks.co.uk/help/nnet/ref/trainlm.html
[4] Bayesian. (n.d.). Retrieved from Mathworks:
http://www.mathworks.co.uk/help/nnet/ref/trainbr.html
[5] Improved Generalization. (n.d.). Retrieved from Mathworks:
http://www.mathworks.co.uk/help/nnet/ug/improving-generalization.html
[6] Machine Learning Methods in the Environnemental Sciences, Neural Networks and
Kernels, William W. Hsieh, Cambridge University Press
Fig. 1 - A representation of a feed forward ANN with
connections between layers: 2 hidden layers
with 6 neurons each and an output layer.
Fig. 2 - Simple Neuron Diagram: input vector is
sent to the neuron and processed by a transfer
function, resulting in the neural network output
0,00
2,00
4,00
6,00
8,00
10,00
12,00
14,00
16,00
18,00
1966
1967
1968
1969
1970
1971
1972
1973
1974
1975
1976
1977
1978
1979
1980
1981
1982
1983
1984
1985
1986
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
2000
2001
2002
2003
2004
2005
2006
2007
2008
PredictionHorizoninyears
End date of training sample
Neural Network Long Term Prediction
NN1 Best
NN1 Average
NN2 Best
NN2 Average
Gambis Model
Fig.3 - Long Term Prediction using ANN compared to the Least Squares Method (Gambis Model)
Fig.4 – Network 2 Prediction at cycle 21 Fig.5 – Network 1 Prediction at cycle 27: Tangent Prediction
---- Targets function
---- Training results
---- Prediction
---- Targets function
---- Training results
---- Prediction
Presentation:
The prediction horizon is estimated from 5 to 25 days using different kinds of neural networks.
The goal is to minimize Root Mean Squared Error (RMSE) by manipulating several aspects of the networks, such as
training sample size and number of neurons.
Network 1 with Increasing sample size and proportional
number of neurons
Network 2 with Increasing sample size and constant number of
neurons
2 hidden layers with increasing number of neurons from
1 to 12 (the number of training months)
Network Parameters:
• Multi-Layer Perceptron networks (MLP)
• Neural Network Structure: Feed Forward
• Tangent-Sigmoid transfer function
• Training algorithms Levenberg-Marquardt (LM) back propagation
• Random Data Division: removes data division, all data is used for
training
• Weights and bias initialization: random values
2 hidden layers with 4 neurons each
Network Parameters:
• Multi-Layer Perceptron networks (MLP)
• Neural Network Structure: Feed Forward
• Tangent-Sigmoid transfer function
• Training algorithms Levenberg-Marquardt (LM) back propagation
• Random Data Division: removes data division, all data is used for
training
• Weights and bias initialization: random values
Network 3: daily training, 2 hidden layers of 2 neurons
Using the daily data of UT1-TAI for training – Prediction horizon until 10 days
Method: training with different sizes from 4, 10, 20 and 365 days
 These training samples of constant size are trained over 100 steps and for each step 10 trials are performed.
 From the 10 trials, the best is chosen, based on the lowest RMS, which represents the trial with the least error.
 This trial will then represent the prediction error of that step. The same is done for each of the next steps.
 Based on the prediction error of the 10 steps we calculate the RMS for that sample (table 1).
Sample size RMS in miliseconds for a prediction horizon of:
Days 1 days 2 days 3 days 4 days 5 days 6 days 7 days 8 days 9 days 10 days
4 (Série 1) 0,913575 1,69958 2,953057 4,30566 5,6858689 7,088269 8,505233 9,92678 11,35163 12,78
10 (Série 2) 0,336091 0,706539 1,041066 1,37853 1,8503255 2,51942 3,357175 4,32092 5,387449 6,5438
20 (Série 3) 0,185875 0,322345 0,491731 0,67442 0,8713904 1,101273 1,38747 1,74448 2,174952 2,6722
365 (Série 4) 0,651144 0,790716 0,954395 1,14036 1,3516377 1,586719 1,845221 2,12814 2,433499 2,7617
Results
Network 1: The evolution of the RMS for the
smaller samples vs. larger samples suggests that
the dependency between sample size and neurons
is not linear and as such there might be too many
neurons for the larger samples.
Network 2: Based on Network 1 results the RMS
for the larger samples was improved reducing the
number of neurons. The increase in RMS for the
smaller samples supports the previous conclusion.
Network 3: Results show again the necessity of
finding proper balance between sample size and
number of neurons. With the 20 days sample
returning very similar results to the “3 Months”
sample (18 points) of Network 1 at 5 and 10 day
horizon.
Table 4 – Daily predictions for different training samples sizes.
0
2
4
6
8
10
12
14
1 days 2 days 3 days 4 days 5 days 6 days 7 days 8 days 9 days 10 days
RMS(ms)
Série1
Série2
Série3
Série4
Prediction Horizon
Fig. 8 - RMS for daily training 2 hidden layers of 2 neurons
Procedure:
• Iteration of training data size from 1 month to 12 months with 1 month steps
• For each training sample 10 trials are performed and the best is chosen in terms of MSE for each
prediction horizon (from 5 to 25 days by 5 days stepping)
• the training sample window is shifted by 1 month N times in order to collect the RMS for each horizon for
different samples of constant size.
0,00
5,00
10,00
15,00
20,00
25,00
30,00
35,00
40,00
5 days 10 days 15 days 20 days 25 days
RMSinms
Prediction Horizon
RMS vs Prediction Horizon
1 Month
2 Months
3 Months
4 Months
5 Months
6 Months
7 Months
8 Months
9 Months
10 Months
11 Months
12 Months
Fig. 6 - RMS for an increasing sample and number of neurons Fig. 7 - RMS for increasing sample and constant number of neurons
0,00
5,00
10,00
15,00
20,00
25,00
30,00
35,00
5 days 10 days 15 days 20 days 25 days
RMSinms
Prediction Horizon
RMS vs Prediction Horizon
1 Month
2 Months
3 Months
4 Months
5 Months
6 Months
7 Months
8 Months
9 Months
10 Months
11 Months
12 Months
CONCLUSION
• For Long Term Predictions standard MLP appear to have a maximum prediction horizon of 1 year before the
predictions diverge too much from the real function. This and the long computation time that is taken to run such large
training samples makes them very limited option for our applications.
• Short Term Predictions, however, can be applied and are generally faster to compute when using small samples. In
order for them to be accurate, it requires a proper balance between sample size and number of neurons in order to
minimize the RMS.
• Although the standard MLP fails to return reliable long term predictions, Matlab Neural Network Toolbox has more
advanced networks available that could be of interest, such as dynamic networks called NARX (Nonlinear
AutoRegressive with EXternal Input) that use their own output to predict the next values of the series. Their predictive
potential is seen when used to predict a Sinus function, being able to predict the function over 6 periods whereas the
standard MLP only predicts up to a quarter of a period, similar to the limit of long term predictions seen above.
Table 2: ANN 1 & 2 simulated for short term predictions
Table 3 – ANN 3 for daily predictions.
INTRODUCTION

More Related Content

What's hot

Electricity Demand Forecasting Using ANN
Electricity Demand Forecasting Using ANNElectricity Demand Forecasting Using ANN
Electricity Demand Forecasting Using ANNNaren Chandra Kattla
 
PR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networks
PR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networksPR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networks
PR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networksTaesu Kim
 
Learning Methods in a Neural Network
Learning Methods in a Neural NetworkLearning Methods in a Neural Network
Learning Methods in a Neural NetworkSaransh Choudhary
 
08 neural networks
08 neural networks08 neural networks
08 neural networksankit_ppt
 
House Price Estimation as a Function Fitting Problem with using ANN Approach
House Price Estimation as a Function Fitting Problem with using ANN ApproachHouse Price Estimation as a Function Fitting Problem with using ANN Approach
House Price Estimation as a Function Fitting Problem with using ANN ApproachYusuf Uzun
 
Meta learning with memory augmented neural network
Meta learning with memory augmented neural networkMeta learning with memory augmented neural network
Meta learning with memory augmented neural networkKaty Lee
 
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...inventionjournals
 
Regularization in deep learning
Regularization in deep learningRegularization in deep learning
Regularization in deep learningKien Le
 
Activation functions and Training Algorithms for Deep Neural network
Activation functions and Training Algorithms for Deep Neural networkActivation functions and Training Algorithms for Deep Neural network
Activation functions and Training Algorithms for Deep Neural networkGayatri Khanvilkar
 
Neural network image recognition
Neural network image recognitionNeural network image recognition
Neural network image recognitionOleksii Sekundant
 
A tutorial on applying artificial neural networks and geometric brownian moti...
A tutorial on applying artificial neural networks and geometric brownian moti...A tutorial on applying artificial neural networks and geometric brownian moti...
A tutorial on applying artificial neural networks and geometric brownian moti...eSAT Journals
 
Expert system design for elastic scattering neutrons optical model using bpnn
Expert system design for elastic scattering neutrons optical model using bpnnExpert system design for elastic scattering neutrons optical model using bpnn
Expert system design for elastic scattering neutrons optical model using bpnnijcsa
 
IRJET- Mango Classification using Convolutional Neural Networks
IRJET- Mango Classification using Convolutional Neural NetworksIRJET- Mango Classification using Convolutional Neural Networks
IRJET- Mango Classification using Convolutional Neural NetworksIRJET Journal
 
Meta-GMVAE: Mixture of Gaussian VAE for Unsupervised Meta-Learning
Meta-GMVAE: Mixture of Gaussian VAE for Unsupervised Meta-LearningMeta-GMVAE: Mixture of Gaussian VAE for Unsupervised Meta-Learning
Meta-GMVAE: Mixture of Gaussian VAE for Unsupervised Meta-LearningMLAI2
 
Basic Learning Algorithms of ANN
Basic Learning Algorithms of ANNBasic Learning Algorithms of ANN
Basic Learning Algorithms of ANNwaseem khan
 
[딥논읽] Meta-Transfer Learning for Zero-Shot Super-Resolution paper review
[딥논읽] Meta-Transfer Learning for Zero-Shot Super-Resolution paper review[딥논읽] Meta-Transfer Learning for Zero-Shot Super-Resolution paper review
[딥논읽] Meta-Transfer Learning for Zero-Shot Super-Resolution paper reviewtaeseon ryu
 

What's hot (20)

Electricity Demand Forecasting Using ANN
Electricity Demand Forecasting Using ANNElectricity Demand Forecasting Using ANN
Electricity Demand Forecasting Using ANN
 
PR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networks
PR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networksPR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networks
PR12-094: Model-Agnostic Meta-Learning for fast adaptation of deep networks
 
Learning Methods in a Neural Network
Learning Methods in a Neural NetworkLearning Methods in a Neural Network
Learning Methods in a Neural Network
 
08 neural networks
08 neural networks08 neural networks
08 neural networks
 
House Price Estimation as a Function Fitting Problem with using ANN Approach
House Price Estimation as a Function Fitting Problem with using ANN ApproachHouse Price Estimation as a Function Fitting Problem with using ANN Approach
House Price Estimation as a Function Fitting Problem with using ANN Approach
 
Meta learning with memory augmented neural network
Meta learning with memory augmented neural networkMeta learning with memory augmented neural network
Meta learning with memory augmented neural network
 
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
Artificial Neural Network and Multi-Response Optimization in Reliability Meas...
 
Regularization in deep learning
Regularization in deep learningRegularization in deep learning
Regularization in deep learning
 
Activation functions and Training Algorithms for Deep Neural network
Activation functions and Training Algorithms for Deep Neural networkActivation functions and Training Algorithms for Deep Neural network
Activation functions and Training Algorithms for Deep Neural network
 
Neural network image recognition
Neural network image recognitionNeural network image recognition
Neural network image recognition
 
A tutorial on applying artificial neural networks and geometric brownian moti...
A tutorial on applying artificial neural networks and geometric brownian moti...A tutorial on applying artificial neural networks and geometric brownian moti...
A tutorial on applying artificial neural networks and geometric brownian moti...
 
Expert system design for elastic scattering neutrons optical model using bpnn
Expert system design for elastic scattering neutrons optical model using bpnnExpert system design for elastic scattering neutrons optical model using bpnn
Expert system design for elastic scattering neutrons optical model using bpnn
 
Itc stock slide
Itc stock slideItc stock slide
Itc stock slide
 
IRJET- Mango Classification using Convolutional Neural Networks
IRJET- Mango Classification using Convolutional Neural NetworksIRJET- Mango Classification using Convolutional Neural Networks
IRJET- Mango Classification using Convolutional Neural Networks
 
Meta-GMVAE: Mixture of Gaussian VAE for Unsupervised Meta-Learning
Meta-GMVAE: Mixture of Gaussian VAE for Unsupervised Meta-LearningMeta-GMVAE: Mixture of Gaussian VAE for Unsupervised Meta-Learning
Meta-GMVAE: Mixture of Gaussian VAE for Unsupervised Meta-Learning
 
Basic Learning Algorithms of ANN
Basic Learning Algorithms of ANNBasic Learning Algorithms of ANN
Basic Learning Algorithms of ANN
 
01 Introduction to Machine Learning
01 Introduction to Machine Learning01 Introduction to Machine Learning
01 Introduction to Machine Learning
 
Concurrent Programming
Concurrent ProgrammingConcurrent Programming
Concurrent Programming
 
[딥논읽] Meta-Transfer Learning for Zero-Shot Super-Resolution paper review
[딥논읽] Meta-Transfer Learning for Zero-Shot Super-Resolution paper review[딥논읽] Meta-Transfer Learning for Zero-Shot Super-Resolution paper review
[딥논읽] Meta-Transfer Learning for Zero-Shot Super-Resolution paper review
 
Complex system
Complex systemComplex system
Complex system
 

Similar to Poster_Reseau_Neurones_Journees_2013

BACKPROPAGATION LEARNING ALGORITHM BASED ON LEVENBERG MARQUARDT ALGORITHM
BACKPROPAGATION LEARNING ALGORITHM BASED ON LEVENBERG MARQUARDT ALGORITHMBACKPROPAGATION LEARNING ALGORITHM BASED ON LEVENBERG MARQUARDT ALGORITHM
BACKPROPAGATION LEARNING ALGORITHM BASED ON LEVENBERG MARQUARDT ALGORITHMcscpconf
 
Black-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX modelBlack-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX modelIJECEIAES
 
Predicting rainfall using ensemble of ensembles
Predicting rainfall using ensemble of ensemblesPredicting rainfall using ensemble of ensembles
Predicting rainfall using ensemble of ensemblesVarad Meru
 
Comparison of Neural Network Training Functions for Hematoma Classification i...
Comparison of Neural Network Training Functions for Hematoma Classification i...Comparison of Neural Network Training Functions for Hematoma Classification i...
Comparison of Neural Network Training Functions for Hematoma Classification i...IOSR Journals
 
Comparative Study of Pre-Trained Neural Network Models in Detection of Glaucoma
Comparative Study of Pre-Trained Neural Network Models in Detection of GlaucomaComparative Study of Pre-Trained Neural Network Models in Detection of Glaucoma
Comparative Study of Pre-Trained Neural Network Models in Detection of GlaucomaIRJET Journal
 
Power system transient stability margin estimation using artificial neural ne...
Power system transient stability margin estimation using artificial neural ne...Power system transient stability margin estimation using artificial neural ne...
Power system transient stability margin estimation using artificial neural ne...elelijjournal
 
AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...
AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...
AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...IAEME Publication
 
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...Disease Classification using ECG Signal Based on PCA Feature along with GA & ...
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...IRJET Journal
 
Applications of Artificial Neural Networks in Cancer Prediction
Applications of Artificial Neural Networks in Cancer PredictionApplications of Artificial Neural Networks in Cancer Prediction
Applications of Artificial Neural Networks in Cancer PredictionIRJET Journal
 
Large Scale Kernel Learning using Block Coordinate Descent
Large Scale Kernel Learning using Block Coordinate DescentLarge Scale Kernel Learning using Block Coordinate Descent
Large Scale Kernel Learning using Block Coordinate DescentShaleen Kumar Gupta
 
Artificial Neural Networks (ANNS) For Prediction of California Bearing Ratio ...
Artificial Neural Networks (ANNS) For Prediction of California Bearing Ratio ...Artificial Neural Networks (ANNS) For Prediction of California Bearing Ratio ...
Artificial Neural Networks (ANNS) For Prediction of California Bearing Ratio ...IJMER
 
Implementation Of Back-Propagation Neural Network For Isolated Bangla Speech ...
Implementation Of Back-Propagation Neural Network For Isolated Bangla Speech ...Implementation Of Back-Propagation Neural Network For Isolated Bangla Speech ...
Implementation Of Back-Propagation Neural Network For Isolated Bangla Speech ...ijistjournal
 
Implementation Of Back-Propagation Neural Network For Isolated Bangla Speech ...
Implementation Of Back-Propagation Neural Network For Isolated Bangla Speech ...Implementation Of Back-Propagation Neural Network For Isolated Bangla Speech ...
Implementation Of Back-Propagation Neural Network For Isolated Bangla Speech ...ijistjournal
 
Hidden Layer Leraning Vector Quantizatio
Hidden Layer Leraning Vector Quantizatio Hidden Layer Leraning Vector Quantizatio
Hidden Layer Leraning Vector Quantizatio Armando Vieira
 

Similar to Poster_Reseau_Neurones_Journees_2013 (20)

F017533540
F017533540F017533540
F017533540
 
BACKPROPAGATION LEARNING ALGORITHM BASED ON LEVENBERG MARQUARDT ALGORITHM
BACKPROPAGATION LEARNING ALGORITHM BASED ON LEVENBERG MARQUARDT ALGORITHMBACKPROPAGATION LEARNING ALGORITHM BASED ON LEVENBERG MARQUARDT ALGORITHM
BACKPROPAGATION LEARNING ALGORITHM BASED ON LEVENBERG MARQUARDT ALGORITHM
 
Black-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX modelBlack-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX model
 
Predicting rainfall using ensemble of ensembles
Predicting rainfall using ensemble of ensemblesPredicting rainfall using ensemble of ensembles
Predicting rainfall using ensemble of ensembles
 
24csit38.pdf
24csit38.pdf24csit38.pdf
24csit38.pdf
 
Comparison of Neural Network Training Functions for Hematoma Classification i...
Comparison of Neural Network Training Functions for Hematoma Classification i...Comparison of Neural Network Training Functions for Hematoma Classification i...
Comparison of Neural Network Training Functions for Hematoma Classification i...
 
Comparative Study of Pre-Trained Neural Network Models in Detection of Glaucoma
Comparative Study of Pre-Trained Neural Network Models in Detection of GlaucomaComparative Study of Pre-Trained Neural Network Models in Detection of Glaucoma
Comparative Study of Pre-Trained Neural Network Models in Detection of Glaucoma
 
A0270107
A0270107A0270107
A0270107
 
Power system transient stability margin estimation using artificial neural ne...
Power system transient stability margin estimation using artificial neural ne...Power system transient stability margin estimation using artificial neural ne...
Power system transient stability margin estimation using artificial neural ne...
 
AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...
AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...
AN IMPROVED METHOD FOR IDENTIFYING WELL-TEST INTERPRETATION MODEL BASED ON AG...
 
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...Disease Classification using ECG Signal Based on PCA Feature along with GA & ...
Disease Classification using ECG Signal Based on PCA Feature along with GA & ...
 
Applications of Artificial Neural Networks in Cancer Prediction
Applications of Artificial Neural Networks in Cancer PredictionApplications of Artificial Neural Networks in Cancer Prediction
Applications of Artificial Neural Networks in Cancer Prediction
 
Jf3515881595
Jf3515881595Jf3515881595
Jf3515881595
 
Large Scale Kernel Learning using Block Coordinate Descent
Large Scale Kernel Learning using Block Coordinate DescentLarge Scale Kernel Learning using Block Coordinate Descent
Large Scale Kernel Learning using Block Coordinate Descent
 
Ajila (1)
Ajila (1)Ajila (1)
Ajila (1)
 
J04401066071
J04401066071J04401066071
J04401066071
 
Artificial Neural Networks (ANNS) For Prediction of California Bearing Ratio ...
Artificial Neural Networks (ANNS) For Prediction of California Bearing Ratio ...Artificial Neural Networks (ANNS) For Prediction of California Bearing Ratio ...
Artificial Neural Networks (ANNS) For Prediction of California Bearing Ratio ...
 
Implementation Of Back-Propagation Neural Network For Isolated Bangla Speech ...
Implementation Of Back-Propagation Neural Network For Isolated Bangla Speech ...Implementation Of Back-Propagation Neural Network For Isolated Bangla Speech ...
Implementation Of Back-Propagation Neural Network For Isolated Bangla Speech ...
 
Implementation Of Back-Propagation Neural Network For Isolated Bangla Speech ...
Implementation Of Back-Propagation Neural Network For Isolated Bangla Speech ...Implementation Of Back-Propagation Neural Network For Isolated Bangla Speech ...
Implementation Of Back-Propagation Neural Network For Isolated Bangla Speech ...
 
Hidden Layer Leraning Vector Quantizatio
Hidden Layer Leraning Vector Quantizatio Hidden Layer Leraning Vector Quantizatio
Hidden Layer Leraning Vector Quantizatio
 

Poster_Reseau_Neurones_Journees_2013

  • 1. Presentation: This section aims to study the performance of a feed forward network when used with real data with a time step of 5 days, trying to approximate and predict the UT1-TAI from 1962 to 2013, in order to estimate the maximum prediction horizon possible with ANN. Two kinds of networks have been simulated and summarized in table 1. Table 1: ANN 1 & 2 simulated for long term predictions Training data is used only for training, adjusting weights and bias Validation set is used for training and to stop it if the errors increase 6 times in a row (validation stop) Test data is only used to compare models and represents the answer of the network having no effect in training Performance evaluation: using Mean Square Error & Regression Analysis Procedure: • Each training sample goes through N trials, defined by the user (set to 10 in this experimentation), where the sample is used to train a new network and generate new predictions N times. All network parameters and results are saved in an .mat file after each trial. • Training sample is increased with one year of data and the procedure repeats this behavior until the training sample size reaches its maximum. Results Figure 3 shows the long term prediction limit where |UT1-UTC| reaches 0,9s using Neural Network 1 & 2 and the Least Squares Method. Fig. 4 shows the prediction of the network 2 at cycle 21 (MJD=47519), and although it manages to begin to predict a small sinusoidal signal, it fails to actually approximate the real data and diverges quickly soon after. The occasional good and long prediction showed by these networks seem to be more a product of chance than a good result where the prediction turns out to be a curve, without changes and tangent to the target data (see Fig.5 for cycle 27 of Network 1). Network 1 with validation checks Network 2 without validation checks 2 hidden layers with 6 neurons each Network parameters: • Multi-Layer Perceptron networks (MLP) • Neural Network Structure: Feed Forward • Tangent-Sigmoid transfer function • Training algorithms Levenberg-Marquardt (LM) back propagation • Random Data Division: division splits the input data into three sub datasets: 70% for Training, 15% for Validation and 15% for Testing • Weights and bias initialization: random values 2 hidden layers with 12 neurons each Network parameters: • Multi-Layer Perceptron networks (MLP) • Neural Network Structure: Feed Forward • Tangent-Sigmoid transfer function • Training Algorithm: Bayesian Regularization (BR) • Random Data Division disabled: all the training sample will be used for training and there will be no validation stop. This means that each trial will take all 10000 iterations to finish or training will be stopped when one of the other parameters converge • Weights and bias initialization: random values Pedro Lopes1, Jean-Yves Richard1, Christophe Barache1, Daniel Gambis1 , Christian Bizouard1 Prediction of EOPs using Artificial Neural Networks (1) Observatoire de Paris SYRTE / GRGS, Paris, France (Pedro.Lopes@obspm.fr ) Session 4: Earth Rotation - Theory, modelling and observations ABSTRACT The monitoring of the Earth Orientation Parameters (EOP) variations is the main task of the IERS Earth Orientation Centre. EOP have applications on navigation, precise orbit determination or leap seconds announcements, meaning short and long term predictions are required. Currently, the method applied for predictions is based on deterministic processes, namely Least Square fitting and autoregressive filtering. Here, we present Artificial Neural Networks (ANN) as an alternative. These have successfully been applied for pattern recognition, and have been tested by various authors for EOP predictions. Though, no real improvement was shown when compared with the current methods. Howbeit, recent mechanisms allow reconsidering the use of ANN for EOP predictions. Hence, a series of simulations for short and long term predictions; as well as statistics and comparisons with the current methods are presented. Short Term Prediction Bibliography Long Term Prediction This project’s objective is to study the use of Neural Networks, namely Multi-Layer Perceptron (MLP), to perform predictions of one of Earth’s rotation parameters, UT1 and to verify if it is possible to use those networks to predict the Leap Second. Artificial Neural Networks – ANNs Like the human brain, the ANN learns by example and can perform tasks such as function approximation, pattern recognition or prediction of future events. Fig. 1 shows a typical example. Single Neuron A single neuron follows a simple mathematical equation, where the output is the result of the product between inputs and weights (Fig.2 and equation 1). 𝑜𝑢𝑡𝑝𝑢𝑡 = 𝑓 𝑾𝒑 + 𝒃 (1) The input vector, p of dimensions R, is multiplied by the weight matrix, W. This matrix is randomly generated by the toolbox every time a new network is created and trained. The result of this dot product is called weighted input that is added to the vector b, bias, to form the net input, n. 𝑛 = 𝑊1,1 𝑝1 + 𝑊1,2 𝑝2 + ⋯ 𝑊1,𝑅 𝑝 𝑅 + 𝑏 (2) This net input is then passed through the transfer function, f, which will in turn produce the network’s output. The most common and the one used for this project is the tangent-sigmoid function. Training the Neural Network During training, the values of weights and bias are changed to increase the network’s performance, which is measured by the network performance function - mean squared error (MSE) between output and targets. [1] SYRTE. (n.d.). Retrieved from http://hpiers.obspm.fr/eop- pc/index.php?index=leapsecond&lang=en [2] Mark Hudson Beale, M. T. (2013). Neural Network Toolbox™ User's Guide. Retrieved from http://www.mathworks.co.uk/help/nnet/index.html [3] Levenberg. (n.d.). Retrieved from Mathworks: http://www.mathworks.co.uk/help/nnet/ref/trainlm.html [4] Bayesian. (n.d.). Retrieved from Mathworks: http://www.mathworks.co.uk/help/nnet/ref/trainbr.html [5] Improved Generalization. (n.d.). Retrieved from Mathworks: http://www.mathworks.co.uk/help/nnet/ug/improving-generalization.html [6] Machine Learning Methods in the Environnemental Sciences, Neural Networks and Kernels, William W. Hsieh, Cambridge University Press Fig. 1 - A representation of a feed forward ANN with connections between layers: 2 hidden layers with 6 neurons each and an output layer. Fig. 2 - Simple Neuron Diagram: input vector is sent to the neuron and processed by a transfer function, resulting in the neural network output 0,00 2,00 4,00 6,00 8,00 10,00 12,00 14,00 16,00 18,00 1966 1967 1968 1969 1970 1971 1972 1973 1974 1975 1976 1977 1978 1979 1980 1981 1982 1983 1984 1985 1986 1987 1988 1989 1990 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000 2001 2002 2003 2004 2005 2006 2007 2008 PredictionHorizoninyears End date of training sample Neural Network Long Term Prediction NN1 Best NN1 Average NN2 Best NN2 Average Gambis Model Fig.3 - Long Term Prediction using ANN compared to the Least Squares Method (Gambis Model) Fig.4 – Network 2 Prediction at cycle 21 Fig.5 – Network 1 Prediction at cycle 27: Tangent Prediction ---- Targets function ---- Training results ---- Prediction ---- Targets function ---- Training results ---- Prediction Presentation: The prediction horizon is estimated from 5 to 25 days using different kinds of neural networks. The goal is to minimize Root Mean Squared Error (RMSE) by manipulating several aspects of the networks, such as training sample size and number of neurons. Network 1 with Increasing sample size and proportional number of neurons Network 2 with Increasing sample size and constant number of neurons 2 hidden layers with increasing number of neurons from 1 to 12 (the number of training months) Network Parameters: • Multi-Layer Perceptron networks (MLP) • Neural Network Structure: Feed Forward • Tangent-Sigmoid transfer function • Training algorithms Levenberg-Marquardt (LM) back propagation • Random Data Division: removes data division, all data is used for training • Weights and bias initialization: random values 2 hidden layers with 4 neurons each Network Parameters: • Multi-Layer Perceptron networks (MLP) • Neural Network Structure: Feed Forward • Tangent-Sigmoid transfer function • Training algorithms Levenberg-Marquardt (LM) back propagation • Random Data Division: removes data division, all data is used for training • Weights and bias initialization: random values Network 3: daily training, 2 hidden layers of 2 neurons Using the daily data of UT1-TAI for training – Prediction horizon until 10 days Method: training with different sizes from 4, 10, 20 and 365 days  These training samples of constant size are trained over 100 steps and for each step 10 trials are performed.  From the 10 trials, the best is chosen, based on the lowest RMS, which represents the trial with the least error.  This trial will then represent the prediction error of that step. The same is done for each of the next steps.  Based on the prediction error of the 10 steps we calculate the RMS for that sample (table 1). Sample size RMS in miliseconds for a prediction horizon of: Days 1 days 2 days 3 days 4 days 5 days 6 days 7 days 8 days 9 days 10 days 4 (Série 1) 0,913575 1,69958 2,953057 4,30566 5,6858689 7,088269 8,505233 9,92678 11,35163 12,78 10 (Série 2) 0,336091 0,706539 1,041066 1,37853 1,8503255 2,51942 3,357175 4,32092 5,387449 6,5438 20 (Série 3) 0,185875 0,322345 0,491731 0,67442 0,8713904 1,101273 1,38747 1,74448 2,174952 2,6722 365 (Série 4) 0,651144 0,790716 0,954395 1,14036 1,3516377 1,586719 1,845221 2,12814 2,433499 2,7617 Results Network 1: The evolution of the RMS for the smaller samples vs. larger samples suggests that the dependency between sample size and neurons is not linear and as such there might be too many neurons for the larger samples. Network 2: Based on Network 1 results the RMS for the larger samples was improved reducing the number of neurons. The increase in RMS for the smaller samples supports the previous conclusion. Network 3: Results show again the necessity of finding proper balance between sample size and number of neurons. With the 20 days sample returning very similar results to the “3 Months” sample (18 points) of Network 1 at 5 and 10 day horizon. Table 4 – Daily predictions for different training samples sizes. 0 2 4 6 8 10 12 14 1 days 2 days 3 days 4 days 5 days 6 days 7 days 8 days 9 days 10 days RMS(ms) Série1 Série2 Série3 Série4 Prediction Horizon Fig. 8 - RMS for daily training 2 hidden layers of 2 neurons Procedure: • Iteration of training data size from 1 month to 12 months with 1 month steps • For each training sample 10 trials are performed and the best is chosen in terms of MSE for each prediction horizon (from 5 to 25 days by 5 days stepping) • the training sample window is shifted by 1 month N times in order to collect the RMS for each horizon for different samples of constant size. 0,00 5,00 10,00 15,00 20,00 25,00 30,00 35,00 40,00 5 days 10 days 15 days 20 days 25 days RMSinms Prediction Horizon RMS vs Prediction Horizon 1 Month 2 Months 3 Months 4 Months 5 Months 6 Months 7 Months 8 Months 9 Months 10 Months 11 Months 12 Months Fig. 6 - RMS for an increasing sample and number of neurons Fig. 7 - RMS for increasing sample and constant number of neurons 0,00 5,00 10,00 15,00 20,00 25,00 30,00 35,00 5 days 10 days 15 days 20 days 25 days RMSinms Prediction Horizon RMS vs Prediction Horizon 1 Month 2 Months 3 Months 4 Months 5 Months 6 Months 7 Months 8 Months 9 Months 10 Months 11 Months 12 Months CONCLUSION • For Long Term Predictions standard MLP appear to have a maximum prediction horizon of 1 year before the predictions diverge too much from the real function. This and the long computation time that is taken to run such large training samples makes them very limited option for our applications. • Short Term Predictions, however, can be applied and are generally faster to compute when using small samples. In order for them to be accurate, it requires a proper balance between sample size and number of neurons in order to minimize the RMS. • Although the standard MLP fails to return reliable long term predictions, Matlab Neural Network Toolbox has more advanced networks available that could be of interest, such as dynamic networks called NARX (Nonlinear AutoRegressive with EXternal Input) that use their own output to predict the next values of the series. Their predictive potential is seen when used to predict a Sinus function, being able to predict the function over 6 periods whereas the standard MLP only predicts up to a quarter of a period, similar to the limit of long term predictions seen above. Table 2: ANN 1 & 2 simulated for short term predictions Table 3 – ANN 3 for daily predictions. INTRODUCTION