SlideShare a Scribd company logo
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Cyrus Vahid - Principal Architect – AWS Deep Learning
Amazon Web Services
Multivariate Time Series
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Autoregressive Models
• Hyndman[1] defines autoregressive models as:
’’ In an autoregression model, we forecast the variable of
interest using a linear combination of past values of the
variable. The term autoregression indicates that it is a
regression of the variable against itself.’’
• AR(p) model:
𝑦𝑡 = 𝑐 + 𝜙1 𝑦𝑡−1 + 𝜙𝑦𝑡−2 + … + 𝜙𝑦𝑡−𝑝 + 𝑒𝑡
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Auto Regressive Models
𝑦𝑡 = 18 − 0.8𝑦𝑡−1 + 𝑒𝑡 𝑦𝑡 = 8 + 1.3𝑦𝑡 − 1 − 0.7 𝑦𝑡−2 − 2 + 𝑒𝑡
• Autoregressive models are remarkably flexible at handling a wide range of
different time series patterns.
𝑟𝑒𝑓: 𝐻𝑦𝑛𝑑𝑚𝑎𝑛 [1]
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Challenges faced by existing models
• Most methods are designed to forecasting individual
series or small groups. New set of problems have
emerged:
• Forecasting a large number of individual or grouped time
series.
• Trying to learn a global model facing the difficulty of dealing
with scale of different time-series that would otherwise be
related.
• Many older models cannot account for environmental inputs.
• Cold start problem for new items to be included in the forecast.
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Goal
• Ability to learn and generalized from similar series
provides us with the ability to learn more complex models
without overfitting.
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
DeepAR
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Solution
• DeepAR is a forecasting model based on autoregressive
RNNs, which learns a global model from historical data of
all time series in all datasets.[2]
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
DeepAR Advantages
• Minimal manual feature engineering
• Ability to provide forecast for series with little or no history.
• Ability to incorporate a wide range of likelihood models.
• Provides consistent estimates for subgroups.
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
DeepAR Model
• Goal: Given observed values of a series 𝑖 for 𝑡 time-steps, estimating
probability distribution of next 𝑇 steps; more formally, modeling the below
conditional distribution is the goal: 𝑃 𝑧𝑖,𝑡0:𝑇 𝑧𝑖,1:𝑡0
, 𝑥𝑖,1:𝑇
• Parameterized by output of an AR RNN.
𝑄Θ 𝑧𝑖,𝑡0:𝑇 𝑧𝑖,1:𝑡0
, 𝑥𝑖,1:𝑇 =
𝑡=𝑡0
𝑇
𝑄Θ 𝓏𝑖,𝑡 𝑧𝑖,1:𝑡−1, 𝑥𝑖,1:𝑇 =
𝑡=𝑡0
𝑇
ℓ(𝓏𝑖,𝑡|𝜃(𝒉𝑖,𝑡, Θ))
𝒉𝑖,𝑡 = h(𝒉𝑖,𝑡−1, 𝓏𝑖,𝑡−1, 𝑥𝑖,𝑡, Θ)
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
DeepAR Architecture
• DeepAR is an encoder decode architecture, taking a
number of input steps, output from encoder, and
covariates, and predicts for the number of steps indicated
as horizon.
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Likelihood Model – Gaussian
• Gaussian likelihood for real-valued Data
ℓ 𝐺 𝓏 𝜇, 𝜎 = 2𝜋𝜎2 −
1
2 𝑒
−
𝓏−𝜇 2
2𝜎2
𝜇 𝒉𝑖,𝑡 = 𝑤𝜇
𝑇 𝒉𝑖,𝑡 + 𝑏 𝜇
𝜎 𝒉𝑖,𝑡 = log 1 + 𝑒 𝑤 𝜇
𝑇 𝒉𝑖,𝑡+𝑏 𝜎
Softplus activation
Network output
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Likelihood Model – Negative Bionomial
• Negative-binomial likelihood for positive count data. The
Negative Binomial distribution is the distribution that
underlies the stochasticity in over-dispersed count data.[3]
ℓ 𝑁𝐵 𝓏 𝜇, 𝛼 =
Γ 𝓏 +
1
𝛼
Γ 𝓏 + 1 Γ
1
𝛼
1
1 + 𝛼𝜇
1
𝛼 𝛼𝜇
1 + 𝛼𝜇
𝓏
𝜇 𝒉𝑖,𝑡 = log 1 + 𝑒 𝑤 𝜇
𝑇 𝒉𝑖,𝑡+𝑏 𝜇
𝛼 𝒉𝑖,𝑡 = log 1 + 𝑒 𝑤 𝛼
𝑇 𝒉𝑖,𝑡+𝑏 𝛼
• 𝜇 𝑎𝑛𝑑 𝛼𝑎𝑟𝑒 𝑏𝑜𝑡ℎ 𝑜𝑢𝑡𝑝𝑢𝑡 𝑜𝑓 𝑎
𝑑𝑒𝑛𝑠𝑒 𝑙𝑎𝑦𝑒𝑟 𝑤𝑖𝑡ℎ
𝑠𝑜𝑓𝑡𝑝𝑙𝑢𝑠 𝑎𝑐𝑡𝑖𝑣𝑎𝑡𝑖𝑜𝑛
• 𝛼 𝑠𝑐𝑎𝑙𝑒𝑠 𝑣𝑎𝑟𝑖𝑎𝑛𝑐𝑒 𝑟𝑒𝑙𝑎𝑡𝑖𝑣𝑒 𝑡𝑜
𝑡ℎ𝑒 𝑚𝑒𝑎𝑛
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Scaling
• Non-linearity results in loss of scale.
• Solution:
• Dividing AR inputs by item-dependent scale factor.
• Multiplying scale-dependent likelihood by the same factor.
• 𝑣𝑖 = 1 +
1
𝑡0
𝑡=1
𝑡0
𝓏𝑖,𝑡
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Comparison
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Code
https://github.com/awslabs/amazon-sagemaker-
examples/blob/master/introduction_to_amazon_algorithms/deepar_electricity/DeepAR-
Electricity.ipynb
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
LSTNet
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Challenge
• Autoregressive models may fail to capture mixture of long
and short term patterns.`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
`
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Solution – LSTNet[4]
• Long and Short Terms Time-series Networks is designed
to capture mix long- and short-term patterns in data for
multivariate time-series.
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Concept
• Using CNN to discover local dependencies
• RNNs to capture long-term dependencies
• Autoregressive model to handle scale.
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Problem Formulation
• Given 𝑌 = 𝑦1, 𝑦2, … , 𝑦 𝑇 where 𝑦𝑡 𝜖ℝ 𝑛
and 𝑛 is the variable
dimension, the aim is to predict 𝑦 𝑇+ℎ, and h is the horizon.
• Similarly, given 𝑌 = 𝑦1, 𝑦2, … , 𝑦 𝑇+1 , we want to predict
𝑦 𝑇+1+ ℎ
• The input matrix is denoted as 𝑋 = 𝑦1, 𝑦2, … , 𝑦 𝑇 𝜖ℝ 𝑛×𝑇
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Architecture
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Convolutional Component
• Extract short-term patterns in the time dimension as well
as local dependencies between variables.
• Multiple filters of width 𝜔 and height 𝑛 = 𝑛𝑢𝑚_𝑣𝑎𝑟
• ℎ 𝑘 = 𝑅𝐸𝐿𝑈(𝑊𝑘 ∗ 𝑋 + 𝑏 𝑘)
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Recurrent Component
• The output of the Conv layer is simultaneously fed to
Recurrent and Recurrent-skip layers (next slide).
• RNN component is GRU layer with RELU activation.*
𝑟𝑡 = 𝜎 𝑥 𝑡 𝑊𝑥𝑟 + ℎ 𝑡−1 𝑊ℎ𝑟 + 𝑏 𝑟
𝑢 𝑡 = 𝜎 𝑥 𝑡 𝑊𝑥𝑢 + ℎ 𝑡−1 𝑊ℎ𝑢 + 𝑏 𝑢
𝑐𝑡 = 𝑅𝐸𝐿𝑈 𝑥 𝑡 𝑊𝑥𝑐 + 𝑟𝑡 ⊙ (ℎ 𝑡−1 𝑊𝑐𝑟) + 𝑏 𝑐
ℎ 𝑡 = 1 − 𝑢 𝑡 ⊙ ℎ 𝑡−1 + 𝑢 𝑡 ⊙ 𝑐𝑡
* The implementation of the paper is using tanh, but the authors claim is that RELU performs better than tanh
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Recurrent-skip Component
• Recurrent skip component is a recurrent layer that
captures lagged long-term dependencies according to the
appropriate lag. For instance hourly electricity
consumption would have a lag of 24 time steps.
𝑟𝑡 = 𝜎 𝑥 𝑡 𝑊𝑥𝑟 + ℎ 𝑡−𝑝 𝑊ℎ𝑟 + 𝑏 𝑟
𝑢 𝑡 = 𝜎 𝑥 𝑡 𝑊𝑥𝑢 + ℎ 𝑡−𝑝 𝑊ℎ𝑢 + 𝑏 𝑢
𝑐𝑡 = 𝑅𝐸𝐿𝑈 𝑥 𝑡 𝑊𝑥𝑐 + 𝑟𝑡 ⊙ (ℎ 𝑡−𝑝 𝑊𝑐𝑟) + 𝑏 𝑐
ℎ 𝑡 = 1 − 𝑢 𝑡 ⊙ ℎ 𝑡−𝑝 + 𝑢 𝑡 ⊙ 𝑐𝑡
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Combining Recurrent and Recurrent-skip Outputs
• A Dense layer combines the output from recurrent layers.
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Temporal Attention Layer
• In case of non-seasonal data skip step p is not useful.
• In such cases an attention mechanism is used, which
learns the weighted combination of hidden
representations at each window position of the input
matrix.
𝛼 𝑡 = 𝐴𝑡𝑡𝑛𝑆𝑐𝑜𝑟𝑒 𝐻 𝑇
𝑅
, ℎ 𝑇−1
𝑅
; 𝛼 𝑡 𝜖ℝ 𝑞
: 𝐴𝑡𝑡𝑛. 𝑤𝑒𝑖𝑔ℎ𝑡𝑠
𝐻 𝑇
𝑅
= ℎ 𝑡−𝑞
𝑅
, … , ℎ 𝑡−1
𝑅
: 𝑠𝑡𝑎𝑐𝑘𝑖𝑛𝑔 ℎ𝑖𝑑𝑑𝑒𝑛 𝑠𝑡𝑎𝑡𝑒𝑠 𝑐𝑜𝑙𝑢𝑚𝑛 − 𝑤𝑖𝑠𝑒𝑙𝑦
𝑐𝑡 = 𝐻𝑡 𝛼 𝑡: context vector
ℎ 𝑡
𝐷
= 𝑊 𝑐𝑡; ℎ 𝑡−1
𝑅
+ 𝑏: 𝑜𝑢𝑡𝑝𝑢𝑡 𝑖𝑠 𝑐𝑜𝑛𝑐𝑎𝑡𝑖𝑛𝑎𝑡𝑖𝑜𝑛 𝑜𝑓 𝑐 𝑎𝑛𝑑 𝑙𝑎𝑠𝑡 𝑤𝑖𝑛𝑑𝑜𝑤 ℎ𝑖𝑑𝑑𝑒𝑛 𝑟𝑒𝑝𝑟𝑒𝑠𝑒𝑛𝑡𝑎𝑡𝑖𝑛𝑜
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Autoregressive Component
• ARC overcomes loss of scale, cased by DNN non-
linearity.
• ARC is a linear AR.
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Final Output
• Final output is obtained by integrating AR and DNN
outputs.
𝑌𝑡 = ℎ 𝑡
𝐷
+ ℎ 𝑡
𝐿
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Objective Function
• The paper suggests using either L1 or L2 loss function.
𝐹: 𝐹𝑟𝑜𝑏𝑒𝑛𝑖𝑜𝑢𝑠 𝑁𝑜𝑟𝑚: 𝐴 𝐹
=
𝑖=1
𝑚
𝑗=1
𝑛
|𝑎𝑖𝑗|2
ℎ: ℎ𝑜𝑟𝑖𝑧𝑜𝑛
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Metrics
• Root Relative Squared Error (RSE): We want lower error.
• Empirical Correlation Coefficient (CORR): We want higher
correlation.
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Comparison
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
Code
https://github.com/safrooze/LSTNet-Gluon
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.
References
1. Forecasting: Principles and Practice – Rob J Hyndman, George Athanasopoulos https://www.otexts.org/fpp/8/3
2. DeepAR: Probabilistic Forecasting with Autoregressive Recurrent Networks - Valentin Flunkert , David Salinas , Jan
Gasthaus. https://arxiv.org/abs/1704.04110
3. http://sherrytowers.com/2014/07/11/negative-binomial-likelihood/
4. Modeling Long- and Short-Term Temporal Patterns with Deep Neural Networks, Guokun Lai et. Al
https://arxiv.org/pdf/1703.07015.pdf
© 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.

More Related Content

What's hot

Time Series Analysis - 2 | Time Series in R | ARIMA Model Forecasting | Data ...
Time Series Analysis - 2 | Time Series in R | ARIMA Model Forecasting | Data ...Time Series Analysis - 2 | Time Series in R | ARIMA Model Forecasting | Data ...
Time Series Analysis - 2 | Time Series in R | ARIMA Model Forecasting | Data ...
Simplilearn
 
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
Simplilearn
 

What's hot (20)

Time-series Analysis in Minutes
Time-series Analysis in MinutesTime-series Analysis in Minutes
Time-series Analysis in Minutes
 
Lstm
LstmLstm
Lstm
 
Optimization for Deep Learning
Optimization for Deep LearningOptimization for Deep Learning
Optimization for Deep Learning
 
Temporal based Recommendation System
Temporal based Recommendation SystemTemporal based Recommendation System
Temporal based Recommendation System
 
Deep Learning for Time Series Data
Deep Learning for Time Series DataDeep Learning for Time Series Data
Deep Learning for Time Series Data
 
Hands-On Machine Learning with Scikit-Learn and TensorFlow - Chapter8
Hands-On Machine Learning with Scikit-Learn and TensorFlow - Chapter8Hands-On Machine Learning with Scikit-Learn and TensorFlow - Chapter8
Hands-On Machine Learning with Scikit-Learn and TensorFlow - Chapter8
 
Time Series Analysis/ Forecasting
Time Series Analysis/ Forecasting  Time Series Analysis/ Forecasting
Time Series Analysis/ Forecasting
 
LSTM Basics
LSTM BasicsLSTM Basics
LSTM Basics
 
Time Series Analysis - 2 | Time Series in R | ARIMA Model Forecasting | Data ...
Time Series Analysis - 2 | Time Series in R | ARIMA Model Forecasting | Data ...Time Series Analysis - 2 | Time Series in R | ARIMA Model Forecasting | Data ...
Time Series Analysis - 2 | Time Series in R | ARIMA Model Forecasting | Data ...
 
rnn BASICS
rnn BASICSrnn BASICS
rnn BASICS
 
Attention
AttentionAttention
Attention
 
Recurrent Neural Networks, LSTM and GRU
Recurrent Neural Networks, LSTM and GRURecurrent Neural Networks, LSTM and GRU
Recurrent Neural Networks, LSTM and GRU
 
Thai Text processing by Transfer Learning using Transformer (Bert)
Thai Text processing by Transfer Learning using Transformer (Bert)Thai Text processing by Transfer Learning using Transformer (Bert)
Thai Text processing by Transfer Learning using Transformer (Bert)
 
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
Recurrent Neural Network (RNN) | RNN LSTM Tutorial | Deep Learning Course | S...
 
Recurrent neural networks rnn
Recurrent neural networks   rnnRecurrent neural networks   rnn
Recurrent neural networks rnn
 
Dimensionality Reduction
Dimensionality ReductionDimensionality Reduction
Dimensionality Reduction
 
Time series analysis
Time series analysisTime series analysis
Time series analysis
 
Model selection
Model selectionModel selection
Model selection
 
1. Linear Algebra for Machine Learning: Linear Systems
1. Linear Algebra for Machine Learning: Linear Systems1. Linear Algebra for Machine Learning: Linear Systems
1. Linear Algebra for Machine Learning: Linear Systems
 
Timeseries forecasting
Timeseries forecastingTimeseries forecasting
Timeseries forecasting
 

Similar to Deep ar presentation

Similar to Deep ar presentation (20)

Multivariate Time Series
Multivariate Time SeriesMultivariate Time Series
Multivariate Time Series
 
Automatic Model Tuning Using Amazon SageMaker (AIM412) - AWS re:Invent 2018
Automatic Model Tuning Using Amazon SageMaker (AIM412) - AWS re:Invent 2018Automatic Model Tuning Using Amazon SageMaker (AIM412) - AWS re:Invent 2018
Automatic Model Tuning Using Amazon SageMaker (AIM412) - AWS re:Invent 2018
 
Get the Most out of Your Amazon Elasticsearch Service Domain (ANT334-R1) - AW...
Get the Most out of Your Amazon Elasticsearch Service Domain (ANT334-R1) - AW...Get the Most out of Your Amazon Elasticsearch Service Domain (ANT334-R1) - AW...
Get the Most out of Your Amazon Elasticsearch Service Domain (ANT334-R1) - AW...
 
Achieving Global Consistency Using AWS CloudFormation StackSets - AWS Online ...
Achieving Global Consistency Using AWS CloudFormation StackSets - AWS Online ...Achieving Global Consistency Using AWS CloudFormation StackSets - AWS Online ...
Achieving Global Consistency Using AWS CloudFormation StackSets - AWS Online ...
 
Accelerate Your Analytic Queries with Amazon Aurora Parallel Query (DAT362) -...
Accelerate Your Analytic Queries with Amazon Aurora Parallel Query (DAT362) -...Accelerate Your Analytic Queries with Amazon Aurora Parallel Query (DAT362) -...
Accelerate Your Analytic Queries with Amazon Aurora Parallel Query (DAT362) -...
 
Build Deep Learning Applications Using Apache MXNet - Featuring Chick-fil-A (...
Build Deep Learning Applications Using Apache MXNet - Featuring Chick-fil-A (...Build Deep Learning Applications Using Apache MXNet - Featuring Chick-fil-A (...
Build Deep Learning Applications Using Apache MXNet - Featuring Chick-fil-A (...
 
16th Athens Big Data Meetup - 1st Talk - An Introduction to Machine Learning ...
16th Athens Big Data Meetup - 1st Talk - An Introduction to Machine Learning ...16th Athens Big Data Meetup - 1st Talk - An Introduction to Machine Learning ...
16th Athens Big Data Meetup - 1st Talk - An Introduction to Machine Learning ...
 
Serverless State Management & Orchestration for Modern Apps (API302) - AWS re...
Serverless State Management & Orchestration for Modern Apps (API302) - AWS re...Serverless State Management & Orchestration for Modern Apps (API302) - AWS re...
Serverless State Management & Orchestration for Modern Apps (API302) - AWS re...
 
Build Deep Learning Applications Using Apache MXNet, Featuring Workday (AIM40...
Build Deep Learning Applications Using Apache MXNet, Featuring Workday (AIM40...Build Deep Learning Applications Using Apache MXNet, Featuring Workday (AIM40...
Build Deep Learning Applications Using Apache MXNet, Featuring Workday (AIM40...
 
Deep Learning in Java with Apache MXNet
Deep Learning in Java with Apache MXNetDeep Learning in Java with Apache MXNet
Deep Learning in Java with Apache MXNet
 
Using Java to deploy Deep Learning models with MXNet
Using Java to deploy Deep Learning models with MXNetUsing Java to deploy Deep Learning models with MXNet
Using Java to deploy Deep Learning models with MXNet
 
Deep Dive on Amazon Neptune (DAT403) - AWS re:Invent 2018
Deep Dive on Amazon Neptune (DAT403) - AWS re:Invent 2018Deep Dive on Amazon Neptune (DAT403) - AWS re:Invent 2018
Deep Dive on Amazon Neptune (DAT403) - AWS re:Invent 2018
 
Advanced Design Patterns for Amazon DynamoDB - Workshop (DAT404-R1) - AWS re:...
Advanced Design Patterns for Amazon DynamoDB - Workshop (DAT404-R1) - AWS re:...Advanced Design Patterns for Amazon DynamoDB - Workshop (DAT404-R1) - AWS re:...
Advanced Design Patterns for Amazon DynamoDB - Workshop (DAT404-R1) - AWS re:...
 
Build Deep Learning Applications Using MXNet and Amazon SageMaker (AIM418) - ...
Build Deep Learning Applications Using MXNet and Amazon SageMaker (AIM418) - ...Build Deep Learning Applications Using MXNet and Amazon SageMaker (AIM418) - ...
Build Deep Learning Applications Using MXNet and Amazon SageMaker (AIM418) - ...
 
Deep Learning with MXNet
Deep Learning with MXNetDeep Learning with MXNet
Deep Learning with MXNet
 
Optimising your Amazon Redshift Cluster for Peak Performance
Optimising your Amazon Redshift Cluster for Peak PerformanceOptimising your Amazon Redshift Cluster for Peak Performance
Optimising your Amazon Redshift Cluster for Peak Performance
 
Building Applications with Apache MXNet
Building Applications with Apache MXNetBuilding Applications with Apache MXNet
Building Applications with Apache MXNet
 
How Intuit TurboTax Ran Entirely on AWS for 2017 Taxes (ARC307) - AWS re:Inve...
How Intuit TurboTax Ran Entirely on AWS for 2017 Taxes (ARC307) - AWS re:Inve...How Intuit TurboTax Ran Entirely on AWS for 2017 Taxes (ARC307) - AWS re:Inve...
How Intuit TurboTax Ran Entirely on AWS for 2017 Taxes (ARC307) - AWS re:Inve...
 
[REPEAT] Deep Learning for Developers: An Introduction, Featuring Samsung SDS...
[REPEAT] Deep Learning for Developers: An Introduction, Featuring Samsung SDS...[REPEAT] Deep Learning for Developers: An Introduction, Featuring Samsung SDS...
[REPEAT] Deep Learning for Developers: An Introduction, Featuring Samsung SDS...
 
Analyze Slide Images and Process Phenotypic Assays at Scale on AWS (CMP358) -...
Analyze Slide Images and Process Phenotypic Assays at Scale on AWS (CMP358) -...Analyze Slide Images and Process Phenotypic Assays at Scale on AWS (CMP358) -...
Analyze Slide Images and Process Phenotypic Assays at Scale on AWS (CMP358) -...
 

Recently uploaded

The importance of continents, oceans and plate tectonics for the evolution of...
The importance of continents, oceans and plate tectonics for the evolution of...The importance of continents, oceans and plate tectonics for the evolution of...
The importance of continents, oceans and plate tectonics for the evolution of...
Sérgio Sacani
 
Gliese 12 b: A Temperate Earth-sized Planet at 12 pc Ideal for Atmospheric Tr...
Gliese 12 b: A Temperate Earth-sized Planet at 12 pc Ideal for Atmospheric Tr...Gliese 12 b: A Temperate Earth-sized Planet at 12 pc Ideal for Atmospheric Tr...
Gliese 12 b: A Temperate Earth-sized Planet at 12 pc Ideal for Atmospheric Tr...
Sérgio Sacani
 
Seminar on Halal AGriculture and Fisheries.pptx
Seminar on Halal AGriculture and Fisheries.pptxSeminar on Halal AGriculture and Fisheries.pptx
Seminar on Halal AGriculture and Fisheries.pptx
RUDYLUMAPINET2
 
Aerodynamics. flippatterncn5tm5ttnj6nmnynyppt
Aerodynamics. flippatterncn5tm5ttnj6nmnynypptAerodynamics. flippatterncn5tm5ttnj6nmnynyppt
Aerodynamics. flippatterncn5tm5ttnj6nmnynyppt
sreddyrahul
 
THYROID-PARATHYROID medical surgical nursing
THYROID-PARATHYROID medical surgical nursingTHYROID-PARATHYROID medical surgical nursing
THYROID-PARATHYROID medical surgical nursing
Jocelyn Atis
 
Pests of Green Manures_Bionomics_IPM_Dr.UPR.pdf
Pests of Green Manures_Bionomics_IPM_Dr.UPR.pdfPests of Green Manures_Bionomics_IPM_Dr.UPR.pdf
Pests of Green Manures_Bionomics_IPM_Dr.UPR.pdf
PirithiRaju
 
Anemia_ different types_causes_ conditions
Anemia_ different types_causes_ conditionsAnemia_ different types_causes_ conditions
Anemia_ different types_causes_ conditions
muralinath2
 

Recently uploaded (20)

Richard's entangled aventures in wonderland
Richard's entangled aventures in wonderlandRichard's entangled aventures in wonderland
Richard's entangled aventures in wonderland
 
NuGOweek 2024 full programme - hosted by Ghent University
NuGOweek 2024 full programme - hosted by Ghent UniversityNuGOweek 2024 full programme - hosted by Ghent University
NuGOweek 2024 full programme - hosted by Ghent University
 
Topography and sediments of the floor of the Bay of Bengal
Topography and sediments of the floor of the Bay of BengalTopography and sediments of the floor of the Bay of Bengal
Topography and sediments of the floor of the Bay of Bengal
 
The importance of continents, oceans and plate tectonics for the evolution of...
The importance of continents, oceans and plate tectonics for the evolution of...The importance of continents, oceans and plate tectonics for the evolution of...
The importance of continents, oceans and plate tectonics for the evolution of...
 
Gliese 12 b: A Temperate Earth-sized Planet at 12 pc Ideal for Atmospheric Tr...
Gliese 12 b: A Temperate Earth-sized Planet at 12 pc Ideal for Atmospheric Tr...Gliese 12 b: A Temperate Earth-sized Planet at 12 pc Ideal for Atmospheric Tr...
Gliese 12 b: A Temperate Earth-sized Planet at 12 pc Ideal for Atmospheric Tr...
 
Seminar on Halal AGriculture and Fisheries.pptx
Seminar on Halal AGriculture and Fisheries.pptxSeminar on Halal AGriculture and Fisheries.pptx
Seminar on Halal AGriculture and Fisheries.pptx
 
GLOBAL AND LOCAL SCENARIO OF FOOD AND NUTRITION.pptx
GLOBAL AND LOCAL SCENARIO OF FOOD AND NUTRITION.pptxGLOBAL AND LOCAL SCENARIO OF FOOD AND NUTRITION.pptx
GLOBAL AND LOCAL SCENARIO OF FOOD AND NUTRITION.pptx
 
Aerodynamics. flippatterncn5tm5ttnj6nmnynyppt
Aerodynamics. flippatterncn5tm5ttnj6nmnynypptAerodynamics. flippatterncn5tm5ttnj6nmnynyppt
Aerodynamics. flippatterncn5tm5ttnj6nmnynyppt
 
FAIRSpectra - Towards a common data file format for SIMS images
FAIRSpectra - Towards a common data file format for SIMS imagesFAIRSpectra - Towards a common data file format for SIMS images
FAIRSpectra - Towards a common data file format for SIMS images
 
THYROID-PARATHYROID medical surgical nursing
THYROID-PARATHYROID medical surgical nursingTHYROID-PARATHYROID medical surgical nursing
THYROID-PARATHYROID medical surgical nursing
 
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.
 
Transport in plants G1.pptx Cambridge IGCSE
Transport in plants G1.pptx Cambridge IGCSETransport in plants G1.pptx Cambridge IGCSE
Transport in plants G1.pptx Cambridge IGCSE
 
INSIGHT Partner Profile: Tampere University
INSIGHT Partner Profile: Tampere UniversityINSIGHT Partner Profile: Tampere University
INSIGHT Partner Profile: Tampere University
 
Pests of Green Manures_Bionomics_IPM_Dr.UPR.pdf
Pests of Green Manures_Bionomics_IPM_Dr.UPR.pdfPests of Green Manures_Bionomics_IPM_Dr.UPR.pdf
Pests of Green Manures_Bionomics_IPM_Dr.UPR.pdf
 
Anemia_ different types_causes_ conditions
Anemia_ different types_causes_ conditionsAnemia_ different types_causes_ conditions
Anemia_ different types_causes_ conditions
 
SCHIZOPHRENIA Disorder/ Brain Disorder.pdf
SCHIZOPHRENIA Disorder/ Brain Disorder.pdfSCHIZOPHRENIA Disorder/ Brain Disorder.pdf
SCHIZOPHRENIA Disorder/ Brain Disorder.pdf
 
EY - Supply Chain Services 2018_template.pptx
EY - Supply Chain Services 2018_template.pptxEY - Supply Chain Services 2018_template.pptx
EY - Supply Chain Services 2018_template.pptx
 
National Biodiversity protection initiatives and Convention on Biological Di...
National Biodiversity protection initiatives and  Convention on Biological Di...National Biodiversity protection initiatives and  Convention on Biological Di...
National Biodiversity protection initiatives and Convention on Biological Di...
 
GEOLOGICAL FIELD REPORT On Kaptai Rangamati Road-Cut Section.pdf
GEOLOGICAL FIELD REPORT  On  Kaptai Rangamati Road-Cut Section.pdfGEOLOGICAL FIELD REPORT  On  Kaptai Rangamati Road-Cut Section.pdf
GEOLOGICAL FIELD REPORT On Kaptai Rangamati Road-Cut Section.pdf
 
insect taxonomy importance systematics and classification
insect taxonomy importance systematics and classificationinsect taxonomy importance systematics and classification
insect taxonomy importance systematics and classification
 

Deep ar presentation

  • 1. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Cyrus Vahid - Principal Architect – AWS Deep Learning Amazon Web Services Multivariate Time Series
  • 2. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Autoregressive Models • Hyndman[1] defines autoregressive models as: ’’ In an autoregression model, we forecast the variable of interest using a linear combination of past values of the variable. The term autoregression indicates that it is a regression of the variable against itself.’’ • AR(p) model: 𝑦𝑡 = 𝑐 + 𝜙1 𝑦𝑡−1 + 𝜙𝑦𝑡−2 + … + 𝜙𝑦𝑡−𝑝 + 𝑒𝑡
  • 3. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Auto Regressive Models 𝑦𝑡 = 18 − 0.8𝑦𝑡−1 + 𝑒𝑡 𝑦𝑡 = 8 + 1.3𝑦𝑡 − 1 − 0.7 𝑦𝑡−2 − 2 + 𝑒𝑡 • Autoregressive models are remarkably flexible at handling a wide range of different time series patterns. 𝑟𝑒𝑓: 𝐻𝑦𝑛𝑑𝑚𝑎𝑛 [1]
  • 4. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Challenges faced by existing models • Most methods are designed to forecasting individual series or small groups. New set of problems have emerged: • Forecasting a large number of individual or grouped time series. • Trying to learn a global model facing the difficulty of dealing with scale of different time-series that would otherwise be related. • Many older models cannot account for environmental inputs. • Cold start problem for new items to be included in the forecast.
  • 5. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Goal • Ability to learn and generalized from similar series provides us with the ability to learn more complex models without overfitting.
  • 6. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. DeepAR
  • 7. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Solution • DeepAR is a forecasting model based on autoregressive RNNs, which learns a global model from historical data of all time series in all datasets.[2]
  • 8. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. DeepAR Advantages • Minimal manual feature engineering • Ability to provide forecast for series with little or no history. • Ability to incorporate a wide range of likelihood models. • Provides consistent estimates for subgroups.
  • 9. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. DeepAR Model • Goal: Given observed values of a series 𝑖 for 𝑡 time-steps, estimating probability distribution of next 𝑇 steps; more formally, modeling the below conditional distribution is the goal: 𝑃 𝑧𝑖,𝑡0:𝑇 𝑧𝑖,1:𝑡0 , 𝑥𝑖,1:𝑇 • Parameterized by output of an AR RNN. 𝑄Θ 𝑧𝑖,𝑡0:𝑇 𝑧𝑖,1:𝑡0 , 𝑥𝑖,1:𝑇 = 𝑡=𝑡0 𝑇 𝑄Θ 𝓏𝑖,𝑡 𝑧𝑖,1:𝑡−1, 𝑥𝑖,1:𝑇 = 𝑡=𝑡0 𝑇 ℓ(𝓏𝑖,𝑡|𝜃(𝒉𝑖,𝑡, Θ)) 𝒉𝑖,𝑡 = h(𝒉𝑖,𝑡−1, 𝓏𝑖,𝑡−1, 𝑥𝑖,𝑡, Θ)
  • 10. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. DeepAR Architecture • DeepAR is an encoder decode architecture, taking a number of input steps, output from encoder, and covariates, and predicts for the number of steps indicated as horizon.
  • 11. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Likelihood Model – Gaussian • Gaussian likelihood for real-valued Data ℓ 𝐺 𝓏 𝜇, 𝜎 = 2𝜋𝜎2 − 1 2 𝑒 − 𝓏−𝜇 2 2𝜎2 𝜇 𝒉𝑖,𝑡 = 𝑤𝜇 𝑇 𝒉𝑖,𝑡 + 𝑏 𝜇 𝜎 𝒉𝑖,𝑡 = log 1 + 𝑒 𝑤 𝜇 𝑇 𝒉𝑖,𝑡+𝑏 𝜎 Softplus activation Network output
  • 12. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Likelihood Model – Negative Bionomial • Negative-binomial likelihood for positive count data. The Negative Binomial distribution is the distribution that underlies the stochasticity in over-dispersed count data.[3] ℓ 𝑁𝐵 𝓏 𝜇, 𝛼 = Γ 𝓏 + 1 𝛼 Γ 𝓏 + 1 Γ 1 𝛼 1 1 + 𝛼𝜇 1 𝛼 𝛼𝜇 1 + 𝛼𝜇 𝓏 𝜇 𝒉𝑖,𝑡 = log 1 + 𝑒 𝑤 𝜇 𝑇 𝒉𝑖,𝑡+𝑏 𝜇 𝛼 𝒉𝑖,𝑡 = log 1 + 𝑒 𝑤 𝛼 𝑇 𝒉𝑖,𝑡+𝑏 𝛼 • 𝜇 𝑎𝑛𝑑 𝛼𝑎𝑟𝑒 𝑏𝑜𝑡ℎ 𝑜𝑢𝑡𝑝𝑢𝑡 𝑜𝑓 𝑎 𝑑𝑒𝑛𝑠𝑒 𝑙𝑎𝑦𝑒𝑟 𝑤𝑖𝑡ℎ 𝑠𝑜𝑓𝑡𝑝𝑙𝑢𝑠 𝑎𝑐𝑡𝑖𝑣𝑎𝑡𝑖𝑜𝑛 • 𝛼 𝑠𝑐𝑎𝑙𝑒𝑠 𝑣𝑎𝑟𝑖𝑎𝑛𝑐𝑒 𝑟𝑒𝑙𝑎𝑡𝑖𝑣𝑒 𝑡𝑜 𝑡ℎ𝑒 𝑚𝑒𝑎𝑛
  • 13. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Scaling • Non-linearity results in loss of scale. • Solution: • Dividing AR inputs by item-dependent scale factor. • Multiplying scale-dependent likelihood by the same factor. • 𝑣𝑖 = 1 + 1 𝑡0 𝑡=1 𝑡0 𝓏𝑖,𝑡
  • 14. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Comparison
  • 15. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Code https://github.com/awslabs/amazon-sagemaker- examples/blob/master/introduction_to_amazon_algorithms/deepar_electricity/DeepAR- Electricity.ipynb
  • 16. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. LSTNet
  • 17. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Challenge • Autoregressive models may fail to capture mixture of long and short term patterns.` ` ` ` ` ` ` ` ` ` ` ` ` ` ` ` ` ` ` ` ` `
  • 18. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Solution – LSTNet[4] • Long and Short Terms Time-series Networks is designed to capture mix long- and short-term patterns in data for multivariate time-series.
  • 19. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Concept • Using CNN to discover local dependencies • RNNs to capture long-term dependencies • Autoregressive model to handle scale.
  • 20. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Problem Formulation • Given 𝑌 = 𝑦1, 𝑦2, … , 𝑦 𝑇 where 𝑦𝑡 𝜖ℝ 𝑛 and 𝑛 is the variable dimension, the aim is to predict 𝑦 𝑇+ℎ, and h is the horizon. • Similarly, given 𝑌 = 𝑦1, 𝑦2, … , 𝑦 𝑇+1 , we want to predict 𝑦 𝑇+1+ ℎ • The input matrix is denoted as 𝑋 = 𝑦1, 𝑦2, … , 𝑦 𝑇 𝜖ℝ 𝑛×𝑇
  • 21. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Architecture
  • 22. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Convolutional Component • Extract short-term patterns in the time dimension as well as local dependencies between variables. • Multiple filters of width 𝜔 and height 𝑛 = 𝑛𝑢𝑚_𝑣𝑎𝑟 • ℎ 𝑘 = 𝑅𝐸𝐿𝑈(𝑊𝑘 ∗ 𝑋 + 𝑏 𝑘)
  • 23. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Recurrent Component • The output of the Conv layer is simultaneously fed to Recurrent and Recurrent-skip layers (next slide). • RNN component is GRU layer with RELU activation.* 𝑟𝑡 = 𝜎 𝑥 𝑡 𝑊𝑥𝑟 + ℎ 𝑡−1 𝑊ℎ𝑟 + 𝑏 𝑟 𝑢 𝑡 = 𝜎 𝑥 𝑡 𝑊𝑥𝑢 + ℎ 𝑡−1 𝑊ℎ𝑢 + 𝑏 𝑢 𝑐𝑡 = 𝑅𝐸𝐿𝑈 𝑥 𝑡 𝑊𝑥𝑐 + 𝑟𝑡 ⊙ (ℎ 𝑡−1 𝑊𝑐𝑟) + 𝑏 𝑐 ℎ 𝑡 = 1 − 𝑢 𝑡 ⊙ ℎ 𝑡−1 + 𝑢 𝑡 ⊙ 𝑐𝑡 * The implementation of the paper is using tanh, but the authors claim is that RELU performs better than tanh
  • 24. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Recurrent-skip Component • Recurrent skip component is a recurrent layer that captures lagged long-term dependencies according to the appropriate lag. For instance hourly electricity consumption would have a lag of 24 time steps. 𝑟𝑡 = 𝜎 𝑥 𝑡 𝑊𝑥𝑟 + ℎ 𝑡−𝑝 𝑊ℎ𝑟 + 𝑏 𝑟 𝑢 𝑡 = 𝜎 𝑥 𝑡 𝑊𝑥𝑢 + ℎ 𝑡−𝑝 𝑊ℎ𝑢 + 𝑏 𝑢 𝑐𝑡 = 𝑅𝐸𝐿𝑈 𝑥 𝑡 𝑊𝑥𝑐 + 𝑟𝑡 ⊙ (ℎ 𝑡−𝑝 𝑊𝑐𝑟) + 𝑏 𝑐 ℎ 𝑡 = 1 − 𝑢 𝑡 ⊙ ℎ 𝑡−𝑝 + 𝑢 𝑡 ⊙ 𝑐𝑡
  • 25. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Combining Recurrent and Recurrent-skip Outputs • A Dense layer combines the output from recurrent layers.
  • 26. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Temporal Attention Layer • In case of non-seasonal data skip step p is not useful. • In such cases an attention mechanism is used, which learns the weighted combination of hidden representations at each window position of the input matrix. 𝛼 𝑡 = 𝐴𝑡𝑡𝑛𝑆𝑐𝑜𝑟𝑒 𝐻 𝑇 𝑅 , ℎ 𝑇−1 𝑅 ; 𝛼 𝑡 𝜖ℝ 𝑞 : 𝐴𝑡𝑡𝑛. 𝑤𝑒𝑖𝑔ℎ𝑡𝑠 𝐻 𝑇 𝑅 = ℎ 𝑡−𝑞 𝑅 , … , ℎ 𝑡−1 𝑅 : 𝑠𝑡𝑎𝑐𝑘𝑖𝑛𝑔 ℎ𝑖𝑑𝑑𝑒𝑛 𝑠𝑡𝑎𝑡𝑒𝑠 𝑐𝑜𝑙𝑢𝑚𝑛 − 𝑤𝑖𝑠𝑒𝑙𝑦 𝑐𝑡 = 𝐻𝑡 𝛼 𝑡: context vector ℎ 𝑡 𝐷 = 𝑊 𝑐𝑡; ℎ 𝑡−1 𝑅 + 𝑏: 𝑜𝑢𝑡𝑝𝑢𝑡 𝑖𝑠 𝑐𝑜𝑛𝑐𝑎𝑡𝑖𝑛𝑎𝑡𝑖𝑜𝑛 𝑜𝑓 𝑐 𝑎𝑛𝑑 𝑙𝑎𝑠𝑡 𝑤𝑖𝑛𝑑𝑜𝑤 ℎ𝑖𝑑𝑑𝑒𝑛 𝑟𝑒𝑝𝑟𝑒𝑠𝑒𝑛𝑡𝑎𝑡𝑖𝑛𝑜
  • 27. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Autoregressive Component • ARC overcomes loss of scale, cased by DNN non- linearity. • ARC is a linear AR.
  • 28. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Final Output • Final output is obtained by integrating AR and DNN outputs. 𝑌𝑡 = ℎ 𝑡 𝐷 + ℎ 𝑡 𝐿
  • 29. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Objective Function • The paper suggests using either L1 or L2 loss function. 𝐹: 𝐹𝑟𝑜𝑏𝑒𝑛𝑖𝑜𝑢𝑠 𝑁𝑜𝑟𝑚: 𝐴 𝐹 = 𝑖=1 𝑚 𝑗=1 𝑛 |𝑎𝑖𝑗|2 ℎ: ℎ𝑜𝑟𝑖𝑧𝑜𝑛
  • 30. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Metrics • Root Relative Squared Error (RSE): We want lower error. • Empirical Correlation Coefficient (CORR): We want higher correlation.
  • 31. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Comparison
  • 32. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. Code https://github.com/safrooze/LSTNet-Gluon
  • 33. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved. References 1. Forecasting: Principles and Practice – Rob J Hyndman, George Athanasopoulos https://www.otexts.org/fpp/8/3 2. DeepAR: Probabilistic Forecasting with Autoregressive Recurrent Networks - Valentin Flunkert , David Salinas , Jan Gasthaus. https://arxiv.org/abs/1704.04110 3. http://sherrytowers.com/2014/07/11/negative-binomial-likelihood/ 4. Modeling Long- and Short-Term Temporal Patterns with Deep Neural Networks, Guokun Lai et. Al https://arxiv.org/pdf/1703.07015.pdf
  • 34. © 2018, Amazon Web Services, Inc. or its affiliates. All rights reserved.