Algorithm and its Properties
Computational Complexity
TIME COMPLEXITY
SPACE COMPLEXITY
Complexity Analysis and Asymptotic notations.
Big-oh-notation (O)
Omega-notation (Ω)
Theta-notation (Θ)
The Best, Average, and Worst Case Analyses.
COMPLEXITY Analyses EXAMPLES.
Comparing GROWTH RATES
Deep Water - GPU Deep Learning for H2O - Arno CandelSri Ambati
Deep Water brings the latest and greatest in the Deep Learning space all under the H2O hood. Use Tensorflow, Mxnet & Caffe all from standard H2O interface's including R, Python & Flow. Also deploy your models easily using the H2O platform.
- Powered by the open source machine learning software H2O.ai. Contributors welcome at: https://github.com/h2oai
- To view videos on H2O open source machine learning software, go to: https://www.youtube.com/user/0xdata
H2O Deep Water - Making Deep Learning Accessible to EveryoneSri Ambati
Deep Water is H2O's integration with multiple open source deep learning libraries such as TensorFlow, MXNet and Caffe. On top of the performance gains from GPU backends, Deep Water naturally inherits all H2O properties in scalability. ease of use and deployment. In this talk, I will go through the motivation and benefits of Deep Water. After that, I will demonstrate how to build and deploy deep learning models with or without programming experience using H2O's R/Python/Flow (Web) interfaces.
Jo-fai (or Joe) is a data scientist at H2O.ai. Before joining H2O, he was in the business intelligence team at Virgin Media in UK where he developed data products to enable quick and smart business decisions. He also worked remotely for Domino Data Lab in the US as a data science evangelist promoting products via blogging and giving talks at meetups. Joe has a background in water engineering. Before his data science journey, he was an EngD research engineer at STREAM Industrial Doctorate Centre working on machine learning techniques for drainage design optimization. Prior to that, he was an asset management consultant specialized in data mining and constrained optimization for the utilities sector in the UK and abroad. He also holds an MSc in Environmental Management and a BEng in Civil Engineering.
Deep Water - Bringing Tensorflow, Caffe, Mxnet to H2OSri Ambati
Arno Candel introduces Deep Water, which brings Tensorflow, Caffe, Mxnet to H2O. It also brings support for GPUs, image classification, NLP and much more to H2O.
- Powered by the open source machine learning software H2O.ai. Contributors welcome at: https://github.com/h2oai
- To view videos on H2O open source machine learning software, go to: https://www.youtube.com/user/0xdata
H2O.ai basic components and model deployment pipeline presented. Benchmark for scalability, speed and accuracy of machine learning libraries for classification presented from https://github.com/szilard/benchm-ml.
Algorithm and its Properties
Computational Complexity
TIME COMPLEXITY
SPACE COMPLEXITY
Complexity Analysis and Asymptotic notations.
Big-oh-notation (O)
Omega-notation (Ω)
Theta-notation (Θ)
The Best, Average, and Worst Case Analyses.
COMPLEXITY Analyses EXAMPLES.
Comparing GROWTH RATES
Deep Water - GPU Deep Learning for H2O - Arno CandelSri Ambati
Deep Water brings the latest and greatest in the Deep Learning space all under the H2O hood. Use Tensorflow, Mxnet & Caffe all from standard H2O interface's including R, Python & Flow. Also deploy your models easily using the H2O platform.
- Powered by the open source machine learning software H2O.ai. Contributors welcome at: https://github.com/h2oai
- To view videos on H2O open source machine learning software, go to: https://www.youtube.com/user/0xdata
H2O Deep Water - Making Deep Learning Accessible to EveryoneSri Ambati
Deep Water is H2O's integration with multiple open source deep learning libraries such as TensorFlow, MXNet and Caffe. On top of the performance gains from GPU backends, Deep Water naturally inherits all H2O properties in scalability. ease of use and deployment. In this talk, I will go through the motivation and benefits of Deep Water. After that, I will demonstrate how to build and deploy deep learning models with or without programming experience using H2O's R/Python/Flow (Web) interfaces.
Jo-fai (or Joe) is a data scientist at H2O.ai. Before joining H2O, he was in the business intelligence team at Virgin Media in UK where he developed data products to enable quick and smart business decisions. He also worked remotely for Domino Data Lab in the US as a data science evangelist promoting products via blogging and giving talks at meetups. Joe has a background in water engineering. Before his data science journey, he was an EngD research engineer at STREAM Industrial Doctorate Centre working on machine learning techniques for drainage design optimization. Prior to that, he was an asset management consultant specialized in data mining and constrained optimization for the utilities sector in the UK and abroad. He also holds an MSc in Environmental Management and a BEng in Civil Engineering.
Deep Water - Bringing Tensorflow, Caffe, Mxnet to H2OSri Ambati
Arno Candel introduces Deep Water, which brings Tensorflow, Caffe, Mxnet to H2O. It also brings support for GPUs, image classification, NLP and much more to H2O.
- Powered by the open source machine learning software H2O.ai. Contributors welcome at: https://github.com/h2oai
- To view videos on H2O open source machine learning software, go to: https://www.youtube.com/user/0xdata
H2O.ai basic components and model deployment pipeline presented. Benchmark for scalability, speed and accuracy of machine learning libraries for classification presented from https://github.com/szilard/benchm-ml.
Time Series Forecasting Using Recurrent Neural Network and Vector Autoregress...Databricks
Given the resurgence of neural network-based techniques in recent years, it is important for data science practitioner to understand how to apply these techniques and the tradeoffs between neural network-based and traditional statistical methods.
This lecture discusses two specific techniques: Vector Autoregressive (VAR) Models and Recurrent Neural Network (RNN). The former is one of the most important class of multivariate time series statistical models applied in finance while the latter is a neural network architecture that is suitable for time series forecasting. I’ll demonstrate how they are implemented in practice and compares their advantages and disadvantages. Real-world applications, demonstrated using python and Spark, are used to illustrate these techniques. While not the focus in this lecture, exploratory time series data analysis using time-series plot, plots of autocorrelation (i.e. correlogram), plots of partial autocorrelation, plots of cross-correlations, histogram, and kernel density plot, will also be included in the demo.
The attendees will learn – the formulation of a time series forecasting problem statement in context of VAR and RNN – the application of Recurrent Neural Network-based techniques in time series forecasting – the application of Vector Autoregressive Models in multivariate time series forecasting – the pros and cons of using VAR and RNN-based techniques in the context of financial time series forecasting – When to use VAR and when to use RNN-based techniques
FellowBuddy.com is an innovative platform that brings students together to share notes, exam papers, study guides, project reports and presentation for upcoming exams.
We connect Students who have an understanding of course material with Students who need help.
Benefits:-
# Students can catch up on notes they missed because of an absence.
# Underachievers can find peer developed notes that break down lecture and study material in a way that they can understand
# Students can earn better grades, save time and study effectively
Our Vision & Mission – Simplifying Students Life
Our Belief – “The great breakthrough in your life comes when you realize it, that you can learn anything you need to learn; to accomplish any goal that you have set for yourself. This means there are no limits on what you can be, have or do.”
Like Us - https://www.facebook.com/FellowBuddycom
Online learning with structured streaming, spark summit brussels 2016Ram Sriharsha
Structured Streaming is a new API in Spark 2.0 that simplifies the end to end development of continuous applications. One such continuous application is online model updates: Online models are incrementally updated with new data and can be continuously queried while being updated. As a result, they can be fast to train and leverage new data faster than offline algorithms. In this talk, we give a brief introduction the area of online learning and describe how online model updates can be built using structured streaming APIs. The end result is a robust pipeline for updating models that is scalable, fast and fault tolerant.
Introduction of Chainer, a framework for neural networks, v1.11. Slides used for the student seminar on July 20, 2016, at Sugiyama-Sato lab in the Univ. of Tokyo.
Remember what you learn “A trainable spaced repetition model for online skill...Anni Sapountzi
While acquiring a new skill, it is frustrating and hard to recall what was recently practiced. Space repetition models advocate this by scheduling practice opportunities. The schedule considers the decline of memory retention over time. Despite the success of these models, their parameters are often chosen by humans.
What if we could learn these parameters in a data driven way? Then we could improve the accuracy of the scheduling decisions. Additionally, it could allow the model to adapt towards the individuals' needs. In this talk, we will explore the Half-Life Regression algorithm and the use of essential python libraries to fit large scale learners' data.
Artificial Intelligence Course: Linear models ananth
In this presentation we present the linear models: Regression and Classification. We illustrate with several examples. Concepts such as underfitting (Bias) and overfitting (Variance) are presented. Linear models can be used as stand alone classifiers for simple cases and they are essential building blocks as a part of larger deep learning networks
Choosing the right process improvement tool for your project.
Learn how an experienced engineer decides when simulation is the right tool for his projects,
and when it isn't.
With the evolution of process improvement software, it can be difficult to decide the right tool for the job. Using something too powerful and complex can be a lengthy and unnecessary process, but underestimating the depth of analysis required and choosing something too simplistic early in a project can result in repeated work later.
Time Series Forecasting Using Recurrent Neural Network and Vector Autoregress...Databricks
Given the resurgence of neural network-based techniques in recent years, it is important for data science practitioner to understand how to apply these techniques and the tradeoffs between neural network-based and traditional statistical methods.
This lecture discusses two specific techniques: Vector Autoregressive (VAR) Models and Recurrent Neural Network (RNN). The former is one of the most important class of multivariate time series statistical models applied in finance while the latter is a neural network architecture that is suitable for time series forecasting. I’ll demonstrate how they are implemented in practice and compares their advantages and disadvantages. Real-world applications, demonstrated using python and Spark, are used to illustrate these techniques. While not the focus in this lecture, exploratory time series data analysis using time-series plot, plots of autocorrelation (i.e. correlogram), plots of partial autocorrelation, plots of cross-correlations, histogram, and kernel density plot, will also be included in the demo.
The attendees will learn – the formulation of a time series forecasting problem statement in context of VAR and RNN – the application of Recurrent Neural Network-based techniques in time series forecasting – the application of Vector Autoregressive Models in multivariate time series forecasting – the pros and cons of using VAR and RNN-based techniques in the context of financial time series forecasting – When to use VAR and when to use RNN-based techniques
FellowBuddy.com is an innovative platform that brings students together to share notes, exam papers, study guides, project reports and presentation for upcoming exams.
We connect Students who have an understanding of course material with Students who need help.
Benefits:-
# Students can catch up on notes they missed because of an absence.
# Underachievers can find peer developed notes that break down lecture and study material in a way that they can understand
# Students can earn better grades, save time and study effectively
Our Vision & Mission – Simplifying Students Life
Our Belief – “The great breakthrough in your life comes when you realize it, that you can learn anything you need to learn; to accomplish any goal that you have set for yourself. This means there are no limits on what you can be, have or do.”
Like Us - https://www.facebook.com/FellowBuddycom
Online learning with structured streaming, spark summit brussels 2016Ram Sriharsha
Structured Streaming is a new API in Spark 2.0 that simplifies the end to end development of continuous applications. One such continuous application is online model updates: Online models are incrementally updated with new data and can be continuously queried while being updated. As a result, they can be fast to train and leverage new data faster than offline algorithms. In this talk, we give a brief introduction the area of online learning and describe how online model updates can be built using structured streaming APIs. The end result is a robust pipeline for updating models that is scalable, fast and fault tolerant.
Introduction of Chainer, a framework for neural networks, v1.11. Slides used for the student seminar on July 20, 2016, at Sugiyama-Sato lab in the Univ. of Tokyo.
Remember what you learn “A trainable spaced repetition model for online skill...Anni Sapountzi
While acquiring a new skill, it is frustrating and hard to recall what was recently practiced. Space repetition models advocate this by scheduling practice opportunities. The schedule considers the decline of memory retention over time. Despite the success of these models, their parameters are often chosen by humans.
What if we could learn these parameters in a data driven way? Then we could improve the accuracy of the scheduling decisions. Additionally, it could allow the model to adapt towards the individuals' needs. In this talk, we will explore the Half-Life Regression algorithm and the use of essential python libraries to fit large scale learners' data.
Artificial Intelligence Course: Linear models ananth
In this presentation we present the linear models: Regression and Classification. We illustrate with several examples. Concepts such as underfitting (Bias) and overfitting (Variance) are presented. Linear models can be used as stand alone classifiers for simple cases and they are essential building blocks as a part of larger deep learning networks
Choosing the right process improvement tool for your project.
Learn how an experienced engineer decides when simulation is the right tool for his projects,
and when it isn't.
With the evolution of process improvement software, it can be difficult to decide the right tool for the job. Using something too powerful and complex can be a lengthy and unnecessary process, but underestimating the depth of analysis required and choosing something too simplistic early in a project can result in repeated work later.
Building LLM Solutions using Open Source and Closed Source Solutions in Coher...Sri Ambati
Sandeep Singh, Head of Applied AI Computer Vision, Beans.ai
H2O Open Source GenAI World SF 2023
In the modern era of machine learning, leveraging both open-source and closed-source solutions has become paramount for achieving cutting-edge results. This talk delves into the intricacies of seamlessly integrating open-source Large Language Model (LLM) solutions like Vicuna, Falcon, and Llama with industry giants such as ChatGPT and Google's Palm. As the demand for fine-tuned and specialized datasets grows, it is imperative to understand the synergy between these tools. Attendees will gain insights into best practices for building and enriching datasets tailored for fine-tuning tasks, ensuring that their LLM projects are both robust and efficient. Through real-world examples and hands-on demonstrations, this talk will equip attendees with the knowledge to harness the power of both open and closed-source tools in a coherent and effective manner.
Patrick Hall, Professor, AI Risk Management, The George Washington University
H2O Open Source GenAI World SF 2023
Language models are incredible engineering breakthroughs but require auditing and risk management before productization. These systems raise concerns about toxicity, transparency and reproducibility, intellectual property licensing and ownership, disinformation and misinformation, supply chains, and more. How can your organization leverage these new tools without taking on undue or unknown risks? While language models and associated risk management are in their infancy, a small number of best practices in governance and risk are starting to emerge. If you have a language model use case in mind, want to understand your risks, and do something about them, this presentation is for you!
Dr. Alexy Khrabrov, Open Source Science Community Director, IBM
H2O Open Source GenAI World SF 2023
In this talk, Dr. Alexy Khrabrov, recently elected Chair of the new Generative AI Commons at Linux Foundation for AI & Data, outlines the OSS AI landscape, challenges, and opportunities. With new models and frameworks being unveiled weekly, one thing remains constant: community building and validation of all aspects of AI is key to reliable and responsible AI we can use for business and society needs. Industrial AI is one key area where such community validation can prove invaluable.
Michelle Tanco, Head of Product, H2O.ai
H2O Open Source GenAI World SF 2023
Learn how the makers at H2O.ai are building internal tools to solve real use cases using H2O Wave and h2oGPT. We will walk through an end-to-end use case and discuss how to incorporate business rules and generated content to rapidly develop custom AI apps using only Python APIs.
Applied Gen AI for the Finance Vertical Sri Ambati
Megan Kurka, Vice President, Customer Data Scientist, H2O.ai
H2O Open Source GenAI World SF 2023
Discover the transformative power of Applied Gen AI. Learn how the H2O team builds customized applications and workflows that integrate capabilities of Gen AI and AutoML specifically designed to address and enhance financial use cases. Explore real world examples, learn best practices, and witness firsthand how our innovative solutions are reshaping the landscape of finance technology.
Practitioner's Guide to LLMs: Exploring Use Cases and a Glimpse Beyond Curren...Sri Ambati
Pascal Pfeiffer, Principal Data Scientist, H2O.ai
H2O Open Source GenAI World SF 2023
This talk dives into the expansive ecosystem of Large Language Models (LLMs), offering practitioners an insightful guide to various relevant applications, from natural language understanding to creative content generation. While exploring use cases across different industries, it also honestly addresses the current limitations of LLMs and anticipates future advancements.
Introducción al Aprendizaje Automatico con H2O-3 (1)Sri Ambati
En esta reunión virtual, damos una introducción a la plataforma de aprendizaje automático de código abierto número 1, H2O-3 y te mostramos cómo puedes usarla para desarrollar modelos para resolver diferentes casos de uso.
From Rapid Prototypes to an end-to-end Model Deployment: an AI Hedge Fund Use...Sri Ambati
Numerai is an open, crowd-sourced hedge fund powered by predictions from data scientists around the world. In return, participants are rewarded with weekly payouts in crypto.
In this talk, Joe will give an overview of the Numerai tournament based on his own experience. He will then explain how he automates the time-consuming tasks such as testing different modelling strategies, scoring new datasets, submitting predictions to Numerai as well as monitoring model performance with H2O Driverless AI and R.
AI Foundations Course Module 1 - Shifting to the Next Step in Your AI Transfo...Sri Ambati
In this session, you will learn about what you should do after you’ve taken an AI transformation baseline. Over the span of this session, we will discuss the next steps in moving toward AI readiness through alignment of talent and tools to drive successful adoption and continuous use within an organization.
To find additional videos on AI courses, earn badges, join the courses at H2O.ai Learning Center: https://training.h2o.ai/products/ai-foundations-course
To find the Youtube video about this presentation: https://youtu.be/K1Cl3x3rd8g
Speaker:
Chemere Davis (H2O.ai - Senior Data Scientist Training Specialist)
GraphRAG is All You need? LLM & Knowledge GraphGuy Korland
Guy Korland, CEO and Co-founder of FalkorDB, will review two articles on the integration of language models with knowledge graphs.
1. Unifying Large Language Models and Knowledge Graphs: A Roadmap.
https://arxiv.org/abs/2306.08302
2. Microsoft Research's GraphRAG paper and a review paper on various uses of knowledge graphs:
https://www.microsoft.com/en-us/research/blog/graphrag-unlocking-llm-discovery-on-narrative-private-data/
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Elevating Tactical DDD Patterns Through Object CalisthenicsDorra BARTAGUIZ
After immersing yourself in the blue book and its red counterpart, attending DDD-focused conferences, and applying tactical patterns, you're left with a crucial question: How do I ensure my design is effective? Tactical patterns within Domain-Driven Design (DDD) serve as guiding principles for creating clear and manageable domain models. However, achieving success with these patterns requires additional guidance. Interestingly, we've observed that a set of constraints initially designed for training purposes remarkably aligns with effective pattern implementation, offering a more ‘mechanical’ approach. Let's explore together how Object Calisthenics can elevate the design of your tactical DDD patterns, offering concrete help for those venturing into DDD for the first time!
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
UiPath Test Automation using UiPath Test Suite series, part 3DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 3. In this session, we will cover desktop automation along with UI automation.
Topics covered:
UI automation Introduction,
UI automation Sample
Desktop automation flow
Pradeep Chinnala, Senior Consultant Automation Developer @WonderBotz and UiPath MVP
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
Essentials of Automations: Optimizing FME Workflows with Parameters
Using Machine Learning For Solving Time Series Probelms
1. H2O Time Series
Using machine learning for solving time series
problems
By Danilo Bassi
2. Machine learning with time series
Typical steps:
1. Conversion data to regular time series (if needed)
• Sampling rate or period to use
• Convert irregular time data (or events) into regular sampled
• Resample (in time) and convert to numeric categorical/symbolic variables
2. Dynamic preprocessing (optional): convert dynamic data into static pattern
• Tapped delay line (time shift), fixed
• Apply dynamic filter (time convolution), adjustable
• Use transformed domain
3. Apply machine learning
• Static (standard) if dynamic processing is available
• Dynamic models: recurrent neural networks and variants
H2O Time Series 2
3. Machine learning with time series
Conversion data to regular time series
• Define convenient sampling rate or period (time resolution), according to
problem need and variability (analyze autocorrelation)
• Regularize time series:
• Convert irregular time data (or events) into regular sampled
• Complete when needed
• Convert categorical/symbolic into numeric (optional)
• Reshape :
• Under sampling: aggregation
• Over sampling: smoothing
H2O Time Series 3
4. Machine learning with time series
Dynamic preprocessing
• Needed for static machine learning model to deal with past information
• Simple processing may be convenient: smoothing, differentiating, etc.
• Other transformation may be used (non linear like log, etc.)
• Complex dynamic processing (array):
• Time domain:
• Tapped delay line: time shifts (past information over a window)
• Dynamic filter array: time convolution (may be adjustable)
• Transformed domain (over window): FFT, wavelet transform, etc.
• Result: each time series variable converted in array suitable for input at the
machine learning process
H2O Time Series 4
5. Machine learning with time series
Dynamic processing array
H2O Time Series 5
H1(z) (𝑥 ∗ ℎ1)[𝑖]
𝑥[𝑖]
H2(z)
H3(z)
(𝑥 ∗ ℎ2)[𝑖]
(𝑥 ∗ ℎ3)[𝑖]
Tapped delay line
Time shift
Dynamic filters array
Time convolution
z -1
z -1
z -1
𝑥[𝑖]
𝑥[𝑖 − 1]
𝑥[𝑖]
𝑥[𝑖 − 2]
𝑥[𝑖 − 3]
6. Machine learning with time series
Application of machine learning
• Style of learning:
• Long term for fixed model: single process off-line training, then use with no adjustment
• Short term for (time) variant model: initial off-line training (optional) followed by use
with on-line training (adaptive)
• Type of machine learning model:
• Static: output is function of present input only, dynamic aspect of times series must be
captured by dynamic preprocessing. Most machine learning models.
• Dynamic: output is function of actual input and past history. Recurrent neural networks
(RNN) and variants (Long short-term memory neural network, etc.)
H2O Time Series 6
7. Machine learning with time series
Machine Learning with dynamic preprocessing
Machine Learning with dynamic system (like recurrent neural network)
H2O Time Series 7
Dynamic
processing
Static
Machine
Learning
Recurrent
Neural
Network
Dynamic
Processing
(optional)
𝑣[i]
𝑥[i]
𝑦[𝑖]𝑥[𝑖]
𝑥[𝑖] 𝑦[𝑖]
8. Machine learning with time series
Use cases
• Prediction or forecasting:
• Given a time series (and optional inputs), produce expected future value(s):
𝑥 −∞: 𝑖 , 𝑣 −∞: 𝑖 ⟹ 𝑥 𝑖 + 𝑑 , 𝑑 ≥ 1
• Examples:
• Simple one-step forecasting (single variable)
𝑥 −∞: 𝑖 ⟹ 𝑥 𝑖 + 1
• Soft sensor: target values are known only at training
𝑣 −∞: 𝑖 ⟹ 𝑥 𝑖
• Anomaly detector: detection of unexpected values
𝑥 −∞: 𝑖 , 𝑣 −∞: 𝑖 ⟹ ε 𝑖 = 𝑥 𝑖 − 𝑥 𝑖
H2O Time Series 8
9. Machine learning with time series
Use cases
• Classification:
• Given a time series with a known classification (supervised)
𝑥 −∞: 𝑖 , 𝑐 −∞: 𝑖 − 1 ⟹ 𝑐 𝑖 , training phase
𝑥 −∞: 𝑖 ⟹ 𝑐 𝑖 , application phase
• Given a time series with unknown classification (unsupervised)
𝑥 −∞: 𝑖 ⟹ 𝑐 𝑖
H2O Time Series 9
10. Machine learning with time series
Use cases
• Control or Decision Making
• For a dynamic system H with input x, output y:
𝑥 −∞: 𝑖
𝐻
𝑦 𝑖
• Define a controller G to produce suitable input into H:
H2O Time Series 10
𝑥[𝑖]
H
𝑦[𝑖]
G
11. Machine learning with time series
Control or Decision Making
1. Reference problem
Achieve or follow a desired output (target or reference): 𝑦
• Define synthetic model G such that 𝐺𝜊𝐻 𝑦 = 𝑦:
{ 𝑦 −∞: 𝑖 , 𝑦 −∞: 𝑖 − 1 }
𝐺
𝑥 −∞: 𝑖
𝐻
𝑦 𝑖 | 𝑦 𝑖 + 𝑑 = 𝑦 𝑖 + 𝑑 , 𝑑 ≥ 1
• Train 𝐺, may use other model to predict H (predictive control)
H2O Time Series 11
𝑥[𝑖]
H
𝑦[𝑖]
G
𝑦 −∞: 𝑖 − 1
𝑦 −∞: 𝑖
12. Machine learning with time series
Control or Decision Making
2. Optimization problem
Optimize (min or max) an objective function the output: 𝑞 = 𝑄(𝑦)
• Define synthetic model G such that 𝐺 ∘ 𝐻 ∘ 𝑄 𝑑 = 𝑞 𝑖 + 𝑑 be optimal
{𝑞 −∞: 𝑖 − 1 }
𝐺
𝑥 −∞: 𝑖
𝐻
𝑦 −∞: 𝑖
𝑄
𝑞 𝑖 | 𝑀𝑎𝑥 𝐺 𝑞[𝑖 + 𝑑] (or Min)
• Train 𝐺, may use other model to predict H and Q
H2O Time Series 12
𝑥[𝑖]
H
𝑦[𝑖]
G𝑞 −∞: 𝑖 − 1 Q
𝑞[𝑖]