This document proposes combining the Kalman filter and ADWIN algorithm (Adaptive Windowing algorithm) to create an algorithm called K-ADWIN for learning from data streams. The Kalman filter is used as an estimator to estimate statistics from data streams, while ADWIN acts as a change detector to detect changes in the data distribution over time. This allows K-ADWIN to adaptively update its learning window size based on the detected changes, providing more accurate estimations that account for concept drift in non-stationary data streams. The document outlines the key components of K-ADWIN and provides experimental validation of its effectiveness.
Direct tall-and-skinny QR factorizations in MapReduce architecturesDavid Gleich
This document discusses tall-and-skinny QR factorizations in MapReduce. It begins by reviewing the basics of QR factorization and how it can be used for regression. It then describes how tall-and-skinny matrices arise in applications like regression with many samples, iterative methods, and model reduction. The document proposes using MapReduce to compute the QR factorization of very large tall-and-skinny matrices from applications like dynamic mode decomposition on large-scale simulations. It provides a brief overview of MapReduce and gives examples of computing variance in MapReduce.
This document summarizes a presentation on implementing a fast adaptive Kalman filter algorithm for speech enhancement. The presentation was given by 5 students from the Department of Electrical and Electronics Engineering at P.B.R Visvodaya Institute of Technology and Science. It provides background on speech processing and enhancement, an overview of linear predictive coding and Kalman filtering techniques, and details on how the Kalman filter was implemented to model and reconstruct a speech signal as an autoregressive process. Key steps in the Kalman filter implementation included expressing the speech signal in state space form and running the filter in a looping method to iteratively estimate and update the state over multiple iterations.
Kalman filter - Applications in Image processingRavi Teja
The document discusses the Kalman filter, an algorithm used to estimate the state of a linear dynamic system from a series of noisy measurements. It describes how the Kalman filter uses a predictor-corrector approach with time update and measurement update equations to estimate the true state. The filter is applied to image processing to reduce noise by modeling the image as an autoregressive process and using the Kalman filter estimates. Extensions to nonlinear and complex systems using the extended and complex Kalman filters are also covered.
Troubleshooting Coherent Optical Communication SystemsCPqD
1) The document discusses troubleshooting of coherent optical communication systems. It covers topics such as market trends driving higher data rates, challenges in coherent measurements, and typical tests and impairments in coherent transmission systems.
2) The document outlines techniques used to maximize transmission capacity within the physical layer, including higher order modulation formats, time-domain pulse shaping, and polarization division multiplexing. It also discusses requirements for test instruments to test these advanced modulation schemes.
3) The presentation provides an overview of using arbitrary waveform generators to emulate optical distortions in the electrical domain, allowing for deterministic and precise testing with complex impairments like phase noise and polarization mode dispersion.
This document outlines the contents and topics that will be covered in a course on digital signal processing and digital filters taught by Professor A.G. Constantinides. The course will cover introductions to digital signal processing, multirate techniques and wavelets, classical and modern spectrum estimation methods, adaptive signal processing, and applications of digital signal processing. It also provides an overview of the key concepts and techniques that will be discussed regarding digital filters.
Data Con LA 2022 - Pre - recorded - Quantum Computing, The next new technolog...Data Con LA
Mark Jackson, Quantum Evangelist at Quantinuum
Emerging Tech
Quantum computing is rapidly becoming commercially feasible. Many tech giants - Google, IBM, Honeywell, and Microsoft - are spending billions to far outpace Moore's Law. Last year achieved the major milestone of Quantum Supremacy where it was shown that a quantum computer could greatly outperform a classical computer. Quantum computing offers the promise of solving problems which would be impossible for a classical computer including optimization, anomaly detection, and material design. It also allows unhackable communication. In this presentation I will summarize what quantum computing is and why it is so important. I will sketch the landscape of the field including the hardware, software, and major customers at present. The tool most critical for data analysis - quantum machine learning - will be explained, along with the type of applications it is best suited for. Finally I will explain how you can take the first steps into leveraging quantum computing for your enterprise's benefit. -What is quantum computing -Who are the major players in the field -What is quantum machine learning and what types of problems can it address -How your company can take advantage of this
Optical Signal Property Synthesis at Runtime – An new approach for coherent t...CPqD
This document describes a new approach for coherent optical receiver stress testing using digital signal processing (DSP) in an arbitrary waveform generator (AWG)-based optical signal synthesizer. The new approach allows for real-time generation and control of optical signal properties like phase noise, polarization, and polarization mode dispersion that overcome limitations of pre-calculating entire waveforms. This provides a flexible test platform for systematically evaluating receivers under worst-case impairment conditions.
FR1.L09 - PREDICTIVE QUANTIZATION OF DECHIRPED SPOTLIGHT-MODE SAR RAW DATA IN...grssieee
This document presents methods for predictive quantization of dechirped spotlight-mode synthetic aperture radar (SAR) raw data in the transform domain. It discusses previous work on SAR data compression, analyzes the characteristics of spotlight SAR data in the inverse discrete Fourier transform (IDFT) domain, and proposes three predictive encoding schemes - transform domain block predictive quantization (TD-BPQ), transform domain block predictive vector quantization (TD-BPVQ), and predictive trellis coded quantization (TD-PTCQ) - to take advantage of correlations in the transformed data. Numerical results on an example dataset show SNR improvements of up to 6 dB compared to baseline block adaptive quantization.
Direct tall-and-skinny QR factorizations in MapReduce architecturesDavid Gleich
This document discusses tall-and-skinny QR factorizations in MapReduce. It begins by reviewing the basics of QR factorization and how it can be used for regression. It then describes how tall-and-skinny matrices arise in applications like regression with many samples, iterative methods, and model reduction. The document proposes using MapReduce to compute the QR factorization of very large tall-and-skinny matrices from applications like dynamic mode decomposition on large-scale simulations. It provides a brief overview of MapReduce and gives examples of computing variance in MapReduce.
This document summarizes a presentation on implementing a fast adaptive Kalman filter algorithm for speech enhancement. The presentation was given by 5 students from the Department of Electrical and Electronics Engineering at P.B.R Visvodaya Institute of Technology and Science. It provides background on speech processing and enhancement, an overview of linear predictive coding and Kalman filtering techniques, and details on how the Kalman filter was implemented to model and reconstruct a speech signal as an autoregressive process. Key steps in the Kalman filter implementation included expressing the speech signal in state space form and running the filter in a looping method to iteratively estimate and update the state over multiple iterations.
Kalman filter - Applications in Image processingRavi Teja
The document discusses the Kalman filter, an algorithm used to estimate the state of a linear dynamic system from a series of noisy measurements. It describes how the Kalman filter uses a predictor-corrector approach with time update and measurement update equations to estimate the true state. The filter is applied to image processing to reduce noise by modeling the image as an autoregressive process and using the Kalman filter estimates. Extensions to nonlinear and complex systems using the extended and complex Kalman filters are also covered.
Troubleshooting Coherent Optical Communication SystemsCPqD
1) The document discusses troubleshooting of coherent optical communication systems. It covers topics such as market trends driving higher data rates, challenges in coherent measurements, and typical tests and impairments in coherent transmission systems.
2) The document outlines techniques used to maximize transmission capacity within the physical layer, including higher order modulation formats, time-domain pulse shaping, and polarization division multiplexing. It also discusses requirements for test instruments to test these advanced modulation schemes.
3) The presentation provides an overview of using arbitrary waveform generators to emulate optical distortions in the electrical domain, allowing for deterministic and precise testing with complex impairments like phase noise and polarization mode dispersion.
This document outlines the contents and topics that will be covered in a course on digital signal processing and digital filters taught by Professor A.G. Constantinides. The course will cover introductions to digital signal processing, multirate techniques and wavelets, classical and modern spectrum estimation methods, adaptive signal processing, and applications of digital signal processing. It also provides an overview of the key concepts and techniques that will be discussed regarding digital filters.
Data Con LA 2022 - Pre - recorded - Quantum Computing, The next new technolog...Data Con LA
Mark Jackson, Quantum Evangelist at Quantinuum
Emerging Tech
Quantum computing is rapidly becoming commercially feasible. Many tech giants - Google, IBM, Honeywell, and Microsoft - are spending billions to far outpace Moore's Law. Last year achieved the major milestone of Quantum Supremacy where it was shown that a quantum computer could greatly outperform a classical computer. Quantum computing offers the promise of solving problems which would be impossible for a classical computer including optimization, anomaly detection, and material design. It also allows unhackable communication. In this presentation I will summarize what quantum computing is and why it is so important. I will sketch the landscape of the field including the hardware, software, and major customers at present. The tool most critical for data analysis - quantum machine learning - will be explained, along with the type of applications it is best suited for. Finally I will explain how you can take the first steps into leveraging quantum computing for your enterprise's benefit. -What is quantum computing -Who are the major players in the field -What is quantum machine learning and what types of problems can it address -How your company can take advantage of this
Optical Signal Property Synthesis at Runtime – An new approach for coherent t...CPqD
This document describes a new approach for coherent optical receiver stress testing using digital signal processing (DSP) in an arbitrary waveform generator (AWG)-based optical signal synthesizer. The new approach allows for real-time generation and control of optical signal properties like phase noise, polarization, and polarization mode dispersion that overcome limitations of pre-calculating entire waveforms. This provides a flexible test platform for systematically evaluating receivers under worst-case impairment conditions.
FR1.L09 - PREDICTIVE QUANTIZATION OF DECHIRPED SPOTLIGHT-MODE SAR RAW DATA IN...grssieee
This document presents methods for predictive quantization of dechirped spotlight-mode synthetic aperture radar (SAR) raw data in the transform domain. It discusses previous work on SAR data compression, analyzes the characteristics of spotlight SAR data in the inverse discrete Fourier transform (IDFT) domain, and proposes three predictive encoding schemes - transform domain block predictive quantization (TD-BPQ), transform domain block predictive vector quantization (TD-BPVQ), and predictive trellis coded quantization (TD-PTCQ) - to take advantage of correlations in the transformed data. Numerical results on an example dataset show SNR improvements of up to 6 dB compared to baseline block adaptive quantization.
Artificial intelligence and data stream miningAlbert Bifet
Big Data and Artificial Intelligence have the potential to
fundamentally shift the way we interact with our surroundings. The
challenge of deriving insights from data streams has been recognized
as one of the most exciting and key opportunities for both academia
and industry. Advanced analysis of big data streams from sensors and
devices is bound to become a key area of artificial intelligence
research as the number of applications requiring such processing
increases. Dealing with the evolution over time of such data streams,
i.e., with concepts that drift or change completely, is one of the
core issues in stream mining. In this talk, I will present an overview
of data stream mining, industrial applications, open source tools, and
current challenges of data stream mining.
MOA is a framework for online machine learning from data streams. It includes algorithms for classification, regression, clustering and frequent pattern mining that can incorporate data and update models on the fly. MOA is closely related to WEKA and includes tools for evaluating streaming algorithms on data from sensors and IoT devices. It provides an environment for designing and running experiments on streaming machine learning algorithms at massive scales.
Mining Big Data Streams with APACHE SAMOAAlbert Bifet
In this talk, we present Apache SAMOA, an open-source platform for
mining big data streams with Apache Flink, Storm and Samza. Real time analytics is
becoming the fastest and most efficient way to obtain useful knowledge
from what is happening now, allowing organizations to react quickly
when problems appear or to detect new trends helping to improve their
performance. Apache SAMOA includes algorithms for the most common
machine learning tasks such as classification and clustering. It
provides a pluggable architecture that allows it to run on Apache
Flink, but also with other several distributed stream processing
engines such as Storm and Samza.
Efficient Online Evaluation of Big Data Stream ClassifiersAlbert Bifet
The evaluation of classifiers in data streams is fundamental so that poorly-performing models can be identified, and either improved or replaced by better-performing models. This is an increasingly relevant and important task as stream data is generated from more sources, in real-time, in large quantities, and is now considered the largest source of big data. Both researchers and practitioners need to be able to effectively evaluate the performance of the methods they employ. However, there are major challenges for evaluation in a stream. Instances arriving in a data stream are usually time-dependent, and the underlying concept that they represent may evolve over time. Furthermore, the massive quantity of data also tends to exacerbate issues such as class imbalance. Current frameworks for evaluating streaming and online algorithms are able to give predictions in real-time, but as they use a prequential setting, they build only one model, and are thus not able to compute the statistical significance of results in real-time. In this paper we propose a new evaluation methodology for big data streams. This methodology addresses unbalanced data streams, data where change occurs on different time scales, and the question of how to split the data between training and testing, over multiple models.
Apache Samoa: Mining Big Data Streams with Apache FlinkAlbert Bifet
1) Apache SAMOA is a platform for mining big data streams in real-time that provides algorithms, libraries, and frameworks.
2) It allows researchers to develop and compare stream mining algorithms and practitioners to easily apply state-of-the-art algorithms to problems like sentiment analysis, spam detection, and recommendations.
3) A key challenge addressed by SAMOA is how to perform distributed stream mining on high-volume, high-velocity data streams at low latency using approaches like Apache Flink that can scale to handle large, fast data.
Data science involves extracting insights from large volumes of data. It is an interdisciplinary field that uses techniques from statistics, machine learning, and other domains. The document provides examples of classification algorithms like k-nearest neighbors, naive Bayes, and perceptrons that are commonly used in data science to build models for tasks like spam filtering or sentiment analysis. It also discusses clustering, frequent pattern mining, and other machine learning concepts.
This document provides an introduction to big data and MapReduce frameworks. It discusses:
- What big data is and examples of large datasets.
- An overview of MapReduce, including how it allows programmers to break problems into parallelizable map and reduce tasks.
- Details of how MapReduce frameworks like Apache Hadoop work, including distributed processing, fault tolerance, and the roles of mappers, reducers, and other components.
The document discusses data stream classification and algorithms for handling data streams. It begins with an introduction to data stream characteristics and challenges. It then discusses approximation algorithms for data streams, including maintaining statistics over sliding windows. Classification algorithms for data streams discussed include Naive Bayes classifiers, perceptrons, and Hoeffding trees, which are decision trees adapted for data streams using the Hoeffding bound inequality to determine the optimal split attribute.
The document discusses real-time big data management and Apache Flink. It provides an overview of Apache Flink, including its architecture, components, and APIs for batch and streaming data processing. It also provides examples of word count programs in Java, Scala, and Java 8 that demonstrate how to write Flink programs for batch and streaming data.
1. Real-time analytics of social networks can help companies detect new business opportunities by understanding customer needs and reactions in real-time.
2. MOA and SAMOA are frameworks for analyzing massive online and distributed data streams. MOA deals with evolving data streams using online learning algorithms. SAMOA provides a programming model for distributed, real-time machine learning on data streams.
3. Both tools allow companies to gain insights from social network and other real-time data to understand customers and react to opportunities.
Multi-label Classification with Meta-labelsAlbert Bifet
The area of multi-label classification has rapidly developed in recent years. It has become widely known that the baseline binary relevance approach suffers from class imbalance and a restricted hypothesis space that negatively affects its predictive performance, and can easily be outperformed by methods which learn labels together. A number of methods have grown around the label powerset approach, which models label combinations together as class values in a multi-class problem. We describe the label-powerset-based solutions under a general framework of \emph{meta-labels}. We provide theoretical justification for this framework which has been lacking, by viewing meta-labels as a hidden layer in an artificial neural network. We explain how meta-labels essentially allow a random projection into a space where non-linearities can easily be tackled with established linear learning algorithms. The proposed framework enables comparison and combination of related approaches to different multi-label problems. Indeed, we present a novel model in the framework and evaluate it empirically against several high-performing methods, with respect to predictive performance and scalability, on a number of datasets and evaluation metrics. Our deployment of an ensemble of meta-label classifiers obtains competitive accuracy for a fraction of the computation required by the current meta-label methods for multi-label classification.
Pitfalls in benchmarking data stream classification and how to avoid themAlbert Bifet
This document discusses pitfalls in benchmarking data stream classification and proposes ways to avoid them. It analyzes the electricity market dataset, a popular benchmark, and finds that it exhibits temporal dependence that favors classifiers that simply predict the previous value. It introduces new evaluation metrics like kappa plus that account for temporal dependence by comparing to a "no change" classifier. It also proposes a temporally aware classifier called SWT that incorporates previous labels into its predictions. Experiments on electricity and forest cover datasets show SWT and the new metrics better capture classifier performance on temporally dependent streaming data.
STRIP: stream learning of influence probabilities.Albert Bifet
This document presents a method called STRIP (Streaming Learning of Influence Probabilities) for learning influence probabilities between users in a social network from a streaming log of propagations. It describes three solutions: (1) storing the whole social graph in memory, (2) using min-wise independent hashing to estimate probabilities while using sublinear space, and (3) estimating probabilities only for the most active users to be more space efficient. Experimental results on a Twitter dataset showed these solutions provided good approximations while using reasonable memory and processing time.
Efficient Data Stream Classification via Probabilistic Adaptive WindowsAlbert Bifet
This document discusses efficient data stream classification using probabilistic adaptive windows. It introduces the concept of data streams which have potentially infinite sequences of high-speed data that must be processed in real-time with limited memory. It then describes the probabilistic approximate window (PAW) algorithm, which maintains a sample of data instances in logarithmic memory by giving greater weight to newer instances. The document evaluates several data stream classification methods on real and synthetic data streams and finds that k-nearest neighbors with PAW has higher accuracy and lower memory usage than other methods.
Big Data is a new term used to identify datasets that we can not manage with current methodologies or data mining software tools due to their large size and complexity. Big Data mining is the capability of extracting useful information from these large datasets or streams of data. New mining techniques are necessary due to the volume, variability, and velocity, of such data.
Streaming data analysis in real time is becoming the fastest and most efficient way to obtain useful knowledge from what is happening now, allowing organizations to react quickly when problems appear or to detect new trends helping to improve their performance. Evolving data streams are contributing to the growth of data created over the last few years. We are creating the same quantity of data every two days, as we created from the dawn of time up until 2003. Evolving data streams methods are becoming a low-cost, green methodology for real time online prediction and analysis. We discuss the current and future trends of mining evolving data streams, and the challenges that the field will have to overcome during the next years.
Mining Frequent Closed Graphs on Evolving Data StreamsAlbert Bifet
Graph mining is a challenging task by itself, and even more so when processing data streams which evolve in real-time. Data stream mining faces hard constraints regarding time and space for processing, and also needs to provide for concept drift detection. In this talk we present a framework for studying graph pattern mining on time-varying streams and large datasets.
The document outlines a tutorial on handling concept drift in machine learning. It discusses the challenges of concept drift when applying supervised learning algorithms to streaming data where the underlying data distribution changes over time. The tutorial aims to provide an integrated view of adaptive learning methods and how they can handle concept drift. It covers topics such as the problem of concept drift, techniques for handling drift, evaluating adaptive learning approaches, and applications that experience concept drift.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Artificial intelligence and data stream miningAlbert Bifet
Big Data and Artificial Intelligence have the potential to
fundamentally shift the way we interact with our surroundings. The
challenge of deriving insights from data streams has been recognized
as one of the most exciting and key opportunities for both academia
and industry. Advanced analysis of big data streams from sensors and
devices is bound to become a key area of artificial intelligence
research as the number of applications requiring such processing
increases. Dealing with the evolution over time of such data streams,
i.e., with concepts that drift or change completely, is one of the
core issues in stream mining. In this talk, I will present an overview
of data stream mining, industrial applications, open source tools, and
current challenges of data stream mining.
MOA is a framework for online machine learning from data streams. It includes algorithms for classification, regression, clustering and frequent pattern mining that can incorporate data and update models on the fly. MOA is closely related to WEKA and includes tools for evaluating streaming algorithms on data from sensors and IoT devices. It provides an environment for designing and running experiments on streaming machine learning algorithms at massive scales.
Mining Big Data Streams with APACHE SAMOAAlbert Bifet
In this talk, we present Apache SAMOA, an open-source platform for
mining big data streams with Apache Flink, Storm and Samza. Real time analytics is
becoming the fastest and most efficient way to obtain useful knowledge
from what is happening now, allowing organizations to react quickly
when problems appear or to detect new trends helping to improve their
performance. Apache SAMOA includes algorithms for the most common
machine learning tasks such as classification and clustering. It
provides a pluggable architecture that allows it to run on Apache
Flink, but also with other several distributed stream processing
engines such as Storm and Samza.
Efficient Online Evaluation of Big Data Stream ClassifiersAlbert Bifet
The evaluation of classifiers in data streams is fundamental so that poorly-performing models can be identified, and either improved or replaced by better-performing models. This is an increasingly relevant and important task as stream data is generated from more sources, in real-time, in large quantities, and is now considered the largest source of big data. Both researchers and practitioners need to be able to effectively evaluate the performance of the methods they employ. However, there are major challenges for evaluation in a stream. Instances arriving in a data stream are usually time-dependent, and the underlying concept that they represent may evolve over time. Furthermore, the massive quantity of data also tends to exacerbate issues such as class imbalance. Current frameworks for evaluating streaming and online algorithms are able to give predictions in real-time, but as they use a prequential setting, they build only one model, and are thus not able to compute the statistical significance of results in real-time. In this paper we propose a new evaluation methodology for big data streams. This methodology addresses unbalanced data streams, data where change occurs on different time scales, and the question of how to split the data between training and testing, over multiple models.
Apache Samoa: Mining Big Data Streams with Apache FlinkAlbert Bifet
1) Apache SAMOA is a platform for mining big data streams in real-time that provides algorithms, libraries, and frameworks.
2) It allows researchers to develop and compare stream mining algorithms and practitioners to easily apply state-of-the-art algorithms to problems like sentiment analysis, spam detection, and recommendations.
3) A key challenge addressed by SAMOA is how to perform distributed stream mining on high-volume, high-velocity data streams at low latency using approaches like Apache Flink that can scale to handle large, fast data.
Data science involves extracting insights from large volumes of data. It is an interdisciplinary field that uses techniques from statistics, machine learning, and other domains. The document provides examples of classification algorithms like k-nearest neighbors, naive Bayes, and perceptrons that are commonly used in data science to build models for tasks like spam filtering or sentiment analysis. It also discusses clustering, frequent pattern mining, and other machine learning concepts.
This document provides an introduction to big data and MapReduce frameworks. It discusses:
- What big data is and examples of large datasets.
- An overview of MapReduce, including how it allows programmers to break problems into parallelizable map and reduce tasks.
- Details of how MapReduce frameworks like Apache Hadoop work, including distributed processing, fault tolerance, and the roles of mappers, reducers, and other components.
The document discusses data stream classification and algorithms for handling data streams. It begins with an introduction to data stream characteristics and challenges. It then discusses approximation algorithms for data streams, including maintaining statistics over sliding windows. Classification algorithms for data streams discussed include Naive Bayes classifiers, perceptrons, and Hoeffding trees, which are decision trees adapted for data streams using the Hoeffding bound inequality to determine the optimal split attribute.
The document discusses real-time big data management and Apache Flink. It provides an overview of Apache Flink, including its architecture, components, and APIs for batch and streaming data processing. It also provides examples of word count programs in Java, Scala, and Java 8 that demonstrate how to write Flink programs for batch and streaming data.
1. Real-time analytics of social networks can help companies detect new business opportunities by understanding customer needs and reactions in real-time.
2. MOA and SAMOA are frameworks for analyzing massive online and distributed data streams. MOA deals with evolving data streams using online learning algorithms. SAMOA provides a programming model for distributed, real-time machine learning on data streams.
3. Both tools allow companies to gain insights from social network and other real-time data to understand customers and react to opportunities.
Multi-label Classification with Meta-labelsAlbert Bifet
The area of multi-label classification has rapidly developed in recent years. It has become widely known that the baseline binary relevance approach suffers from class imbalance and a restricted hypothesis space that negatively affects its predictive performance, and can easily be outperformed by methods which learn labels together. A number of methods have grown around the label powerset approach, which models label combinations together as class values in a multi-class problem. We describe the label-powerset-based solutions under a general framework of \emph{meta-labels}. We provide theoretical justification for this framework which has been lacking, by viewing meta-labels as a hidden layer in an artificial neural network. We explain how meta-labels essentially allow a random projection into a space where non-linearities can easily be tackled with established linear learning algorithms. The proposed framework enables comparison and combination of related approaches to different multi-label problems. Indeed, we present a novel model in the framework and evaluate it empirically against several high-performing methods, with respect to predictive performance and scalability, on a number of datasets and evaluation metrics. Our deployment of an ensemble of meta-label classifiers obtains competitive accuracy for a fraction of the computation required by the current meta-label methods for multi-label classification.
Pitfalls in benchmarking data stream classification and how to avoid themAlbert Bifet
This document discusses pitfalls in benchmarking data stream classification and proposes ways to avoid them. It analyzes the electricity market dataset, a popular benchmark, and finds that it exhibits temporal dependence that favors classifiers that simply predict the previous value. It introduces new evaluation metrics like kappa plus that account for temporal dependence by comparing to a "no change" classifier. It also proposes a temporally aware classifier called SWT that incorporates previous labels into its predictions. Experiments on electricity and forest cover datasets show SWT and the new metrics better capture classifier performance on temporally dependent streaming data.
STRIP: stream learning of influence probabilities.Albert Bifet
This document presents a method called STRIP (Streaming Learning of Influence Probabilities) for learning influence probabilities between users in a social network from a streaming log of propagations. It describes three solutions: (1) storing the whole social graph in memory, (2) using min-wise independent hashing to estimate probabilities while using sublinear space, and (3) estimating probabilities only for the most active users to be more space efficient. Experimental results on a Twitter dataset showed these solutions provided good approximations while using reasonable memory and processing time.
Efficient Data Stream Classification via Probabilistic Adaptive WindowsAlbert Bifet
This document discusses efficient data stream classification using probabilistic adaptive windows. It introduces the concept of data streams which have potentially infinite sequences of high-speed data that must be processed in real-time with limited memory. It then describes the probabilistic approximate window (PAW) algorithm, which maintains a sample of data instances in logarithmic memory by giving greater weight to newer instances. The document evaluates several data stream classification methods on real and synthetic data streams and finds that k-nearest neighbors with PAW has higher accuracy and lower memory usage than other methods.
Big Data is a new term used to identify datasets that we can not manage with current methodologies or data mining software tools due to their large size and complexity. Big Data mining is the capability of extracting useful information from these large datasets or streams of data. New mining techniques are necessary due to the volume, variability, and velocity, of such data.
Streaming data analysis in real time is becoming the fastest and most efficient way to obtain useful knowledge from what is happening now, allowing organizations to react quickly when problems appear or to detect new trends helping to improve their performance. Evolving data streams are contributing to the growth of data created over the last few years. We are creating the same quantity of data every two days, as we created from the dawn of time up until 2003. Evolving data streams methods are becoming a low-cost, green methodology for real time online prediction and analysis. We discuss the current and future trends of mining evolving data streams, and the challenges that the field will have to overcome during the next years.
Mining Frequent Closed Graphs on Evolving Data StreamsAlbert Bifet
Graph mining is a challenging task by itself, and even more so when processing data streams which evolve in real-time. Data stream mining faces hard constraints regarding time and space for processing, and also needs to provide for concept drift detection. In this talk we present a framework for studying graph pattern mining on time-varying streams and large datasets.
The document outlines a tutorial on handling concept drift in machine learning. It discusses the challenges of concept drift when applying supervised learning algorithms to streaming data where the underlying data distribution changes over time. The tutorial aims to provide an integrated view of adaptive learning methods and how they can handle concept drift. It covers topics such as the problem of concept drift, techniques for handling drift, evaluating adaptive learning approaches, and applications that experience concept drift.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!SOFTTECHHUB
As the digital landscape continually evolves, operating systems play a critical role in shaping user experiences and productivity. The launch of Nitrux Linux 3.5.0 marks a significant milestone, offering a robust alternative to traditional systems such as Windows 11. This article delves into the essence of Nitrux Linux 3.5.0, exploring its unique features, advantages, and how it stands as a compelling choice for both casual users and tech enthusiasts.
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
Introducing Milvus Lite: Easy-to-Install, Easy-to-Use vector database for you...Zilliz
Join us to introduce Milvus Lite, a vector database that can run on notebooks and laptops, share the same API with Milvus, and integrate with every popular GenAI framework. This webinar is perfect for developers seeking easy-to-use, well-integrated vector databases for their GenAI apps.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
Presentation of the OECD Artificial Intelligence Review of Germany
Kalman Filters and Adaptive Windows for Learning in Data Streams
1. Kalman Filters and Adaptive Windows for Learning
in Data Streams
Albert Bifet Ricard Gavaldà
Universitat Politècnica de Catalunya
DS’06 Barcelona
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams Barcelona
DS’06 1 / 29
2. Outline
1 Introduction
2 The Kalman Filter and the CUSUM Test
3 The ADWIN Algorithm
4 General Framework
5 K-ADWIN
6 Experimental Validation of K-ADWIN
7 Conclusions
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams Barcelona
DS’06 2 / 29
3. Introduction
Introduction
Data Streams
Sequence potentially infinite
High amount of data:
sublinear space
High Speed of arrival:
small constant time per example
Estimation and prediction
Distribution and concept drift
K-ADWIN : Combination
Kalman filter
ADWIN : Adaptive window of recently seen data items.
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams Barcelona
DS’06 3 / 29
4. Introduction
Introduction
Problem
Given an input sequence x1 , x2 , . . . , xt , . . . we want to output at instant t
a prediction xt+1 minimizing prediction error:
|xt+1 − xt+1 |
considering distribution changes overtime.
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams Barcelona
DS’06 4 / 29
5. Introduction
Time Change Detectors and Predictors: A General
Framework
Estimation
-
xt
-
Estimator
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams Barcelona
DS’06 5 / 29
6. Introduction
Time Change Detectors and Predictors: A General
Framework
Estimation
-
xt
-
Estimator Alarm
- -
Change Detect.
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams Barcelona
DS’06 5 / 29
7. Introduction
Time Change Detectors and Predictors: A General
Framework
Estimation
-
xt
-
Estimator Alarm
- -
Change Detect.
6
6
?
-
Memory
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams Barcelona
DS’06 5 / 29
8. Introduction
Introduction
Our particular proposal: K-ADWIN
Our generic proposal:
Kalman filter as estimator
Use change detector
Use ADWIN as change
Use memory
detector with memory [BG06]
Application
Estimate statistics from data streams
In Data Mining Algorithms based on counters, replace them for
estimators.
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams Barcelona
DS’06 6 / 29
9. Introduction
Data Mining Algorithms with Concept Drift
No Concept Drift Concept drift
DM Algorithm
DM Algorithm
input output input output
-
Counter5 - - -
Counter4 Static Model
Counter3 6
Counter2
Counter1 -
Change Detect.
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams Barcelona
DS’06 7 / 29
10. Introduction
Data Mining Algorithms with Concept Drift
No Concept Drift Concept Drift
DM Algorithm DM Algorithm
input output input output
-
Counter5 - - Estimator5 -
Counter4 Estimator4
Counter3 Estimator3
Counter2 Estimator2
Counter1 Estimator1
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams Barcelona
DS’06 7 / 29
11. The Kalman Filter and the CUSUM Test
The Kalman Filter
Optimal recursive algorithm
Minimum mean-square error estimator
Estimate the state x ∈ n of a discrete-time controlled process
xk = Axk −1 + Buk + wk −1
with a measurement z ∈ m that is
Zk = Hxk + vk .
The random variables wk and vk represent the process and
measurement noise (respectively). They are assumed to be
independent (of each other), white, and with normal probability
distributions
p(w) ∼ N(0, Q) p(v ) ∼ N(0, R).
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams Barcelona
DS’06 8 / 29
12. The Kalman Filter and the CUSUM Test
The Kalman Filter
The difference equation of our discrete-time controlled process is
Kk = Pk −1 /(Pk −1 + R)
Xk = Xk −1 + Kk (zk − Xk −1 )
Pk = Pk (1 − Kk ) + Q
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams Barcelona
DS’06 9 / 29
13. The Kalman Filter and the CUSUM Test
The Kalman Filter
The difference equation of our discrete-time controlled process is
Kk = Pk −1 /(Pk −1 + R)
Xk = Xk −1 + Kk (zk − Xk −1 )
Pk = Pk (1 − Kk ) + Q
The performance of the Kalman filter depends on the accuracy of the
a-priori assumptions:
linearity of the difference stochastic equation
estimation of covariances Q and R, assumed to be fixed, known,
and follow normal distributions with zero mean.
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams Barcelona
DS’06 9 / 29
14. The Kalman Filter and the CUSUM Test
The CUSUM Test
The cumulative sum (CUSUM algorithm),is a change detection
algorithm that gives an alarm when the mean of the input data is
significantly different from zero.
The CUSUM test is memoryless, and its accuracy depends on the
choice of parameters υ and h. It is as follows:
g0 = 0, gt = max (0, gt−1 + t − υ)
if gt h then alarm and gt = 0
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 10 / 29
15. The ADWIN Algorithm
Algorithm ADWIN [BG06]
Example
W = 101010110111111
ADWIN: A DAPTIVE W INDOWING A LGORITHM
1 Initialize Window W
2 for each t 0
3 do W ← W ∪ {xt } (i.e., add xt to the head of W )
4 repeat Drop elements from the tail of W
5 until |ˆW0 − µW1 | ≥ c holds
µ ˆ
6 for every split of W into W = W0 · W1
7 Output µW
ˆ
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 11 / 29
16. The ADWIN Algorithm
Algorithm ADWIN [BG06]
Example
W = 101010110111111
W0 = 1 W1 = 01010110111111
ADWIN: A DAPTIVE W INDOWING A LGORITHM
1 Initialize Window W
2 for each t 0
3 do W ← W ∪ {xt } (i.e., add xt to the head of W )
4 repeat Drop elements from the tail of W
5 until |ˆW0 − µW1 | ≥ c holds
µ ˆ
6 for every split of W into W = W0 · W1
7 Output µW
ˆ
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 11 / 29
17. The ADWIN Algorithm
Algorithm ADWIN [BG06]
Example
W = 101010110111111
W0 = 10 W1 = 1010110111111
ADWIN: A DAPTIVE W INDOWING A LGORITHM
1 Initialize Window W
2 for each t 0
3 do W ← W ∪ {xt } (i.e., add xt to the head of W )
4 repeat Drop elements from the tail of W
5 until |ˆW0 − µW1 | ≥ c holds
µ ˆ
6 for every split of W into W = W0 · W1
7 Output µW
ˆ
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 11 / 29
18. The ADWIN Algorithm
Algorithm ADWIN [BG06]
Example
W = 101010110111111
W0 = 101 W1 = 010110111111
ADWIN: A DAPTIVE W INDOWING A LGORITHM
1 Initialize Window W
2 for each t 0
3 do W ← W ∪ {xt } (i.e., add xt to the head of W )
4 repeat Drop elements from the tail of W
5 until |ˆW0 − µW1 | ≥ c holds
µ ˆ
6 for every split of W into W = W0 · W1
7 Output µW
ˆ
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 11 / 29
19. The ADWIN Algorithm
Algorithm ADWIN [BG06]
Example
W = 101010110111111
W0 = 1010 W1 = 10110111111
ADWIN: A DAPTIVE W INDOWING A LGORITHM
1 Initialize Window W
2 for each t 0
3 do W ← W ∪ {xt } (i.e., add xt to the head of W )
4 repeat Drop elements from the tail of W
5 until |ˆW0 − µW1 | ≥ c holds
µ ˆ
6 for every split of W into W = W0 · W1
7 Output µW
ˆ
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 11 / 29
20. The ADWIN Algorithm
Algorithm ADWIN [BG06]
Example
W = 101010110111111
W0 = 10101 W1 = 0110111111
ADWIN: A DAPTIVE W INDOWING A LGORITHM
1 Initialize Window W
2 for each t 0
3 do W ← W ∪ {xt } (i.e., add xt to the head of W )
4 repeat Drop elements from the tail of W
5 until |ˆW0 − µW1 | ≥ c holds
µ ˆ
6 for every split of W into W = W0 · W1
7 Output µW
ˆ
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 11 / 29
21. The ADWIN Algorithm
Algorithm ADWIN [BG06]
Example
W = 101010110111111
W0 = 101010 W1 = 110111111
ADWIN: A DAPTIVE W INDOWING A LGORITHM
1 Initialize Window W
2 for each t 0
3 do W ← W ∪ {xt } (i.e., add xt to the head of W )
4 repeat Drop elements from the tail of W
5 until |ˆW0 − µW1 | ≥ c holds
µ ˆ
6 for every split of W into W = W0 · W1
7 Output µW
ˆ
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 11 / 29
22. The ADWIN Algorithm
Algorithm ADWIN [BG06]
Example
W = 101010110111111
W0 = 1010101 W1 = 10111111
ADWIN: A DAPTIVE W INDOWING A LGORITHM
1 Initialize Window W
2 for each t 0
3 do W ← W ∪ {xt } (i.e., add xt to the head of W )
4 repeat Drop elements from the tail of W
5 until |ˆW0 − µW1 | ≥ c holds
µ ˆ
6 for every split of W into W = W0 · W1
7 Output µW
ˆ
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 11 / 29
23. The ADWIN Algorithm
Algorithm ADWIN [BG06]
Example
W = 101010110111111
W0 = 10101011 W1 = 0111111
ADWIN: A DAPTIVE W INDOWING A LGORITHM
1 Initialize Window W
2 for each t 0
3 do W ← W ∪ {xt } (i.e., add xt to the head of W )
4 repeat Drop elements from the tail of W
5 until |ˆW0 − µW1 | ≥ c holds
µ ˆ
6 for every split of W into W = W0 · W1
7 Output µW
ˆ
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 11 / 29
24. The ADWIN Algorithm
Algorithm ADWIN [BG06]
Example
W = 101010110111111 |ˆW0 − µW1 | ≥
µ ˆ c : CHANGE DETECTED!
W0 = 101010110 W1 = 111111
ADWIN: A DAPTIVE W INDOWING A LGORITHM
1 Initialize Window W
2 for each t 0
3 do W ← W ∪ {xt } (i.e., add xt to the head of W )
4 repeat Drop elements from the tail of W
5 until |ˆW0 − µW1 | ≥ c holds
µ ˆ
6 for every split of W into W = W0 · W1
7 Output µW
ˆ
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 11 / 29
25. The ADWIN Algorithm
Algorithm ADWIN [BG06]
Example
W = 101010110111111 Drop elements from the tail of W
W0 = 101010110 W1 = 111111
ADWIN: A DAPTIVE W INDOWING A LGORITHM
1 Initialize Window W
2 for each t 0
3 do W ← W ∪ {xt } (i.e., add xt to the head of W )
4 repeat Drop elements from the tail of W
5 until |ˆW0 − µW1 | ≥ c holds
µ ˆ
6 for every split of W into W = W0 · W1
7 Output µW
ˆ
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 11 / 29
26. The ADWIN Algorithm
Algorithm ADWIN [BG06]
Example
W = 01010110111111 Drop elements from the tail of W
W0 = 101010110 W1 = 111111
ADWIN: A DAPTIVE W INDOWING A LGORITHM
1 Initialize Window W
2 for each t 0
3 do W ← W ∪ {xt } (i.e., add xt to the head of W )
4 repeat Drop elements from the tail of W
5 until |ˆW0 − µW1 | ≥ c holds
µ ˆ
6 for every split of W into W = W0 · W1
7 Output µW
ˆ
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 11 / 29
27. The ADWIN Algorithm
Window Management Models
W = 101010110111111
Equal fixed size subwindows Total window against subwindow
1010 1011011 1111 10101011011 1111
D. Kifer, S. Ben-David, and J. J. Gama, P. Medas, G. Castillo,
Gehrke. Detecting change in data and P. Rodrigues. Learning with
streams. 2004 drift detection. 2004
ADWIN: All Adjacent subwindows
1 01010110111111
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 12 / 29
28. The ADWIN Algorithm
Window Management Models
W = 101010110111111
Equal fixed size subwindows Total window against subwindow
1010 1011011 1111 10101011011 1111
D. Kifer, S. Ben-David, and J. J. Gama, P. Medas, G. Castillo,
Gehrke. Detecting change in data and P. Rodrigues. Learning with
streams. 2004 drift detection. 2004
ADWIN: All Adjacent subwindows
10 1010110111111
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 12 / 29
29. The ADWIN Algorithm
Window Management Models
W = 101010110111111
Equal fixed size subwindows Total window against subwindow
1010 1011011 1111 10101011011 1111
D. Kifer, S. Ben-David, and J. J. Gama, P. Medas, G. Castillo,
Gehrke. Detecting change in data and P. Rodrigues. Learning with
streams. 2004 drift detection. 2004
ADWIN: All Adjacent subwindows
101 010110111111
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 12 / 29
30. The ADWIN Algorithm
Window Management Models
W = 101010110111111
Equal fixed size subwindows Total window against subwindow
1010 1011011 1111 10101011011 1111
D. Kifer, S. Ben-David, and J. J. Gama, P. Medas, G. Castillo,
Gehrke. Detecting change in data and P. Rodrigues. Learning with
streams. 2004 drift detection. 2004
ADWIN: All Adjacent subwindows
1010 10110111111
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 12 / 29
31. The ADWIN Algorithm
Window Management Models
W = 101010110111111
Equal fixed size subwindows Total window against subwindow
1010 1011011 1111 10101011011 1111
D. Kifer, S. Ben-David, and J. J. Gama, P. Medas, G. Castillo,
Gehrke. Detecting change in data and P. Rodrigues. Learning with
streams. 2004 drift detection. 2004
ADWIN: All Adjacent subwindows
10101 0110111111
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 12 / 29
32. The ADWIN Algorithm
Window Management Models
W = 101010110111111
Equal fixed size subwindows Total window against subwindow
1010 1011011 1111 10101011011 1111
D. Kifer, S. Ben-David, and J. J. Gama, P. Medas, G. Castillo,
Gehrke. Detecting change in data and P. Rodrigues. Learning with
streams. 2004 drift detection. 2004
ADWIN: All Adjacent subwindows
101010 110111111
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 12 / 29
33. The ADWIN Algorithm
Window Management Models
W = 101010110111111
Equal fixed size subwindows Total window against subwindow
1010 1011011 1111 10101011011 1111
D. Kifer, S. Ben-David, and J. J. Gama, P. Medas, G. Castillo,
Gehrke. Detecting change in data and P. Rodrigues. Learning with
streams. 2004 drift detection. 2004
ADWIN: All Adjacent subwindows
1010101 10111111
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 12 / 29
34. The ADWIN Algorithm
Window Management Models
W = 101010110111111
Equal fixed size subwindows Total window against subwindow
1010 1011011 1111 10101011011 1111
D. Kifer, S. Ben-David, and J. J. Gama, P. Medas, G. Castillo,
Gehrke. Detecting change in data and P. Rodrigues. Learning with
streams. 2004 drift detection. 2004
ADWIN: All Adjacent subwindows
10101011 0111111
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 12 / 29
35. The ADWIN Algorithm
Window Management Models
W = 101010110111111
Equal fixed size subwindows Total window against subwindow
1010 1011011 1111 10101011011 1111
D. Kifer, S. Ben-David, and J. J. Gama, P. Medas, G. Castillo,
Gehrke. Detecting change in data and P. Rodrigues. Learning with
streams. 2004 drift detection. 2004
ADWIN: All Adjacent subwindows
101010110 111111
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 12 / 29
36. The ADWIN Algorithm
Window Management Models
W = 101010110111111
Equal fixed size subwindows Total window against subwindow
1010 1011011 1111 10101011011 1111
D. Kifer, S. Ben-David, and J. J. Gama, P. Medas, G. Castillo,
Gehrke. Detecting change in data and P. Rodrigues. Learning with
streams. 2004 drift detection. 2004
ADWIN: All Adjacent subwindows
1010101101 11111
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 12 / 29
37. The ADWIN Algorithm
Window Management Models
W = 101010110111111
Equal fixed size subwindows Total window against subwindow
1010 1011011 1111 10101011011 1111
D. Kifer, S. Ben-David, and J. J. Gama, P. Medas, G. Castillo,
Gehrke. Detecting change in data and P. Rodrigues. Learning with
streams. 2004 drift detection. 2004
ADWIN: All Adjacent subwindows
10101011011 1111
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 12 / 29
38. The ADWIN Algorithm
Window Management Models
W = 101010110111111
Equal fixed size subwindows Total window against subwindow
1010 1011011 1111 10101011011 1111
D. Kifer, S. Ben-David, and J. J. Gama, P. Medas, G. Castillo,
Gehrke. Detecting change in data and P. Rodrigues. Learning with
streams. 2004 drift detection. 2004
ADWIN: All Adjacent subwindows
101010110111 111
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 12 / 29
39. The ADWIN Algorithm
Window Management Models
W = 101010110111111
Equal fixed size subwindows Total window against subwindow
1010 1011011 1111 10101011011 1111
D. Kifer, S. Ben-David, and J. J. Gama, P. Medas, G. Castillo,
Gehrke. Detecting change in data and P. Rodrigues. Learning with
streams. 2004 drift detection. 2004
ADWIN: All Adjacent subwindows
1010101101111 11
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 12 / 29
40. The ADWIN Algorithm
Window Management Models
W = 101010110111111
Equal fixed size subwindows Total window against subwindow
1010 1011011 1111 10101011011 1111
D. Kifer, S. Ben-David, and J. J. Gama, P. Medas, G. Castillo,
Gehrke. Detecting change in data and P. Rodrigues. Learning with
streams. 2004 drift detection. 2004
ADWIN: All Adjacent subwindows
10101011011111 1
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 12 / 29
41. The ADWIN Algorithm
Algorithm ADWIN [BG06]
ADWIN has rigorous guarantees
On ratio of false positives
On ratio of false negatives
On the relation of the size of the current window and change rates
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 13 / 29
42. The ADWIN Algorithm
Algorithm ADWIN [BG06]
Theorem
At every time step we have:
1 (Few false positives guarantee) If µt remains constant within W ,
the probability that ADWIN shrinks the window at this step is at
most δ.
2 (Few false negatives guarantee) If for any partition W in two parts
W0 W1 (where W1 contains the most recent items) we have
|µW0 − µW1 | , and if
3 max{µW0 , µW1 } 4n
≥4· ln
min{n0 , n1 } δ
then with probability 1 − δ ADWIN shrinks W to W1 , or shorter.
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 14 / 29
43. The ADWIN Algorithm
Data Streams Algorithm ADWIN2 [BG06]
ADWIN2 using a Data Stream Sliding Window Model,
can provide the exact counts of 1’s in O(1) time per point.
tries O(log W ) cutpoints
uses O( 1 log W ) memory words
the processing time per example is O(log W ) (amortized) and
O(log2 W ) (worst-case).
Sliding Window Model
1010101 101 11 1 1
Content: 4 2 2 1 1
Capacity: 7 3 2 1 1
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 15 / 29
44. The ADWIN Algorithm
Algorithm ADWIN2
ADWIN2 using a Data Stream Sliding Window Model,
can provide the exact counts of 1’s in O(1) time per point.
tries O(log W ) cutpoints
uses O( 1 log W ) memory words
the processing time per example is O(log W ) (amortized) and
O(log2 W ) (worst-case).
Insert new Item
1010101 101 11 1 1
Content: 4 2 2 1 1
Capacity: 7 3 2 1 1
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 16 / 29
45. The ADWIN Algorithm
Algorithm ADWIN2
ADWIN2 using a Data Stream Sliding Window Model,
can provide the exact counts of 1’s in O(1) time per point.
tries O(log W ) cutpoints
uses O( 1 log W ) memory words
the processing time per example is O(log W ) (amortized) and
O(log2 W ) (worst-case).
Insert new Item
1010101 101 11 1 1 1
Content: 4 2 2 1 1 1
Capacity: 7 3 2 1 1 1
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 16 / 29
46. The ADWIN Algorithm
Algorithm ADWIN2
ADWIN2 using a Data Stream Sliding Window Model,
can provide the exact counts of 1’s in O(1) time per point.
tries O(log W ) cutpoints
uses O( 1 log W ) memory words
the processing time per example is O(log W ) (amortized) and
O(log2 W ) (worst-case).
Compressing Buckets
1010101 101 11 1 1 1
Content: 4 2 2 1 1 1
Capacity: 7 3 2 1 1 1
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 16 / 29
47. The ADWIN Algorithm
Algorithm ADWIN2
ADWIN2 using a Data Stream Sliding Window Model,
can provide the exact counts of 1’s in O(1) time per point.
tries O(log W ) cutpoints
uses O( 1 log W ) memory words
the processing time per example is O(log W ) (amortized) and
O(log2 W ) (worst-case).
Compressing Buckets
1010101 101 11 11 1
Content: 4 2 2 2 1
Capacity: 7 3 2 2 1
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 17 / 29
48. The ADWIN Algorithm
Algorithm ADWIN2
ADWIN2 using a Data Stream Sliding Window Model,
can provide the exact counts of 1’s in O(1) time per point.
tries O(log W ) cutpoints
uses O( 1 log W ) memory words
the processing time per example is O(log W ) (amortized) and
O(log2 W ) (worst-case).
Compressing Buckets
1010101 101 11 11 1
Content: 4 2 2 2 1
Capacity: 7 3 2 2 1
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 17 / 29
49. The ADWIN Algorithm
Algorithm ADWIN2
ADWIN2 using a Data Stream Sliding Window Model,
can provide the exact counts of 1’s in O(1) time per point.
tries O(log W ) cutpoints
uses O( 1 log W ) memory words
the processing time per example is O(log W ) (amortized) and
O(log2 W ) (worst-case).
Compressing Buckets
1010101 10111 11 1
Content: 4 4 2 1
Capacity: 7 5 2 1
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 18 / 29
50. The ADWIN Algorithm
Algorithm ADWIN2
ADWIN2 using a Data Stream Sliding Window Model,
can provide the exact counts of 1’s in O(1) time per point.
tries O(log W ) cutpoints
uses O( 1 log W ) memory words
the processing time per example is O(log W ) (amortized) and
O(log2 W ) (worst-case).
Detecting Change: Delete last Bucket
1010101 10111 11 1
Content: 4 4 2 1
Capacity: 7 5 2 1
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 19 / 29
51. The ADWIN Algorithm
Algorithm ADWIN2
ADWIN2 using a Data Stream Sliding Window Model,
can provide the exact counts of 1’s in O(1) time per point.
tries O(log W ) cutpoints
uses O( 1 log W ) memory words
the processing time per example is O(log W ) (amortized) and
O(log2 W ) (worst-case).
Detecting Change: Delete last Bucket
10111 11 1
Content: 4 2 1
Capacity: 5 2 1
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 19 / 29
52. General Framework
General Framework Time Change Detectors and
Predictors : Type I
Example (Kalman Filter)
Estimation
-
xt
-
Estimator
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 20 / 29
53. General Framework
General Framework Time Change Detectors and
Predictors : Type II
Example (Kalman Filter + CUSUM)
Estimation
-
xt
-
Estimator Alarm
- -
Change Detect.
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 20 / 29
54. General Framework
General Framework Time Change Detectors and
Predictors : Type III
Example (Adaptive Kalman Filter)
Estimation
-
xt
-
Estimator
6
-
Memory
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 20 / 29
55. General Framework
General Framework Time Change Detectors and
Predictors : Type IV
Example (ADWIN, Kalman Filter+ADWIN)
Estimation
-
xt
-
Estimator Alarm
- -
Change Detect.
6
6
?
-
Memory
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 20 / 29
56. General Framework
Time Change Detectors and Predictors: A General
Framework
No memory Memory
No Change Type I Type III
Detector Kalman Filter Adaptive Kalman Filter
Change Type II Type IV
Detector Kalman Filter + CUSUM ADWIN
Kalman Filter + ADWIN
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 21 / 29
57. General Framework
Time Change Detectors and Predictors: A General
Framework
No memory Memory
No Change Type I Type III
Detector Kalman Filter Adaptive Kalman Filter
Q,R estimated from window
Change Type II Type IV
Detector Kalman Filter + CUSUM ADWIN
Kalman Filter + ADWIN
Q,R estimated from window
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 21 / 29
58. K-ADWIN
K-ADWIN = ADWIN + Kalman Filtering
Estimation
-
xt
-
Kalman Alarm
- -
ADWIN
6
6
?
-
ADWIN Memory
R = W 2 /50 and Q = 200/W , where W is the length of the window
maintained by ADWIN.
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 22 / 29
59. Experimental Validation of K-ADWIN
Tracking Experiments
KALMAN: R=1000;Q=1 Error= 854.97
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 23 / 29
60. Experimental Validation of K-ADWIN
Tracking Experiments
ADWIN : Error= 674.66
KALMAN: R=1000;Q=1 Error= 854.97
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 23 / 29
61. Experimental Validation of K-ADWIN
Tracking Experiments
K-ADWIN Error= 530.13
ADWIN : Error= 674.66
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 23 / 29
62. Experimental Validation of K-ADWIN
Naïve Bayes
Example
Data set that outlook temp. humidity windy play
describes the sunny hot high false no
weather sunny hot high true no
conditions for overcast hot high false yes
playing some rainy mild high false yes
game. rainy cool normal false yes
rainy cool normal true no
overcast cool normal true yes
Assume we have to classify the following new instance:
outlook temp. humidity windy play
sunny cool high true ?
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 24 / 29
63. Experimental Validation of K-ADWIN
Naïve Bayes
Assume we have to classify the following new instance:
outlook temp. humidity windy play
sunny cool high true ?
We classify the new instance:
νNB = arg max P(νj )P(sunny |νj )P(cool|νj )P(high|νj )P(true|νj )
ν∈{yes,no}
Conditional probabilities can be estimated directly as frequencies:
number of instances with attribute ai and class νj
P(ai |νj ) =
total number of training instances with class νj
Create one estimator for each frequence that needs estimation
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 24 / 29
64. Experimental Validation of K-ADWIN
Experimental Validation of K-ADWIN
We test Naïve Bayes Predictor and k-means clustering
Method: replace counters by estimators
Synthetic data where change is controllable
Naïve Bayes: We compare accuracy of
Static model: Training of 1000 samples every instant
Dynamic model: replace probabilities counters by estimators
%Dynamic
computing the ratio Static with tests using 2000 samples.
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 25 / 29
65. Experimental Validation of K-ADWIN
Naïve Bayes Predictor
Width %Static %Dynamic % Dynamic/Static
ADWIN 83,36% 80,30% 96,33%
Kalman Q = 1, R = 1000 83,22% 71,13% 85,48%
Kalman Q = 1, R = 1 83,21% 56,91% 68,39%
Kalman Q = .25, R = .25 83,26% 56,91% 68,35%
Adaptive Kalman 83,24% 76,21% 91,56%
CUSUM Kalman 83,30% 50,65% 60,81%
K-ADWIN 83,24% 81,39% 97,77%
Fixed-sized Window 32 83,28% 67,64% 81,22%
Fixed-sized Window 128 83,30% 75,40% 90,52%
Fixed-sized Window 512 83,28% 80,47% 96,62%
Fixed-sized Window 2048 83,24% 82,19% 98,73%
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 26 / 29
66. Experimental Validation of K-ADWIN
k -means Clustering
σ = 0.15
Width Static Dynamic
ADWIN 9,72 21,54
Kalman Q = 1, R = 1000 9,72 19,72
Kalman Q = 1, R = 100 9,71 17,60
Kalman Q = .25, R = .25 9,71 22,63
Adaptive Kalman 9,72 18,98
CUSUM Kalman 9,72 18,29
K-ADWIN 9,72 17,30
Fixed-sized Window 32 9,72 25,70
Fixed-sized Window 128 9,72 36,42
Fixed-sized Window 512 9,72 38,75
Fixed-sized Window 2048 9,72 39,64
Fixed-sized Window 8192 9,72 43,39
Fixed-sized Window 32768 9,72 53,82
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 27 / 29
67. Experimental Validation of K-ADWIN
Results
No estimator ever does much better than K-ADWIN
K-ADWIN does much better than every other estimators in at least
one context.
Tracking problem K-ADWIN and ADWIN automatically do about as
well as the Kalman filter with the best set of fixed covariance
parameters.
Naïve Bayes and k -means: K-ADWIN does somewhat better than
ADWIN and far better than any memoryless Kalman filter.
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 28 / 29
68. Conclusions
Conclusions and Future Work
K-ADWIN tunes itself to the data stream at hand, with no need for
the user to hardwire or precompute parameters.
Better results than either memoryless Kalman Filtering or sliding
windows with linear estimators.
Future work :
Tests on real-world, not only synthetic data.
Other learning algorithms: algorithms for induction of decision
trees.
A. Bifet, R. Gavaldà (UPC) Kalman Filters and Adaptive Windows for Learning in Data Streams
DS’06 Barcelona 29 / 29