The document presents an approach to using data mining techniques to build predictive models for predicting the need to replace aircraft components using data collected from aircraft sensors. It addresses four key challenges: selecting relevant data from the multiple datasets generated by aircraft, automatically labeling examples with classification values, evaluating models while accounting for dependencies between examples, and combining results from models built on different datasets. The approach was applied to predict problems for various aircraft components using over 3 years of data from 34 aircraft.
Data-Driven Hydrocarbon Production Forecasting Using Machine Learning Techniques (pp. 65-72)
Masoud Safari Zanjani, Mohammad Abdus Salam, Osman Kandara
Department of Computer Science, Southern University, Baton Rouge, Louisiana, USA.
Vol. 18 No. 6 JUNE 2020 International Journal of Computer Science and Information Security
https://sites.google.com/site/ijcsis/vol-18-no-6-jun-2020
Presented in this short document is a description of our three separate techniques to analyze the data by checking, clustering and componentizing it before it is used by other IMPL’s routines especially in on-line/real-time decision-making applications. We also have other data consistency or analysis techniques which have been described in other IMPL documents and these relate to the application of data reconciliation and regression with diagnostics but require an explicit model (model-based) whereas the techniques below do not i.e., they are data-based techniques.
TECHNICAL REVIEW: PERFORMANCE OF EXISTING IMPUTATION METHODS FOR MISSING DATA...IJDKP
Incomplete data is present in many study contents. This incomplete or uncollected data information is named as missing data (values), and considered as vital problem for various researchers. Even this missing data problem is faced more in air pollution monitoring stations, where data is collected from multiple monitoring stations widespread across various locations. In literature, various imputation methods for missing data are proposed, however, in this research we considered only existing imputation methods for missing data and recorded their performance in ensemble creation. The five existing imputation methods for missing data deployed in this research are series mean method, mean of nearby points, median of nearby points, linear trend at a point and linear interpolation respectively. Series mean (SM) method demonstrated comparatively better to other imputation methods with least mean absolute error and better performance accuracy for SVM ensemble creation on CO data set using bagging and boosting algorithms.
Data-Driven Hydrocarbon Production Forecasting Using Machine Learning Techniques (pp. 65-72)
Masoud Safari Zanjani, Mohammad Abdus Salam, Osman Kandara
Department of Computer Science, Southern University, Baton Rouge, Louisiana, USA.
Vol. 18 No. 6 JUNE 2020 International Journal of Computer Science and Information Security
https://sites.google.com/site/ijcsis/vol-18-no-6-jun-2020
Presented in this short document is a description of our three separate techniques to analyze the data by checking, clustering and componentizing it before it is used by other IMPL’s routines especially in on-line/real-time decision-making applications. We also have other data consistency or analysis techniques which have been described in other IMPL documents and these relate to the application of data reconciliation and regression with diagnostics but require an explicit model (model-based) whereas the techniques below do not i.e., they are data-based techniques.
TECHNICAL REVIEW: PERFORMANCE OF EXISTING IMPUTATION METHODS FOR MISSING DATA...IJDKP
Incomplete data is present in many study contents. This incomplete or uncollected data information is named as missing data (values), and considered as vital problem for various researchers. Even this missing data problem is faced more in air pollution monitoring stations, where data is collected from multiple monitoring stations widespread across various locations. In literature, various imputation methods for missing data are proposed, however, in this research we considered only existing imputation methods for missing data and recorded their performance in ensemble creation. The five existing imputation methods for missing data deployed in this research are series mean method, mean of nearby points, median of nearby points, linear trend at a point and linear interpolation respectively. Series mean (SM) method demonstrated comparatively better to other imputation methods with least mean absolute error and better performance accuracy for SVM ensemble creation on CO data set using bagging and boosting algorithms.
Predicting and visualizing the heart diseases by machine learning algorithms ...KannanRamasamy25
Arthur Samuel (1959) :
"Field of study that gives computers the ability to learn without being explicitly programmed“
Tom Mitchell (1998) :
“A computer program is said to learn from experience E with respect to some task T and some performance measure P, if its performance on T, as measured by P, improves with experience E”.
There are several ways to implement machine learning algorithms such as,
Automating automation
Getting computers to program themselves
Mx net image segmentation to predict and diagnose the cardiac diseases karp...KannanRamasamy25
Powerful open-source deep learning framework instrument
MXNet supports multiple languages like C++, Python, R, Julia, Perl etc
MXNet supported by Intel, Dato, Baidu, Microsoft, Wolfram Research, and research institutions such as Carnegie Mellon, MIT, the University of Washington, and the Hong Kong University of Science and Technology
Symbolic Execution: Static symbolic graph executor, which provides efficient symbolic graph execution and optimization.
Supports an efficient deployment of a trained model to low-end devices for inference, such as mobile devices, IoT devices (using AWS Greengrass), Serverless (Using AWS Lambda) or containers.
The International Journal of Chaos, Control, Modeling and Simulation is a Quarterly open access peer-reviewed journal that publishes articles which contribute new results in all areas of Chaos Theory, Control Systems, Scientific Modeling and Computer Simulation. In the last few decades, chaos theory has found important applications in secure communication devices and secure data encryption. Control methods are very useful in several applied areas like Chaos, Automation, Communication, Robotics, Power Systems, and Biomedical Instrumentation. The journal focuses on all technical and practical aspects of Chaos, Control, Modeling and Simulation. The goal of this journal is to bring together researchers and practitioners from academia and industry to focus on chaotic systems and applications, control theory, process control, automation, modeling concepts, computational methods, computer simulation and establishing new collaborations in these areas.
Authors are solicited to contribute to this journal by submitting articles that illustrate research results, projects, surveying works and industrial experiences that describe significant advances in Chaos Theory, Control Systems, Modeling and Simulation Techniques.
Techniques for the evaluation of complex polynomials with one and two variables are
introduced.Polynomials arise in may areas such as control systems, image and signal processing, coding
theory,electrical networks, etc., and their evaluations are time consuming. This paper introduces new
evaluationalgorithms that are straightforward with fewer arithmetic operations and a fast matrix
exponentiation technique.
Cluster Computing Environment for On - line Static Security Assessment of lar...IDES Editor
The increased size of modern power systems
demand faster and accurate means for the security assessment,
so that the decisions for reliable and secure operation planning
could be drawn in a systematic manner. Large computational
overhead is the major impediment in preventing the power
system security assessment (PSSA) from on-line use. To
mitigate this problem, this paper proposes, a cluster computing
based architecture for power system static security assessment,
utilizing the tools in the open source domain. A variant of the
master/slave pattern is used for deploying the cluster of
workstations (COW), which act as the computational engine
for the on-line PSSA. The security assessment is performed
utilizing the developed composite security index that can
accurately differentiate the secure and non-secure cases and
has been defined as a function of bus voltage and line flow
limit violations. Due to the inherent parallel structure of
security assessment algorithm and to exploit the potential of
distributed computing, domain decomposition is employed for
parallelizing the sequential algorithm. Extensive
experimentations were carried out on IEEE 57 bus and IEEE
145-bus 50 machine standard test systems for demonstrating
the validity of the proposed architecture.
Software CrashLocator: Locating the Faulty Functions by Analyzing the Crash S...INFOGAIN PUBLICATION
In recent years, studies have been dedicated mainly in the analysis, of crashes in real-world related to large-scale software systems. A crash in terms of computing can be termed as a computer program such as a software application that stops functioning properly. Software crash is a serious problem in production environment. When crash happens, the crash report with the stack trace of software at time of crash is sent to the developer team. Software development team may receive hundreds of stack traces from all deployment sites and many stack traces may be due to same problem. If the developer starts analyzing each trace, it may take a longer duration of time and redundancy many happen in terms of two developers fixing the same problem. This motivates us to present the solution to analyze the stack traces and find the important functions responsible for crash and rank them, so that development resources can be optimized. In this paper we have proposed the solution to solve the problem by developing Software CrashLocator.
MACHINE LEARNING TECHNIQUES FOR ANALYSIS OF EGYPTIAN FLIGHT DELAYIJDKP
Flight delay has been the fiendish problem to the world's aviation industry, so there is very important
significance to research for computer system predicting flight delay propagation. Extraction of hidden
information from large datasets of raw data could be one of the ways for building predictive model. This
paper describes the application of classification techniques for analysing the Flight delay pattern in Egypt
Airline’s Flight dataset. In this work, four decision tree classifiers were evaluated and results show that the
REPTree have the best accuracy 80.3% with respect to Forest, Stump and J48. However, four rules based
classifiers were compared and results show that PART provides best accuracy among studied rule-based
classifiers with accuracy of 83.1%. By analysing running time for all classifiers, the current work
concluded that REPtree is the most efficient classifier with respect to accuracy and running time. Also, the
current work is extended to apply of Apriori association technique to extract some important information
about flight delay. Association rules are presented and association technique is evaluated.
Keynote Predictive Maintenance in Aviation Sander De Bree
Using big data in aircraft maintenance – a practical guide on how data in aircraft maintenance can be used for predictive analytical purposes.
Over recent years the topic of data being produced by aircraft has taken centre stage in the pursuit of enabling predictive analytical capabilities in aircraft maintenance. During this presentation we will dive into various aspects of data of aircraft, get familiar with some terminology and how data can be used to enable predictive analytics within maintenance & engineering. In addition, we will look at a real airline business use case that has applied the full methodology and how they have benefitted from this.
Using Machine Learning to Quantify the Impact of Heterogeneous Data on Transf...Power System Operation
Using large-scale distributed computing and a variety of heterogeneous data sources including real-time sensor measurements, dissolved gas measurements, and localized historical weather, we construct a predictive model that allows us to accurately predict remaining useful life and failure probabilities for a fleet of network transformers. Our model is robust to highly variable data types, including both static and dynamic data, sparse and dense time series, and measurements of internal and external processes (such as weather). By comparing the predictive performance of models built on different combinations of these data sources, we can quantify the marginal benefit of including each additional data source in our model.
In order to relate each type of data to the risk of failure across a fleet of transformers, we have developed a novel class of survival models, the convex latent variable (CLV) model. This type of specialized survival model has several advantages. Rather than an opaque and subjective "health index", it produces interpretable predictions like the probability of failure within a given time window or the expected RUL of an asset. Our framework supports accurate estimates of the risk of equipment failure across a wide range of time-scales, from a few weeks to many years in the future, and can model not just the instantaneous risk of failure due to an event like a storm, but also the long-term impact on the risk of failure.
An intrusion detection algorithm for amiIJCI JOURNAL
Nowadays, using the smart metering devices for energy users to manage a wide variety of subscribers,
reading devices for measuring, billing, disconnection and connection of subscribers’ connection
management is an important issue. The performance of these intelligent systems is based on information
transfer in the context of information technology, so reported data from network should be managed to
avoid the malicious activities that including the issues that could affect the quality of service the system. In
this paper for control of the reported data and to ensure the veracity of the obtained information, using
intrusion detection system is proposed based on the support vector machine and principle component
analysis (PCA) to recognize and identify the intrusions and attacks in the smart grid. Here, the operation of
intrusion detection systems for different kernel of SVM when using support vector machine (SVM) and PCA
simultaneously is studied. To evaluate the algorithm, based on data KDD99, numerical simulation is done
on five different kernels for an intrusion detection system using support vector machine with PCA
simultaneously. Also comparison analysis is investigated for presented intrusion detection algorithm in
terms of time - response, rate of increase network efficiency and increase system error and differences in
the use or lack of use PCA. The results indicate that correct detection rate and the rate of attack error
detection have best value when PCA is used, and when the core of algorithm is radial type, in SVM
algorithm reduces the time for data analysis and enhances performance of intrusion detection.
A Framework for Dead stock Management System for in-house computer engineerin...theijes
Equipment management is an important issue for the safety and cost in an institute. In addition, the use of an efficient information system effectively promotes the processing performance. The maintenance of deadstock’s information using paperwork is very difficult in terms of time. Deadstock management system to be used for in-house computer engineering department. The system is web based and uses intranet approach for communication between different users of the system. Through the related application, it has efficiently improved operation such as addition, modification and deletion of dead stock information. The system also generates report useful for arranging equipment’s for different purpose like practical exam and workshops. System also sends notification to users for effective and consistence maintenance of data. The system can be used to improve the work quality, reduce the maintenance cost and promote the safety of all equipment
Predicting and visualizing the heart diseases by machine learning algorithms ...KannanRamasamy25
Arthur Samuel (1959) :
"Field of study that gives computers the ability to learn without being explicitly programmed“
Tom Mitchell (1998) :
“A computer program is said to learn from experience E with respect to some task T and some performance measure P, if its performance on T, as measured by P, improves with experience E”.
There are several ways to implement machine learning algorithms such as,
Automating automation
Getting computers to program themselves
Mx net image segmentation to predict and diagnose the cardiac diseases karp...KannanRamasamy25
Powerful open-source deep learning framework instrument
MXNet supports multiple languages like C++, Python, R, Julia, Perl etc
MXNet supported by Intel, Dato, Baidu, Microsoft, Wolfram Research, and research institutions such as Carnegie Mellon, MIT, the University of Washington, and the Hong Kong University of Science and Technology
Symbolic Execution: Static symbolic graph executor, which provides efficient symbolic graph execution and optimization.
Supports an efficient deployment of a trained model to low-end devices for inference, such as mobile devices, IoT devices (using AWS Greengrass), Serverless (Using AWS Lambda) or containers.
The International Journal of Chaos, Control, Modeling and Simulation is a Quarterly open access peer-reviewed journal that publishes articles which contribute new results in all areas of Chaos Theory, Control Systems, Scientific Modeling and Computer Simulation. In the last few decades, chaos theory has found important applications in secure communication devices and secure data encryption. Control methods are very useful in several applied areas like Chaos, Automation, Communication, Robotics, Power Systems, and Biomedical Instrumentation. The journal focuses on all technical and practical aspects of Chaos, Control, Modeling and Simulation. The goal of this journal is to bring together researchers and practitioners from academia and industry to focus on chaotic systems and applications, control theory, process control, automation, modeling concepts, computational methods, computer simulation and establishing new collaborations in these areas.
Authors are solicited to contribute to this journal by submitting articles that illustrate research results, projects, surveying works and industrial experiences that describe significant advances in Chaos Theory, Control Systems, Modeling and Simulation Techniques.
Techniques for the evaluation of complex polynomials with one and two variables are
introduced.Polynomials arise in may areas such as control systems, image and signal processing, coding
theory,electrical networks, etc., and their evaluations are time consuming. This paper introduces new
evaluationalgorithms that are straightforward with fewer arithmetic operations and a fast matrix
exponentiation technique.
Cluster Computing Environment for On - line Static Security Assessment of lar...IDES Editor
The increased size of modern power systems
demand faster and accurate means for the security assessment,
so that the decisions for reliable and secure operation planning
could be drawn in a systematic manner. Large computational
overhead is the major impediment in preventing the power
system security assessment (PSSA) from on-line use. To
mitigate this problem, this paper proposes, a cluster computing
based architecture for power system static security assessment,
utilizing the tools in the open source domain. A variant of the
master/slave pattern is used for deploying the cluster of
workstations (COW), which act as the computational engine
for the on-line PSSA. The security assessment is performed
utilizing the developed composite security index that can
accurately differentiate the secure and non-secure cases and
has been defined as a function of bus voltage and line flow
limit violations. Due to the inherent parallel structure of
security assessment algorithm and to exploit the potential of
distributed computing, domain decomposition is employed for
parallelizing the sequential algorithm. Extensive
experimentations were carried out on IEEE 57 bus and IEEE
145-bus 50 machine standard test systems for demonstrating
the validity of the proposed architecture.
Software CrashLocator: Locating the Faulty Functions by Analyzing the Crash S...INFOGAIN PUBLICATION
In recent years, studies have been dedicated mainly in the analysis, of crashes in real-world related to large-scale software systems. A crash in terms of computing can be termed as a computer program such as a software application that stops functioning properly. Software crash is a serious problem in production environment. When crash happens, the crash report with the stack trace of software at time of crash is sent to the developer team. Software development team may receive hundreds of stack traces from all deployment sites and many stack traces may be due to same problem. If the developer starts analyzing each trace, it may take a longer duration of time and redundancy many happen in terms of two developers fixing the same problem. This motivates us to present the solution to analyze the stack traces and find the important functions responsible for crash and rank them, so that development resources can be optimized. In this paper we have proposed the solution to solve the problem by developing Software CrashLocator.
MACHINE LEARNING TECHNIQUES FOR ANALYSIS OF EGYPTIAN FLIGHT DELAYIJDKP
Flight delay has been the fiendish problem to the world's aviation industry, so there is very important
significance to research for computer system predicting flight delay propagation. Extraction of hidden
information from large datasets of raw data could be one of the ways for building predictive model. This
paper describes the application of classification techniques for analysing the Flight delay pattern in Egypt
Airline’s Flight dataset. In this work, four decision tree classifiers were evaluated and results show that the
REPTree have the best accuracy 80.3% with respect to Forest, Stump and J48. However, four rules based
classifiers were compared and results show that PART provides best accuracy among studied rule-based
classifiers with accuracy of 83.1%. By analysing running time for all classifiers, the current work
concluded that REPtree is the most efficient classifier with respect to accuracy and running time. Also, the
current work is extended to apply of Apriori association technique to extract some important information
about flight delay. Association rules are presented and association technique is evaluated.
Keynote Predictive Maintenance in Aviation Sander De Bree
Using big data in aircraft maintenance – a practical guide on how data in aircraft maintenance can be used for predictive analytical purposes.
Over recent years the topic of data being produced by aircraft has taken centre stage in the pursuit of enabling predictive analytical capabilities in aircraft maintenance. During this presentation we will dive into various aspects of data of aircraft, get familiar with some terminology and how data can be used to enable predictive analytics within maintenance & engineering. In addition, we will look at a real airline business use case that has applied the full methodology and how they have benefitted from this.
Using Machine Learning to Quantify the Impact of Heterogeneous Data on Transf...Power System Operation
Using large-scale distributed computing and a variety of heterogeneous data sources including real-time sensor measurements, dissolved gas measurements, and localized historical weather, we construct a predictive model that allows us to accurately predict remaining useful life and failure probabilities for a fleet of network transformers. Our model is robust to highly variable data types, including both static and dynamic data, sparse and dense time series, and measurements of internal and external processes (such as weather). By comparing the predictive performance of models built on different combinations of these data sources, we can quantify the marginal benefit of including each additional data source in our model.
In order to relate each type of data to the risk of failure across a fleet of transformers, we have developed a novel class of survival models, the convex latent variable (CLV) model. This type of specialized survival model has several advantages. Rather than an opaque and subjective "health index", it produces interpretable predictions like the probability of failure within a given time window or the expected RUL of an asset. Our framework supports accurate estimates of the risk of equipment failure across a wide range of time-scales, from a few weeks to many years in the future, and can model not just the instantaneous risk of failure due to an event like a storm, but also the long-term impact on the risk of failure.
An intrusion detection algorithm for amiIJCI JOURNAL
Nowadays, using the smart metering devices for energy users to manage a wide variety of subscribers,
reading devices for measuring, billing, disconnection and connection of subscribers’ connection
management is an important issue. The performance of these intelligent systems is based on information
transfer in the context of information technology, so reported data from network should be managed to
avoid the malicious activities that including the issues that could affect the quality of service the system. In
this paper for control of the reported data and to ensure the veracity of the obtained information, using
intrusion detection system is proposed based on the support vector machine and principle component
analysis (PCA) to recognize and identify the intrusions and attacks in the smart grid. Here, the operation of
intrusion detection systems for different kernel of SVM when using support vector machine (SVM) and PCA
simultaneously is studied. To evaluate the algorithm, based on data KDD99, numerical simulation is done
on five different kernels for an intrusion detection system using support vector machine with PCA
simultaneously. Also comparison analysis is investigated for presented intrusion detection algorithm in
terms of time - response, rate of increase network efficiency and increase system error and differences in
the use or lack of use PCA. The results indicate that correct detection rate and the rate of attack error
detection have best value when PCA is used, and when the core of algorithm is radial type, in SVM
algorithm reduces the time for data analysis and enhances performance of intrusion detection.
A Framework for Dead stock Management System for in-house computer engineerin...theijes
Equipment management is an important issue for the safety and cost in an institute. In addition, the use of an efficient information system effectively promotes the processing performance. The maintenance of deadstock’s information using paperwork is very difficult in terms of time. Deadstock management system to be used for in-house computer engineering department. The system is web based and uses intranet approach for communication between different users of the system. Through the related application, it has efficiently improved operation such as addition, modification and deletion of dead stock information. The system also generates report useful for arranging equipment’s for different purpose like practical exam and workshops. System also sends notification to users for effective and consistence maintenance of data. The system can be used to improve the work quality, reduce the maintenance cost and promote the safety of all equipment
Predicting the Maintenance of Aircraft Engines using LSTMijtsrd
What if apart of aircraft could let you know when the aircraft component needed to be replaced or repaired It can be done with continuous data collection, monitoring, and advanced analytics. In the aviation industry, predictive maintenance promises increased reliability as well as improved supply chain and operational performance. The main goal is to ensure that the engines work correctly under all conditions and there is no risk of failure. If an effective method for predicting failures is applied, maintenance may be improved. The main source of data regarding the health of the engines is measured during the flights. Several variables are calculated, including fan speed, core speed, quantity and oil pressure and, environmental variables such as outside temperature, aircraft speed, altitude, and so on. Sensor data obtained in real time can be used to model component deterioration. To predict the maintenance of an aircraft engine, LSTM networks is used in this paper. A sequential input file is dealt with by the LSTM model. The training of LSTM networks was carried out on a high performance large scale processing engine. Machines, data, ideas, and people must all be brought together to understand the importance of predictive maintenance and achieve business results that matter. Nitin Prasad | Dr. A Rengarajan "Predicting the Maintenance of Aircraft Engines using LSTM" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-4 , June 2021, URL: https://www.ijtsrd.compapers/ijtsrd41288.pdf Paper URL: https://www.ijtsrd.comcomputer-science/data-processing/41288/predicting-the-maintenance-of-aircraft-engines-using-lstm/nitin-prasad
ESTIMATING HANDLING TIME OF SOFTWARE DEFECTScsandit
The problem of accurately predicting handling time for software defects is of great practical
importance. However, it is difficult to suggest a practical generic algorithm for such estimates,
due in part to the limited information available when opening a defect and the lack of a uniform
standard for defect structure. We suggest an algorithm to address these challenges that is
implementable over different defect management tools. Our algorithm uses machine learning
regression techniques to predict the handling time of defects based on past behaviour of similar
defects. The algorithm relies only on a minimal set of assumptions about the structure of the
input data. We show how an implementation of this algorithm predicts defect handling time with
promising accuracy results
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Data mining-for-prediction-of-aircraft-component-replacement
1. Data Mining For Prediction of Aircraft
Component Replacement
Sylvain Letourneau1 Fazel Famili1 Stan Matwin2
1Institute for Information Technology
National Research Council of Canada, Ottawa
fsylvain.letourneau, fazel.familig@iit.nrc.ca
2School of Information Technology and Engineering
University of Ottawa, Canada
stan@site.uottawa.ca
August 3, 1999
Abstract
The operation and maintenance of modern sensor-equipped systems
such as passenger aircraft generate vast amounts of numerical and sym-
bolic data. Learning models from this data to predict problems with com-
ponent may lead to considerable savings, reducing the number of delays,
and increasing the overall level of safety. Several data mining techniques
exist to learn models from vast amounts of data. However, the use of these
techniques to infer the desired models from the data obtained during the
operation and maintenance of aircraft is extremely challenging. Dicul-
ties that need to be addressed include: data gathering, data labeling,
data and model integration, and model evaluation. This paper presents
an approach that addresses these issues. We also report results from the
application of this approach to build models that predict problems for a
variety of aircraft components.
Keywords: Data mining, machine learning, aircraft health monitoring,
component failure prediction.
1 Introduction
The operation and maintenance of modern sensor-equipped systems such as
aircraft generates vast amounts of numerical and symbolic data. The data is
1
2. generated by thousands of sensors installed in various components of the air-
craft; it is then sent in real-time to ground stations and then stored in relational
databases. Before being transmitted to the ground, a number of on-board com-
puter systems monitor and analyze the data in order to make sure that various
systems of the aircraft operate properly. However, once the data is stored in
central databases, further data analysis is rarely performed. This paper presents
an approach that makes use of this data in order to develop models to predict
the need for replacement of various aircraft components before they become
non-operational. The end goal is to implement these models in a
ight data
monitoring system that will receive as input (in real-time) the data from a
eet
of commercial aircraft, will analyze it, and as output it will alert appropriate
sta when there is a need for a component replacement. The monitoring sys-
tem will use the automatically sensor data from the aircraft, and a predictive
model described in this paper, to detect component problems and recommend
their replacement. Such a system could help improving the airline's operation
by: reducing the number of delays, reducing the maintenance costs, helping to
obtain better maintenance planning, and increasing the level of safety.
The approach proposed in this paper applies techniques from the
3. elds of
Machine Learning and Data Mining on huge amounts of complex historical data
in order to develop the predictive models required by the monitoring system.
The approach described addresses four fundamental diculties with existing
data mining approaches: automatic selection of relevant data, automatic label-
ing of instances, evaluation method that accounts for dependencies between the
instances, and a scoring function measuring the extent to which the results
4. t
the domain requirements. By addressing these four issues, we believe that the
2
5. proposed approach will help extending the range of potential applications for
data mining techniques. Examples of other applications that can bene
6. t from
the approach developed in this paper are: prediction of problems in complex sys-
tems (e.g.: trucks, ships, trains, and cars), prediction of problems with complex
industrial equipment for which a lot of data is continuously acquired, and pre-
diction of critical events in medical applications (e.g. Emergency Room care).
The fact that the proposed approach relies on a minimal amount of domain
speci
7. c information will also facilitate the adaptation to other applications.
2 The application and the data used
The aim of the approach described in this paper is to generate a valid set of
models to predict the need for replacement of an aircraft component. These
models will have to accurately recognize particular patterns in the data that
indicate upcoming problems with a component. Our approachmakes use of data
mining techniques to infer these models from the available data. Ideally, the
models developed will be able to recognize problems within a reasonable period
of time prior to the actual occurrence of the problem. For most components,
a period of 1-3 weeks in advance is appropriate. For components that are very
expensive or dicult to obtain, it may be preferable to receive alerts even before
3 weeks in order to allow enough time for proper actions.
There are two reasons to replace an aircraft component: either the compo-
nent has to be replaced as part of regular maintenance (imposed by aerospace
regulations or airline's policy) or it is in a deteriorated condition and needs to
be replaced before it fails. Only the second type of replacement requires the de-
velopment of predictive models because maintenance sta already know about
3
8. regular maintenance requirements. It is also important to remember that no
accurate predictions can be expected for components for which relevant data is
not available. For instance, there is usually no data related to the quality of
passenger seats, therefore one will not be able to predict the need for replace-
ment of passenger seats. Since the majority of the data available for this project
is related to the engines of the aircraft, we focus on models to predict problems
with engine's components.
The approach presented in this paper does not attempt to capture all phe-
nomena that may lead to a component failure. Two types of component failures
that our approach is very likely to miss are: failure because of a problematic
maintenance action, and failure due to a design problem with a speci
9. c compo-
nent. Currently, we do not have sucient data about the maintenance actions
and design of the components to identify such phenomena. Fortunately, these
problems do not represent a very high percentage of all component problems.
More than 3 years of data from a
eet of 34 Airbus A320 is available for
this project and new data is continuously acquired. Each Airbus A320 gener-
ates around 1MB of data per month. The data consists of two major parts:
(i) textual descriptions of all repair actions taken on these aircraft, and (ii)
all parametric data acquired from sensors during the operation of the aircraft.
The parametric data is obtained in the form of reports at dierent stages of
operation. For example, during takeo a report which consists of about 100
measurements is generated, in stable cruising conditions another report is gen-
erated with a dierent set of measurements. etc. An Airbus A320 generates up
to 19 types of reports. These are generated at dierent frequencies and contain
between 20 and 150 parameters (numeric and symbolic).
4
10. 3 Application challenges
The idea of using data mining techniques to infer models from historical data
is not new and has already been applied in other applications. On the other
hand, our approach contributes to the domain by addressing a number of issues
that are not taken into account in these classical applications of data mining.
The
11. rst two issues are related to data preparation requirements prior to the
application of data mining techniques, the third issue involves adequate evalu-
ation of the results for the application considered, and
12. nally, the fourth issue
is about the fusion of results obtained during the analysis.
Data Gathering Most Data mining techniques require as input one dataset
containing a set of examples described by a vector of attribute values. Modern
aircraft such as Airbus A320 do not generate one, but up to 19 dierent datasets
reporting the status of the aircraft in dierent phases of operation. The number
of examples in each dataset varies considerably from one dataset to another.
Given a component of interest, the
13. rst problem is to decide on which dataset(s)
to use to develop the predictive models.
Once a dataset has been chosen, we must select the subset of instances to be
included in the analysis. Considering the number of instances available in each
dataset, it will be very inecient to try to learn the models using all instances.
Simple solutions such as random sampling are not appropriate either. In order
to build the desired predictive models, we need to focus the analysis on the data
generated around each occurrence of component replacement. The approach to
retrieve this data is explained in Section 4.
Data Labeling Data mining approaches are typically classi
14. ed in two cat-
egories: supervised and unsupervised [1]. The two approaches dier in the data
5
15. they require as input and the type of tasks they can address. For instance, a su-
pervised learning approach will require each instance to be de
16. ned by a number
of attributes along with its class membership. Given this data as input, the su-
pervised algorithm will try to learn models to predict the class of each instance
using the other attributes. Supervised learning approaches are useful in both
classi
17. cation and forecasting tasks. On the other hand, unsupervised learning
approach will take as input as set of examples without any class information
and will try to
18. nd groups of similar instances.
Since there exists a variety of robust classi
20. cation task
with two class values: (i) there is a need to replace the component or (ii) there
is no need for replacement. The model learned from the data will have to
classify each new report received from the aircraft in one of these two classes.
Supervised approaches are appropriate for this kind of tasks. However, the
data that we receive from the aircraft cannot be directly used by a supervised
learning approach because it does not contain the class attribute. The solution
is to have a pre-processing step that automatically computes the membership
value of all examples that will be used during the analysis. This process is called
Data Labeling and it is described in Section 5.
Model Evaluation Proper evaluation of the results is a key factor in a
practical application such as the one described here. The evaluation approach
must respect two criteria: (i) provide a fair estimate of the performance of the
model when applied to new data, and (ii) take into account important domain
speci
21. c requirements. In Section 6, we present an approach to adequately come
up with a good estimate for the performance of a model. This approach ex-
6
22. tends the well known cross-validation technique. We also discuss an evaluation
function that accounts for important application constrains. This new evalu-
ation function diers considerably from traditional evaluation criteria such as
accuracy, and recall/precision.
Exploiting multiple sources of information Data mining algorithms are
designed to take as input one dataset in a given format. As discussed earlier,
among all datasets generated by the aircraft, more than one could be poten-
tially useful to predict the need for replacement of the component of interest.
The question is how could we eciently combine the information from the dif-
ferent datasets that are all relevant. Two basic approaches exist: (i) merge
the datasets into a new dataset during a pre-processing step and then we build
the models from this new dataset, or (ii) build independent predictive models
using dierent datasets and then we combine the output of these models to
get the
23. nal prediction. The former approach is dicult to implement for a
number of reasons. First, there is no obvious merging strategy because dierent
datasets may contain a very dierent number of examples. Moreover, by merg-
ing datasets which already have a signi
24. cant number of attributes we are likely
to end up with a very complex dataset. This will contribute to make learn-
ing of the models even more dicult. Our experiments have been focusing on
the second approach which turned out to be easier to implement while helping
to improve the results. Models are developed independently from the dierent
datasets, then we combine the output of these initial models to get the
25. nal
prediction. In this paper, we do not fully address this fourth issue (merging of
results) because we are still experimenting with dierent strategies to combine
the models.
7
26. Building models to predict need for component replacement
APU reports
ETO reports
ECR reports
EOR reports
Data
Gathering
Data
Labeling
Modeling
Evaluation
Model
Fusion
Final
Models
Repair
descriptions
Figure 1: The 4 steps process to build models for component replacements.
It is important to note that the
27. rst three issues must be addressed in order
to apply data mining techniques to predict the need for aircraft component
replacement. On the other hand, the fourth issue is only related to optimization
and as we argue in Section 6, promising results can be obtained without fully
addressing it. Figure 1 presents the main steps of the approach developed to
build models for predicting the need for component replacement. The
28. rst three
processes are further explained in the following sections.
4 Data Gathering
There are two tasks to perform during this step. The
29. rst one is to decide which
datasets should be used during the analysis. We rely on the descriptions of the
datasets and advices from domain experts to determine which dataset(s) may
be relevant to the component under study.
The second task consists of retrieving all relevant instances from the selected
dataset(s). Figure 2 illustrates the process followed to retrieve these instances.
The
30. rst and the most dicult challenge of the Data Gathering step is to retrieve
the required information about all occurrences of replacement of a given compo-
8
31. nent. The information is retrieved from a database which contains descriptions
of all maintenance actions performed on the aircraft. These descriptions are
provided by the engineers and technicians who perform the actions. Examples
of information used to describe the maintenance actions are: the date on which
the action has been performed, the identi
33. er(s)
for part(s)removed and/or installed, and textual explanation of the problem
and work carried out. Ideally, a simple query returning all maintenance reports
for which the Part Removed identi
35. er for the compo-
nent of interest will be enough to return all the desired information about all
replacement occurrences.
Unfortunately, this is not possible because the
41. er. This implies that
the textual explanation of the repairs also need to be processed in order to
42. nd
all occurrences of replacement of a given component. Automatic processing
of the textual explanations is very dicult because of the open language used
by the maintenance sta. There are many abbreviations and acronyms, major
grammatical and syntactical problems (e.g. no delimiters between prepositions,
absence of verbs, absence of articles and complements), typing mistakes, and
inconsistent use of abbreviations among reports. Manual analysis of the texts is
not a viable solution either, there are simply too many reports that would need
to be analyzed (more than 65000).
The approach we adopt to address this problem is represented by the
43. rst
three processes in Figure 2. This approach is built around the Information Re-
9
44. trieval paradigm. First, we query the maintenance database in order to retrieve
the information about all replacements for which the part removed identi
46. er of the part of interest. Second, we use a tuned version
of a keyword generation system called Extractor [2] to extract the key phrases
from the textual explanations for these replacements. These key phrases are
then used to extend the initial query which was uniquely based on the part
identi
47. er. The reasoning behind this approach is as follow: if the key phrases
obtained correspond to the vocabulary used by the maintenance sta to talk
about the component of interest, then these key phrases are very likely to help
in
54. eld or any of the these three key phrases in the textual
descriptions. As shown in Figure 2, this will probably add a number of new
maintenance reports to the initial ones obtained from the
55. rst query. These
new reports may talk about the component of interest but without necessarily
referring to its replacement. A manual validation of the new maintenance re-
ports is required in order to take out the reports not describing a replacement
of the given component. Note that the number of reports that require manual
validation is very small in comparison with the overall number of maintenance
reports available in the database. In that sense, the proposed approach helps in
reducing the amount of manual analysis by automatically pre-selecting poten-
tially related reports. On the other hand, there is no guarantee that all related
reports will be selected by the proposed mechanism. This depends on the com-
10
56. pleteness of the set of key phrases outputs by the Information Retrieval system.
To evaluate the coverage of the proposed method, we asked a domain specialist
to comment on the number of replacements found for 5 randomly selected com-
ponents. In all cases, the expert concluded that the number of replacements
found with our approach is very close to the exact number of replacements. In
other words, for these 5 components, all or almost all occurrences of replacement
have been retrieved by the proposed approach.
Once we have the date and aircraft identi
57. er for each replacement of the
component of interest, we can retrieve the relevant instances from the selected
dataset(s). We consider relevant all instances obtained around the time of the
replacement. In particular, for each selected dataset and for each occurrence
of replacement, we retrieve the data obtained during m days prior to the re-
placement and n days after the replacement. The numbers m and n depend on
the datasets and the component of interest. Our strategy is to select m so that
we have at least 200 instances available for learning the patterns. For instance,
with a dataset for which we have an average of two reports per aircraft per day,
m will be set to be at least 100. For n, we simply set it as n = :15 m. As we
discuss in Section 5 (Data Labeling), the selected values for m and n in
uenced
the setting of the k parameter.
In our approach, we use days as time metric instead of cycles (pairs of
takeos and landings) or hours of operation mainly because the maintenance
sta which are the target users prefer to obtain predictions about the need for
replacement in days. From a data analysis point of view, this choice may not
seem appropriate because one would think that repairs are more related to the
use of a plane than to time. We veri
59. Select data around ea failure
from relevant dataset(s)
Add parameters: “Problem Id” and
“time from replacement”
Aircraft
datasets
Select * where
PartOff# =
target#
Find keyphrases
from descriptions
Extend query
Manual
validation
Details about
replacements
Dataset(s) for ...
learning
Repair
Descriptions
Figure 2: Retrieving relevant data for the analysis.
using cycles instead of days, but the results were quite similar. Consequently,
we decided to only use days.
Finally, two new attributes are added to the initial raw measurements: the
time between the observation is collected and the actual replacement time (time
from failure), and a tag identifying each observation with a speci
60. c replacement
case (problem Id). The instance from a given selected dataset are combined
to create a dataset that will be used to build the predictive model. The number
of output datasets is equal to the number of dataset(s)selected as potentially
useful for the given component.
5 Data Labeling
As explained earlier, the data mining techniques we intend to use to build the
models fall into the category of supervised learning algorithms. These algo-
12
61. rithms require the input dataset to contain a class attribute (also called label or
membership attribute). In our case, the class attribute may take two dierent
values (0 indicating no need for replacement of the component and 1 indicat-
ing a need). In our approach, we set the class attribute to 1 for all instances
obtained between the time of the replacement and the preceding k days (these
k days de
62. ne the window that we target for the predictions of the need for
component replacement), and set the class attribute to 0 for all other instances
observed outside that period of time. Obviously, k has to be smaller than m
(see Section 4). Figure 3 presents the label value obtained from the labeling
algorithm as a function of the number of days from replacement.
The interval for positive examples is predetermined for each component of
interest. The choice for k is a function of:
1. Target period for prediction of the need for replacement of the component
of interest. Dierent components may have dierent target periods.
2. Proportion of positive and negative examples. Without a signi
63. cant pro-
portion of positive examples, several data mining approaches are expected
to have diculties learning the desired models. To obtain at least 10% of
positive examples, we use the following constraint k :1 (m+ n).
3. The complexity of the patterns to model. Some components may start
giving indications of deterioration a while before they actually need to be
replaced. The labeling approach should be adjusted accordingly.
For each component studied, we have performed a number of experiments with
dierent values of k in order to satisfy the above criteria. Values obtained are
between 10 and 40, depending on the component and the dataset used.
13
64. Figure 3: Labeling the instances.
6 Building and Evaluating models
The datasets obtained after the application of the labeling process can be used
to learn the desired models. Several data mining techniques are available to
infer these models such as Decision Tree, Instance Based, Naive-Bayes, Rough
Sets, Regression, and Neural Networks. Depending on the technique used, one
may need to further pre-process the data in order to: (i) select the most suit-
able subset of attributes, (ii) normalize the initial attributes, (iii) create new
attributes from the initial ones, or (iv) discretize the continuous attributes. Ap-
proaches to perform these tasks are proposed in dierent areas of research such
as machine learning, statistics, pattern recognition, and data mining [3].
A critical issue in any data mining application is the evaluation of the learned
models. In order to decide on the suitability of a model we need to get a good
estimate of its expected performance on operational data (i.e. new data that
will be provided as input to the model once it has been deployed). There are a
14
65. number of methods to compute the expected performance; these include hold-
out validation, cross-validation, and bootstrapping.
Hold-out validation consists of randomly splitting the data into a training
and a testing set. In this approach, one would
66. rst use the training set to build
the model, and then estimate its performance by using the model to classify the
instances in the testing set. The performance on the testing set becomes the
expected performance of the model on new cases. This method considerably
reduces the number of instances available to build your models. The method is
therefore unsuitable when only a small dataset is available. Hold-out validation
assumes that the instances are independent of each other.
Cross-validation makes use of the available data in a more ecient manner
than hold-out validation. Cross-validation starts by randomly splitting the data
into k subsets of approximately equal size. Then, you train the model using k1
subsets and evaluate it on the remaining one. The previous step is repeated
until each subset has been used for testing once. The expected performance
is
67. nally computed from the testing results obtained from the dierent runs.
When k equals the number of instances, the process is called leave-one-out
cross-validation. Cross-validation requires more computer time than hold-out
validation but provides a more robust estimation of the expected performance.
Bootstrapping is even more demanding in terms of computation time than
cross-validation. In its simplest form, you repeat the analysis (train and vali-
date) by randomly sampling (with replacement) the test and training instances
from the population. Thousands of repetitions are usually required. Experimen-
tal comparisons of bootstrapping and cross-validation tend to show that boot-
strapping is better than cross-validation with some learning approaches such as
15
69. cantly worse than cross-validation with other
learning approaches such as decision tree [5].
Unfortunately, none of these three approaches is adequate for the applica-
tion considered here. All of these approaches rely on random sampling at some
point to select instances from the population. The problem with random sam-
pling is that it implicitly assumes that the instances are independent. Whenever
this assumption is not too severely violated, repeated analysis methods such as
cross-validation and bootstrapping will probably lead to reasonably good results.
However, with our application this assumption is simply not viable. Because
of variations in the construction, operation, and maintenance of the aircraft,
it is clear that any two instances from a given aircraft are not as independent
as any two instances from two dierent aircraft. This means that if we use
data from the same aircraft for training and validation the likelihood of success
would be much higher than if we validate using data from a dierent aircraft.
Similar reasoning applies to the data related to a given occurrence of component
replacement versus data related to dierent occurrences of component replace-
ment; two instances related to a given occurrence of a problem will be more
related than two instances related to two dierent problems. Considering the
fact that the number of occurrences of component replacement is signi
70. cantly
less then the total number of instances, random sampling is very likely to gener-
ate training and validation sets that contain data related to the same occurrence
of component replacement. The end result will be an over optimistic evaluation
of the performance of the model.
A simple solution to this problem consists of splitting the data such that the
training instances come from a sub-group of occurrences of component replace-
16
71. ment and validation instances come from another sub-group of replacement
occurrences. This splitting is easily achieved using the Problem Id attribute
added during the Data Gathering process.
A more robust estimate could be obtained by adapting the cross-validation
method so that it accounts for the above splitting constraint. The process is as
follow:
1. Split the data into batches (one for each failure case). This is done using
the parameter Problem Id.
2. Keep one batch for validation and train using all other batches. Repeat
this step until each batch has been used for testing once.
3. Get
72. nal evaluation from validation results over the dierent runs.
We have implemented this approach using the SAS system and MLC++ [6]. The
term LOBO (leave-one-batch-out) has been used to describe a similar evaluation
method in [7]. We have been using this method to evaluate the performance of
a variety of learning algorithms for dierent aircraft components.
6.1 Evaluation function
In the previous section, we explained a method to come up with a fair estimate
for the performance of a model, but we did not explain the actual performance
criteria we are trying to optimize. In machine learning research, the reliability
of a classi
73. er is often summarized by either its error-rate or accuracy. The
error-rate is de
79. er reliability proposed the use of ROC
(Receiver Operating Characteristics) curves [8]. Other related metrics from the
information retrieval community are the precision and recall. In this context,
the recall is de
80. ned as the ratio of relevant documents retrieved for a given
query over the number of relevant documents in the database, and the precision
is the ratio of relevant documents retrieved over the total number of documents
retrieved.
Unfortunately, all of these metrics fail to capture two important aspects
of our application. The
81. rst one is the relation between the usefulness of a
prediction and the time that separates it from the replacement. A too early
warning about a potential failure will lead to a non-optimal use of the component
while a too late advice may not let enough time for proper planing of the repair.
What we need for this application is an evaluation method that takes into
account the timeliness of the alerts. The second aspect is related to the coverage
of potential failures. Because the learned model would be used to classify each
report coming from the aircraft into one of the two possible categories (the
component needs to be replaced or not), there is a potential for a model to
generate several alerts before the component was replaced. More alerts may
sometimes mean a higher con
82. dence in the prediction. However, it is clear
that we prefer a model that can generate at least one alert for most of the
component failures, than a model that generates several alerts for only few
failure cases. In oder words, the coverage of the model is very important because
we want to minimize the number of unexpected failures. To take into account
this preference, we need an overall scoring metric that considers the distribution
18
83. of the alerts over the various failure cases. The next section presents a reward
function to take into account the timeliness of the alerts while the following
section introduces a new scoring metric that addresses the coverage issue.
6.2 Reward function
The criterion we have developed is based on the notion of reward. Predicting the
correct outcome for an instance generates a reward. In its basic form, reward
thresholds are
85. ned for
each possible class (or membership) value. In our application, we have extended
this framework to accept: i) varying reward thresholds for dierent instances
and ii) any real number as reward threshold. In particular, for each component
of interest, we de
86. ne a function that computes the reward for prediction of
a positive instance based on the number of days between the time at which
the instance is generated and the actual time of the replacement. Figure 4
presents the graph of such a function. For this example, the maximum gain
is obtained when predicting the need for replacement between 5 and 40 days
prior to the replacement of the component. We also observed that predicting a
need for replacement of the component outside that target period may lead to
a negative reward threshold since such a prediction corresponds to a misleading
advice. According to this function, false positive predictions are penalized by a
reward of -1.5 in comparison to a reward of 1.0 for true positive predictions.
We note that the target period for the reward function does not need to be
the same as the labeling period, but the two are de
88. rst determine the target period for the reward function according to the user
requirements and then we
89. x the labeling period. For instance, if the user says
that predictions for a given component should be within 1-3 weeks in advance
19
90. Reward thresholds for positive predictions over time from replacement
Time from
replacement
1
-1.5
-60 -30 -5 0
Reward
Figure 4: Example of a reward function for prediction of positive instances.
then we set the target period for the reward function between -7 to -21 days.
Then we experiment with various labeling periods around this target period and
take the one that leads to the best result. Setting the labeling period as the
target period is often a good strategy.
The reward function illustrated in Figure 4 follows a piecewise
91. rst order
polynomial model. Such a model is convenient because it is comprehensible
for domain people and suciently complex to capture the relevant information.
There are several ways to improve the precision of the function. One possibility
is to use higher order polynomials instead of straight lines. As another possi-
bility, one could try to smooth the overall function. However, domain experts
often think that the increase in complexity is not justi
92. ed.
6.3 Scoring metric
The above reward assignment method accounts for the timeliness of the alerts.
However, to evaluate the coverage of a model, we also need to look at the
distribution of the alerts over the dierent failure cases. The overall performance
metric we propose to evaluate a model is:
score = (
p
X
i=1
scorei) (NbrDetected=NbrOfCases)SignOfSumOfScores (1)
20
93. where p is the number of positive predictions made by the model on validation
instances during the application of the LOBO evaluation method, scorei is
the score from the reward function for the ith instance classi
94. ed as positive,
NbrDectected is the number of replacement cases for which we have at least 1
positive prediction in the target interval (e.g -30 to -5 days), NbrOfCases is the
total number of replacement occurrences we have for the given component, and
SignOfSumOfScores is the sign of the
95. rst term in the expression. The term
(NbrDetected=NbrOfCases)SignOfSumOfScores is introduced in the evaluation
function in order to favor models that optimize the recall1. Let us observe that
the value ofPp
i=1 scorei makes the
96. nal score sensitive to precision of the model.
7 Does it work? Experimental results
In the previous sections we presented solutions to fundamental issues in applying
data mining techniques to predict the need for aircraft component replacement.
A possible approach to evaluate the proposed solutions would be to compare
each solution individually with alternative methods proposed in the literature.
This is not realistic in our application for various reasons. First, we are ad-
dressing problems for which it is very dicult to
97. nd comparable methods in
the literature. For instance, to our knowledge no one has proposed methods
for data gathering and data labeling that can substitute the ones we proposed.
Moreover, it is important to note that the proposed methods are very related.
Therefore, evaluating solutions individually would not necessarily shed light on
the usefulness of the overall approach proposed. These observations lead us to
a dierent evaluation strategy in which we globally evaluate the potential of
1The term (NbrDetected=NbrOfCases)SignOfSumOfScores is set to zero when
NbrDetected = 0 and SignOfSumOfScores is negative
21
98. the proposed approach by applying it in a large-scale experiment. In this ex-
periment, we test the potential of the approach to learn models for 16 aircraft
engine components.
Since the aim was not to develop optimal models, we decided to simplify the
process in various ways. First, we did apply pre-processing techniques usually
required in order to optimize the results such as: cleaning the data by removing
outliers, normalization of the data, creation of new parameters, and selection
of the best subset of parameters (the same subset of attributes has been used
for all learning approaches and all components). We only used one potentially
useful dataset. There are at least 4 other relevant datasets that could have
been used to infer models for these components. Moreover, the same settings
have been used for the various parameters (i.e. m = 100,n = 30, and k = 20).
Optimization of these parameters could have signi
99. cantly improved the results.
For each of the 16 components, we have been using three dierent data
mining techniques to learn the desired models: decision tree with C4.5 [9],
Instance-Based (one nearest-neighbor) and Naive Bayes both implemented in
MLC++ [6]. The evaluation has been performed using the methods described
in Section 6. We used the same reward function for all components. The selected
reward function is very similar to the one presented in Figure 4.
Table 1 presents the results obtained from this large-scale experimentation.
The
101. cation. For each part, we ran the
three data mining approach as indicated in the second column. The third col-
umn indicates the number of folds used during the cross-validation experiment.
As explained in Section 6, the number of folds used is equal to the number of
occurrences of replacement we have found during the data gathering step. Fi-
22
102. nally, the fourth column shows the overall score of each model on the testing
data. This score comes from the formula presented in Section 6.3.
Even if the data mining process has been substantially simpli
103. ed during
this experiment, we observe very interesting results. First, we note that we
got at least one positive score for 112 components out of the 16 under study.
This result clearly demonstrates the potential of the approach. Second, from
the scores we can see that no classi
104. er outperforms the other two for all com-
ponents. This supports the idea that experiments with several data mining
approaches are generally required in order to
105. nd the most suitable approach
for the problem addressed. In our experiment, it is just like we have investigated
16 dierent tasks (one for each component) and each of these is likely to have
its own optimal con
106. guration in terms of techniques and parameter settings.
It is therefore reasonable to think that these already promising results could be
further improved.
An interesting aspect of the results obtained from the various classi
108. cation errors seem to be independent. In
fact, correlation analysis of the predictions obtained from the three classi
109. ers
used lead to the conclusion that the dierent approaches tend to make mistakes
on dierent cases. This observation is very important because it indicates a po-
tential for improving the performance by combining predictions from the various
classi
110. ers. One should also remember that valuable models can also be devel-
oped using the other potentially relevant dataset(s) available and, therefore,
further enhance the overall performance of the model.
2excluding the score value of 0:2 obtained for part 12
23
111. Part Id Approach Nbr of folds Score
part 1 ib 16 16.9
c4.5-rules 16 -33.8667
naive-bayes 16 -103.383
part 2 ib 25 16.1133
c4.5-rules 25 6.468
naive-bayes 25 -111.813
part 3 ib 17 -22.5
c4.5-rules 17 -34.5857
naive-bayes 17 -75.6533
part 4 ib 36 140.195
c4.5-rules 36 89.5222
naive-bayes 36 -18.6225
part 5 ib 27 48.2733
c4.5-rules 27 9.87677
naive-bayes 27 -11.3261
part 6 ib 19 77.2133
naive-bayes 19 48.8168
c4.5-rules 19 15.5684
part 7 ib 31 10.1678
c4.5-rules 31 -1.67809
naive-bayes 31 -45.5741
part 8 c4.5-rules 41 -13.6536
naive-bayes 41 -32.4486
ib 41 40.2209
part 9 naive-bayes 12 9.9267
c4.5-rules 12 2.98223
ib 12 12.6
part 10 ib 28 42.16
naive-bayes 28 0.52143
c4.5-rules 28 -3.06133
part 11 naive-bayes 2 0.00
c4.5-rules 2 -2.1467
ib 2 -5.5667
part 12 c4.5-rules 16 0.2
ib 16 -4.3733
naive-bayes 16 -80.96
part 13 ib 53 48.088
naive-bayes 53 1.45456
c4.5-rules 53 -110.593
part 14 naive-bayes 40 -15.2416
ib 40 -18.8133
c4.5-rules 40 -43.8596
part 15 c4.5-rules 26 9.12
naive-bayes 26 2.98458
ib 26 -69.68
part 16 ib 31 99.0959
c4.5-rules 24
31 9.40712
naive-bayes 31 -40.0467
Table 1: Prediction results for 16 components from 3 data mining approaches
obtained using one dataset. High score values mean good performance.
112. 8 Related work
In terms of the tasks addressed, the closest works are found in the area of Re-
liability Analysis. These works concentrate on the development of models to
assess the reliability of individual units or complex systems. The models devel-
oped may then be used to predict the probability that a speci
113. c component(or
system) will fail within a given period of time. Predicting the failure of a com-
ponent is obviously very similar to the prediction of the need for replacement
of a component.
In reliability analysis, the data is usually obtained through a planned ex-
periment in which one observes the evolution of a sample of units and record
the failure times. The analysis almost always starts by
114. tting the failure time
values to a known probability distribution. Knowledge about this distribution
is then used to infer models to predict the time to failure of similar units. The
approach introduced in this paper diers considerably from this process. First,
the data available in our application does not come from an experiment and we
have no control on the variables measured and on the quality of the data re-
ceived. As a consequence, major pre-processing work (as explained in Sections 4
and 5) had to performed in order to retrieve the desired data for the analysis.
Second, the parameter time from replacement does not play such an important
role during the modeling step. On the other hand, we extensively use it during
pre-processing of the data. This was required in order to compensate for the fact
that most data mining approaches are not designed to bene
115. t from time infor-
mation. Finally, the evaluation methods are also dierent. In our application,
we rely on a novel metric that meets speci
116. c domain requirements to evaluate
the models. In reliability analysis, the evaluation is mostly based on traditional
25
117. statistical measures such as mean-squared error. Crowder et al. [10] present an
up-to-date account of the analysis of reliability data. Researchers in the area of
Survival Analysis make use of techniques similar to the ones
118. nd in reliability
analysis, but mainly in the context of medical and biological applications [11].
There has been research on applying data mining techniques to identify
problems with aircraft gas turbine (e.g. [12, 13, 14]). These works are, however,
limited to the identi
119. cation of current problem(s) with the engine and do not
try to predict potential problems in advance. The data used is also dierent;
our approach makes use data obtained during normal operation of the engines
while the above works are based on data acquired in engine test cells. The
applicability of the results is therefore very dierent.
9 Conclusion and further work
We have presented an application in which data mining techniques are applied
on operational and maintenance data. Our work has been data-driven. Hav-
ing decided the kind of model to be developed (a classi
120. er), we have focused
on addressing the complex characteristics of the data. Our data is hetero-
geneous (dierent data sets, some data is numerical, some is text, and some
needs to be interpreted as messages in natural language). The data is also
time-sensitive (most of it comes with a time-stamp which is relevant for model
building). Moreover, the data is unprepared for the development of the model:
it lacks class labels or ranking. Finally, it was clear from the outset that ad-
equate methods for the evaluation of models for this kind of application were
lacking.
The contribution of this paper is therefore in the areas of data gathering,
26
121. data labeling, and in its novel approach to evaluation. In data gathering, we
show how a simple process based on the use of o-the-shelf NLP tools can help
with data cleaning and integration. In data labeling, we propose an empirical,
time-sensitive approach which trades o the proximity of an alarm to the time
of replacement against the requirement of a reasonable number of training in-
stances. In evaluation, we suggest an evaluation function which combines the
timeliness of an alarm, the recall of positive training instances, as well as their
precision.
In the course of this project, we have produced several dierent models
(alarm systems) for our application. Originally, we expected to experiment
with dierent models and choose the one with the best performance. However,
our work to date has shown that it is more appropriate to consider model fusion.
We are currently working on model fusion, combining all or some of the models
in order to improve the performance of the resulting classi
122. er. In this work, we
are exploring some of the ideas on the ensembles of classi
123. ers [15]. Since, as
mentioned in Section 6 , the errors of dierent classi
124. ers are uncorrelated, we
expect an improved performance from a set of combined classi
125. ers.
Acknowledgment The authors are grateful to Air Canada for supplying the
data and explaining dierent aspects of aircraft health monitoring, and to the
Natural Sciences and Engineering Resarch Council of Canada for the support
to the
126. rst and third author.
References
[1] R.O. Duda and P.E. Hart. Pattern Classi
128. [2] P. Turney. Extraction of keyphrases from text: evaluation of four al-
gorithms. Technical Report NRC 41550, National Research Council of
Canada, 1997.
[3] D. Pyle. Data Preparation for Data Mining. Morgan Kaufmann, San
Mateo, CA, 1999.
[4] B. Efron. Estimating the error rate of a prediction rule: Improvement on
cross-validation. Journal of the American Statistical Association, 78:316{
331, 1983.
[5] L. Breiman, J. Friedman, R. Olshen, and C. Stone. Classi
129. cation and
Regression Trees. Wadsworth International Group, 1984.
[6] R. Kohavi, D. Sommer
130. eld, and J. Dougherty. Data mining using MLC++:
A machine learning library in C++. In Tools with Arti
131. cial Intelligence,
pages 234{245. IEEE Computer Society Press, 1996. Received the best
paper award.
[7] M. Kubat, R.C. Holte, and S. Matwin. Machine learning for the detection
of oil spills in satellite radar images. Machine Learning, 30:195{215, 1998.
[8] F. Provost and T. Fawcett. Analysis and visualization of classi
132. er perfor-
mance: Comparison under imprecise class and cost distributions. In David
Heckerman, Heikki Mannila, Daryl Pregibon, and Ramasamy Uthurusamy,
editors, Proceedings of the Third International Conference on Knowledge
Discovery and Data Mining, page 43. AAAI Press, 1997.
[9] J.R. Quinlan. C4.5: Programs for Machine Learning. Morgan Kaufmann,
San Mateo, CA, 1993.
28
133. [10] M.J. Crowder, A.C. Kimber, R.L. Smith, and T.J. Sweeting. Statistical
Analysis of Reliability Data. Chapman Hall, London, 1991.
[11] D. Collett. Modelling Survival Data in Medical Research. Chapman Hall,
London, 1994.
[12] T.W. Brotherton and G. Charddedon. Automated rule extraction for engine
heath monitoring. In Evolutionary Programming VII: 7th International
Conference, EP98, pages 725{734, 1998.
[13] P. Turney and M. Halasz. Contextual normalization applied to aircraft gas
turbine engine diagnosis. Applied Intelligence, 1993.
[14] W.E. Dietz, E.L. Kiech, and M. Ali. Classi
134. cation of data patterns using
an autoassociative neural network topology. 1989.
[15] T.G. Dietterich. Machine learning research: Four current directions. AI
Magazine, 18(4):97{136, 1997.
29