This paper presents the MG-CHARM algorithm for mining association rules from transactional datasets. MG-CHARM aims to improve upon the Apriori and CHARM algorithms by mining only the minimal generators of frequent closed itemsets, rather than all frequent itemsets, resulting in faster performance. The algorithm works by first identifying frequent closed itemsets that satisfy a minimum support threshold, then mining just the minimal generators of those closed itemsets to derive association rules. An experimental evaluation shows MG-CHARM performs faster and uses less memory than Apriori and CHARM for the same datasets.
Multiple Minimum Support Implementations with Dynamic Matrix Apriori Algorith...ijsrd.com
Data mining can be defined as the process of uncovering hidden patterns in random data that are potentially useful. The discovery of interesting association relationships among large amounts of business transactions is currently vital for making appropriate business decisions. Association rule analysis is the task of discovering association rules that occur frequently in a given transaction data set. Its task is to find certain relationships among a set of data (itemset) in the database. It has two measurements: Support and confidence values. Confidence value is a measure of rule’s strength, while support value corresponds to statistical significance. There are currently a variety of algorithms to discover association rules. Some of these algorithms depend on the use of minimum support to weed out the uninteresting rules. Other algorithms look for highly correlated items, that is, rules with high confidence. Traditional association rule mining techniques employ predefined support and confidence values. However, specifying minimum support value of the mined rules in advance often leads to either too many or too few rules, which negatively impacts the performance of the overall system. This work proposes a way to efficiently mine association rules over dynamic databases using Dynamic Matrix Apriori technique and Multiple Support Apriori (MSApriori). A modification for Matrix Apriori algorithm to accommodate this modification is proposed. Experiments on large set of data bases have been conducted to validate the proposed framework. The achieved results show that there is a remarkable improvement in the overall performance of the system in terms of run time, the number of generated rules, and number of frequent items used.
IRJET- Fault Detection and Prediction of Failure using Vibration AnalysisIRJET Journal
This document discusses fault detection and prediction of failures in rotating equipment using vibration analysis. It begins by introducing vibration analysis as a method to monitor machines and detect faults in rotating components that may cause failures. It then discusses how motor vibration is measured and analyzed using techniques like spectrum analysis to identify faults like unbalance, bearing issues, or broken rotor bars. The document proposes decomposing vibration signals using intrinsic mode functions and calculating the Gabor representation's frequency marginal to identify fault types using classifiers like support vector machines or random forests. It provides context on data mining techniques relevant to this type of fault prediction problem.
A Survey on the Clustering Algorithms in Sales Data MiningEditor IJCATR
This document discusses clustering techniques that can be used for analyzing sales data. It begins by introducing the importance of clustering large sales databases to extract useful knowledge that can help senior management with decision making. It then provides an overview of different clustering algorithms like hierarchical, partitioning, grid-based, and density-based clustering. The document also discusses the goals of clustering sales data, which include predicting customer purchasing behavior and improving knowledge discovery. It outlines the typical stages of sales data clustering as feature selection, validation of results, and interpretation of results. Finally, it reviews several papers that have used clustering and other techniques like association rule mining to analyze retail sales data.
Study of Data Mining Methods and its ApplicationsIRJET Journal
This document discusses data mining methods and their applications. It begins by defining data mining as the process of extracting useful patterns from large amounts of data. The document then outlines the typical steps in the knowledge discovery process, including data selection, preprocessing, transformation, mining, and evaluation. It classifies data mining techniques into predictive and descriptive methods. Specific techniques discussed include classification, clustering, prediction, and association rule mining. Finally, the document discusses applications of data mining in fields like healthcare, biology, retail, and banking.
SURVEY ON CLASSIFICATION ALGORITHMS USING BIG DATASETEditor IJMTER
Data mining environment produces a large amount of data that need to be analyzed.
Using traditional databases and architectures, it has become difficult to process, manage and analyze
patterns. To gain knowledge about the Big Data a proper architecture should be understood.
Classification is an important data mining technique with broad applications to classify the various
kinds of data used in nearly every field of our life. Classification is used to classify the item
according to the features of the item with respect to the predefined set of classes. This paper put a
light on various classification algorithms including j48, C4.5, Naive Bayes using large dataset.
A Trinity Construction for Web Extraction Using Efficient AlgorithmIOSR Journals
This document describes a proposed system for web extraction using a trinity construction and efficient algorithms. It begins with an abstract discussing how trinity characteristics can be used to automatically extract content from websites in a sequential tree structure. It then discusses the existing system which uses trinity tree and prefix/suffix sorting but has limitations. The proposed system introduces fuzzy logic for multi-perspective crawling across multiple websites. A genetic algorithm is used to load extracted content into the trinity structure and remove unwanted data. Finally, an ant colony optimization algorithm is used to obtain an effective structure and suggest optimized solutions.
Review on: Techniques for Predicting Frequent Itemsvivatechijri
Electronic commerce(E- Commerce) is the trading or facilitation of trading in products or services
using computer networks, such as the Internet. It comes under a part of Data Mining which takes large amount
of data and extracts them. The paper uses the information about the techniques and methods used in the
shopping cart for prediction of product that the customer wants to buy or will buy and shows the relevant
products according to the cost of the product. The paper also summarizes the descriptive methods with
examples. For predicting the frequent pattern of itemset, many prediction algorithms, rule mining techniques
and various methods have already been designed for use of retail market. This paper examines literature
analysis on several techniques for mining frequent itemsets.The survey comprises various tree formations like
Partial tree, IT tree and algorithms with its advantages and its limitations.
A model for profit pattern mining based on genetic algorithmeSAT Journals
Abstract
Mining profit oriented patterns is a novel technique of association rule mining in data mining, which basically focuses on important issues related with business. As it is well known that every business aims to generate the profit and find the ways to improve the same. In earlier days association rule mining was used for market basket analysis and targeted only some of the business and commercial aspects. Afterwards the researchers started to aim the most prominent element of any business i.e. Profit, and determined the innovative way to generate the association rules based on profit. Profit oriented patterns mining approach combines the statistic based pattern mining with value-based decision making to generate those patterns with the maximum profit and some ways to generate recommenders for future strategy. To achieve the desired goal the traditional association rule mining alone is not effectual, so we combine the strength of genetic algorithm with association rule mining to enhance its capability. The study shows that Genetic Algorithm improves the effectiveness and efficiency of association rule mining outcome, since genetic algorithms are competent to handle the problems related with the uncertainty, multi-dimensional, non-differential, non-continuous, and non-parametrical, non-linearity constraint and multi-objective optimization problems. In this paper we apply the concept of profit pattern mining with genetic algorithm to generate profit oriented pattern which help out in future business expansion and fulfill the business objective.
Keywords: Data Mining, Association Rule Mining, Profit Pattern Mining, Genetic Algorithm
Multiple Minimum Support Implementations with Dynamic Matrix Apriori Algorith...ijsrd.com
Data mining can be defined as the process of uncovering hidden patterns in random data that are potentially useful. The discovery of interesting association relationships among large amounts of business transactions is currently vital for making appropriate business decisions. Association rule analysis is the task of discovering association rules that occur frequently in a given transaction data set. Its task is to find certain relationships among a set of data (itemset) in the database. It has two measurements: Support and confidence values. Confidence value is a measure of rule’s strength, while support value corresponds to statistical significance. There are currently a variety of algorithms to discover association rules. Some of these algorithms depend on the use of minimum support to weed out the uninteresting rules. Other algorithms look for highly correlated items, that is, rules with high confidence. Traditional association rule mining techniques employ predefined support and confidence values. However, specifying minimum support value of the mined rules in advance often leads to either too many or too few rules, which negatively impacts the performance of the overall system. This work proposes a way to efficiently mine association rules over dynamic databases using Dynamic Matrix Apriori technique and Multiple Support Apriori (MSApriori). A modification for Matrix Apriori algorithm to accommodate this modification is proposed. Experiments on large set of data bases have been conducted to validate the proposed framework. The achieved results show that there is a remarkable improvement in the overall performance of the system in terms of run time, the number of generated rules, and number of frequent items used.
IRJET- Fault Detection and Prediction of Failure using Vibration AnalysisIRJET Journal
This document discusses fault detection and prediction of failures in rotating equipment using vibration analysis. It begins by introducing vibration analysis as a method to monitor machines and detect faults in rotating components that may cause failures. It then discusses how motor vibration is measured and analyzed using techniques like spectrum analysis to identify faults like unbalance, bearing issues, or broken rotor bars. The document proposes decomposing vibration signals using intrinsic mode functions and calculating the Gabor representation's frequency marginal to identify fault types using classifiers like support vector machines or random forests. It provides context on data mining techniques relevant to this type of fault prediction problem.
A Survey on the Clustering Algorithms in Sales Data MiningEditor IJCATR
This document discusses clustering techniques that can be used for analyzing sales data. It begins by introducing the importance of clustering large sales databases to extract useful knowledge that can help senior management with decision making. It then provides an overview of different clustering algorithms like hierarchical, partitioning, grid-based, and density-based clustering. The document also discusses the goals of clustering sales data, which include predicting customer purchasing behavior and improving knowledge discovery. It outlines the typical stages of sales data clustering as feature selection, validation of results, and interpretation of results. Finally, it reviews several papers that have used clustering and other techniques like association rule mining to analyze retail sales data.
Study of Data Mining Methods and its ApplicationsIRJET Journal
This document discusses data mining methods and their applications. It begins by defining data mining as the process of extracting useful patterns from large amounts of data. The document then outlines the typical steps in the knowledge discovery process, including data selection, preprocessing, transformation, mining, and evaluation. It classifies data mining techniques into predictive and descriptive methods. Specific techniques discussed include classification, clustering, prediction, and association rule mining. Finally, the document discusses applications of data mining in fields like healthcare, biology, retail, and banking.
SURVEY ON CLASSIFICATION ALGORITHMS USING BIG DATASETEditor IJMTER
Data mining environment produces a large amount of data that need to be analyzed.
Using traditional databases and architectures, it has become difficult to process, manage and analyze
patterns. To gain knowledge about the Big Data a proper architecture should be understood.
Classification is an important data mining technique with broad applications to classify the various
kinds of data used in nearly every field of our life. Classification is used to classify the item
according to the features of the item with respect to the predefined set of classes. This paper put a
light on various classification algorithms including j48, C4.5, Naive Bayes using large dataset.
A Trinity Construction for Web Extraction Using Efficient AlgorithmIOSR Journals
This document describes a proposed system for web extraction using a trinity construction and efficient algorithms. It begins with an abstract discussing how trinity characteristics can be used to automatically extract content from websites in a sequential tree structure. It then discusses the existing system which uses trinity tree and prefix/suffix sorting but has limitations. The proposed system introduces fuzzy logic for multi-perspective crawling across multiple websites. A genetic algorithm is used to load extracted content into the trinity structure and remove unwanted data. Finally, an ant colony optimization algorithm is used to obtain an effective structure and suggest optimized solutions.
Review on: Techniques for Predicting Frequent Itemsvivatechijri
Electronic commerce(E- Commerce) is the trading or facilitation of trading in products or services
using computer networks, such as the Internet. It comes under a part of Data Mining which takes large amount
of data and extracts them. The paper uses the information about the techniques and methods used in the
shopping cart for prediction of product that the customer wants to buy or will buy and shows the relevant
products according to the cost of the product. The paper also summarizes the descriptive methods with
examples. For predicting the frequent pattern of itemset, many prediction algorithms, rule mining techniques
and various methods have already been designed for use of retail market. This paper examines literature
analysis on several techniques for mining frequent itemsets.The survey comprises various tree formations like
Partial tree, IT tree and algorithms with its advantages and its limitations.
A model for profit pattern mining based on genetic algorithmeSAT Journals
Abstract
Mining profit oriented patterns is a novel technique of association rule mining in data mining, which basically focuses on important issues related with business. As it is well known that every business aims to generate the profit and find the ways to improve the same. In earlier days association rule mining was used for market basket analysis and targeted only some of the business and commercial aspects. Afterwards the researchers started to aim the most prominent element of any business i.e. Profit, and determined the innovative way to generate the association rules based on profit. Profit oriented patterns mining approach combines the statistic based pattern mining with value-based decision making to generate those patterns with the maximum profit and some ways to generate recommenders for future strategy. To achieve the desired goal the traditional association rule mining alone is not effectual, so we combine the strength of genetic algorithm with association rule mining to enhance its capability. The study shows that Genetic Algorithm improves the effectiveness and efficiency of association rule mining outcome, since genetic algorithms are competent to handle the problems related with the uncertainty, multi-dimensional, non-differential, non-continuous, and non-parametrical, non-linearity constraint and multi-objective optimization problems. In this paper we apply the concept of profit pattern mining with genetic algorithm to generate profit oriented pattern which help out in future business expansion and fulfill the business objective.
Keywords: Data Mining, Association Rule Mining, Profit Pattern Mining, Genetic Algorithm
PERFORMING DATA MINING IN (SRMS) THROUGH VERTICAL APPROACH WITH ASSOCIATION R...Editor IJMTER
This system technique is used for efficient data mining in SRMS (Student Records
Management System) through vertical approach with association rules in distributed databases. The
current leading technique is that of Kantarcioglu and Clifton[1]. In this system I deal with two
challenges or issues, one that computes the union of private subsets that each of the interacting users
hold, and another that tests the inclusion of an element held by one user in a subset held by another.
The existing system uses different techniques for data mining purpose like Apriori algorithm. The
Fast Distributed Mining (FDM) algorithm of Cheung et al. [2], which is an unsecured distributed
version of the Apriori algorithm. Proposed system offers enhanced privacy and data mining with
respect to the Encryption techniques and Association rule with Fp-Growth Algorithm in private
cloud (system contains different files of subjects with respect to their branches). Due to this above
techniques the expected effect on this system is that, it is simpler and more efficient in terms of
communication cost and combinational cost. Due to these techniques it will affect the parameter like
time consumption for execution, length of the code is decrease, find the data fast, extracting hidden
predictive information from large databases and the efficiency of this proposed system should
increase by the 20%.
This document summarizes a research paper that proposes using a genetic algorithm to generate high-quality association rules for measuring data quality. The genetic algorithm evaluates rules based on four metrics: confidence, completeness, comprehensibility, and interestingness. It aims to discover high-level prediction rules that perform better than traditional greedy rule induction algorithms at handling attribute interactions. The genetic algorithm represents rules as chromosomes and uses the four metrics as an objective fitness function to evaluate the quality of each rule.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Comparison Between High Utility Frequent Item sets Mining Techniquesijsrd.com
Data Mining can be defined as an activity that extracts some new nontrivial information contained in large databases. Traditional data mining techniques have focused largely on detecting the statistical correlations between the items that are more frequent in the transaction databases. Also termed as frequent itemsets mining, these techniques were based on the rationale that itemsets which appear more frequently must be of more importance to the user from the business perspective .In this paper we throw light upon an emerging area called Utility Mining which not only considers the frequency of the itemsets but also considers the utility associated with the itemsets. The term utility refers to the importance or the usefulness of the appearance of the itemset in transactions quantified in terms like profit , sales or any other user preferences. This paper presents a novel efficient algorithm FUFM (Fast Utility-Frequent Mining) which finds all utility-frequent itemsets within the given utility and support constraints threshold. It is faster and simpler than the original 2P-UF algorithm (2 Phase Utility-Frequent), as it is based on efficient methods for frequent itemset mining. Experimental evaluation on artificial datasets shown here, in contrast with 2P-UF, our algorithm can also be applied to mine large databases.
A SURVEY ON PRIVACY PRESERVING ASSOCIATION RULE MININGijdkp
This document summarizes a survey on privacy preserving association rule mining techniques. It discusses the goals of association rule hiding which aim to prevent sensitive rules from being revealed while preserving non-sensitive rules. The techniques can be categorized as heuristic-based using data perturbation or blocking, or reconstruction-based where the data is perturbed and distributions are reconstructed. Common heuristic approaches modify transactions to hide sensitive rules with minimal side effects, while reconstruction approaches first perturb data and then reconstruct distributions for rule mining. The document analyzes various privacy preserving association rule mining algorithms.
A new hybrid algorithm for business intelligence recommender systemIJNSA Journal
Business Intelligence is a set of methods, process and technologies that transform raw data into meaningful
and useful information. Recommender system is one of business intelligence system that is used to obtain
knowledge to the active user for better decision making. Recommender systems apply data mining
techniques to the problem of making personalized recommendations for information. Due to the growth in
the number of information and the users in recent years offers challenges in recommender systems.
Collaborative, content, demographic and knowledge-based are four different types of recommendations
systems. In this paper, a new hybrid algorithm is proposed for recommender system which combines
knowledge based, profile of the users and most frequent item mining technique to obtain intelligence.
This document summarizes a study analyzing sales data at PT. Panca Putra Solusindo using the Apriori and FP-Growth algorithms. The study aimed to determine the most frequently purchased electronic products to help develop marketing strategies. Transaction data from 2016 was analyzed using the Apriori and FP-Growth algorithms in Weka software. The results showed the top selling items, with PCs and HP notebooks having the highest support over 46% and 30%. The algorithms helped identify best-selling products to optimize promotions.
The International Journal of Engineering and Sciencetheijes
This document summarizes a research paper on discovering actionable knowledge through multi-step data mining. The paper proposes a framework that combines multiple data sources, mining methods, and features to generate comprehensive patterns. This approach aims to provide more reliable and dependable intelligence than single-step mining. The framework integrates multi-source, multi-method, and multi-feature combined mining techniques. A prototype application demonstrated the effectiveness of the proposed combined mining approach for generating actionable knowledge from complex enterprise data.
This document provides a literature review on data mining with Oracle 10g using clustering and classification algorithms. It discusses the data mining process and common algorithms used, including Naive Bayes, decision trees, k-means clustering, and neural networks. The review categorizes data mining techniques into supervised learning (classification, prediction) and unsupervised learning (clustering, association rule mining). It also outlines the typical 4-step data mining process of problem definition, data preparation, model building and evaluation, and knowledge deployment.
This document discusses how data mining is used in the retail industry to gain insights about customers from large datasets. It explains that data mining can help retailers identify high-value customers, determine which new products customers may be interested in, and enable better decision making. Specific techniques discussed include market basket analysis to find common purchasing patterns, association rule mining to link frequently bought item combinations, and k-means clustering to organize customers into groups. The goal of these applications is to support customer relationship management and improve business strategies.
The document summarizes a novel approach for privacy preserving data mining on continuous and discrete data sets. The approach converts original sample data sets into a group of "unreal" data sets from which the original samples cannot be reconstructed. However, an accurate decision tree can still be built directly from the unreal data sets. This protects privacy while maintaining data mining utility. The approach determines information entropy and generates a decision tree using the unreal data sets and a perturbing set, without reconstructing the original samples.
This document summarizes an article that proposes a new algorithm for efficiently mining both positive and negative association rules from transactional databases. The algorithm first constructs a frequent pattern tree (FP-tree) to store the transaction information. It then uses an FP-growth approach to iteratively find frequent patterns and generate the positive and negative association rules without candidate generation. The algorithm aims to overcome limitations of previous methods and efficiently find all valid comparative association rules.
Irjet v4 iA Survey on FP (Growth) Tree using Association Rule Mining7351IRJET Journal
This document summarizes a survey on the FP-Growth algorithm for association rule mining. It discusses how FP-Growth is an effective algorithm that examines the transaction database only twice to generate frequent itemsets, unlike other algorithms. The document also reviews other association rule mining techniques like Apriori and discusses how FP-Growth avoids generating numerous conditional FP-trees.
IRJET- Retrieval of Images & Text using Data Mining TechniquesIRJET Journal
This document discusses using data mining techniques like clustering and association rule mining for image retrieval. It proposes a system that extracts both visual features (e.g. color, texture) and textual features from images. The features are clustered separately, then association rules are mined by fusing the clusters. Strong association rules are selected as training data. A query image's features are mined to find matching rules to retrieve semantically related images from the database. This combines content-based and text-based retrieval to address limitations of each approach individually.
Study on Data Mining Suitability for Intrusion Detection System (IDS)ijdmtaiir
Intrusion Detection System used to discover attacks
against computers and network Infrastructures. There are many
techniques used to determine the IDS such as Outlier Detection
Schemes for Anomaly Detection, K-Mean Clustering of
monitoring data, classification detection and outlier detection.
The data mining approaches help to determine what meets the
criteria as an intrusion versus normal traffic, whether a system
uses anomaly detection, misuse detection, target monitoring, or
stealth probes. This paper attempts to evaluate, categorize,
compares and summarizes the performance of data mining
techniques to detect the intrusion
This document analyzes wind power density in Azmar Mountain located in Sulaimani region of northern Iraq. Hourly wind speed data was collected over one year at a height of 5 meters above ground. The mean wind speed ranged from 4.65 to 11.86 m/s at 5 meters and 6.41 to 16.36 m/s at 50 meters after extrapolation. The Weibull distribution parameters c and k varied between months with c ranging from 5.21 to 13.46 m/s and k ranging from 1.60 to 3.47. The mean wind power density ranged from 70.89 to 1559.16 W/m2 at 5 meters and 186.71 to 4096.20 W/
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
This document summarizes research on improving login authentication by providing a graphical password system. It discusses limitations of traditional text passwords, such as being vulnerable to guessing attacks. The proposed system allows users to select a personal image during registration and authenticate by selecting a region of that image. An experiment found the system reduced guessing attacks compared to text passwords. The document provides details on implementing the system, including storing selected image coordinates during registration and verifying the coordinates during login. It presents results showing the system encouraged more secure passwords compared to text passwords.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
This document analyzes stresses in a rotating circular disc with a central hole and symmetrical array of non-central holes using finite element analysis. It investigates the effect of geometric parameters such as R2/R1 ratio of inner to outer disc radius, d/2R1 ratio of hole diameter to disc radius, Rb/(R2-R1) ratio of pitch circle radius to annular width, and number of holes N. Stress concentration factors are derived for these parameters. As number of holes increases, the stress concentration factor decreases. Results are verified against analytical solutions. Contour plots show highest stresses occur near holes.
PERFORMING DATA MINING IN (SRMS) THROUGH VERTICAL APPROACH WITH ASSOCIATION R...Editor IJMTER
This system technique is used for efficient data mining in SRMS (Student Records
Management System) through vertical approach with association rules in distributed databases. The
current leading technique is that of Kantarcioglu and Clifton[1]. In this system I deal with two
challenges or issues, one that computes the union of private subsets that each of the interacting users
hold, and another that tests the inclusion of an element held by one user in a subset held by another.
The existing system uses different techniques for data mining purpose like Apriori algorithm. The
Fast Distributed Mining (FDM) algorithm of Cheung et al. [2], which is an unsecured distributed
version of the Apriori algorithm. Proposed system offers enhanced privacy and data mining with
respect to the Encryption techniques and Association rule with Fp-Growth Algorithm in private
cloud (system contains different files of subjects with respect to their branches). Due to this above
techniques the expected effect on this system is that, it is simpler and more efficient in terms of
communication cost and combinational cost. Due to these techniques it will affect the parameter like
time consumption for execution, length of the code is decrease, find the data fast, extracting hidden
predictive information from large databases and the efficiency of this proposed system should
increase by the 20%.
This document summarizes a research paper that proposes using a genetic algorithm to generate high-quality association rules for measuring data quality. The genetic algorithm evaluates rules based on four metrics: confidence, completeness, comprehensibility, and interestingness. It aims to discover high-level prediction rules that perform better than traditional greedy rule induction algorithms at handling attribute interactions. The genetic algorithm represents rules as chromosomes and uses the four metrics as an objective fitness function to evaluate the quality of each rule.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Comparison Between High Utility Frequent Item sets Mining Techniquesijsrd.com
Data Mining can be defined as an activity that extracts some new nontrivial information contained in large databases. Traditional data mining techniques have focused largely on detecting the statistical correlations between the items that are more frequent in the transaction databases. Also termed as frequent itemsets mining, these techniques were based on the rationale that itemsets which appear more frequently must be of more importance to the user from the business perspective .In this paper we throw light upon an emerging area called Utility Mining which not only considers the frequency of the itemsets but also considers the utility associated with the itemsets. The term utility refers to the importance or the usefulness of the appearance of the itemset in transactions quantified in terms like profit , sales or any other user preferences. This paper presents a novel efficient algorithm FUFM (Fast Utility-Frequent Mining) which finds all utility-frequent itemsets within the given utility and support constraints threshold. It is faster and simpler than the original 2P-UF algorithm (2 Phase Utility-Frequent), as it is based on efficient methods for frequent itemset mining. Experimental evaluation on artificial datasets shown here, in contrast with 2P-UF, our algorithm can also be applied to mine large databases.
A SURVEY ON PRIVACY PRESERVING ASSOCIATION RULE MININGijdkp
This document summarizes a survey on privacy preserving association rule mining techniques. It discusses the goals of association rule hiding which aim to prevent sensitive rules from being revealed while preserving non-sensitive rules. The techniques can be categorized as heuristic-based using data perturbation or blocking, or reconstruction-based where the data is perturbed and distributions are reconstructed. Common heuristic approaches modify transactions to hide sensitive rules with minimal side effects, while reconstruction approaches first perturb data and then reconstruct distributions for rule mining. The document analyzes various privacy preserving association rule mining algorithms.
A new hybrid algorithm for business intelligence recommender systemIJNSA Journal
Business Intelligence is a set of methods, process and technologies that transform raw data into meaningful
and useful information. Recommender system is one of business intelligence system that is used to obtain
knowledge to the active user for better decision making. Recommender systems apply data mining
techniques to the problem of making personalized recommendations for information. Due to the growth in
the number of information and the users in recent years offers challenges in recommender systems.
Collaborative, content, demographic and knowledge-based are four different types of recommendations
systems. In this paper, a new hybrid algorithm is proposed for recommender system which combines
knowledge based, profile of the users and most frequent item mining technique to obtain intelligence.
This document summarizes a study analyzing sales data at PT. Panca Putra Solusindo using the Apriori and FP-Growth algorithms. The study aimed to determine the most frequently purchased electronic products to help develop marketing strategies. Transaction data from 2016 was analyzed using the Apriori and FP-Growth algorithms in Weka software. The results showed the top selling items, with PCs and HP notebooks having the highest support over 46% and 30%. The algorithms helped identify best-selling products to optimize promotions.
The International Journal of Engineering and Sciencetheijes
This document summarizes a research paper on discovering actionable knowledge through multi-step data mining. The paper proposes a framework that combines multiple data sources, mining methods, and features to generate comprehensive patterns. This approach aims to provide more reliable and dependable intelligence than single-step mining. The framework integrates multi-source, multi-method, and multi-feature combined mining techniques. A prototype application demonstrated the effectiveness of the proposed combined mining approach for generating actionable knowledge from complex enterprise data.
This document provides a literature review on data mining with Oracle 10g using clustering and classification algorithms. It discusses the data mining process and common algorithms used, including Naive Bayes, decision trees, k-means clustering, and neural networks. The review categorizes data mining techniques into supervised learning (classification, prediction) and unsupervised learning (clustering, association rule mining). It also outlines the typical 4-step data mining process of problem definition, data preparation, model building and evaluation, and knowledge deployment.
This document discusses how data mining is used in the retail industry to gain insights about customers from large datasets. It explains that data mining can help retailers identify high-value customers, determine which new products customers may be interested in, and enable better decision making. Specific techniques discussed include market basket analysis to find common purchasing patterns, association rule mining to link frequently bought item combinations, and k-means clustering to organize customers into groups. The goal of these applications is to support customer relationship management and improve business strategies.
The document summarizes a novel approach for privacy preserving data mining on continuous and discrete data sets. The approach converts original sample data sets into a group of "unreal" data sets from which the original samples cannot be reconstructed. However, an accurate decision tree can still be built directly from the unreal data sets. This protects privacy while maintaining data mining utility. The approach determines information entropy and generates a decision tree using the unreal data sets and a perturbing set, without reconstructing the original samples.
This document summarizes an article that proposes a new algorithm for efficiently mining both positive and negative association rules from transactional databases. The algorithm first constructs a frequent pattern tree (FP-tree) to store the transaction information. It then uses an FP-growth approach to iteratively find frequent patterns and generate the positive and negative association rules without candidate generation. The algorithm aims to overcome limitations of previous methods and efficiently find all valid comparative association rules.
Irjet v4 iA Survey on FP (Growth) Tree using Association Rule Mining7351IRJET Journal
This document summarizes a survey on the FP-Growth algorithm for association rule mining. It discusses how FP-Growth is an effective algorithm that examines the transaction database only twice to generate frequent itemsets, unlike other algorithms. The document also reviews other association rule mining techniques like Apriori and discusses how FP-Growth avoids generating numerous conditional FP-trees.
IRJET- Retrieval of Images & Text using Data Mining TechniquesIRJET Journal
This document discusses using data mining techniques like clustering and association rule mining for image retrieval. It proposes a system that extracts both visual features (e.g. color, texture) and textual features from images. The features are clustered separately, then association rules are mined by fusing the clusters. Strong association rules are selected as training data. A query image's features are mined to find matching rules to retrieve semantically related images from the database. This combines content-based and text-based retrieval to address limitations of each approach individually.
Study on Data Mining Suitability for Intrusion Detection System (IDS)ijdmtaiir
Intrusion Detection System used to discover attacks
against computers and network Infrastructures. There are many
techniques used to determine the IDS such as Outlier Detection
Schemes for Anomaly Detection, K-Mean Clustering of
monitoring data, classification detection and outlier detection.
The data mining approaches help to determine what meets the
criteria as an intrusion versus normal traffic, whether a system
uses anomaly detection, misuse detection, target monitoring, or
stealth probes. This paper attempts to evaluate, categorize,
compares and summarizes the performance of data mining
techniques to detect the intrusion
This document analyzes wind power density in Azmar Mountain located in Sulaimani region of northern Iraq. Hourly wind speed data was collected over one year at a height of 5 meters above ground. The mean wind speed ranged from 4.65 to 11.86 m/s at 5 meters and 6.41 to 16.36 m/s at 50 meters after extrapolation. The Weibull distribution parameters c and k varied between months with c ranging from 5.21 to 13.46 m/s and k ranging from 1.60 to 3.47. The mean wind power density ranged from 70.89 to 1559.16 W/m2 at 5 meters and 186.71 to 4096.20 W/
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
This document summarizes research on improving login authentication by providing a graphical password system. It discusses limitations of traditional text passwords, such as being vulnerable to guessing attacks. The proposed system allows users to select a personal image during registration and authenticate by selecting a region of that image. An experiment found the system reduced guessing attacks compared to text passwords. The document provides details on implementing the system, including storing selected image coordinates during registration and verifying the coordinates during login. It presents results showing the system encouraged more secure passwords compared to text passwords.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
This document analyzes stresses in a rotating circular disc with a central hole and symmetrical array of non-central holes using finite element analysis. It investigates the effect of geometric parameters such as R2/R1 ratio of inner to outer disc radius, d/2R1 ratio of hole diameter to disc radius, Rb/(R2-R1) ratio of pitch circle radius to annular width, and number of holes N. Stress concentration factors are derived for these parameters. As number of holes increases, the stress concentration factor decreases. Results are verified against analytical solutions. Contour plots show highest stresses occur near holes.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
This document proposes a new approach called Local Active Pixels (LAPP) for face recognition that reduces computational resources compared to Local Binary Patterns (LBP). LAPP identifies "active pixels" that contain essential image information using Brody Transform thresholds. Only active pixels are used for feature extraction and recognition, reducing features without sacrificing accuracy. The document describes LBP and Brody Transform, then presents the LAPP approach and experimental evaluation using several face datasets to demonstrate its performance.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
This document compares the Selected Mapping (SLM) and Partial Transmit Sequence (PTS) techniques for reducing peak-to-average power ratio (PAPR) in multicarrier code division multiple access (MC-CDMA) systems. Through MATLAB simulations, the authors found that:
1) PTS partitions the input data block into disjoint subblocks, weights the subcarriers in each subblock with phase factors, and selects the combination with the lowest PAPR for transmission.
2) SLM generates multiple alternative signals by multiplying the input with different phase sequences before inverse fast Fourier transform. It selects the signal with the lowest PAPR for transmission.
3) PTS has lower computational
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
This document presents a new approach to redesigning the basic operators used in parallel prefix adders when implementing them in FPGAs. Parallel prefix adders are commonly used for high-speed binary addition. They work by using generate and propagate signals to compute the carry inputs at each stage in parallel. Typically, two operators called "black" and "gray" are used to compute the generate and propagate signals. The authors propose redesigning these basic operators by using multiplexers instead of logic gates, to better match the structure of FPGA slices. They implement several parallel prefix adders, including Brent-Kung, Skalansky, and Kogge-Stone adders, using the original and redesigned operators on
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
PUNTO DE ACUERDO PARA INVESTIGAR A FUNCIONARIOS DE BENITO JUÁREZErnestina Godoy
a efecto de que los ciudadanos del Distrito Federal estén en posibilidad de contar con información veraz, proveniente de las propias autoridades, se solicita que se requiera al Ciudadano Delegado en Benito Juárez, Jorge Romero, informe sobre viaje de funcionarios de esa delegación a Brasil.
Go through our presentation on Why aren't you using Quality Monitoring and learn more about quality monitoring and its benefits. To learn more, visit us at www.inin.com
This short document promotes creating presentations using Haiku Deck on SlideShare. It encourages the reader to get started making their own Haiku Deck presentation by providing a button to click to begin the process. The document is advertising the creation of presentations on Haiku Deck and SlideShare.
Improved Snr Of Mst Radar Signals By Kaiser-Hamming And Cosh-Hamming Window F...IJERA Editor
In this paper the effect of window shape parameter „α‟ in Kaiser-Hamming and Cosh-Hamming window functions on the signal to noise ratio (SNR) values of the Indian Mesosphere-Stratosphere-Troposphere (MST) radar is computed. The six parts of multibeam observations of the lower atmosphere made by the MST radar are utilized for the analysis of results. Prior to the Fourier transformation, the in-phase and quadrature components of radar echo samples are weighted with proposed windows based on the Kaiser-Hamming and Cosh-Hamming Window functions. The effects of data weighting with the change of the window shape parameter „α‟ of the Kaiser-Hamming and Cosh-Hamming Window functions are given in it. It is noted that the increase of variable window shape parameter „α‟ increases the signal to noise ratio values and a better improvement is reported. For Kaiser-Hamming and Cosh-Hamming Window functions are proposed to analyze the MST radar return signals to obtain optimum values of the window shape parameter. The results shows the improvement of signal to noise ratio of noisy data due to the effect of side lobe reduction and demands for the design of the optimal window functions.
A NOVEL APPROACH TO MINE FREQUENT PATTERNS FROM LARGE VOLUME OF DATASET USING...IAEME Publication
In this paper, MDL based reduction in frequent pattern is presented. The ideal outcome of any pattern mining process is to explore the data in new insights. And also, we need to eliminate the non-interesting patterns that describe noise. The major problem in frequent pattern mining is to identify the interesting patterns. Instead of performing association rule mining on all the frequent item sets, it is feasible to select a sub set of frequent item sets and perform the mining task. Selecting a small set of frequent item sets from large amount of interesting ones is a difficult task. In our approach, MDL based algorithm is used for reducing the number of frequent item sets to be used for association rule mining is presented.
IRJET- Classification of Pattern Storage System and Analysis of Online Shoppi...IRJET Journal
This document discusses classifying patterns from online shopping data using data mining techniques. It proposes using the Apriori algorithm to mine frequent patterns from transaction data stored in a data warehouse. Patterns mined from the data warehouse using Apriori would then be stored in a pattern warehouse. This would allow users to view product details and related patterns when browsing items online. The system aims to efficiently analyze large amounts of user data to discover useful patterns for improving the online shopping experience.
Mining Frequent Item set Using Genetic Algorithmijsrd.com
By applying rule mining algorithms, frequent itemsets are generated from large data sets e.g. Apriori algorithm. It takes so much computer time to compute all frequent itemsets. We can solve this problem much efficiently by using Genetic Algorithm(GA). GA performs global search and the time complexity is less compared to other algorithms. Genetic Algorithms (GAs) are adaptive heuristic search & optimization method for solving both constrained and unconstrained problems based on the evolutionary ideas of natural selection and genetic. The main aim of this work is to find all the frequent itemsets from given data sets using genetic algorithm & compare the results generated by GA with other algorithms. Population size, number of generation, crossover probability, and mutation probability are the parameters of GA which affect the quality of result and time of calculation.
COMPARATIVE STUDY OF DISTRIBUTED FREQUENT PATTERN MINING ALGORITHMS FOR BIG S...IAEME Publication
The document presents a comparative study of distributed frequent pattern mining algorithms CDA, FDM, and MR-DARM for mining big sales data. It first preprocesses a large AMUL dairy sales dataset using Hadoop MapReduce to convert it to transactions. It then applies the MR-DARM, CDA, and FDM algorithms to find frequent itemsets and compare their performance. Experimental results on the dairy dataset show that MR-DARM has lower execution times than CDA and FDM, especially when the data is distributed across multiple nodes.
In the recent years the scope of data mining has evolved into an active area of research because of the previously unknown and interesting knowledge from very large database collection. The data mining is applied on a variety of applications in multiple domains like in business, IT and many more sectors. In Data Mining the major problem which receives great attention by the community is the classification of the data. The classification of data should be such that it could be they can be easily verified and should be easily interpreted by the humans. In this paper we would be studying various data mining techniques so that we can find few combinations for enhancing the hybrid technique which would be having multiple techniques involved so enhance the usability of the application. We would be studying CHARM Algorithm, CM-SPAM Algorithm, Apriori Algorithm, MOPNAR Algorithm and the Top K Rules.
This document describes a proposed intelligent supermarket system that uses a distributed implementation of the Apriori algorithm (called Dipriori) to analyze transaction data from multiple supermarket locations and identify frequent item sets and association rules. The system would have client machines at each location running Apriori locally on their transaction data and sending the frequent item sets to a central global server. The server would calculate average support counts and identify globally frequent item sets. Association rules would then be generated and used to provide insights into customer purchasing behaviors and inform targeted marketing strategies. The authors implemented this approach and found it helped improve the time efficiency of analyzing large, distributed transaction datasets.
Frequent pattern mining techniques helpful to find interesting trends or patterns in
massive data. Prior domain knowledge leads to decide appropriate minimum support threshold. This
review article show different frequent pattern mining techniques based on apriori or FP-tree or user
define techniques under different computing environments like parallel, distributed or available data
mining tools, those helpful to determine interesting frequent patterns/itemsets with or without prior
domain knowledge. Proposed review article helps to develop efficient and scalable frequent pattern
mining techniques.
To Get any Project for CSE, IT ECE, EEE Contact Me @ 09666155510, 09849539085 or mail us - ieeefinalsemprojects@gmail.com-Visit Our Website: www.finalyearprojects.org
To Get any Project for CSE, IT ECE, EEE Contact Me @ 09666155510, 09849539085 or mail us - ieeefinalsemprojects@gmail.com-Visit Our Website: www.finalyearprojects.org
A New Data Stream Mining Algorithm for Interestingness-rich Association RulesVenu Madhav
Frequent itemset mining and association rule generation is
a challenging task in data stream. Even though, various algorithms
have been proposed to solve the issue, it has been found
out that only frequency does not decides the significance
interestingness of the mined itemset and hence the association
rules. This accelerates the algorithms to mine the association
rules based on utility i.e. proficiency of the mined rules. However,
fewer algorithms exist in the literature to deal with the utility
as most of them deals with reducing the complexity in frequent
itemset/association rules mining algorithm. Also, those few
algorithms consider only the overall utility of the association
rules and not the consistency of the rules throughout a defined
number of periods. To solve this issue, in this paper, an enhanced
association rule mining algorithm is proposed. The algorithm
introduces new weightage validation in the conventional
association rule mining algorithms to validate the utility and
its consistency in the mined association rules. The utility is
validated by the integrated calculation of the cost/price efficiency
of the itemsets and its frequency. The consistency validation
is performed at every defined number of windows using the
probability distribution function, assuming that the weights are
normally distributed. Hence, validated and the obtained rules
are frequent and utility efficient and their interestingness are
distributed throughout the entire time period. The algorithm is
implemented and the resultant rules are compared against the
rules that can be obtained from conventional mining algorithms
IRJET- Improving the Performance of Smart Heterogeneous Big DataIRJET Journal
This document discusses improving the performance of smart heterogeneous big data. It begins by defining key concepts like big data, data mining, and the challenges of analyzing large, complex datasets. It then describes two common association rule mining algorithms - Apriori and FP-Growth - that are used to extract patterns from big data. The document proposes using principal component analysis as a feature selection method to improve the performance of these algorithms. It finds that this proposed approach reduces execution time compared to the original algorithms when processing big data.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
DEVELOPING A NOVEL MULTIDIMENSIONAL MULTIGRANULARITY DATA MINING APPROACH FOR...cscpconf
Data Mining is one of the most significant tools for discovering association patterns that are useful for many knowledge domains. Yet, there are some drawbacks in existing mining techniques. Three main weaknesses of current data-mining techniques are: 1) re-scanning of the entire database must be done whenever new attributes are added. 2) An association rule may be true on a certain granularity but fail on a smaller ones and vise verse. 3) Current methods can only be used to find either frequent rules or infrequent rules, but not both at the same time. This research proposes a novel data schema and an algorithm that solves the above weaknesses while improving on the efficiency and effectiveness of data mining strategies. Crucial mechanisms in each step will be clarified in this paper. Finally, this paper presents experimental results regarding efficiency, scalability, information loss, etc. of the proposed approach to prove its advantages.
The document describes a self-learning real-time expert system developed for fault diagnosis in power plants. It uses historic operational data and machine learning techniques to automatically generate new rules for the expert system's knowledge base. The key steps are: (1) preprocessing historical data, (2) applying an association algorithm (Tertius was found most suitable) to generate new rules, (3) validating and formatting new rules, and (4) updating the knowledge base. This allows the expert system to continuously improve its diagnostic ability by learning from plant operational data.
This document describes a proposed system for web extraction using a trinity construction and efficient algorithms. It begins with an abstract discussing how trinity characteristics can be used to automatically extract content from websites in a sequential tree structure. It then discusses the existing system which uses trinity tree and prefix/suffix sorting but has limitations. The proposed system introduces fuzzy logic for multi-perspective crawling across multiple websites, genetic algorithms to load extracted content into the trinity structure, and an ant colony optimization algorithm to obtain an effective structure and suggest optimized solutions. Key aspects of fuzzy logic, genetic algorithms, and ant colony optimization are also summarized.
International journal of computer science and innovation vol 2015-n1-paper4sophiabelthome
This paper discusses applications of association rule mining (ARM). Section 1 provides an overview of ARM and describes it as an important data mining technique. Section 2 reviews literature on ARM and discusses various ARM algorithms. Section 3 describes several applications of ARM in detail, including market basket analysis, building intelligent transportation systems, medical diagnosis, and web log analysis. ARM is useful for discovering frequent patterns and relationships in large datasets across many domains.
1) The document discusses the Cyclic Model Analysis (CMA) technique for sequential pattern mining which aims to predict customer purchasing behavior.
2) CMA calculates the Trend Distribution Function from sequential patterns to model purchasing trends over time. It then uses Generalized Periodicity Detection and Trend Modeling to identify periodic patterns and construct an approximating model.
3) The Cyclic Model Analysis algorithm is applied to further analyze the patterns, dividing the domain into segments where the distribution function is increasing or decreasing and applying the other techniques recursively to fully model the cyclic behavior.
A NEW HYBRID ALGORITHM FOR BUSINESS INTELLIGENCE RECOMMENDER SYSTEMIJNSA Journal
Business Intelligence is a set of methods, process and technologies that transform raw data into meaningful and useful information. Recommender system is one of business intelligence system that is used to obtain knowledge to the active user for better decision making. Recommender systems apply data mining techniques to the problem of making personalized recommendations for information. Due to the growth in the number of information and the users in recent years offers challenges in recommender systems. Collaborative, content, demographic and knowledge-based are four different types of recommendations systems. In this paper, a new hybrid algorithm is proposed for recommender system which combines knowledge based, profile of the users and most frequent item mining technique to obtain intelligence.
In a Power plant with a Distributed Control System ( DCS ), process parameters are continuously stored in
databases at discrete intervals. The data contained in these databases may not appear to contain valuable
relational information but practically such a relation exists. The large number of process parameter values
are changing with time in a Power Plant. These parameters are part of rules framed by domain experts for
the expert system. With the changes in parameters there is a quite high possibility to form new rules using
the dynamics of the process itself. We present an efficient algorithm that generates all significant rules
based on the real data. The association based algorithms were compared and the best suited algorithm for
this process application was selected. The application for the Learning system is studied in a Power Plant
domain. The SCADA interface was developed to acquire online plant data.
An Efficient Compressed Data Structure Based Method for Frequent Item Set Miningijsrd.com
Frequent pattern mining is very important for business organizations. The major applications of frequent pattern mining include disease prediction and analysis, rain forecasting, profit maximization, etc. In this paper, we are presenting a new method for mining frequent patterns. Our method is based on a new compact data structure. This data structure will help in reducing the execution time.
Similar to Adaptive and Fast Predictions by Minimal Itemsets Creation (20)
Prediction of Electrical Energy Efficiency Using Information on Consumer's Ac...PriyankaKilaniya
Energy efficiency has been important since the latter part of the last century. The main object of this survey is to determine the energy efficiency knowledge among consumers. Two separate districts in Bangladesh are selected to conduct the survey on households and showrooms about the energy and seller also. The survey uses the data to find some regression equations from which it is easy to predict energy efficiency knowledge. The data is analyzed and calculated based on five important criteria. The initial target was to find some factors that help predict a person's energy efficiency knowledge. From the survey, it is found that the energy efficiency awareness among the people of our country is very low. Relationships between household energy use behaviors are estimated using a unique dataset of about 40 households and 20 showrooms in Bangladesh's Chapainawabganj and Bagerhat districts. Knowledge of energy consumption and energy efficiency technology options is found to be associated with household use of energy conservation practices. Household characteristics also influence household energy use behavior. Younger household cohorts are more likely to adopt energy-efficient technologies and energy conservation practices and place primary importance on energy saving for environmental reasons. Education also influences attitudes toward energy conservation in Bangladesh. Low-education households indicate they primarily save electricity for the environment while high-education households indicate they are motivated by environmental concerns.
Digital Twins Computer Networking Paper Presentation.pptxaryanpankaj78
A Digital Twin in computer networking is a virtual representation of a physical network, used to simulate, analyze, and optimize network performance and reliability. It leverages real-time data to enhance network management, predict issues, and improve decision-making processes.
Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024Sinan KOZAK
Sinan from the Delivery Hero mobile infrastructure engineering team shares a deep dive into performance acceleration with Gradle build cache optimizations. Sinan shares their journey into solving complex build-cache problems that affect Gradle builds. By understanding the challenges and solutions found in our journey, we aim to demonstrate the possibilities for faster builds. The case study reveals how overlapping outputs and cache misconfigurations led to significant increases in build times, especially as the project scaled up with numerous modules using Paparazzi tests. The journey from diagnosing to defeating cache issues offers invaluable lessons on maintaining cache integrity without sacrificing functionality.
Accident detection system project report.pdfKamal Acharya
The Rapid growth of technology and infrastructure has made our lives easier. The
advent of technology has also increased the traffic hazards and the road accidents take place
frequently which causes huge loss of life and property because of the poor emergency facilities.
Many lives could have been saved if emergency service could get accident information and
reach in time. Our project will provide an optimum solution to this draw back. A piezo electric
sensor can be used as a crash or rollover detector of the vehicle during and after a crash. With
signals from a piezo electric sensor, a severe accident can be recognized. According to this
project when a vehicle meets with an accident immediately piezo electric sensor will detect the
signal or if a car rolls over. Then with the help of GSM module and GPS module, the location
will be sent to the emergency contact. Then after conforming the location necessary action will
be taken. If the person meets with a small accident or if there is no serious threat to anyone’s
life, then the alert message can be terminated by the driver by a switch provided in order to
avoid wasting the valuable time of the medical rescue team.
Open Channel Flow: fluid flow with a free surfaceIndrajeet sahu
Open Channel Flow: This topic focuses on fluid flow with a free surface, such as in rivers, canals, and drainage ditches. Key concepts include the classification of flow types (steady vs. unsteady, uniform vs. non-uniform), hydraulic radius, flow resistance, Manning's equation, critical flow conditions, and energy and momentum principles. It also covers flow measurement techniques, gradually varied flow analysis, and the design of open channels. Understanding these principles is vital for effective water resource management and engineering applications.
Home security is of paramount importance in today's world, where we rely more on technology, home
security is crucial. Using technology to make homes safer and easier to control from anywhere is
important. Home security is important for the occupant’s safety. In this paper, we came up with a low cost,
AI based model home security system. The system has a user-friendly interface, allowing users to start
model training and face detection with simple keyboard commands. Our goal is to introduce an innovative
home security system using facial recognition technology. Unlike traditional systems, this system trains
and saves images of friends and family members. The system scans this folder to recognize familiar faces
and provides real-time monitoring. If an unfamiliar face is detected, it promptly sends an email alert,
ensuring a proactive response to potential security threats.
DEEP LEARNING FOR SMART GRID INTRUSION DETECTION: A HYBRID CNN-LSTM-BASED MODELijaia
As digital technology becomes more deeply embedded in power systems, protecting the communication
networks of Smart Grids (SG) has emerged as a critical concern. Distributed Network Protocol 3 (DNP3)
represents a multi-tiered application layer protocol extensively utilized in Supervisory Control and Data
Acquisition (SCADA)-based smart grids to facilitate real-time data gathering and control functionalities.
Robust Intrusion Detection Systems (IDS) are necessary for early threat detection and mitigation because
of the interconnection of these networks, which makes them vulnerable to a variety of cyberattacks. To
solve this issue, this paper develops a hybrid Deep Learning (DL) model specifically designed for intrusion
detection in smart grids. The proposed approach is a combination of the Convolutional Neural Network
(CNN) and the Long-Short-Term Memory algorithms (LSTM). We employed a recent intrusion detection
dataset (DNP3), which focuses on unauthorized commands and Denial of Service (DoS) cyberattacks, to
train and test our model. The results of our experiments show that our CNN-LSTM method is much better
at finding smart grid intrusions than other deep learning algorithms used for classification. In addition,
our proposed approach improves accuracy, precision, recall, and F1 score, achieving a high detection
accuracy rate of 99.50%.
Use PyCharm for remote debugging of WSL on a Windo cf5c162d672e4e58b4dde5d797...shadow0702a
This document serves as a comprehensive step-by-step guide on how to effectively use PyCharm for remote debugging of the Windows Subsystem for Linux (WSL) on a local Windows machine. It meticulously outlines several critical steps in the process, starting with the crucial task of enabling permissions, followed by the installation and configuration of WSL.
The guide then proceeds to explain how to set up the SSH service within the WSL environment, an integral part of the process. Alongside this, it also provides detailed instructions on how to modify the inbound rules of the Windows firewall to facilitate the process, ensuring that there are no connectivity issues that could potentially hinder the debugging process.
The document further emphasizes on the importance of checking the connection between the Windows and WSL environments, providing instructions on how to ensure that the connection is optimal and ready for remote debugging.
It also offers an in-depth guide on how to configure the WSL interpreter and files within the PyCharm environment. This is essential for ensuring that the debugging process is set up correctly and that the program can be run effectively within the WSL terminal.
Additionally, the document provides guidance on how to set up breakpoints for debugging, a fundamental aspect of the debugging process which allows the developer to stop the execution of their code at certain points and inspect their program at those stages.
Finally, the document concludes by providing a link to a reference blog. This blog offers additional information and guidance on configuring the remote Python interpreter in PyCharm, providing the reader with a well-rounded understanding of the process.
Discover the latest insights on Data Driven Maintenance with our comprehensive webinar presentation. Learn about traditional maintenance challenges, the right approach to utilizing data, and the benefits of adopting a Data Driven Maintenance strategy. Explore real-world examples, industry best practices, and innovative solutions like FMECA and the D3M model. This presentation, led by expert Jules Oudmans, is essential for asset owners looking to optimize their maintenance processes and leverage digital technologies for improved efficiency and performance. Download now to stay ahead in the evolving maintenance landscape.
Software Engineering and Project Management - Software Testing + Agile Method...Prakhyath Rai
Software Testing: A Strategic Approach to Software Testing, Strategic Issues, Test Strategies for Conventional Software, Test Strategies for Object -Oriented Software, Validation Testing, System Testing, The Art of Debugging.
Agile Methodology: Before Agile – Waterfall, Agile Development.
Adaptive and Fast Predictions by Minimal Itemsets Creation
1. Ms. Sonali Vaidya et al. Int. Journal of Engineering Research and Applications www.ijera.com
ISSN : 2248-9622, Vol. 5, Issue 3, ( Part -3) March 2015, pp.27-31
www.ijera.com 27 | P a g e
Adaptive and Fast Predictions by Minimal Itemsets Creation
Ms. Sonali Vaidya*, Mr. Struan Souto**, Mr. Keegan Pereira**, Mr. Vincent
Soares**, Mr. Neil Rego**
*,**(Department of Information Technology, Mumbai University, St. Francis Institute of Technology, Mumbai)
ABSTRACT
Association rules in data mining are useful for the analysis and prediction of an individual user’s behavior which
facilitates the data analysis on a regular basis for market basket data, clustering of products, designing catalogs
and playing an immense role for store layout setting. This paper presents a successor of the Apriori and variation
of the CHARM algorithm, which is the MG-CHARM algorithm that is used for finding relationships between
certain attributes instead of the whole dataset. The MG-CHARM algorithm is an Association Rule Mining
(ARM) algorithm that is used to select the target database which in turn takes less time to find the desired
association rules. In the proposed implementation, the actions performed by a user will be effectively recorded
in the target database. The datasets that are generated will have to be fed to the semi - automated, adaptive
software which will fetch the output based on ARM algorithm. The algorithm will determine the relations for
different datasets by mining Minimal Generators (mGs) from Frequently Closed Itemsets (FCI’s) to carry out
decision making and pattern analysis of the Itemsets.
Keywords–Apriori, CHARM, MG-CHARM, ARM, mG’s, FCI’s
I. INTRODUCTION
Programmers use association rules to build
programs capable of machine learning. This
Association Rule Mining (ARM) algorithm is used
to select the target database which in turn takes less
time to find the desired association rules. The
incorporating user’s preference in selection of target
attribute also helps to search the association rules
efficiently both in terms of space and time.
II. NATURE OF PROBLEM
The key element that makes association rule
mining practical is the minimal support and
threshold which is used to prune the search space
and to limit the number of frequent itemsets and
rules generated. Some of the major drawbacks of
association rule algorithms are huge number of
considered rules, generation of non-interesting rules
and low algorithm performance. At present, almost
all algorithms for mining Minimal Generators of
FCIs are based on Apriori algorithm. The first
method to find mGs extended from Apriori
algorithm found candidates that are mGs, and then
defined their closures to find out frequent closed
itemsets. Firstly, all frequent closed itemsets were
found using CHARM algorithm. Then using level-
wise method all mGs that correspond to each closed
itemset was found. Both of these methods have
disadvantage in large size of frequent itemsets since
the number of considered candidates is large. Also
the time and space involved is too large since all the
itemsets are taken into consideration. Also
association rule mining of different datasets taking
into consideration a dynamic environment is
troublesome.
III. PREVIOUS WORK
The past few years have seen a tremendous
interest in the area of data mining. Data mining is
generally thought of as the process of finding the
information associated or included in a large
collection of data which is thought to be hidden,
non-trivial and previously unknown. Finding
regularities data patterns can be done by a trivial
data mining component and an important class of
methods referred to as the association rules.
Association mining has been used in many
application domains and therefore it paves its way as
the most important model invented and extensively
studied by databases and data mining community.
Discovering of purchase patterns or association
between products is very useful for decision making
and effective marketing which is an immense part of
the business field. Biological databases pattern
exploration, software engineering metrics has
inclusive knowledge that can be extracted or
retrieved periodically, personalization of web
content, and mining of textual data are some of the
applications recently developed taking into
consideration the field of data mining. Most of the
research efforts in the scope of association rules
have been oriented to simplify the rule set and
improve the performance of the algorithm. But these
are not the only problems that can be found when
rules are generated and applied in different domains.
RESEARCH ARTICLE OPEN ACCESS
2. Ms. Sonali Vaidya et al. Int. Journal of Engineering Research and Applications www.ijera.com
ISSN : 2248-9622, Vol. 5, Issue 3, ( Part -3) March 2015, pp.27-31
www.ijera.com 28 | P a g e
Troubleshooting for them should also be taken into
consideration. The purpose of association model and
data they come from also need to be understood.
IV. PURPOSE
The proposed solution focuses on overcoming
the drawbacks of the Apriori and CHARM
algorithms. In this technique the number of frequent
closed itemsets (FCIs) generated is usually fewer
than the number of frequent itemsets and therefore it
is necessary to find Minimal Generators (mGs) for
mining association rule from them. Exploring of one
or more mG’s depends on the approaches based on
generating candidate lose timeliness when the
number of frequent closed itemsets is large. Finding
all mGs of frequent closed itemsets is done
efficiently and effectively by the MG-CHARM in a
closed hierarchical manner. Using the ARM
algorithm, an adaptive system which does not
generate candidates, by mining directly the mGs of
frequent closed itemsets during the mining of FCI’s
can be developed. Thus, the time for finding mGs of
frequent closed itemsets is insignificant. Also
mining (minimal) non-redundant association rules is
possible. The system can effectively handle different
datasets in order to provide an output of association
rules, thereby facilitating data visualization in order
to manage the data in an organized manner. It will
also work to provide greater accuracy and will
perform reliably at a given point of time.
V. CONTRIBUTION OF PAPER
This paper aims to use and detect the frequency
patterns i.e. the minimally frequent itemsets taking
into consideration a large number of datasets.
However to effectively provide suggestions in real
time, the datasets need to fed to the software
separately at regular intervals of time to generate the
necessary output. Association Rule Mining and Data
Visualization will be carried out for the same and
yield the necessary output. Mining of any dataset is
possible. Also the software is semi-automated when
it comes to finding out association rules for a given
support and confidence and thus time and work
effort is also reduced.
VI. FIGURES AND TABLES
Fig. 1
Fig.1 speaks about the consideration and the mining
of association rules from datasets mined or fetched
from the target database.
The system implementation can be carried out in two
parts:
1. User to database: To record user action(s).
2. Database to software: To analyze, visualize,
generate and return an association rule output.
The ARM algorithm will determine effectively the
relations for different datasets by mining Minimal
Generators (MGs) from Frequently Closed Itemsets
(FCI’s).
Data Visualization can be effectively carried out to
study and visualize the data. This in turn will aid in
effective decision making. The output of the ARM
algorithm will be returned back to the database.
In order to prevent system bottlenecks and allow the
system to continue its tasks without having to stop or
suspend its working, the mining will have to be
carried out periodically depending on the urge and
need of the organization to analyze the data. The
idea of mining the FCI’s periodically is to enable the
system to yield suggestions with ease and guide the
user, thereby avoiding any difficulty while accessing
the system at any point in time.
The system will hold true for any dataset. This will
therefore give an organization a clear cut idea of
how to use the data provided as well as the data
generated in an efficient manner.
3. Ms. Sonali Vaidya et al. Int. Journal of Engineering Research and Applications www.ijera.com
ISSN : 2248-9622, Vol. 5, Issue 3, ( Part -3) March 2015, pp.27-31
www.ijera.com 29 | P a g e
Fig. 2
Fig.2 represents a flow chart of the system in which
a determined support and confidence will give the
required output for a recorded user action provided
the details of the user activity for example: A user
transaction for certain items at a store, is recorded in
the database. The itemset relation is obtained by
pruning which is immensely important and saves
time and memory.
VII. THE ALGORITHM
Input: The database D and support threshold
minSup
Output: all FCI satisfy minSup and their mG
Method:
MG-CHARM(D, minSup)
[Φ]={lj× t(li) ,(Ii): Ii ∈I/ σ (li) ≥minSup}
MG-CHARM-EXTEND([Φ], C = Φ )
Return C
MG-CHARM-EXTEND([P], C)
for each li × t(lj), mG(li) in [P] do
Pi= Pi ∪ li and [Pi]=Φ
for each lj× t(lj), mG(lj) in [P], with j > i do
X = lj and Y=t(li) ∩ t(lj)
MG-CHARM-PROPERTY(X ×Y,li,lj,Pi, [Pi],[Pj)
SUBSUMPTION-CHECK(C, Pi)
MG-CIIARM-EXTEND([Pi],C)
MG-CHARM-PROPERTY(X ×Y,li,lj,Pi, [Pi],[Pj)
ifσ (X) ≥minSup then
if t(li) = t(lj) then // property I
Remove lj from P
Pi=Pi ∪lj
mG(Pi) = mG(Pi) + mG(lj)
else if t(li) ⊂ t(lj) then // property 2
Pi=Pi ∪lj
else if t(li) ⊂ t(lj) then // property 3
Remove Ij from [P]
Add X ×Y,mG(Ij) to [Pi]
else if t(li) ≠ t(1j) then // property 4
Add X ×Y ,∪ [mG(li), mG(lj)]to [Pi]
The MG-CHARM(D, minSup) function takes in
input Database D and the minimum support
necessary for the item sets . [Φ]- denotes the set of
all Item-Transaction pair having minimum support
The MG-CHARM-EXTEND([P], C) function is is
Recursive function until all possible mG’s are
generated. This function calls the MG-CHARM-
PROPERTY(XXY,li,lj,Pi, [Pi],[Pj])
Function where it states that
If t(Xi) = t(Xj) then c(Xi) = c(Xj) = c(Xi∪Xj).
If t(Xi) ⊂ t(Xj) then c(Xi) ≠ c(Xj) but c(Xi) =
c(Xi∪Xj).
If t(Xi) ⊃ t(Xj) then c(Xi) ≠ c(Xj) but c(Xj) =
c(Xi∪Xj).
lf t(Xi) ⊄ t(Xj) and t(Xj) ⊄ t(Xi) then c(Xi) ≠ c(Xj)
≠ c(Xi∪Xj).
The SUBSUMPTION-CHECK function checks
whether frequent itemset Pi is closed or not. If it is,
add it into C, otherwise it will be removed and its
mGs ,will be added to its parent closed. It uses hash-
table to store C, therefore the time of checking is
insignificant.
∪ [mG(li), mG(lj)] are minimal generators of Xi
∪Xj.
The algorithm stated in [11] has been used for the
effective implementation of the adaptive system in
data mining to predict and analyze the output as per
requirement
VIII. EXPERIMENTAL RESULTS AND
COMPARISONS
The experiment has been performed on a JAVA
Swing application coded on a Windows 7 OS, CPU:
2.93 GHz, 2 GB RAM
Let I = {i1, i2, …, in} be a set of items, T = {t1, t2,
…, tm} be a set of transaction identifiers (tids or
tidset) in a database D. The input database is a
binary relation I T. If an item i occurs in a
transaction t, we write it as (i,t) or it.
Consider the following transaction database
Transaction Items Bought
1 1 3 4
2 2 3 5
3 1 2 3 5
4 2 5
5 1 2 3 5
Table 1
4. Ms. Sonali Vaidya et al. Int. Journal of Engineering Research and Applications www.ijera.com
ISSN : 2248-9622, Vol. 5, Issue 3, ( Part -3) March 2015, pp.27-31
www.ijera.com 30 | P a g e
For the given transaction database the assumed
support is 50 % and confidence is 50 %. The
association rules generated for the assumption is
shown in Table 2:
Table 2
The results in Table 2 show a mere comparison of
the association rule mining algorithms in terms of
total time required for mining rules, total memory
required and the number of association rules
generated. MG-CHARM is faster than CHARM
which in turn is faster than the traditional Apriori
algorithm. The time required for mining FCI’s and
mining association from the FCI’s is sufficiently less
in case of MG-CHARM which makes it faster than
the prior algorithms. Also the space required is less
for the same datasets considered and for the assumed
support and confidence. Mining of non-redundant
association rules is also possible using the MG-
CHARM algorithm which facilitates generation of
non-redundant queries.
IX. RELATED WORK
3.1 In [10] the proposed method for mining mG’s of
FCI’s needs not generate candidates. Experiments
showed that the time of updating mG’s of frequent
closed itemsets is insignificant. Especially, in case of
the large size of closed itemsets, the time of updating
is very fewer. Some applications of mGs in mining
non-redundant association rules (NARs) based on
methods presented in [3,5].
3.2 Method of Bastide et al
In this section, we present a method for mining
minimal NARs (MINARs) based on FCI’s (Bastide
et al [3]). Assume we had all FCI’s from database D
satisfy minSup. Now, we want to mine MINARs. As
we mentioned above, MINARs only generate from
X →Y, where X is a minimal generator and Y is a
FCI. Bastide et al divided mining MINARs problem
into two sub-problems. Mining MINARs with the
confidence = 100% and mining MINARs with the
confidence < 100%.
Phase 1: Mining MINARs with the confidence =
100% for all g belongs to E mG's (in ascending
length), if g →y(g) then generate the rule { g →y(g)
g} (y(g) is closure of g).
Phase 2: Mining MINARs with the confidence <
100% for k = 1 to µ-1 do // µ is longest FCI’s for all
g belong to mG’s with |g| = k do generate rules from
g →Y g, where g is a subset of Y and Y is a FCI
(Y is not equal to (g)).
3.3 METHOD OF ZAKI
Zaki [5] mined the two kinds of rules. i) Rules with
the confidence = 100%: Self-rules (the rules that
generating from mG’s(X) to X, where X is a FCI)
and Down-rules (the rules that generating from
mG’s(Y) to mG’s(X), where X, Y are FCI’s, X is a
subset of Y). ii) Rules with the confidence < 100%:
From mG’s(X) to mG’s(Y), where X, Y belongs to
FCI’s, X is a subset of Y. Number of rules was
generated by this approach is linear with FCI’s.
X. CONCLUSION
In this paper, the presented MG-CHARM
algorithm is used for mining minimal generators of
frequent closed itemsets. By mere comparisons,
experimental study, and paper research, the
bottleneck identified with the CHARM algorithm is
that the number of frequent items is large and it
takes more time. To solve this problem the numbers
of items were decreased the iterations and new
comparison methodologies were used by enhancing
CHARM. The implementation proposed defines a
generic basis for exact association rules and
transitive reduction of the informative basis for
approximate association rules, non-redundant in
nature and does not represent any loss of information
from the user’s point of view. Visualization and
analysis of the association rules is possible.
Altogether the intended system can be made to work
and function in a dynamic environment. The project
has future scope when very large numbers of
datasets need to be taken into consideration. The
software will be fruitful and show effective
performance when used with a Big Data Database.
Providing suggestions to the user on the basis of a
dynamic scenario will be put into effect in the future.
Acknowledgements
We hereby take the privilege to present our project
report on “Adaptive and Fast Predictions by
Minimal Itemsets Creation”. We are very grateful
5. Ms. Sonali Vaidya et al. Int. Journal of Engineering Research and Applications www.ijera.com
ISSN : 2248-9622, Vol. 5, Issue 3, ( Part -3) March 2015, pp.27-31
www.ijera.com 31 | P a g e
to our Project Supervisor Ms. Sonali Vaidya for
contributing her valuable moments in the Project
from her busy and hectic schedule right from the
Project’s inception. Being after us like a true mentor
and a great academic parent.
We are very thankful to Ms. Sonali Vaidya whose
guidance and support was an immense motivation
for us to carry on with our Project. She has been a
constant source of inspiration to us. Her suggestions
have greatly contributed for the betterment of our
project.
Our special thanks to the Head of Department Mr.
Pramod Shanbhag, staff members and lab
assistants for their co-operation.
REFERENCES
[1] R. Agrawal, R. Srikant: Fast algorithms for
mining association rules. In: VLDB'94,
pp(1994).
[2] R. Agrawal, T. Imielinski, A. Swami: Mining
association rules between sets of items in large
databases. In: SIGMOD'93, pp. 207 - 216
(1993).
[3] Y. Bastide, N. Pasquier, R. Taouil, G. Stumme,
L.Lakhal: Mining Minimal Non-Redundant
Association Rules using Closed Frequent
Itemssets. In: 1st International Conference on
Computational Logic, pp. 972 - 986 (2000).
[4] M. 1. Zaki, C.J. Hsiao: Efficient Algorithms for
Mining Closed Itemsets and Their Lattice
Structure. In: IEEE Transactions on
Knowledge and Data Engineering, Vol. 17, No
4, April 2005, pp. 462-478 (2005).
[5] M. 1. Zaki: Mining Non-Redundant
Association Rules, Data Mining and
Knowledge Discovery, 9, 223-248, 2004
Kluwer Academic Publishers. Manufactured in
The Netherlands, pp. 223-248 (2004).
[6] M. J. Zaki, K. Gouda: Fast Vertical Mining
Using Diffsets. In: Proc. Ninth ACM SIGKDD
Int'l Conf.Knowledge Discovery and Data
Mining, pp. 326 335 (2003).
[7] M. J. Zaki, B. Phoophakdee: MIRAGE: A
Framework for Mining, Exploring, and
Visualizing Minimal Association Rules. In:
Technical Report 03-4, Computer Science
Dept., Rensselaer Polytechnic Inst., July
(2003).
[8] M. J. Zaki: Generating Non-Redundant
Association Rules. In: 6th ACM SIGKDD Intl
Conf. Knowledge Discovery and Data Mining,
(2000).
[9] S. B. Yahia, T. Hamrouni, E. M. Nguifo:
Frequent Closed Itemset based Algorithms: A
thorough Structural and Analytical Survey. In:
ACM SIGKDD Explorations Newsletter 8 (1),
pp. 93 - 104 (2006).
[10]Bay Vo, Bac Le: Fast Algorithm for Mining
Minimal Generators of Frequent Closed
Itemsets and Their Applications: A thorough
Structural and Analytical Survey. year-2009.