The document summarizes a study that models the view selection problem in data warehousing as a weighted constraint satisfaction problem. The researchers encode the view selection problem as a weighted constraint satisfaction problem to find an optimal set of views to materialize. They use the multiple view processing plan framework as the search space and propose a genetic algorithm to select the views to materialize. The experimental results show that the proposed algorithm can effectively select appropriate views for materialization.
MULTI-OBJECTIVE ENERGY EFFICIENT OPTIMIZATION ALGORITHM FOR COVERAGE CONTROL ...ijcseit
Many studies have been done in the area of Wireless Sensor Networks (WSNs) in recent years. In this kind of networks, some of the key objectives that need to be satisfied are area coverage, number of active sensors and energy consumed by nodes. In this paper, we propose a NSGA-II based multi-objective algorithm for optimizing all of these objectives simultaneously. The efficiency of our algorithm is demonstrated in the simulation results. This efficiency can be shown as finding the optimal balance point among the maximum coverage rate, the least energy consumption, and the minimum number of active nodes while maintaining the connectivity of the network
Decision tree clustering a columnstores tuple reconstructioncsandit
Column-Stores has gained market share due to promi
sing physical storage alternative for
analytical queries. However, for multi-attribute qu
eries column-stores pays performance
penalties due to on-the-fly tuple reconstruction. T
his paper presents an adaptive approach for
reducing tuple reconstruction time. Proposed approa
ch exploits decision tree algorithm to
cluster attributes for each projection and also eli
minates frequent database scanning.
Experimentations with TPC-H data shows the effectiv
eness of proposed approach.
Reversible Enhancement of Reversibly Data Embedded Images using HVS Character...CSCJournals
In the recent, HVS based reversible data embedding is emerging. In a reversible data embedding scheme, in the process of achieving reversibility and embedding the data, it induces some distortions into marked image/video. If the distortions in marked image are random due to reversible embedding, it may not reveal much information about the existence of hidden data; on the contrary, if the distortions are repetitive geometric patterns which arises when an embedding scheme's emphasis is on capacity and reversibility, then the marked media with geometric distortions may reveal the existence of the hidden data. Further, in applications involving satellite images, medical, military, fine arts images, etc. which involves reversible embedding, it requires to show the details for inspection. Due to repetitive geometric distortions as in [6] and [7], the visual quality of image degrades. Hence, there is a need for enhancing the reversibly embedded images. To tackle this issue, if any exiting image enhancement scheme is used, it incur in loss of hidden data as well as the restoration of image to original form is not possible; hence, violating the requirement of reversibility for such applications. To address this problem we propose a reversible image enhancement scheme for reversibly data embedded images by using HVS characteristics of DCT. The experimental results show that the visual quality has been improved in terms of PSNR, PSNR-HVS, PSNR-HVS-M, and MSSIM when compared to S. Gujjunoori et al. scheme [6] and S. Gujjunoori et al. scheme [7], which contains repetitive geometrical distortions in marked media.
MULTI-OBJECTIVE ENERGY EFFICIENT OPTIMIZATION ALGORITHM FOR COVERAGE CONTROL ...ijcseit
Many studies have been done in the area of Wireless Sensor Networks (WSNs) in recent years. In this kind of networks, some of the key objectives that need to be satisfied are area coverage, number of active sensors and energy consumed by nodes. In this paper, we propose a NSGA-II based multi-objective algorithm for optimizing all of these objectives simultaneously. The efficiency of our algorithm is demonstrated in the simulation results. This efficiency can be shown as finding the optimal balance point among the maximum coverage rate, the least energy consumption, and the minimum number of active nodes while maintaining the connectivity of the network
Decision tree clustering a columnstores tuple reconstructioncsandit
Column-Stores has gained market share due to promi
sing physical storage alternative for
analytical queries. However, for multi-attribute qu
eries column-stores pays performance
penalties due to on-the-fly tuple reconstruction. T
his paper presents an adaptive approach for
reducing tuple reconstruction time. Proposed approa
ch exploits decision tree algorithm to
cluster attributes for each projection and also eli
minates frequent database scanning.
Experimentations with TPC-H data shows the effectiv
eness of proposed approach.
Reversible Enhancement of Reversibly Data Embedded Images using HVS Character...CSCJournals
In the recent, HVS based reversible data embedding is emerging. In a reversible data embedding scheme, in the process of achieving reversibility and embedding the data, it induces some distortions into marked image/video. If the distortions in marked image are random due to reversible embedding, it may not reveal much information about the existence of hidden data; on the contrary, if the distortions are repetitive geometric patterns which arises when an embedding scheme's emphasis is on capacity and reversibility, then the marked media with geometric distortions may reveal the existence of the hidden data. Further, in applications involving satellite images, medical, military, fine arts images, etc. which involves reversible embedding, it requires to show the details for inspection. Due to repetitive geometric distortions as in [6] and [7], the visual quality of image degrades. Hence, there is a need for enhancing the reversibly embedded images. To tackle this issue, if any exiting image enhancement scheme is used, it incur in loss of hidden data as well as the restoration of image to original form is not possible; hence, violating the requirement of reversibility for such applications. To address this problem we propose a reversible image enhancement scheme for reversibly data embedded images by using HVS characteristics of DCT. The experimental results show that the visual quality has been improved in terms of PSNR, PSNR-HVS, PSNR-HVS-M, and MSSIM when compared to S. Gujjunoori et al. scheme [6] and S. Gujjunoori et al. scheme [7], which contains repetitive geometrical distortions in marked media.
Introduction to Multi-Objective Clustering EnsembleIJSRD
Association rule mining is a popular and well researched method for discovering interesting relations between variables in large databases. In this paper we introduce the concept of Data mining, Association rule and Multilevel association rule with different algorithm, its advantage and concept of Fuzzy logic and Genetic Algorithm. Multilevel association rules can be mined efficiently using concept hierarchies under a support-confidence framework.
Improved probabilistic distance based locality preserving projections method ...IJECEIAES
In this paper, a dimensionality reduction is achieved in large datasets using the proposed distance based Non-integer Matrix Factorization (NMF) technique, which is intended to solve the data dimensionality problem. Here, NMF and distance measurement aim to resolve the non-orthogonality problem due to increased dataset dimensionality. It initially partitions the datasets, organizes them into a defined geometric structure and it avoids capturing the dataset structure through a distance based similarity measurement. The proposed method is designed to fit the dynamic datasets and it includes the intrinsic structure using data geometry. Therefore, the complexity of data is further avoided using an Improved Distance based Locality Preserving Projection. The proposed method is evaluated against existing methods in terms of accuracy, average accuracy, mutual information and average mutual information.
Support Vector Machine–Based Prediction System for a Football Match Resultiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Securing Biometric Images using Reversible WatermarkingCSCJournals
Biometric security is a fast growing area. Protecting biometric data is very important since it can be misused by attackers. In order to increase security of biometric data there are different methods in which watermarking is widely accepted. A more acceptable, new important development in this area is reversible watermarking in which the original image can be completely restored and the watermark can be retrieved. But reversible watermarking in biometrics is an understudied area. Reversible watermarking maintains high quality of biometric data. This paper proposes Rotational Replacement of LSB as a reversible watermarking scheme for biometric images. PSNR is the regular method used for quality measurement of biometric data. In this paper we also prove that SSIM Index can be used as a better alternate for effective quality assessment for reversible watermarked biometric data by comparing with the well known reversible watermarking scheme using Difference Expansion.
Among many data clustering approaches available today, mixed data set of numeric and category data
poses a significant challenge due to difficulty of an appropriate choice and employment of
distance/similarity functions for clustering and its verification. Unsupervised learning models for
artificial neural network offers an alternate means for data clustering and analysis. The objective of this
study is to highlight an approach and its associated considerations for mixed data set clustering with
Adaptive Resonance Theory 2 (ART-2) artificial neural network model and subsequent validation of the
clusters with dimensionality reduction using Autoencoder neural network model.
Web image annotation by diffusion maps manifold learning algorithmijfcstjournal
Automatic image annotation is one of the most challenging problems in machine vision areas. The goal of this task is to predict number of keywords automatically for images captured in real data. Many methods are based on visual features in order to calculate similarities between image samples. But the computation cost of these approaches is very high. These methods require many training samples to be stored in memory. To lessen thisburden, a number of techniques have been developed to reduce the number
of features in a dataset. Manifold learning is a popular approach to nonlinear dimensionality reduction. In
this paper, we investigate Diffusion maps manifold learning method for webimage auto-annotation task.Diffusion maps
manifold learning method isused to reduce the dimension of some visual features. Extensive experiments and analysis onNUS-WIDE-LITE web image dataset with
different visual featuresshow how this manifold learning dimensionality reduction method can be applied effectively to image annotation.
Most of the existing co-segmentation methods are usually
complex, and require pre-grouping of images, fine-tuning a
few parameters and initial segmentation masks etc. These
limitations become serious concerns for their application on
large scale datasets. In this paper, Group Saliency Propagation
(GSP) model is proposed where a single group saliency
map is developed, which can be propagated to segment the
entire group. In addition, it is also shown how a pool of these
group saliency maps can help in quickly segmenting new
input images. Experiments demonstrate that the proposed
method can achieve competitive performance on several
benchmark co-segmentation datasets including ImageNet,
with the added advantage of speed up.
A parsimonious SVM model selection criterion for classification of real-world ...o_almasi
This paper proposes and optimizes a two-term cost function consisting of a sparseness term and a generalized v-fold cross-validation term by a new adaptive particle swarm optimization (APSO). APSO updates its parameters adaptively based on a dynamic feedback from the success rate of the each particle’s personal best. Since the proposed cost function is based on the choosing fewer numbers of support vectors, the complexity of SVM models decreased while the accuracy remains in an acceptable range. Therefore, the testing time decreases and makes SVM more applicable for practical applications in real data sets. A comparative study on data sets of UCI database is performed between the proposed cost function and conventional cost function to demonstrate the effectiveness of the proposed cost function.
MARKOV CHAIN FOR THE RECOMMENDATION OF MATERIALIZED VIEWS IN REAL-TIME DATA W...IJCSEA Journal
In this paper we propose an approach, which is based on Markov Chain, to cluster and recommend candidate views for the selection algorithm of materialized views. Our idea is to intervene at regular period of time in order to filter the candidate views which will be used by an algorithm for the selection of materialized views in real-time data warehouse. The aim is to reduce the complexity and the execution cost of the online selection of materialized views. Our experiment results have shown that our solution is very efficient to specify the more profitable views and to improve the query response time.
Markov Chain for the Recommendation of Materialized Views in Real-Time Data W...IJCSEA Journal
In this paper we propose an approach, which is based on Markov Chain, to cluster and recommend candidate views for the selection algorithm of materialized views. Our idea is to intervene at regular period of time in order to filter the candidate views which will be used by an algorithm for the selection of materialized views in real-time data warehouse. The aim is to reduce the complexity and the execution cost of the online selection of materialized views. Our experiment results have shown that our solution is very efficient to specify the more profitable views and to improve the query response time.
In this paper we propose an approach, which is based on Markov Chain, to cluster and recommend
candidate views for the selection algorithm of materialized views. Our idea is to intervene at regular period
of time in order to filter the candidate views which will be used by an algorithm for the selection of
materialized views in real-time data warehouse. The aim is to reduce the complexity and the execution cost
of the online selection of materialized views. Our experiment results have shown that our solution is very
efficient to specify the more profitable views and to improve the query response time.
TASK SCHEDULING USING AMALGAMATION OF MET HEURISTICS SWARM OPTIMIZATION ALGOR...Journal For Research
Cloud Computing is the latest networking technology and also popular archetype for hosting the application and delivering of services over the network. The foremost technology of the cloud computing is virtualization which enables of building the applications, dynamically sharing of resources and providing diverse services to the cloud users. With virtualization, a service provider can guarantee Quality of Service to the user at the same time as achieving higher server consumption and energy competence. One of the most important challenges in the cloud computing environment is the VM placemnt and task scheduling problem. This paper focus on Metaheuristic Swarm Optimisation Algorithms(MSOA) deals with the problem of VM placement and Task scheduling in cloud environment. The MSOA is a simple parallel algorithm that can be applied in different ways to resolve the task scheduling problems. The proposed algorithm is considered an amalgamation of the SO algorithm and the Cuckoo search (CS) algorithm; called MSOACS. The proposed algorithm is evaluated using Cloudsim Simulator. The results proves the reduction of the makespan and increase the utilization ratio of the proposed MSOACS algorithm compared with SOA algorithms and Randomised Allocation Allocation (RA).
A quality of service management in distributed feedback control scheduling ar...csandit
In our days, there are many real-time applications that use data which are geographically
dispersed. The distributed real-time database management systems (DRTDBMS) have been used
to manage large amount of distributed data while meeting the stringent temporal requirements
in real-time applications. Providing Quality of Service (QoS) guarantees in DRTDBMSs is a
challenging task. To address this problem, different research works are often based on
distributed feedback control real-time scheduling architecture (DFCSA). Data replication is an
efficient method to help DRTDBMS meeting the stringent temporal requirements of real-time
applications. In literature, many works have designed algorithms that provide QoS guarantees
in distributed real-time databases using only full temporal data replication policy.
In this paper, we have applied two data replication policies in a distributed feedback control
scheduling architecture to manage a QoS performance for DRTDBMS. The first proposed data
replication policy is called semi-total replication, and the second is called partial replication
policy.
QUALITY OF SERVICE MANAGEMENT IN DISTRIBUTED FEEDBACK CONTROL SCHEDULING ARCH...cscpconf
In our days, there are many real-time applications that use data which are geographically
dispersed. The distributed real-time database management systems (DRTDBMS) have been used
to manage large amount of distributed data while meeting the stringent temporal requirements
in real-time applications. Providing Quality of Service (QoS) guarantees in DRTDBMSs is a
challenging task. To address this problem, different research works are often based on
distributed feedback control real-time scheduling architecture (DFCSA). Data replication is an
efficient method to help DRTDBMS meeting the stringent temporal requirements of real-time
applications. In literature, many works have designed algorithms that provide QoS guarantees
in distributed real-time databases using only full temporal data replication policy.
In this paper, we have applied two data replication policies in a distributed feedback control
scheduling architecture to manage a QoS performance for DRTDBMS. The first proposed data
replication policy is called semi-total replication, and the second is called partial replication
policy.
Introduction to Multi-Objective Clustering EnsembleIJSRD
Association rule mining is a popular and well researched method for discovering interesting relations between variables in large databases. In this paper we introduce the concept of Data mining, Association rule and Multilevel association rule with different algorithm, its advantage and concept of Fuzzy logic and Genetic Algorithm. Multilevel association rules can be mined efficiently using concept hierarchies under a support-confidence framework.
Improved probabilistic distance based locality preserving projections method ...IJECEIAES
In this paper, a dimensionality reduction is achieved in large datasets using the proposed distance based Non-integer Matrix Factorization (NMF) technique, which is intended to solve the data dimensionality problem. Here, NMF and distance measurement aim to resolve the non-orthogonality problem due to increased dataset dimensionality. It initially partitions the datasets, organizes them into a defined geometric structure and it avoids capturing the dataset structure through a distance based similarity measurement. The proposed method is designed to fit the dynamic datasets and it includes the intrinsic structure using data geometry. Therefore, the complexity of data is further avoided using an Improved Distance based Locality Preserving Projection. The proposed method is evaluated against existing methods in terms of accuracy, average accuracy, mutual information and average mutual information.
Support Vector Machine–Based Prediction System for a Football Match Resultiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Securing Biometric Images using Reversible WatermarkingCSCJournals
Biometric security is a fast growing area. Protecting biometric data is very important since it can be misused by attackers. In order to increase security of biometric data there are different methods in which watermarking is widely accepted. A more acceptable, new important development in this area is reversible watermarking in which the original image can be completely restored and the watermark can be retrieved. But reversible watermarking in biometrics is an understudied area. Reversible watermarking maintains high quality of biometric data. This paper proposes Rotational Replacement of LSB as a reversible watermarking scheme for biometric images. PSNR is the regular method used for quality measurement of biometric data. In this paper we also prove that SSIM Index can be used as a better alternate for effective quality assessment for reversible watermarked biometric data by comparing with the well known reversible watermarking scheme using Difference Expansion.
Among many data clustering approaches available today, mixed data set of numeric and category data
poses a significant challenge due to difficulty of an appropriate choice and employment of
distance/similarity functions for clustering and its verification. Unsupervised learning models for
artificial neural network offers an alternate means for data clustering and analysis. The objective of this
study is to highlight an approach and its associated considerations for mixed data set clustering with
Adaptive Resonance Theory 2 (ART-2) artificial neural network model and subsequent validation of the
clusters with dimensionality reduction using Autoencoder neural network model.
Web image annotation by diffusion maps manifold learning algorithmijfcstjournal
Automatic image annotation is one of the most challenging problems in machine vision areas. The goal of this task is to predict number of keywords automatically for images captured in real data. Many methods are based on visual features in order to calculate similarities between image samples. But the computation cost of these approaches is very high. These methods require many training samples to be stored in memory. To lessen thisburden, a number of techniques have been developed to reduce the number
of features in a dataset. Manifold learning is a popular approach to nonlinear dimensionality reduction. In
this paper, we investigate Diffusion maps manifold learning method for webimage auto-annotation task.Diffusion maps
manifold learning method isused to reduce the dimension of some visual features. Extensive experiments and analysis onNUS-WIDE-LITE web image dataset with
different visual featuresshow how this manifold learning dimensionality reduction method can be applied effectively to image annotation.
Most of the existing co-segmentation methods are usually
complex, and require pre-grouping of images, fine-tuning a
few parameters and initial segmentation masks etc. These
limitations become serious concerns for their application on
large scale datasets. In this paper, Group Saliency Propagation
(GSP) model is proposed where a single group saliency
map is developed, which can be propagated to segment the
entire group. In addition, it is also shown how a pool of these
group saliency maps can help in quickly segmenting new
input images. Experiments demonstrate that the proposed
method can achieve competitive performance on several
benchmark co-segmentation datasets including ImageNet,
with the added advantage of speed up.
A parsimonious SVM model selection criterion for classification of real-world ...o_almasi
This paper proposes and optimizes a two-term cost function consisting of a sparseness term and a generalized v-fold cross-validation term by a new adaptive particle swarm optimization (APSO). APSO updates its parameters adaptively based on a dynamic feedback from the success rate of the each particle’s personal best. Since the proposed cost function is based on the choosing fewer numbers of support vectors, the complexity of SVM models decreased while the accuracy remains in an acceptable range. Therefore, the testing time decreases and makes SVM more applicable for practical applications in real data sets. A comparative study on data sets of UCI database is performed between the proposed cost function and conventional cost function to demonstrate the effectiveness of the proposed cost function.
MARKOV CHAIN FOR THE RECOMMENDATION OF MATERIALIZED VIEWS IN REAL-TIME DATA W...IJCSEA Journal
In this paper we propose an approach, which is based on Markov Chain, to cluster and recommend candidate views for the selection algorithm of materialized views. Our idea is to intervene at regular period of time in order to filter the candidate views which will be used by an algorithm for the selection of materialized views in real-time data warehouse. The aim is to reduce the complexity and the execution cost of the online selection of materialized views. Our experiment results have shown that our solution is very efficient to specify the more profitable views and to improve the query response time.
Markov Chain for the Recommendation of Materialized Views in Real-Time Data W...IJCSEA Journal
In this paper we propose an approach, which is based on Markov Chain, to cluster and recommend candidate views for the selection algorithm of materialized views. Our idea is to intervene at regular period of time in order to filter the candidate views which will be used by an algorithm for the selection of materialized views in real-time data warehouse. The aim is to reduce the complexity and the execution cost of the online selection of materialized views. Our experiment results have shown that our solution is very efficient to specify the more profitable views and to improve the query response time.
In this paper we propose an approach, which is based on Markov Chain, to cluster and recommend
candidate views for the selection algorithm of materialized views. Our idea is to intervene at regular period
of time in order to filter the candidate views which will be used by an algorithm for the selection of
materialized views in real-time data warehouse. The aim is to reduce the complexity and the execution cost
of the online selection of materialized views. Our experiment results have shown that our solution is very
efficient to specify the more profitable views and to improve the query response time.
TASK SCHEDULING USING AMALGAMATION OF MET HEURISTICS SWARM OPTIMIZATION ALGOR...Journal For Research
Cloud Computing is the latest networking technology and also popular archetype for hosting the application and delivering of services over the network. The foremost technology of the cloud computing is virtualization which enables of building the applications, dynamically sharing of resources and providing diverse services to the cloud users. With virtualization, a service provider can guarantee Quality of Service to the user at the same time as achieving higher server consumption and energy competence. One of the most important challenges in the cloud computing environment is the VM placemnt and task scheduling problem. This paper focus on Metaheuristic Swarm Optimisation Algorithms(MSOA) deals with the problem of VM placement and Task scheduling in cloud environment. The MSOA is a simple parallel algorithm that can be applied in different ways to resolve the task scheduling problems. The proposed algorithm is considered an amalgamation of the SO algorithm and the Cuckoo search (CS) algorithm; called MSOACS. The proposed algorithm is evaluated using Cloudsim Simulator. The results proves the reduction of the makespan and increase the utilization ratio of the proposed MSOACS algorithm compared with SOA algorithms and Randomised Allocation Allocation (RA).
A quality of service management in distributed feedback control scheduling ar...csandit
In our days, there are many real-time applications that use data which are geographically
dispersed. The distributed real-time database management systems (DRTDBMS) have been used
to manage large amount of distributed data while meeting the stringent temporal requirements
in real-time applications. Providing Quality of Service (QoS) guarantees in DRTDBMSs is a
challenging task. To address this problem, different research works are often based on
distributed feedback control real-time scheduling architecture (DFCSA). Data replication is an
efficient method to help DRTDBMS meeting the stringent temporal requirements of real-time
applications. In literature, many works have designed algorithms that provide QoS guarantees
in distributed real-time databases using only full temporal data replication policy.
In this paper, we have applied two data replication policies in a distributed feedback control
scheduling architecture to manage a QoS performance for DRTDBMS. The first proposed data
replication policy is called semi-total replication, and the second is called partial replication
policy.
QUALITY OF SERVICE MANAGEMENT IN DISTRIBUTED FEEDBACK CONTROL SCHEDULING ARCH...cscpconf
In our days, there are many real-time applications that use data which are geographically
dispersed. The distributed real-time database management systems (DRTDBMS) have been used
to manage large amount of distributed data while meeting the stringent temporal requirements
in real-time applications. Providing Quality of Service (QoS) guarantees in DRTDBMSs is a
challenging task. To address this problem, different research works are often based on
distributed feedback control real-time scheduling architecture (DFCSA). Data replication is an
efficient method to help DRTDBMS meeting the stringent temporal requirements of real-time
applications. In literature, many works have designed algorithms that provide QoS guarantees
in distributed real-time databases using only full temporal data replication policy.
In this paper, we have applied two data replication policies in a distributed feedback control
scheduling architecture to manage a QoS performance for DRTDBMS. The first proposed data
replication policy is called semi-total replication, and the second is called partial replication
policy.
An ontological approach to handle multidimensional schema evolution for data ...ijdms
In recent years, the number of digital information storage and retrieval systems has increased immensely.
Data warehousing has been found to be an extremely useful technology for integrating such heterogeneous
and autonomous information sources. Data within the data warehouse is modelled in the form of a star or
snowflake schema which facilitates business analysis in a multidimensional perspective. As user
requirements are interesting measures of business processes, the data warehouse schema is derived from
the information sources and business requirements. Due to the changing business scenario, the information
sources not only change their data, but also change their schema structure. In addition to the source
changes the business requirements for data warehouse may also change. Both these changes results in data
warehouse schema evolution. These changes can be handled either by just updating it in the DW model, or
can be developed as a new version of the DW structure. Existing approaches either deal with source
changes or requirements changes in a manual way and changes to the data warehouse schema is carried
out at the physical level. This may induce high maintenance costs and complex OLAP server
administration. As ontology seems to be a promising solution for the data warehouse research, in this
paper an ontological approach to automate the evolution of a data warehouse schema is proposed. This
method assists the data warehouse designer in handling evolution at the ontological level based on which
decision can be made to carry out the changes at the physical level. We evaluate the proposed ontological
approach with the existing method of manual adaptation of data warehouse schema.
The role of materialized views is becoming vital in today’s distributed Data warehouses. Materialization is
where parts of the data cube are pre-computed. Some of the real time distributed architectures are
maintaining materialization transparencies in the sense the users are not known with the materialization at
a node. Usually what all followed by them is a cache maintenance mechanism where the query results are
cached. When a query requesting materialization arrives at a distributed node it checks in its cache and if
the materialization is available answers the query. What if materialization is not available- the node
communicates the query in the network until a node answering the requested materialization is available.
This type of network communication increases the number of query forwarding’s between nodes. The aim
of this paper is to reduce the multiple redirects. In this paper we propose a new CB-pattern tree indexing to
identify the exact distributed node where the needed materialization is available.
FAST ALGORITHMS FOR UNSUPERVISED LEARNING IN LARGE DATA SETScsandit
The ability to mine and extract useful information automatically, from large datasets, is a
common concern for organizations (having large datasets), over the last few decades. Over the
internet, data is vastly increasing gradually and consequently the capacity to collect and store
very large data is significantly increasing.
Existing clustering algorithms are not always efficient and accurate in solving clustering
problems for large datasets.
However, the development of accurate and fast data classification algorithms for very large
scale datasets is still a challenge. In this paper, various algorithms and techniques especially,
approach using non-smooth optimization formulation of the clustering problem, are proposed
for solving the minimum sum-of-squares clustering problems in very large datasets. This
research also develops accurate and real time L2-DC algorithm based with the incremental
approach to solve the minimum
A BI-OBJECTIVE MODEL FOR SVM WITH AN INTERACTIVE PROCEDURE TO IDENTIFY THE BE...gerogepatton
A support vector machine (SVM) learns the decision surface from two different classes of the input points, there are misclassifications in some of the input points in several applications. In this paper a bi-objective quadratic programming model is utilized and different feature quality measures are optimized simultaneously using the weighting method for solving our bi-objective quadratic programming problem. An important contribution will be added for the proposed bi-objective quadratic programming model by getting different efficient support vectors due to changing the weighting values. The numerical examples, give evidence of the effectiveness of the weighting parameters on reducing the misclassification between two classes of the input points. An interactive procedure will be added to identify the best compromise solution from the generated efficient solutions.
A BI-OBJECTIVE MODEL FOR SVM WITH AN INTERACTIVE PROCEDURE TO IDENTIFY THE BE...ijaia
A support vector machine (SVM) learns the decision surface from two different classes of the input points, there are misclassifications in some of the input points in several applications. In this paper a bi-objective quadratic programming model is utilized and different feature quality measures are optimized simultaneously using the weighting method for solving our bi-objective quadratic programming problem. An important contribution will be added for the proposed bi-objective quadratic programming model by getting different efficient support vectors due to changing the weighting values. The numerical examples, give evidence of the effectiveness of the weighting parameters on reducing the misclassification between two classes of the input points. An interactive procedure will be added to identify the best compromise solution from the generated efficient solutions.
Query optimization in oodbms identifying subquery for query managementijdms
This paper is based on relatively newer approach for query optimization in object databases, which uses
query decomposition and cached query results to improve execution a query. Issues that are focused here is
fast retrieval and high reuse of cached queries, Decompose Query into Sub query, Decomposition of
complex queries into smaller for fast retrieval of result.
Here we try to address another open area of query caching like handling wider queries. By using some
parts of cached results helpful for answering other queries (wider Queries) and combining many cached
queries while producing the result.
Multiple experiments were performed to prove the productivity of this newer way of optimizing a query.
The limitation of this technique is that it’s useful especially in scenarios where data manipulation rate is
very low as compared to data retrieval rate.
Unit 8 - Information and Communication Technology (Paper I).pdfThiyagu K
This slides describes the basic concepts of ICT, basics of Email, Emerging Technology and Digital Initiatives in Education. This presentations aligns with the UGC Paper I syllabus.
Read| The latest issue of The Challenger is here! We are thrilled to announce that our school paper has qualified for the NATIONAL SCHOOLS PRESS CONFERENCE (NSPC) 2024. Thank you for your unwavering support and trust. Dive into the stories that made us stand out!
The French Revolution, which began in 1789, was a period of radical social and political upheaval in France. It marked the decline of absolute monarchies, the rise of secular and democratic republics, and the eventual rise of Napoleon Bonaparte. This revolutionary period is crucial in understanding the transition from feudalism to modernity in Europe.
For more information, visit-www.vavaclasses.com
Palestine last event orientationfvgnh .pptxRaedMohamed3
An EFL lesson about the current events in Palestine. It is intended to be for intermediate students who wish to increase their listening skills through a short lesson in power point.
Model Attribute Check Company Auto PropertyCeline George
In Odoo, the multi-company feature allows you to manage multiple companies within a single Odoo database instance. Each company can have its own configurations while still sharing common resources such as products, customers, and suppliers.
Macroeconomics- Movie Location
This will be used as part of your Personal Professional Portfolio once graded.
Objective:
Prepare a presentation or a paper using research, basic comparative analysis, data organization and application of economic information. You will make an informed assessment of an economic climate outside of the United States to accomplish an entertainment industry objective.
Operation “Blue Star” is the only event in the history of Independent India where the state went into war with its own people. Even after about 40 years it is not clear if it was culmination of states anger over people of the region, a political game of power or start of dictatorial chapter in the democratic setup.
The people of Punjab felt alienated from main stream due to denial of their just demands during a long democratic struggle since independence. As it happen all over the word, it led to militant struggle with great loss of lives of military, police and civilian personnel. Killing of Indira Gandhi and massacre of innocent Sikhs in Delhi and other India cities was also associated with this movement.
Instructions for Submissions thorugh G- Classroom.pptxJheel Barad
This presentation provides a briefing on how to upload submissions and documents in Google Classroom. It was prepared as part of an orientation for new Sainik School in-service teacher trainees. As a training officer, my goal is to ensure that you are comfortable and proficient with this essential tool for managing assignments and fostering student engagement.
Instructions for Submissions thorugh G- Classroom.pptx
WEIGHTED CONSTRAINT SATISFACTION AND GENETIC ALGORITHM TO SOLVE THE VIEW SELECTION PROBLEM
1. International Journal of Database Management Systems ( IJDMS ) Vol.9, No.4, June 2017
DOI : 10.5121/ijdms.2017.9402 11
WEIGHTED CONSTRAINT SATISFACTION AND
GENETIC ALGORITHM TO SOLVE THE VIEW
SELECTION PROBLEM
Mohammed El Alaoui1
, Karim El moutaouakil2
and Mohamed Ettaouil1
1
Modelling and Scientific Computing Laboratory University Sidi Mohammed Ben
Abdellah, Fez, MOROCCO
2
National school of applied sciences Al-Hoceima (ENSAH) BP 03, Ajdir Al-Hoceima
ABSTRACT
A Data warehouse is a tool that is used by big companies, and it gathers data coming from different
sources. The main goal of a data warehouse is not only to store data, but also to help companies to make
decisions. The huge volume of data makes processing queries complex and time-consuming. In order to
solve this problem, the materialization of views is suggested as a solution to improve the processing of the
queries. The materialization of views aims to optimize an objective function which is a compromise
between the cost of processing queries and the cost of maintenance under a storage space constraint. In
this work, we modelled the view selection problem as a weight constraint satisfaction problem. In
addition, we use the used the multiple view processing plans (MVPP) Framework as a search space, and
we call genetic algorithm to select views to be materialized. According to experimental result, the
proposed algorithm has been used to show the quality of the appropriate materialized views selection.
KEYWORDS
Data warehouse, multiple view processing plans, genetic algorithm, weighted constraint satisfaction
problem, storage space constraint.
1. INTRODUCTION
Data Warehouse DW can be defined by different ways, but the definition that is most
commonly used consists of considering it as being subject-oriented, integrated, non-volatile and
time-variant collection of data that supports management's decision of any given entity. It helps
store a vast volume of data to be used across the enterprise for decision making. Some
intermediate results in the query processing are stored in the DW in order to increase the
efficiency of the queries requested to a DW and as well as to avoid accessing the original data
sources continuously. These intermediate results are known as materialized views, which help
providing faster access to data, thus enhance the query performance. Undoubtedly,
materialization of views minimizes query response time. However, materializing all possible
views or leaving them all virtual can lead to two extreme and opposite outcomes. In fact, if all
views are materialized in a DW, that may give the best performance but at the highest cost of
view maintenance. Whereas, if all views are left virtual (i.e. no intermediate result saved in the
DW), that will have the lowest view maintenance cost but the poorest query performance.
Hence, the optimal choice between the two options is to materialize some of the views while
leaving others virtual. At this point, we should pose the following question: what are all views
that must be materialized to get the most optimal (or almost optimal) solution? One of the
solutions consists of selecting a set of derived views to materialize that simultaneously
minimizes the sum of total query response time as well as the maintenance cost of the selected
2. International Journal of Database Management Systems ( IJDMS ) Vol.9, No.4, June 2017
12
views. This is known as View Selection Problem VSP. The selection of views to be materialized
is one of the crucial decisions in the design of data warehouse. That is in fact NP-hard problem
because of the fact that the solution space grows exponentially as the problem size increases
[1],[2]. Four search space representations have been proposed, for the selection of views to
materialize in data warehouse, such as lattice representations of views[3],[4],[5], [6],[7],[8],[9].
AND-OR view graph representations of views [10],[11],[12]. Multiple View Processing Plan
representation [13],[14] and data mining representation [15],[16]. Multidimensional Lattice
representation of views is used to express dependencies among views; [3] used a lattice
representation to express dependencies among different views of the data cube to handle the
problem of view selection for materializing in data warehouse. Using this framework a greedy
algorithm is used to pick a set of views with the maximum benefit. [17] proposed Particle
Swarm optimization algorithm for materialized cube selection. AND-OR view graphs
representations were introduced to represent all the possible ways to generate warehouse views
such that the best query path can be utilized to optimize query. [10] have chosen this
framework for the selection of materialized views. Multiple views processing plan (MVPP)
representation is defined by [18] for the same problem but their works ignored storage costs.
[14] Proposed simulated annealing for materialized view selection in data warehousing
environment. Data mining representation, this approach is based on detection of common sub-
expressions, and represented workload as a binary matrix, in this matrix, each row represents a
query and each column is an attribute. This data mining techniques is used for view selection
problem. [15] have used data mining techniques are applied to this matrix to obtain a set of
candidate views for materializing. In this work, we used MVPP as a search space to obtain an
optimal materialized view selection, such that the storage cost is addressed in this work.
Multiple views processing plan representation is used to exploit the common sub-expressions
that can be detected among the queries. The leaf nodes correspond to the base relations and the
root nodes corresponds to warehouse queries. In this step of building MVPP search space we
generate the global plan of all queries to find an appropriate set of views a set of views to
materialize. The basic idea behind a multi-query graph is to create a single connection graph
which represents a set of queries. A multi-query graph facilitates the detection of common sub-
expressions among the queries represented by the graph[19]. The selection-view problem can be
divided into two phases. The first phase is the identification of common tasks among a set of
queries and prepares a small set of alternative plans for each query. The second phase generates
a global execution plan that will produce the answers for all the queries when it is executed. In
this paper we introduce a new approach for materialized view selection in conjunction with the
use of a MVPP. The rest of the paper is organized as follows, In Section 2, we encoding view
selection problem into weighted constraint satisfaction problem. In Section 3, we built our
search space to solve materialized view selection. Section 4, describes our experimental results,
and Section 5, concludes the paper.
2. ENCODING VIEW SELECTION PROBLEM INTO WEIGHTED CONSTRAINT
SATISFACTION PROBLEM.
In this part, we propose original weighted constraint satisfaction problems (WCSP) that encode
the view selection problem (VSP). WCSP is a constraint satisfaction problem (CSP) in which
preferences among solutions can be expressed. In the last few years, the CSP framework has
been extended to the soft constraints permits to express preferences among solutions[20]. Soft
constraint frameworks associate costs to tuples and the goal is to find a complete assignment
with minimum aggregated cost. Costs from different constraints are aggregated with a domain
dependent operator. This case is known as the weighted constraint satisfaction problems WCSP.
2.1. Weighted constraint satisfaction problem WCSP
A weighted CSP instance is a quadruplet ( )P X, D, C,W= where:
3. International Journal of Database Management Systems ( IJDMS ) Vol.9, No.4, June 2017
13
• { }X x , x , . . . , x1 2 n= : is a finite set of n variables,
• D : is a function that maps each variable x
i to the set ( )D xi of possible values it can take
(domain of x
i ),
• { }C C , C , . . . , C1 2 m= : is a finite set of constraints.
Each constraint C
jis a relation over an ordered set ( )Var Cj of variables from X , i.e.,
for ( ) { }Var C y , y , . . . , yj 1 2 k= , ( ) ( ) ( )C D y D y . . . D y .1 2 kj
⊆ × × ×
The elements of the set C
jare called satisfying tuples of C
jand k is called the arity of the
constraint C
j.
• [ ]{ }w : C | j 1,2, ,m Rj
+
∈ →LL : is a function that assigns a positive real value to each constraint C
j
of P . w(C )
j : is called the weight of constraint C
j.
2.2. Proposed WCSP for the VSP
A number of parameters such as query cost, query frequencies, maintenance cost, frequency of
updating, frequency query are used to select an appropriate set of views to be materialized. First,
to encode the SVP in terms of a WCSP, we define the following concepts:
• R : is the set of base relations.
• Q {q , q , , q }
1 2 n
= LL : is the set of queries on a given data warehouse;
• f
q : is query frequency associated with the query q;
• M : is the set of intermediate nodes to be materialized whose the elements are obtained by
applying the operations: select or join or project or aggregation;
• S: is the cost of storage;
• Sv : is the cost of storage corresponding to view v.
• f
u
: is update frequency associated with the view v.
Variables: the variables of the weighted constraint satisfaction problem considered in this paper
are given by:
• Mat(v) : is a decision variable that equals to 1 if the view v is materialized, 0 else;
• Cost (v)
q : is the cost of the query q corresponding to the view v;
• Cost (v)
m : is the cost of maintenance when v is materialized.
Domains: the domain of a given variable x , denoted by D(x), is the set of the values that this
variable can take.
As the variable Mat(v) is a binary variable, we have D(Mat(v)) {0,1}.=
The other two variables Cost (v)
q and Cost (v)
m have the two domains:
D(Cost (v)) N and D(Cost (v)) N.q m⊂ ⊂
Constraints: constraints are divided into two groups:
Binary constraints: The constraint between the variables is presented as follows:
Query cost:
f (the cost of the path from q to v) if Mat(v) 1q*
Cost (v)
q f *(the cost of the tree of root q) otherwiseq
=
=
If v is materialized, the cost query of q is sum cost paths from q to views v. If v is not
materialized, the cost query of q is sum cost path from q to base relations.
4. International Journal of Database Management Systems ( IJDMS ) Vol.9, No.4, June 2017
14
The query processing cost of an MVPP DAG can be obtained by multiplying query frequency
by query processing cost.
The total query cost: Cost Cost (v)qQ
q Q
∑=
∈
Maintenance cost:
cost of the path from v to relationsf *(the if Mat(v) 1u
Cost (v)
m 0 otherwise
R ) =
=
If v is not materialized, the cost maintenance is null, else maintenance cost is sum cost from v
materialized to base relations.
Materialized view maintenance cost of an MVPP representation can be obtained by multiplying
update frequency by maintenance cost.
The total maintenance cost: Cost Cost (v)mM
m M
∑=
∈
Unary constraints:
The unary constraint corresponds to the minimization of cost maintenance query. The
variable,Cost (v),
m value is minimized by a unary constraint with
weight β: (value of variable maint withenace 0 1) β ≤ β ≤× .
Minimize the total processing query cost is query processing cost add to maintenance cost.
minimize(Cost Cost Cost ) under the constraint that S SvT Q M v M
∑= + ≤
∈
3. MULTIPLE VIEW PROCESSING PLAN
Multiple views processing plan MVPP is constructed from the workloads queries supported in a
data warehouse system. Each node in the MVPP represents a view that could be selected for
materialization. The cost metrics for selecting materialized views are based on the following
costs: Cost of a query access is the number of rows in the table used to answer this query. Cost
of query processing is the frequency of the query multiplied by cost of query access from
materialized views. Cost of view maintenance is equal to the cost of constructing the view in
response to the changes in the base relation. Total cost is equal to the sum of the cost of query
processing and the cost of view maintenance. In this sense, MVPP is a graphic representation of
a given set of queries. The root (source) nodes are the queries themselves, the leaf (sink) nodes
are the base relations, and all other intermediate nodes are selection, projection, join and
aggregation function views that contribute to the construction of a given query in a bottom-up
manner.
The view selection process is based on two steeps:
• Construct the best MVPP
• Select views to materialize from the best MVPP
In the first step, MVPP global execution plan is obtained by selection of exactly one plan for
each query and identify sub-expressions in which sharing is possible. The selected plans should
minimize the cost of global execution plan[21].The second step, so we have to select
an appropriate set of materialized view from the best MVPP. The algorithm for generating the
MVPP graph is as follows. First, create all alternate execution plans for each query; and then,
individual query processing plans are merged for generate MVPP graphs, ultimately the goal is
to find the best MVPP, such as the cost of global execution plan is minimized.
Example:
In this part, an application example is provided to illustrate the process of generating multiple
views processing plan by considering two queries Q1 and Q2. These queries were selected
from Star Schema Benchmark(SSB) [22].
Q1. select c_nation, s_nation, d_year,
sum(lo_revenue)
from customer, lineorder, supplier, dates
5. International Journal of Database Management Systems ( IJDMS ) Vol.9, No.4, June 2017
15
where lo_custkey = c_customerkey
and lo_suppkey = s_suppkey
and lo_orderdatekey = d_datekey
and c_region = 'ASIA'
and s_region = 'ASIA'
and d_year >= 1992 and d_year <= 1997
group by c_nation, s_nation, d_year ;
Q2. select c_city, s_city, d_year,
sum(lo_revenue)
from customer, lineorder, supplier, dates
where lo_custkey = c_customerkey
and lo_suppkey = s_suppkey
and lo_orderdatekey = d_datekey
and c_nation = 'UNITED STATES'
and s_region = 'ASIA'
and d_year >= 1992 and d_year <= 1997
group by c_city, s_city, d_year ;
In Figure 1, the global MVPP is constructed by combining the local query plan the first query
with second query. Each query has multiple execution plans. One execution plan is choose for
each query to obtain the global MVPP. The best global MVPP is corresponding to best total
query costs. In each graph, the query access frequencies are labelled on the top of each query
node. And for each node except the root (query node) and leaf (base relation node) nodes, there
are two data associated with it. The left stands for the query operator and the right stands for the
cost to generate the nodes from base relations.
Figure 1: MVPP after all Query Q1-Q2 are merged
The view selection problem, based on framework MVPP, is to choose a set of views to
materialize such that the sum cost for query processing and view maintenance is minimal. The
sum cost is calculated from the possible combination of nodes. A framework MVPP of n nodes,
6. International Journal of Database Management Systems ( IJDMS ) Vol.9, No.4, June 2017
16
then we have to try 2n
combinations of nodes to use as a view to materialize. Suppose that there
are some intermediate nodes to be materialized. For each query, the cost of query processing is
query frequency multiplied by the cost of query access from the materialized node(s). In Figure
1, if node tmp4 is chosen to be materialized, the query processing cost for Q1 is 1*(246764+
593389744+ 2192+ 2556+ 8151483630+ 30000+ 30000+ 1347130). The maintenance cost for
the materialized view is the cost for the process of updating a materialized view in response to
the change in the base relations. The view maintenance cost of temp4 is 2*( 2694000000+ 449+
2000+ 6000000). The goal of solving view selection problem is to find an appropriate of nodes
to be materialized in data warehouse.
4. EXPERIMENTAL RESULTS
In this section, we present some computational results. The numerical experiments are carried
out using TPC-H database with scale factors of 1GB and the TPC-H queries as our workload.
The components of the TPC-H database are defined to consist of eight tables including Region,
Nation, Customer, Supplier, Part, Partsupp, Lineitem and Order. TPC-H is the benchmark of the
Transaction Processing Performance Council (TPC) for decision support[22]. In order to
determine a suitable set of views that minimizes the total cost associated with the materialized
views, in conjunction with MVPP framework. In order to evaluate the performance of the
proposed algorithm, we compared the result obtained with another algorithm. The proposed
algorithm is called genetic algorithm (GA) and the second called hill-climbing (HC) algorithm.
The basic idea of the comparison test is to compare the obtained results for materialized views.
Based on the analysis of Figure 3, it can be seen, that the result obtained by the proposed
approach is better than the HC algorithm.
Figure 2 : MVPP graph using TPC-H benchmark data warehouse with materialized views
7. International Journal of Database Management Systems ( IJDMS ) Vol.9, No.4, June 2017
17
Table 1: MVPP, cost of query processing, cost of maintenance and total cost
Cost of query
processing
Cost of
maintenance
Total cost
All virtual views 17585777978944 0 17585777978944
All materialized views 12014999 10995789073264 10995801088263
Optimal set materialized
views obtained by (GA)
1451491204595 7559210580088 9010701784683
Optimal set materialized
views obtained by (HC)
1451508620875 7902161180088 9353669800963
The result from Table 1 shows the total cost of optimal set materialized views, obtained by
using proposed approach, is better than all virtual views, all materialized views and also of the
HC algorithm. The solution of the proposed approach, based on MVPP framework, correspond
to materialize the views tmp28, tmp26, tmp16, tmp18, tmp14, tmp22, tmp13 which have
appeared in Figure 2. The total cost of the optimal set materialized views refers to the sum of
total query processing & maintenance cost of the selected views.
Figure 3: the execution times of the queries (in ms)
In Figure 3, we show how the execution times of the queries workload changes if we consider
views materialized. As can be seen, there is a considerable difference in the time of execution
for the all queries, mainly due to there are many materialized views that were created to help
that the queries. The proposed approach gives better execution time compared with HC
algorithm. Finally, something quite important to mention is that all queries, have benefited from
materialized views either partially or fully.
8. International Journal of Database Management Systems ( IJDMS ) Vol.9, No.4, June 2017
18
Figure 4: comparison of the average execution time of different weight β.
In Figure 4, shows the average execution time of proposed algorithm is better than HC
algorithm on eight queries of different weight β. It is clear from figure 4, that there is a
significant decrease in average execution time while changing the value of variable maintenance
by weight β, with {20%, 40%,60%,80%,100%}β = . As it was expected, the final population
produced by our method contains only a feasible solution of the weighted constraint satisfaction
problem (proposed in the section 2).
Figure 5: Query processing and maintenance cost vs. storage cost
The objective of the storage cost analysis is to minimize the total cost, which is the sum of two
costs: the cost associated with query processing and the cost associated with maintenance. The
cost of query processing is the cost to queries that have benefited or not from the materialization
of the views. The maintenance cost is the cost associated with updating the materialized views.
9. International Journal of Database Management Systems ( IJDMS ) Vol.9, No.4, June 2017
19
The intimate objective of storage cost analysis is to find a compromise between the cost
associated with the query processing and the maintenance cost. Note that when the cost of
storage increases, the total cost decreases. Figure 5, illustrates this concept well.
5. CONCLUSION
The selection of views to materialize is one of the most discussed topics in data warehouse.
View selection problem (VSP) can be solved by the construction of a multiple view processing
plan (MVPP) search space based on multiple query optimizations (MQO) technique. In this
work, we have extended the usage of VSP in conjunction with the MVPP framework. The first
extension is to model the VSP as WCSP. A second extension is to incorporate storage cost into
the objective function to select optimal set of views to materialization. In order to find most
optimized solution, the proposed algorithm is better in term of selection of suitable set of
materialized views as compared to HC algorithm. In future work, we will extend this work by
applying the neural network approach and weighted constraint satisfaction problem to solve the
view selection problem.
REFERENCES
[1] H. GUPTA AND I. S. MUMICK, “SELECTION OF VIEWS TO MATERIALIZE UNDER A
MAINTENANCE COST CONSTRAINT,” PROC. 7TH INT. CONF. DATABASE THEORY, VOL.
13, PP. 453–470, 1999.
[2] H. GUPTA AND I. S. MUMICK, “SELECTION OF VIEWS TO MATERIALIZE IN A DATA
WAREHOUSE,” IEEE TRANS. KNOWL. DATA ENG., VOL. 17, NO. 1, PP. 24–43, 2005.
[3] V.HARINARAYAN, “IMPLEMENTING DATA CUBES EFFICIENTLY,” PROC.ACM SIGMOD
INT. CONF. MANAG. DATA, PP. 205–216, 1996.
[4] D. YANG, M. HUANG, AND M. HUNG, “EFFICIENT UTILIZATION OF MATERIALIZED
VIEWS IN A DATA WAREHOUSE,” PAKDD 2002 ADV. KNOWL. DISCOV. DATA MIN., PP.
393–404, 2002.
[5] G. GOU, J. X. YU, AND H. LU, “A* SEARCH: AN EFFICIENT AND FLEXIBLE APPROACH
TO MATERIALIZED VIEW SELECTION,” IEEE TRANS. SYST. MAN CYBERN. PART C
APPL. REV., VOL. 36, NO. 3, PP. 411–425, 2006.
[6] T. V. V. KUMAR AND S. KUMAR, “MATERIALIZED VIEW SELECTION USING SIMULAT-
ED ANNEALING,” INT. CONF. BIG DATA ANAL., PP. 168–179, 2012.
[7] C. S. PARK, M. H. KIM, AND Y. J. LEE, “FINDING AN EFFICIENT REWRITING OF OLAP
QUERIES USING MATERIALIZED VIEWS IN DATA WAREHOUSES,” DECIS. SUPPORT
SYST., VOL. 32, NO. 4, PP. 379–399, 2002.
[8] J. CHANG AND S. LEE, “EXTENDED CONDITIONS FOR ANSWERING AN AGGREGATE
QUERY USING MATERIALIZED VIEWS,” INF. PROCESS. LETT., VOL. 72, PP. 205–212,
1999.
[9] E. SOUTYRINA AND F. FOTOUHI, “OPTIMAL VIEW SELECTION FOR MULTI DIMENSI-
ONAL DATABASE SYSTEMS,” IDEAS, PP. 309–318, 1997.
[10] I. MAMI, R. COLETTA, AND Z. BELLAHSENE, “MODELING VIEW SELECTION AS A
CONSTRAINT SATISFACTION PROBLEM,” LECT. NOTES COMPUT. SCI. (INCLUDING
SUBSER. LECT. NOTES ARTIF. INTELL. LECT. NOTES BIOINFORMATICS), VOL. 6861
LNCS, NO. PART 2, PP. 396–410, 2011.
10. International Journal of Database Management Systems ( IJDMS ) Vol.9, No.4, June 2017
20
[11] D. THEODORATOS, “DETECTING REDUNDANT MATERIALIZED VIEWS IN DATA
WAREHOUSE EVOLUTION,” INF. SYST., VOL. 26, NO. 5, PP. 363–381, 2001.
[12] T.V.V.KUMAR AND S.KUMAR, “MATERIALISED VIEW SELECTION USING DIFFER-
ENTIAL EVOLUTION,” INT. J. INNOV. COMPUT. APPL., VOL. 6, NO. 2, PP. 102–113, 2014.
[13] J. YANG, K. KARLAPALEM, AND Q. LI, “ALGORITHMS FOR MATERIALIZED VIEW
DESIGN IN DATA WAREHOUSING ENVIRONMENT,” VLDB, NO. 1, PP. 136–145, 1997.
[14] R. DERAKHSHAN AND F. DEHNE, “SIMULATED ANNEALING FOR MATERIALIZED
VIEW SELECTION IN DATA WAREHOUSING ENVIRONMENT.,” 24TH IASTED INT. CONF.
DATABASE APPL., PP. 89–94, 2006.
[15] K. AOUICHE AND J. DARMONT, “DATA MINING-BASED MATERIALIZED VIEW AND
INDEX SELECTION IN DATA WAREHOUSES,” J. INTELL. INF. SYST., VOL. 33, NO. 1, PP.
65–93, 2009.
[16] K. AOUICHE, P.-E. JOUVE, AND J. DARMONT, “CLUSTERING-BASED MATERIALIZED
VIEW SELECTION IN DATA WAREHOUSES,” LECT. NOTES COMPUT. SCI. INCL. SUBSER.
LECT. NOTES ARTIF. INTELL. LECT. NOTES BIOINFORMA., VOL. 4152 LNCS, NO. 1, PP.
81–95, 2007.
[17] A. GOSAIN AND HEENA, “MATERIALIZED CUBE SELECTION USING PARTICLE SWARM
OPTIMIZATION ALGORITHM,” PROCEDIA COMPUT. SCI., VOL. 79, PP. 2–7, 2016.
[18] J. YANG, K. KARLAPALEM, AND Q. LI, “A FRAMEWORK FOR DESIGNING MATERIAL-
IZED VIEWS IN DATA WAREHOUSING ENVIRONMENT,” ICDCS’97, PP. 458–465, 1997.
[19] T. K. SELLIS AND S. GHOSH, “ON THE MULTIPLE-QUERY OPTIMIZATION PROBLEM,”
{IEEE} TRANS. KNOWL. DATA ENG., VOL. 2, NO. 2, PP. 262–266, 1990.
[20] K. HADDOUCH, K. ELMOUTAOUKIL, AND M. ETTAOUIL, “SOLVING THE WEIGHTED
CONSTRAINT SATISFACTION PROBLEMS VIA THE NEURAL NETWORK APPROACH,”
INT. J. INTERACT. MULTIMED. ARTIF. INTELL., VOL. 4, NO. 1, P. 56, 2016.
[21] S. K. MISHRA AND S. PATTNAIK, “EVALUATION OF GENE VALUE AND HEURISTIC
FUNCTION OF ALTERNATE PLANS IN MULTI DATABASE SYSTEM USING GENETIC
ALGORITHM,” INT. J. DATABASE MANAG. SYST., VOL. 3, NO. 3, PP. 78–86, 2011.
[22] P. O. NEIL, B. O. NEIL, AND X. CHEN, “STAR SCHEMA BENCHMARK - REVISION 3,”
TECH. REP., 2009.