This document proposes a new approach for processing multiple queries for a patented medical database that handles temporal domain events more efficiently. The approach uses three techniques: automatic error correction, topic relevant query suggestion, and extended query augmentation. Queries are first used to retrieve coherent clusters from topic-classified medical data. Relevant results from each cluster are then combined to generate the top K answers for the user. The techniques aim to better define the user's information need and improve retrieval speed and memory usage when searching large medical databases. An experimental evaluation found the approach improved retrieval quality and efficiency compared to existing methods.
Query-Based Retrieval of Annotated DocumentIRJET Journal
1. The document discusses a new technique that combines the Collaborative Adaptive Data Sharing (CADS) framework and USHER technique to improve information quality and recommendation of attributes for annotated document retrieval.
2. The technique uses CADS to first build the structure and provide attribute recommendations, then applies USHER's probabilistic model to automatically generate query forms and minimize questions, improving information quality with lower cost.
3. By jointly using CADS for structure design and USHER for applying probabilistic models, the proposed dual approach achieves more effective results from both frameworks for enhanced data search.
WebSite Visit Forecasting Using Data Mining TechniquesChandana Napagoda
Data mining is a technique which is used for identifying relationships between various large amounts of data in many areas including scientific research, business planning, traffic analysis, clinical trial data mining etc. This research will be researching applicability of data mining techniques in web site visit prediction domain. Here we will be concentrating on time series regression techniques which will be used to analyse and forecast time dependent data points. Then how those techniques will be applied to forecast web site visits will be explained.
INTELLIGENT INFORMATION RETRIEVAL WITHIN DIGITAL LIBRARY USING DOMAIN ONTOLOGYcscpconf
A digital library is a type of information retrieval (IR) system. The existing information retrieval
methodologies generally have problems on keyword-searching. We proposed a model to solve
the problem by using concept-based approach (ontology) and metadata case base. This model
consists of identifying domain concepts in user’s query and applying expansion to them. The
system aims at contributing to an improved relevance of results retrieved from digital libraries
by proposing a conceptual query expansion for intelligent concept-based retrieval. We need to
import the concept of ontology, making use of its advantage of abundant semantics and
standard concept. Domain specific ontology can be used to improve information retrieval from
traditional level based on keyword to the lay based on knowledge (or concept) and change the
process of retrieval from traditional keyword matching to semantics matching. One approach is
query expansion techniques using domain ontology and the other would be introducing a case
based similarity measure for metadata information retrieval using Case Based Reasoning
(CBR) approach. Results show improvements over classic method, query expansion using
general purpose ontology and a number of other approaches.
Enhancement techniques for data warehouse staging areaIJDKP
This document discusses techniques for enhancing the performance of data warehouse staging areas. It proposes two algorithms: 1) A semantics-based extraction algorithm that reduces extraction time by pruning useless data using semantic information. 2) A semantics-based transformation algorithm that similarly aims to reduce transformation time. It also explores three scheduling techniques (FIFO, minimum cost, round robin) for loading data into the data warehouse and experimentally evaluates their performance. The goal is to enhance each stage of the ETL process to maximize overall performance.
This document summarizes a project to apply machine learning models to predict outcomes of cases in the US Court of Appeals based on historical data. It describes the data source and characteristics, data preprocessing steps including handling missing values and complex variables, dimensionality reduction techniques, models selected like random forest, neural networks and XGBoost, and the results of model tuning and testing. The neural network achieved an accuracy of 85% on test data after oversampling the training data, while XGBoost achieved 91% accuracy after tuning. The most significant variables identified for prediction included the circuit court, verdict of the previous court, nature of the appellant, and directionality of the third judge.
The document describes developing machine learning models to predict outcomes of cases in the US Court of Appeals using historical data. Neural networks and XGBoost classifiers were developed and evaluated on a dataset of over 2,000 court cases. The models achieved encouraging accuracy levels between 36-98%, though limitations included the small dataset size. Significant factors identified included the appeals court, nature of prior rulings, applicant type, and judges' ideological leanings. Future work could focus on decision trees to characterize outcomes and simplify the models.
Introduction to feature subset selection methodIJSRD
Data Mining is a computational progression to ascertain patterns in hefty data sets. It has various important techniques and one of them is Classification which is receiving great attention recently in the database community. Classification technique can solve several problems in different fields like medicine, industry, business, science. PSO is based on social behaviour for optimization problem. Feature Selection (FS) is a solution that involves finding a subset of prominent features to improve predictive accuracy and to remove the redundant features. Rough Set Theory (RST) is a mathematical tool which deals with the uncertainty and vagueness of the decision systems.
Knowledge discovery is the process of adding knowledge from a large amount of data. The quality of knowledge generated from the process of knowledge discovery greatly affects the results of the decisions obtained. Existing data must be qualified and tested to ensure knowledge discovery processes can produce knowledge or information that is useful and feasible. It deals with strategic decision making for an organization. Combining multiple operational databases and external data create data warehouse. This treatment is very vulnerable to incomplete, inconsistent, and noisy data. Data mining provides a mechanism to clear this deficiency before finally stored in the data warehouse. This research tries to give technique to improve the quality of information in the data warehouse.
Query-Based Retrieval of Annotated DocumentIRJET Journal
1. The document discusses a new technique that combines the Collaborative Adaptive Data Sharing (CADS) framework and USHER technique to improve information quality and recommendation of attributes for annotated document retrieval.
2. The technique uses CADS to first build the structure and provide attribute recommendations, then applies USHER's probabilistic model to automatically generate query forms and minimize questions, improving information quality with lower cost.
3. By jointly using CADS for structure design and USHER for applying probabilistic models, the proposed dual approach achieves more effective results from both frameworks for enhanced data search.
WebSite Visit Forecasting Using Data Mining TechniquesChandana Napagoda
Data mining is a technique which is used for identifying relationships between various large amounts of data in many areas including scientific research, business planning, traffic analysis, clinical trial data mining etc. This research will be researching applicability of data mining techniques in web site visit prediction domain. Here we will be concentrating on time series regression techniques which will be used to analyse and forecast time dependent data points. Then how those techniques will be applied to forecast web site visits will be explained.
INTELLIGENT INFORMATION RETRIEVAL WITHIN DIGITAL LIBRARY USING DOMAIN ONTOLOGYcscpconf
A digital library is a type of information retrieval (IR) system. The existing information retrieval
methodologies generally have problems on keyword-searching. We proposed a model to solve
the problem by using concept-based approach (ontology) and metadata case base. This model
consists of identifying domain concepts in user’s query and applying expansion to them. The
system aims at contributing to an improved relevance of results retrieved from digital libraries
by proposing a conceptual query expansion for intelligent concept-based retrieval. We need to
import the concept of ontology, making use of its advantage of abundant semantics and
standard concept. Domain specific ontology can be used to improve information retrieval from
traditional level based on keyword to the lay based on knowledge (or concept) and change the
process of retrieval from traditional keyword matching to semantics matching. One approach is
query expansion techniques using domain ontology and the other would be introducing a case
based similarity measure for metadata information retrieval using Case Based Reasoning
(CBR) approach. Results show improvements over classic method, query expansion using
general purpose ontology and a number of other approaches.
Enhancement techniques for data warehouse staging areaIJDKP
This document discusses techniques for enhancing the performance of data warehouse staging areas. It proposes two algorithms: 1) A semantics-based extraction algorithm that reduces extraction time by pruning useless data using semantic information. 2) A semantics-based transformation algorithm that similarly aims to reduce transformation time. It also explores three scheduling techniques (FIFO, minimum cost, round robin) for loading data into the data warehouse and experimentally evaluates their performance. The goal is to enhance each stage of the ETL process to maximize overall performance.
This document summarizes a project to apply machine learning models to predict outcomes of cases in the US Court of Appeals based on historical data. It describes the data source and characteristics, data preprocessing steps including handling missing values and complex variables, dimensionality reduction techniques, models selected like random forest, neural networks and XGBoost, and the results of model tuning and testing. The neural network achieved an accuracy of 85% on test data after oversampling the training data, while XGBoost achieved 91% accuracy after tuning. The most significant variables identified for prediction included the circuit court, verdict of the previous court, nature of the appellant, and directionality of the third judge.
The document describes developing machine learning models to predict outcomes of cases in the US Court of Appeals using historical data. Neural networks and XGBoost classifiers were developed and evaluated on a dataset of over 2,000 court cases. The models achieved encouraging accuracy levels between 36-98%, though limitations included the small dataset size. Significant factors identified included the appeals court, nature of prior rulings, applicant type, and judges' ideological leanings. Future work could focus on decision trees to characterize outcomes and simplify the models.
Introduction to feature subset selection methodIJSRD
Data Mining is a computational progression to ascertain patterns in hefty data sets. It has various important techniques and one of them is Classification which is receiving great attention recently in the database community. Classification technique can solve several problems in different fields like medicine, industry, business, science. PSO is based on social behaviour for optimization problem. Feature Selection (FS) is a solution that involves finding a subset of prominent features to improve predictive accuracy and to remove the redundant features. Rough Set Theory (RST) is a mathematical tool which deals with the uncertainty and vagueness of the decision systems.
Knowledge discovery is the process of adding knowledge from a large amount of data. The quality of knowledge generated from the process of knowledge discovery greatly affects the results of the decisions obtained. Existing data must be qualified and tested to ensure knowledge discovery processes can produce knowledge or information that is useful and feasible. It deals with strategic decision making for an organization. Combining multiple operational databases and external data create data warehouse. This treatment is very vulnerable to incomplete, inconsistent, and noisy data. Data mining provides a mechanism to clear this deficiency before finally stored in the data warehouse. This research tries to give technique to improve the quality of information in the data warehouse.
A time efficient and accurate retrieval of range aggregate queries using fuzz...IJECEIAES
This document presents a new approach called Fuzzy Clustering Means (FCM) to efficiently retrieve range aggregate queries from big data. Existing approaches have issues with inefficient retrieval times and clustering inaccuracies for large datasets. The FCM approach first partitions big data into independent partitions using balanced partitioning. It then creates an estimation sketch for each partition. When a range query is received, it estimates the result from each partition and summarizes the local estimates to provide the final output. Analysis on a dataset of 200,000 records shows the FCM approach has higher accuracy, lower error rates, and faster execution times for queries compared to existing approaches. Future work will investigate extending this solution to handle more complex query formats and using FCM to boost general
This document reviews the use of data mining and neural network techniques for stock market prediction. It discusses how data mining can extract hidden patterns from large datasets and neural networks can handle nonlinear and uncertain financial data. Specifically, it examines how a combination of data mining and neural networks may improve the reliability of stock predictions by leveraging their complementary strengths. The document also provides an overview of common data mining and neural network methods used for this purpose, such as statistical data mining, neural network-based data processing, clustering, and fuzzy logic. It reviews several previous studies that found neural networks and other nonlinear techniques often outperform traditional statistical models at predicting stock prices and indices.
This document presents a case study on applying a data analytics approach to conducting a systematic literature review on master data management. It outlines the steps taken, including defining review questions, searching multiple databases and sources, combining and preprocessing the data, and performing descriptive and text analyses. The analyses addressed questions about trends in publications over time, primary databases, publication types, and frequent keywords. This provided insights into the progress and topics within the master data management research domain. The presented structured approach aims to improve the replicability of systematic literature reviews.
This document provides an overview of an information retrieval system. It defines an information retrieval system as a system capable of storing, retrieving, and maintaining information such as text, images, audio, and video. The objectives of an information retrieval system are to minimize the overhead for a user to locate needed information. The document discusses functions like search, browse, indexing, cataloging, and various capabilities to facilitate querying and retrieving relevant information from the system.
A Survey on Automatically Mining Facets for Queries from their Search ResultsIRJET Journal
This document summarizes research on automatically mining query facets from search results. Query facets provide useful summaries of a query by grouping related terms and phrases. The document reviews existing methods for query recommendation and facet extraction. It also proposes an unsupervised technique to mine query facets from top search results without additional domain knowledge. The technique aims to help users better understand queries and explore information through faceted search.
One of the most important problems in modern finance is finding efficient ways to summarize and visualize
the stock market data to give individuals or institutions useful information about the market behavior for
investment decisions Therefore, Investment can be considered as one of the fundamental pillars of national
economy. So, at the present time many investors look to find criterion to compare stocks together and
selecting the best and also investors choose strategies that maximize the earning value of the investment
process. Therefore the enormous amount of valuable data generated by the stock market has attracted
researchers to explore this problem domain using different methodologies. Therefore research in data
mining has gained a high attraction due to the importance of its applications and the increasing generation
information. So, Data mining tools such as association rule, rule induction method and Apriori algorithm
techniques are used to find association between different scripts of stock market, and also much of the
research and development has taken place regarding the reasons for fluctuating Indian stock exchange.
But, now days there are two important factors such as gold prices and US Dollar Prices are more
dominating on Indian Stock Market and to find out the correlation between gold prices, dollar prices and
BSE index statistical correlation is used and this helps the activities of stock operators, brokers, investors
and jobbers. They are based on the forecasting the fluctuation of index share prices, gold prices, dollar
prices and transactions of customers. Hence researcher has considered these problems as a topic for
research.
A Semantic Retrieval System for Extracting Relationships from Biological Corpusijcsit
The World Wide Web holds a large size of different information. Sometimes while searching the World Wide Web, users always do not gain the type of information they expect. In the subject of information extraction, extracting semantic relationships between terms from documents become a challenge. This
paper proposes a system helps in retrieving documents based on the query expansion and tackles the extracting of semantic relationships from biological documents. This system retrieved documents that are relevant to the input terms then it extracts the existence of a relationship. In this system, we use Boolean
model and the pattern recognition which helps in determining the relevant documents and determining the place of the relationship in the biological document. The system constructs a term-relation table that accelerates the relation extracting part. The proposed method offers another usage of the system so the
researchers can use it to figure out the relationship between two biological terms through the available information in the biological documents. Also for the retrieved documents, the system measures the percentage of the precision and recall.
Survey on scalable continual top k keyword search in relational databaseseSAT Journals
Abstract Keyword search in relational database is a technique that has higher relevance in the present world. Extracting data from a large number of sets of database is very important .Because it reduces the usage of man power and time consumption. Data extraction from a large database using the relevant keyword based on the information needed is a very interactive and user friendly. Without knowing any database schemas or query languages like sql the user can get information. By using keyword in relational database data extraction will be simpler. The user doesn’t want to know the query language for search. But the database content is always changing for real time application for example database which store the data of publication data. When new publications arrive it should be added to database so the database content changes according to time. Because the database is updated frequently the result should change. In order to handle the database updation takes the top-k result from the currently updated data for each search. Top-k keyword search means take greatest k results based on the relevance of document. Keyword search in relational database means to find structural information from tuples from the database. Two types of keyword search are schema-based method and graph based approach. Using top-k keyword search instead of executing all query results taking highest k queries. By handling database updation try to find the new results and remove expired one
A statistical data fusion technique in virtual data integration environmentIJDKP
Data fusion in the virtual data integration environment starts after detecting and clustering duplicated
records from the different integrated data sources. It refers to the process of selecting or fusing attribute
values from the clustered duplicates into a single record representing the real world object. In this paper, a
statistical technique for data fusion is introduced based on some probabilistic scores from both data
sources and clustered duplicates
Novel Ensemble Tree for Fast Prediction on Data StreamsIJERA Editor
Data Streams are sequential set of data records. When data appears at highest speed and constantly, so predicting
the class accordingly to the time is very essential. Currently Ensemble modeling techniques are growing
speedily in Classification of Data Stream. Ensemble learning will be accepted since its benefit to manage huge
amount of data stream, means it will manage the data in a large size and also it will be able to manage concept
drifting. Prior learning, mostly focused on accuracy of ensemble model, prediction efficiency has not considered
much since existing ensemble model predicts in linear time, which is enough for small applications and
accessible models workings on integrating some of the classifier. Although real time application has huge
amount of data stream so we required base classifier to recognize dissimilar model and make a high grade
ensemble model. To fix these challenges we developed Ensemble tree which is height balanced tree indexing
structure of base classifier for quick prediction on data streams by ensemble modeling techniques. Ensemble
Tree manages ensembles as geodatabases and it utilizes R tree similar to structure to achieve sub linear time
complexity
Associative Regressive Decision Rule Mining for Predicting Customer Satisfact...csandit
Opinion mining also known as sentiment analysis, involves customer satisfactory patterns,
sentiments and attitudes toward entities, products, services and their attributes. With the rapid
development in the field of Internet, potential customer’s provides a satisfactory level of
product/service reviews. The high volume of customer reviews were developed for
product/review through taxonomy-aware processing but, it was difficult to identify the best
reviews. In this paper, an Associative Regression Decision Rule Mining (ARDRM) technique is
developed to predict the pattern for service provider and to improve customer satisfaction based
on the review comments. Associative Regression based Decision Rule Mining performs twosteps
for improving the customer satisfactory level. Initially, the Machine Learning Bayes
Sentiment Classifier (MLBSC) is used to classify the class labels for each service reviews. After
that, Regressive factor of the opinion words and Class labels were checked for Association
between the words by using various probabilistic rules. Based on the probabilistic rules, the
opinion and sentiments effect on customer reviews, are analyzed to arrive at specific set of
service preferred by the customers with their review comments. The Associative Regressive
Decision Rule helps the service provider to take decision on improving the customer satisfactory
level. The experimental results reveal that the Associative Regression Decision Rule Mining
(ARDRM) technique improved the performance in terms of true positive rate, Associative
Regression factor, Regressive Decision Rule Generation time and Review Detection Accuracy of
similar pattern.
This document summarizes various techniques for scalable continual top-k keyword search in relational databases. There are two main approaches: schema-based and graph-based. Schema-based methods generate candidate networks from the database schema and evaluate them. Graph-based methods represent the database as a graph and use techniques like bidirectional expansion. Top-k keyword search finds the highest scoring k results instead of all results. Methods like the Global Pipeline algorithm and Skyline-Sweeping algorithm efficiently process top-k queries over multiple candidate networks. Techniques for updating results with database changes include maintaining an initial top-k and recalculating scores. Lattice-based methods share computational costs for keyword search in data streams.
Document Classification Using Expectation Maximization with Semi Supervised L...ijsc
As the amount of online document increases, the demand for document classification to aid the analysis and management of document is increasing. Text is cheap, but information, in the form of knowing what classes a document belongs to, is expensive. The main purpose of this paper is to explain the expectation maximization technique of data mining to classify the document and to learn how to improve the accuracy while using semi-supervised approach. Expectation maximization algorithm is applied with both supervised and semi-supervised approach. It is found that semi-supervised approach is more accurate and effective. The main advantage of semi supervised approach is “DYNAMICALLY GENERATION OF NEW CLASS”. The algorithm first trains a classifier using the labeled document and probabilistically classifies the
unlabeled documents. The car dataset for the evaluation purpose is collected from UCI repository dataset in which some changes have been done from our side.
The document discusses multidimensional databases. It defines multidimensional databases as systems designed to efficiently store and retrieve large volumes of related data that can be viewed from different perspectives or dimensions. It provides an example using automobile sales data that can be analyzed based on dimensions like model, color, dealership, and time. Multidimensional databases allow for interactive analysis of data from multiple angles, unlike relational databases that are slower for such analyses.
Internet becomes the most popular surfing environment which increases the
service oriented data size. As the data size grows, finding and retrieving the most
similar data from the large volume of data would become more difficult task. This
problem is focused in the various research methods, which attempts to cluster the
large volume of data. In the existing research method Clustering-based Collaborative
Filtering approach (ClubCF) is introduced whose main goal is to cluster the similar
kind of data together, so that retrieval time cost can be reduced considerably.
However, existing research methods cannot find the similar reviews accurately which
needs to be focused more for efficient and accurate recommendation system. This is
ensured in the proposed research method by introducing the novel research technique
namely Modified Collaborative Filtering and Clustering with Regression (MoCFCR).
In this research method, initially k means algorithm is used to cluster the similar
movie reviewer together, so that recommendation process can be done in the easier
way. In order to handle the large volume of data this research work adapts the map
reduce framework which will divide the entire data into subsets which will assigned
on separate nodes with individual key values. After clustering, the clustered outcome
is merged together using inverted index procedure in which similarity between movies
would be calculated. Here collaborative filtering is applied to remove the movies that
are not relevant to input. Finally recommendations of movies are made in the accurate
way by using the logistic regression method. The overall evaluation of the proposed
research method is done in Hadoop from which it can be proved that the proposed
research technique can lead to provide better outcome than the existing research
techniques
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
A time efficient and accurate retrieval of range aggregate queries using fuzz...IJECEIAES
This document presents a new approach called Fuzzy Clustering Means (FCM) to efficiently retrieve range aggregate queries from big data. Existing approaches have issues with inefficient retrieval times and clustering inaccuracies for large datasets. The FCM approach first partitions big data into independent partitions using balanced partitioning. It then creates an estimation sketch for each partition. When a range query is received, it estimates the result from each partition and summarizes the local estimates to provide the final output. Analysis on a dataset of 200,000 records shows the FCM approach has higher accuracy, lower error rates, and faster execution times for queries compared to existing approaches. Future work will investigate extending this solution to handle more complex query formats and using FCM to boost general
This document reviews the use of data mining and neural network techniques for stock market prediction. It discusses how data mining can extract hidden patterns from large datasets and neural networks can handle nonlinear and uncertain financial data. Specifically, it examines how a combination of data mining and neural networks may improve the reliability of stock predictions by leveraging their complementary strengths. The document also provides an overview of common data mining and neural network methods used for this purpose, such as statistical data mining, neural network-based data processing, clustering, and fuzzy logic. It reviews several previous studies that found neural networks and other nonlinear techniques often outperform traditional statistical models at predicting stock prices and indices.
This document presents a case study on applying a data analytics approach to conducting a systematic literature review on master data management. It outlines the steps taken, including defining review questions, searching multiple databases and sources, combining and preprocessing the data, and performing descriptive and text analyses. The analyses addressed questions about trends in publications over time, primary databases, publication types, and frequent keywords. This provided insights into the progress and topics within the master data management research domain. The presented structured approach aims to improve the replicability of systematic literature reviews.
This document provides an overview of an information retrieval system. It defines an information retrieval system as a system capable of storing, retrieving, and maintaining information such as text, images, audio, and video. The objectives of an information retrieval system are to minimize the overhead for a user to locate needed information. The document discusses functions like search, browse, indexing, cataloging, and various capabilities to facilitate querying and retrieving relevant information from the system.
A Survey on Automatically Mining Facets for Queries from their Search ResultsIRJET Journal
This document summarizes research on automatically mining query facets from search results. Query facets provide useful summaries of a query by grouping related terms and phrases. The document reviews existing methods for query recommendation and facet extraction. It also proposes an unsupervised technique to mine query facets from top search results without additional domain knowledge. The technique aims to help users better understand queries and explore information through faceted search.
One of the most important problems in modern finance is finding efficient ways to summarize and visualize
the stock market data to give individuals or institutions useful information about the market behavior for
investment decisions Therefore, Investment can be considered as one of the fundamental pillars of national
economy. So, at the present time many investors look to find criterion to compare stocks together and
selecting the best and also investors choose strategies that maximize the earning value of the investment
process. Therefore the enormous amount of valuable data generated by the stock market has attracted
researchers to explore this problem domain using different methodologies. Therefore research in data
mining has gained a high attraction due to the importance of its applications and the increasing generation
information. So, Data mining tools such as association rule, rule induction method and Apriori algorithm
techniques are used to find association between different scripts of stock market, and also much of the
research and development has taken place regarding the reasons for fluctuating Indian stock exchange.
But, now days there are two important factors such as gold prices and US Dollar Prices are more
dominating on Indian Stock Market and to find out the correlation between gold prices, dollar prices and
BSE index statistical correlation is used and this helps the activities of stock operators, brokers, investors
and jobbers. They are based on the forecasting the fluctuation of index share prices, gold prices, dollar
prices and transactions of customers. Hence researcher has considered these problems as a topic for
research.
A Semantic Retrieval System for Extracting Relationships from Biological Corpusijcsit
The World Wide Web holds a large size of different information. Sometimes while searching the World Wide Web, users always do not gain the type of information they expect. In the subject of information extraction, extracting semantic relationships between terms from documents become a challenge. This
paper proposes a system helps in retrieving documents based on the query expansion and tackles the extracting of semantic relationships from biological documents. This system retrieved documents that are relevant to the input terms then it extracts the existence of a relationship. In this system, we use Boolean
model and the pattern recognition which helps in determining the relevant documents and determining the place of the relationship in the biological document. The system constructs a term-relation table that accelerates the relation extracting part. The proposed method offers another usage of the system so the
researchers can use it to figure out the relationship between two biological terms through the available information in the biological documents. Also for the retrieved documents, the system measures the percentage of the precision and recall.
Survey on scalable continual top k keyword search in relational databaseseSAT Journals
Abstract Keyword search in relational database is a technique that has higher relevance in the present world. Extracting data from a large number of sets of database is very important .Because it reduces the usage of man power and time consumption. Data extraction from a large database using the relevant keyword based on the information needed is a very interactive and user friendly. Without knowing any database schemas or query languages like sql the user can get information. By using keyword in relational database data extraction will be simpler. The user doesn’t want to know the query language for search. But the database content is always changing for real time application for example database which store the data of publication data. When new publications arrive it should be added to database so the database content changes according to time. Because the database is updated frequently the result should change. In order to handle the database updation takes the top-k result from the currently updated data for each search. Top-k keyword search means take greatest k results based on the relevance of document. Keyword search in relational database means to find structural information from tuples from the database. Two types of keyword search are schema-based method and graph based approach. Using top-k keyword search instead of executing all query results taking highest k queries. By handling database updation try to find the new results and remove expired one
A statistical data fusion technique in virtual data integration environmentIJDKP
Data fusion in the virtual data integration environment starts after detecting and clustering duplicated
records from the different integrated data sources. It refers to the process of selecting or fusing attribute
values from the clustered duplicates into a single record representing the real world object. In this paper, a
statistical technique for data fusion is introduced based on some probabilistic scores from both data
sources and clustered duplicates
Novel Ensemble Tree for Fast Prediction on Data StreamsIJERA Editor
Data Streams are sequential set of data records. When data appears at highest speed and constantly, so predicting
the class accordingly to the time is very essential. Currently Ensemble modeling techniques are growing
speedily in Classification of Data Stream. Ensemble learning will be accepted since its benefit to manage huge
amount of data stream, means it will manage the data in a large size and also it will be able to manage concept
drifting. Prior learning, mostly focused on accuracy of ensemble model, prediction efficiency has not considered
much since existing ensemble model predicts in linear time, which is enough for small applications and
accessible models workings on integrating some of the classifier. Although real time application has huge
amount of data stream so we required base classifier to recognize dissimilar model and make a high grade
ensemble model. To fix these challenges we developed Ensemble tree which is height balanced tree indexing
structure of base classifier for quick prediction on data streams by ensemble modeling techniques. Ensemble
Tree manages ensembles as geodatabases and it utilizes R tree similar to structure to achieve sub linear time
complexity
Associative Regressive Decision Rule Mining for Predicting Customer Satisfact...csandit
Opinion mining also known as sentiment analysis, involves customer satisfactory patterns,
sentiments and attitudes toward entities, products, services and their attributes. With the rapid
development in the field of Internet, potential customer’s provides a satisfactory level of
product/service reviews. The high volume of customer reviews were developed for
product/review through taxonomy-aware processing but, it was difficult to identify the best
reviews. In this paper, an Associative Regression Decision Rule Mining (ARDRM) technique is
developed to predict the pattern for service provider and to improve customer satisfaction based
on the review comments. Associative Regression based Decision Rule Mining performs twosteps
for improving the customer satisfactory level. Initially, the Machine Learning Bayes
Sentiment Classifier (MLBSC) is used to classify the class labels for each service reviews. After
that, Regressive factor of the opinion words and Class labels were checked for Association
between the words by using various probabilistic rules. Based on the probabilistic rules, the
opinion and sentiments effect on customer reviews, are analyzed to arrive at specific set of
service preferred by the customers with their review comments. The Associative Regressive
Decision Rule helps the service provider to take decision on improving the customer satisfactory
level. The experimental results reveal that the Associative Regression Decision Rule Mining
(ARDRM) technique improved the performance in terms of true positive rate, Associative
Regression factor, Regressive Decision Rule Generation time and Review Detection Accuracy of
similar pattern.
This document summarizes various techniques for scalable continual top-k keyword search in relational databases. There are two main approaches: schema-based and graph-based. Schema-based methods generate candidate networks from the database schema and evaluate them. Graph-based methods represent the database as a graph and use techniques like bidirectional expansion. Top-k keyword search finds the highest scoring k results instead of all results. Methods like the Global Pipeline algorithm and Skyline-Sweeping algorithm efficiently process top-k queries over multiple candidate networks. Techniques for updating results with database changes include maintaining an initial top-k and recalculating scores. Lattice-based methods share computational costs for keyword search in data streams.
Document Classification Using Expectation Maximization with Semi Supervised L...ijsc
As the amount of online document increases, the demand for document classification to aid the analysis and management of document is increasing. Text is cheap, but information, in the form of knowing what classes a document belongs to, is expensive. The main purpose of this paper is to explain the expectation maximization technique of data mining to classify the document and to learn how to improve the accuracy while using semi-supervised approach. Expectation maximization algorithm is applied with both supervised and semi-supervised approach. It is found that semi-supervised approach is more accurate and effective. The main advantage of semi supervised approach is “DYNAMICALLY GENERATION OF NEW CLASS”. The algorithm first trains a classifier using the labeled document and probabilistically classifies the
unlabeled documents. The car dataset for the evaluation purpose is collected from UCI repository dataset in which some changes have been done from our side.
The document discusses multidimensional databases. It defines multidimensional databases as systems designed to efficiently store and retrieve large volumes of related data that can be viewed from different perspectives or dimensions. It provides an example using automobile sales data that can be analyzed based on dimensions like model, color, dealership, and time. Multidimensional databases allow for interactive analysis of data from multiple angles, unlike relational databases that are slower for such analyses.
Internet becomes the most popular surfing environment which increases the
service oriented data size. As the data size grows, finding and retrieving the most
similar data from the large volume of data would become more difficult task. This
problem is focused in the various research methods, which attempts to cluster the
large volume of data. In the existing research method Clustering-based Collaborative
Filtering approach (ClubCF) is introduced whose main goal is to cluster the similar
kind of data together, so that retrieval time cost can be reduced considerably.
However, existing research methods cannot find the similar reviews accurately which
needs to be focused more for efficient and accurate recommendation system. This is
ensured in the proposed research method by introducing the novel research technique
namely Modified Collaborative Filtering and Clustering with Regression (MoCFCR).
In this research method, initially k means algorithm is used to cluster the similar
movie reviewer together, so that recommendation process can be done in the easier
way. In order to handle the large volume of data this research work adapts the map
reduce framework which will divide the entire data into subsets which will assigned
on separate nodes with individual key values. After clustering, the clustered outcome
is merged together using inverted index procedure in which similarity between movies
would be calculated. Here collaborative filtering is applied to remove the movies that
are not relevant to input. Finally recommendations of movies are made in the accurate
way by using the logistic regression method. The overall evaluation of the proposed
research method is done in Hadoop from which it can be proved that the proposed
research technique can lead to provide better outcome than the existing research
techniques
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
International Journal of Engineering and Science Invention (IJESI)inventionjournals
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online
Performance Evaluation of Query Processing Techniques in Information Retrievalidescitation
The first element of the search process is the query.
The user query being on an average restricted to two or three
keywords makes the query ambiguous to the search engine.
Given the user query, the goal of an Information Retrieval
[IR] system is to retrieve information which might be useful
or relevant to the information need of the user. Hence, the
query processing plays an important role in IR system.
The query processing can be divided into four categories
i.e. query expansion, query optimization, query classification and
query parsing. In this paper an attempt is made to evaluate the
performance of query processing algorithms in each of the
category. The evaluation was based on dataset as specified by
Forum for Information Retrieval [FIRE15]. The criteria used
for evaluation are precision and relative recall. The analysis is
based on the importance of each step in query processing. The
experimental results show that the significance of each step
in query processing and also the relevance of web semantics
and spelling correction in the user query.
QUERY OPTIMIZATION IN OODBMS: IDENTIFYING SUBQUERY FOR COMPLEX QUERY MANAGEMENTcsandit
This document discusses query optimization in object-oriented database management systems (OODBMS) using query decomposition and caching. It proposes an approach that decomposes complex queries into smaller subqueries for faster retrieval of cached results. The approach aims to reuse parts of cached results to answer wider queries by combining multiple cached queries. Experiments showed this approach improved query optimization performance especially when data manipulation rates were low compared to data retrieval rates. Key aspects included decomposing queries, caching subquery results, and reusing cached results to answer other queries.
Open domain question answering system using semantic role labelingeSAT Publishing House
1. The document describes a proposed open domain question answering system that uses semantic role labeling to extract answers from documents retrieved from the web.
2. The system consists of three modules: question processing, document retrieval, and answer extraction. Semantic role labeling is used in the answer extraction module to identify answers based on the question type.
3. An evaluation of the proposed system showed it achieved higher accuracy compared to a baseline system using only pattern matching for answer extraction.
Query optimization in oodbms identifying subquery for query managementijdms
This paper is based on relatively newer approach for query optimization in object databases, which uses
query decomposition and cached query results to improve execution a query. Issues that are focused here is
fast retrieval and high reuse of cached queries, Decompose Query into Sub query, Decomposition of
complex queries into smaller for fast retrieval of result.
Here we try to address another open area of query caching like handling wider queries. By using some
parts of cached results helpful for answering other queries (wider Queries) and combining many cached
queries while producing the result.
Multiple experiments were performed to prove the productivity of this newer way of optimizing a query.
The limitation of this technique is that it’s useful especially in scenarios where data manipulation rate is
very low as compared to data retrieval rate.
1) The document discusses a review of semantic approaches for nearest neighbor search. It describes using an ontology to add a semantic layer to an information retrieval system to relate concepts using query words.
2) A technique called spatial inverted index is proposed to locate multidimensional information and handle nearest neighbor queries by finding the hospitals closest to a given address.
3) Several semantic approaches are described including using clustering measures, specificity measures, link analysis, and relation-based page ranking to improve search and interpret hidden concepts behind keywords.
A signature based indexing method for efficient content-based retrieval of re...Mumbai Academisc
This document summarizes a proposed system for efficiently retrieving subsets of large collections of temporal patterns using signature-based indexing. The system aims to address the challenge of interpreting large numbers of patterns generated by data mining algorithms. It proposes converting temporal patterns to equivalent sets and then using a signature file (bit representation) and bit-sliced signature files to enable fast content-based querying. The system would allow subpattern, superpattern, equality, and k-nearest neighbor queries on temporal patterns. It is intended to optimize storage and retrieval of large numbers of temporal patterns discovered during data mining.
Query Evaluation Techniques for Large Databases.pdfRayWill4
This document surveys techniques for efficiently executing queries over large databases. It describes algorithms for sorting, hashing, aggregation, joins and other operations. It also discusses parallel query execution, complex query plans, and techniques for non-traditional data models. The goal is to provide a foundation for designing query execution facilities in new database management systems.
This document summarizes an algorithm for efficiently refining why-not questions on top-k queries. It begins by executing a top-k query and identifying a missing object m. It then samples potential replacements for m from a restricted sample space and computes m's new ranking if modified to each sample value. The refined query with the smallest penalty, which returns m in the results, is returned as the answer. The algorithm improves efficiency by skipping unnecessary ranking computations and is tested on a basketball player database, demonstrating effectiveness. It answers why-not questions faster than without optimization techniques.
Efficient Refining Of Why-Not Questions on Top-K Queriesiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Predictive job scheduling in a connection limited system using parallel genet...Mumbai Academisc
The document discusses predictive job scheduling in a connection limited system using parallel genetic algorithms. It introduces the problem of job scheduling in parallel computing systems and describes existing non-predictive greedy algorithms. The proposed approach uses genetic algorithms to develop a predictive model for job scheduling that learns from previous experiences to improve scheduling efficiency over time. The goal is to schedule jobs in a way that optimizes system metrics like utilization and throughput while minimizing user metrics like turnaround time.
The activity of finding significant data identified with a particular subject is troublesome in web because of the immensity of web information. This situation makes website streamlining strategies into an irreplaceable technique according to analysts, academicians, and industrialists. Inquiry history investigation is the definite examination of web information from various clients with the end goal of comprehension and upgrading web taking care of. Inquiry log or client seek history incorporates clients' beforehand submitted inquiries and their comparing clicked reports or locales' URLs. Accordingly question log investigation is considered as the most utilized technique for improving the clients' pursuit encounter. The proposed strategy investigates and groups client scan histories with the end goal of website streamlining. In this approach, the issue of getting sorted out clients' verifiable questions into bunches in a dynamic and robotized design is examined. The consequently arranged inquiry gatherings will help in various website streamlining systems like question proposal, item re-positioning, question adjustments and so on. The proposed strategy considers a question aggregate as an accumulation of inquiries together with the comparing set of clicked URLs that are identified with each other around a general data require. This technique proposes another strategy for joining word likeness measures alongside report similitude measures to frame a consolidated comparability measure. In the proposed strategy other question importance measures, for example, inquiry reformulation and clicked URL idea are likewise considered. Assessment comes about show how the proposed technique outflanks existing strategies.
Classification-based Retrieval Methods to Enhance Information Discovery on th...IJMIT JOURNAL
The widespread adoption of the World-Wide Web (the Web) has created challenges both for society as a whole and for the technology used to build and maintain the Web. The ongoing struggle of information retrieval systems is to wade through this vast pile of data and satisfy users by presenting them with information that most adequately it’s their needs. On a societal level, the Web is expanding faster than we can comprehend its implications or develop rules for its use. The ubiquitous use of the Web has raised important social concerns in the areas of privacy, censorship, and access to information. On a technical level, the novelty of the Web and the pace of its growth have created challenges not only in the development of new applications that realize the power of the Web, but also in the technology needed to scale applications to accommodate the resulting large data sets and heavy loads. This thesis presents searching algorithms and hierarchical classification techniques for increasing a search service's understanding of web queries. Existing search services rely solely on a query's occurrence in the document collection to locate relevant documents. They typically do not perform any task or topic-based analysis of queries using other available resources, and do not leverage changes in user query patterns over time. Provided within are a set of techniques and metrics for performing temporal analysis on query logs. Our log analyses are shown to be reasonable and informative, and can be used to detect changing trends and patterns in the query stream, thus providing valuable data to a search service.
Elimination of data redundancy before persisting into dbms using svm classifi...nalini manogaran
Elimination of data redundancy before persisting into dbms using svm classification,
Data Base Management System is one of the
growing fields in computing world. Grid computing, internet
sharing, distributed computing, parallel processing and cloud
are the areas store their huge amount of data in a DBMS to
maintain the structure of the data. Memory management is
one of the major portions in DBMS due to edit, delete, recover
and commit operations used on the records. To improve the
memory utilization efficiently, the redundant data should be
eliminated accurately. In this paper, the redundant data is
fetched by the Quick Search Bad Character (QSBC) function
and intimate to the DB admin to remove the redundancy.
QSBC function compares the entire data with patterns taken
from index table created for all the data persisted in the
DBMS to easy comparison of redundant (duplicate) data in
the database. This experiment in examined in SQL server
software on a university student database and performance is
evaluated in terms of time and accuracy. The database is
having 15000 students data involved in various activities.
Keywords—Data redundancy, Data Base Management System,
Support Vector Machine, Data Duplicate.
I. INTRODUCTION
The growing (prenominal) mass of information
present in digital media has become a resistive problem for
data administrators. Usually, shaped on data congregate
from distinct origin, data repositories such as those used by
digital libraries and e-commerce agent based records with
disparate schemata and structures. Also problems regarding
to low response time, availability, security and quality
assurance become more troublesome to manage as the
amount of data grow larger. It is practicable to specimen
that the peculiarity of the data that an association uses in its
systems is relative to its efficiency for offering beneficial
services to their users. In this environment, the
determination of maintenance repositories with “dirty” data
(i.e., with replicas, identification errors, equal patterns,
etc.) goes greatly beyond technical discussion such as the
everywhere quickness or accomplishment of data
administration systems.
Nalini.M, nalini.tptwin@gmail.com, Anbu.S, anomaly detection,
data mining
big data
dbms
intrusion detection
dublicate detection
data cleaning
data redundancy
data replication, redundancy removel, QSBC, Duplicate detection, error correction, de-duplication, Data cleaning, Dbms, Data sets
This document summarizes a research paper that proposes a new approach for web content extraction using soft computing algorithms and a trinity structure. The proposed system uses fuzzy logic for multi-website crawling, genetic algorithms to load extracted data into a trinity structure, and ant colony optimization for accurate data extraction without NP-complete problems. It aims to more efficiently extract exact web documents through the use of a decision tree algorithm along with the trinity search approach.
A Web Extraction Using Soft Algorithm for Trinity Structureiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Missing Value Evaluation in SQL Queries: A SurveyIRJET Journal
This document summarizes research on evaluating missing values, or "why-not" questions, in SQL queries. It surveys techniques used to answer why-not questions for both numeric and non-numeric data. The paper compares strategies like query refinement, index-based algorithms, and ranking functions. It also outlines future work on social and graph queries before concluding that research on answering why and why-not questions in different data settings can make databases more interactive and transparent for users.
IRJET- Missing Value Evaluation in SQL Queries: A SurveyIRJET Journal
This document summarizes research on evaluating missing values, or "why-not" questions, in SQL queries. It surveys techniques used to answer why-not questions for both numeric and non-numeric data. The document compares strategies like query refinement, index-based algorithms, and ranking functions. It also outlines future work on applying these techniques to social network and graph queries. The goal is to make database systems more transparent and interactive by supporting exploratory analysis of missing answers.
Similar to International Journal of Engineering and Science Invention (IJESI) (20)
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
GraphRAG for Life Science to increase LLM accuracyTomaz Bratanic
GraphRAG for life science domain, where you retriever information from biomedical knowledge graphs using LLMs to increase the accuracy and performance of generated answers
HCL Notes und Domino Lizenzkostenreduzierung in der Welt von DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-und-domino-lizenzkostenreduzierung-in-der-welt-von-dlau/
DLAU und die Lizenzen nach dem CCB- und CCX-Modell sind für viele in der HCL-Community seit letztem Jahr ein heißes Thema. Als Notes- oder Domino-Kunde haben Sie vielleicht mit unerwartet hohen Benutzerzahlen und Lizenzgebühren zu kämpfen. Sie fragen sich vielleicht, wie diese neue Art der Lizenzierung funktioniert und welchen Nutzen sie Ihnen bringt. Vor allem wollen Sie sicherlich Ihr Budget einhalten und Kosten sparen, wo immer möglich. Das verstehen wir und wir möchten Ihnen dabei helfen!
Wir erklären Ihnen, wie Sie häufige Konfigurationsprobleme lösen können, die dazu führen können, dass mehr Benutzer gezählt werden als nötig, und wie Sie überflüssige oder ungenutzte Konten identifizieren und entfernen können, um Geld zu sparen. Es gibt auch einige Ansätze, die zu unnötigen Ausgaben führen können, z. B. wenn ein Personendokument anstelle eines Mail-Ins für geteilte Mailboxen verwendet wird. Wir zeigen Ihnen solche Fälle und deren Lösungen. Und natürlich erklären wir Ihnen das neue Lizenzmodell.
Nehmen Sie an diesem Webinar teil, bei dem HCL-Ambassador Marc Thomas und Gastredner Franz Walder Ihnen diese neue Welt näherbringen. Es vermittelt Ihnen die Tools und das Know-how, um den Überblick zu bewahren. Sie werden in der Lage sein, Ihre Kosten durch eine optimierte Domino-Konfiguration zu reduzieren und auch in Zukunft gering zu halten.
Diese Themen werden behandelt
- Reduzierung der Lizenzkosten durch Auffinden und Beheben von Fehlkonfigurationen und überflüssigen Konten
- Wie funktionieren CCB- und CCX-Lizenzen wirklich?
- Verstehen des DLAU-Tools und wie man es am besten nutzt
- Tipps für häufige Problembereiche, wie z. B. Team-Postfächer, Funktions-/Testbenutzer usw.
- Praxisbeispiele und Best Practices zum sofortigen Umsetzen
Let's Integrate MuleSoft RPA, COMPOSER, APM with AWS IDP along with Slackshyamraj55
Discover the seamless integration of RPA (Robotic Process Automation), COMPOSER, and APM with AWS IDP enhanced with Slack notifications. Explore how these technologies converge to streamline workflows, optimize performance, and ensure secure access, all while leveraging the power of AWS IDP and real-time communication via Slack notifications.
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Removing Uninteresting Bytes in Software FuzzingAftab Hussain
Imagine a world where software fuzzing, the process of mutating bytes in test seeds to uncover hidden and erroneous program behaviors, becomes faster and more effective. A lot depends on the initial seeds, which can significantly dictate the trajectory of a fuzzing campaign, particularly in terms of how long it takes to uncover interesting behaviour in your code. We introduce DIAR, a technique designed to speedup fuzzing campaigns by pinpointing and eliminating those uninteresting bytes in the seeds. Picture this: instead of wasting valuable resources on meaningless mutations in large, bloated seeds, DIAR removes the unnecessary bytes, streamlining the entire process.
In this work, we equipped AFL, a popular fuzzer, with DIAR and examined two critical Linux libraries -- Libxml's xmllint, a tool for parsing xml documents, and Binutil's readelf, an essential debugging and security analysis command-line tool used to display detailed information about ELF (Executable and Linkable Format). Our preliminary results show that AFL+DIAR does not only discover new paths more quickly but also achieves higher coverage overall. This work thus showcases how starting with lean and optimized seeds can lead to faster, more comprehensive fuzzing campaigns -- and DIAR helps you find such seeds.
- These are slides of the talk given at IEEE International Conference on Software Testing Verification and Validation Workshop, ICSTW 2022.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
Essentials of Automations: The Art of Triggers and Actions in FMESafe Software
In this second installment of our Essentials of Automations webinar series, we’ll explore the landscape of triggers and actions, guiding you through the nuances of authoring and adapting workspaces for seamless automations. Gain an understanding of the full spectrum of triggers and actions available in FME, empowering you to enhance your workspaces for efficient automation.
We’ll kick things off by showcasing the most commonly used event-based triggers, introducing you to various automation workflows like manual triggers, schedules, directory watchers, and more. Plus, see how these elements play out in real scenarios.
Whether you’re tweaking your current setup or building from the ground up, this session will arm you with the tools and insights needed to transform your FME usage into a powerhouse of productivity. Join us to discover effective strategies that simplify complex processes, enhancing your productivity and transforming your data management practices with FME. Let’s turn complexity into clarity and make your workspaces work wonders!
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Cosa hanno in comune un mattoncino Lego e la backdoor XZ?Speck&Tech
ABSTRACT: A prima vista, un mattoncino Lego e la backdoor XZ potrebbero avere in comune il fatto di essere entrambi blocchi di costruzione, o dipendenze di progetti creativi e software. La realtà è che un mattoncino Lego e il caso della backdoor XZ hanno molto di più di tutto ciò in comune.
Partecipate alla presentazione per immergervi in una storia di interoperabilità, standard e formati aperti, per poi discutere del ruolo importante che i contributori hanno in una comunità open source sostenibile.
BIO: Sostenitrice del software libero e dei formati standard e aperti. È stata un membro attivo dei progetti Fedora e openSUSE e ha co-fondato l'Associazione LibreItalia dove è stata coinvolta in diversi eventi, migrazioni e formazione relativi a LibreOffice. In precedenza ha lavorato a migrazioni e corsi di formazione su LibreOffice per diverse amministrazioni pubbliche e privati. Da gennaio 2020 lavora in SUSE come Software Release Engineer per Uyuni e SUSE Manager e quando non segue la sua passione per i computer e per Geeko coltiva la sua curiosità per l'astronomia (da cui deriva il suo nickname deneb_alpha).
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdfMalak Abu Hammad
Discover how MongoDB Atlas and vector search technology can revolutionize your application's search capabilities. This comprehensive presentation covers:
* What is Vector Search?
* Importance and benefits of vector search
* Practical use cases across various industries
* Step-by-step implementation guide
* Live demos with code snippets
* Enhancing LLM capabilities with vector search
* Best practices and optimization strategies
Perfect for developers, AI enthusiasts, and tech leaders. Learn how to leverage MongoDB Atlas to deliver highly relevant, context-aware search results, transforming your data retrieval process. Stay ahead in tech innovation and maximize the potential of your applications.
#MongoDB #VectorSearch #AI #SemanticSearch #TechInnovation #DataScience #LLM #MachineLearning #SearchTechnology
Unlock the Future of Search with MongoDB Atlas_ Vector Search Unleashed.pdf
International Journal of Engineering and Science Invention (IJESI)
1. International Journal of Engineering Science Invention
ISSN (Online): 2319 – 6734, ISSN (Print): 2319 – 6726
www.ijesi.org Volume 2 Issue 10ǁ October 2013 ǁ PP.56-61
An Expressive Multiple Query Processing For The Patented
Medical Database in Handling the Temporal Domain Event
1 T.Yogameera, 2 Dr.D.Shanthi Saravanan
1
1,2,
2
PG Student, Professor,
Department of Computer Science and Engineering, P.S.N.A College of Engineering and Technology,
Tamilnadu.
ABSTRACT: The Intellectual query processing has become mandatory for efficient information retrieval .The
traditional approaches such as Try and See approach, Prior-Art- Search, As You Type approach, Fuzzy
Approach, Filtering algorithms, Graph Methods were not sufficiently proven worth in the upcoming temporal
events when implied in a context of data mart or an enterprise database. This paper suggest a innovative
methodology with three streamlined techniques namely Automatic Error Correction, Topic Relevant Query
Suggestion with extended Query Augmentation to enhance the functionality of patented data search in high
dimensional databases. The patented data from the sources are first clustered into topics and classes, when
given a query the highly coherent cluster partitions are recovered. The upshots in each coherent cluster are
combined generating top K relevant answers for the examiners from the database. After going through a
detailed study about the different literatures on search and retrieval of information, it is decided to propose a
new novel approach that amplifies the user’s intention contour and enhances the retrieval time with more
efficient memory management of the database. Further this technique can be implemented in patented
medical databases which would give better results with an economical Query processing in accurate and
proficient electronic data systems.
KEY WORDS: Automatic Error Correction, Query Augmentations, Query Analysis, Referential
Medical database, Patented Structure.
I.
INTRODUCTION
The Patented Medical Databases have now been used for referential report generations with detailed
and analyzed metadata structure. The Research is underway to implement the referential analysis with the
automated machines and with human robots, so that the process can be with accurate analyzation and speed up
the further treatment after observations recorded. In accurate syndrome cases and other immediately diagnostic
needed cases like echo cardio problem, cancer etc. We are in need of proper less time consuming appraised
information retrieval tool to be designed. Thus relating this module to the domain of data mining the query
processing becomes the core of attention needed for automated/expert referential activity. The common
approaches like As You Type, Prior Art Relevancy, Graph Method, Type Ahead Search, Click through Data,
SVM Ranking have not proven worth efficient in this era of medical treatments. Seeking the problem,
the proposed 3 proven streamlined techniques Automatic Error correction, Topic Relevant Query suggestion
with extended query augmentation helps in précised query contour and quick retrieval of referential data from
the patented medical databases in the hospital data mart which holds big data of a medical forum. More over
when implemented in cloud the multiple queries processing with the foreign enhanced metadata analysis
provides the best way in making proper decisions for the expert doctors.
1.https://www.google.com/patents.
2.www.bmj,com/content..
3.www.nationalarchieves.gov.uk/informati on-management.
4. www.annauniv.edu/ipr.
II.
OVER VIEW ON RELATED TECHNIQUES:
2.1 Query search techniques: Click through data [1]: It finds the subset of the surveying data, the Boolean
operators used for scaling has only three criteria (Disagree, Neutral, and Agree). The technique consumes
much time and effort due to lack of understanding of functionality of search context. In As You Type[9]
www.ijesi.org
56 | Page
2. An Expressive Multiple Query Processing…
approach, letter by letter query suggestion with topic relevancy is provided, the user gratification found by
the trie structure currently by suffix part of keyword explores a huge search space thus it need to be dealt with,
moreover the error correction becomes a trial approach only. SVM ranking [12] of top k answers, concerned
with all context of information retrieval performed with the citation edges in the graph, word net when easy to
maintain in the homogeneous database. The performance is not sensitive to heterogeneous database with
tunable parameter (α). More over the complex conceptual indexing based on large scale database and backend
algorithms with AND/OR semantics need to be concentrated. Prior-Art-Search [11] Presence of mismatch and
vague terms was found by the pseudo relevance feedback and automatically select better match, but there is a
need for enhancement of this mile stone approach with extended query augmentation in statistical distribution
which now deals only with less skewed retrieval.
2.2 Query processing techniques:Pattern Matching [1] NFA computation for dealing with temporal events
must concentrate on shared buffer and database with current version states and points to recent events, for
future edge evaluation with both logical and temporal decision making. Regular expression matching [3] the
queries are converted into regular expressions with NFA binary logic, The Field Programmable Gate Array
(FPGA) extended to self reconfigurable GA for the configuration bit generation reduces the number of state
traversals thus speed up the row and column traversal and search operations.
2.3 Information Retrieval strategies for the given Processed Query: Backtracking algorithm [10] the
processing of the query using selection- join-aggregation was enhanced with run time efficiency with this
algorithm, It is an apt logical programming algorithm for the constrain based satisfaction. It finds all the possible
solutions within the time bound as search space has been pruned with the invalid branch optimization. Trie
Structure Analysis [2] the current retrieval based on prefix part in the sub-linear search algorithm need to
reconsider in the temporal and structured high dimensional data marts. The length matching and loop
processing when taken into account may result in fast pruning of search space that needs to be dealt with
inverted indexing Multiple Query Optimizations[4] (MQO) The spanning of multiple events with parameterized
scalability needs unambiguous indexing. In addition the query rewriting must be performed to speed up
subscriptions and publish the notifications of the related events after finding the filtered commonalities and
merging them for the efficient retrieval.
III.
RELATED WORKS AND DISCUSSION ON FUTURE WORKS
Larkey[11] has studied the problem of patent classification but neglected the Prior Art search which
is present milestone. X.Xue & W.B Croft [5] discusses about the query generation in the patent for finding the
referential answer. Our problem constrained here with the relevancy of the retrieved results.
Azzorpadi[6]surveyed 8 patents(including Medical DB) to obtain preference and functionality oriented finding
for the given query with two approaches prior art search & sophisticated method of content analysis proving
better retrieval of results with citation analysis with indexed SVM ranking[12]. Yan Cao, Jufan & Gu.Li[7] has
discussed about the automatic error correction approach with the partial or full keyword relevancies and
retrieving top K answers from the partitions Our proposed system differs from existing system when
implemented in the medical data analysis keeping all the advantages of the above approaches with enhanced
user friendly and expressive processing of high dimensional data analysis in temporal events.
FIG 1: A low level architecture, describes the information management system in the hospital enterprise with
efficient query processing.
www.ijesi.org
57 | Page
3. An Expressive Multiple Query Processing…
The information management structure in the medical database used in appraised query processing has
been modeled. The posted query by the expert will first be verified and automatic error correction is
performed to enter into correct consolidated record links then the topic based suggestion such as echo,
cancer, liver disorder etc based on the temporal events updated is provided. Further the query is expanded and
rewritten by the system which uses inverted index to capture the upshots. The top k answers are retrieved
based on the citation edges and query keyword, and visualized to the expert.The existing technologies
discussed in the related works need to be further concentrated with the temporal high dimensional database
events. The clustering and classification technique is the perfect management strategy to handle the upcoming
Big data that are topically related. The query when proposed to the system must be deeply processed with
both topic relevancy and query key relevancy with automated error correction techniques made possible.Though
the recently available trends mines the query results based on the content, it is indeed necessary to step into the
smart mining with improved vision on intelligent information retrieval from the engineered database. Thus the
points to concentrate on include the retrieval of query results that are highly relevant are the following
The relevant results within accurate
user‟s contour.
The speedup of appraisal in the big data analyzation
Clouding
the
structured(or) patented databases that are clustered based on the topics and related
based on their inner class partitions
Finding of positive partition classes for the given query.
Data visualization in ranked order with pattern matching made more efficient.
Our paper concentrates in these points. The qualified technique Automatic Error correction now not
mostly available in medical database is quintessential for accurate result retrieval. The Backtracking
algorithm is used for a quick test whether the partial solution can be a valid solution or not. The recursive
depth first search strategy in our tree helps in pruning the irrelevant search space and in determining
whether the branch is valid or not thus ongoing with nano unit time operations. Topic based
query suggestion can be achieved by our proposed temporal pattern search algorithm that arrays the
events of same type with the time stamp value, this is useful to skip unnecessary histories and
encapsulate the users recall view of records. The recently enhanced Faster B-Tree algorithm is used to
make a sorted data storage and perform the updating and retrieval of the temporal events in
logarithmic time, thus helps in query expansion and suggestion that makes a friendly interface for the
user.
IV.
EXPERIMENTAL RESULTS:
We have implemented our proposed techniques. We compared with the prior art technique and
SVM ranking in the retrieval of information in the simulated medical database. The obtained results were
satisfying in the advanced information retrieval in respect of relevancy, quality and ranking within the bounded
time factor.
4.1 Relevancy of documents:
We evaluate the effect on k (the number of selected partitions). We partitioned the records into topics
and classes and evaluated the effectiveness and quality by varying the value of k. To evaluate the result quality,
we used the milestone technique p@k, where the precision is the ratio of the number of retrieved relevant results
to the number of retrieved results, and p@k is the precision of the top-k results.
www.ijesi.org
58 | Page
4. An Expressive Multiple Query Processing…
Fig2b.Comparision with different approaches (with p@k)
Fig. 2a shows the results. The main reason is as follows: First, the more records used to answers a query, the
more relevant answers, and thus the higher precision. Second, as each query usually belongs to limited number
of topics, only several partitions are relevant to the query. Thus, the precision is stable when „k‟ is large enough.
For example, where k „>‟ 10, they achieved nearly the same precision. Then, we evaluated the efficiency. Fig.
2b shows the results. We see that with the increase of k, the elapsed time increased. This is because the more the
categories used to answer a query the efficiency increases.
4.2 Precision Comparison: In this section, we compare the result quality. Table 1 shows the experimental
results. We can see that our three techniques can improve the result quality. For example, for p@50, error
correction improved to 0.83, query suggestion increases the precision to 0.82, and query expansion can
improve the precision to 0.88. More importantly, our method by combing the three methods can improve the
precision to 0.88, and the improvement ratio is about 0.84, achieved by query rewriting. The main reason is as
follows: First, the automatic error correction can provide users accurate keywords based on users inputs.
Second, the query expansion can suggest relevant keywords. Third, topic-based query suggestion can provide
users topic- relevant keywords.
Table 1: Quality comparison
4.3 Efficiency Comparison : In this section, we compare the efficiency. We partitioned the data into 24
partitions. We used three computers to manage the data and each node managed three partitions. For each
partition, we built the corresponding inverted indexes. We first compared the two methods by varying the
number of keywords. We can see that for different numbers of keywords, our method always outperformed the
existing method SVMPR, with the speedup ratio about 8. This reflects that our method achieved high
efficiency since we employ an effective partition-based method. We then evaluated the two methods by
varying the number of returned answers. We can see that for different values, our method always
outperformed SVMPR. This is because, we partition the data into eight partitions and each partition was
inversely indexed and searched by different cores. More importantly, our partition-based method can prune
the search space and thus can improve the performance significantly.
www.ijesi.org
59 | Page
5. An Expressive Multiple Query Processing…
4.4 Scalability :We also evaluate the scalability of our method. Fig. 4a shows the experimental results by
varying the number of patents. Fig. 4a shows the scalability on quality and we can see that with the increase of
the number of records, the precision of query suggestion and query expansion increased slightly. This is
because we can utilize
more data to select the topic of each record and find more relevant keyword pairs. The more data used the higher
precision of the topic model. For query correction, the precision nearly kept the same as we only used the trie
structure to correct the keywords but the quality increased as we prposed the prefix based depth first search. On
the other hand, our method scaled very well.
V.
CONCLUSION:
The user friendly expert analysis query processing techniques discussed so far had made milestones in
the domain of information retrieval and management. But our proposed suggestion may make a perfect
benchmark. The patented data from the sources are first clustered into topics and classes, when given a query
the highly coherent cluster partitions are recovered. The upshots in each coherent cluster are combined
generating top K relevant answers for the examiners from the database. The ongoing implementation of
our method had shown high efficiency and increased quality of the partitioned based search strategy with the
simulated results. The techniques streamlined namely query suggestions,
Topic
relevancy,
query
augmentation when handshake with existing prior art relevancy may prove marvelous enhancements
in knowledge engineering.
VI.
Acknowledgement:
It is my pleasure to acknowledge Mr.J.Venkatesan Prabhu, Head, Kaashiv- InfoTech, Chennai for his
support in implementation part and for his guidance throughout this course of work.
REFERENCES
[1]
[2]
[4]
[5]
[6]
[7]
[8]
[9]
[10]
[11]
J. Agrawal, Y. Diao, D. Gyllstrom, and N. Immerman, “Efficient Pattern Matching over Event Streams,” Proc. ACM
SIGMOD Int‟l Conf. Management of Data, pp. 147-160, 2008.
R.S. Boyer and J.S. Moore, “A Fast String-Searching Algorithm,” Comm. ACM, vol. 20, no. 10, pp. 762-772, 1977. [3] R.
Cox, “Regular Expression Matching Can
Be
Simple
and
Fast,” http://swtch.com/rsc/regexp/regexp1.html,
2007.
A. Demers, J. Gehrke, M. Hong, M. Riedewald, and W. White, “Towards Expressive Publish/Subscribe Systems,” Proc. 10th
Int‟l Conf. Extending Database Technology (EDBT), pp. 627-644, 2006
X. Xue and W.B Croft, “Automatic Query Generation for Patent Search,” Proc. ACM Conf. Information and Knowledge
Management (CIKM), pp.
2037-2040, 2009.
L. Azzopardi, W. Vanderbauwhede, and H. Joho, “Search System Requirements of Patent Analysts,” Proc. 33rd Int‟l ACM
SIGIR Conf. Researchand Development in Information Retrieval (SIGIR), pp.775-776, 2010.
A User-Friendly Patent Search Paradigm Yang Cao, Ju Fan, and Guoliang Li IEEE Transaction on knowledge and data
engineering, vol. 25, no. 6, 2013.
L.S. Larkey, “A Patent Search and Classification System,” Proc. Fourth ACM Conf. Digital Libraries, pp. 179-187,
1999.
J. Fan, H. Wu, G. Li, and L. Zhou, “Suggesting Topic-Based Query Terms as You Type,” Proc. Int‟l Asia Pacific Web Conf.
(APWEB), pp. 61-67, 2010.
G. Li, J. Feng, and C. Li, “Supporting Search-As-You-Type Using SQL in Databases,” IEEE Trans. Knowledge and Data Eng.,
vol. 25, no. 2, pp. 461-475, Feb. 2013.
S. Bashir and A. Rauber, “Improving Retrievability of Patents in Prior-Art Search,” Proc. European Conf. Information
Retrieval (ECIR), pp. 457470, 2010.
www.ijesi.org
60 | Page
6. An Expressive Multiple Query Processing…
[12]
Y. Guo and C.P. Gomes, “Ranking Structured Documents: A Large Margin Based Approach for Patent Prior Art Search,” Proc.
Int‟l Joint Conf. Artificial Intelligence (IJCAI), pp. 1058-1064, 2009.
AUTHOR DETAILS:
Ms.T.Yogameera completed her B.E Computer science and Engineering in 2012 with first
class from R.V.S College of Engineering and Technology, Dindigul, TamilNadu and pursuing M.E. Computer
Science and Engineering in P.S.N.A College of Engineering and Technology, Dindigul, TamilNadu. She has
secured best academician award 2 times during her school days in Shri Maharishi Vidya Mandir(CBSE),
Dindigul, TamilNadu, she has grabbed the first best project award in the international science and technology
contest during her 9TH, has secured state 2ND rank during her 11TH , School second rank in her 12TH . She
has participated in two international symposiums presenting her own paper and won second prize in both, she
has attended 3 national level workshops and 1 international workshop, 2 national seminars, and 15 days
internship program on mobile operating system and has been acknowledged as the class representative for 4
continuous semesters during her B.E. Her domains of interest are data mining, information retrieval, query
processing, database management, currently concentrating on information management in medical systems.
Dr.D.Shanthi Saravanan received her B.E Computer Science and Engineering in 1992
from Thiagaraja College of Engineering, Madurai, TamilNadu and M.E Computer Science and Engineering
from Manonmaniam Sundaranar University, TamilNadu and PhD in Soft Computing from Brila Institute
of Technology, Ranchi. She is currently working as a Professor in Department of Computer Science and
Engineering, PSNA College of Engineering and Technology. She has more than 21 years of Teaching and
Programming Experience. She is a member of various professional societies like IEEE, CSTA, IAENG and
IACSIT. Her research interest includes Genetic Algorithm, Neural Networks, and Intelligent Systems, Image
processing, Embedded System, Machine Learning and Green Computing. She has published more than 25
papers in international journals and conferences and also 5 books in computing and applications. She is a
reviewer of various international journals.
www.ijesi.org
61 | Page