This document discusses the use of fuzzy queries to retrieve information from databases. Fuzzy queries allow for imprecise or vague terms to be used in queries, similar to natural language. The document first provides background on limitations of traditional database queries. It then discusses how fuzzy set theory and membership functions can be applied to queries and data to handle uncertain terms. The proposed approach applies fuzzy queries to a relational database, defining linguistic variables and membership functions. This allows information to be retrieved based on fuzzy criteria and improves the ability to query databases using human-like terms. Benefits of fuzzy queries include more natural interaction and accounting for real-world data imperfections.
Performance Evaluation of Query Processing Techniques in Information Retrievalidescitation
The first element of the search process is the query.
The user query being on an average restricted to two or three
keywords makes the query ambiguous to the search engine.
Given the user query, the goal of an Information Retrieval
[IR] system is to retrieve information which might be useful
or relevant to the information need of the user. Hence, the
query processing plays an important role in IR system.
The query processing can be divided into four categories
i.e. query expansion, query optimization, query classification and
query parsing. In this paper an attempt is made to evaluate the
performance of query processing algorithms in each of the
category. The evaluation was based on dataset as specified by
Forum for Information Retrieval [FIRE15]. The criteria used
for evaluation are precision and relative recall. The analysis is
based on the importance of each step in query processing. The
experimental results show that the significance of each step
in query processing and also the relevance of web semantics
and spelling correction in the user query.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Text mining efforts to innovate new, previous unknown or hidden data by automatically extracting
collection of information from various written resources. Applying knowledge detection method to
formless text is known as Knowledge Discovery in Text or Text data mining and also called Text Mining.
Most of the techniques used in Text Mining are found on the statistical study of a term either word or
phrase. There are different algorithms in Text mining are used in the previous method. For example
Single-Link Algorithm and Self-Organizing Mapping(SOM) is introduces an approach for visualizing
high-dimensional data and a very useful tool for processing textual data based on Projection method.
Genetic and Sequential algorithms are provide the capability for multiscale representation of datasets and
fast to compute with less CPU time based on the Isolet Reduces subsets in Unsupervised Feature
Selection. We are going to propose the Vector Space Model and Concept based analysis algorithm it will
improve the text clustering quality and a better text clustering result may achieve. We think it is a good
behavior of the proposed algorithm is in terms of toughness and constancy with respect to the formation of
Neural Network.
Seeds Affinity Propagation Based on Text ClusteringIJRES Journal
The objective is to find among all partitions of the data set, best publishing according to some quality measure. Affinity propagation is a low error, high speed, flexible, and remarkably simple clustering algorithm that may be used in forming teams of participants for business simulations and experiential exercises, and in organizing participant’s preferences for the parameters of simulations. This paper proposes an efficient Affinity Propagation algorithm that guarantees the same clustering result as the original algorithm after convergence. The heart of our approach is (1) to prune unnecessary message exchanges in the iterations and (2) to compute the convergence values of pruned messages after the iterations to determine clusters.
This paper proposes Natural language based Discourse Analysis method used for extracting
information from the news article of different domain. The Discourse analysis used the Rhetorical Structure
theory which is used to find coherent group of text which are most prominent for extracting information
from text. RST theory used the Nucleus- Satellite concept for finding most prominent text from the text
document. After Discourse analysis the text analysis has been done for extracting domain related object
and relates this object. For extracting the information knowledge based system has been used which
consist of domain dictionary .The domain dictionary has a bag of words for domain. The system is
evaluated according gold-of-art analysis and human decision for extracted information.
Performance Evaluation of Query Processing Techniques in Information Retrievalidescitation
The first element of the search process is the query.
The user query being on an average restricted to two or three
keywords makes the query ambiguous to the search engine.
Given the user query, the goal of an Information Retrieval
[IR] system is to retrieve information which might be useful
or relevant to the information need of the user. Hence, the
query processing plays an important role in IR system.
The query processing can be divided into four categories
i.e. query expansion, query optimization, query classification and
query parsing. In this paper an attempt is made to evaluate the
performance of query processing algorithms in each of the
category. The evaluation was based on dataset as specified by
Forum for Information Retrieval [FIRE15]. The criteria used
for evaluation are precision and relative recall. The analysis is
based on the importance of each step in query processing. The
experimental results show that the significance of each step
in query processing and also the relevance of web semantics
and spelling correction in the user query.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Text mining efforts to innovate new, previous unknown or hidden data by automatically extracting
collection of information from various written resources. Applying knowledge detection method to
formless text is known as Knowledge Discovery in Text or Text data mining and also called Text Mining.
Most of the techniques used in Text Mining are found on the statistical study of a term either word or
phrase. There are different algorithms in Text mining are used in the previous method. For example
Single-Link Algorithm and Self-Organizing Mapping(SOM) is introduces an approach for visualizing
high-dimensional data and a very useful tool for processing textual data based on Projection method.
Genetic and Sequential algorithms are provide the capability for multiscale representation of datasets and
fast to compute with less CPU time based on the Isolet Reduces subsets in Unsupervised Feature
Selection. We are going to propose the Vector Space Model and Concept based analysis algorithm it will
improve the text clustering quality and a better text clustering result may achieve. We think it is a good
behavior of the proposed algorithm is in terms of toughness and constancy with respect to the formation of
Neural Network.
Seeds Affinity Propagation Based on Text ClusteringIJRES Journal
The objective is to find among all partitions of the data set, best publishing according to some quality measure. Affinity propagation is a low error, high speed, flexible, and remarkably simple clustering algorithm that may be used in forming teams of participants for business simulations and experiential exercises, and in organizing participant’s preferences for the parameters of simulations. This paper proposes an efficient Affinity Propagation algorithm that guarantees the same clustering result as the original algorithm after convergence. The heart of our approach is (1) to prune unnecessary message exchanges in the iterations and (2) to compute the convergence values of pruned messages after the iterations to determine clusters.
This paper proposes Natural language based Discourse Analysis method used for extracting
information from the news article of different domain. The Discourse analysis used the Rhetorical Structure
theory which is used to find coherent group of text which are most prominent for extracting information
from text. RST theory used the Nucleus- Satellite concept for finding most prominent text from the text
document. After Discourse analysis the text analysis has been done for extracting domain related object
and relates this object. For extracting the information knowledge based system has been used which
consist of domain dictionary .The domain dictionary has a bag of words for domain. The system is
evaluated according gold-of-art analysis and human decision for extracted information.
AUTOMATED INFORMATION RETRIEVAL MODEL USING FP GROWTH BASED FUZZY PARTICLE SW...ijcseit
To mine out relevant facts at the time of need from web has been a tenuous task. Research on diverse fields
are fine tuning methodologies toward these goals that extracts the best of information relevant to the users
search query. In the proposed methodology discussed in this paper find ways to ease the search complexity
tackling the severe issues hindering the performance of traditional approaches in use. The proposed
methodology find effective means to find all possible semantic relatable frequent sets with FP Growth
algorithm. The outcome of which is the further source of fuel for Bio inspired Fuzzy PSO to find the optimal
attractive points for the web documents to get clustered meeting the requirement of the search query
without losing the relevance. On the whole the proposed system optimizes the objective function of
minimizing the intra cluster differences and maximizes the inter cluster distances along with retention of all
possible relationships with the search context intact. The major contribution being the system finds all
possible combinations matching the user search transaction and thereby making the system more
meaningful. These relatable sets form the set of particles for Fuzzy Clustering as well as PSO and thus
being unbiased and maintains a innate behaviour for any number of new additions to follow the herd
behaviour’s evaluations reveals the proposed methodology fares well as an optimized and effective
enhancements over the conventional approaches.
Text Mining is the technique that helps users to find out useful information from a large amount of text documents on the web or database. Most popular text mining and classification methods have adopted term-based approaches. The term based approaches and the pattern-based method describing user preferences. This review paper analyse how the text mining work on the three level i.e sentence level, document level and feature level. In this paper we review the related work which is previously done. This paper also demonstrated that what are the problems arise while doing text mining done at the feature level. This paper presents the technique to text mining for the compound sentences.
ONTOLOGY-DRIVEN INFORMATION RETRIEVAL FOR HEALTHCARE INFORMATION SYSTEM : ...IJNSA Journal
In health research, one of the major tasks is to retrieve, and analyze heterogeneous databases containing
one single patient’s information gathered from a large volume of data over a long period of time. The
main objective of this paper is to represent our ontology-based information retrieval approach for
clinical Information System. We have performed a Case Study in the real life hospital settings. The results
obtained illustrate the feasibility of the proposed approach which significantly improved the information
retrieval process on a large volume of data over a long period of time from August 2011 until January
2012
The enormous amount of information stored in unstructured texts cannot simply be used for further
processing by computers, which typically handle text as simple sequences of character strings. Therefore, specific
(pre-) processing methods and algorithms are required in order to extract useful patterns. Text Mining is the
discovery of valuable, yet hidden, information from the text document. Text classification (Also called Text
Categorization) is one of the important research issues in the field of text mining. It is necessary to
classify/categorize large texts (documents) into specific classes. Text Classification assigns a text document to one of a
set of predefined classes. This paper covers different text classification techniques and also includes Classifier
Architecture and Text Classification Applications.
Hybrid approach for generating non overlapped substring using genetic algorithmeSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Novel Database-Centric Framework for Incremental Information Extractionijsrd.com
Information extraction (IE) has been an active research area that seeks techniques to uncover information from a large collection of text. IE is the task of automatically extracting structured information from unstructured and/or semi structured machine-readable documents. In most of the cases this activity concerns processing human language texts by means of natural language processing (NLP). Recent activities in document processing like automatic annotation and content extraction could be seen as information extraction. Many applications call for methods to enable automatic extraction of structured information from unstructured natural language text. Due to the inherent challenges of natural language processing, most of the existing methods for information extraction from text tend to be domain specific. In this project a new paradigm for information extraction. In this extraction framework, intermediate output of each text processing component is stored so that only the improved component has to be deployed to the entire corpus. Extraction is then performed on both the previously processed data from the unchanged components as well as the updated data generated by the improved component. Performing such kind of incremental extraction can result in a tremendous reduction of processing time and there is a mechanism to generate extraction queries from both labeled and unlabeled data. Query generation is critical so that casual users can specify their information needs without learning the query language.
A novel approach for text extraction using effective pattern matching techniqueeSAT Journals
Abstract
There are many data mining techniques have been proposed for mining useful patterns from documents. Still, how to effectively use and update discovered patterns is open for future research , especially in the field of text mining. As most existing text mining methods adopted term-based approaches, they all suffer from the problems of polysemy(words have multiple meanings) and synonymy(multiple words have same meaning). People have held hypothesis that pattern-based approaches should perform better than the term-based, but many experiments does not support this hypothesis. This paper presents an innovative and effective pattern discovery technique which includes the processes of pattern deploying and pattern matching, to improve the effective use of discovered patterns.
Keywords: Pattern Mining, Pattern Taxonomy Model, Inner Pattern Evolving, TF-IDF, NLP etc.
Prediction of Answer Keywords using Char-RNNIJECEIAES
Generating sequences of characters using a Recurrent Neural Network (RNN) is a tried and tested method for creating unique and context aware words, and is fundamental in Natural Language Processing tasks. These type of Neural Networks can also be used a question-answering system. The main drawback of most of these systems is that they work from a factoid database of information, and when queried about new and current information, the responses are usually bleak. In this paper, the author proposes a novel approach to finding answer keywords from a given body of news text or headline, based on the query that was asked, where the query would be of the nature of current affairs or recent news, with the use of Gated Recurrent Unit (GRU) variant of RNNs. Thus, this ensures that the answers provided are relevant to the content of query that was put forth.
Semantic Based Model for Text Document Clustering with IdiomsWaqas Tariq
Text document clustering has become an increasingly important problem in recent years because of the tremendous amount of unstructured data which is available in various forms in online forums such as the web, social networks, and other information networks. Clustering is a very powerful data mining technique to organize the large amount of information on the web. Traditionally, document clustering methods do not consider the semantic structure of the document. This paper addresses the task of developing an effective and efficient method to improve the semantic structure of the text documents. A method has been developed that performs the following: tag the documents for parsing, replacement of idioms with their original meaning, semantic weights calculation for document words and apply semantic grammar. The similarity measure is obtained between the documents and then the documents are clustered using Hierarchical clustering algorithm. The method adopted in this work is evaluated on different data sets with standard performance measures and the effectiveness of the method to develop in meaningful clusters has been proved.
Information Retrieval System is an effective process that helps a user to trace relevant information by Natural Language Processing (NLP). In this research paper, we have presented present an algorithmic Information Retrieval System(BIRS) based on information and the system is significant mathematically and statistically. This paper is demonstrated by two algorithms for finding out the lemmatization of Bengali words such as Trie and Dictionary Based Search by Removing Affix (DBSRA) as well as compared with Edit Distance for the exact lemmatization. We have presented the Bengali Anaphora resolution system using the Hobbs’ algorithm to get the correct expression of information. As the actions of questions answering algorithms, the TF-IDF and Cosine Similarity are developed to find out the accurate answer from the documents. In this study, we have introduced a Bengali Language Toolkit (BLTK) and Bengali Language Expression (BRE) that make the easiest implication of our task. We have also developed Bengali root word’s corpus, synonym word’s corpus, stop word’s corpus and gathered 672 articles from the popular Bengali newspapers ‘The Daily Prothom Alo’ which is our inserted information. For testing this system, we have created 19335 questions from the introduced information and got 97.22% accurate answer.
The Statement of Conjunctive and Disjunctive Queries in Object Oriented Datab...Editor IJCATR
Entrance of object orienting concept in database caused the relation database gradually to replace with object oriented
database in various fields. On the other hand for solving the problem of real world uncertain data, several methods were presented.
One of these methods for modeling database is an approach wich couples object-oriented database modeling with fuzzy logic. Many
queries that users to pose are expressed on the basis of linguistic variables. Because of classical databases are not able to support these
variables, leads to fuzzy approaches are considered. We investigate databases queries in this study both simple and complex ways. In
the complex way, we use conjunctive and disjunctive queries. In the following, we use the XML labels to express inqueries into fuzzy.
We can also communicate with other sections of software by entering into XML world as the most reliable opportunity. Also we want
to correct conjunctive and disjunctive queries related to fuzzy object oriented database using the concept of dependency measure and
weight, and weight be assigned to different phrases of a query based on user emphasis. The other aim of this research is mapping fuzzy
queries to fuzzy-XML. It is expected to be simple implement of query, and output of execution of queries be greatly closer to users'
needs and fulfill her expect. The results show that the proposed method explains the possible conjunctive and disjunctive queries the
database in the form of Fuzzy-XML.
Feature selection, optimization and clustering strategies of text documentsIJECEIAES
Clustering is one of the most researched areas of data mining applications in the contemporary literature. The need for efficient clustering is observed across wide sectors including consumer segmentation, categorization, shared filtering, document management, and indexing. The research of clustering task is to be performed prior to its adaptation in the text environment. Conventional approaches typically emphasized on the quantitative information where the selected features are numbers. Efforts also have been put forward for achieving efficient clustering in the context of categorical information where the selected features can assume nominal values. This manuscript presents an in-depth analysis of challenges of clustering in the text environment. Further, this paper also details prominent models proposed for clustering along with the pros and cons of each model. In addition, it also focuses on various latest developments in the clustering task in the social network and associated environments.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
A template based algorithm for automatic summarization and dialogue managemen...eSAT Journals
Abstract This paper describes an automated approach for extracting significant and useful events from unstructured text. The goal of research is to come out with a methodology which helps in extracting important events such as dates, places, and subjects of interest. It would be also convenient if the methodology helps in presenting the users with a shorter version of the text which contain all non-trivial information. We also discuss implementation of algorithms which exactly does this task, developed by us. Key Words: Cosine Similarity, Information, Natural Language, Summarization, Text Mining
This study investigate the characteristics of a gear system including contact stresses, bending
stresses, and the transmission errors of gears in mesh. Gearing is one of most critical component in
mechanical power transmission systems. The bending stresses in the tooth root were examined using 3D
model.
Current methods of calculating gear contact stresses use Hertz’s equations and Lewis Equation
which were originally derived for contact between two cylinders. To enable the investigation of contact
problems with FEM, the stiffness relationship between the two contact areas is usually established
through a spring placed between the two contacting areas. This can be achieve by inserting a contact
element in between the two areas where contact occurs. The results of the three dimensional FEM analysis
from ANSYS are presented. These stresses were compared with theoretical values. Both results agree very
well. This indicates that the FEM model is accurate.
This report also considers the variations of the whole body stiffness arising from gear body
rotation due to bending deflection, shearing displacement and contact deformation. Many different
positions within the meshing cycle were investigated. Investigation of contact and bending stress
characteristics of spur gears continues to be immense attention to both engineers and researchers in spite
of many studies in the past.
This is because of the advances in the engineering technology that demands for gears with ever
increasing load capacities and speeds with high reliability, the designers need to be able to accurately
predict the stresses experienced the stresses by the loaded gears.
As Diabetes Mellitus combined with other ailments will become a deadly combination, hence there
is an urgent need to break the link between diabetes and its related complications. For this purpose image
processing based analysis can potentially be helpful for earlier detection, education and treatment. Medical
image analysis of Diabetic patients with its related complications such as DR, CVD & Diabetic
Myonecrosis (i.e. on Retinal Images, Coronary angiographs, Electron micrographs, MRI etc) is to be the
aprioristic because of its more prevalence. Thus the main work of this paper is on literature review about
Diabetes and Imaging such as the Prevalence, Classification, Causes and Medical Imaging & Survey of
Image processing methods applied on Diabetic Related Causes.
Keywords — Image, segmentation, retinopathy, Myonecrosis,
AUTOMATED INFORMATION RETRIEVAL MODEL USING FP GROWTH BASED FUZZY PARTICLE SW...ijcseit
To mine out relevant facts at the time of need from web has been a tenuous task. Research on diverse fields
are fine tuning methodologies toward these goals that extracts the best of information relevant to the users
search query. In the proposed methodology discussed in this paper find ways to ease the search complexity
tackling the severe issues hindering the performance of traditional approaches in use. The proposed
methodology find effective means to find all possible semantic relatable frequent sets with FP Growth
algorithm. The outcome of which is the further source of fuel for Bio inspired Fuzzy PSO to find the optimal
attractive points for the web documents to get clustered meeting the requirement of the search query
without losing the relevance. On the whole the proposed system optimizes the objective function of
minimizing the intra cluster differences and maximizes the inter cluster distances along with retention of all
possible relationships with the search context intact. The major contribution being the system finds all
possible combinations matching the user search transaction and thereby making the system more
meaningful. These relatable sets form the set of particles for Fuzzy Clustering as well as PSO and thus
being unbiased and maintains a innate behaviour for any number of new additions to follow the herd
behaviour’s evaluations reveals the proposed methodology fares well as an optimized and effective
enhancements over the conventional approaches.
Text Mining is the technique that helps users to find out useful information from a large amount of text documents on the web or database. Most popular text mining and classification methods have adopted term-based approaches. The term based approaches and the pattern-based method describing user preferences. This review paper analyse how the text mining work on the three level i.e sentence level, document level and feature level. In this paper we review the related work which is previously done. This paper also demonstrated that what are the problems arise while doing text mining done at the feature level. This paper presents the technique to text mining for the compound sentences.
ONTOLOGY-DRIVEN INFORMATION RETRIEVAL FOR HEALTHCARE INFORMATION SYSTEM : ...IJNSA Journal
In health research, one of the major tasks is to retrieve, and analyze heterogeneous databases containing
one single patient’s information gathered from a large volume of data over a long period of time. The
main objective of this paper is to represent our ontology-based information retrieval approach for
clinical Information System. We have performed a Case Study in the real life hospital settings. The results
obtained illustrate the feasibility of the proposed approach which significantly improved the information
retrieval process on a large volume of data over a long period of time from August 2011 until January
2012
The enormous amount of information stored in unstructured texts cannot simply be used for further
processing by computers, which typically handle text as simple sequences of character strings. Therefore, specific
(pre-) processing methods and algorithms are required in order to extract useful patterns. Text Mining is the
discovery of valuable, yet hidden, information from the text document. Text classification (Also called Text
Categorization) is one of the important research issues in the field of text mining. It is necessary to
classify/categorize large texts (documents) into specific classes. Text Classification assigns a text document to one of a
set of predefined classes. This paper covers different text classification techniques and also includes Classifier
Architecture and Text Classification Applications.
Hybrid approach for generating non overlapped substring using genetic algorithmeSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Novel Database-Centric Framework for Incremental Information Extractionijsrd.com
Information extraction (IE) has been an active research area that seeks techniques to uncover information from a large collection of text. IE is the task of automatically extracting structured information from unstructured and/or semi structured machine-readable documents. In most of the cases this activity concerns processing human language texts by means of natural language processing (NLP). Recent activities in document processing like automatic annotation and content extraction could be seen as information extraction. Many applications call for methods to enable automatic extraction of structured information from unstructured natural language text. Due to the inherent challenges of natural language processing, most of the existing methods for information extraction from text tend to be domain specific. In this project a new paradigm for information extraction. In this extraction framework, intermediate output of each text processing component is stored so that only the improved component has to be deployed to the entire corpus. Extraction is then performed on both the previously processed data from the unchanged components as well as the updated data generated by the improved component. Performing such kind of incremental extraction can result in a tremendous reduction of processing time and there is a mechanism to generate extraction queries from both labeled and unlabeled data. Query generation is critical so that casual users can specify their information needs without learning the query language.
A novel approach for text extraction using effective pattern matching techniqueeSAT Journals
Abstract
There are many data mining techniques have been proposed for mining useful patterns from documents. Still, how to effectively use and update discovered patterns is open for future research , especially in the field of text mining. As most existing text mining methods adopted term-based approaches, they all suffer from the problems of polysemy(words have multiple meanings) and synonymy(multiple words have same meaning). People have held hypothesis that pattern-based approaches should perform better than the term-based, but many experiments does not support this hypothesis. This paper presents an innovative and effective pattern discovery technique which includes the processes of pattern deploying and pattern matching, to improve the effective use of discovered patterns.
Keywords: Pattern Mining, Pattern Taxonomy Model, Inner Pattern Evolving, TF-IDF, NLP etc.
Prediction of Answer Keywords using Char-RNNIJECEIAES
Generating sequences of characters using a Recurrent Neural Network (RNN) is a tried and tested method for creating unique and context aware words, and is fundamental in Natural Language Processing tasks. These type of Neural Networks can also be used a question-answering system. The main drawback of most of these systems is that they work from a factoid database of information, and when queried about new and current information, the responses are usually bleak. In this paper, the author proposes a novel approach to finding answer keywords from a given body of news text or headline, based on the query that was asked, where the query would be of the nature of current affairs or recent news, with the use of Gated Recurrent Unit (GRU) variant of RNNs. Thus, this ensures that the answers provided are relevant to the content of query that was put forth.
Semantic Based Model for Text Document Clustering with IdiomsWaqas Tariq
Text document clustering has become an increasingly important problem in recent years because of the tremendous amount of unstructured data which is available in various forms in online forums such as the web, social networks, and other information networks. Clustering is a very powerful data mining technique to organize the large amount of information on the web. Traditionally, document clustering methods do not consider the semantic structure of the document. This paper addresses the task of developing an effective and efficient method to improve the semantic structure of the text documents. A method has been developed that performs the following: tag the documents for parsing, replacement of idioms with their original meaning, semantic weights calculation for document words and apply semantic grammar. The similarity measure is obtained between the documents and then the documents are clustered using Hierarchical clustering algorithm. The method adopted in this work is evaluated on different data sets with standard performance measures and the effectiveness of the method to develop in meaningful clusters has been proved.
Information Retrieval System is an effective process that helps a user to trace relevant information by Natural Language Processing (NLP). In this research paper, we have presented present an algorithmic Information Retrieval System(BIRS) based on information and the system is significant mathematically and statistically. This paper is demonstrated by two algorithms for finding out the lemmatization of Bengali words such as Trie and Dictionary Based Search by Removing Affix (DBSRA) as well as compared with Edit Distance for the exact lemmatization. We have presented the Bengali Anaphora resolution system using the Hobbs’ algorithm to get the correct expression of information. As the actions of questions answering algorithms, the TF-IDF and Cosine Similarity are developed to find out the accurate answer from the documents. In this study, we have introduced a Bengali Language Toolkit (BLTK) and Bengali Language Expression (BRE) that make the easiest implication of our task. We have also developed Bengali root word’s corpus, synonym word’s corpus, stop word’s corpus and gathered 672 articles from the popular Bengali newspapers ‘The Daily Prothom Alo’ which is our inserted information. For testing this system, we have created 19335 questions from the introduced information and got 97.22% accurate answer.
The Statement of Conjunctive and Disjunctive Queries in Object Oriented Datab...Editor IJCATR
Entrance of object orienting concept in database caused the relation database gradually to replace with object oriented
database in various fields. On the other hand for solving the problem of real world uncertain data, several methods were presented.
One of these methods for modeling database is an approach wich couples object-oriented database modeling with fuzzy logic. Many
queries that users to pose are expressed on the basis of linguistic variables. Because of classical databases are not able to support these
variables, leads to fuzzy approaches are considered. We investigate databases queries in this study both simple and complex ways. In
the complex way, we use conjunctive and disjunctive queries. In the following, we use the XML labels to express inqueries into fuzzy.
We can also communicate with other sections of software by entering into XML world as the most reliable opportunity. Also we want
to correct conjunctive and disjunctive queries related to fuzzy object oriented database using the concept of dependency measure and
weight, and weight be assigned to different phrases of a query based on user emphasis. The other aim of this research is mapping fuzzy
queries to fuzzy-XML. It is expected to be simple implement of query, and output of execution of queries be greatly closer to users'
needs and fulfill her expect. The results show that the proposed method explains the possible conjunctive and disjunctive queries the
database in the form of Fuzzy-XML.
Feature selection, optimization and clustering strategies of text documentsIJECEIAES
Clustering is one of the most researched areas of data mining applications in the contemporary literature. The need for efficient clustering is observed across wide sectors including consumer segmentation, categorization, shared filtering, document management, and indexing. The research of clustering task is to be performed prior to its adaptation in the text environment. Conventional approaches typically emphasized on the quantitative information where the selected features are numbers. Efforts also have been put forward for achieving efficient clustering in the context of categorical information where the selected features can assume nominal values. This manuscript presents an in-depth analysis of challenges of clustering in the text environment. Further, this paper also details prominent models proposed for clustering along with the pros and cons of each model. In addition, it also focuses on various latest developments in the clustering task in the social network and associated environments.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
A template based algorithm for automatic summarization and dialogue managemen...eSAT Journals
Abstract This paper describes an automated approach for extracting significant and useful events from unstructured text. The goal of research is to come out with a methodology which helps in extracting important events such as dates, places, and subjects of interest. It would be also convenient if the methodology helps in presenting the users with a shorter version of the text which contain all non-trivial information. We also discuss implementation of algorithms which exactly does this task, developed by us. Key Words: Cosine Similarity, Information, Natural Language, Summarization, Text Mining
This study investigate the characteristics of a gear system including contact stresses, bending
stresses, and the transmission errors of gears in mesh. Gearing is one of most critical component in
mechanical power transmission systems. The bending stresses in the tooth root were examined using 3D
model.
Current methods of calculating gear contact stresses use Hertz’s equations and Lewis Equation
which were originally derived for contact between two cylinders. To enable the investigation of contact
problems with FEM, the stiffness relationship between the two contact areas is usually established
through a spring placed between the two contacting areas. This can be achieve by inserting a contact
element in between the two areas where contact occurs. The results of the three dimensional FEM analysis
from ANSYS are presented. These stresses were compared with theoretical values. Both results agree very
well. This indicates that the FEM model is accurate.
This report also considers the variations of the whole body stiffness arising from gear body
rotation due to bending deflection, shearing displacement and contact deformation. Many different
positions within the meshing cycle were investigated. Investigation of contact and bending stress
characteristics of spur gears continues to be immense attention to both engineers and researchers in spite
of many studies in the past.
This is because of the advances in the engineering technology that demands for gears with ever
increasing load capacities and speeds with high reliability, the designers need to be able to accurately
predict the stresses experienced the stresses by the loaded gears.
As Diabetes Mellitus combined with other ailments will become a deadly combination, hence there
is an urgent need to break the link between diabetes and its related complications. For this purpose image
processing based analysis can potentially be helpful for earlier detection, education and treatment. Medical
image analysis of Diabetic patients with its related complications such as DR, CVD & Diabetic
Myonecrosis (i.e. on Retinal Images, Coronary angiographs, Electron micrographs, MRI etc) is to be the
aprioristic because of its more prevalence. Thus the main work of this paper is on literature review about
Diabetes and Imaging such as the Prevalence, Classification, Causes and Medical Imaging & Survey of
Image processing methods applied on Diabetic Related Causes.
Keywords — Image, segmentation, retinopathy, Myonecrosis,
For Image Authentication Problem using Encryption Technique and LDPC Source Coding is necessary in Content
Delivery via unsecure medium, Like Peer-To-Peer (P2P) File Sharing. These transferring Digital Files from one Computer to
another. Images are the Most Important Utility of our life. They are used in many applications. There are Two Main Goals of
Image Security: Image Encryption and Authentication. More different encoded versions of the original image available.In
addition, unsecure medium might tamper with the contents.. We propose an efficient, accurate, reliable process using
encryption and LDPC source coding for the image authentication problem. The key idea is to provide a Slepian-Wolf encoded
as authentication data which is encrypted using cryptography key before ready to send. The key used for encryption is usually
independent of the Plain-Image. This can be decoded with side information of an authentic image.
Software keyloggers are a fast growing class of invasive software often used to harvest confidential
information. One of the main reasons for this rapid growth is the possibility for unprivileged programs
running in user space to eavesdrop and record all thekeystrokes typed by the users of a system. The ability
to run in unprivileged mode facilitates their implementation and distribution, but,at the same time, allows
one to understand and model their behavior in detail. Leveraging this characteristic, we propose a new
detection technique that simulates carefully crafted keystroke sequences in input and observes the behavior
of the keylogger in output to unambiguously identify it among all the running processes. We have
prototyped our technique as an unprivileged application, hence matching the same ease of deployment of a
keylogger executing in unprivileged mode. We have successfully evaluated the underlying technique
against the most common free keyloggers. This confirms the viability of our approach in practical
scenarios. We have also devised potential evasion techniques that may be adopted to circumvent our
approach and proposed a heuristic to strengthen the effectiveness of our solution against more elaborated
attacks. Extensive experimental results confirm that our technique is robust to both false positives and
false negatives in realistic settings.
As the rising of transportation system design, so many data logger design was developed for safety. In our country, there are
many accidents on highways. Both driver’s faults and road construction cause accidents. To reduce these condition some safety
system such as obstacle detection system, vehicle declination alarm system, temperature and smoke level display unit,
signboard warning on road sides should be used for both driver and passengers. In addition data logger system for whole
vehicle must be equipped for safety. With the implementation of PIC microcontroller as an Embedded device, this logger
design was constructed with many sensors and C# service-based database. Using Arduino boards, vehicle detection sensing
circuit, Check point radio signal sensing circuit for dangerous road sector, hall-effect magnetic wheel revolution sensing circuit
were designed to be connected with main PIC microcontroller and Personal Computer. Real time result was displayed on C#
Graphical User Interface and Vehicle data log could be easily exported to Microsoft Excel report.
Keywords -Arduino, alarm and alert system,C# service-based database, PC based control system, Vehicle data
logger.
This paper intends to construct and design of the fuel monitoring and electronics control of dispenser for fuel station. It
consists of two sections; hardware implementation for control section and software implementation section. Some fuel stations
which used the flow rate sensors cannot get satisfactions both customers and dealers. The reason is that the fuel volume can
increase or decrease due to the temperature changes depending on the places and times. The system can solve the
disadvantages of it. The system controls for the compensation in volume flow rate with controller. In order to solve this
problem, PIC16F877A is used as a controller to control the time taken for each data representation customer requests from the
keypad and LCD (liquid crystal display) is also used as display unit in this system. And then computer is also used as a control
station by using VB.Net. The status of the fuel levels is also monitored from the computer. MikroC is used for software
implementation of the system In order to control time taken.
Keywords — fuel compensation system, volume flow rate, controller, and software.
This research paper presents a new technique for the synthesis of complex planar mechanisms. The author
calls this technique ‘nomogram-based synthesis’ since it depends on a kinematic nomogram facilitating the
synthesis of complex planar mechanisms without need to complex optimization procedures. A procedure of five
steps is presented to synthesize the mechanism under study for time ratio up to 4.3, normalized stroke up to 3.33.
The nomogram based synthesis can maintain transmission angle to be within a range from 94 to 120 degrees
indicating the effectiveness of the synthesis approach presented.
Keywords — Nomogram-based mechanism synthesis, 6 bar-2 sliders planar mechanism, successful
mechanism performance.
This study aims to enlighten the researchers about the details of process mining. As process mining is a new research area, it includes process modelling and process analysis, as well as business intelligence and data mining. Also it is used as a tool that gives information about procedures. In this paper classification of process mining techniques, different process mining algorithms, challenges and area of application have been explained.Therefore, it was concluded that process mining can be a useful technique with faster results and ability to check conformance and compliance.
In real world applications, most of the optimization problems involve more than one objective to
be optimized. The objectives in most of engineering problems are often conflicting, i.e., maximize
performance, minimize cost, maximize reliability, etc. In the case, one extreme solution would not satisfy
both objective functions and the optimal solution of one objective will not necessary be the best solution
for other objective(s). Therefore different solutions will produce trade-offs between different objectives
and a set of solutions is required to represent the optimal solutions of all objectives. Multi-objective
formulations are realistic models for many complex engineering optimization problems. Customized
genetic algorithms have been demonstrated to be particularly effective to determine excellent solutions to
these problems. A reasonable solution to a multi-objective problem is to investigate a set of solutions, each
of which satisfies the objectives at an acceptable level without being dominated by any other solution. In
this paper, an overview is presented describing various multi objective genetic algorithms developed to
handle different problems with multiple objectives.
To develop an Application to visualize the key board of computer with the concept of image processing. The virtual
keyboard should be accessible and functioning. The keyboard must give input to computer. With the help of camera, image
of keyboard will be fetched. The typing will be captured by camera, as we type on cardboard simply drawn on paper. Camera
will capture finger movement while typing. So basically this is giving the virtual keyboard.
As the technology advances, more and more systems are introduced which will look after the users comfort. Few
years before hard switches were used as keys. Traditional QWERTY keyboards are bulky and offer very little in terms of
enhancements. Now-a-days soft touch keypads are much popular in the market. These keypads give an elegant look and a
better feel. Currently keyboards are static and their interactivity and usability would increase if they were made dynamic and
adaptable. Various on-screen virtual keyboards are available but it is difficult to accommodate full sized keyboard on the
screen as it creates hindrance to see the documents being typed. Virtual Keyboard has no physical appearance. Although
other forms of Virtual Keyboards exist; they provide solutions using specialized devices such as 3D cameras. Due to this, a
practical implementation of such keyboards is not feasible. The Virtual Keyboard that we propose uses only a standard web
camera, with no additional hardware. Thus we see that the new technology always has more Benefits and is more userfriendly.
Recently technological and population development, the usage of vehicles is rapidly increasing and at the
same time the occurrence accident is also increased. Hence, the value of human life is ignored. No one can prevent
the accident, but can save their life by expediting the ambulance to the hospital in time. The objective of this
scheme is to minimize the delay caused by traffic congestion and to provide the smooth flow of emergency
vehicles. The concept of this scheme is to green the traffic signal in the path of ambulance automatically so that
the ambulance can reach the spot in time and human life can be saved. The main server finds the ambulance
through mail. At the same time, it controls the traffic light according to the ambulance location and thus arriving at
the hospital safely. This scheme is fully automated, thus it locates emergency vehicle and controls the traffic
lights, provide the shortest path to reach the hospital in time.
Osteoarthritis (OA) is the most common form of arthritis seen in aged or older populations. It is caused
because of a degeneration of articular cartilage, which functions as shock absorption cushion in knee joint. OA
also leads sliding of bones together, cause swelling, pain, eventually and loss of motion. Nowadays, magnetic
resonance imaging (MRI) technique is widely used in the progression of osteoarthritis diagnosis due to the ability
to display the contrast between bone and cartilage. Usually, analysis of MRI image is done manually by a
physician which is very unpredictable, subjective and time consuming. Hence, there is need to develop automated
system to reduce the processing time. In this paper, a new automatic knee OA detection system based on feature
extraction and artificial neural network is developed. The different features viz GLCM texture, statistical, shape
etc. is extracted by using different image processing algorithms. This detection system consists of 4 stages, which
are pre-processing with ROI cropping, segmentation, feature extraction, and classification by neural network. This
technique results 98.5% of classification accuracy at training stage and 92% at testing stage.
Keywords — Artificial Neural Network (ANN), Gray Level Co-occurrence Matrix (GLCM),Knee
Joint, Magnetic Resonance Imaging (MRI), Osteoarthritis(OA).
rring of useable data, for example a sensor in a room to monitor and control the temperature. It is estimated
that by 2020 there will be about 50 billion internet-enabled devices. The Internet of things presently is being used
in the fields of automobiles, agriculture, security surveillance, building management, smart-homes, and health
care. The IOT expects to use low-cost computing devices where there is less energy consumption impact to the environment[1]. and limited
This paper aims to describe a way for giving security to IT companies, scouting units, business organizations and
volunteer groups. Among the person identification methods, face recognition is known to be the most natural
ones, since the face modality is the modality that uses to identify people in everyday lives. This face detection
differentiates faces from non-faces and is therefore essential for accurate security. The other strategy involves face
recognition for marking the employees. The Raspberry pi module is used for face detection & recognition. The
camera will be connected to the Raspberry pi module. The employees database is collected. The database includes
name of the employees, there images & ID number[2].
This RFID reader module will be installed at the front side of organizations in such a way that all employees
provided by the RFID cards. Raspberry pi system have the database of the employees so comparing the database if
employees details are matched employ can entered into company. Thus with the help of this system, time will be
saved and it is so convenient to record employees. And the details of the employees will be sent to the
corresponding head of organization using IOT technology
The content based Image Retrieval is the restoration of images with respect to the visual appearances
like texture, shape and color.The methods, components and the algorithms adopted in this content based
retrieval of images were commonly derived from the areas like pattern identification, signal progressing
and the computer vision. Moreover the shape and the color features were abstracted in the course of
wavelet transformation and color histogram. Thus the new content based retrieval is proposed in this
research paper.In this paper the algorithms were required to propose with regards to the shape, shade and
texture feature abstraction .The concept of discrete wavelet transform to be implemented in order to
compute the Euclidian distance.The calculation of clusters was made with the help of the modified KMeans
clustering technique. Thus the analysis is made in among the query image and the database
image.The MATLAB software is implemented to execute the queries. The K-Means of abstraction is
proposed by performing fragmentation and grid-means module, feature extraction and K- nearest neighbor
clustering algorithms to construct the content based image retrieval system.Thus the obtained result are
made to compute and compared to all other algorithm for the retrieval of quality image features
Free Space Optics is a medium with high bandwidth which has maximum data rate. Demand for
large data speed capacity has been increasing exponentially due to the massive spread of internet
So with the growing transmission rate and demand in the field of optical communication, the electronic
regeneration has become more expensive. With the introduction of power optical amplifiers the cost of
converting optical signals to electronic
combinations of hybrid amplifiers have been studied and emerged in FSO system. Their performances
have been compared on the basis of transmission distance
Abstract:
E-commerce stands for electronic commerce. E-commerce is doing business online and electronically. This paper attempts to highlight the different challenges faced by e-commerce in India and to understand the essential growth factors required for e-commerce. This paper describes the different services and opportunities offered by E-commerce to business, Producers, Distributers and Customers.
Keywords: - E-commerce, Challenges, Growth Factors.
Generally the prediction of behaviour of material at high temperature is very difficult. During design of
components which are subjected to or working at high temperature must consider the testing at elevated
temperature. Hot tensile testing (HTT) is the method of tensile testing of material at elevated temperature. The
materials used for automotive or aerospace applications are mostly subject to cyclic loading, high temperature
and sometimes involve high frequency vibrations. High strength aluminium alloys are one class of materials that
are widely used in the automotive and aerospace industries .In this work I test A413 material for HTT at different
temperature and strain rate, which can be used for piston.
Keywords — HTT, high temperature, strain rate, piston, automotive or aerospace.
A mediator person is required for communication between deaf person and a second person. But a
mediator should know the sign language used by deaf person. But this is also not possible always since there are
multiple sign languages for multiple languages. It is difficult for a deaf person to understand what a second
person speaks. And therefore deaf person should keep track of lip movements of second person in order to know
what he is speaking. But the lip movements do not give proper efficiency and accuracy since the facial
expressions and speech might not match. To overcome the above problems we have proposed a system, an
Android Application for recognizing sign language using hand gesture with the facility for user to define and
upload their own sign language into the system. The features of this system are the real time conversion of
gesture to text and speech. For two-way communication between deaf person and second person, the speech of
second person is converted into text. The processing steps include: gesture extraction, gesture matching and
conversion of text to speech and vice-versa. The system is not only useful for deaf community but can also be
used by common people who migrate to different regions and do not know local language.
ONTOLOGY-DRIVEN INFORMATION RETRIEVAL FOR HEALTHCARE INFORMATION SYSTEM : A C...IJNSA Journal
In health research, one of the major tasks is to retrieve, and analyze heterogeneous databases containing one single patient’s information gathered from a large volume of data over a long period of time. The main objective of this paper is to represent our ontology-based information retrieval approach for clinical Information System. We have performed a Case Study in the real life hospital settings. The results obtained illustrate the feasibility of the proposed approach which significantly improved the information retrieval process on a large volume of data over a long period of time from August 2011 until January 2012.
Document Retrieval System, a Case StudyIJERA Editor
In this work we have proposed a method for automatic indexing and retrieval. This method will provide as a
result the most likelihood document which is related to the input query. The technique used in this project is
known as singular-value decomposition, in this method a large term by document matrix is analyzed and
decomposed into 100 factors. Documents are represented by 100 item vector of factor weights. On the other
hand queries are represented as pseudo-document vectors, which are formed from weighed combinations of
terms.
Ontology Based Approach for Semantic Information Retrieval SystemIJTET Journal
Abstract—The Information retrieval system is taking an important role in current search engine which performs searching operation based on keywords which results in an enormous amount of data available to the user, from which user cannot figure out the essential and most important information. This limitation may be overcome by a new web architecture known as the semantic web which overcome the limitation of the keyword based search technique called the conceptual or the semantic search technique. Natural language processing technique is mostly implemented in a QA system for asking user’s questions and several steps are also followed for conversion of questions to the query form for retrieving an exact answer. In conceptual search, search engine interprets the meaning of the user’s query and the relation among the concepts that document contains with respect to a particular domain that produces specific answers instead of showing lists of answers. In this paper, we proposed the ontology based semantic information retrieval system and the Jena semantic web framework in which, the user enters an input query which is parsed by Standford Parser then the triplet extraction algorithm is used. For all input queries, the SPARQL query is formed and further, it is fired on the knowledge base (Ontology) which finds appropriate RDF triples in knowledge base and retrieve the relevant information using the Jena framework.
INTELLIGENT INFORMATION RETRIEVAL WITHIN DIGITAL LIBRARY USING DOMAIN ONTOLOGYcscpconf
A digital library is a type of information retrieval (IR) system. The existing information retrieval
methodologies generally have problems on keyword-searching. We proposed a model to solve
the problem by using concept-based approach (ontology) and metadata case base. This model
consists of identifying domain concepts in user’s query and applying expansion to them. The
system aims at contributing to an improved relevance of results retrieved from digital libraries
by proposing a conceptual query expansion for intelligent concept-based retrieval. We need to
import the concept of ontology, making use of its advantage of abundant semantics and
standard concept. Domain specific ontology can be used to improve information retrieval from
traditional level based on keyword to the lay based on knowledge (or concept) and change the
process of retrieval from traditional keyword matching to semantics matching. One approach is
query expansion techniques using domain ontology and the other would be introducing a case
based similarity measure for metadata information retrieval using Case Based Reasoning
(CBR) approach. Results show improvements over classic method, query expansion using
general purpose ontology and a number of other approaches.
A Domain Based Approach to Information Retrieval in Digital Libraries - Rotel...University of Bari (Italy)
The current abundance of electronic documents requires automatic techniques that support the users in understanding their content and extracting useful information. To this aim, improving the retrieval performance must necessarily go beyond simple lexical interpretation of the user queries, and pass through an understanding of their semantic content and aims. It goes without saying that any digital library would take enormous advantage from the availability of effective Information Retrieval techniques to provide to their users. This paper proposes an approach to Information Retrieval based on a correspondence of the domain of discourse between the query and the documents in the repository. Such an association is based on standard general-purpose linguistic resources (WordNet and WordNet Domains) and on a novel similarity assessment technique. Although the work is at a preliminary stage, interesting initial results suggest to go on extending and improving the approach.
MULTILINGUAL INFORMATION RETRIEVAL BASED ON KNOWLEDGE CREATION TECHNIQUESijcseit
As the information access across languages increases, the importance of a system that supports querybased
searching with the presence of multilingual also grows. Gathering the information in different
natural language is the most difficult task, which requires huge resources like database and digital
libraries. Cross language information retrieval (CLIR) enables to search in multilingual document
collections using the native language which can be supported by the different data mining techniques. This
paper deals with various data mining techniques that can be used for solving the problems encountered in
CLIR.
Great model a model for the automatic generation of semantic relations betwee...ijcsity
The
large
a
v
ailable
am
ou
n
t
of
non
-
structured
texts
that
b
e
-
long
to
differe
n
t
domains
su
c
h
as
healthcare
(e.g.
medical
records),
justice
(e.g.
l
a
ws,
declarations),
insurance
(e.g.
declarations),
etc. increases
the
effort
required
for
the
analysis
of
information
in
a
decision making
pro
-
cess.
Differe
n
t
pr
o
jects
and t
o
ols
h
av
e
pro
p
osed
strategies
to
reduce
this
complexi
t
y
b
y
classifying,
summarizing
or
annotating
the
texts.
P
artic
-
ularl
y
,
text
summary
strategies
h
av
e
pr
ov
en
to
b
e
v
ery
useful
to
pr
o
vide
a
compact
view
of
an
original
text.
H
ow
e
v
er,
the
a
v
ailable
strategies
to
generate
these
summaries
do
not
fit
v
ery
w
ell
within
the
domains
that
require
ta
k
e
i
n
to
consideration
the
tem
p
oral
dimension
of
the
text
(e.g.
a
rece
n
t
piece
of
text
in
a
medical
record
is
more
im
p
orta
n
t
than
a
pre
-
vious
one)
and
the
profile
of
the
p
erson
who
requires
the
summary
(e.g
the
medical
s
p
ecialization).
T
o
co
p
e with
these
limitations
this
pa
p
er
prese
n
ts
”GRe
A
T”
a
m
o
del
for
automatic
summary
generation
that
re
-
lies
on
natural
language
pr
o
cessing
and
text
mining
te
c
hniques
to
extract
the
most
rele
v
a
n
t
information
from
narrati
v
e
texts
and
disc
o
v
er
new
in
-
formation
from
the
detection
of
related
information. GRe
A
T
M
o
del
w
as impleme
n
ted
on
sof
tw
are
to
b
e
v
alidated
in
a
health
institution
where
it
has
sh
o
wn
to
b
e
v
ery
useful
to displ
a
y
a
preview
of
the
information
a
b
ou
t
medical
health
records
and
disc
o
v
er
new
facts
and
h
y
p
otheses
within
the
information.
Se
v
eral
tests
w
ere
executed
su
c
h
as
F
unctional
-
i
t
y
,
Usabili
t
y
and
P
erformance
regarding
to
the
impleme
n
ted
sof
t
w
are.
In
addition,
precision
and
recall
measures
w
ere
applied
on
the
results
ob
-
tained
through
the
impleme
n
ted
t
o
ol,
as
w
ell
as
on
the
loss
of
information
obtained
b
y
pr
o
viding
a
text
more
shorter than
the
original
A NOVEL APPROACH OF CLASSIFICATION TECHNIQUES FOR CLIRcscpconf
Recent and continuing advances in online information systems are creating many opportunities
and also new problems in information retrieval. Gathering the information in different natural
language is the most difficult task, which often requires huge resources. Cross-language
information retrieval (CLIR) is the retrieval of information for a query written in the native
language. This paper deals with various classification techniques that can be used for solving
the problems encountered in CLIR.
Data mining , knowledge discovery is the process
of analyzing data from different perspectives and summarizing it
into useful information - information that can be used to increase
revenue, cuts costs, or both. Data mining software is one of a
number of analytical tools for analyzing data. It allows users to
analyze data from many different dimensions or angles, categorize
it, and summarize the relationships identified. Technically, data
mining is the process of finding correlations or patterns among
dozens of fields in large relational databases. The goal of
clustering is to determine the intrinsic grouping in a set of
unlabeled data. But how to decide what constitutes a good
clustering? It can be shown that there is no absolute “best”
criterion which would be independent of the final aim of the
clustering. Consequently, it is the user which must supply this
criterion, in such a way that the result of the clustering will suit
their needs.
For instance, we could be interested in finding
representatives for homogeneous groups (data reduction), in
finding “natural clusters” and describe their unknown properties
(“natural” data types), in finding useful and suitable groupings
(“useful” data classes) or in finding unusual data objects (outlier
detection).Of late, clustering techniques have been applied in the
areas which involve browsing the gathered data or in categorizing
the outcome provided by the search engines for the reply to the
query raised by the users. In this paper, we are providing a
comprehensive survey over the document clustering.
Similar to [IJET-V2I3P19] Authors: Priyanka Sharma (20)
These days we have an increased number of heart diseases including increased risk of heart attacks. Our proposed system users sensors that allow to detect heart rate of a person using heartbeat sensing even if the person is at home. The sensor is then interfaced to a microcontroller that allows checking heart rate readings and transmitting them over internet. The user may set the high as well as low levels of heart beat limit. After setting these limits, the system starts monitoring and as soon as patient heart beat goes above a certain limit, the system sends an alert to the controller which then transmits this over the internet and alerts the doctors as well as concerned users. Also the system alerts for lower heartbeats. Whenever the user logs on for monitoring, the system also displays the live heart rate of the patient. Thus concerned ones may monitor heart rate as well get an alert of heart attack to the patient immediately from anywhere and the person can be saved on time.This value will continue to grow if no proper solution is found. Internet of Things (IoT) technology developments allows humans to control a variety of high-tech equipment in our daily lives. One of these is the ease of checking health using gadgets, either a phone, tablet or laptop. we mainly focused on the safety measures for both driver and vehicle by using three types of sensors: Heartbeat sensor, Traffic light sensor and Level sensor. Heartbeat sensor is used to monitor heartbeat rate of the driver constantly and prevents from the accidents by controlling through IOT.
ABSTRACT The success of the cloud computing paradigm is due to its on-demand, self-service, and pay-by-use nature. Public key encryption with keyword search applies only to the certain circumstances that keyword cipher text can only be retrieved by a specific user and only supports single-keyword matching. In the existing searchable encryption schemes, either the communication mode is one-to-one, or only single-keyword search is supported. This paper proposes a searchable encryption that is based on attributes and supports multi-keyword search. Searchable encryption is a primitive, which not only protects data privacy of data owners but also enables data users to search over the encrypted data. Most existing searchable encryption schemes are in the single-user setting. There are only few schemes in the multiple data users setting, i.e., encrypted data sharing. Among these schemes, most of the early techniques depend on a trusted third party with interactive search protocols or need cumbersome key management. To remedy the defects, the most recent approaches borrow ideas from attribute-based encryption to enable attribute-based keyword search (ABKS
Cloud computing is the one of the emerging techniques to process the big data. Large collection of set or large
volume of data is known as big data. Processing of big data (MRI images and DICOM images) normally takes
more time compare with other data. The main tasks such as handling big data can be solved by using the concepts
of hadoop. Enhancing the hadoop concept it will help the user to process the large set of images or data. The
Advanced Hadoop Distributed File System (AHDF) and MapReduce are the two default main functions which
are used to enhance hadoop. HDF method is a hadoop file storing system, which is used for storing and retrieving
the data. MapReduce is the combinations of two functions namely maps and reduce. Map is the process of
splitting the inputs and reduce is the process of integrating the output of map’s input. Recently, in medical fields
the experienced problems like machine failure and fault tolerance while processing the result for the scanned
data. A unique optimized time scheduling algorithm, called Advanced Dynamic Handover Reduce Function
(ADHRF) algorithm is introduced in the reduce function. Enhancement of hadoop and cloud introduction of
ADHRF helps to overcome the processing risks, to get optimized result with less waiting time and reduction in
error percentage of the output image
Text mining has turned out to be one of the in vogue handle that has been joined in a few research
fields, for example, computational etymology, Information Retrieval (IR) and data mining. Natural
Language Processing (NLP) methods were utilized to extricate learning from the textual text that is
composed by people. Text mining peruses an unstructured form of data to give important
information designs in a most brief day and age. Long range interpersonal communication locales
are an awesome wellspring of correspondence as the vast majority of the general population in this
day and age utilize these destinations in their everyday lives to keep associated with each other. It
turns into a typical practice to not compose a sentence with remedy punctuation and spelling. This
training may prompt various types of ambiguities like lexical, syntactic, and semantic and because of
this kind of indistinct data; it is elusive out the genuine data arrange. As needs be, we are directing
an examination with the point of searching for various text mining techniques to get different
textual requests via web-based networking media sites. This review expects to depict how
contemplates in online networking have utilized text investigation and text mining methods to
identify the key topics in the data. This study concentrated on examining the text mining
contemplates identified with Facebook and Twitter; the two prevailing web-based social networking
on the planet. Aftereffects of this overview can fill in as the baselines for future text mining research.
Colorectal cancer (CRC) has potential to spread within the peritoneal cavity, and this transcoelomic
dissemination is termed “peritoneal metastases” (PM).The aim of this article was to summarise the current
evidence regarding CRC patients at high risk of PM. Colorectal cancer is the second most common cause of cancer
death in the UK. Prompt investigation of suspicious symptoms is important, but there is increasing evidence that
screening for the disease can produce significant reductions in mortality.High quality surgery is of paramount
importance in achieving good outcomes, particularly in rectal cancer, but adjuvant radiotherapy and chemotherapy
have important parts to play. The treatment of advanced disease is still essentially palliative, although surgery for
limited hepatic metastases may be curative in a small proportion of patients.
Heat transfer in pipes is a distinctive kind of procedure employed in heat exchanger which transfers great
deal of heat because of the impact of capillary action and phase change heat transfer principle. Late improvement
in the heat pipe incorporates high thermal conductivity liquids like Nano liquids, fixed inside to extricate the most
extreme heat. This paper audits, impact of different factors, for example, thermal pipe tilt edge, charged measure
of working liquid, nano particles sort, size, and mass/volume part and its impact on the change of thermal
proficiency, thermal exchange limit and decrease in thermal protection. The Nano liquid arrangement and the
examination of its thermal attributes likewise have been investigated. The retained sun oriented vitality is
exchanged to the working liquid streaming in the pipe. The execution of the framework is affected by thermal
exchange from tube to working liquid, with least convective misfortunes, which must be considered as one of the
essential plan factor. In tube and channel streams, to improve the rate of heat exchange to the working liquid,
detached enlargement methods, for example, contorted tapes and swirl generators are employed from the fluid
flow path. The variation of heat transfer coefficient and pressure drop in the pipe flow for water and water based
Al2O3 Nano fluids at different volume concentrations and twisted tapes are studied.
Now-a-day’s pedal powered grinding machine is used only for grinding purpose. Also, it requires lots of efforts
and limited for single application use. Another problem in existing model is that it consumed more time and also has
lower efficiency. Our aim is to design a human powered grinding machine which can also be used for many purposes
like pumping, grinding, washing, cutting, etc. it can carry water to a height 8 meter and produces 4 ampere of electricity
in most effective way. The system is also useful for the health conscious work out purpose. The purpose of this technical
study is to increase the performance and output capacity of pedal powered grinding machine.
This project proposes a distributed control approach to coordinate multiple energy storage units
(ESUs) to avoid violation of voltage and network load constraints ESU as a buffer can be a promising
solution which can store surplus power during the peak generation periods and use it in peak load
periods.In ESU converters both active and reactive power are used to deal with the power quality
issues in distribution network ESU’s reactive power is proposed to be used for voltage support, while
the active power is to be utilized in managing network loading.
The steady increase in non-linear loads on the power supply network such as, AC variable speed drives,
DC variable Speed drives, UPS, Inverter and SMPS raises issues about power quality and reliability. In this
subject, attention has been focused on harmonics . Harmonics overload the power system network and cause
reliability problems on equipment and system and also waste energy. Passive and active harmonic filters are
used to mitigate harmonic problems. The use of both active and passive filter is justified to mitigate the
harmonics. The difficulty for practicing engineers is to select and deploy correct harmonic filters , This paper
explains which solutions are suitable when it comes to choosing active and passive harmonic filters and also
explains the mistakes need to be avoided.
This Paper is aimed at analyzing the few important Power System equipment failures generally
occurring in the Industrial Power Distribution system. Many such general problems if not resolved it may
lead to huge production stoppage and unforeseen equipment damages. We can improve the reliability of
Power system by simply applying the problem solving tool for every case study and finding out the root cause
of the problem, validation of root cause and elimination by corrective measures. This problem solving
approach to be practiced by every day to improve the power system reliability. This paper will throw the light
and will be a guide for the Practicing Electrical Engineers to find out the solution for every problem which
they come across in their day to day maintenance activity.
More from IJET - International Journal of Engineering and Techniques (20)
Final project report on grocery store management system..pdfKamal Acharya
In today’s fast-changing business environment, it’s extremely important to be able to respond to client needs in the most effective and timely manner. If your customers wish to see your business online and have instant access to your products or services.
Online Grocery Store is an e-commerce website, which retails various grocery products. This project allows viewing various products available enables registered users to purchase desired products instantly using Paytm, UPI payment processor (Instant Pay) and also can place order by using Cash on Delivery (Pay Later) option. This project provides an easy access to Administrators and Managers to view orders placed using Pay Later and Instant Pay options.
In order to develop an e-commerce website, a number of Technologies must be studied and understood. These include multi-tiered architecture, server and client-side scripting techniques, implementation technologies, programming language (such as PHP, HTML, CSS, JavaScript) and MySQL relational databases. This is a project with the objective to develop a basic website where a consumer is provided with a shopping cart website and also to know about the technologies used to develop such a website.
This document will discuss each of the underlying technologies to create and implement an e- commerce website.
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...Amil Baba Dawood bangali
Contact with Dawood Bhai Just call on +92322-6382012 and we'll help you. We'll solve all your problems within 12 to 24 hours and with 101% guarantee and with astrology systematic. If you want to take any personal or professional advice then also you can call us on +92322-6382012 , ONLINE LOVE PROBLEM & Other all types of Daily Life Problem's.Then CALL or WHATSAPP us on +92322-6382012 and Get all these problems solutions here by Amil Baba DAWOOD BANGALI
#vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore#blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #blackmagicforlove #blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #Amilbabainuk #amilbabainspain #amilbabaindubai #Amilbabainnorway #amilbabainkrachi #amilbabainlahore #amilbabaingujranwalan #amilbabainislamabad
Hierarchical Digital Twin of a Naval Power SystemKerry Sado
A hierarchical digital twin of a Naval DC power system has been developed and experimentally verified. Similar to other state-of-the-art digital twins, this technology creates a digital replica of the physical system executed in real-time or faster, which can modify hardware controls. However, its advantage stems from distributing computational efforts by utilizing a hierarchical structure composed of lower-level digital twin blocks and a higher-level system digital twin. Each digital twin block is associated with a physical subsystem of the hardware and communicates with a singular system digital twin, which creates a system-level response. By extracting information from each level of the hierarchy, power system controls of the hardware were reconfigured autonomously. This hierarchical digital twin development offers several advantages over other digital twins, particularly in the field of naval power systems. The hierarchical structure allows for greater computational efficiency and scalability while the ability to autonomously reconfigure hardware controls offers increased flexibility and responsiveness. The hierarchical decomposition and models utilized were well aligned with the physical twin, as indicated by the maximum deviations between the developed digital twin hierarchy and the hardware.
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Dr.Costas Sachpazis
Terzaghi's soil bearing capacity theory, developed by Karl Terzaghi, is a fundamental principle in geotechnical engineering used to determine the bearing capacity of shallow foundations. This theory provides a method to calculate the ultimate bearing capacity of soil, which is the maximum load per unit area that the soil can support without undergoing shear failure. The Calculation HTML Code included.
CFD Simulation of By-pass Flow in a HRSG module by R&R Consult.pptxR&R Consult
CFD analysis is incredibly effective at solving mysteries and improving the performance of complex systems!
Here's a great example: At a large natural gas-fired power plant, where they use waste heat to generate steam and energy, they were puzzled that their boiler wasn't producing as much steam as expected.
R&R and Tetra Engineering Group Inc. were asked to solve the issue with reduced steam production.
An inspection had shown that a significant amount of hot flue gas was bypassing the boiler tubes, where the heat was supposed to be transferred.
R&R Consult conducted a CFD analysis, which revealed that 6.3% of the flue gas was bypassing the boiler tubes without transferring heat. The analysis also showed that the flue gas was instead being directed along the sides of the boiler and between the modules that were supposed to capture the heat. This was the cause of the reduced performance.
Based on our results, Tetra Engineering installed covering plates to reduce the bypass flow. This improved the boiler's performance and increased electricity production.
It is always satisfying when we can help solve complex challenges like this. Do your systems also need a check-up or optimization? Give us a call!
Work done in cooperation with James Malloy and David Moelling from Tetra Engineering.
More examples of our work https://www.r-r-consult.dk/en/cases-en/
Student information management system project report ii.pdfKamal Acharya
Our project explains about the student management. This project mainly explains the various actions related to student details. This project shows some ease in adding, editing and deleting the student details. It also provides a less time consuming process for viewing, adding, editing and deleting the marks of the students.
Student information management system project report ii.pdf
[IJET-V2I3P19] Authors: Priyanka Sharma
1. International Journal of Engineering and Techniques - Volume 2 Issue 3, May – June 2016
ISSN: 2395-1303 http://www.ijetjournal.org Page 118
Retrieval of Information Using Fuzzy Queries
Priyanka Sharma1
1(CSE, Advanced Institute of Technology and Management, Palwal)
I. INTRODUCTION
So far we have been implementing the classical
or simple queries on classical database but here we
are actually implementing the fuzzy queries on the
classical because sometimes we are likely to use
vague terms in our query. So our implementation in
this paper is imparting the users an efficiency to
make use of data using natural language. Classical
SQL queries have outstanding capabilities in terms
of data retrieval and gathering of answer from
information stored at completely dispersed
databases.
Human queries are rarely precise which poses
challenges in efficient answer formation and data
retrieval. These are based on human perception
which is most of the time unclear and imprecise
depending on world knowledge. The primary
concentration of fuzzy logic is always on human
language where reasoning with imprecise or inexact
proposition is much more difficult, not user friendly.
As it is known that a database management system
includes in itself a catalogue of data which is
related to one another in some form and a collection
of programs for the purpose of accessing that
Interrelated data. And searching in a database is
an important operation among other operations in
database management system. As the size of
database increases, the searching time also
increases. So numbers of algorithms have been
developed to improve the performance but these are
only for the classical database [4]. This research
aims to integrate query language with fuzzy logic in
order to improve the data retrieval. Information
retrieval is seen as the very active or progressive
area while using fuzzy set theory. The fuzzy
information retrieval makes use of the fuzzy sets to
depict the documents and the level of membership
to show the aptness of terms and vague
compatibility measures with the aim of having
status value for the retrieved document. This system
gives references to documents on the basis of two
main components and these two components are the
document representation and the use query. Here
the document representation is applied to those
terms which are considered as the atomic part of the
documents. And user query is explained using any
language for query which is depending on these
RESEARCH ARTICLE OPEN ACCESS
Abstract:
The classical or traditional information system provides answer after a user submits a complete query. It is even
noticed that presently, almost all the relational database systems rely on the query which has syntax and semantics
defined completely to access data. But often it is the case that we are willing to use vague terms in our query. The main
objective of database management system is to provide an environment that is both convenient and efficient for people
to use in storing and retrieving information. A recent trend of supporting auto complete is a first step to cope up with
this problem. We can have design of both classical and fuzzy database and can use effectively fuzzy queries on these
databases. Fuzzy databases are developed to manipulate the incomplete, unclear and vague data such as low, fast, very
high, about etc. The primary focus of fuzzy logic is on the natural language. This Paper provides the users the flexibility
or freedom to query database using natural language. Here this paper implements “interactive fuzzy search”. This
framework for interactive fuzzy search permits the user to explore the data as they type even in the presence of some
minor errors. This paper applies fuzzy queries on relational database so that it is possible to have the precise result as
well as the output for the uncertain terms we generally use based on some membership function
Keywords — vague, fuzzy logic, interactive search, membership function, fuzzy variables
2. International Journal of Engineering and Techniques - Volume 2 Issue 3, May – June 2016
ISSN: 2395-1303 http://www.ijetjournal.org Page 119
terms and also permits the concatenation of the
customer needs with the AND, OR, and NOT
logical operator.
In the Boolean information retrieval model, a term
which is appearing in a document is either relevant
or it is irrelevant. Information retrieval is shown by
inexactness in this model [3].
The main disadvantage of this model is that it
cannot consider a semi degree of relevance. And the
method that also combines the elements depending
on their contents, it is assumed that document
depiction is imprecise. So the users who are having
any knowledge regarding data would like to use
their knowledge of importance and the relevance of
terms in any particular document. The value of
membership function for any object in a non fuzzy
set is considered to be 0 or 1 where 0 value
specifies that the object is not the part of set and the
value 1 indicates that the object is fully belonging
to the set. A fuzzy set is considered as the
generalization of the traditional set having the range
of [0, 1].
It is the set to which an object can belong to some
degree of membership. And it is assumed that if
more objects are part of this set then the degree of
membership will be higher. A fuzzy set F is taken
as the subset of the union of discourse T{t1,
t2,.....................tn}. It is specified by a membership
function µ. And range always lies between [0, 1].
Where µF(ti) is showing the membership degree of
universe of discourse belonging to the fuzzy set F.
II. LITERATURE SURVEY
A far-reaching literature survey has been done to
arrive at this approach and this literature survey has
been very beneficial from the all the aspects
considered. This extensive literature survey belong
to the below mentioned research papers.
These research papers were highly beneficial and
we have taken many approaches into account and
derived this idea. The literature review is presented
as follows:
In the paper of Juan Carlos Santamarina and Jjean-
lou Chameau, it is presented that the fuzziness
which is present in comprehension is most of the
time is given or shown in the outputs; the terms like
as complexity, inexactness , and the structure of
membership functions are purely related to each
other in one or other way. And also the changes or
variations between crisp extremes through the
vagueness have characteristic manifestations. He
has presented the emphasis on the practical
implications of these ideas. Most of the applications
are including developing standard and type h fuzzy
sets, which are helpful in increasing the exactness
and consistency of presented data, and also having
the support for preparation of knowledge-based
systems which are using fuzzy sets for their proper
functioning.
In the paper of Abhijeet R Raipurkar,
G.R.Bamnote on fuzzy logic based query
optimization in distributed database, this paper it is
given that the query optimization is a complicated
task in a distributed environment having both
client/server environment because the location of
the data becomes a major factor. So by integrating
query processing subsystem with a distributed
database management system and also having fuzzy
logic with it is used for the estimation of response
time of the query, this paper concludes that the by
using the fuzzy logic based method for query
optimization in distributed database have very
significant effect on the efficiency of query
processing of distributed database.
In the paper of Miroslav Hudec on “approach to
fuzzy database querying, analysis and realisation “it
is presented that querying fuzzy variables is not
only a simple tool for query; it also enhances the
main idea of a query and also retrieves the extra
priceless information along with it. Fuzzy method
has also few shortcomings in itself that are clearly
seen while querying a process. This paper is
basically presenting these shortcomings of the
fuzzy method and also the concepts or ideas which
can be used for resolving those limitations.
In the paper of Kevin Woods, Diane Cook,
Lawrence Hall on “learning membership functions
in a function-based object recognition system they
have presented the learning component for GRUFF
system called OMLET that has the ability to learn
the membership by itself from the number of
objects which are also named with their category
measures.
3. International Journal of Engineering and Techniques - Volume 2 Issue 3, May – June 2016
ISSN: 2395-1303 http://www.ijetjournal.org Page 120
In the paper of Sujit Kumar Mondal, Joyassree Sen,
md. Rabiul Islam, md. Shamim Hossian on
“performance comparison of fuzzy queries on fuzzy
database and classical database” it is given that the
users have the efficiency and flexibility to interact
with the database by using human language. The
time and cost of both the classical and fuzzy query
over traditional and fuzzy database (db) is also
calculated in the paper and is showing the decrease
in time cost while fuzzy querying on fuzzy
database as compared to classical database.
III. BENEFITS OF FUZZY QUERIES
While putting one’s thought as query in terms of
natural languages into database, a lot of problems
are experienced due to inefficiency of relational
database to handle such query. As it is not possible
for computers to use natural language, there is
number of ways which are having their own
advantages and disadvantages. A fuzzy querying
interface as meant here makes it possible for a user
to employ his or her own dictionary of linguistic
terms to be used in queries and these linguistic terms
provide for a direct representation of presumed
vagueness and imprecision of query. It provides the
better representation for the user requirement by
expressing imprecise condition through linguistic
terms. The terms in fuzzy queries include certain
vagueness or uncertainties that information system
operated on two valued logic usually do not
understand and this is the reason they cannot use
these terms. For example: A customer of real estate
agency looking for a house would rather express his
or her own criteria using imprecise description as
cheap, large garden only. And this possibility to use
fuzzy linguistic term should be attractive not only
for a non expert casual user. Similarly the query
“give me the name of young people who are
suffering from severe cancer” cannot be executed or
processed directly by SQL since it contains the lot
vagueness as young people and severe cancer. The
purpose of using linguistic variable is to provide a
means of approximate characterization of
phenomenon that is not properly defined. Fuzzy
logic also reduces the complication of mathematical
analysis of difficulties in comparison to traditional
queries or precise query. Whenever classical queries
on classical database are applied then a kind of lack
of expressiveness is seen because the real world data
is not always very clear but simply has
imperfections or vagueness. But on the other side
the implementation of fuzzy queries on classical
database leads to the more interactivity of human
expression without having any effect on searching
time. But on further proceedings it is seen that we
move further we see that when we are implementing
the fuzzy queries on fuzzy database then the cost
here is less than the fuzzy queries on classical
database because here fuzzy value is calculated
during the building of fuzzy database so extra time
to compute fuzzy value is not required here. Next
this implementation gives the user the flexibility to
query database using natural language. The lack of
expressiveness can be entirely eliminated and
searching time can be reduced by fuzzy database
management system and using fuzzy queries on it.
This implementation provides an easy to use
functionality for the purpose of extracting or
retrieving data similar to conventional structured
query.
In the traditional information systems user is usually
unaware about the data specially whenever there is
the case that they are having less knowledge about
the underlying data and is forced to make use of the
traditional try and see approach to extract the
information so we can say that this paper is
presenting an interactive fuzzy search. The vast
expansion of internet gives rise to a significant
growth of a large number of databases. This
framework has made the user to explore the data as
they type even in the presence of some small errors.
Frequently most of the relational database systems
use query which are having all syntax and semantics
which have been defined previously to retrieve data
but sometimes we are likely to use uncertain terms
in our query. Secondly if we use distributed
database it permits much faster execution of the
local queries and also make it possible to have the
reduction in the traffic. But there is the main issue of
maintaining data integrity. This method of fuzzy set
theory with its membership function is widely used
and is popular to form realistic explanations of
evaluation. So this implementation provides the
method to capture the uncertainty situation of the
real world data.
IV. PROPOSED WORK
4. International Journal of Engineering and Techniques - Volume 2 Issue 3, May – June 2016
ISSN: 2395-1303 http://www.ijetjournal.org Page 121
This paper is implementing the fuzzy query on
relational database to have information that is the
most relevant on some fuzzy basis. So for this
purpose first of all we have to create a database and
that can be created in any language either in SQL or
in oracle. Then we have to create a search interface
and again that can be in any language. And this
interface must be created in such a way that it
should connect with the database we have made in
SQL or in oracle and when we fire any query on this
interface, it should retrieve the result from the
database made in SQL or oracle.
Connection to For fields
Database
Define
Ranges
For terms
Membership
is
Calculated
Execute
Queries
With
Membership
Figure1: Architecture of information retrieval using fuzzy query
So here it is started with the database that has been
created in SQL language and after that a search
interface is created in java that is all coding will be
done in java. Then this all coding will be used in
eclipse software which has some inbuilt functions.
This software makes work easier.
Then the coding will connect to our database of
SQL and fired query will retrieve the results from
that database in SQL.
A. CALCULATION OF MEMBERSHIP OF
LINGUISTIC VARIABLES
The membership of the linguistic variables is
evaluated by a function known as “Membership
Function”. The membership function is denoted by
π.
If there are three linguistic variables Low, Moderate
and High and their range is between 10,000 to
25,000, 20,000 to 100000 and 70,000 to 130,000
respectively.
The membership for variable Low is given as:
µ (Low) =
It means that the membership of the employee for
salary low will be 0 when the salary is less than
10,000 or it is greater than 25,000. The membership
of the employee for salary low will be 1 when the
salary is greater than or equals to 10,000 and it is
less than or equals to 20,000.
And membership of the employee for salary low
will be (25,000 – salary)/ (25,000 – 20,000).
Similarly the membership for variable Moderate is
calculated based on the below method.
µ (Moderate) =
It means that the membership of the employee for
salary moderate will be 0 when the salary is less
than 20,000 or it is greater than 100000. The
membership of the employee for salary moderate
will be 1 when the salary is greater than or equals to
25,000 and it is less than or equals to 70,000.And
Create
database
Define
linguistic
terms
Define
membership
function
Extract
results from
the database
Create a search
interface
5. International Journal of Engineering and Techniques - Volume 2 Issue 3, May – June 2016
ISSN: 2395-1303 http://www.ijetjournal.org Page 122
membership of the employee for salary moderate
will be (salary – 20,000)/(25,000 – 20,000).
And membership of the employee for salary
moderate will be (100000-salary) / (100000 –
70,000).
Now the membership for variable High is calculated
as given below. The membership for variable high is
as shown:
µ (High) =
It means that the membership of the employee for
salary high will be 0 when the salary is less than
70,000 or it is greater than 130000. The membership
of the employee for salary high will be 1 when the
salary is greater than or equals to 100000 and it is
less than or equals to 130,000.And membership of
the employee for salary moderate will be
( salary – 70,000) / (100000 – 70,000).
V. CONCLUSION
This Paper has provided an overview of retrieval of
relevant information using fuzzy queries and then
discussed some approaches of applying fuzzy
queries to the database. By trying different
techniques and algorithms or by comparing
different techniques or by combining the algorithms,
data analyst gives the best result. One approach is
to apply fuzzy queries on fuzzy database and other
one is to apply fuzzy queries on classical database.
Here we applied fuzzy queries on relational
database and we find some meaningful results after
applying queries. After applying these fuzzy queries
on the database, it is concluded that it is extremely
helpful in reducing the lack of expressiveness in the
queries and in handling the real world problems
related with the inexact or vague data. It has
provided us with a flexible system to deal with the
most common human queries.
VI. FUTURE ENHANCEMENT
Handling of vague or inexact data is very difficult
but it is an important thing to do and this retrieving
information using fuzzy query has helped a lot to
solve this problem. In future, fuzzy querying
techniques need to be integrated with traditional
techniques with the help of disciplines like machine
learning, operation research, statistics and
simulation. The main goal is to hide the complexity
of real world data from end user and visualize the
data which is easily read and understand by any
novice user.
REFERENCES
1. George, J. Klir, Bo Yuan, “Fuzzy Sets and
Fuzzy Logic Theory and Applications”Prentice
Hall of India, 1997.
2. C. T. Yu, K. C. Guh, W. Zhang, M. Templeton,
D. Brill, and A. L. P. Chen, "An Integrated
Algorithm for Distributed Query Processing,"
in IFIP Conference on Distributed
processingAmsterdam, 1987.
3. D. Chiang, L.R. Chow and N. Hsien, “Fuzzy
information in extended fuzzy relational
databases”, Fuzzy Sets and Systems 92, pp.1-
10. (1997).
4. Sujit kumar Mondal,Joyassree Sen, Md. Rabiul
Islam, Md. Shamim Hossian,” performance
comparison of fuzzy queries on fuzzy database
and classical database” in International Journal
of Computer Application Issue 3, Volume 1
(February 2013)
5. Abhijeet Raipurkar, G.R. Bamnote, ”Fuzzy
Logic Based Query Optimization in
Distributed Database” in International Journal
of Innovative Research in Computer and
Communication Engineering Vol. 1, Issue 2,
April 2013.
6. J. Kacprzyk and S. Zadrozny, “On
summarization of large datasets via a fuzzy-
logic- based querying add-on to Microsoft
Access. In: Intelligent Information Systems VII,
Malbork, IPI PAN, Warsaw 1998, pp. 249-258.