SlideShare a Scribd company logo
International Journal of Research in Advent Technology, Vol.2, No.8, August 2014 
E-ISSN: 2321-9637 
Question Classification: Using Support Vector Machine 
77 
and Lexical, Semantic and Sytactic Features 
Kiran Yadav, Megha Mishra 
M.E scholar sscet Bhilai, Prof. sscet Bhilai 
Yadavkiran64@gmail.com 
Abstract Question classification is play important role in the question answering system. The results of the 
question classification find out the quality of the question answering system. In this paper, a question 
classification algorithm based on SVM and feature, Support Vector Machine model is take to train a classifier 
on coarse categories, there features also use for classify the category. SVM has been used for question 
classification and have a good results. We use SVM as the classifier. The experiment results show that the 
feature extraction can perform well with SVM and our approach can reach classification accuracy. 
Index Terms- 
Question answering, text classification, machine learning, support vector machine. 
1. INTRODUCTION 
In this work, we use a machine learning approach to 
question classification. Task of question classification 
as a supervised learning classification. In order to 
prepare the learning model, we designed a deep 
position of features that are prognostic of question 
categories . 
In this paper work this classification has two 
purposes. It provides constraints on the answer types 
that provide foster processing to just site and verify 
the answer. Which city has the largest population? we 
do not want to test each phrase in a document to look 
that it gives an answer . 
However, there characteristics of question 
classification that mark it from the common work. On 
one hand, questions are relatively short and contain 
less word-based information equate with classifying 
the entire text. On the other hand, small questions are 
amenable for more correct and deeper-level In this 
way, this work on question classification can be also 
see as a case study is take semantic information to text 
classification. Similar to syntactic information such as 
part-of-speech tags, clear notion of how to use lexical 
semantic information is to replace or augment each 
word by its semantic class in the given context, then 
generate a feature-based representation and learn a 
mapping from this representation to the desired 
property. This general scheme leaves several issues 
open that make the analogy to syntactic categories 
nontrivial. 
First, it is not open which semantic category 
is allow and how to develop them. Second, it is not 
open how to hold the more dissimilar problem of 
semantic when decisied the delegacy of a sentence. 
Merge these three features and increase the accuracy 
of the question classification by using these features. 
Question classification plays an important role in 
question answering. Features are the key to obtain an 
accurate question classifier. 
Question answering systems deal defferent it this 
problem, by giving natural language de in which users 
can explain their information required form of a 
natural language question. Retrieve the exact answer 
to that very same question in place of a set of 
documents. natural language, from a (typically large) 
collection of documents, such as the WWW. 
The developing period of the q/a system in different 
field is too long and recycle rate is so low. Developed 
a state of the art machine leaning based question 
classifier that use a rich a set of lexical, syntactic and 
semantic features. 
2. QUESTION CLASSIFICATION 
Question Classification means it helps for give the 
result of given question .It is mainly use for the 
question answer system. It work category wise 
example if any type of question it there and find the 
answer in category it give fast result. When we search 
any thing it search engine like google then it gives all 
things which are related to that word which is in 
search. But it gives the answer in category wise. 
because of only the question`s answer is presented. 
Table 1. The coarse question categories 
Coarse 
ABBR 
DESC 
ENTY 
HUM 
LOC 
NUM
International Journal of Research in Advent Technology, Vol.2, No.8, August 2014 
E-ISSN: 2321-9637 
78 
To simplify the following experiments, we assume 
that one Question resides in only one category. That is 
to say, unambiguous question is labeled with its most 
probable category. 
2.1 Question types 
What is the fastest fish in the world? 
What’s the colored part of the eye called? 
What color is Mr. Spock’s blood? 
Name a novel written by John Steinbeck. 
What currency is used in Australia? 
What is the fear of cockroaches called? 
What are the historical trials following WorldWar II 
called? 
What is the world ’s best selling cookie? 
What instrument is Ray Charles best known for 
playing? 
What language is mostly spoken in Brazil? 
What letter adorns the flag of Rwanda? 
What’s the highest hand in straight poker? 
What is the state tree of Nebraska? 
What is the best brand for a laptop computer? 
What religion has the most members? 
What game is Garry Kasparov really good at? 
3. RELATED WORK 
Hand-made Rule-based show on extracting names 
using many of human-made rules set. basically the 
systems consist of a set of patterns using grammatical 
(e.g. part of speech), syntactic (e.g. word precedence) 
and orthographic features (e.g. capitalization) in 
combination with dictionaries An example for this 
type of system is: "President rao said bankers talks 
will make discussions on private, U.S. forces to leave 
Iraq". In this example a proper noun follows a 
person's title(president), then noun is a person's name 
and proper noun that is started with capital character 
(Iraq) after the verb (to leave) is a Location's name. 
In this family of approaches, Appelt , propose a name 
identification system based on carefully handcrafted 
regular expression called FASTUS. They divided the 
task into three steps: Recognizing Phrases, 
Recognizing Patterns and Merging incidents These 
approaches are relying on manually coded rules and 
manually compiled corpora. These kinds of models 
have better results for restricted domains, are capable 
of detecting complex entities that learning models 
have difficulty with. However, the rule-based NE 
systems lack the ability of portability robustness, and 
furthermore the high cost of the rule maintains 
increases even though the data is slightly changed. 
These type of approaches are often domain and 
language specific and do 
not necessarily adapt well to new domains and 
languages. 
In Machine Learning-based NER system, 
the purpose of Named Entity Recognition approach is 
converting identification problem into a classification 
problem and employs a classification statistical model 
to solve it. In this type of approach, the systems look 
for patterns and relationships into text to make a 
model using statistical models and machine learning 
algorithms. The systems identify and classify nouns 
into particular classes such as persons, locations, 
times, etc base on this model, using machine learning 
algorithms. There are two types of machine learning 
model that are use for NER. Supervised and 
Unsupervised machine learning model. Supervised 
learning involves using a program that can learn to 
classify a given set of labeled examples that are made 
up of the same number of features. 
Each example is thus represented with 
respect to the different feature spaces. The learning 
process is called supervised, because the people who 
marked up the training examples are teaching the 
program the right distinctions. The supervised 
learning approach requires preparing labeled training 
data to construct a statistical model, but it cannot 
achieve a good performance without a large amount 
of training data, because of data sparseness problem. 
In recent years several statistical methods based on 
supervised learning method were proposed. Bikel et. 
al. propose a learning name-finder base on hidden 
Markov model [8] called Nymbel, while Borthwick 
et. al. investigates exploiting diverse knowledge 
sources via maximum entropy in named entity 
recognition [9,10]. A tagging of unknown proper 
names system with Decision Tree model was 
proposed by Bechet et. al. [5], while Wuet. al. 
presented a named entity recognition system based on 
support vector machines [2]. Unsupervised learning 
method is another type of machine learning model, 
where an unsupervised model learns without any 
feedback. In unsupervised learning, the goal of the 
program is to build representations from data. These 
representations can then be used for data compression, 
classifying, decision making, and other purposes. 
Unsupervised learning is not a very popular approach 
for NER and the systems that do use unsupervised 
learning are usually not completely unsupervised. In 
these types of approach, Collins et. al. discusses an 
unsupervised model for named entity classification by 
use of unlabeled examples of data [7], 
Koimetal. Proposes an unsupervised named entity 
classification models and their ensembles that uses a 
small-scale named entity dictionary and an unlabeled 
corpus for classifying named entities [4]. Unlike the 
rulebased method, these types of approaches can be 
easily port to different domain or languages. In 
Hybrid NER system, the approach is to combine 
rulebased and machine learning-based methods, and 
make new methods using strongest points from each 
method.
International Journal of Research in Advent Technology, Vol.2, No.8, August 2014 
E-ISSN: 2321-9637 
79 
4. QUESTION FEATURES 
One of the main challenges in developing a 
supervised classifier for a particular domain is to 
identifyand design a rich set of features – a process 
which is generally referred to as feature engineering. 
In the subsections that follow, we present the different 
types of features that were used in the question 
classifier, and how they are extracted from a given 
question. 
4.1Lexical features 
Lexical features refer to word related features that are 
extracted directly from the question. In this work,we 
use word level n-grams as lexical features. We also 
include in this section the techniques of stemming and 
stop word removal, which can be used to reduce the 
dimensionality of the feature set. 
4.1.1 Stemming and Stop word removal 
Stemming is a technique that reduces words to their 
grammatical roots or stems, by removing their affixes. 
For instance, after applying stemming, the words 
inventing and invented both become invent. We 
exploit this technique in our question classifier in the 
following manner. First, we represent the question 
using the bag-of-words model as previously 
described. Second, we apply Porter’s stemming 
algorithm (Porter, 1980) to transform each word into 
its stem. The following two examples depict a 
question before and after stemming are applied, 
respectively. 
(1) Which countries are bordered by France? 
(2) Which country are border by Franc? 
Another related technique is to remove stop words, 
which are frequently occurring words with no 
semantic value, such as the articles the and an. Both 
of these techniques are mainly used to reduce the 
feature space of the classifier – i.e., the number of 
total features that need to be considered. This is 
achieved by collapsing several different forms of the 
same word into one distinct term by applying 
stemming; or by eliminating words which are likely to 
be present in most questions – stop words –, and 
which do not provide useful information for the 
classifier. 
4.2 Syntactic Features 
In addition to the information that is readily available 
in the input instance, it is common in natural language 
processing tasks to augment sentence representation 
with syntactic categories, under the assumption that 
the sought-after property, for which we seek the 
classifier, depends on the syntactic role of a word in 
the sentence rather than the specific word . 
4.2.1Question headword 
The question headword 1 is a word in a given 
question that represents the information that is being 
Sought after. In the following examples, the headword 
is in bold face: 
(1) What is Australia’s national flower? 
(2) Name an American made motorcycle. 
(3) Which country are Godiva chocolates from? 
(4) What is the name of the highest mountain in 
Africa? 
In Example 1,2,3, the 
headword flower provides the classifier with an 
important clue to correctly classify the question to 
ENTITY:PLANT. By the same token, motorcycle in 
Example 4 renders hints that help classify the question 
to ENTITY:VEHICLE. Indeed, the aforementioned 
examples’ entire headword serves as an important 
feature to unveil the question’s category, which is 
why we dedicate a great effort to its accurate 
extraction. Our baseline classifier makes use of the 
standard POS information and phrase information 
extracted by a shallow parser. Specifically, we use 
chunks (non overlapping phrases) and head chunks, 
.The following example illustrates the information 
available when generating the syntax-augmented 
feature-based representation. Question: Who was the 
first woman killed in the Vietnam War? Chunking: 
[NP Who] [VP was] [NP the first woman] [VP killed] 
[PP in] [NP the Vietnam War] ? 
The head chunks 
denote the first noun or verb chunk after the question 
word in a question. For example, in the above 
question, the first noun chunk after the question word 
who is ‘the first woman’. The features are represented 
as abstract tags in each example. 
4.3 Semantic Features 
Similar logic can be applied to semantic categories. In 
many cases, the property seems not depend on the 
specific word used in the sentence – that could be 
replaced without affecting this property – but rather 
on its ‘meaning’. For example, given the question: 
What Cuban dictator did Fidel Castro force out of 
power in 1958?, we would like to determine that its 
answer 
Should be a name of a person. Knowing that dictator 
refers to a person is essential to correct classification. 
This work systematically 
studies four semantic information sources and their 
contribution to classification: (1) automatically 
acquired named entity categories -NE, (2) word 
senses in WordNet -SemWN, (3) manually 
constructed word lists related to specific categories of 
interest -SemCSR, and (4) automatically generated
International Journal of Research in Advent Technology, Vol.2, No.8, August 2014 
E-ISSN: 2321-9637 
80 
semantically similar word lists (Zhang, D., & Lee, W. 
S, 2003) -SemSWL. 
For the four external semantic 
information sources, we define semantic categories of 
words and incorporate the information into question 
classification in the 
same way: if a word w occurs in a question, the 
question representation is augmented with the 
semantic category(ies), of the word. For example, in 
the question: What is the state flower of California? 
given that plant (for example) is the only semantic 
class of flower, the feature extractor adds plant, an 
abstract label to the question representation. 
4.3.1 Named Entities 
A named entity (NE) recognizer assigns a semantic 
category to some of the noun phrases in the question. 
The scope of the categories used here is broader than 
the common named entity recognizer. With additional 
categories that could help question answering, such as 
profession, event, holiday, plant, sport, medical etc., 
we redefine our task in the direction of semantic 
categorization. The named entity recognizer was built 
on the shallow parser described in (Voorhees, E. M. 
(2004).), and was trained to categorize noun phrases 
into one of 34 different semantic categories of varying 
specificity. Its overall accuracy (F¯ =1) is above 90%. 
For the question Who was the woman killed in the 
Vietnam War ?, the named entity tagger will return: 
NE: Who was the [Num first] woman killed in the 
[Event Vietnam War] ? As described above, the 
identified named entities are added to the question 
representation. 
4.3.2WordNet Senses 
In WordNet (C. Peters,2005)words are organized 
according to their ‘senses’ (meanings). Words of the 
same sense can, in principle, be exchanged in some 
contexts. The senses are organized in a hierarchy of 
hypernyms and hyponyms. Word senses provide 
another effective way to describe the semantic 
category of a word. For example, in WordNet 1.7, the 
word water belongs to 5 senses. The first two senses 
are: 
Sense 1: binary compound that occurs at room 
temperature 
as a colorless odorless liquid; 
Sense 2: body of water. 
Sense 1 contains words fH2O, water} while Sense 2 
contains 
water, body of water. Sense 1 has a hypernym 
Sense 3: binary compound); and one hyponym of 
Sense 2 is (Sense 4: tap water). For each word in a 
question, all of its sense IDs and direct hypernym and 
hyponym IDs are extracted as features. 
This approach possibly introduces 
significant noise to classification since only a small 
proportion of senses are really related. 
5 SUPPORT VECTOR MACHINE 
Machine learning tasks can be of several forms. 
In supervised learning, the computer is presented with 
example inputs and their desired outputs, given by a 
"teacher", and the goal is to learn a general rule 
that maps inputs to outputs. Spam filtering is an 
example of supervised learning, 
particular classification, where the learning algorithm 
is presented with email (or other) messages labeled 
beforehand as "spam" or "not spam", to produce a 
computer program that labels unseen messages as 
either spam or not. 
In unsupervised learning, no labels 
are given to the learning algorithm, leaving it on its 
own to groups of similar inputs (clustering),density 
estimates or projections of high-dimensional data that 
can be visualised effectively.[2]:3 Unsupervised 
learning can be a goal in itself (discovering hidden 
patterns in data) or a means towards an end. Topic 
modeling is an example of unsupervised learning, 
where a program is given a list of human 
language documents and is tasked to find out which 
documents cover similar topics Supervised learning is 
the machine learning task of inferring a function from 
labeled training data.[1] The training data consist of a 
set of training examples. In supervised learning, each 
example is a pair consisting of an input object 
(typically a vector) and a desired output value (also 
called the supervisory signal). A supervised learning 
algorithm analyzes the training data and produces an 
inferred function, which can be used for mapping new 
examples. An optimal scenario will allow for the 
algorithm to correctly determine the class labels for 
unseen instances. This requires the learning algorithm 
to generalize from the training data to unseen 
situations in a "reasonable" way (see inductive bias). 
In machine learning, the problem of unsupervised 
learning is that of trying to find hidden structure in 
unlabeled data. Since the examples given to the 
learner are unlabeled, there is no error or reward 
signal to evaluate a potential solution. This 
distinguishes unsupervised learning from supervised 
learning and reinforcement learning. 
Unsupervised 
learning is closely related to the problem of density 
estimation in statistics.[1] However unsupervised 
learning also encompasses many other techniques that 
seek to summarize and explain key features of the 
data. Many methods employed in unsupervised 
learning
International Journal of Research in Advent Technology, Vol.2, No.8, August 2014 
E-ISSN: 2321-9637 
81 
6 CONCLUSION 
In this paper we presented a detailed overview on 
learning-based question classification approaches. 
Question classification is a hard problem. In fact the 
machine need to understand the question and classify 
it to the right category. This is done by a series of 
complicated steps. In this paper we reviewed different 
learning methods and feature extraction techniques for 
question classification. Deciding for the best model 
and optimal set of features is not a simple problem. 
Enhancing the feature space with syntactic and 
semantic features can usually improve the 
classification accuracy. 
. 
7 FUTURE WORK 
In the question classification task, we have shown that 
a machine learning-based classifier using solely 
superficial features . Increase the accuracy in question 
answer system with the combination of the three 
feature by using svm (support vector system) method. 
8 RESULT 
It increases the accuracy of the answer detection. It 
give 95.2% of accuracy. 
Acknowledgements 
I thank PROF .Megha Mishra for several valuable 
suggestions and the entire SSCET team for help with 
various components, feature suggestions and 
guidance. 
REFERENCES 
[1] Question classification using support vector 
machines. By Zhang, D., & Lee, W. S. (2003). In 
Proceedings of the 26th annual international acm 
sigir conference on researc and 
developmentininformaionretrieval(pp.26–32). 
[2] Voorhees, E. M. (2004). Overview of the trec 
2004 question answering track. In E. M. 
Voorhees & L. P.Buckland (Eds.), Trec (Vol. 
Special Publication 500-261). National Institute 
of Standards and Technology(NIST). 
[3] Wang, Y.-C., Wu, J.-C., Liang, T., & Chang, J. S. 
(2005). Web-based unsupervised learning for 
queryformulation in question answering. In Ijcnlp 
(p. 519-529). 
[4] Accessingmultilingualinformation2005multilingu 
alquestion answering track. In C. Peters (Ed.), 
repositories.Berlin, Heidelberg: Springer-Verlag. 
[5] Adaptive information extraction. ACM Comput. 
Surv.,Turmo, J., Ageno, A., & Catal`a, N. (2006). 
38(2), 4.Vallin, A., Magnini, B., Giampiccolo, 
D., Aunimo, L., & Ayache, C. (2006). 
[6] Improved inference for unlexicalized parsing by 
Petrov, S., & Klein, D. (2007, April).. In Human 
language technologies2007: The conference of 
the north american chapter of the association for 
computational . 
[7] Question classification with semantic tree kernel. 
Pan, Y., Tang, Y., Lin, L., & Luo, Y. (2008). 
InProceedings of the 31st annual international 
acm sigir conference on research and 
development in information retrieval (pp. 837– 
838). New York, NY, USA: AC 
[8] Designing an interactive open-domain question 
answering 
[9] System by Quarteroni, S&Manandhar, S. 
(2009)..forthcoming,Journal of Natural Language 
Engineering,Volume 15 Issue 1. 
[10]Biomedical Semantics by Chanlekha and Collier 
(2010)Journalof,1:3 
http://www.jbiomedsem.com/content/1/1/3 
[11]Document Classification with Support Vector 
Machines By Konstantin Mertsalov Principal 
Scientist, Machine and Computational Learning 
Rational Retention, LLC 
kmertsalov@rationalretention.com January 2009 
[12] Information Processing and Management journal 
Trento, Italy(2011) homepage: 
www.elsevier.com/ locate/ infoproman Linguistic 
kernels for answer re-ranking in question 
answering systems Alessandro Moschitti, Silvia 
Quarteroni University of Trento, Via Sommarive 
14, 38050 Povo.

More Related Content

What's hot

A Fuzzy Logic Intelligent Agent for Information Extraction
A Fuzzy Logic Intelligent Agent for Information ExtractionA Fuzzy Logic Intelligent Agent for Information Extraction
A Fuzzy Logic Intelligent Agent for Information Extraction
TarekMourad8
 
AUTOMATED SHORT ANSWER GRADER USING FRIENDSHIP GRAPHS
AUTOMATED SHORT ANSWER GRADER USING FRIENDSHIP GRAPHSAUTOMATED SHORT ANSWER GRADER USING FRIENDSHIP GRAPHS
AUTOMATED SHORT ANSWER GRADER USING FRIENDSHIP GRAPHS
csandit
 
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUE
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUECOMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUE
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUE
Journal For Research
 
Ijarcet vol-2-issue-4-1363-1367
Ijarcet vol-2-issue-4-1363-1367Ijarcet vol-2-issue-4-1363-1367
Ijarcet vol-2-issue-4-1363-1367
Editor IJARCET
 
Using Decision Tree for Automatic Identification of Bengali Noun-Noun Compounds
Using Decision Tree for Automatic Identification of Bengali Noun-Noun CompoundsUsing Decision Tree for Automatic Identification of Bengali Noun-Noun Compounds
Using Decision Tree for Automatic Identification of Bengali Noun-Noun Compounds
idescitation
 
NLP_Project_Paper_up276_vec241
NLP_Project_Paper_up276_vec241NLP_Project_Paper_up276_vec241
NLP_Project_Paper_up276_vec241
Urjit Patel
 
OOAD - UML - Sequence and Communication Diagrams - Lab
OOAD - UML - Sequence and Communication Diagrams - LabOOAD - UML - Sequence and Communication Diagrams - Lab
OOAD - UML - Sequence and Communication Diagrams - Lab
Victer Paul
 
Assessment of Programming Language Reliability Utilizing Soft-Computing
Assessment of Programming Language Reliability Utilizing Soft-ComputingAssessment of Programming Language Reliability Utilizing Soft-Computing
Assessment of Programming Language Reliability Utilizing Soft-Computing
ijcsa
 
TEXT SENTIMENTS FOR FORUMS HOTSPOT DETECTION
TEXT SENTIMENTS FOR FORUMS HOTSPOT DETECTIONTEXT SENTIMENTS FOR FORUMS HOTSPOT DETECTION
TEXT SENTIMENTS FOR FORUMS HOTSPOT DETECTION
ijistjournal
 
A Survey on Sentiment Categorization of Movie Reviews
A Survey on Sentiment Categorization of Movie ReviewsA Survey on Sentiment Categorization of Movie Reviews
A Survey on Sentiment Categorization of Movie Reviews
Editor IJMTER
 
Implementation of Semantic Analysis Using Domain Ontology
Implementation of Semantic Analysis Using Domain OntologyImplementation of Semantic Analysis Using Domain Ontology
Implementation of Semantic Analysis Using Domain Ontology
IOSR Journals
 
QUESTION ANALYSIS FOR ARABIC QUESTION ANSWERING SYSTEMS
QUESTION ANALYSIS FOR ARABIC QUESTION ANSWERING SYSTEMS QUESTION ANALYSIS FOR ARABIC QUESTION ANSWERING SYSTEMS
QUESTION ANALYSIS FOR ARABIC QUESTION ANSWERING SYSTEMS
ijnlc
 
Sentiment analysis using naive bayes classifier
Sentiment analysis using naive bayes classifier Sentiment analysis using naive bayes classifier
Sentiment analysis using naive bayes classifier
Dev Sahu
 
D017422528
D017422528D017422528
D017422528
IOSR Journals
 
Suitability of naïve bayesian methods for paragraph level text classification...
Suitability of naïve bayesian methods for paragraph level text classification...Suitability of naïve bayesian methods for paragraph level text classification...
Suitability of naïve bayesian methods for paragraph level text classification...
ijaia
 
Domain Specific Named Entity Recognition Using Supervised Approach
Domain Specific Named Entity Recognition Using Supervised ApproachDomain Specific Named Entity Recognition Using Supervised Approach
Domain Specific Named Entity Recognition Using Supervised Approach
Waqas Tariq
 
Novel Scoring System for Identify Accurate Answers for Factoid Questions
Novel Scoring System for Identify Accurate Answers for Factoid QuestionsNovel Scoring System for Identify Accurate Answers for Factoid Questions
Novel Scoring System for Identify Accurate Answers for Factoid Questions
International Journal of Science and Research (IJSR)
 
Modeling Text Independent Speaker Identification with Vector Quantization
Modeling Text Independent Speaker Identification with Vector QuantizationModeling Text Independent Speaker Identification with Vector Quantization
Modeling Text Independent Speaker Identification with Vector Quantization
TELKOMNIKA JOURNAL
 
INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS
INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS
INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS
cscpconf
 

What's hot (19)

A Fuzzy Logic Intelligent Agent for Information Extraction
A Fuzzy Logic Intelligent Agent for Information ExtractionA Fuzzy Logic Intelligent Agent for Information Extraction
A Fuzzy Logic Intelligent Agent for Information Extraction
 
AUTOMATED SHORT ANSWER GRADER USING FRIENDSHIP GRAPHS
AUTOMATED SHORT ANSWER GRADER USING FRIENDSHIP GRAPHSAUTOMATED SHORT ANSWER GRADER USING FRIENDSHIP GRAPHS
AUTOMATED SHORT ANSWER GRADER USING FRIENDSHIP GRAPHS
 
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUE
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUECOMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUE
COMPREHENSIVE ANALYSIS OF NATURAL LANGUAGE PROCESSING TECHNIQUE
 
Ijarcet vol-2-issue-4-1363-1367
Ijarcet vol-2-issue-4-1363-1367Ijarcet vol-2-issue-4-1363-1367
Ijarcet vol-2-issue-4-1363-1367
 
Using Decision Tree for Automatic Identification of Bengali Noun-Noun Compounds
Using Decision Tree for Automatic Identification of Bengali Noun-Noun CompoundsUsing Decision Tree for Automatic Identification of Bengali Noun-Noun Compounds
Using Decision Tree for Automatic Identification of Bengali Noun-Noun Compounds
 
NLP_Project_Paper_up276_vec241
NLP_Project_Paper_up276_vec241NLP_Project_Paper_up276_vec241
NLP_Project_Paper_up276_vec241
 
OOAD - UML - Sequence and Communication Diagrams - Lab
OOAD - UML - Sequence and Communication Diagrams - LabOOAD - UML - Sequence and Communication Diagrams - Lab
OOAD - UML - Sequence and Communication Diagrams - Lab
 
Assessment of Programming Language Reliability Utilizing Soft-Computing
Assessment of Programming Language Reliability Utilizing Soft-ComputingAssessment of Programming Language Reliability Utilizing Soft-Computing
Assessment of Programming Language Reliability Utilizing Soft-Computing
 
TEXT SENTIMENTS FOR FORUMS HOTSPOT DETECTION
TEXT SENTIMENTS FOR FORUMS HOTSPOT DETECTIONTEXT SENTIMENTS FOR FORUMS HOTSPOT DETECTION
TEXT SENTIMENTS FOR FORUMS HOTSPOT DETECTION
 
A Survey on Sentiment Categorization of Movie Reviews
A Survey on Sentiment Categorization of Movie ReviewsA Survey on Sentiment Categorization of Movie Reviews
A Survey on Sentiment Categorization of Movie Reviews
 
Implementation of Semantic Analysis Using Domain Ontology
Implementation of Semantic Analysis Using Domain OntologyImplementation of Semantic Analysis Using Domain Ontology
Implementation of Semantic Analysis Using Domain Ontology
 
QUESTION ANALYSIS FOR ARABIC QUESTION ANSWERING SYSTEMS
QUESTION ANALYSIS FOR ARABIC QUESTION ANSWERING SYSTEMS QUESTION ANALYSIS FOR ARABIC QUESTION ANSWERING SYSTEMS
QUESTION ANALYSIS FOR ARABIC QUESTION ANSWERING SYSTEMS
 
Sentiment analysis using naive bayes classifier
Sentiment analysis using naive bayes classifier Sentiment analysis using naive bayes classifier
Sentiment analysis using naive bayes classifier
 
D017422528
D017422528D017422528
D017422528
 
Suitability of naïve bayesian methods for paragraph level text classification...
Suitability of naïve bayesian methods for paragraph level text classification...Suitability of naïve bayesian methods for paragraph level text classification...
Suitability of naïve bayesian methods for paragraph level text classification...
 
Domain Specific Named Entity Recognition Using Supervised Approach
Domain Specific Named Entity Recognition Using Supervised ApproachDomain Specific Named Entity Recognition Using Supervised Approach
Domain Specific Named Entity Recognition Using Supervised Approach
 
Novel Scoring System for Identify Accurate Answers for Factoid Questions
Novel Scoring System for Identify Accurate Answers for Factoid QuestionsNovel Scoring System for Identify Accurate Answers for Factoid Questions
Novel Scoring System for Identify Accurate Answers for Factoid Questions
 
Modeling Text Independent Speaker Identification with Vector Quantization
Modeling Text Independent Speaker Identification with Vector QuantizationModeling Text Independent Speaker Identification with Vector Quantization
Modeling Text Independent Speaker Identification with Vector Quantization
 
INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS
INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS
INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS
 

Viewers also liked

Paper id 27201445
Paper id 27201445Paper id 27201445
Paper id 27201445
IJRAT
 
Paper id 21201414
Paper id 21201414Paper id 21201414
Paper id 21201414
IJRAT
 
Paper id 26201415
Paper id 26201415Paper id 26201415
Paper id 26201415
IJRAT
 
Paper id 21201425
Paper id 21201425Paper id 21201425
Paper id 21201425
IJRAT
 
Paper id 21201418
Paper id 21201418Paper id 21201418
Paper id 21201418
IJRAT
 
Paper id 312201512
Paper id 312201512Paper id 312201512
Paper id 312201512
IJRAT
 
Paper id 27201415
Paper id 27201415Paper id 27201415
Paper id 27201415
IJRAT
 
Paper id 42201625
Paper id 42201625Paper id 42201625
Paper id 42201625
IJRAT
 
Paper id 27201420
Paper id 27201420Paper id 27201420
Paper id 27201420
IJRAT
 
Paper id 26201482
Paper id 26201482Paper id 26201482
Paper id 26201482
IJRAT
 
Paper id 41201609
Paper id 41201609Paper id 41201609
Paper id 41201609
IJRAT
 
Paper id 212014105
Paper id 212014105Paper id 212014105
Paper id 212014105
IJRAT
 
Paper id 2820141
Paper id 2820141Paper id 2820141
Paper id 2820141
IJRAT
 
Paper id 2720146
Paper id 2720146Paper id 2720146
Paper id 2720146
IJRAT
 
Paper id 28201435
Paper id 28201435Paper id 28201435
Paper id 28201435
IJRAT
 
Paper id 2120147
Paper id 2120147Paper id 2120147
Paper id 2120147
IJRAT
 
Paper id 311201535
Paper id 311201535Paper id 311201535
Paper id 311201535
IJRAT
 
Paper id 28201431
Paper id 28201431Paper id 28201431
Paper id 28201431
IJRAT
 
Paper id 21201493
Paper id 21201493Paper id 21201493
Paper id 21201493
IJRAT
 
Paper id 27201418
Paper id 27201418Paper id 27201418
Paper id 27201418
IJRAT
 

Viewers also liked (20)

Paper id 27201445
Paper id 27201445Paper id 27201445
Paper id 27201445
 
Paper id 21201414
Paper id 21201414Paper id 21201414
Paper id 21201414
 
Paper id 26201415
Paper id 26201415Paper id 26201415
Paper id 26201415
 
Paper id 21201425
Paper id 21201425Paper id 21201425
Paper id 21201425
 
Paper id 21201418
Paper id 21201418Paper id 21201418
Paper id 21201418
 
Paper id 312201512
Paper id 312201512Paper id 312201512
Paper id 312201512
 
Paper id 27201415
Paper id 27201415Paper id 27201415
Paper id 27201415
 
Paper id 42201625
Paper id 42201625Paper id 42201625
Paper id 42201625
 
Paper id 27201420
Paper id 27201420Paper id 27201420
Paper id 27201420
 
Paper id 26201482
Paper id 26201482Paper id 26201482
Paper id 26201482
 
Paper id 41201609
Paper id 41201609Paper id 41201609
Paper id 41201609
 
Paper id 212014105
Paper id 212014105Paper id 212014105
Paper id 212014105
 
Paper id 2820141
Paper id 2820141Paper id 2820141
Paper id 2820141
 
Paper id 2720146
Paper id 2720146Paper id 2720146
Paper id 2720146
 
Paper id 28201435
Paper id 28201435Paper id 28201435
Paper id 28201435
 
Paper id 2120147
Paper id 2120147Paper id 2120147
Paper id 2120147
 
Paper id 311201535
Paper id 311201535Paper id 311201535
Paper id 311201535
 
Paper id 28201431
Paper id 28201431Paper id 28201431
Paper id 28201431
 
Paper id 21201493
Paper id 21201493Paper id 21201493
Paper id 21201493
 
Paper id 27201418
Paper id 27201418Paper id 27201418
Paper id 27201418
 

Similar to Paper id 28201441

Lexicon base approch
Lexicon base approchLexicon base approch
Lexicon base approch
anil maurya
 
Question Classification using Semantic, Syntactic and Lexical features
Question Classification using Semantic, Syntactic and Lexical featuresQuestion Classification using Semantic, Syntactic and Lexical features
Question Classification using Semantic, Syntactic and Lexical features
dannyijwest
 
Question Retrieval in Community Question Answering via NON-Negative Matrix Fa...
Question Retrieval in Community Question Answering via NON-Negative Matrix Fa...Question Retrieval in Community Question Answering via NON-Negative Matrix Fa...
Question Retrieval in Community Question Answering via NON-Negative Matrix Fa...
IRJET Journal
 
Generation of Question and Answer from Unstructured Document using Gaussian M...
Generation of Question and Answer from Unstructured Document using Gaussian M...Generation of Question and Answer from Unstructured Document using Gaussian M...
Generation of Question and Answer from Unstructured Document using Gaussian M...
IJACEE IJACEE
 
Open domain question answering system using semantic role labeling
Open domain question answering system using semantic role labelingOpen domain question answering system using semantic role labeling
Open domain question answering system using semantic role labeling
eSAT Publishing House
 
A SURVEY OF SENTIMENT CLASSSIFICTION TECHNIQUES
A SURVEY OF SENTIMENT CLASSSIFICTION TECHNIQUESA SURVEY OF SENTIMENT CLASSSIFICTION TECHNIQUES
A SURVEY OF SENTIMENT CLASSSIFICTION TECHNIQUES
Journal For Research
 
NLP Techniques for Question Answering.docx
NLP Techniques for Question Answering.docxNLP Techniques for Question Answering.docx
NLP Techniques for Question Answering.docx
KevinSims18
 
Architecture of an ontology based domain-specific natural language question a...
Architecture of an ontology based domain-specific natural language question a...Architecture of an ontology based domain-specific natural language question a...
Architecture of an ontology based domain-specific natural language question a...
IJwest
 
A New Active Learning Technique Using Furthest Nearest Neighbour Criterion fo...
A New Active Learning Technique Using Furthest Nearest Neighbour Criterion fo...A New Active Learning Technique Using Furthest Nearest Neighbour Criterion fo...
A New Active Learning Technique Using Furthest Nearest Neighbour Criterion fo...
ijcsa
 
NE7012- SOCIAL NETWORK ANALYSIS
NE7012- SOCIAL NETWORK ANALYSISNE7012- SOCIAL NETWORK ANALYSIS
NE7012- SOCIAL NETWORK ANALYSIS
rathnaarul
 
An Introduction to Machine Learning
An Introduction to Machine LearningAn Introduction to Machine Learning
An Introduction to Machine Learning
Vedaj Padman
 
introduction to machine learning and nlp
introduction to machine learning and nlpintroduction to machine learning and nlp
introduction to machine learning and nlp
Mahmoud Farag
 
Natural Language Processing Through Different Classes of Machine Learning
Natural Language Processing Through Different Classes of Machine LearningNatural Language Processing Through Different Classes of Machine Learning
Natural Language Processing Through Different Classes of Machine Learning
csandit
 
Application of hidden markov model in question answering systems
Application of hidden markov model in question answering systemsApplication of hidden markov model in question answering systems
Application of hidden markov model in question answering systems
ijcsa
 
Semantic based automatic question generation using artificial immune system
Semantic based automatic question generation using artificial immune systemSemantic based automatic question generation using artificial immune system
Semantic based automatic question generation using artificial immune system
Alexander Decker
 
Supervised Sentiment Classification using DTDP algorithm
Supervised Sentiment Classification using DTDP algorithmSupervised Sentiment Classification using DTDP algorithm
Supervised Sentiment Classification using DTDP algorithm
IJSRD
 
The Role of Families and the Community Proposal Template (N.docx
The Role of Families and the Community Proposal Template  (N.docxThe Role of Families and the Community Proposal Template  (N.docx
The Role of Families and the Community Proposal Template (N.docx
ssusera34210
 
A simplified classification computational model of opinion mining using deep ...
A simplified classification computational model of opinion mining using deep ...A simplified classification computational model of opinion mining using deep ...
A simplified classification computational model of opinion mining using deep ...
IJECEIAES
 
An Adaptive Approach for Subjective Answer Evaluation
An Adaptive Approach for Subjective Answer EvaluationAn Adaptive Approach for Subjective Answer Evaluation
An Adaptive Approach for Subjective Answer Evaluation
vivatechijri
 
Deep learning based Arabic short answer grading in serious games
Deep learning based Arabic short answer grading in serious gamesDeep learning based Arabic short answer grading in serious games
Deep learning based Arabic short answer grading in serious games
IJECEIAES
 

Similar to Paper id 28201441 (20)

Lexicon base approch
Lexicon base approchLexicon base approch
Lexicon base approch
 
Question Classification using Semantic, Syntactic and Lexical features
Question Classification using Semantic, Syntactic and Lexical featuresQuestion Classification using Semantic, Syntactic and Lexical features
Question Classification using Semantic, Syntactic and Lexical features
 
Question Retrieval in Community Question Answering via NON-Negative Matrix Fa...
Question Retrieval in Community Question Answering via NON-Negative Matrix Fa...Question Retrieval in Community Question Answering via NON-Negative Matrix Fa...
Question Retrieval in Community Question Answering via NON-Negative Matrix Fa...
 
Generation of Question and Answer from Unstructured Document using Gaussian M...
Generation of Question and Answer from Unstructured Document using Gaussian M...Generation of Question and Answer from Unstructured Document using Gaussian M...
Generation of Question and Answer from Unstructured Document using Gaussian M...
 
Open domain question answering system using semantic role labeling
Open domain question answering system using semantic role labelingOpen domain question answering system using semantic role labeling
Open domain question answering system using semantic role labeling
 
A SURVEY OF SENTIMENT CLASSSIFICTION TECHNIQUES
A SURVEY OF SENTIMENT CLASSSIFICTION TECHNIQUESA SURVEY OF SENTIMENT CLASSSIFICTION TECHNIQUES
A SURVEY OF SENTIMENT CLASSSIFICTION TECHNIQUES
 
NLP Techniques for Question Answering.docx
NLP Techniques for Question Answering.docxNLP Techniques for Question Answering.docx
NLP Techniques for Question Answering.docx
 
Architecture of an ontology based domain-specific natural language question a...
Architecture of an ontology based domain-specific natural language question a...Architecture of an ontology based domain-specific natural language question a...
Architecture of an ontology based domain-specific natural language question a...
 
A New Active Learning Technique Using Furthest Nearest Neighbour Criterion fo...
A New Active Learning Technique Using Furthest Nearest Neighbour Criterion fo...A New Active Learning Technique Using Furthest Nearest Neighbour Criterion fo...
A New Active Learning Technique Using Furthest Nearest Neighbour Criterion fo...
 
NE7012- SOCIAL NETWORK ANALYSIS
NE7012- SOCIAL NETWORK ANALYSISNE7012- SOCIAL NETWORK ANALYSIS
NE7012- SOCIAL NETWORK ANALYSIS
 
An Introduction to Machine Learning
An Introduction to Machine LearningAn Introduction to Machine Learning
An Introduction to Machine Learning
 
introduction to machine learning and nlp
introduction to machine learning and nlpintroduction to machine learning and nlp
introduction to machine learning and nlp
 
Natural Language Processing Through Different Classes of Machine Learning
Natural Language Processing Through Different Classes of Machine LearningNatural Language Processing Through Different Classes of Machine Learning
Natural Language Processing Through Different Classes of Machine Learning
 
Application of hidden markov model in question answering systems
Application of hidden markov model in question answering systemsApplication of hidden markov model in question answering systems
Application of hidden markov model in question answering systems
 
Semantic based automatic question generation using artificial immune system
Semantic based automatic question generation using artificial immune systemSemantic based automatic question generation using artificial immune system
Semantic based automatic question generation using artificial immune system
 
Supervised Sentiment Classification using DTDP algorithm
Supervised Sentiment Classification using DTDP algorithmSupervised Sentiment Classification using DTDP algorithm
Supervised Sentiment Classification using DTDP algorithm
 
The Role of Families and the Community Proposal Template (N.docx
The Role of Families and the Community Proposal Template  (N.docxThe Role of Families and the Community Proposal Template  (N.docx
The Role of Families and the Community Proposal Template (N.docx
 
A simplified classification computational model of opinion mining using deep ...
A simplified classification computational model of opinion mining using deep ...A simplified classification computational model of opinion mining using deep ...
A simplified classification computational model of opinion mining using deep ...
 
An Adaptive Approach for Subjective Answer Evaluation
An Adaptive Approach for Subjective Answer EvaluationAn Adaptive Approach for Subjective Answer Evaluation
An Adaptive Approach for Subjective Answer Evaluation
 
Deep learning based Arabic short answer grading in serious games
Deep learning based Arabic short answer grading in serious gamesDeep learning based Arabic short answer grading in serious games
Deep learning based Arabic short answer grading in serious games
 

More from IJRAT

96202108
9620210896202108
96202108
IJRAT
 
97202107
9720210797202107
97202107
IJRAT
 
93202101
9320210193202101
93202101
IJRAT
 
92202102
9220210292202102
92202102
IJRAT
 
91202104
9120210491202104
91202104
IJRAT
 
87202003
8720200387202003
87202003
IJRAT
 
87202001
8720200187202001
87202001
IJRAT
 
86202013
8620201386202013
86202013
IJRAT
 
86202008
8620200886202008
86202008
IJRAT
 
86202005
8620200586202005
86202005
IJRAT
 
86202004
8620200486202004
86202004
IJRAT
 
85202026
8520202685202026
85202026
IJRAT
 
711201940
711201940711201940
711201940
IJRAT
 
711201939
711201939711201939
711201939
IJRAT
 
711201935
711201935711201935
711201935
IJRAT
 
711201927
711201927711201927
711201927
IJRAT
 
711201905
711201905711201905
711201905
IJRAT
 
710201947
710201947710201947
710201947
IJRAT
 
712201907
712201907712201907
712201907
IJRAT
 
712201903
712201903712201903
712201903
IJRAT
 

More from IJRAT (20)

96202108
9620210896202108
96202108
 
97202107
9720210797202107
97202107
 
93202101
9320210193202101
93202101
 
92202102
9220210292202102
92202102
 
91202104
9120210491202104
91202104
 
87202003
8720200387202003
87202003
 
87202001
8720200187202001
87202001
 
86202013
8620201386202013
86202013
 
86202008
8620200886202008
86202008
 
86202005
8620200586202005
86202005
 
86202004
8620200486202004
86202004
 
85202026
8520202685202026
85202026
 
711201940
711201940711201940
711201940
 
711201939
711201939711201939
711201939
 
711201935
711201935711201935
711201935
 
711201927
711201927711201927
711201927
 
711201905
711201905711201905
711201905
 
710201947
710201947710201947
710201947
 
712201907
712201907712201907
712201907
 
712201903
712201903712201903
712201903
 

Recently uploaded

Recycled Concrete Aggregate in Construction Part III
Recycled Concrete Aggregate in Construction Part IIIRecycled Concrete Aggregate in Construction Part III
Recycled Concrete Aggregate in Construction Part III
Aditya Rajan Patra
 
sieving analysis and results interpretation
sieving analysis and results interpretationsieving analysis and results interpretation
sieving analysis and results interpretation
ssuser36d3051
 
Harnessing WebAssembly for Real-time Stateless Streaming Pipelines
Harnessing WebAssembly for Real-time Stateless Streaming PipelinesHarnessing WebAssembly for Real-time Stateless Streaming Pipelines
Harnessing WebAssembly for Real-time Stateless Streaming Pipelines
Christina Lin
 
IEEE Aerospace and Electronic Systems Society as a Graduate Student Member
IEEE Aerospace and Electronic Systems Society as a Graduate Student MemberIEEE Aerospace and Electronic Systems Society as a Graduate Student Member
IEEE Aerospace and Electronic Systems Society as a Graduate Student Member
VICTOR MAESTRE RAMIREZ
 
BPV-GUI-01-Guide-for-ASME-Review-Teams-(General)-10-10-2023.pdf
BPV-GUI-01-Guide-for-ASME-Review-Teams-(General)-10-10-2023.pdfBPV-GUI-01-Guide-for-ASME-Review-Teams-(General)-10-10-2023.pdf
BPV-GUI-01-Guide-for-ASME-Review-Teams-(General)-10-10-2023.pdf
MIGUELANGEL966976
 
14 Template Contractual Notice - EOT Application
14 Template Contractual Notice - EOT Application14 Template Contractual Notice - EOT Application
14 Template Contractual Notice - EOT Application
SyedAbiiAzazi1
 
[JPP-1] - (JEE 3.0) - Kinematics 1D - 14th May..pdf
[JPP-1] - (JEE 3.0) - Kinematics 1D - 14th May..pdf[JPP-1] - (JEE 3.0) - Kinematics 1D - 14th May..pdf
[JPP-1] - (JEE 3.0) - Kinematics 1D - 14th May..pdf
awadeshbabu
 
bank management system in java and mysql report1.pdf
bank management system in java and mysql report1.pdfbank management system in java and mysql report1.pdf
bank management system in java and mysql report1.pdf
Divyam548318
 
Generative AI leverages algorithms to create various forms of content
Generative AI leverages algorithms to create various forms of contentGenerative AI leverages algorithms to create various forms of content
Generative AI leverages algorithms to create various forms of content
Hitesh Mohapatra
 
22CYT12-Unit-V-E Waste and its Management.ppt
22CYT12-Unit-V-E Waste and its Management.ppt22CYT12-Unit-V-E Waste and its Management.ppt
22CYT12-Unit-V-E Waste and its Management.ppt
KrishnaveniKrishnara1
 
Properties Railway Sleepers and Test.pptx
Properties Railway Sleepers and Test.pptxProperties Railway Sleepers and Test.pptx
Properties Railway Sleepers and Test.pptx
MDSABBIROJJAMANPAYEL
 
Understanding Inductive Bias in Machine Learning
Understanding Inductive Bias in Machine LearningUnderstanding Inductive Bias in Machine Learning
Understanding Inductive Bias in Machine Learning
SUTEJAS
 
CSM Cloud Service Management Presentarion
CSM Cloud Service Management PresentarionCSM Cloud Service Management Presentarion
CSM Cloud Service Management Presentarion
rpskprasana
 
basic-wireline-operations-course-mahmoud-f-radwan.pdf
basic-wireline-operations-course-mahmoud-f-radwan.pdfbasic-wireline-operations-course-mahmoud-f-radwan.pdf
basic-wireline-operations-course-mahmoud-f-radwan.pdf
NidhalKahouli2
 
DfMAy 2024 - key insights and contributions
DfMAy 2024 - key insights and contributionsDfMAy 2024 - key insights and contributions
DfMAy 2024 - key insights and contributions
gestioneergodomus
 
spirit beverages ppt without graphics.pptx
spirit beverages ppt without graphics.pptxspirit beverages ppt without graphics.pptx
spirit beverages ppt without graphics.pptx
Madan Karki
 
132/33KV substation case study Presentation
132/33KV substation case study Presentation132/33KV substation case study Presentation
132/33KV substation case study Presentation
kandramariana6
 
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECT
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECTCHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECT
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECT
jpsjournal1
 
Exception Handling notes in java exception
Exception Handling notes in java exceptionException Handling notes in java exception
Exception Handling notes in java exception
Ratnakar Mikkili
 
Manufacturing Process of molasses based distillery ppt.pptx
Manufacturing Process of molasses based distillery ppt.pptxManufacturing Process of molasses based distillery ppt.pptx
Manufacturing Process of molasses based distillery ppt.pptx
Madan Karki
 

Recently uploaded (20)

Recycled Concrete Aggregate in Construction Part III
Recycled Concrete Aggregate in Construction Part IIIRecycled Concrete Aggregate in Construction Part III
Recycled Concrete Aggregate in Construction Part III
 
sieving analysis and results interpretation
sieving analysis and results interpretationsieving analysis and results interpretation
sieving analysis and results interpretation
 
Harnessing WebAssembly for Real-time Stateless Streaming Pipelines
Harnessing WebAssembly for Real-time Stateless Streaming PipelinesHarnessing WebAssembly for Real-time Stateless Streaming Pipelines
Harnessing WebAssembly for Real-time Stateless Streaming Pipelines
 
IEEE Aerospace and Electronic Systems Society as a Graduate Student Member
IEEE Aerospace and Electronic Systems Society as a Graduate Student MemberIEEE Aerospace and Electronic Systems Society as a Graduate Student Member
IEEE Aerospace and Electronic Systems Society as a Graduate Student Member
 
BPV-GUI-01-Guide-for-ASME-Review-Teams-(General)-10-10-2023.pdf
BPV-GUI-01-Guide-for-ASME-Review-Teams-(General)-10-10-2023.pdfBPV-GUI-01-Guide-for-ASME-Review-Teams-(General)-10-10-2023.pdf
BPV-GUI-01-Guide-for-ASME-Review-Teams-(General)-10-10-2023.pdf
 
14 Template Contractual Notice - EOT Application
14 Template Contractual Notice - EOT Application14 Template Contractual Notice - EOT Application
14 Template Contractual Notice - EOT Application
 
[JPP-1] - (JEE 3.0) - Kinematics 1D - 14th May..pdf
[JPP-1] - (JEE 3.0) - Kinematics 1D - 14th May..pdf[JPP-1] - (JEE 3.0) - Kinematics 1D - 14th May..pdf
[JPP-1] - (JEE 3.0) - Kinematics 1D - 14th May..pdf
 
bank management system in java and mysql report1.pdf
bank management system in java and mysql report1.pdfbank management system in java and mysql report1.pdf
bank management system in java and mysql report1.pdf
 
Generative AI leverages algorithms to create various forms of content
Generative AI leverages algorithms to create various forms of contentGenerative AI leverages algorithms to create various forms of content
Generative AI leverages algorithms to create various forms of content
 
22CYT12-Unit-V-E Waste and its Management.ppt
22CYT12-Unit-V-E Waste and its Management.ppt22CYT12-Unit-V-E Waste and its Management.ppt
22CYT12-Unit-V-E Waste and its Management.ppt
 
Properties Railway Sleepers and Test.pptx
Properties Railway Sleepers and Test.pptxProperties Railway Sleepers and Test.pptx
Properties Railway Sleepers and Test.pptx
 
Understanding Inductive Bias in Machine Learning
Understanding Inductive Bias in Machine LearningUnderstanding Inductive Bias in Machine Learning
Understanding Inductive Bias in Machine Learning
 
CSM Cloud Service Management Presentarion
CSM Cloud Service Management PresentarionCSM Cloud Service Management Presentarion
CSM Cloud Service Management Presentarion
 
basic-wireline-operations-course-mahmoud-f-radwan.pdf
basic-wireline-operations-course-mahmoud-f-radwan.pdfbasic-wireline-operations-course-mahmoud-f-radwan.pdf
basic-wireline-operations-course-mahmoud-f-radwan.pdf
 
DfMAy 2024 - key insights and contributions
DfMAy 2024 - key insights and contributionsDfMAy 2024 - key insights and contributions
DfMAy 2024 - key insights and contributions
 
spirit beverages ppt without graphics.pptx
spirit beverages ppt without graphics.pptxspirit beverages ppt without graphics.pptx
spirit beverages ppt without graphics.pptx
 
132/33KV substation case study Presentation
132/33KV substation case study Presentation132/33KV substation case study Presentation
132/33KV substation case study Presentation
 
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECT
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECTCHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECT
CHINA’S GEO-ECONOMIC OUTREACH IN CENTRAL ASIAN COUNTRIES AND FUTURE PROSPECT
 
Exception Handling notes in java exception
Exception Handling notes in java exceptionException Handling notes in java exception
Exception Handling notes in java exception
 
Manufacturing Process of molasses based distillery ppt.pptx
Manufacturing Process of molasses based distillery ppt.pptxManufacturing Process of molasses based distillery ppt.pptx
Manufacturing Process of molasses based distillery ppt.pptx
 

Paper id 28201441

  • 1. International Journal of Research in Advent Technology, Vol.2, No.8, August 2014 E-ISSN: 2321-9637 Question Classification: Using Support Vector Machine 77 and Lexical, Semantic and Sytactic Features Kiran Yadav, Megha Mishra M.E scholar sscet Bhilai, Prof. sscet Bhilai Yadavkiran64@gmail.com Abstract Question classification is play important role in the question answering system. The results of the question classification find out the quality of the question answering system. In this paper, a question classification algorithm based on SVM and feature, Support Vector Machine model is take to train a classifier on coarse categories, there features also use for classify the category. SVM has been used for question classification and have a good results. We use SVM as the classifier. The experiment results show that the feature extraction can perform well with SVM and our approach can reach classification accuracy. Index Terms- Question answering, text classification, machine learning, support vector machine. 1. INTRODUCTION In this work, we use a machine learning approach to question classification. Task of question classification as a supervised learning classification. In order to prepare the learning model, we designed a deep position of features that are prognostic of question categories . In this paper work this classification has two purposes. It provides constraints on the answer types that provide foster processing to just site and verify the answer. Which city has the largest population? we do not want to test each phrase in a document to look that it gives an answer . However, there characteristics of question classification that mark it from the common work. On one hand, questions are relatively short and contain less word-based information equate with classifying the entire text. On the other hand, small questions are amenable for more correct and deeper-level In this way, this work on question classification can be also see as a case study is take semantic information to text classification. Similar to syntactic information such as part-of-speech tags, clear notion of how to use lexical semantic information is to replace or augment each word by its semantic class in the given context, then generate a feature-based representation and learn a mapping from this representation to the desired property. This general scheme leaves several issues open that make the analogy to syntactic categories nontrivial. First, it is not open which semantic category is allow and how to develop them. Second, it is not open how to hold the more dissimilar problem of semantic when decisied the delegacy of a sentence. Merge these three features and increase the accuracy of the question classification by using these features. Question classification plays an important role in question answering. Features are the key to obtain an accurate question classifier. Question answering systems deal defferent it this problem, by giving natural language de in which users can explain their information required form of a natural language question. Retrieve the exact answer to that very same question in place of a set of documents. natural language, from a (typically large) collection of documents, such as the WWW. The developing period of the q/a system in different field is too long and recycle rate is so low. Developed a state of the art machine leaning based question classifier that use a rich a set of lexical, syntactic and semantic features. 2. QUESTION CLASSIFICATION Question Classification means it helps for give the result of given question .It is mainly use for the question answer system. It work category wise example if any type of question it there and find the answer in category it give fast result. When we search any thing it search engine like google then it gives all things which are related to that word which is in search. But it gives the answer in category wise. because of only the question`s answer is presented. Table 1. The coarse question categories Coarse ABBR DESC ENTY HUM LOC NUM
  • 2. International Journal of Research in Advent Technology, Vol.2, No.8, August 2014 E-ISSN: 2321-9637 78 To simplify the following experiments, we assume that one Question resides in only one category. That is to say, unambiguous question is labeled with its most probable category. 2.1 Question types What is the fastest fish in the world? What’s the colored part of the eye called? What color is Mr. Spock’s blood? Name a novel written by John Steinbeck. What currency is used in Australia? What is the fear of cockroaches called? What are the historical trials following WorldWar II called? What is the world ’s best selling cookie? What instrument is Ray Charles best known for playing? What language is mostly spoken in Brazil? What letter adorns the flag of Rwanda? What’s the highest hand in straight poker? What is the state tree of Nebraska? What is the best brand for a laptop computer? What religion has the most members? What game is Garry Kasparov really good at? 3. RELATED WORK Hand-made Rule-based show on extracting names using many of human-made rules set. basically the systems consist of a set of patterns using grammatical (e.g. part of speech), syntactic (e.g. word precedence) and orthographic features (e.g. capitalization) in combination with dictionaries An example for this type of system is: "President rao said bankers talks will make discussions on private, U.S. forces to leave Iraq". In this example a proper noun follows a person's title(president), then noun is a person's name and proper noun that is started with capital character (Iraq) after the verb (to leave) is a Location's name. In this family of approaches, Appelt , propose a name identification system based on carefully handcrafted regular expression called FASTUS. They divided the task into three steps: Recognizing Phrases, Recognizing Patterns and Merging incidents These approaches are relying on manually coded rules and manually compiled corpora. These kinds of models have better results for restricted domains, are capable of detecting complex entities that learning models have difficulty with. However, the rule-based NE systems lack the ability of portability robustness, and furthermore the high cost of the rule maintains increases even though the data is slightly changed. These type of approaches are often domain and language specific and do not necessarily adapt well to new domains and languages. In Machine Learning-based NER system, the purpose of Named Entity Recognition approach is converting identification problem into a classification problem and employs a classification statistical model to solve it. In this type of approach, the systems look for patterns and relationships into text to make a model using statistical models and machine learning algorithms. The systems identify and classify nouns into particular classes such as persons, locations, times, etc base on this model, using machine learning algorithms. There are two types of machine learning model that are use for NER. Supervised and Unsupervised machine learning model. Supervised learning involves using a program that can learn to classify a given set of labeled examples that are made up of the same number of features. Each example is thus represented with respect to the different feature spaces. The learning process is called supervised, because the people who marked up the training examples are teaching the program the right distinctions. The supervised learning approach requires preparing labeled training data to construct a statistical model, but it cannot achieve a good performance without a large amount of training data, because of data sparseness problem. In recent years several statistical methods based on supervised learning method were proposed. Bikel et. al. propose a learning name-finder base on hidden Markov model [8] called Nymbel, while Borthwick et. al. investigates exploiting diverse knowledge sources via maximum entropy in named entity recognition [9,10]. A tagging of unknown proper names system with Decision Tree model was proposed by Bechet et. al. [5], while Wuet. al. presented a named entity recognition system based on support vector machines [2]. Unsupervised learning method is another type of machine learning model, where an unsupervised model learns without any feedback. In unsupervised learning, the goal of the program is to build representations from data. These representations can then be used for data compression, classifying, decision making, and other purposes. Unsupervised learning is not a very popular approach for NER and the systems that do use unsupervised learning are usually not completely unsupervised. In these types of approach, Collins et. al. discusses an unsupervised model for named entity classification by use of unlabeled examples of data [7], Koimetal. Proposes an unsupervised named entity classification models and their ensembles that uses a small-scale named entity dictionary and an unlabeled corpus for classifying named entities [4]. Unlike the rulebased method, these types of approaches can be easily port to different domain or languages. In Hybrid NER system, the approach is to combine rulebased and machine learning-based methods, and make new methods using strongest points from each method.
  • 3. International Journal of Research in Advent Technology, Vol.2, No.8, August 2014 E-ISSN: 2321-9637 79 4. QUESTION FEATURES One of the main challenges in developing a supervised classifier for a particular domain is to identifyand design a rich set of features – a process which is generally referred to as feature engineering. In the subsections that follow, we present the different types of features that were used in the question classifier, and how they are extracted from a given question. 4.1Lexical features Lexical features refer to word related features that are extracted directly from the question. In this work,we use word level n-grams as lexical features. We also include in this section the techniques of stemming and stop word removal, which can be used to reduce the dimensionality of the feature set. 4.1.1 Stemming and Stop word removal Stemming is a technique that reduces words to their grammatical roots or stems, by removing their affixes. For instance, after applying stemming, the words inventing and invented both become invent. We exploit this technique in our question classifier in the following manner. First, we represent the question using the bag-of-words model as previously described. Second, we apply Porter’s stemming algorithm (Porter, 1980) to transform each word into its stem. The following two examples depict a question before and after stemming are applied, respectively. (1) Which countries are bordered by France? (2) Which country are border by Franc? Another related technique is to remove stop words, which are frequently occurring words with no semantic value, such as the articles the and an. Both of these techniques are mainly used to reduce the feature space of the classifier – i.e., the number of total features that need to be considered. This is achieved by collapsing several different forms of the same word into one distinct term by applying stemming; or by eliminating words which are likely to be present in most questions – stop words –, and which do not provide useful information for the classifier. 4.2 Syntactic Features In addition to the information that is readily available in the input instance, it is common in natural language processing tasks to augment sentence representation with syntactic categories, under the assumption that the sought-after property, for which we seek the classifier, depends on the syntactic role of a word in the sentence rather than the specific word . 4.2.1Question headword The question headword 1 is a word in a given question that represents the information that is being Sought after. In the following examples, the headword is in bold face: (1) What is Australia’s national flower? (2) Name an American made motorcycle. (3) Which country are Godiva chocolates from? (4) What is the name of the highest mountain in Africa? In Example 1,2,3, the headword flower provides the classifier with an important clue to correctly classify the question to ENTITY:PLANT. By the same token, motorcycle in Example 4 renders hints that help classify the question to ENTITY:VEHICLE. Indeed, the aforementioned examples’ entire headword serves as an important feature to unveil the question’s category, which is why we dedicate a great effort to its accurate extraction. Our baseline classifier makes use of the standard POS information and phrase information extracted by a shallow parser. Specifically, we use chunks (non overlapping phrases) and head chunks, .The following example illustrates the information available when generating the syntax-augmented feature-based representation. Question: Who was the first woman killed in the Vietnam War? Chunking: [NP Who] [VP was] [NP the first woman] [VP killed] [PP in] [NP the Vietnam War] ? The head chunks denote the first noun or verb chunk after the question word in a question. For example, in the above question, the first noun chunk after the question word who is ‘the first woman’. The features are represented as abstract tags in each example. 4.3 Semantic Features Similar logic can be applied to semantic categories. In many cases, the property seems not depend on the specific word used in the sentence – that could be replaced without affecting this property – but rather on its ‘meaning’. For example, given the question: What Cuban dictator did Fidel Castro force out of power in 1958?, we would like to determine that its answer Should be a name of a person. Knowing that dictator refers to a person is essential to correct classification. This work systematically studies four semantic information sources and their contribution to classification: (1) automatically acquired named entity categories -NE, (2) word senses in WordNet -SemWN, (3) manually constructed word lists related to specific categories of interest -SemCSR, and (4) automatically generated
  • 4. International Journal of Research in Advent Technology, Vol.2, No.8, August 2014 E-ISSN: 2321-9637 80 semantically similar word lists (Zhang, D., & Lee, W. S, 2003) -SemSWL. For the four external semantic information sources, we define semantic categories of words and incorporate the information into question classification in the same way: if a word w occurs in a question, the question representation is augmented with the semantic category(ies), of the word. For example, in the question: What is the state flower of California? given that plant (for example) is the only semantic class of flower, the feature extractor adds plant, an abstract label to the question representation. 4.3.1 Named Entities A named entity (NE) recognizer assigns a semantic category to some of the noun phrases in the question. The scope of the categories used here is broader than the common named entity recognizer. With additional categories that could help question answering, such as profession, event, holiday, plant, sport, medical etc., we redefine our task in the direction of semantic categorization. The named entity recognizer was built on the shallow parser described in (Voorhees, E. M. (2004).), and was trained to categorize noun phrases into one of 34 different semantic categories of varying specificity. Its overall accuracy (F¯ =1) is above 90%. For the question Who was the woman killed in the Vietnam War ?, the named entity tagger will return: NE: Who was the [Num first] woman killed in the [Event Vietnam War] ? As described above, the identified named entities are added to the question representation. 4.3.2WordNet Senses In WordNet (C. Peters,2005)words are organized according to their ‘senses’ (meanings). Words of the same sense can, in principle, be exchanged in some contexts. The senses are organized in a hierarchy of hypernyms and hyponyms. Word senses provide another effective way to describe the semantic category of a word. For example, in WordNet 1.7, the word water belongs to 5 senses. The first two senses are: Sense 1: binary compound that occurs at room temperature as a colorless odorless liquid; Sense 2: body of water. Sense 1 contains words fH2O, water} while Sense 2 contains water, body of water. Sense 1 has a hypernym Sense 3: binary compound); and one hyponym of Sense 2 is (Sense 4: tap water). For each word in a question, all of its sense IDs and direct hypernym and hyponym IDs are extracted as features. This approach possibly introduces significant noise to classification since only a small proportion of senses are really related. 5 SUPPORT VECTOR MACHINE Machine learning tasks can be of several forms. In supervised learning, the computer is presented with example inputs and their desired outputs, given by a "teacher", and the goal is to learn a general rule that maps inputs to outputs. Spam filtering is an example of supervised learning, particular classification, where the learning algorithm is presented with email (or other) messages labeled beforehand as "spam" or "not spam", to produce a computer program that labels unseen messages as either spam or not. In unsupervised learning, no labels are given to the learning algorithm, leaving it on its own to groups of similar inputs (clustering),density estimates or projections of high-dimensional data that can be visualised effectively.[2]:3 Unsupervised learning can be a goal in itself (discovering hidden patterns in data) or a means towards an end. Topic modeling is an example of unsupervised learning, where a program is given a list of human language documents and is tasked to find out which documents cover similar topics Supervised learning is the machine learning task of inferring a function from labeled training data.[1] The training data consist of a set of training examples. In supervised learning, each example is a pair consisting of an input object (typically a vector) and a desired output value (also called the supervisory signal). A supervised learning algorithm analyzes the training data and produces an inferred function, which can be used for mapping new examples. An optimal scenario will allow for the algorithm to correctly determine the class labels for unseen instances. This requires the learning algorithm to generalize from the training data to unseen situations in a "reasonable" way (see inductive bias). In machine learning, the problem of unsupervised learning is that of trying to find hidden structure in unlabeled data. Since the examples given to the learner are unlabeled, there is no error or reward signal to evaluate a potential solution. This distinguishes unsupervised learning from supervised learning and reinforcement learning. Unsupervised learning is closely related to the problem of density estimation in statistics.[1] However unsupervised learning also encompasses many other techniques that seek to summarize and explain key features of the data. Many methods employed in unsupervised learning
  • 5. International Journal of Research in Advent Technology, Vol.2, No.8, August 2014 E-ISSN: 2321-9637 81 6 CONCLUSION In this paper we presented a detailed overview on learning-based question classification approaches. Question classification is a hard problem. In fact the machine need to understand the question and classify it to the right category. This is done by a series of complicated steps. In this paper we reviewed different learning methods and feature extraction techniques for question classification. Deciding for the best model and optimal set of features is not a simple problem. Enhancing the feature space with syntactic and semantic features can usually improve the classification accuracy. . 7 FUTURE WORK In the question classification task, we have shown that a machine learning-based classifier using solely superficial features . Increase the accuracy in question answer system with the combination of the three feature by using svm (support vector system) method. 8 RESULT It increases the accuracy of the answer detection. It give 95.2% of accuracy. Acknowledgements I thank PROF .Megha Mishra for several valuable suggestions and the entire SSCET team for help with various components, feature suggestions and guidance. REFERENCES [1] Question classification using support vector machines. By Zhang, D., & Lee, W. S. (2003). In Proceedings of the 26th annual international acm sigir conference on researc and developmentininformaionretrieval(pp.26–32). [2] Voorhees, E. M. (2004). Overview of the trec 2004 question answering track. In E. M. Voorhees & L. P.Buckland (Eds.), Trec (Vol. Special Publication 500-261). National Institute of Standards and Technology(NIST). [3] Wang, Y.-C., Wu, J.-C., Liang, T., & Chang, J. S. (2005). Web-based unsupervised learning for queryformulation in question answering. In Ijcnlp (p. 519-529). [4] Accessingmultilingualinformation2005multilingu alquestion answering track. In C. Peters (Ed.), repositories.Berlin, Heidelberg: Springer-Verlag. [5] Adaptive information extraction. ACM Comput. Surv.,Turmo, J., Ageno, A., & Catal`a, N. (2006). 38(2), 4.Vallin, A., Magnini, B., Giampiccolo, D., Aunimo, L., & Ayache, C. (2006). [6] Improved inference for unlexicalized parsing by Petrov, S., & Klein, D. (2007, April).. In Human language technologies2007: The conference of the north american chapter of the association for computational . [7] Question classification with semantic tree kernel. Pan, Y., Tang, Y., Lin, L., & Luo, Y. (2008). InProceedings of the 31st annual international acm sigir conference on research and development in information retrieval (pp. 837– 838). New York, NY, USA: AC [8] Designing an interactive open-domain question answering [9] System by Quarteroni, S&Manandhar, S. (2009)..forthcoming,Journal of Natural Language Engineering,Volume 15 Issue 1. [10]Biomedical Semantics by Chanlekha and Collier (2010)Journalof,1:3 http://www.jbiomedsem.com/content/1/1/3 [11]Document Classification with Support Vector Machines By Konstantin Mertsalov Principal Scientist, Machine and Computational Learning Rational Retention, LLC kmertsalov@rationalretention.com January 2009 [12] Information Processing and Management journal Trento, Italy(2011) homepage: www.elsevier.com/ locate/ infoproman Linguistic kernels for answer re-ranking in question answering systems Alessandro Moschitti, Silvia Quarteroni University of Trento, Via Sommarive 14, 38050 Povo.