SlideShare a Scribd company logo
1 of 10
Download to read offline
International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015
DOI : 10.5121/ijnlc.2015.4105 51
EVALUATION OF SUBJECTIVE ANSWERS USING GLSA
ENHANCED WITH CONTEXTUAL SYNONYMY
Parag A. Guruji1
, Mrunal M. Pagnis2
, Sayali M. Pawar3
and Prakash J. Kulkarni4
Department of Computer Science and Engineering,
Walchand College of Engineering, Sangli, India 416415
ABSTRACT
Evaluation of subjective answers submitted in an exam is an essential but one of the most resource
consuming educational activity. This paper details experiments conducted under our project to build a
software that evaluates the subjective answers of informative nature in a given knowledge domain. The
paper first summarizes the techniques such as Generalized Latent Semantic Analysis (GLSA) and Cosine
Similarity that provide basis for the proposed model. The further sections point out the areas of
improvement in the previous work and describe our approach towards the solutions of the same. We then
discuss the implementation details of the project followed by the findings that show the improvements
achieved. Our approach focuses on comprehending the various forms of expressing same entity and
thereby capturing the subjectivity of text into objective parameters. The model is tested by evaluating
answers submitted by 61 students of Third Year B. Tech. CSE class of Walchand College of Engineering
Sangli in a test on Database Engineering.
KEYWORDS
N-gram, Synonym-Set, natural language processing, GLSA, SVD, Wordnet
1.INTRODUCTION
The progress in computer network technologies and increased inclination towards making human
interactions „paperless‟, online test systems are being adopted widely. Automation of the
assessment activity comes to a great help in saving resources (time, efforts & cost) and in
increasing the consistency of the assessment by mitigating potential human biases of intangible
nature. Realization of Computer Aided Assessment (CAA) for objective types of answers is a
well addressed deterministic problem which involves „matching‟ of submitted answers to the
corresponding predefined expected answers. Realization of the CAA in case of descriptive
answers becomes comparatively more challenging by virtue of the subjectivity of the answers. It
involves „mapping‟ of underlying semantics of submitted answers to that of the expected
answers. Pattern recognition, artificial cognition, and natural language processing are required in
such process of CAA.
Significant work has been done in this field. Project Essay Grader (PEG) was developed by Ellis
Page in 1966. PEG ignored the semantic aspect of essays and relied on the surface structures, for
which it had been criticized [1, 2]. Thomas K. Landauer, Boulder & Peter W. Foltz developed
International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015
52
Latent Semantics Analysis (LSA) [1, 2] – the technique widely used in automated assessment.
Group lead by Jill Burstein designed E-rater that deals with automatedessay scoring of GMAT
[3]. Automated Text Maker (ATM) is another computer aided assessment system and can be
adopted to address different disciplines [4]. Md. Monjurul Islam and A. S. M. Latiful Hoque
came up with Automated Essay Scoring Using Generalized Latent Semantic Analysis in 2010 [5]
and also applied GLSA for scoring essays in Bangla Language in 2013 [6]. M. S. Devi and
Himani Mittal applied the LSA Technique to evaluate technical answers in 2013 [7] wherein they
have used single word as atomic unit for computations and suggested in future work that use of
synonymy would enhance the results.
In this paper, automation of assessment of technical answers written by undergraduate students is
further enhanced by modifying GLSA based techniques. Section II outlines the fundamental
model and describes fundamental techniques used. In Section III, we discuss the enhancements
made to the basic model. It includes the use of N-grams as atomic computation unit, capturing
context based synonymy, the concept of Synonym-Sets and change in document representation
and the N-gram formation over Synonym-Sets which enables representation of many by one.
2. THE FUNDAMENTAL MODEL
The existing model of the system for evaluating subjective answers that uses Generalized Latent
Semantic Analysis (GLSA) is discussed in this section. We have referred to it as the fundamental
model.
2.1. GLSA over LSA:
Normally LSA represents documents and their word content in a large two-dimensional matrix
semantic space. Using a matrix algebra technique known as Singular Value Decomposition
(SVD), new relationships between words and documents are uncovered, and existing relationship
are modified to more accurately represent their true significance.[1] Each word represents a row
in the matrix, while each column represents the sentences, paragraphs and other subdivisions of
the context in which the word occurs. The traditional word by document matrix creation of LSA
does not consider word sequence in a document [2]. Here the formation of word by document
matrix the word pair “concurrent transactions” makes the same result of “transactions
concurrent”. Thus, LSA fails to capture the sematic effect of collocations in the document.
GLSA addresses this issue by considering n-gram as atomic unit of the document instead of
individual word. An n-gram is a contiguous sequence of n words from a given sequence of text.
Thus now, “concurrent transactions” is not recognized same as “transactions concurrent”.
2.2. System Architecture Overview:
The system works in two phases – training set generation and testing of newly submitted answers.
Pre-processing and frequency computation operations to be done on an individual document are
common in both the phases.
International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015
53
2.3. Pre-processing:
The pre-processing is comprised of stopwords removal, word-stemming and lingual error
handling in the fundamental model. We have decided to eliminate the step of word-stemming
because we found that dealing with contextual synonymy (discussed in section III) with the
stemmed words as inputs introduces too much of noise and thus inaccuracies.
The pre-processing thus now consists of two steps - stopwords removal and lingual errors
handling.
2.4. Phase 1: Training set generation:
The input for this phase is set of pre-graded answers as per earlier model. We include the
question-domain in the input set for the purpose discussed in section III. Output of this phase is
the training-set.
After pre-processing every pre-graded answer – a training document, n-grams created from all the
training documents and their respective frequencies for corresponding documents are computed
to form the n-gram-by-document vectors. Each such vector represents one training document. The
n-gram-by-document matrix is formed by putting together all such vectors as columns of the
matrix. Every row of the matrix represents an n-gram and every column represents a training-
document. Every cell of this matrix holds the frequency value of the n-gram corresponding to its
row-index in the training document that corresponds to its column index. The n-gram-by-
document matrix then undergoes Singular Value Decomposition (SVD) to generate the singular
value matrices. According to SVD a matrix At×n is decomposed as follows:
At×n = Ut×n × Sn×n × Vt×n
T
…(1)
Here, A is n-gram by documents matrix, U is an orthogonal matrix, S is diagonal matrix and VT
is the transpose of an orthogonal matrix V. After applying dimensionality reduction on matrices
U, S and V we get reduced singular value matrices that are used to generate training vectors. The
dimensionality reduction eliminates one or more least significant singular values in order to
reduce the noise involved in the sematic space developed in the form of SVD matrices. Training
vector corresponding to every pre-graded answer is then formed as per the following equation:
dj
‟ = dj
T
× Ut×k × Sk×k
-1
…(2)
Here, dj
T
is the transpose of document vector dj, Ut×k is truncated left orthogonal matrix and Sk×k is
truncated singular matrix from truncated SVD matrices. The training vectors dj
‘
along with
human grades of pre-graded answers makes the training set.
Phase 2: Evaluation of submitted answers:
The input for this phase is set of submitted answers for corresponding question which will be
evaluated by the system. For the purpose of testing the system, these answers are also evaluated
International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015
54
by the human graders. Output of this phase is the machine-generated grades for the submitted
answers/(s).
Submitted answer first undergoes pre-processing. Frequency values for all the n-grams (generated
in training phase) for this document are computed to form the n-gram-by-document vector q of
the submitted answer. This vector is used to generate the query vector for submitted answer as
follows:
q‟ = q × Ut×k × Sk×k
-1
…(3)
Here, q is query matrix Ut×k is truncated left orthogonal matrix and Sk×k is truncated singular
matrix generated from SVD in training phase. Similarity between query vector q' and training set
vectors dj
‟
is calculated by Cosine similarity [1]
Here, wqj is the jth
weight of query vector q' and dij is the ith
weight of training essay set vectors dj
‟.
The highest correlation value from the Cosine of query vector and the training essay vector is
used for grading the submitted essay.
3. THE ENHANCEMENTS
In the n-gram by document matrix, each row represents an n-gram and each column represents a
document. Every cell contains a value that represents the frequency of n-gram associated with the
row index of that cell in the document associated with the column index of the same cell. We use
the Term Frequency/Inverse Document Frequency (TFIDF) notation to compute this frequency
value.
Now, to capture the significance of a given n-gram, the system depends on the frequency value
associated with it. The subjectivity of the natural language text has an inherent feature which is
generally known as “saying same thing in different words”. For instance, while talking about
networks, somebody may refer to a term, say, „connected nodes‟; whereas, in some other text,
same entity might have been referred to as „joint peers‟ or „connected peers‟ or „joint nodes‟.
Now, all these four expressions represent same entity, given the same context – networks. But, in
the existing model that uses GLSA, all of them are treated as independent bigrams and thus, their
frequencies are computed separately. This introduces inaccuracy in the computed significance of
the given entity if the same entity is expressed in different forms at its different occurrences in the
text. We propose a solution to this problem with the use of Synonym-Set.
International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015
55
3.1. Concept of Synonym-Set & Document redefinition:
For a given n-gram, say, N = (W1 W2 … Wi … Wn)c which represents certain entity, say E in a
given context C, if W11, W12, ... , W1S1 are synonyms representing W1 then, we define Synonym-
Set Sic as
Sic = (Wi, {Wi1, Wi2, ... , WiSi})
The document d is defined as the sequence of words (W1, W2, ... , Wj, ... Wn). We redefine d as
d = (S1c, S2c, ... Sjc, ..., Snc)
Every Synonym-Set Sic is associated with a unique integer say Iic. Therefore the document d is
represented as
d = (I1c, I2c, ... Ijc, ..., Inc)
Now, n-grams are formed over this redefined representation of the document. One such n-gram
now represents a set of all such permutations of expressing the entity represented by that n-gram
which are valid in the given context. In the example stated previously,
C = „
computer networks‟,
S1c = (I1c, {„
connected’, „
joint’}),
S2c = (I2c, {„
node‟, „
peer‟
}),
Nold = { „connected node‟, „connected peer‟, „joint node‟, „joint peer‟}
Nnew = ( I1c I2c)
Thus, a single new n-gram now represents all P permutations of forms in which E can be
expressed in the subjective text in context C, where P is given by:
P = S1 × S2 × ... × Si × ... × Sn
Now, occurrence of any of the P forms of expressing E will be identified by the system as single
entity represented by Nnew. Hence, while computing the significance of any n-gram (and in turn
the entity which it represents) using TFIDF, the weightage of the entity is now converged and put
together under single unique entity unlike earlier models where it was getting scattered among P
¬ways of expressing E. This increases the precision of the system and deals with an inherent
feature of subjectivity – „saying same thing in different words’.
3.2. Contextual Synonymy:
In the solution discussed above, there lies an assumption that a synonym-set will consist of words
having same semantics in a given context, i.e. every word in a given synonym-set shares a
relation of contextual synonymy with the key-word of that synonym-set.
The number of false positive n-grams representing different ways of expressing same entity is
thus inversely proportional to the accuracy of formation of the contextual synonym-sets. For
instance, word „node‟ is related also to the word „knot‟ by the relation of synonymy. But, in given
International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015
56
context of „computer networks‟, „knot‟ does not mean „node‟. Hence, „joint knots‟ cannot be a
valid form of representing the entity „connected nodes‟. So, the implicit challenge in making the
concept of synonym-set and document redefinition functional is to device a process that
accurately forms the contextual synonym-set for every word in the pre-processed training text.
Our method to prepare such synonym-sets uses the database provided by the Wordnet. Along
with inter-relations between words such as synonymy, hyponym, etc., the Wordnet also provides
definitions of every word. We observed that the strings of context-specific definitions in Wordnet
show certain pattern wherein the context or the domain of the definition is specified at the
beginning of the definition in parentheses.
We first fetch all the synonyms of the word under consideration. Then, to filter out only
contextual synonyms, we make use of the aforesaid pattern in the Wordnet definitions and apply
string matching techniques along with the context of the question to which the subjective answers
are written to identify the contexts or the domains of the words which are initially fetched as
synonyms of the key-word.
The presence of domain-specific definition in Wordnet for a given context also serves as a
qualifying criterion for any word to be identified as a terminological, and hence significant, word
in the context of the question. Since, a subjective text of informative nature which talks about
certain domain holds the major part of its semantics in the form of terminological entities, the
process discussed above proves to be effective in dealing with informative answers pertaining to
an identified domain.
3.3. Feedback-driven incremental advancement:
We attempt to incorporate an aspect of human behavior into our model by using a feedback
system. The objective of this feedback system is to make the system avoid same mistakes twice.
The system can make mistake in grading an answer which has content semantically or
syntactically „unknown’ to the system, i.e., it contains large number of new entities which were
not part of the training set at all. Such cases are identified as outliers and the grades assigned to
them are forwarded for human moderation. Identifying potential outliers is important for the
System’s reliability. To identify outliers from a given set of submitted answers which is normally
distributed in terms of probability of presence of outliers, we use the characteristic of an outlier
by virtue of which, even the highest amongst the cosine correlations of the query vector of outlier
with the training vectors has comparatively low value than the highest of cosine correlations of an
average case test answer (graded with acceptable reliability) with the training vectors. Using this
property, we identify an outlier as the answer in case of which, the highest cosine correlation of
the query vector with the training vectors is below the average of such highest cosine correlations
of query vectors of all answers in the given set.
We feedback every such outlier along with its human-moderated grade to the system for its
inclusion in the training set. This leads to addition of new n-grams and re-computation of the
SVD matrices which in turn modifies the training set.
Figure1 shows the overview of our system model.
International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015
57
Figure 1. System overview
4. EXPERIMENT AND RESULTS
We have experimented on our model by conducting an online test on Database Engineering
course for a class of 61 students in our departmental lab. We collected the human grades for each
answer from the concerned faculty. Further, 30 answers were used for training the system and rest
for testing. Following (Fig. 2) is the graph comparing the machine generated grades and
corresponding average human grades for each test answer. Three human graders have graded the
answer and we have used the average of their scores. These results are obtained for value of
n=1,2,3.
Figure 2 Graph: Machine grade Vs Human grade
Table 1. Variation in Training Time, Testing Times and Standard Deviation with various
combinations of n-grams
International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015
58
T1 : Training time in minutes with 30 answers
T2 : Testing time in minutes with 30 answers
Table 1 shows the comparison of various values of n and the time taken for training 30 answers,
testing answers, number of n-grams and the standard deviation.
Figure 3 Graph: Variation in Standard Deviation
Graph in Figure 3 shows that when we increase the value of n in n-gram we get improved results.
The standard deviation approached zero as we increased n. However, we observed that if we try
to improve the results with 4-grams then it has a reverse effect. This is because average number
of words in a sentence (for third year undergraduate students) is about 6-7 after stop word
removal. So when we have 4-grams then we are actually trying to find phrases as long as four
words. This introduces noise in the result. This is the major reason for limiting the value of n to 3
International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015
59
Figure 4 Graph: Variation in Training Time with N
Graph in Figure 4 shows that the training time increases with number of n-grams, which is
dependent on the number of training documents. Thus, we can say that the number of training
documents determines the time taken for training.
As a part of reliability measure we compute the error with respect to human grades. When error
goes beyond a threshold we add the outliers to SVD and recomputed the matrix to improve the
results.
5. CONCLUSIONS
We have developed a system that focuses on the concept of synonyms while it grades an answer.
This enhancement has given significant results in comparisons with the recently proposed models
that use latent semantic analysis without incorporating synonyms. Section 2 is the basic model
that currently is considered successful for grading answers. Section 3 introduces a new aspect due
to which the model can perform better. Implementation and results have proved that the
enhancement costs negligible resources in terms of time and space. The improvement in
performance definitely outweighs the negligible costs. Section 4 shows the experimental results
taken on collected data. The feedback driven model improves the results every time we train the
system. This regulates error periodically and also improves the results after each time we train.
ACKNOWLEDGEMENTS
The authors would like to thank firstly the Department of Computer Science and Engineering,
Walchand College of Engineering, Sangli for facilitating the experimentation by providing
required infrastructure. The authors also thank in particular the esteemed faculty members of
Dept. of CSE, WCE, Sangli Mrs. Hetal Gandhi and Dr. Smriti Bhandari, for their valuable
assistance in human expert grading of test answers used in training and testing of the system.
Lastly the authors thank all the students of 2013-14‟s Third Year B.Tech. CSE class of WCE,
Sangli who participated in the experiments as test takers. Without their sincere voluntary
participation, the experiments could not have proceeded.
International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015
60
REFERENCES
[1] Md. Monjurul Islam and A. S. M. Latiful Hoque, “Automated Essay Scoring Using Generalized
Latent Semantic Analysis,” in Proceedings of 13th International Conference on Computer and
Information Technology (ICCIT 2010), 23-25 December, Dhaka, Bangladesh
[2] M. Syamala Devi and Himani Mittal, “Subjective Evaluation using LSA Technique,” in International
Journal of Computers and Distributed Systems Vol. No.3, Issue I, April-May 2013 ISSN: 2278-5183
(www.ijcdsonline.com)
[3] S. Valenti, F. Neri, and A. Cucchiarelli, “An overview of current research on automated essay
grading,” in Journal of Information Technology Education, vol. 2, 2003, pp. 319-330
[4] Md. Monjurul Islam and A. S. M. Latiful Hoque, “Automated Bangla Essay Scoring System:
ABESS,” in 978-1-4799-0400-6/13/$31.00 ©2013 IEEE
[5] “Java API for WordNet Searching (JAWS)” [Online] Available:
http://lyle.smu.edu/~tspell/jaws/index.html
[6] Mark Watson, “Practical Artificial Intelligence Programming with Java”, Ed. 3, 11/11/2008 [Online]
Available:
http://archive.org/stream/PracticalArtificialIntelligenceProgrammingWithJava/JavaAI3rd_djvu.txt
[7] JAMA: A Java Matrix Package [Online]. Available: http://math.nist.gov/javanumerics/jama/
Authors
Parag A. Guruji
Data Scientist at Zlemma Analytics Inc, Pune, India.
Former student of Walchand College of Engineering, Sangli, India.
Received Bachelor of Technology Degree in Computer Science and Engineering in May 2014.
Team Member of the project discussed in this paper.
Mrunal M. Pagnis
Graduate Student at Indiana University at Bloomington, Indiana, USA.
Former student of Walchand College of Engineering, Sangli, India.
Received Bachelor of Technology Degree in Computer Science and Engineering in May 2014.
Team Member of the project discussed in this paper.
Sayali M. Pawar
Graduate Engineer Trainee at Tata Consultancy Services, Pune, India.
Former student of Walchand College of Engineering, Sangli, India.
Received Bachelor of Technology Degree in Computer Science and Engineering in May 2014.
Team Member of the project discussed in this paper.
Dr. Prakash J. Kulkarni
Professor in Computer Science and Engineering and Deputy Director at Walchand College of Engineering,
Sangli, India.
Faculty Mentor of the project discussed in this paper.

More Related Content

What's hot

Current Approaches in Search Result Diversification
Current Approaches in Search Result DiversificationCurrent Approaches in Search Result Diversification
Current Approaches in Search Result DiversificationMario Sangiorgio
 
Basic review on topic modeling
Basic review on  topic modelingBasic review on  topic modeling
Basic review on topic modelingHiroyuki Kuromiya
 
Deep Reinforcement Learning with Distributional Semantic Rewards for Abstract...
Deep Reinforcement Learning with Distributional Semantic Rewards for Abstract...Deep Reinforcement Learning with Distributional Semantic Rewards for Abstract...
Deep Reinforcement Learning with Distributional Semantic Rewards for Abstract...Deren Lei
 
Different Similarity Measures for Text Classification Using Knn
Different Similarity Measures for Text Classification Using KnnDifferent Similarity Measures for Text Classification Using Knn
Different Similarity Measures for Text Classification Using KnnIOSR Journals
 
Machine Learning Techniques with Ontology for Subjective Answer Evaluation
Machine Learning Techniques with Ontology for Subjective Answer EvaluationMachine Learning Techniques with Ontology for Subjective Answer Evaluation
Machine Learning Techniques with Ontology for Subjective Answer Evaluationijnlc
 
Supervised WSD Using Master- Slave Voting Technique
Supervised WSD Using Master- Slave Voting TechniqueSupervised WSD Using Master- Slave Voting Technique
Supervised WSD Using Master- Slave Voting Techniqueiosrjce
 
EXPERT OPINION AND COHERENCE BASED TOPIC MODELING
EXPERT OPINION AND COHERENCE BASED TOPIC MODELINGEXPERT OPINION AND COHERENCE BASED TOPIC MODELING
EXPERT OPINION AND COHERENCE BASED TOPIC MODELINGijnlc
 
From_seq2seq_to_BERT
From_seq2seq_to_BERTFrom_seq2seq_to_BERT
From_seq2seq_to_BERTHuali Zhao
 
Decision-Making Model for Student Assessment by Unifying Numerical and Lingui...
Decision-Making Model for Student Assessment by Unifying Numerical and Lingui...Decision-Making Model for Student Assessment by Unifying Numerical and Lingui...
Decision-Making Model for Student Assessment by Unifying Numerical and Lingui...IJECEIAES
 
Neural Network in Knowledge Bases
Neural Network in Knowledge BasesNeural Network in Knowledge Bases
Neural Network in Knowledge BasesKushal Arora
 
Reasoning Over Knowledge Base
Reasoning Over Knowledge BaseReasoning Over Knowledge Base
Reasoning Over Knowledge BaseShubham Agarwal
 
Generation of Question and Answer from Unstructured Document using Gaussian M...
Generation of Question and Answer from Unstructured Document using Gaussian M...Generation of Question and Answer from Unstructured Document using Gaussian M...
Generation of Question and Answer from Unstructured Document using Gaussian M...IJACEE IJACEE
 
Modeling Text Independent Speaker Identification with Vector Quantization
Modeling Text Independent Speaker Identification with Vector QuantizationModeling Text Independent Speaker Identification with Vector Quantization
Modeling Text Independent Speaker Identification with Vector QuantizationTELKOMNIKA JOURNAL
 
[UMAP 2015] Integrating Context Similarity with Sparse Linear Recommendation ...
[UMAP 2015] Integrating Context Similarity with Sparse Linear Recommendation ...[UMAP 2015] Integrating Context Similarity with Sparse Linear Recommendation ...
[UMAP 2015] Integrating Context Similarity with Sparse Linear Recommendation ...YONG ZHENG
 
Bert pre_training_of_deep_bidirectional_transformers_for_language_understanding
Bert  pre_training_of_deep_bidirectional_transformers_for_language_understandingBert  pre_training_of_deep_bidirectional_transformers_for_language_understanding
Bert pre_training_of_deep_bidirectional_transformers_for_language_understandingThyrixYang1
 
Implicit regularization of SGD in NLP
Implicit regularization of SGD in NLPImplicit regularization of SGD in NLP
Implicit regularization of SGD in NLPDeren Lei
 

What's hot (18)

Current Approaches in Search Result Diversification
Current Approaches in Search Result DiversificationCurrent Approaches in Search Result Diversification
Current Approaches in Search Result Diversification
 
Basic review on topic modeling
Basic review on  topic modelingBasic review on  topic modeling
Basic review on topic modeling
 
J045055560
J045055560J045055560
J045055560
 
Deep Reinforcement Learning with Distributional Semantic Rewards for Abstract...
Deep Reinforcement Learning with Distributional Semantic Rewards for Abstract...Deep Reinforcement Learning with Distributional Semantic Rewards for Abstract...
Deep Reinforcement Learning with Distributional Semantic Rewards for Abstract...
 
Different Similarity Measures for Text Classification Using Knn
Different Similarity Measures for Text Classification Using KnnDifferent Similarity Measures for Text Classification Using Knn
Different Similarity Measures for Text Classification Using Knn
 
Machine Learning Techniques with Ontology for Subjective Answer Evaluation
Machine Learning Techniques with Ontology for Subjective Answer EvaluationMachine Learning Techniques with Ontology for Subjective Answer Evaluation
Machine Learning Techniques with Ontology for Subjective Answer Evaluation
 
Supervised WSD Using Master- Slave Voting Technique
Supervised WSD Using Master- Slave Voting TechniqueSupervised WSD Using Master- Slave Voting Technique
Supervised WSD Using Master- Slave Voting Technique
 
Nn kb
Nn kbNn kb
Nn kb
 
EXPERT OPINION AND COHERENCE BASED TOPIC MODELING
EXPERT OPINION AND COHERENCE BASED TOPIC MODELINGEXPERT OPINION AND COHERENCE BASED TOPIC MODELING
EXPERT OPINION AND COHERENCE BASED TOPIC MODELING
 
From_seq2seq_to_BERT
From_seq2seq_to_BERTFrom_seq2seq_to_BERT
From_seq2seq_to_BERT
 
Decision-Making Model for Student Assessment by Unifying Numerical and Lingui...
Decision-Making Model for Student Assessment by Unifying Numerical and Lingui...Decision-Making Model for Student Assessment by Unifying Numerical and Lingui...
Decision-Making Model for Student Assessment by Unifying Numerical and Lingui...
 
Neural Network in Knowledge Bases
Neural Network in Knowledge BasesNeural Network in Knowledge Bases
Neural Network in Knowledge Bases
 
Reasoning Over Knowledge Base
Reasoning Over Knowledge BaseReasoning Over Knowledge Base
Reasoning Over Knowledge Base
 
Generation of Question and Answer from Unstructured Document using Gaussian M...
Generation of Question and Answer from Unstructured Document using Gaussian M...Generation of Question and Answer from Unstructured Document using Gaussian M...
Generation of Question and Answer from Unstructured Document using Gaussian M...
 
Modeling Text Independent Speaker Identification with Vector Quantization
Modeling Text Independent Speaker Identification with Vector QuantizationModeling Text Independent Speaker Identification with Vector Quantization
Modeling Text Independent Speaker Identification with Vector Quantization
 
[UMAP 2015] Integrating Context Similarity with Sparse Linear Recommendation ...
[UMAP 2015] Integrating Context Similarity with Sparse Linear Recommendation ...[UMAP 2015] Integrating Context Similarity with Sparse Linear Recommendation ...
[UMAP 2015] Integrating Context Similarity with Sparse Linear Recommendation ...
 
Bert pre_training_of_deep_bidirectional_transformers_for_language_understanding
Bert  pre_training_of_deep_bidirectional_transformers_for_language_understandingBert  pre_training_of_deep_bidirectional_transformers_for_language_understanding
Bert pre_training_of_deep_bidirectional_transformers_for_language_understanding
 
Implicit regularization of SGD in NLP
Implicit regularization of SGD in NLPImplicit regularization of SGD in NLP
Implicit regularization of SGD in NLP
 

Viewers also liked

S URVEY O N M ACHINE T RANSLITERATION A ND M ACHINE L EARNING M ODELS
S URVEY  O N M ACHINE  T RANSLITERATION A ND  M ACHINE L EARNING M ODELSS URVEY  O N M ACHINE  T RANSLITERATION A ND  M ACHINE L EARNING M ODELS
S URVEY O N M ACHINE T RANSLITERATION A ND M ACHINE L EARNING M ODELSijnlc
 
An exhaustive font and size invariant classification scheme for ocr of devana...
An exhaustive font and size invariant classification scheme for ocr of devana...An exhaustive font and size invariant classification scheme for ocr of devana...
An exhaustive font and size invariant classification scheme for ocr of devana...ijnlc
 
S ENTIMENT A NALYSIS F OR M ODERN S TANDARD A RABIC A ND C OLLOQUIAl
S ENTIMENT A NALYSIS  F OR M ODERN S TANDARD  A RABIC  A ND  C OLLOQUIAlS ENTIMENT A NALYSIS  F OR M ODERN S TANDARD  A RABIC  A ND  C OLLOQUIAl
S ENTIMENT A NALYSIS F OR M ODERN S TANDARD A RABIC A ND C OLLOQUIAlijnlc
 
Conceptual framework for abstractive text summarization
Conceptual framework for abstractive text summarizationConceptual framework for abstractive text summarization
Conceptual framework for abstractive text summarizationijnlc
 
IMPLEMENTATION OF NLIZATION FRAMEWORK FOR VERBS, PRONOUNS AND DETERMINERS WIT...
IMPLEMENTATION OF NLIZATION FRAMEWORK FOR VERBS, PRONOUNS AND DETERMINERS WIT...IMPLEMENTATION OF NLIZATION FRAMEWORK FOR VERBS, PRONOUNS AND DETERMINERS WIT...
IMPLEMENTATION OF NLIZATION FRAMEWORK FOR VERBS, PRONOUNS AND DETERMINERS WIT...ijnlc
 
An expert system for automatic reading of a text written in standard arabic
An expert system for automatic reading of a text written in standard arabicAn expert system for automatic reading of a text written in standard arabic
An expert system for automatic reading of a text written in standard arabicijnlc
 
Hybrid part of-speech tagger for non-vocalized arabic text
Hybrid part of-speech tagger for non-vocalized arabic textHybrid part of-speech tagger for non-vocalized arabic text
Hybrid part of-speech tagger for non-vocalized arabic textijnlc
 
Verb based manipuri sentiment analysis
Verb based manipuri sentiment analysisVerb based manipuri sentiment analysis
Verb based manipuri sentiment analysisijnlc
 
HANDLING UNKNOWN WORDS IN NAMED ENTITY RECOGNITION USING TRANSLITERATION
HANDLING UNKNOWN WORDS IN NAMED ENTITY RECOGNITION USING TRANSLITERATIONHANDLING UNKNOWN WORDS IN NAMED ENTITY RECOGNITION USING TRANSLITERATION
HANDLING UNKNOWN WORDS IN NAMED ENTITY RECOGNITION USING TRANSLITERATIONijnlc
 
A MULTI-STREAM HMM APPROACH TO OFFLINE HANDWRITTEN ARABIC WORD RECOGNITION
A MULTI-STREAM HMM APPROACH TO OFFLINE HANDWRITTEN ARABIC WORD RECOGNITIONA MULTI-STREAM HMM APPROACH TO OFFLINE HANDWRITTEN ARABIC WORD RECOGNITION
A MULTI-STREAM HMM APPROACH TO OFFLINE HANDWRITTEN ARABIC WORD RECOGNITIONijnlc
 
CLUSTERING WEB SEARCH RESULTS FOR EFFECTIVE ARABIC LANGUAGE BROWSING
CLUSTERING WEB SEARCH RESULTS FOR EFFECTIVE ARABIC LANGUAGE BROWSINGCLUSTERING WEB SEARCH RESULTS FOR EFFECTIVE ARABIC LANGUAGE BROWSING
CLUSTERING WEB SEARCH RESULTS FOR EFFECTIVE ARABIC LANGUAGE BROWSINGijnlc
 
Building a vietnamese dialog mechanism for v dlg~tabl system
Building a vietnamese dialog mechanism for v dlg~tabl systemBuilding a vietnamese dialog mechanism for v dlg~tabl system
Building a vietnamese dialog mechanism for v dlg~tabl systemijnlc
 
Smart grammar a dynamic spoken language understanding grammar for inflective ...
Smart grammar a dynamic spoken language understanding grammar for inflective ...Smart grammar a dynamic spoken language understanding grammar for inflective ...
Smart grammar a dynamic spoken language understanding grammar for inflective ...ijnlc
 
Chunking in manipuri using crf
Chunking in manipuri using crfChunking in manipuri using crf
Chunking in manipuri using crfijnlc
 
Developemnt and evaluation of a web based question answering system for arabi...
Developemnt and evaluation of a web based question answering system for arabi...Developemnt and evaluation of a web based question answering system for arabi...
Developemnt and evaluation of a web based question answering system for arabi...ijnlc
 
NERHMM: A TOOL FOR NAMED ENTITY RECOGNITION BASED ON HIDDEN MARKOV MODEL
NERHMM: A TOOL FOR NAMED ENTITY RECOGNITION BASED ON HIDDEN MARKOV MODELNERHMM: A TOOL FOR NAMED ENTITY RECOGNITION BASED ON HIDDEN MARKOV MODEL
NERHMM: A TOOL FOR NAMED ENTITY RECOGNITION BASED ON HIDDEN MARKOV MODELijnlc
 
An improved apriori algorithm for association rules
An improved apriori algorithm for association rulesAn improved apriori algorithm for association rules
An improved apriori algorithm for association rulesijnlc
 
A comparative analysis of particle swarm optimization and k means algorithm f...
A comparative analysis of particle swarm optimization and k means algorithm f...A comparative analysis of particle swarm optimization and k means algorithm f...
A comparative analysis of particle swarm optimization and k means algorithm f...ijnlc
 
IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...
IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...
IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...ijnlc
 
G2 pil a grapheme to-phoneme conversion tool for the italian language
G2 pil a grapheme to-phoneme conversion tool for the italian languageG2 pil a grapheme to-phoneme conversion tool for the italian language
G2 pil a grapheme to-phoneme conversion tool for the italian languageijnlc
 

Viewers also liked (20)

S URVEY O N M ACHINE T RANSLITERATION A ND M ACHINE L EARNING M ODELS
S URVEY  O N M ACHINE  T RANSLITERATION A ND  M ACHINE L EARNING M ODELSS URVEY  O N M ACHINE  T RANSLITERATION A ND  M ACHINE L EARNING M ODELS
S URVEY O N M ACHINE T RANSLITERATION A ND M ACHINE L EARNING M ODELS
 
An exhaustive font and size invariant classification scheme for ocr of devana...
An exhaustive font and size invariant classification scheme for ocr of devana...An exhaustive font and size invariant classification scheme for ocr of devana...
An exhaustive font and size invariant classification scheme for ocr of devana...
 
S ENTIMENT A NALYSIS F OR M ODERN S TANDARD A RABIC A ND C OLLOQUIAl
S ENTIMENT A NALYSIS  F OR M ODERN S TANDARD  A RABIC  A ND  C OLLOQUIAlS ENTIMENT A NALYSIS  F OR M ODERN S TANDARD  A RABIC  A ND  C OLLOQUIAl
S ENTIMENT A NALYSIS F OR M ODERN S TANDARD A RABIC A ND C OLLOQUIAl
 
Conceptual framework for abstractive text summarization
Conceptual framework for abstractive text summarizationConceptual framework for abstractive text summarization
Conceptual framework for abstractive text summarization
 
IMPLEMENTATION OF NLIZATION FRAMEWORK FOR VERBS, PRONOUNS AND DETERMINERS WIT...
IMPLEMENTATION OF NLIZATION FRAMEWORK FOR VERBS, PRONOUNS AND DETERMINERS WIT...IMPLEMENTATION OF NLIZATION FRAMEWORK FOR VERBS, PRONOUNS AND DETERMINERS WIT...
IMPLEMENTATION OF NLIZATION FRAMEWORK FOR VERBS, PRONOUNS AND DETERMINERS WIT...
 
An expert system for automatic reading of a text written in standard arabic
An expert system for automatic reading of a text written in standard arabicAn expert system for automatic reading of a text written in standard arabic
An expert system for automatic reading of a text written in standard arabic
 
Hybrid part of-speech tagger for non-vocalized arabic text
Hybrid part of-speech tagger for non-vocalized arabic textHybrid part of-speech tagger for non-vocalized arabic text
Hybrid part of-speech tagger for non-vocalized arabic text
 
Verb based manipuri sentiment analysis
Verb based manipuri sentiment analysisVerb based manipuri sentiment analysis
Verb based manipuri sentiment analysis
 
HANDLING UNKNOWN WORDS IN NAMED ENTITY RECOGNITION USING TRANSLITERATION
HANDLING UNKNOWN WORDS IN NAMED ENTITY RECOGNITION USING TRANSLITERATIONHANDLING UNKNOWN WORDS IN NAMED ENTITY RECOGNITION USING TRANSLITERATION
HANDLING UNKNOWN WORDS IN NAMED ENTITY RECOGNITION USING TRANSLITERATION
 
A MULTI-STREAM HMM APPROACH TO OFFLINE HANDWRITTEN ARABIC WORD RECOGNITION
A MULTI-STREAM HMM APPROACH TO OFFLINE HANDWRITTEN ARABIC WORD RECOGNITIONA MULTI-STREAM HMM APPROACH TO OFFLINE HANDWRITTEN ARABIC WORD RECOGNITION
A MULTI-STREAM HMM APPROACH TO OFFLINE HANDWRITTEN ARABIC WORD RECOGNITION
 
CLUSTERING WEB SEARCH RESULTS FOR EFFECTIVE ARABIC LANGUAGE BROWSING
CLUSTERING WEB SEARCH RESULTS FOR EFFECTIVE ARABIC LANGUAGE BROWSINGCLUSTERING WEB SEARCH RESULTS FOR EFFECTIVE ARABIC LANGUAGE BROWSING
CLUSTERING WEB SEARCH RESULTS FOR EFFECTIVE ARABIC LANGUAGE BROWSING
 
Building a vietnamese dialog mechanism for v dlg~tabl system
Building a vietnamese dialog mechanism for v dlg~tabl systemBuilding a vietnamese dialog mechanism for v dlg~tabl system
Building a vietnamese dialog mechanism for v dlg~tabl system
 
Smart grammar a dynamic spoken language understanding grammar for inflective ...
Smart grammar a dynamic spoken language understanding grammar for inflective ...Smart grammar a dynamic spoken language understanding grammar for inflective ...
Smart grammar a dynamic spoken language understanding grammar for inflective ...
 
Chunking in manipuri using crf
Chunking in manipuri using crfChunking in manipuri using crf
Chunking in manipuri using crf
 
Developemnt and evaluation of a web based question answering system for arabi...
Developemnt and evaluation of a web based question answering system for arabi...Developemnt and evaluation of a web based question answering system for arabi...
Developemnt and evaluation of a web based question answering system for arabi...
 
NERHMM: A TOOL FOR NAMED ENTITY RECOGNITION BASED ON HIDDEN MARKOV MODEL
NERHMM: A TOOL FOR NAMED ENTITY RECOGNITION BASED ON HIDDEN MARKOV MODELNERHMM: A TOOL FOR NAMED ENTITY RECOGNITION BASED ON HIDDEN MARKOV MODEL
NERHMM: A TOOL FOR NAMED ENTITY RECOGNITION BASED ON HIDDEN MARKOV MODEL
 
An improved apriori algorithm for association rules
An improved apriori algorithm for association rulesAn improved apriori algorithm for association rules
An improved apriori algorithm for association rules
 
A comparative analysis of particle swarm optimization and k means algorithm f...
A comparative analysis of particle swarm optimization and k means algorithm f...A comparative analysis of particle swarm optimization and k means algorithm f...
A comparative analysis of particle swarm optimization and k means algorithm f...
 
IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...
IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...
IMPROVING THE QUALITY OF GUJARATI-HINDI MACHINE TRANSLATION THROUGH PART-OF-S...
 
G2 pil a grapheme to-phoneme conversion tool for the italian language
G2 pil a grapheme to-phoneme conversion tool for the italian languageG2 pil a grapheme to-phoneme conversion tool for the italian language
G2 pil a grapheme to-phoneme conversion tool for the italian language
 

Similar to Automated Evaluation of Subjective Answers Using Contextual Synonyms

Automated Essay Scoring Using Generalized Latent Semantic Analysis
Automated Essay Scoring Using Generalized Latent Semantic AnalysisAutomated Essay Scoring Using Generalized Latent Semantic Analysis
Automated Essay Scoring Using Generalized Latent Semantic AnalysisGina Rizzo
 
Doc format.
Doc format.Doc format.
Doc format.butest
 
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali TextChunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali Textkevig
 
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali TextChunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali Textkevig
 
Subjective evaluation answer ppt
Subjective evaluation answer pptSubjective evaluation answer ppt
Subjective evaluation answer pptMrunal Pagnis
 
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESTHE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESkevig
 
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESTHE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESkevig
 
Question Classification using Semantic, Syntactic and Lexical features
Question Classification using Semantic, Syntactic and Lexical featuresQuestion Classification using Semantic, Syntactic and Lexical features
Question Classification using Semantic, Syntactic and Lexical featuresdannyijwest
 
20433-39028-3-PB.pdf
20433-39028-3-PB.pdf20433-39028-3-PB.pdf
20433-39028-3-PB.pdfIjictTeam
 
Max stable set problem to found the initial centroids in clustering problem
Max stable set problem to found the initial centroids in clustering problemMax stable set problem to found the initial centroids in clustering problem
Max stable set problem to found the initial centroids in clustering problemnooriasukmaningtyas
 
Centralized Class Specific Dictionary Learning for Wearable Sensors based phy...
Centralized Class Specific Dictionary Learning for Wearable Sensors based phy...Centralized Class Specific Dictionary Learning for Wearable Sensors based phy...
Centralized Class Specific Dictionary Learning for Wearable Sensors based phy...sherinmm
 
Discovering Novel Information with sentence Level clustering From Multi-docu...
Discovering Novel Information with sentence Level clustering  From Multi-docu...Discovering Novel Information with sentence Level clustering  From Multi-docu...
Discovering Novel Information with sentence Level clustering From Multi-docu...irjes
 
T EXT M INING AND C LASSIFICATION OF P RODUCT R EVIEWS U SING S TRUCTURED S U...
T EXT M INING AND C LASSIFICATION OF P RODUCT R EVIEWS U SING S TRUCTURED S U...T EXT M INING AND C LASSIFICATION OF P RODUCT R EVIEWS U SING S TRUCTURED S U...
T EXT M INING AND C LASSIFICATION OF P RODUCT R EVIEWS U SING S TRUCTURED S U...csandit
 
Taxonomy extraction from automotive natural language requirements using unsup...
Taxonomy extraction from automotive natural language requirements using unsup...Taxonomy extraction from automotive natural language requirements using unsup...
Taxonomy extraction from automotive natural language requirements using unsup...ijnlc
 
Novel text categorization by amalgamation of augmented k nearest neighbourhoo...
Novel text categorization by amalgamation of augmented k nearest neighbourhoo...Novel text categorization by amalgamation of augmented k nearest neighbourhoo...
Novel text categorization by amalgamation of augmented k nearest neighbourhoo...ijcsity
 
Multi label classification of
Multi label classification ofMulti label classification of
Multi label classification ofijaia
 
A ROBUST JOINT-TRAINING GRAPHNEURALNETWORKS MODEL FOR EVENT DETECTIONWITHSYMM...
A ROBUST JOINT-TRAINING GRAPHNEURALNETWORKS MODEL FOR EVENT DETECTIONWITHSYMM...A ROBUST JOINT-TRAINING GRAPHNEURALNETWORKS MODEL FOR EVENT DETECTIONWITHSYMM...
A ROBUST JOINT-TRAINING GRAPHNEURALNETWORKS MODEL FOR EVENT DETECTIONWITHSYMM...kevig
 
A Robust Joint-Training Graph Neural Networks Model for Event Detection with ...
A Robust Joint-Training Graph Neural Networks Model for Event Detection with ...A Robust Joint-Training Graph Neural Networks Model for Event Detection with ...
A Robust Joint-Training Graph Neural Networks Model for Event Detection with ...kevig
 
Audit report[rollno 49]
Audit report[rollno 49]Audit report[rollno 49]
Audit report[rollno 49]RAHULROHAM2
 

Similar to Automated Evaluation of Subjective Answers Using Contextual Synonyms (20)

Automated Essay Scoring Using Generalized Latent Semantic Analysis
Automated Essay Scoring Using Generalized Latent Semantic AnalysisAutomated Essay Scoring Using Generalized Latent Semantic Analysis
Automated Essay Scoring Using Generalized Latent Semantic Analysis
 
Doc format.
Doc format.Doc format.
Doc format.
 
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali TextChunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
 
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali TextChunker Based Sentiment Analysis and Tense Classification for Nepali Text
Chunker Based Sentiment Analysis and Tense Classification for Nepali Text
 
Subjective evaluation answer ppt
Subjective evaluation answer pptSubjective evaluation answer ppt
Subjective evaluation answer ppt
 
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESTHE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
 
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIESTHE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
THE ABILITY OF WORD EMBEDDINGS TO CAPTURE WORD SIMILARITIES
 
Question Classification using Semantic, Syntactic and Lexical features
Question Classification using Semantic, Syntactic and Lexical featuresQuestion Classification using Semantic, Syntactic and Lexical features
Question Classification using Semantic, Syntactic and Lexical features
 
20433-39028-3-PB.pdf
20433-39028-3-PB.pdf20433-39028-3-PB.pdf
20433-39028-3-PB.pdf
 
Max stable set problem to found the initial centroids in clustering problem
Max stable set problem to found the initial centroids in clustering problemMax stable set problem to found the initial centroids in clustering problem
Max stable set problem to found the initial centroids in clustering problem
 
Centralized Class Specific Dictionary Learning for Wearable Sensors based phy...
Centralized Class Specific Dictionary Learning for Wearable Sensors based phy...Centralized Class Specific Dictionary Learning for Wearable Sensors based phy...
Centralized Class Specific Dictionary Learning for Wearable Sensors based phy...
 
Discovering Novel Information with sentence Level clustering From Multi-docu...
Discovering Novel Information with sentence Level clustering  From Multi-docu...Discovering Novel Information with sentence Level clustering  From Multi-docu...
Discovering Novel Information with sentence Level clustering From Multi-docu...
 
T EXT M INING AND C LASSIFICATION OF P RODUCT R EVIEWS U SING S TRUCTURED S U...
T EXT M INING AND C LASSIFICATION OF P RODUCT R EVIEWS U SING S TRUCTURED S U...T EXT M INING AND C LASSIFICATION OF P RODUCT R EVIEWS U SING S TRUCTURED S U...
T EXT M INING AND C LASSIFICATION OF P RODUCT R EVIEWS U SING S TRUCTURED S U...
 
Taxonomy extraction from automotive natural language requirements using unsup...
Taxonomy extraction from automotive natural language requirements using unsup...Taxonomy extraction from automotive natural language requirements using unsup...
Taxonomy extraction from automotive natural language requirements using unsup...
 
Novel text categorization by amalgamation of augmented k nearest neighbourhoo...
Novel text categorization by amalgamation of augmented k nearest neighbourhoo...Novel text categorization by amalgamation of augmented k nearest neighbourhoo...
Novel text categorization by amalgamation of augmented k nearest neighbourhoo...
 
call for papers, research paper publishing, where to publish research paper, ...
call for papers, research paper publishing, where to publish research paper, ...call for papers, research paper publishing, where to publish research paper, ...
call for papers, research paper publishing, where to publish research paper, ...
 
Multi label classification of
Multi label classification ofMulti label classification of
Multi label classification of
 
A ROBUST JOINT-TRAINING GRAPHNEURALNETWORKS MODEL FOR EVENT DETECTIONWITHSYMM...
A ROBUST JOINT-TRAINING GRAPHNEURALNETWORKS MODEL FOR EVENT DETECTIONWITHSYMM...A ROBUST JOINT-TRAINING GRAPHNEURALNETWORKS MODEL FOR EVENT DETECTIONWITHSYMM...
A ROBUST JOINT-TRAINING GRAPHNEURALNETWORKS MODEL FOR EVENT DETECTIONWITHSYMM...
 
A Robust Joint-Training Graph Neural Networks Model for Event Detection with ...
A Robust Joint-Training Graph Neural Networks Model for Event Detection with ...A Robust Joint-Training Graph Neural Networks Model for Event Detection with ...
A Robust Joint-Training Graph Neural Networks Model for Event Detection with ...
 
Audit report[rollno 49]
Audit report[rollno 49]Audit report[rollno 49]
Audit report[rollno 49]
 

Recently uploaded

Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machinePadma Pradeep
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024BookNet Canada
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr BaganFwdays
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Commit University
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Scott Keck-Warren
 
APIForce Zurich 5 April Automation LPDG
APIForce Zurich 5 April  Automation LPDGAPIForce Zurich 5 April  Automation LPDG
APIForce Zurich 5 April Automation LPDGMarianaLemus7
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Enterprise Knowledge
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek SchlawackFwdays
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024Lorenzo Miniero
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsSergiu Bodiu
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piececharlottematthew16
 
costume and set research powerpoint presentation
costume and set research powerpoint presentationcostume and set research powerpoint presentation
costume and set research powerpoint presentationphoebematthew05
 
Pigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions
 
"Debugging python applications inside k8s environment", Andrii Soldatenko
"Debugging python applications inside k8s environment", Andrii Soldatenko"Debugging python applications inside k8s environment", Andrii Soldatenko
"Debugging python applications inside k8s environment", Andrii SoldatenkoFwdays
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024Scott Keck-Warren
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLScyllaDB
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationRidwan Fadjar
 

Recently uploaded (20)

Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machine
 
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
New from BookNet Canada for 2024: BNC CataList - Tech Forum 2024
 
"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan"ML in Production",Oleksandr Bagan
"ML in Production",Oleksandr Bagan
 
Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!Nell’iperspazio con Rocket: il Framework Web di Rust!
Nell’iperspazio con Rocket: il Framework Web di Rust!
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024
 
APIForce Zurich 5 April Automation LPDG
APIForce Zurich 5 April  Automation LPDGAPIForce Zurich 5 April  Automation LPDG
APIForce Zurich 5 April Automation LPDG
 
DMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special EditionDMCC Future of Trade Web3 - Special Edition
DMCC Future of Trade Web3 - Special Edition
 
Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024Designing IA for AI - Information Architecture Conference 2024
Designing IA for AI - Information Architecture Conference 2024
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
"Subclassing and Composition – A Pythonic Tour of Trade-Offs", Hynek Schlawack
 
SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024SIP trunking in Janus @ Kamailio World 2024
SIP trunking in Janus @ Kamailio World 2024
 
DevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platformsDevEX - reference for building teams, processes, and platforms
DevEX - reference for building teams, processes, and platforms
 
Story boards and shot lists for my a level piece
Story boards and shot lists for my a level pieceStory boards and shot lists for my a level piece
Story boards and shot lists for my a level piece
 
Hot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort Service
Hot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort ServiceHot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort Service
Hot Sexy call girls in Panjabi Bagh 🔝 9953056974 🔝 Delhi escort Service
 
costume and set research powerpoint presentation
costume and set research powerpoint presentationcostume and set research powerpoint presentation
costume and set research powerpoint presentation
 
Pigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food ManufacturingPigging Solutions in Pet Food Manufacturing
Pigging Solutions in Pet Food Manufacturing
 
"Debugging python applications inside k8s environment", Andrii Soldatenko
"Debugging python applications inside k8s environment", Andrii Soldatenko"Debugging python applications inside k8s environment", Andrii Soldatenko
"Debugging python applications inside k8s environment", Andrii Soldatenko
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024
 
Developer Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQLDeveloper Data Modeling Mistakes: From Postgres to NoSQL
Developer Data Modeling Mistakes: From Postgres to NoSQL
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 Presentation
 

Automated Evaluation of Subjective Answers Using Contextual Synonyms

  • 1. International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015 DOI : 10.5121/ijnlc.2015.4105 51 EVALUATION OF SUBJECTIVE ANSWERS USING GLSA ENHANCED WITH CONTEXTUAL SYNONYMY Parag A. Guruji1 , Mrunal M. Pagnis2 , Sayali M. Pawar3 and Prakash J. Kulkarni4 Department of Computer Science and Engineering, Walchand College of Engineering, Sangli, India 416415 ABSTRACT Evaluation of subjective answers submitted in an exam is an essential but one of the most resource consuming educational activity. This paper details experiments conducted under our project to build a software that evaluates the subjective answers of informative nature in a given knowledge domain. The paper first summarizes the techniques such as Generalized Latent Semantic Analysis (GLSA) and Cosine Similarity that provide basis for the proposed model. The further sections point out the areas of improvement in the previous work and describe our approach towards the solutions of the same. We then discuss the implementation details of the project followed by the findings that show the improvements achieved. Our approach focuses on comprehending the various forms of expressing same entity and thereby capturing the subjectivity of text into objective parameters. The model is tested by evaluating answers submitted by 61 students of Third Year B. Tech. CSE class of Walchand College of Engineering Sangli in a test on Database Engineering. KEYWORDS N-gram, Synonym-Set, natural language processing, GLSA, SVD, Wordnet 1.INTRODUCTION The progress in computer network technologies and increased inclination towards making human interactions „paperless‟, online test systems are being adopted widely. Automation of the assessment activity comes to a great help in saving resources (time, efforts & cost) and in increasing the consistency of the assessment by mitigating potential human biases of intangible nature. Realization of Computer Aided Assessment (CAA) for objective types of answers is a well addressed deterministic problem which involves „matching‟ of submitted answers to the corresponding predefined expected answers. Realization of the CAA in case of descriptive answers becomes comparatively more challenging by virtue of the subjectivity of the answers. It involves „mapping‟ of underlying semantics of submitted answers to that of the expected answers. Pattern recognition, artificial cognition, and natural language processing are required in such process of CAA. Significant work has been done in this field. Project Essay Grader (PEG) was developed by Ellis Page in 1966. PEG ignored the semantic aspect of essays and relied on the surface structures, for which it had been criticized [1, 2]. Thomas K. Landauer, Boulder & Peter W. Foltz developed
  • 2. International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015 52 Latent Semantics Analysis (LSA) [1, 2] – the technique widely used in automated assessment. Group lead by Jill Burstein designed E-rater that deals with automatedessay scoring of GMAT [3]. Automated Text Maker (ATM) is another computer aided assessment system and can be adopted to address different disciplines [4]. Md. Monjurul Islam and A. S. M. Latiful Hoque came up with Automated Essay Scoring Using Generalized Latent Semantic Analysis in 2010 [5] and also applied GLSA for scoring essays in Bangla Language in 2013 [6]. M. S. Devi and Himani Mittal applied the LSA Technique to evaluate technical answers in 2013 [7] wherein they have used single word as atomic unit for computations and suggested in future work that use of synonymy would enhance the results. In this paper, automation of assessment of technical answers written by undergraduate students is further enhanced by modifying GLSA based techniques. Section II outlines the fundamental model and describes fundamental techniques used. In Section III, we discuss the enhancements made to the basic model. It includes the use of N-grams as atomic computation unit, capturing context based synonymy, the concept of Synonym-Sets and change in document representation and the N-gram formation over Synonym-Sets which enables representation of many by one. 2. THE FUNDAMENTAL MODEL The existing model of the system for evaluating subjective answers that uses Generalized Latent Semantic Analysis (GLSA) is discussed in this section. We have referred to it as the fundamental model. 2.1. GLSA over LSA: Normally LSA represents documents and their word content in a large two-dimensional matrix semantic space. Using a matrix algebra technique known as Singular Value Decomposition (SVD), new relationships between words and documents are uncovered, and existing relationship are modified to more accurately represent their true significance.[1] Each word represents a row in the matrix, while each column represents the sentences, paragraphs and other subdivisions of the context in which the word occurs. The traditional word by document matrix creation of LSA does not consider word sequence in a document [2]. Here the formation of word by document matrix the word pair “concurrent transactions” makes the same result of “transactions concurrent”. Thus, LSA fails to capture the sematic effect of collocations in the document. GLSA addresses this issue by considering n-gram as atomic unit of the document instead of individual word. An n-gram is a contiguous sequence of n words from a given sequence of text. Thus now, “concurrent transactions” is not recognized same as “transactions concurrent”. 2.2. System Architecture Overview: The system works in two phases – training set generation and testing of newly submitted answers. Pre-processing and frequency computation operations to be done on an individual document are common in both the phases.
  • 3. International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015 53 2.3. Pre-processing: The pre-processing is comprised of stopwords removal, word-stemming and lingual error handling in the fundamental model. We have decided to eliminate the step of word-stemming because we found that dealing with contextual synonymy (discussed in section III) with the stemmed words as inputs introduces too much of noise and thus inaccuracies. The pre-processing thus now consists of two steps - stopwords removal and lingual errors handling. 2.4. Phase 1: Training set generation: The input for this phase is set of pre-graded answers as per earlier model. We include the question-domain in the input set for the purpose discussed in section III. Output of this phase is the training-set. After pre-processing every pre-graded answer – a training document, n-grams created from all the training documents and their respective frequencies for corresponding documents are computed to form the n-gram-by-document vectors. Each such vector represents one training document. The n-gram-by-document matrix is formed by putting together all such vectors as columns of the matrix. Every row of the matrix represents an n-gram and every column represents a training- document. Every cell of this matrix holds the frequency value of the n-gram corresponding to its row-index in the training document that corresponds to its column index. The n-gram-by- document matrix then undergoes Singular Value Decomposition (SVD) to generate the singular value matrices. According to SVD a matrix At×n is decomposed as follows: At×n = Ut×n × Sn×n × Vt×n T …(1) Here, A is n-gram by documents matrix, U is an orthogonal matrix, S is diagonal matrix and VT is the transpose of an orthogonal matrix V. After applying dimensionality reduction on matrices U, S and V we get reduced singular value matrices that are used to generate training vectors. The dimensionality reduction eliminates one or more least significant singular values in order to reduce the noise involved in the sematic space developed in the form of SVD matrices. Training vector corresponding to every pre-graded answer is then formed as per the following equation: dj ‟ = dj T × Ut×k × Sk×k -1 …(2) Here, dj T is the transpose of document vector dj, Ut×k is truncated left orthogonal matrix and Sk×k is truncated singular matrix from truncated SVD matrices. The training vectors dj ‘ along with human grades of pre-graded answers makes the training set. Phase 2: Evaluation of submitted answers: The input for this phase is set of submitted answers for corresponding question which will be evaluated by the system. For the purpose of testing the system, these answers are also evaluated
  • 4. International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015 54 by the human graders. Output of this phase is the machine-generated grades for the submitted answers/(s). Submitted answer first undergoes pre-processing. Frequency values for all the n-grams (generated in training phase) for this document are computed to form the n-gram-by-document vector q of the submitted answer. This vector is used to generate the query vector for submitted answer as follows: q‟ = q × Ut×k × Sk×k -1 …(3) Here, q is query matrix Ut×k is truncated left orthogonal matrix and Sk×k is truncated singular matrix generated from SVD in training phase. Similarity between query vector q' and training set vectors dj ‟ is calculated by Cosine similarity [1] Here, wqj is the jth weight of query vector q' and dij is the ith weight of training essay set vectors dj ‟. The highest correlation value from the Cosine of query vector and the training essay vector is used for grading the submitted essay. 3. THE ENHANCEMENTS In the n-gram by document matrix, each row represents an n-gram and each column represents a document. Every cell contains a value that represents the frequency of n-gram associated with the row index of that cell in the document associated with the column index of the same cell. We use the Term Frequency/Inverse Document Frequency (TFIDF) notation to compute this frequency value. Now, to capture the significance of a given n-gram, the system depends on the frequency value associated with it. The subjectivity of the natural language text has an inherent feature which is generally known as “saying same thing in different words”. For instance, while talking about networks, somebody may refer to a term, say, „connected nodes‟; whereas, in some other text, same entity might have been referred to as „joint peers‟ or „connected peers‟ or „joint nodes‟. Now, all these four expressions represent same entity, given the same context – networks. But, in the existing model that uses GLSA, all of them are treated as independent bigrams and thus, their frequencies are computed separately. This introduces inaccuracy in the computed significance of the given entity if the same entity is expressed in different forms at its different occurrences in the text. We propose a solution to this problem with the use of Synonym-Set.
  • 5. International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015 55 3.1. Concept of Synonym-Set & Document redefinition: For a given n-gram, say, N = (W1 W2 … Wi … Wn)c which represents certain entity, say E in a given context C, if W11, W12, ... , W1S1 are synonyms representing W1 then, we define Synonym- Set Sic as Sic = (Wi, {Wi1, Wi2, ... , WiSi}) The document d is defined as the sequence of words (W1, W2, ... , Wj, ... Wn). We redefine d as d = (S1c, S2c, ... Sjc, ..., Snc) Every Synonym-Set Sic is associated with a unique integer say Iic. Therefore the document d is represented as d = (I1c, I2c, ... Ijc, ..., Inc) Now, n-grams are formed over this redefined representation of the document. One such n-gram now represents a set of all such permutations of expressing the entity represented by that n-gram which are valid in the given context. In the example stated previously, C = „ computer networks‟, S1c = (I1c, {„ connected’, „ joint’}), S2c = (I2c, {„ node‟, „ peer‟ }), Nold = { „connected node‟, „connected peer‟, „joint node‟, „joint peer‟} Nnew = ( I1c I2c) Thus, a single new n-gram now represents all P permutations of forms in which E can be expressed in the subjective text in context C, where P is given by: P = S1 × S2 × ... × Si × ... × Sn Now, occurrence of any of the P forms of expressing E will be identified by the system as single entity represented by Nnew. Hence, while computing the significance of any n-gram (and in turn the entity which it represents) using TFIDF, the weightage of the entity is now converged and put together under single unique entity unlike earlier models where it was getting scattered among P ¬ways of expressing E. This increases the precision of the system and deals with an inherent feature of subjectivity – „saying same thing in different words’. 3.2. Contextual Synonymy: In the solution discussed above, there lies an assumption that a synonym-set will consist of words having same semantics in a given context, i.e. every word in a given synonym-set shares a relation of contextual synonymy with the key-word of that synonym-set. The number of false positive n-grams representing different ways of expressing same entity is thus inversely proportional to the accuracy of formation of the contextual synonym-sets. For instance, word „node‟ is related also to the word „knot‟ by the relation of synonymy. But, in given
  • 6. International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015 56 context of „computer networks‟, „knot‟ does not mean „node‟. Hence, „joint knots‟ cannot be a valid form of representing the entity „connected nodes‟. So, the implicit challenge in making the concept of synonym-set and document redefinition functional is to device a process that accurately forms the contextual synonym-set for every word in the pre-processed training text. Our method to prepare such synonym-sets uses the database provided by the Wordnet. Along with inter-relations between words such as synonymy, hyponym, etc., the Wordnet also provides definitions of every word. We observed that the strings of context-specific definitions in Wordnet show certain pattern wherein the context or the domain of the definition is specified at the beginning of the definition in parentheses. We first fetch all the synonyms of the word under consideration. Then, to filter out only contextual synonyms, we make use of the aforesaid pattern in the Wordnet definitions and apply string matching techniques along with the context of the question to which the subjective answers are written to identify the contexts or the domains of the words which are initially fetched as synonyms of the key-word. The presence of domain-specific definition in Wordnet for a given context also serves as a qualifying criterion for any word to be identified as a terminological, and hence significant, word in the context of the question. Since, a subjective text of informative nature which talks about certain domain holds the major part of its semantics in the form of terminological entities, the process discussed above proves to be effective in dealing with informative answers pertaining to an identified domain. 3.3. Feedback-driven incremental advancement: We attempt to incorporate an aspect of human behavior into our model by using a feedback system. The objective of this feedback system is to make the system avoid same mistakes twice. The system can make mistake in grading an answer which has content semantically or syntactically „unknown’ to the system, i.e., it contains large number of new entities which were not part of the training set at all. Such cases are identified as outliers and the grades assigned to them are forwarded for human moderation. Identifying potential outliers is important for the System’s reliability. To identify outliers from a given set of submitted answers which is normally distributed in terms of probability of presence of outliers, we use the characteristic of an outlier by virtue of which, even the highest amongst the cosine correlations of the query vector of outlier with the training vectors has comparatively low value than the highest of cosine correlations of an average case test answer (graded with acceptable reliability) with the training vectors. Using this property, we identify an outlier as the answer in case of which, the highest cosine correlation of the query vector with the training vectors is below the average of such highest cosine correlations of query vectors of all answers in the given set. We feedback every such outlier along with its human-moderated grade to the system for its inclusion in the training set. This leads to addition of new n-grams and re-computation of the SVD matrices which in turn modifies the training set. Figure1 shows the overview of our system model.
  • 7. International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015 57 Figure 1. System overview 4. EXPERIMENT AND RESULTS We have experimented on our model by conducting an online test on Database Engineering course for a class of 61 students in our departmental lab. We collected the human grades for each answer from the concerned faculty. Further, 30 answers were used for training the system and rest for testing. Following (Fig. 2) is the graph comparing the machine generated grades and corresponding average human grades for each test answer. Three human graders have graded the answer and we have used the average of their scores. These results are obtained for value of n=1,2,3. Figure 2 Graph: Machine grade Vs Human grade Table 1. Variation in Training Time, Testing Times and Standard Deviation with various combinations of n-grams
  • 8. International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015 58 T1 : Training time in minutes with 30 answers T2 : Testing time in minutes with 30 answers Table 1 shows the comparison of various values of n and the time taken for training 30 answers, testing answers, number of n-grams and the standard deviation. Figure 3 Graph: Variation in Standard Deviation Graph in Figure 3 shows that when we increase the value of n in n-gram we get improved results. The standard deviation approached zero as we increased n. However, we observed that if we try to improve the results with 4-grams then it has a reverse effect. This is because average number of words in a sentence (for third year undergraduate students) is about 6-7 after stop word removal. So when we have 4-grams then we are actually trying to find phrases as long as four words. This introduces noise in the result. This is the major reason for limiting the value of n to 3
  • 9. International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015 59 Figure 4 Graph: Variation in Training Time with N Graph in Figure 4 shows that the training time increases with number of n-grams, which is dependent on the number of training documents. Thus, we can say that the number of training documents determines the time taken for training. As a part of reliability measure we compute the error with respect to human grades. When error goes beyond a threshold we add the outliers to SVD and recomputed the matrix to improve the results. 5. CONCLUSIONS We have developed a system that focuses on the concept of synonyms while it grades an answer. This enhancement has given significant results in comparisons with the recently proposed models that use latent semantic analysis without incorporating synonyms. Section 2 is the basic model that currently is considered successful for grading answers. Section 3 introduces a new aspect due to which the model can perform better. Implementation and results have proved that the enhancement costs negligible resources in terms of time and space. The improvement in performance definitely outweighs the negligible costs. Section 4 shows the experimental results taken on collected data. The feedback driven model improves the results every time we train the system. This regulates error periodically and also improves the results after each time we train. ACKNOWLEDGEMENTS The authors would like to thank firstly the Department of Computer Science and Engineering, Walchand College of Engineering, Sangli for facilitating the experimentation by providing required infrastructure. The authors also thank in particular the esteemed faculty members of Dept. of CSE, WCE, Sangli Mrs. Hetal Gandhi and Dr. Smriti Bhandari, for their valuable assistance in human expert grading of test answers used in training and testing of the system. Lastly the authors thank all the students of 2013-14‟s Third Year B.Tech. CSE class of WCE, Sangli who participated in the experiments as test takers. Without their sincere voluntary participation, the experiments could not have proceeded.
  • 10. International Journal on Natural Language Computing (IJNLC) Vol. 4, No.1, February 2015 60 REFERENCES [1] Md. Monjurul Islam and A. S. M. Latiful Hoque, “Automated Essay Scoring Using Generalized Latent Semantic Analysis,” in Proceedings of 13th International Conference on Computer and Information Technology (ICCIT 2010), 23-25 December, Dhaka, Bangladesh [2] M. Syamala Devi and Himani Mittal, “Subjective Evaluation using LSA Technique,” in International Journal of Computers and Distributed Systems Vol. No.3, Issue I, April-May 2013 ISSN: 2278-5183 (www.ijcdsonline.com) [3] S. Valenti, F. Neri, and A. Cucchiarelli, “An overview of current research on automated essay grading,” in Journal of Information Technology Education, vol. 2, 2003, pp. 319-330 [4] Md. Monjurul Islam and A. S. M. Latiful Hoque, “Automated Bangla Essay Scoring System: ABESS,” in 978-1-4799-0400-6/13/$31.00 ©2013 IEEE [5] “Java API for WordNet Searching (JAWS)” [Online] Available: http://lyle.smu.edu/~tspell/jaws/index.html [6] Mark Watson, “Practical Artificial Intelligence Programming with Java”, Ed. 3, 11/11/2008 [Online] Available: http://archive.org/stream/PracticalArtificialIntelligenceProgrammingWithJava/JavaAI3rd_djvu.txt [7] JAMA: A Java Matrix Package [Online]. Available: http://math.nist.gov/javanumerics/jama/ Authors Parag A. Guruji Data Scientist at Zlemma Analytics Inc, Pune, India. Former student of Walchand College of Engineering, Sangli, India. Received Bachelor of Technology Degree in Computer Science and Engineering in May 2014. Team Member of the project discussed in this paper. Mrunal M. Pagnis Graduate Student at Indiana University at Bloomington, Indiana, USA. Former student of Walchand College of Engineering, Sangli, India. Received Bachelor of Technology Degree in Computer Science and Engineering in May 2014. Team Member of the project discussed in this paper. Sayali M. Pawar Graduate Engineer Trainee at Tata Consultancy Services, Pune, India. Former student of Walchand College of Engineering, Sangli, India. Received Bachelor of Technology Degree in Computer Science and Engineering in May 2014. Team Member of the project discussed in this paper. Dr. Prakash J. Kulkarni Professor in Computer Science and Engineering and Deputy Director at Walchand College of Engineering, Sangli, India. Faculty Mentor of the project discussed in this paper.