Automatic multiple-choice question generation (MCQG) is a useful yet challenging task in Natural Language
Processing (NLP). It is the task of automatic generation of correct and relevant questions from textual data.
Despite its usefulness, manually creating sizeable, meaningful and relevant questions is a time-consuming
and challenging task for teachers. In this paper, we present an NLP-based system for automatic MCQG for
Computer-Based Testing Examination (CBTE).We used NLP technique to extract keywords that are
important words in a given lesson material. To validate that the system is not perverse, five lesson materials
were used to check the effectiveness and efficiency of the system. The manually extracted keywords by the
teacher were compared to the auto-generated keywords and the result shows that the system was capable of
extracting keywords from lesson materials in setting examinable questions. This outcome is presented in a
user-friendly interface for easy accessibility.
Semi Automated Text Categorization Using Demonstration Based Term SetIJCSEA Journal
Manual Analysis of huge amount of textual data requires a tremendous amount of processing time and effort in reading the text and organizing them in required format. In the current scenario, the major problem is with text categorization because of the high dimensionality of feature space. Now-a-days there are many methods available to deal with text feature selection. This paper aims at such semi automated text categorization feature selection methodology to deal with a massive data using one of the phases of David Merrill’s First principles of instruction (FPI). It uses a pre-defined category group by providing them with the proper training set based on the demonstration phase of FPI. The methodology involves the text tokenization, text categorization and text analysis.
A Survey on Sentiment Categorization of Movie ReviewsEditor IJMTER
Sentiment categorization is a process of mining user generated text content and determine
the sentiment of the users towards that particular thing. It is the approach of detecting the sentiment of
the author in regard to some topics. It also known as sentiment detection, sentiment analysis and opinion
mining. It is very useful for movie production companies that interested in knowing how users feel
about their movies. For example word “excellent” indicates that the review gives positive emotion about
particular movie. The same applies to movies, songs, cars, holiday destinations, Political parties, social
network sites, web blogs, discussion forum and so on. Sentiment categorization can be carried out by
using three approaches. First, Supervised machine learning based text classifier on Naïve Bayes,
Maximum Entropy, SVM, kNN classifier, hidden marcov model. Second, Unsupervised Semantic
Orientation scheme of extracting relevant N-grams of the text and then labelling. Third, SentiWordNet
based publicly available library.
TEXT SENTIMENTS FOR FORUMS HOTSPOT DETECTIONijistjournal
The user generated content on the web grows rapidly in this emergent information age. The evolutionary changes in technology make use of such information to capture only the user’s essence and finally the useful information are exposed to information seekers. Most of the existing research on text information processing, focuses in the factual domain rather than the opinion domain. In this paper we detect online hotspot forums by computing sentiment analysis for text data available in each forum. This approach analyses the forum text data and computes value for each word of text. The proposed approach combines K-means clustering and Support Vector Machine with PSO (SVM-PSO) classification algorithm that can be used to group the forums into two clusters forming hotspot forums and non-hotspot forums within the current time span. The proposed system accuracy is compared with the other classification algorithms such as Naïve Bayes, Decision tree and SVM. The experiment helps to identify that K-means and SVM-PSO together achieve highly consistent results.
Feature selection, optimization and clustering strategies of text documentsIJECEIAES
Clustering is one of the most researched areas of data mining applications in the contemporary literature. The need for efficient clustering is observed across wide sectors including consumer segmentation, categorization, shared filtering, document management, and indexing. The research of clustering task is to be performed prior to its adaptation in the text environment. Conventional approaches typically emphasized on the quantitative information where the selected features are numbers. Efforts also have been put forward for achieving efficient clustering in the context of categorical information where the selected features can assume nominal values. This manuscript presents an in-depth analysis of challenges of clustering in the text environment. Further, this paper also details prominent models proposed for clustering along with the pros and cons of each model. In addition, it also focuses on various latest developments in the clustering task in the social network and associated environments.
Modeling Text Independent Speaker Identification with Vector QuantizationTELKOMNIKA JOURNAL
Speaker identification is one of the most important technologies nowadays. Many fields such as
bioinformatics and security are using speaker identification. Also, almost all electronic devices are using
this technology too. Based on number of text, speaker identification divided into text dependent and text
independent. On many fields, text independent is mostly used because number of text is unlimited. So, text
independent is generally more challenging than text dependent. In this research, speaker identification text
independent with Indonesian speaker data was modelled with Vector Quantization (VQ). In this research
VQ with K-Means initialization was used. K-Means clustering also was used to initialize mean and
Hierarchical Agglomerative Clustering was used to identify K value for VQ. The best VQ accuracy was
59.67% when k was 5. According to the result, Indonesian language could be modelled by VQ. This
research can be developed using optimization method for VQ parameters such as Genetic Algorithm or
Particle Swarm Optimization.
Machine Learning Techniques with Ontology for Subjective Answer Evaluationijnlc
Computerized Evaluation of English Essays is performed using Machine learning techniques like Latent
Semantic Analysis (LSA), Generalized LSA, Bilingual Evaluation Understudy and Maximum Entropy.
Ontology, a concept map of domain knowledge, can enhance the performance of these techniques. Use of
Ontology makes the evaluation process holistic as presence of keywords, synonyms, the right word
combination and coverage of concepts can be checked. In this paper, the above mentioned techniques are
implemented both with and without Ontology and tested on common input data consisting of technical
answers of Computer Science. Domain Ontology of Computer Graphics is designed and developed. The
software used for implementation includes Java Programming Language and tools such as MATLAB,
Protégé, etc. Ten questions from Computer Graphics with sixty answers for each question are used for
testing. The results are analyzed and it is concluded that the results are more accurate with use of
Ontology.
Review of Various Text Categorization Methodsiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Semantic Based Model for Text Document Clustering with IdiomsWaqas Tariq
Text document clustering has become an increasingly important problem in recent years because of the tremendous amount of unstructured data which is available in various forms in online forums such as the web, social networks, and other information networks. Clustering is a very powerful data mining technique to organize the large amount of information on the web. Traditionally, document clustering methods do not consider the semantic structure of the document. This paper addresses the task of developing an effective and efficient method to improve the semantic structure of the text documents. A method has been developed that performs the following: tag the documents for parsing, replacement of idioms with their original meaning, semantic weights calculation for document words and apply semantic grammar. The similarity measure is obtained between the documents and then the documents are clustered using Hierarchical clustering algorithm. The method adopted in this work is evaluated on different data sets with standard performance measures and the effectiveness of the method to develop in meaningful clusters has been proved.
Semi Automated Text Categorization Using Demonstration Based Term SetIJCSEA Journal
Manual Analysis of huge amount of textual data requires a tremendous amount of processing time and effort in reading the text and organizing them in required format. In the current scenario, the major problem is with text categorization because of the high dimensionality of feature space. Now-a-days there are many methods available to deal with text feature selection. This paper aims at such semi automated text categorization feature selection methodology to deal with a massive data using one of the phases of David Merrill’s First principles of instruction (FPI). It uses a pre-defined category group by providing them with the proper training set based on the demonstration phase of FPI. The methodology involves the text tokenization, text categorization and text analysis.
A Survey on Sentiment Categorization of Movie ReviewsEditor IJMTER
Sentiment categorization is a process of mining user generated text content and determine
the sentiment of the users towards that particular thing. It is the approach of detecting the sentiment of
the author in regard to some topics. It also known as sentiment detection, sentiment analysis and opinion
mining. It is very useful for movie production companies that interested in knowing how users feel
about their movies. For example word “excellent” indicates that the review gives positive emotion about
particular movie. The same applies to movies, songs, cars, holiday destinations, Political parties, social
network sites, web blogs, discussion forum and so on. Sentiment categorization can be carried out by
using three approaches. First, Supervised machine learning based text classifier on Naïve Bayes,
Maximum Entropy, SVM, kNN classifier, hidden marcov model. Second, Unsupervised Semantic
Orientation scheme of extracting relevant N-grams of the text and then labelling. Third, SentiWordNet
based publicly available library.
TEXT SENTIMENTS FOR FORUMS HOTSPOT DETECTIONijistjournal
The user generated content on the web grows rapidly in this emergent information age. The evolutionary changes in technology make use of such information to capture only the user’s essence and finally the useful information are exposed to information seekers. Most of the existing research on text information processing, focuses in the factual domain rather than the opinion domain. In this paper we detect online hotspot forums by computing sentiment analysis for text data available in each forum. This approach analyses the forum text data and computes value for each word of text. The proposed approach combines K-means clustering and Support Vector Machine with PSO (SVM-PSO) classification algorithm that can be used to group the forums into two clusters forming hotspot forums and non-hotspot forums within the current time span. The proposed system accuracy is compared with the other classification algorithms such as Naïve Bayes, Decision tree and SVM. The experiment helps to identify that K-means and SVM-PSO together achieve highly consistent results.
Feature selection, optimization and clustering strategies of text documentsIJECEIAES
Clustering is one of the most researched areas of data mining applications in the contemporary literature. The need for efficient clustering is observed across wide sectors including consumer segmentation, categorization, shared filtering, document management, and indexing. The research of clustering task is to be performed prior to its adaptation in the text environment. Conventional approaches typically emphasized on the quantitative information where the selected features are numbers. Efforts also have been put forward for achieving efficient clustering in the context of categorical information where the selected features can assume nominal values. This manuscript presents an in-depth analysis of challenges of clustering in the text environment. Further, this paper also details prominent models proposed for clustering along with the pros and cons of each model. In addition, it also focuses on various latest developments in the clustering task in the social network and associated environments.
Modeling Text Independent Speaker Identification with Vector QuantizationTELKOMNIKA JOURNAL
Speaker identification is one of the most important technologies nowadays. Many fields such as
bioinformatics and security are using speaker identification. Also, almost all electronic devices are using
this technology too. Based on number of text, speaker identification divided into text dependent and text
independent. On many fields, text independent is mostly used because number of text is unlimited. So, text
independent is generally more challenging than text dependent. In this research, speaker identification text
independent with Indonesian speaker data was modelled with Vector Quantization (VQ). In this research
VQ with K-Means initialization was used. K-Means clustering also was used to initialize mean and
Hierarchical Agglomerative Clustering was used to identify K value for VQ. The best VQ accuracy was
59.67% when k was 5. According to the result, Indonesian language could be modelled by VQ. This
research can be developed using optimization method for VQ parameters such as Genetic Algorithm or
Particle Swarm Optimization.
Machine Learning Techniques with Ontology for Subjective Answer Evaluationijnlc
Computerized Evaluation of English Essays is performed using Machine learning techniques like Latent
Semantic Analysis (LSA), Generalized LSA, Bilingual Evaluation Understudy and Maximum Entropy.
Ontology, a concept map of domain knowledge, can enhance the performance of these techniques. Use of
Ontology makes the evaluation process holistic as presence of keywords, synonyms, the right word
combination and coverage of concepts can be checked. In this paper, the above mentioned techniques are
implemented both with and without Ontology and tested on common input data consisting of technical
answers of Computer Science. Domain Ontology of Computer Graphics is designed and developed. The
software used for implementation includes Java Programming Language and tools such as MATLAB,
Protégé, etc. Ten questions from Computer Graphics with sixty answers for each question are used for
testing. The results are analyzed and it is concluded that the results are more accurate with use of
Ontology.
Review of Various Text Categorization Methodsiosrjce
IOSR Journal of Computer Engineering (IOSR-JCE) is a double blind peer reviewed International Journal that provides rapid publication (within a month) of articles in all areas of computer engineering and its applications. The journal welcomes publications of high quality papers on theoretical developments and practical applications in computer technology. Original research papers, state-of-the-art reviews, and high quality technical notes are invited for publications.
Semantic Based Model for Text Document Clustering with IdiomsWaqas Tariq
Text document clustering has become an increasingly important problem in recent years because of the tremendous amount of unstructured data which is available in various forms in online forums such as the web, social networks, and other information networks. Clustering is a very powerful data mining technique to organize the large amount of information on the web. Traditionally, document clustering methods do not consider the semantic structure of the document. This paper addresses the task of developing an effective and efficient method to improve the semantic structure of the text documents. A method has been developed that performs the following: tag the documents for parsing, replacement of idioms with their original meaning, semantic weights calculation for document words and apply semantic grammar. The similarity measure is obtained between the documents and then the documents are clustered using Hierarchical clustering algorithm. The method adopted in this work is evaluated on different data sets with standard performance measures and the effectiveness of the method to develop in meaningful clusters has been proved.
Autonomous Educational Testing System Using Unsupervised Feature Learning.IJITE
With the increase of ubiquitous data all over the internet, intelligent classroom systems that integrate
traditional learning techniques with modern e-learning tools have become quite popular and necessary
today. Although a substantial amount of work has been done in the field of e-learning, specifically in
automation of objective question and answer evaluation, personalized learning, adaptive evaluation
systems, the field of qualitative analysis of a student’s subjective paragraph answers remains unexplored to
a large extent.
The traditional board, chalk, talk based classroom scenario involves a teacher setting question papers
based on the concepts taught, checks the answers written by students manually and thus evaluates the
students’ performance. However, setting question papers remains a time consuming process with the
teacher having to bother about question quality, level of difficulty and redundancy. In addition the process
of manually correcting students’ answers is a cumbersome and tedious task especially where the class size
is large.
In this paper, we put forth the design, analysis and implementation details along with some experimental
outputs to build a system that integrates all the above mentioned tasks with minimal teacher involvement
that not only automates the traditional classroom scenario but also overcomes its inherent shortcomings
and fallacies.
Improving Sentiment Analysis of Short Informal Indonesian Product Reviews usi...TELKOMNIKA JOURNAL
Sentiment analysis in short informal texts like product reviews is more challenging. Short texts are
sparse, noisy, and lack of context information. Traditional text classification methods may not be suitable
for analyzing sentiment of short texts given all those difficulties. A common approach to overcome these
problems is to enrich the original texts with additional semantics to make it appear like a large document of
text. Then, traditional classification methods can be applied to it. In this study, we developed an automatic
sentiment analysis system of short informal Indonesian texts using Naïve Bayes and Synonym Based
Feature Expansion. The system consists of three main stages, preprocessing and normalization, features
expansion and classification. After preprocessing and normalization, we utilize Kateglo to find some
synonyms of every words in original texts and append them. Finally, the text is classified using Naïve
Bayes. The experiment shows that the proposed method can improve the performance of sentiment
analysis of short informal Indonesian product reviews. The best sentiment classification performance using
proposed feature expansion is obtained by accuracy of 98%.The experiment also show that feature
expansion will give higher improvement in small number of training data than in the large number of them.
Random forest application on cognitive level classification of E-learning co...IJECEIAES
The e-learning is the primary method of learning for most learners after the regular academics studies. The knowledge delivery through E-learning technologies increased exponentially over the years because of the advancement in internet and e-learning technologies. Knowledge delivery to some people would never have been possible without the e-learning technologies. Most of the working professional do focused studies for carrier advancement, promotion or to improve the domain knowledge. These learner can find many free e-learning web sites from the internet easily in the domain of interest. However it is quite difficult to find the best e-learning content suitable for their learning based on their domain knowledge level. User spent most of the time figuring out the right content from a plethora of available content and end up learning nothing. An intelligent framework using machine learning algorithms with random forest Classifier is proposed to address this issue, which classifies the e-learning content based on its difficulty levels and provide the learner the best content suitable based on the knowledge level .The frame work is trained with the data set collected from multiple popular e-learning web sites. The model is tested with real time e-learning web sites links and found that the e-contents in the web sites are recommended to the user based on its difficulty levels as beginner level, intermediate level and advanced level.
A scoring rubric for automatic short answer grading systemTELKOMNIKA JOURNAL
During the past decades, researches about automatic grading have become an interesting issue. These studies focuses on how to make machines are able to help human on assessing students’ learning outcomes. Automatic grading enables teachers to assess student's answers with more objective, consistent, and faster. Especially for essay model, it has two different types, i.e. long essay and short answer. Almost of the previous researches merely developed automatic essay grading (AEG) instead of automatic short answer grading (ASAG). This study aims to assess the sentence similarity of short answer to the questions and answers in Indonesian without any language semantic's tool. This research uses pre-processing steps consisting of case folding, tokenization, stemming, and stopword removal. The proposed approach is a scoring rubric obtained by measuring the similarity of sentences using the string-based similarity methods and the keyword matching process. The dataset used in this study consists of 7 questions, 34 alternative reference answers and 224 student’s answers. The experiment results show that the proposed approach is able to achieve a correlation value between 0.65419 up to 0.66383 at Pearson's correlation, with Mean Absolute Error (푀퐴퐸) value about 0.94994 until 1.24295. The proposed approach also leverages the correlation value and decreases the error value in each method.
Single document keywords extraction in Bahasa Indonesia using phrase chunkingTELKOMNIKA JOURNAL
Keywords help readers to understand the idea of a document quickly. Unfortunately, considerable time and effort are often needed to come up with a good set of keywords manually. This research focused on generating keywords from a document automatically using phrase chunking. Firstly, we collected part of speech patterns from a collection of documents. Secondly, we used those patterns to extract candidate keywords from the abstract and the content of a document. Finally, keywords are selected from the candidates based on the number of words in the keyword phrases and some scenarios involving candidate reduction and sorting. We evaluated the result of each scenario using precision, recall, and F-measure. The experiment results show: i) shorter-phrase keywords with string reduction extracted from the abstract and sorted by frequency provides the highest score, ii) in every proposed scenario, extracting keywords using the abstract always presents a better result, iii) using shorter-phrase patterns in keywords extraction gives better score in comparison to using all phrase patterns, iv) sorting scenarios based on the multiplication of candidate frequencies and the weight of the phrase patterns offer better results.
INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS cscpconf
In education, the use of electronic (E) examination systems is not a novel idea, as Eexamination systems have been used to conduct objective assessments for the last few years. This research deals with randomly designed E-examinations and proposes an E-assessment system that can be used for subjective questions. This system assesses answers to subjective questions by finding a matching ratio for the keywords in instructor and student answers. The matching ratio is achieved based on semantic and document similarity. The assessment system is composed of four modules: preprocessing, keyword expansion, matching, and grading. A survey and case study were used in the research design to validate the proposed system. The examination assessment system will help instructors to save time, costs, and resources, while increasing efficiency and improving the productivity of exam setting and assessments.
A template based algorithm for automatic summarization and dialogue managemen...eSAT Journals
Abstract This paper describes an automated approach for extracting significant and useful events from unstructured text. The goal of research is to come out with a methodology which helps in extracting important events such as dates, places, and subjects of interest. It would be also convenient if the methodology helps in presenting the users with a shorter version of the text which contain all non-trivial information. We also discuss implementation of algorithms which exactly does this task, developed by us. Key Words: Cosine Similarity, Information, Natural Language, Summarization, Text Mining
New instances classification framework on Quran ontology applied to question ...TELKOMNIKA JOURNAL
Instances classification with the small dataset for Quran ontology is the current research problem which appears in Quran ontology development. The existing classification approach used machine learning: Backpropagation Neural Network. However, this method has a drawback; if the training set amount is small, then the classifier accuracy could decline. Unfortunately, Holy Quran has a small corpus. Based on this problem, our study aims to formulate new instances classification framework for small training corpus applied to semantic question answering system. As a result, the instances classification framework consists of several essential components: pre-processing, morphology analysis, semantic analysis, feature extraction, instances classification with Radial Basis Function Networks algorithm, and the transformation module. This algorithm is chosen since it robustness to noisy data and has an excellent achievement to handle small dataset. Furthermore, document processing module on question answering system is used to access instances classification result in Quran ontology.
Topic Discovery of Online Course Reviews Using LDA with Leveraging Reviews He...IJECEIAES
Despite the popularity of the Massive Open Online Courses, small-scale research has been done to understand the factors that influence the teaching-learning process through the massive online platform. Using topic modeling approach, our results show terms with prior knowledge to understand e.g.: Chuck as the instructor name. So, we proposed the topic modeling approach on helpful subjective reviews. The results show five influential factors: “learn easy excellent class program”, “python learn class easy lot”, “Program learn easy python time game”, and “learn class python time game”. Also, research results showed that the proposed method improved the perplexity score on the LDA model.
COMPARISON INTELLIGENT ELECTRONIC ASSESSMENT WITH TRADITIONAL ASSESSMENT FOR ...cseij
In education, the use of electronic (E) examination systems is not a novel idea, as E-examination systems
have been used to conduct objective assessments for the last few years. This research deals with randomly
designed E-examinations and proposes an E-assessment system that can be used for subjective questions.
This system assesses answers to subjective questions by finding a matching ratio for the keywords in
instructor and student answers.
Opinion mining on newspaper headlines using SVM and NLPIJECEIAES
Opinion Mining also known as Sentiment Analysis, is a technique or procedure which uses Natural Language processing (NLP) to classify the outcome from text. There are various NLP tools available which are used for processing text data. Multiple research have been done in opinion mining for online blogs, Twitter, Facebook etc. This paper proposes a new opinion mining technique using Support Vector Machine (SVM) and NLP tools on newspaper headlines. Relative words are generated using Stanford CoreNLP, which is passed to SVM using count vectorizer. On comparing three models using confusion matrix, results indicate that Tf-idf and Linear SVM provides better accuracy for smaller dataset. While for larger dataset, SGD and linear SVM model outperform other models.
Exploring Semantic Question Generation Methodology and a Case Study for Algor...IJCI JOURNAL
Assessment of student performance is one of the most important tasks in the educational process. Thus, formulating questions and creating tests takes the instructor a lot of time and effort. However, the time spent for learning acquisition and on exam preparation could be utilized in better ways. With the technical development in representing and linking data, ontologies have been used in academic fields to represent the terms in a field by defining concepts and categories classifies the subject. Also, the emergence of such methods that represent the data and link it logically contributed to the creation of methods and tools for creating questions. These tools can be used in existing learning systems to provide effective solutions to assist the teacher in creating test questions. This research paper introduces a semantic methodology for automating question generation in the domain of Algorithms. The primary objective of this approach is to support instructors in effectively incorporating automatically generated questions into their instructional practice, thereby enhancing the teaching and learning experience.
A hybrid composite features based sentence level sentiment analyzerIAESIJAI
Current lexica and machine learning based sentiment analysis approaches
still suffer from a two-fold limitation. First, manual lexicon construction and
machine training is time consuming and error-prone. Second, the
prediction’s accuracy entails sentences and their corresponding training text
should fall under the same domain. In this article, we experimentally
evaluate four sentiment classifiers, namely support vector machines (SVMs),
Naive Bayes (NB), logistic regression (LR) and random forest (RF). We
quantify the quality of each of these models using three real-world datasets
that comprise 50,000 movie reviews, 10,662 sentences, and 300 generic
movie reviews. Specifically, we study the impact of a variety of natural
language processing (NLP) pipelines on the quality of the predicted
sentiment orientations. Additionally, we measure the impact of incorporating
lexical semantic knowledge captured by WordNet on expanding original
words in sentences. Findings demonstrate that the utilizing different NLP
pipelines and semantic relationships impacts the quality of the sentiment
analyzers. In particular, results indicate that coupling lemmatization and
knowledge-based n-gram features proved to produce higher accuracy results.
With this coupling, the accuracy of the SVM classifier has improved to
90.43%, while it was 86.83%, 90.11%, 86.20%, respectively using the three
other classifiers.
Architecture of an ontology based domain-specific natural language question a...IJwest
Question answering (QA) system aims at retrieving precise information from a large collection of
documents against a query. This paper describes the architecture of a Natural Language Question
Answering (NLQA) system for a specifi
c domain based on the ontological information, a step towards
semantic web question answering. The proposed architecture defines four basic modules suitable for
enhancing current QA capabilities with the ability of processing complex questions. The first m
odule was
the question processing, which analyses and classifies the question and also reformulates the user query.
The second module allows the process of retrieving the relevant documents. The next module processes the
retrieved documents, and the last m
odule performs the extraction and generation of a response. Natural
language processing techniques are used for processing the question and documents and also for answer
extraction. Ontology and domain knowledge are used for reformulating queries and ident
ifying the
relations. The aim of the system is to generate short and specific answer to the question that is asked in the
natural language in a specific domain. We have achieved 94 % accuracy of natural language question
answering in our implementation
Development of system for generating questions, answers, distractors using tr...IJECEIAES
The goal of this article is to develop a multiple-choice questions generation system that has a number of advantages, including quick scoring, consistent grading, and a short exam period. To overcome this difficulty, we suggest treating the problem of question creation as a sequence-to-sequence learning problem, where a sentence from a text passage can directly mapped to a question. Our approach is data-driven, which eliminates the need for manual rule implementation. This strategy is more effective and gets rid of potential errors that could result from incorrect human input. Our work on question generation, particularly the usage of the transformer model, has been impacted by recent developments in a number of domains, including neural machine translation, generalization, and picture captioning.
Development of computer-based learning system for learning behavior analyticsnooriasukmaningtyas
This paper aims to analyze the learning behavior of Thai learners by using a computer-based learning system for English writing. Three main objectives were set: the development of a computer-based learning system, automatic behavior data collection, and learning behavior analytics. Firstly, the system is developed under a multidisciplinary idea that is designed to integrate two concepts between the self-regulated learning model and components of natural language processing. The integration design encourages self-learning in the digital learning environment and supports appropriate English writing by the provided component selection. Second, the system automatically collects the writing behavior of a group of Thai learners. The data collected are necessary input for the process of learning analytics. Third, the writing behaviors data were analyzed to find the learning behavioral patterns of the learners. For learning analytics, behavior sequential analysis was used to analyze the learning logs from the system. The 31 undergraduate students are participated to record writing behaviors via the system. The learning patterns in relation to grammatical skills were compared between three groups: basic, intermediate, and advanced levels. The learning behavior patterns of the three groups are different that use for reflecting learners and improving the learning materials or curriculum.
Autonomous Educational Testing System Using Unsupervised Feature Learning.IJITE
With the increase of ubiquitous data all over the internet, intelligent classroom systems that integrate
traditional learning techniques with modern e-learning tools have become quite popular and necessary
today. Although a substantial amount of work has been done in the field of e-learning, specifically in
automation of objective question and answer evaluation, personalized learning, adaptive evaluation
systems, the field of qualitative analysis of a student’s subjective paragraph answers remains unexplored to
a large extent.
The traditional board, chalk, talk based classroom scenario involves a teacher setting question papers
based on the concepts taught, checks the answers written by students manually and thus evaluates the
students’ performance. However, setting question papers remains a time consuming process with the
teacher having to bother about question quality, level of difficulty and redundancy. In addition the process
of manually correcting students’ answers is a cumbersome and tedious task especially where the class size
is large.
In this paper, we put forth the design, analysis and implementation details along with some experimental
outputs to build a system that integrates all the above mentioned tasks with minimal teacher involvement
that not only automates the traditional classroom scenario but also overcomes its inherent shortcomings
and fallacies.
Improving Sentiment Analysis of Short Informal Indonesian Product Reviews usi...TELKOMNIKA JOURNAL
Sentiment analysis in short informal texts like product reviews is more challenging. Short texts are
sparse, noisy, and lack of context information. Traditional text classification methods may not be suitable
for analyzing sentiment of short texts given all those difficulties. A common approach to overcome these
problems is to enrich the original texts with additional semantics to make it appear like a large document of
text. Then, traditional classification methods can be applied to it. In this study, we developed an automatic
sentiment analysis system of short informal Indonesian texts using Naïve Bayes and Synonym Based
Feature Expansion. The system consists of three main stages, preprocessing and normalization, features
expansion and classification. After preprocessing and normalization, we utilize Kateglo to find some
synonyms of every words in original texts and append them. Finally, the text is classified using Naïve
Bayes. The experiment shows that the proposed method can improve the performance of sentiment
analysis of short informal Indonesian product reviews. The best sentiment classification performance using
proposed feature expansion is obtained by accuracy of 98%.The experiment also show that feature
expansion will give higher improvement in small number of training data than in the large number of them.
Random forest application on cognitive level classification of E-learning co...IJECEIAES
The e-learning is the primary method of learning for most learners after the regular academics studies. The knowledge delivery through E-learning technologies increased exponentially over the years because of the advancement in internet and e-learning technologies. Knowledge delivery to some people would never have been possible without the e-learning technologies. Most of the working professional do focused studies for carrier advancement, promotion or to improve the domain knowledge. These learner can find many free e-learning web sites from the internet easily in the domain of interest. However it is quite difficult to find the best e-learning content suitable for their learning based on their domain knowledge level. User spent most of the time figuring out the right content from a plethora of available content and end up learning nothing. An intelligent framework using machine learning algorithms with random forest Classifier is proposed to address this issue, which classifies the e-learning content based on its difficulty levels and provide the learner the best content suitable based on the knowledge level .The frame work is trained with the data set collected from multiple popular e-learning web sites. The model is tested with real time e-learning web sites links and found that the e-contents in the web sites are recommended to the user based on its difficulty levels as beginner level, intermediate level and advanced level.
A scoring rubric for automatic short answer grading systemTELKOMNIKA JOURNAL
During the past decades, researches about automatic grading have become an interesting issue. These studies focuses on how to make machines are able to help human on assessing students’ learning outcomes. Automatic grading enables teachers to assess student's answers with more objective, consistent, and faster. Especially for essay model, it has two different types, i.e. long essay and short answer. Almost of the previous researches merely developed automatic essay grading (AEG) instead of automatic short answer grading (ASAG). This study aims to assess the sentence similarity of short answer to the questions and answers in Indonesian without any language semantic's tool. This research uses pre-processing steps consisting of case folding, tokenization, stemming, and stopword removal. The proposed approach is a scoring rubric obtained by measuring the similarity of sentences using the string-based similarity methods and the keyword matching process. The dataset used in this study consists of 7 questions, 34 alternative reference answers and 224 student’s answers. The experiment results show that the proposed approach is able to achieve a correlation value between 0.65419 up to 0.66383 at Pearson's correlation, with Mean Absolute Error (푀퐴퐸) value about 0.94994 until 1.24295. The proposed approach also leverages the correlation value and decreases the error value in each method.
Single document keywords extraction in Bahasa Indonesia using phrase chunkingTELKOMNIKA JOURNAL
Keywords help readers to understand the idea of a document quickly. Unfortunately, considerable time and effort are often needed to come up with a good set of keywords manually. This research focused on generating keywords from a document automatically using phrase chunking. Firstly, we collected part of speech patterns from a collection of documents. Secondly, we used those patterns to extract candidate keywords from the abstract and the content of a document. Finally, keywords are selected from the candidates based on the number of words in the keyword phrases and some scenarios involving candidate reduction and sorting. We evaluated the result of each scenario using precision, recall, and F-measure. The experiment results show: i) shorter-phrase keywords with string reduction extracted from the abstract and sorted by frequency provides the highest score, ii) in every proposed scenario, extracting keywords using the abstract always presents a better result, iii) using shorter-phrase patterns in keywords extraction gives better score in comparison to using all phrase patterns, iv) sorting scenarios based on the multiplication of candidate frequencies and the weight of the phrase patterns offer better results.
INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS cscpconf
In education, the use of electronic (E) examination systems is not a novel idea, as Eexamination systems have been used to conduct objective assessments for the last few years. This research deals with randomly designed E-examinations and proposes an E-assessment system that can be used for subjective questions. This system assesses answers to subjective questions by finding a matching ratio for the keywords in instructor and student answers. The matching ratio is achieved based on semantic and document similarity. The assessment system is composed of four modules: preprocessing, keyword expansion, matching, and grading. A survey and case study were used in the research design to validate the proposed system. The examination assessment system will help instructors to save time, costs, and resources, while increasing efficiency and improving the productivity of exam setting and assessments.
A template based algorithm for automatic summarization and dialogue managemen...eSAT Journals
Abstract This paper describes an automated approach for extracting significant and useful events from unstructured text. The goal of research is to come out with a methodology which helps in extracting important events such as dates, places, and subjects of interest. It would be also convenient if the methodology helps in presenting the users with a shorter version of the text which contain all non-trivial information. We also discuss implementation of algorithms which exactly does this task, developed by us. Key Words: Cosine Similarity, Information, Natural Language, Summarization, Text Mining
New instances classification framework on Quran ontology applied to question ...TELKOMNIKA JOURNAL
Instances classification with the small dataset for Quran ontology is the current research problem which appears in Quran ontology development. The existing classification approach used machine learning: Backpropagation Neural Network. However, this method has a drawback; if the training set amount is small, then the classifier accuracy could decline. Unfortunately, Holy Quran has a small corpus. Based on this problem, our study aims to formulate new instances classification framework for small training corpus applied to semantic question answering system. As a result, the instances classification framework consists of several essential components: pre-processing, morphology analysis, semantic analysis, feature extraction, instances classification with Radial Basis Function Networks algorithm, and the transformation module. This algorithm is chosen since it robustness to noisy data and has an excellent achievement to handle small dataset. Furthermore, document processing module on question answering system is used to access instances classification result in Quran ontology.
Topic Discovery of Online Course Reviews Using LDA with Leveraging Reviews He...IJECEIAES
Despite the popularity of the Massive Open Online Courses, small-scale research has been done to understand the factors that influence the teaching-learning process through the massive online platform. Using topic modeling approach, our results show terms with prior knowledge to understand e.g.: Chuck as the instructor name. So, we proposed the topic modeling approach on helpful subjective reviews. The results show five influential factors: “learn easy excellent class program”, “python learn class easy lot”, “Program learn easy python time game”, and “learn class python time game”. Also, research results showed that the proposed method improved the perplexity score on the LDA model.
COMPARISON INTELLIGENT ELECTRONIC ASSESSMENT WITH TRADITIONAL ASSESSMENT FOR ...cseij
In education, the use of electronic (E) examination systems is not a novel idea, as E-examination systems
have been used to conduct objective assessments for the last few years. This research deals with randomly
designed E-examinations and proposes an E-assessment system that can be used for subjective questions.
This system assesses answers to subjective questions by finding a matching ratio for the keywords in
instructor and student answers.
Opinion mining on newspaper headlines using SVM and NLPIJECEIAES
Opinion Mining also known as Sentiment Analysis, is a technique or procedure which uses Natural Language processing (NLP) to classify the outcome from text. There are various NLP tools available which are used for processing text data. Multiple research have been done in opinion mining for online blogs, Twitter, Facebook etc. This paper proposes a new opinion mining technique using Support Vector Machine (SVM) and NLP tools on newspaper headlines. Relative words are generated using Stanford CoreNLP, which is passed to SVM using count vectorizer. On comparing three models using confusion matrix, results indicate that Tf-idf and Linear SVM provides better accuracy for smaller dataset. While for larger dataset, SGD and linear SVM model outperform other models.
Exploring Semantic Question Generation Methodology and a Case Study for Algor...IJCI JOURNAL
Assessment of student performance is one of the most important tasks in the educational process. Thus, formulating questions and creating tests takes the instructor a lot of time and effort. However, the time spent for learning acquisition and on exam preparation could be utilized in better ways. With the technical development in representing and linking data, ontologies have been used in academic fields to represent the terms in a field by defining concepts and categories classifies the subject. Also, the emergence of such methods that represent the data and link it logically contributed to the creation of methods and tools for creating questions. These tools can be used in existing learning systems to provide effective solutions to assist the teacher in creating test questions. This research paper introduces a semantic methodology for automating question generation in the domain of Algorithms. The primary objective of this approach is to support instructors in effectively incorporating automatically generated questions into their instructional practice, thereby enhancing the teaching and learning experience.
A hybrid composite features based sentence level sentiment analyzerIAESIJAI
Current lexica and machine learning based sentiment analysis approaches
still suffer from a two-fold limitation. First, manual lexicon construction and
machine training is time consuming and error-prone. Second, the
prediction’s accuracy entails sentences and their corresponding training text
should fall under the same domain. In this article, we experimentally
evaluate four sentiment classifiers, namely support vector machines (SVMs),
Naive Bayes (NB), logistic regression (LR) and random forest (RF). We
quantify the quality of each of these models using three real-world datasets
that comprise 50,000 movie reviews, 10,662 sentences, and 300 generic
movie reviews. Specifically, we study the impact of a variety of natural
language processing (NLP) pipelines on the quality of the predicted
sentiment orientations. Additionally, we measure the impact of incorporating
lexical semantic knowledge captured by WordNet on expanding original
words in sentences. Findings demonstrate that the utilizing different NLP
pipelines and semantic relationships impacts the quality of the sentiment
analyzers. In particular, results indicate that coupling lemmatization and
knowledge-based n-gram features proved to produce higher accuracy results.
With this coupling, the accuracy of the SVM classifier has improved to
90.43%, while it was 86.83%, 90.11%, 86.20%, respectively using the three
other classifiers.
Architecture of an ontology based domain-specific natural language question a...IJwest
Question answering (QA) system aims at retrieving precise information from a large collection of
documents against a query. This paper describes the architecture of a Natural Language Question
Answering (NLQA) system for a specifi
c domain based on the ontological information, a step towards
semantic web question answering. The proposed architecture defines four basic modules suitable for
enhancing current QA capabilities with the ability of processing complex questions. The first m
odule was
the question processing, which analyses and classifies the question and also reformulates the user query.
The second module allows the process of retrieving the relevant documents. The next module processes the
retrieved documents, and the last m
odule performs the extraction and generation of a response. Natural
language processing techniques are used for processing the question and documents and also for answer
extraction. Ontology and domain knowledge are used for reformulating queries and ident
ifying the
relations. The aim of the system is to generate short and specific answer to the question that is asked in the
natural language in a specific domain. We have achieved 94 % accuracy of natural language question
answering in our implementation
Development of system for generating questions, answers, distractors using tr...IJECEIAES
The goal of this article is to develop a multiple-choice questions generation system that has a number of advantages, including quick scoring, consistent grading, and a short exam period. To overcome this difficulty, we suggest treating the problem of question creation as a sequence-to-sequence learning problem, where a sentence from a text passage can directly mapped to a question. Our approach is data-driven, which eliminates the need for manual rule implementation. This strategy is more effective and gets rid of potential errors that could result from incorrect human input. Our work on question generation, particularly the usage of the transformer model, has been impacted by recent developments in a number of domains, including neural machine translation, generalization, and picture captioning.
Development of computer-based learning system for learning behavior analyticsnooriasukmaningtyas
This paper aims to analyze the learning behavior of Thai learners by using a computer-based learning system for English writing. Three main objectives were set: the development of a computer-based learning system, automatic behavior data collection, and learning behavior analytics. Firstly, the system is developed under a multidisciplinary idea that is designed to integrate two concepts between the self-regulated learning model and components of natural language processing. The integration design encourages self-learning in the digital learning environment and supports appropriate English writing by the provided component selection. Second, the system automatically collects the writing behavior of a group of Thai learners. The data collected are necessary input for the process of learning analytics. Third, the writing behaviors data were analyzed to find the learning behavioral patterns of the learners. For learning analytics, behavior sequential analysis was used to analyze the learning logs from the system. The 31 undergraduate students are participated to record writing behaviors via the system. The learning patterns in relation to grammatical skills were compared between three groups: basic, intermediate, and advanced levels. The learning behavior patterns of the three groups are different that use for reflecting learners and improving the learning materials or curriculum.
Convolutional recurrent neural network with template based representation for...IJECEIAES
Complex Question answering system is developed to answer different types of questions accurately. Initially the question from the natural language is transformed to an internal representation which captures the semantics and intent of the question. In the proposed work, internal representation is provided with templates instead of using synonyms or keywords. Then for each internal representation, it is mapped to relevant query against the knowledge base. In present work, the Template representation based Convolutional Recurrent Neural Network (T-CRNN) is proposed for selecting answer in Complex Question Answering (CQA) framework. Recurrent neural network is used to obtain the exact correlation between answers and questions and the semantic matching among the collection of answers. Initially, the process of learning is accomplished through Convolutional Neural Network (CNN) which represents the questions and answers separately. Then the representation with fixed length is produced for each question with the help of fully connected neural network. In order to design the semantic matching between the answers, the representation of Question Answer (QA) pair is given into the Recurrent Neural Network (RNN). Finally, for the given question, the correctly correlated answers are identified with the softmax classifier.
Automatic Generation of Multiple Choice Questions using Surface-based Semanti...CSCJournals
Multiple Choice Questions (MCQs) are a popular large-scale assessment tool. MCQs make it much easier for test-takers to take tests and for examiners to interpret their results; however, they are very expensive to compile manually, and they often need to be produced on a large scale and within short iterative cycles. We examine the problem of automated MCQ generation with the help of unsupervised Relation Extraction, a technique used in a number of related Natural Language Processing problems. Unsupervised Relation Extraction aims to identify the most important named entities and terminology in a document and then recognize semantic relations between them, without any prior knowledge as to the semantic types of the relations or their specific linguistic realization. We investigated a number of relation extraction patterns and tested a number of assumptions about linguistic expression of semantic relations between named entities. Our findings indicate that an optimized configuration of our MCQ generation system is capable of achieving high precision rates, which are much more important than recall in the automatic generation of MCQs. Its enhancement with linguistic knowledge further helps to produce significantly better patterns. We furthermore carried out a user-centric evaluation of the system, where subject domain experts from biomedical domain evaluated automatically generated MCQ items in terms of readability, usefulness of semantic relations, relevance, acceptability of questions and distractors and overall MCQ usability. The results of this evaluation make it possible for us to draw conclusions about the utility of the approach in practical e-Learning applications.
Deep learning based Arabic short answer grading in serious gamesIJECEIAES
Automatic short answer grading (ASAG) has become part of natural language processing problems. Modern ASAG systems start with natural language preprocessing and end with grading. Researchers started experimenting with machine learning in the preprocessing stage and deep learning techniques in automatic grading for English. However, little research is available on automatic grading for Arabic. Datasets are important to ASAG, and limited datasets are available in Arabic. In this research, we have collected a set of questions, answers, and associated grades in Arabic. We have made this dataset publicly available. We have extended to Arabic the solutions used for English ASAG. We have tested how automatic grading works on answers in Arabic provided by schoolchildren in 6th grade in the context of serious games. We found out those schoolchildren providing answers that are 5.6 words long on average. On such answers, deep learning-based grading has achieved high accuracy even with limited training data. We have tested three different recurrent neural networks for grading. With a transformer, we have achieved an accuracy of 95.67%. ASAG for school children will help detect children with learning problems early. When detected early, teachers can solve learning problems easily. This is the main purpose of this research.
CLASSIFICATION OF QUESTIONS AND LEARNING OUTCOME STATEMENTS (LOS) INTO BLOOM’...IJMIT JOURNAL
Bloom’s Taxonomy (BT) have been used to classify the objectives of learning outcome by dividing the
learning into three different domains; the cognitive domain, the effective domain and the psychomotor
domain. In this paper, we are introducing a new approach to classify the questions and learning outcome
statements (LOS) into Blooms taxonomy (BT) and to verify BT verb lists, which are being cited and used by
academicians to write questions and (LOS). An experiment was designed to investigate the semantic
relationship between the action verbs used in both questions and LOS to obtain more accurate
classification of the levels of BT. A sample of 775 different action verbs collected from different universities
allows us to measure an accurate and clear-cut cognitive level for the action verb. It is worth mentioning
that natural language processing techniques were used to develop our rules as to induce the questions into
chunks in order to extract the action verbs. Our proposed solution was able to classify the action verb into
a precise level of the cognitive domain. We, on our side, have tested and evaluated our proposed solution
using confusion matrix. The results of evaluation tests yielded 97% for the macro average of precision and
90% for F1. Thus, the outcome of the research suggests that it is crucial to analyse and verify the action
verbs cited and used by academicians to write LOS and classify their questions based on blooms taxonomy
in order to obtain a definite and more accurate classification.
CLASSIFICATION OF QUESTIONS AND LEARNING OUTCOME STATEMENTS (LOS) INTO BLOOM’...IJMIT JOURNAL
Bloom’s Taxonomy (BT) have been used to classify the objectives of learning outcome by dividing the learning into three different domains; the cognitive domain, the effective domain and the psychomotor domain. In this paper, we are introducing a new approach to classify the questions and learning outcome
statements (LOS) into Blooms taxonomy (BT) and to verify BT verb lists, which are being cited and used by academicians to write questions and (LOS). An experiment was designed to investigate the semantic relationship between the action verbs used in both questions and LOS to obtain more accurate
classification of the levels of BT. A sample of 775 different action verbs collected from different universities allows us to measure an accurate and clear-cut cognitive level for the action verb. It is worth mentioning that natural language processing techniques were used to develop our rules as to induce the questions into
chunks in order to extract the action verbs. Our proposed solution was able to classify the action verb into a precise level of the cognitive domain. We, on our side, have tested and evaluated our proposed solution using confusion matrix. The results of evaluation tests yielded 97% for the macro average of precision and 90% for F1. Thus, the outcome of the research suggests that it is crucial to analyse and verify the action
verbs cited and used by academicians to write LOS and classify their questions based on blooms taxonomy in order to obtain a definite and more accurate classification.
Autonomous Educational Testing System Using Unsupervised Feature LearningIJITE
With the increase of ubiquitous data all over the internet, intelligent classroom systems that integrate
traditional learning techniques with modern e-learning tools have become quite popular and necessary
today. Although a substantial amount of work has been done in the field of e-learning, specifically in
automation of objective question and answer evaluation, personalized learning, adaptive evaluation
systems, the field of qualitative analysis of a student’s subjective paragraph answers remains unexplored to
a large extent.
The traditional board, chalk, talk based classroom scenario involves a teacher setting question papers
based on the concepts taught, checks the answers written by students manually and thus evaluates the
students’ performance. However, setting question papers remains a time consuming process with the
teacher having to bother about question quality, level of difficulty and redundancy. In addition the process
of manually correcting students’ answers is a cumbersome and tedious task especially where the class size
is large.
In this paper, we put forth the design, analysis and implementation details along with some experimental
outputs to build a system that integrates all the above mentioned tasks with minimal teacher involvement
that not only automates the traditional classroom scenario but also overcomes its inherent shortcomings
and fallacies.
Autonomous Educational Testing System Using Unsupervised Feature LearningIJITE
With the increase of ubiquitous data all over the internet, intelligent classroom systems that integrate
traditional learning techniques with modern e-learning tools have become quite popular and necessary
today. Although a substantial amount of work has been done in the field of e-learning, specifically in
automation of objective question and answer evaluation, personalized learning, adaptive evaluation
systems, the field of qualitative analysis of a student’s subjective paragraph answers remains unexplored to
a large extent.
The traditional board, chalk, talk based classroom scenario involves a teacher setting question papers
based on the concepts taught, checks the answers written by students manually and thus evaluates the
students’ performance. However, setting question papers remains a time consuming process with the
teacher having to bother about question quality, level of difficulty and redundancy. In addition the process
of manually correcting students’ answers is a cumbersome and tedious task especially where the class size
is large.
In this paper, we put forth the design, analysis and implementation details along with some experimental
outputs to build a system that integrates all the above mentioned tasks with minimal teacher involvement
that not only automates the traditional classroom scenario but also overcomes its inherent shortcomings
and fallacies.
Knowledge Management in the Cloud: Benefits and RisksEditor IJCATR
The success of organizations largely depends on continual investment in learning and acquiring new knowledge that creates
new businesses and improve existing performance. So, using Knowledge management must result in better achieving, or even
exceeding, organizations objectives. The purpose of knowledge management must not be to just become more knowledgeable, but to
be able to create, transfer and apply knowledge with the purpose of better achieving objectives. As new technologies and paradigms
emerge, businesses have to make new efforts to properly get aligned with them, especially in knowledge management area. Today the
Cloud Computing paradigm is becoming more and more popular, due to the vast decrease in time, cost and effort for meeting software
development needs. It also provides a great means for gathering and redistributing knowledge. In this paper, we will discuss the
benefits and risks of using cloud computing in knowledge management systems.
Effect of Query Formation on Web Search Engine Resultskevig
Query in a search engine is generally based on natural language. A query can be expressed in more than
one way without changing its meaning as it depends on thinking of human being at a particular moment.
Aim of the searcher is to get most relevant results immaterial of how the query has been expressed. In the
present paper, we have examined the results of search engine for change in coverage and similarity of first
few results when a query is entered in two semantically same but in different formats. Searching has been
made through Google search engine. Fifteen pairs of queries have been chosen for the study. The t-test has
been used for the purpose and the results have been checked on the basis of total documents found,
similarity of first five and first ten documents found in the results of a query entered in two different
formats. It has been found that the total coverage is same but first few results are significantly different.
Investigations of the Distributions of Phonemic Durations in Hindi and Dogrikevig
Speech generation is one of the most important areas of research in speech signal processing which is now gaining a serious attention. Speech is a natural form of communication in all living things. Computers with the ability to understand speech and speak with a human like voice are expected to contribute to the development of more natural man-machine interface. However, in order to give those functions that are even closer to those of human beings, we must learn more about the mechanisms by which speech is produced and perceived, and develop speech information processing technologies that can generate a more natural sounding systems. The so described field of stud, also called speech synthesis and more prominently acknowledged as text-to-speech synthesis, originated in the mid eighties because of the emergence of DSP and the rapid advancement of VLSI techniques. To understand this field of speech, it is necessary to understand the basic theory of speech production. Every language has different phonetic alphabets and a different set of possible phonemes and their combinations.
For the analysis of the speech signal, we have carried out the recording of five speakers in Dogri (3 male and 5 females) and eight speakers in Hindi language (4 male and 4 female). For estimating the durational distributions, the mean of mean of ten instances of vowels of each speaker in both the languages has been calculated. Investigations have shown that the two durational distributions differ significantly with respect to mean and standard deviation. The duration of phoneme is speaker dependent. The whole investigation can be concluded with the end result that almost all the Dogri phonemes have shorter duration, in comparison to Hindi phonemes. The period in milli seconds of same phonemes when uttered in Hindi were found to be longer compared to when they were spoken by a person with Dogri as his mother tongue. There are many applications which are directly of indirectly related to the research being carried out. For instance the main application may be for transforming Dogri speech into Hindi and vice versa, and further utilizing this application, we can develop a speech aid to teach Dogri to children. The results may also be useful for synthesizing the phonemes of Dogri using the parameters of the phonemes of Hindi and for building large vocabulary speech recognition systems.
May 2024 - Top10 Cited Articles in Natural Language Computingkevig
Natural Language Processing is a programmed approach to analyze text that is based on both a set of theories and a set of technologies. This forum aims to bring together researchers who have designed and build software that will analyze, understand, and generate languages that humans use naturally to address computers.
Effect of Singular Value Decomposition Based Processing on Speech Perceptionkevig
Speech is an important biological signal for primary mode of communication among human being and also the most natural and efficient form of exchanging information among human in speech. Speech processing is the most important aspect in signal processing. In this paper the theory of linear algebra called singular value decomposition (SVD) is applied to the speech signal. SVD is a technique for deriving important parameters of a signal. The parameters derived using SVD may further be reduced by perceptual evaluation of the synthesized speech using only perceptually important parameters, where the speech signal can be compressed so that the information can be transformed into compressed form without losing its quality. This technique finds wide applications in speech compression, speech recognition, and speech synthesis. The objective of this paper is to investigate the effect of SVD based feature selection of the input speech on the perception of the processed speech signal. The speech signal which is in the form of vowels \a\, \e\, \u\ were recorded from each of the six speakers (3 males and 3 females). The vowels for the six speakers were analyzed using SVD based processing and the effect of the reduction in singular values was investigated on the perception of the resynthesized vowels using reduced singular values. Investigations have shown that the number of singular values can be drastically reduced without significantly affecting the perception of the vowels.
Identifying Key Terms in Prompts for Relevance Evaluation with GPT Modelskevig
Relevance evaluation of a query and a passage is essential in Information Retrieval (IR). Recently, numerous studies have been conducted on tasks related to relevance judgment using Large Language Models (LLMs) such as GPT-4,
demonstrating significant improvements. However, the efficacy of LLMs is considerably influenced by the design of the prompt. The purpose of this paper is to
identify which specific terms in prompts positively or negatively impact relevance
evaluation with LLMs. We employed two types of prompts: those used in previous
research and generated automatically by LLMs. By comparing the performance of
these prompts in both few-shot and zero-shot settings, we analyze the influence of
specific terms in the prompts. We have observed two main findings from our study.
First, we discovered that prompts using the term ‘answer’ lead to more effective
relevance evaluations than those using ‘relevant.’ This indicates that a more direct
approach, focusing on answering the query, tends to enhance performance. Second,
we noted the importance of appropriately balancing the scope of ‘relevance.’ While
the term ‘relevant’ can extend the scope too broadly, resulting in less precise evaluations, an optimal balance in defining relevance is crucial for accurate assessments.
The inclusion of few-shot examples helps in more precisely defining this balance.
By providing clearer contexts for the term ‘relevance,’ few-shot examples contribute
to refine relevance criteria. In conclusion, our study highlights the significance of
carefully selecting terms in prompts for relevance evaluation with LLMs.
Identifying Key Terms in Prompts for Relevance Evaluation with GPT Modelskevig
Relevance evaluation of a query and a passage is essential in Information Retrieval (IR). Recently, numerous studies have been conducted on tasks related to relevance judgment using Large Language Models (LLMs) such as GPT-4, demonstrating significant improvements. However, the efficacy of LLMs is considerably influenced by the design of the prompt. The purpose of this paper is to identify which specific terms in prompts positively or negatively impact relevance evaluation with LLMs. We employed two types of prompts: those used in previous research and generated automatically by LLMs. By comparing the performance of these prompts in both few-shot and zero-shot settings, we analyze the influence of specific terms in the prompts. We have observed two main findings from our study. First, we discovered that prompts using the term ‘answer’ lead to more effective relevance evaluations than those using ‘relevant.’ This indicates that a more direct approach, focusing on answering the query, tends to enhance performance. Second, we noted the importance of appropriately balancing the scope of ‘relevance.’ While the term ‘relevant’ can extend the scope too broadly, resulting in less precise evaluations, an optimal balance in defining relevance is crucial for accurate assessments. The inclusion of few-shot examples helps in more precisely defining this balance. By providing clearer contexts for the term ‘relevance,’ few-shot examples contribute to refine relevance criteria. In conclusion, our study highlights the significance of carefully selecting terms in prompts for relevance evaluation with LLMs.
In recent years, great advances have been made in the speed, accuracy, and coverage of automatic word
sense disambiguator systems that, given a word appearing in a certain context, can identify the sense of
that word. In this paper we consider the problem of deciding whether same words contained in different
documents are related to the same meaning or are homonyms. Our goal is to improve the estimate of the
similarity of documents in which some words may be used with different meanings. We present three new
strategies for solving this problem, which are used to filter out homonyms from the similarity computation.
Two of them are intrinsically non-semantic, whereas the other one has a semantic flavor and can also be
applied to word sense disambiguation. The three strategies have been embedded in an article document
recommendation system that one of the most important Italian ad-serving companies offers to its customers.
Genetic Approach For Arabic Part Of Speech Taggingkevig
With the growing number of textual resources available, the ability to understand them becomes critical.
An essential first step in understanding these sources is the ability to identify the parts-of-speech in each
sentence. Arabic is a morphologically rich language, which presents a challenge for part of speech
tagging. In this paper, our goal is to propose, improve, and implement a part-of-speech tagger based on a
genetic algorithm. The accuracy obtained with this method is comparable to that of other probabilistic
approaches.
Rule Based Transliteration Scheme for English to Punjabikevig
Machine Transliteration has come out to be an emerging and a very important research area in the field of
machine translation. Transliteration basically aims to preserve the phonological structure of words. Proper
transliteration of name entities plays a very significant role in improving the quality of machine translation.
In this paper we are doing machine transliteration for English-Punjabi language pair using rule based
approach. We have constructed some rules for syllabification. Syllabification is the process to extract or
separate the syllable from the words. In this we are calculating the probabilities for name entities (Proper
names and location). For those words which do not come under the category of name entities, separate
probabilities are being calculated by using relative frequency through a statistical machine translation
toolkit known as MOSES. Using these probabilities we are transliterating our input text from English to
Punjabi.
Improving Dialogue Management Through Data Optimizationkevig
In task-oriented dialogue systems, the ability for users to effortlessly communicate with machines and computers through natural language stands as a critical advancement. Central to these systems is the dialogue manager, a pivotal component tasked with navigating the conversation to effectively meet user goals by selecting the most appropriate response. Traditionally, the development of sophisticated dialogue management has embraced a variety of methodologies, including rule-based systems, reinforcement learning, and supervised learning, all aimed at optimizing response selection in light of user inputs. This research casts a spotlight on the pivotal role of data quality in enhancing the performance of dialogue managers. Through a detailed examination of prevalent errors within acclaimed datasets, such as Multiwoz 2.1 and SGD, we introduce an innovative synthetic dialogue generator designed to control the introduction of errors precisely. Our comprehensive analysis underscores the critical impact of dataset imperfections, especially mislabeling, on the challenges inherent in refining dialogue management processes.
Document Author Classification using Parsed Language Structurekevig
Over the years there has been ongoing interest in detecting authorship of a text based on statistical properties of the text, such as by using occurrence rates of noncontextual words. In previous work, these techniques have been used, for example, to determine authorship of all of The Federalist Papers. Such methods may be useful in more modern times to detect fake or AI authorship. Progress in statistical natural language parsers introduces the possibility of using grammatical structure to detect authorship. In this paper we explore a new possibility for detecting authorship using grammatical structural information extracted using a statistical natural language parser. This paper provides a proof of concept, testing author classification based on grammatical structure on a set of “proof texts,” The Federalist Papers and Sanditon which have been as test cases in previous authorship detection studies. Several features extracted from the statisticalnaturallanguage parserwere explored: all subtrees of some depth from any level; rooted subtrees of some depth, part of speech, and part of speech by level in the parse tree. It was found to be helpful to project the features into a lower dimensional space. Statistical experiments on these documents demonstrate that information from a statistical parser can, in fact, assist in distinguishing authors.
Rag-Fusion: A New Take on Retrieval Augmented Generationkevig
Infineon has identified a need for engineers, account managers, and customers to rapidly obtain product information. This problem is traditionally addressed with retrieval-augmented generation (RAG) chatbots, but in this study, I evaluated the use of the newly popularized RAG-Fusion method. RAG-Fusion combines RAG and reciprocal rank fusion (RRF) by generating multiple queries, reranking them with reciprocal scores and fusing the documents and scores. Through manually evaluating answers on accuracy, relevance, and comprehensiveness, I found that RAG-Fusion was able to provide accurate and comprehensive answers due to the generated queries contextualizing the original query from various perspectives. However, some answers strayed off topic when the generated queries' relevance to the original query is insufficient. This research marks significant progress in artificial intelligence (AI) and natural language processing (NLP) applications and demonstrates transformations in a global and multi-industry context.
Performance, Energy Consumption and Costs: A Comparative Analysis of Automati...kevig
The common practice in Machine Learning research is to evaluate the top-performing models based on their performance. However, this often leads to overlooking other crucial aspects that should be given careful consideration. In some cases, the performance differences between various approaches may be insignificant, whereas factors like production costs, energy consumption, and carbon footprint should be taken into account. Large Language Models (LLMs) are widely used in academia and industry to address NLP problems. In this study, we present a comprehensive quantitative comparison between traditional approaches (SVM-based) and more recent approaches such as LLM (BERT family models) and generative models (GPT2 and LLAMA2), using the LexGLUE benchmark. Our evaluation takes into account not only performance parameters (standard indices), but also alternative measures such as timing, energy consumption and costs, which collectively contribute to the carbon footprint. To ensure a complete analysis, we separately considered the prototyping phase (which involves model selection through training-validation-test iterations) and the in-production phases. These phases follow distinct implementation procedures and require different resources. The results indicate that simpler algorithms often achieve performance levels similar to those of complex models (LLM and generative models), consuming much less energy and requiring fewer resources. These findings suggest that companies should consider additional considerations when choosing machine learning (ML) solutions. The analysis also demonstrates that it is increasingly necessary for the scientific world to also begin to consider aspects of energy consumption in model evaluations, in order to be able to give real meaning to the results obtained using standard metrics (Precision, Recall, F1 and so on).
Evaluation of Medium-Sized Language Models in German and English Languagekevig
Large language models (LLMs) have garnered significant attention, but the definition of “large” lacks clarity. This paper focuses on medium-sized language models (MLMs), defined as having at least six billion parameters but less than 100 billion. The study evaluates MLMs regarding zero-shot generative question answering, which requires models to provide elaborate answers without external document retrieval. The paper introduces an own test dataset and presents results from human evaluation. Results show that combining the best answers from different MLMs yielded an overall correct answer rate of 82.7% which is better than the 60.9% of ChatGPT. The best MLM achieved 71.8% and has 33B parameters, which highlights the importance of using appropriate training data for fine-tuning rather than solely relying on the number of parameters. More fine-grained feedback should be used to further improve the quality of answers. The open source community is quickly closing the gap to the best commercial models.
IMPROVING DIALOGUE MANAGEMENT THROUGH DATA OPTIMIZATIONkevig
In task-oriented dialogue systems, the ability for users to effortlessly communicate with machines and
computers through natural language stands as a critical advancement. Central to these systems is the
dialogue manager, a pivotal component tasked with navigating the conversation to effectively meet user
goals by selecting the most appropriate response. Traditionally, the development of sophisticated dialogue
management has embraced a variety of methodologies, including rule-based systems, reinforcement
learning, and supervised learning, all aimed at optimizing response selection in light of user inputs. This
research casts a spotlight on the pivotal role of data quality in enhancing the performance of dialogue
managers. Through a detailed examination of prevalent errors within acclaimed datasets, such as
Multiwoz 2.1 and SGD, we introduce an innovative synthetic dialogue generator designed to control the
introduction of errors precisely. Our comprehensive analysis underscores the critical impact of dataset
imperfections, especially mislabeling, on the challenges inherent in refining dialogue management
processes.
Document Author Classification Using Parsed Language Structurekevig
Over the years there has been ongoing interest in detecting authorship of a text based on statistical properties of the
text, such as by using occurrence rates of noncontextual words. In previous work, these techniques have been used,
for example, to determine authorship of all of The Federalist Papers. Such methods may be useful in more modern
times to detect fake or AI authorship. Progress in statistical natural language parsers introduces the possibility of
using grammatical structure to detect authorship. In this paper we explore a new possibility for detecting authorship
using grammatical structural information extracted using a statistical natural language parser. This paper provides a
proof of concept, testing author classification based on grammatical structure on a set of “proof texts,” The Federalist
Papers and Sanditon which have been as test cases in previous authorship detection studies. Several features extracted
of some depth, part of speech, and part of speech by level in the parse tree. It was found to be helpful to project the
features into a lower dimensional space. Statistical experiments on these documents demonstrate that information
from a statistical parser can, in fact, assist in distinguishing authors.
RAG-FUSION: A NEW TAKE ON RETRIEVALAUGMENTED GENERATIONkevig
Infineon has identified a need for engineers, account managers, and customers to rapidly obtain product
information. This problem is traditionally addressed with retrieval-augmented generation (RAG) chatbots,
but in this study, I evaluated the use of the newly popularized RAG-Fusion method. RAG-Fusion combines
RAG and reciprocal rank fusion (RRF) by generating multiple queries, reranking them with reciprocal
scores and fusing the documents and scores. Through manually evaluating answers on accuracy,
relevance, and comprehensiveness, I found that RAG-Fusion was able to provide accurate and
comprehensive answers due to the generated queries contextualizing the original query from various
perspectives. However, some answers strayed off topic when the generated queries' relevance to the
original query is insufficient. This research marks significant progress in artificial intelligence (AI) and
natural language processing (NLP) applications and demonstrates transformations in a global and multiindustry context
Performance, energy consumption and costs: a comparative analysis of automati...kevig
The common practice in Machine Learning research is to evaluate the top-performing models based on their
performance. However, this often leads to overlooking other crucial aspects that should be given careful
consideration. In some cases, the performance differences between various approaches may be insignificant, whereas factors like production costs, energy consumption, and carbon footprint should be taken into
account. Large Language Models (LLMs) are widely used in academia and industry to address NLP problems. In this study, we present a comprehensive quantitative comparison between traditional approaches
(SVM-based) and more recent approaches such as LLM (BERT family models) and generative models (GPT2 and LLAMA2), using the LexGLUE benchmark. Our evaluation takes into account not only performance
parameters (standard indices), but also alternative measures such as timing, energy consumption and costs,
which collectively contribute to the carbon footprint. To ensure a complete analysis, we separately considered the prototyping phase (which involves model selection through training-validation-test iterations) and
the in-production phases. These phases follow distinct implementation procedures and require different resources. The results indicate that simpler algorithms often achieve performance levels similar to those of
complex models (LLM and generative models), consuming much less energy and requiring fewer resources.
These findings suggest that companies should consider additional considerations when choosing machine
learning (ML) solutions. The analysis also demonstrates that it is increasingly necessary for the scientific
world to also begin to consider aspects of energy consumption in model evaluations, in order to be able to
give real meaning to the results obtained using standard metrics (Precision, Recall, F1 and so on).
EVALUATION OF MEDIUM-SIZED LANGUAGE MODELS IN GERMAN AND ENGLISH LANGUAGEkevig
Large language models (LLMs) have garnered significant attention, but the definition of “large” lacks
clarity. This paper focuses on medium-sized language models (MLMs), defined as having at least six
billion parameters but less than 100 billion. The study evaluates MLMs regarding zero-shot generative
question answering, which requires models to provide elaborate answers without external document
retrieval. The paper introduces an own test dataset and presents results from human evaluation. Results
show that combining the best answers from different MLMs yielded an overall correct answer rate of
82.7% which is better than the 60.9% of ChatGPT. The best MLM achieved 71.8% and has 33B
parameters, which highlights the importance of using appropriate training data for fine-tuning rather than
solely relying on the number of parameters. More fine-grained feedback should be used to further improve
the quality of answers. The open source community is quickly closing the gap to the best commercial
models.
Natural Language Processing is a programmed approach to analyze text that is based on both a set of theories and a set of technologies. This forum aims to bring together researchers who have designed and build software that will analyze, understand, and generate languages that humans use naturally to address computers.
Quality defects in TMT Bars, Possible causes and Potential Solutions.PrashantGoswami42
Maintaining high-quality standards in the production of TMT bars is crucial for ensuring structural integrity in construction. Addressing common defects through careful monitoring, standardized processes, and advanced technology can significantly improve the quality of TMT bars. Continuous training and adherence to quality control measures will also play a pivotal role in minimizing these defects.
COLLEGE BUS MANAGEMENT SYSTEM PROJECT REPORT.pdfKamal Acharya
The College Bus Management system is completely developed by Visual Basic .NET Version. The application is connect with most secured database language MS SQL Server. The application is develop by using best combination of front-end and back-end languages. The application is totally design like flat user interface. This flat user interface is more attractive user interface in 2017. The application is gives more important to the system functionality. The application is to manage the student’s details, driver’s details, bus details, bus route details, bus fees details and more. The application has only one unit for admin. The admin can manage the entire application. The admin can login into the application by using username and password of the admin. The application is develop for big and small colleges. It is more user friendly for non-computer person. Even they can easily learn how to manage the application within hours. The application is more secure by the admin. The system will give an effective output for the VB.Net and SQL Server given as input to the system. The compiled java program given as input to the system, after scanning the program will generate different reports. The application generates the report for users. The admin can view and download the report of the data. The application deliver the excel format reports. Because, excel formatted reports is very easy to understand the income and expense of the college bus. This application is mainly develop for windows operating system users. In 2017, 73% of people enterprises are using windows operating system. So the application will easily install for all the windows operating system users. The application-developed size is very low. The application consumes very low space in disk. Therefore, the user can allocate very minimum local disk space for this application.
Explore the innovative world of trenchless pipe repair with our comprehensive guide, "The Benefits and Techniques of Trenchless Pipe Repair." This document delves into the modern methods of repairing underground pipes without the need for extensive excavation, highlighting the numerous advantages and the latest techniques used in the industry.
Learn about the cost savings, reduced environmental impact, and minimal disruption associated with trenchless technology. Discover detailed explanations of popular techniques such as pipe bursting, cured-in-place pipe (CIPP) lining, and directional drilling. Understand how these methods can be applied to various types of infrastructure, from residential plumbing to large-scale municipal systems.
Ideal for homeowners, contractors, engineers, and anyone interested in modern plumbing solutions, this guide provides valuable insights into why trenchless pipe repair is becoming the preferred choice for pipe rehabilitation. Stay informed about the latest advancements and best practices in the field.
Event Management System Vb Net Project Report.pdfKamal Acharya
In present era, the scopes of information technology growing with a very fast .We do not see any are untouched from this industry. The scope of information technology has become wider includes: Business and industry. Household Business, Communication, Education, Entertainment, Science, Medicine, Engineering, Distance Learning, Weather Forecasting. Carrier Searching and so on.
My project named “Event Management System” is software that store and maintained all events coordinated in college. It also helpful to print related reports. My project will help to record the events coordinated by faculties with their Name, Event subject, date & details in an efficient & effective ways.
In my system we have to make a system by which a user can record all events coordinated by a particular faculty. In our proposed system some more featured are added which differs it from the existing system such as security.
Final project report on grocery store management system..pdfKamal Acharya
In today’s fast-changing business environment, it’s extremely important to be able to respond to client needs in the most effective and timely manner. If your customers wish to see your business online and have instant access to your products or services.
Online Grocery Store is an e-commerce website, which retails various grocery products. This project allows viewing various products available enables registered users to purchase desired products instantly using Paytm, UPI payment processor (Instant Pay) and also can place order by using Cash on Delivery (Pay Later) option. This project provides an easy access to Administrators and Managers to view orders placed using Pay Later and Instant Pay options.
In order to develop an e-commerce website, a number of Technologies must be studied and understood. These include multi-tiered architecture, server and client-side scripting techniques, implementation technologies, programming language (such as PHP, HTML, CSS, JavaScript) and MySQL relational databases. This is a project with the objective to develop a basic website where a consumer is provided with a shopping cart website and also to know about the technologies used to develop such a website.
This document will discuss each of the underlying technologies to create and implement an e- commerce website.
CFD Simulation of By-pass Flow in a HRSG module by R&R Consult.pptxR&R Consult
CFD analysis is incredibly effective at solving mysteries and improving the performance of complex systems!
Here's a great example: At a large natural gas-fired power plant, where they use waste heat to generate steam and energy, they were puzzled that their boiler wasn't producing as much steam as expected.
R&R and Tetra Engineering Group Inc. were asked to solve the issue with reduced steam production.
An inspection had shown that a significant amount of hot flue gas was bypassing the boiler tubes, where the heat was supposed to be transferred.
R&R Consult conducted a CFD analysis, which revealed that 6.3% of the flue gas was bypassing the boiler tubes without transferring heat. The analysis also showed that the flue gas was instead being directed along the sides of the boiler and between the modules that were supposed to capture the heat. This was the cause of the reduced performance.
Based on our results, Tetra Engineering installed covering plates to reduce the bypass flow. This improved the boiler's performance and increased electricity production.
It is always satisfying when we can help solve complex challenges like this. Do your systems also need a check-up or optimization? Give us a call!
Work done in cooperation with James Malloy and David Moelling from Tetra Engineering.
More examples of our work https://www.r-r-consult.dk/en/cases-en/
Welcome to WIPAC Monthly the magazine brought to you by the LinkedIn Group Water Industry Process Automation & Control.
In this month's edition, along with this month's industry news to celebrate the 13 years since the group was created we have articles including
A case study of the used of Advanced Process Control at the Wastewater Treatment works at Lleida in Spain
A look back on an article on smart wastewater networks in order to see how the industry has measured up in the interim around the adoption of Digital Transformation in the Water Industry.
Cosmetic shop management system project report.pdfKamal Acharya
Buying new cosmetic products is difficult. It can even be scary for those who have sensitive skin and are prone to skin trouble. The information needed to alleviate this problem is on the back of each product, but it's thought to interpret those ingredient lists unless you have a background in chemistry.
Instead of buying and hoping for the best, we can use data science to help us predict which products may be good fits for us. It includes various function programs to do the above mentioned tasks.
Data file handling has been effectively used in the program.
The automated cosmetic shop management system should deal with the automation of general workflow and administration process of the shop. The main processes of the system focus on customer's request where the system is able to search the most appropriate products and deliver it to the customers. It should help the employees to quickly identify the list of cosmetic product that have reached the minimum quantity and also keep a track of expired date for each cosmetic product. It should help the employees to find the rack number in which the product is placed.It is also Faster and more efficient way.
Vaccine management system project report documentation..pdfKamal Acharya
The Division of Vaccine and Immunization is facing increasing difficulty monitoring vaccines and other commodities distribution once they have been distributed from the national stores. With the introduction of new vaccines, more challenges have been anticipated with this additions posing serious threat to the already over strained vaccine supply chain system in Kenya.
AN AUTOMATED MULTIPLE-CHOICE QUESTION GENERATION USING NATURAL LANGUAGE PROCESSING TECHNIQUES
1. International Journal on Natural Language Computing (IJNLC) Vol.10, No.2, April 2021
DOI: 10.5121/ijnlc.2021.10201 1
AN AUTOMATED MULTIPLE-CHOICE QUESTION
GENERATION USING NATURAL LANGUAGE
PROCESSING TECHNIQUES
Chidinma A. Nwafor and Ikechukwu E. Onyenwe
Department of Computer Science, Nnamdi Azikiwe University Awka,
Anambra State, Nigeria
ABSTRACT
Automatic multiple-choice question generation (MCQG) is a useful yet challenging task in Natural Language
Processing (NLP). It is the task of automatic generation of correct and relevant questions from textual data.
Despite its usefulness, manually creating sizeable, meaningful and relevant questions is a time-consuming
and challenging task for teachers. In this paper, we present an NLP-based system for automatic MCQG for
Computer-Based Testing Examination (CBTE).We used NLP technique to extract keywords that are
important words in a given lesson material. To validate that the system is not perverse, five lesson materials
were used to check the effectiveness and efficiency of the system. The manually extracted keywords by the
teacher were compared to the auto-generated keywords and the result shows that the system was capable of
extracting keywords from lesson materials in setting examinable questions. This outcome is presented in a
user-friendly interface for easy accessibility.
KEYWORDS
Natural Language Processing,NLP, Computer-Based Test Examination, CBTE, Automatic Question
Generation, AOG, Document, Tokenization, Multiple-Choice Question, MCQ
1. INTRODUCTION
According to [1], automatic question generation (AQG) is the task of automatically generating
syntactically sound, semantically correct and relevant questions from various input formats such
as text, a structured database or a knowledge base. AQG can be naturally applied in many domains
such as Massive Open Online Course (MOOC), setting objective/subjective questions, automated
help systems, search engines, Chabot systems (e.g. for customer interaction), and healthcare for
analyzing mental health [2]. Despite its usefulness, manually creating meaningful and relevant
questions is a time-consuming and challenging task. For example, while evaluating students on
reading comprehension, it is tedious for a teacher to manually create questions, find answers to
those questions, and thereafter evaluates answers.
In the field of education, questioning is widely acknowledged as an effective instructional strategy
to evaluation or assessment of students or the learner at the end of a lesson. Questioning as an
instructional tool can be traced back to the fourth century, when Socrates used questions and
answers to challenge assumptions, expose contradictions, and lead to new knowledge and wisdom.
Since Socrates, questioning was employed to encourage students to use higher-order thinking
processes and it is the dominant mode of teacher-student interaction accounted for almost 80% of
the total interactions [3]. There are three major types of assessment for evaluating students’ learning
capabilities – Formative assessment, Interim assessment and Summative assessment. Generating
questions that covers most areas of a lesson note would help in Formative and Interim assessment
2. International Journal on Natural Language Computing (IJNLC) Vol.10, No.2, April 2021
2
because the Formative assessment helps to improve students’ understanding and performance
while the Interim assessment helps teachers identify gaps in student understanding and instruction.
Multiple-choice Question (MCQ)s that widely cover all the areas of a lesson note is an important
effort to test students on the above assessments and the students’ overall cognitive levels held at
the end of a lesson. Today, MCQs in the area of Computer-Based Testing Examination (CBTE) is
the most commonly used tool for assessing the knowledge capabilities of learners in the field of
learning.
Properly constructed MCQs can assess higher cognitive processing of Bloom’s taxonomy such as
interpretation, synthesis and application of knowledge, instead of just testing recall of isolated facts.
Most institutions and Centre of learning have embraced MCQ testing as the foundation of their
testing systems, thereby saving cost especially where large numbers of candidates are involved—
and it is a format that can provide precision where other measurement options may be lacking such
as observing performance.
Effective style of questioning as described by [4] is always an issue to help students attend to the
desired learning outcome. Furthermore, to make it effective, balancing between lower and higher
level question is a must [4]. Time-consuming factor in setting MCQs hinders achieving the above
assessments and good learning outcomes due to MCQs generated tend to cover limited areas of a
lesson note from which the MCQs are taken from. Therefore, it is necessary for an automatic
system that will bootstrap the existing method of setting MCQs.
Natural Language Processing (NLP) is a branch of artificial intelligence that deals with the
interaction between computers and humans using the natural language. The ultimate objective of
NLP is to read, decipher, understand, and make sense of the written or textual data in natural
languages of human in a manner that is valuable. In this paper, we focus on the setting of MCQs
for CBTE through NLP to improve the method of setting MCQs and modification, and for creating
a viable question bank for subsequent use by the academicians for their learners. This will ensure
an MCQ that includes appropriate questions and options based on the learning objectives and
importance of the topics discussed in a lesson material. We used NLP methods, Term Frequency-
Inverse Document Frequency (TF-IDF) and N-grams, to extract the most significant words present
in a lesson material and the selection does not depend on any vocabulary; these words are keywords
that capture the main topics discussed in a lesson material, they also serve as an indication of
document relevance for users in an information retrieval (IR) environment.
2. RELATED LITERATURE
[5] developed ArikIturri, an AQG for Basque language test questions. The information source for
this question generator consists of linguistically analysed real corpora, represented in XML mark-
up language. [6] researched on using NLP on narrative texts to find discourse connectives for AQG.
Discourse connectives are words or phrases that indicate relationships between two logical
sentences or phrases and suggest the presence of mutually related extended verbal expression. The
questions were generated by first extracting the text from the materials supplied by the user using
text processing concept of NLP. [7] developed a system which can generate various logical
questions from the given text input. The system uses 3-step strategy, viz; Select the best potential
set of sentences from the given text input from which we could generate the questions, find the
subject and context of the sentence to find its core agenda (Gap Selection) and analyze the best
form of question that can be generated from that sentence (Question Formation). [8] proposed a
method for automatically generating distractors for multiple-choice English vocabulary questions.
The proposed method introduces new sources for collecting distractor candidates and utilises
semantic similarity and collocation information when ranking the collected candidates. [9] looked
at the development of an automatic factual open cloze question generation system which can
3. International Journal on Natural Language Computing (IJNLC) Vol.10, No.2, April 2021
3
generate fill-in-the-blank questions without alternatives. The system first extracts a set of
informative sentences from the given input corpus based on Part-of-Speech tagging based rules.
Then, generate questions by omitting the answer-keys which are selected by identifying domain
specific words in the sentences. The system also suggests answer hints for the examinees to reduce
the number of possible answers that make assessment easier.
3. EXPERIMENTAL METHODOLOGY
The paper is based on the generation of MCQG through Natural Language Processing Techniques.
This is used to process the lesson materials/documents fed by the teacher into a multi-choice
questions alongside the answers to each questions. The NLP processes are applied using TF-IDF
and N-gram. This system only applies to the text of the document or lesson material in extracting
keywords presented by the teacher. The document is converted into text file, loaded into the system
memory and stored on the system. The text is split into sentences. The split sentences are tokenized,
from which the corpus are built as TF-IDF and N-gram mode. . The NLP will incorporate text
summarization in word count, keyword/key phrase extraction, stemming, normalization thereby
making the new system to be more efficient and robust in analyzing and classifying the documents
or lesson materials. From Figure 1, the raw data is the original text file from the document (lesson
material). Noise removal is any piece of text which is not relevant to the context of the data such
as stop words. Word Normalization converts high dimensional features or N different features to
the low dimensional space or 1 feature, it includes tokenization and stemming. The cleaned text is
the end result ready for further experimentation and such is used to analyse and auto-generate multi-
choice questions (MCQs). Furthermore, we develop an application interface for easy access and
presentation of the MCQs to the teachers for acceptance or rejection.
Figure 1. Text Preprocessing of Input Data.
3.1. Dataset and Preparation
In other to compute the performance evaluation of this paper to help check that it is not perverse,
keywords extracted by the NLP techniques for accuracy computation was validated. First, five
different lecture materials of varying sizes of sentences was selected (see Table 1).
4. International Journal on Natural Language Computing (IJNLC) Vol.10, No.2, April 2021
4
Table1. Statistics of the Textual Data Used
Lesson Materials Number of
Sentences
Number of
Words
Number of Keywords
by Teachers
Computer Science 39 956 206
Geology 33 903 86
Philosophy 46 820 100
Religion 61 1474 187
History 64 1472 210
In order to evaluate the accuracy of the keywords extraction methods on these lesson materials, it
was necessary to collect manually marked keywords to create a gold standard for the evaluation.
The usual procedure of creating a gold standard of keywords is having one professional indexer
who extracts the keywords based on his expertise and training, but also based on a set of guidelines
defined by the interested parties. However, for this experiment, it was felt that the teachers (T) are
appropriate for setting a gold standard. They know what words are most relevant in their lecture
materials. Each lecture materials were annotated by a T to set a gold standard to which each set of
auto-generated keywords could be compared. The statistics of the data used are stated in Table 1.
Table 2. Descriptive statistics of word counts.
Lesson Materials Minimum Number of
Sentences
Maximum Number
of Words
Mean number
of Sentence
Computer Science 6.0 59.0 24.5
Geology 10.0 43.0 25.8
Philosophy 1.0 39.0 17.8
Religion 1.0 59.0 24.2
History 1.0 79.0 23.0
Table 2 explains the columns “number of Sentences” and “number of words” in Table 1. For
instance
Computer Science in Table 3.2, the least sentence has 6-word length of size while maximum
sentence has 59-word length of size. Summation of all the sizes of the sentences divided by the
total number of sentences gives the mean size of sentences in Computer Science. This helps gain
insight on how words are formed in the sentences of a lecture material. For example, sentences of
“Minumum number of Sentence” < 5 are removed on the preprocessing stage.
3.2. Experimental Tools
The following NLP techniques are used for automated generation, manipulation and analysis of
data from Figure 1. Data here is the lesson materials from teachers henceforth referred to as
documents. In this paper, we represent each sentence in a lesson material as a document. For the
auto-generated MCQ’s application interface, we used Django. It is a high-level Python Web
framework that encourages rapid development and clean, pragmatic design. It also takes care of
much of the hassle of Web development and focuses on writing the API without needing to reinvent
the wheel
3.2.1. Term Frequency-Inverse Document Frequency
Term Frequency-Inverse Document Frequency (TF-IDF) is used for stop-words filtering in various
subject fields including text summarization and classification. The weight is composed of two
terms: Term frequency (TF) and Inverse Document Frequency (IDF).
5. International Journal on Natural Language Computing (IJNLC) Vol.10, No.2, April 2021
5
TF: Term Frequency measures how frequently a term occurs in a document. It is possible
that a term would appear much more times in long documents than shorter ones because
every document is different in length. Thus, it is stated below as a way of normalization:
TFt =
Number of times term t appears in a document
Total number of terms in the document
IDF: Inverse Document Frequency measures how important a term is. While computing TF,
all terms are considered equally important. However, it is known that certain terms like ‘is’,
‘of’, and ‘that’, may appear a lot of times but have little importance. Thus, it is computed as
follows:
IDFt = Loge (
Total number of documents
Number of documents with term t in it
)
The TF-IDFt;d weighting scheme assigns to term t a weight in document d given by
TF-IDFt,d = TFt,d x IDFt
3.2.2. N-gram
This is a contiguous sequence of n items from a given sequence of text or speech. The items can
be phonemes, syllables, letters, words or base pairs according to the application. The n-grams
typically are collected from a text or speech corpus. An n-gram of size 1 is referred to as a
‘unigram’; size 2 is a ‘bigram’; size 3 is a ‘trigram’. Larger sizes are sometimes referred to, by the
value of n, for example, ‘4-gram’, ‘5-gram’, and so on.
4. RESULT DISCUSSION
This discusses the performance of the NLP techniques for MCQs generation applied to the dataset
of different lecture materials contents (see Table 1) and compared with the gold standard created
by T. Table 3 shows the result of application of the above techniques on documents stated in Table
1 after going through preprocessing in Figure 1. Furthermore, MCQs generation performance
scores on the dataset were computed using precision and recall. Precision was to validate the quality
of the keywords by MCQs generation extracted from the lecture materials and recall was used to
measure the quantity.
Precision answers the question – The keywords extracted from the lecture materials are they
important words according to the gold standard marked by the teachers? While recall answers the
question - are all the important words in the lecture materials extracted?
Table 3. Statistics of the Keywords automatically extracted from documents.
Lesson Materials Number of Keywords Auto Extracted
Computer Science 368
Geology 318
Philosophy 254
Religion 488
History 471
For the performance evaluation, the following metrics precision, recall and f-measure are used in
this research.
6. International Journal on Natural Language Computing (IJNLC) Vol.10, No.2, April 2021
6
Precision =
numberCorrect
numberExtracted
(1)
Recall =
numberCorrect
totalExtracted
(2)
where numberCorrect is equal to true positive (TP), that is, important keywords extracted.
numberExtracted in eqn 1 is summation of TP and false positive (FP). False Positive is extracted
keywords that are not important words according to the gold standard. totalExtracted in eqn 2 is
the summation of TP and false negative (FN). FN are keywords not extracted but are important
words. To find the harmonic mean (average) of the precision and recall for computation of balance
between precision (p) and recall (r) in the system:
F − measure = 2 ∗ (
𝑅𝑒𝑐𝑎𝑙𝑙∗𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛
𝑅𝑒𝑐𝑎𝑙𝑙+𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛
) (3)
For each lecture material, for instance Computer Science (CS), there are two sets of data:
Gold standard keywords that are important words in A prepared by the teacher. Henceforth
known as truthset.
The keywords extracted automatically from A. Henceforth known as extractedset.
From the above explanation, assuming truthset and extractedset are sets from CS. TP is a set of
intersection between truthset and extractedset (TP = truthset & extractedset), FP is a set of
keywords that are in extractedset but not in truthset (fp = extractedset – truthset), while FN is a set
of keywords that are in truthset but not in extractedset (fn = truthset – extractedset). Using Table
1, there are 206 truthset and 368 extractedset for CS lesson material.
Table 4. Precision, Recall and F-measure.
Lesson Materials TP FP FN Precision Recall F-measure
Computer Science 162 206 31 0.44 0.84 0.58
Geology 79 293 3 0.25 0.96 0.39
Philosophy 94 160 2 0.37 0.98 0.54
Religion 169 319 1 0.35 0.94 0.51
History 185 286 18 0.39 0.91 0.55
7. International Journal on Natural Language Computing (IJNLC) Vol.10, No.2, April 2021
7
Figure 2. Bar charts of Precision, Recall and F-measure of scores on Table 4.
From Table 4 and Figure 2, it can be observed from the Recall scores that the majority of the
keywords automatically extracted are important words. Although Precision scores show that there
are other keywords automatically extracted that are not important according to the gold standard
set. At the moment we want to ensure that the majority (if not all) the keywords extracted from
lesson notes are in the gold standard sets. The keywords extracted by the NLP processes are used
to generate Multi-Choice Questions (MCQs).
Since teachers are not technical persons, we develop a user-friendly interface where the MCQs
generated are presented for teachers to easily access them. MCQs are generated from the extracted
keywords by locating each keyword position given a document.
Table 5. Sample TF-IDF weightd sentences.
Sentence TF-IDFWeights
The Difference Engine consisted entirely of
mechanical components.
{’entirely’: 0.305, ’consisted’: 0.305, ’components’:
0.305, ’mechanical’: 0.209, ’difference’: 0.209}
0
20
40
60
80
100
120
Computer
Science
Geology Philosophy Religion History
S
C
O
R
E
S
Precision
Recall
f-measure
8. International Journal on Natural Language Computing (IJNLC) Vol.10, No.2, April 2021
8
Figure 3. Figure showing API for lecture material upload
Figure 4. Figure showing API for generated MCQs.
9. International Journal on Natural Language Computing (IJNLC) Vol.10, No.2, April 2021
9
Each keyword of a document has a TF-IDF weight resulting in a composite-weighted sentence;
this is illustrated in Table 5. How the weight in Table 5 work is when it is highest it means that t
occurs many times within a small number of documents (thus lending high discriminating power
to those documents); lower when the t occurs fewer times in a document, or occurs in many
documents (thus offering a less pronounced relevance signal); lowest when the t occurs in virtually
all documents. Given this definition, in a document we extract keywords with top weights as shown
in Table 5. These extracted keywords are mapped to their positions within the document so that
they are replaced with ‘–’ symbols instead. Then, we blindly and randomly select three other
keywords from the pool of extracted keywords plus the main keyword that was replaced with ‘–’
to serve as options from which users are to select the right one. For a document, the number of
multi-choice questions (MCQs) to be generated is dependent on the number of keywords extracted
(see Table 5). These MCQs are regarded as suggested questions for the teachers to accept or reject.
Figure 3 shows the API for lecture material upload while figure 4 shows the API for the generated
MCQs. The accept button is clicked by the teacher to select an MCQ if satisfied with it.
5. CONCLUSION
Multiple-choice questions (MCQs) generation for CBTE using Natural Language Processing
(NLP) techniques is a system that helps teachers set multi-choice questions from their documents
in a text file and thereafter provide answers to the questions generated. Consequently, it can be
used in every area of education sector to test the knowledge or level of the students. However, this
paper is not concerned with the issue of whether multiple-choice tests are better assessment
methodology than other types of tests. What it focuses on is on application of NLP techniques to
generate MCQs about facts explicitly based on keywords as stated in a text format of a given lesson
material. From the results, it has shown that the system was capable of extracting keywords from
lesson materials in setting examinable questions. This outcome is presented in a user-friendly
interface for teachers easy accessibility using Django, a web framework for Python.
ACKNOWLEDGEMENTS
We thank all the teachers and students that contributed to the success of this work. Your feedbacks
are very insightful and helpful towards the development of this work.
REFERENCES
[1] Mannem Prashanth, Rashmi Prasad, and Aravind Joshi. (2010). Question generation from paragraphs
at UPenn: QGSTEC system description. In Proceedings of QG2010: The Third Workshop on Question
Generation, pages 84–91.
[2] Colleen E Crangle and Joyce Brothers Kart. (2015). A questions-based investigation of consumer
mental Health information. PeerJ, 3:e867.
[3] Gall MD. (1970). The use of questions in teaching. Review of Educational Research, 40: 707 720.
[4] Swart (2010). Evaluation of Final Examination Papers in Engineering: A Case Study Using Bloom’s
Taxonomy. IEEE Transactions on Education, (May 2010) Vol. 53, No.2 257-264.
[5] Aldabe, I., De Lacalle, M. L., Maritxalar, M., Martinez, E., and Uria, L. (2006). Arikiturri: an
Automatic question generator based on corpora and nlp techniques. In International Conference on
Intelligent Tutoring Systems (pp. 584-594). Springer, Berlin, Heidelberg.
10. International Journal on Natural Language Computing (IJNLC) Vol.10, No.2, April 2021
10
[6] Folajimi, Y. O., and Omojola, O. E. (2013). Natural language processing techniques for automatic test
Questions generation using discourse connectives. Journal of Computer Science and Its Application,
20(2), 6076.
[7] Aditya S. (2018). Using Natural Language Processing for Smart Question Generation. An Intel
Developer Program Documentation. https://software.intel.com/en-us/articles/using- natural language
processing forsmart-question-generation.
[8] Susanti, Y., Tokunaga, T., Nishikawa, H., and Obari, H. 2018. Automatic distractor generation for
multiple Choice English vocabulary questions. Research and Practice in Technology Enhanced
Learning, 13(1), 15.
[9] Das, B., and Majumder, M. (2017). Factual open cloze question generation for assessment of learne’s
knowledge. International Journal of Educational Technology in Higher Education, 14(1), 24.
AUTHORS
Chidinma A. Nwafor holds a BSc and concluding MSc in Computer Science from and at
Nnamdi Azikiwe University, Awka, Anambra State, Nigeria. She is currently a
lecturer/researcher in the department of Computer Science, Nigerian Army College of
Environmental Science and Technology, Makurdi, Benue State, Nigeria. Her research
interests are AI and Natural Language Processing.
Ikechukwu E. Onyenwe holds a B.Sc. and and M.Sc. in Computer Science from Nnamdi
Azikiwe University, Nigeria; a PHD in Computer science from the University of
Sheffield, UK. He is currently a senior lecturer/researcher in the department of Computer
Science/Linguistics, Nnamdi Azikiwe University, Awka, Anambra State, Nigeria. His
research interests are AI, Natural Language Processing, Computational Sciences,
Machine Learning, Data Science/Communication, Web/Internet Computing.