International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 01 | Jan-2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 1175
An Automatic Question Paper Generation: Using Bloom's Taxonomy
Shubham Dhainje1,Renuka Chatur2,Komal Borse3 ,Vishal Bhamare4
1,2,3,4Student,Dept. of Computer Engineering, R. H. Sapat College of Engineering Management Studies and Research,
Nashik, Maharashtra, India
-----------------------------------------------------------------------***--------------------------------------------------------------------
Abstract - Blooms Taxonomy is a classification of learning
object within education that for students. The cognitive
domain within Bloom Taxonomy is designed to verified a
student’s cognitive level during a written examination.
Educators may sometimes face the challenge in analyzing
whether their examination questions observe within the
requirements of the Blooms taxonomy at different cognitive
levels. This paper proposed an automated analysis of the
exam questions to determine the appropriate category
based on Bloom Taxonomy. This rule-based approach
applied Natural Language Processing (NLP) techniques to
identify important keywords and verbs, which may assist in
the identification of the category of a various questions.
This work focuses on the various subject domain. At present,
a set of 100 questions (70 training set and 30 testing set) is
used in the research. Preliminary results denote that the
rules may successfully assist in the identification of the
Blooms taxonomy category correctly in the exam questions.
We introduce a new algorithm which ensures total
randomization of questions and avoids repetitions or
Duplication. The proposed system can be helpful to many
educational institutes and Universities , NGO based
institutes , Bank sector , Government Exams.
Key Words: Data Mining, Blooms Taxonomy, Natural
Language Processing (NLP), Randomized Algorithm .
1.INTRODUCTION
There are many types of valuation or “testing” to access
student's learning curves. However, written examination is
the most common and important approach used by any
higher education institutions for students' assessment.
Question is an element that is associate with the
examination. Questions raised in the paper plays an
important role in efforts to test the students' overall
cognitive levels held each semester. Effective style of
questioning as described by Swart (2010) is always an
issue to help students attend to the desired learning
outcome. Furthermore, to make it effective, balancing
between lower and higher-level question is a must Swart
(2010). Bloom's Taxonomy, created by Bloom (1956), has
been widely accepted as a guideline in designing
reasonable examination questions belonging to various
cognitive levels. The hierarchical models of Bloom’s are
widely used in education fields (Chang & Chung, 2009)
creating questions (Lister & Leaney, 2003), to ensure
balancing and student cognitive mastery (Oliver et al.,
2004). From the computer science domain itself, the
taxonomy improves curricular design and assessments
(Scoot, 2003).According to Bloom taxonomy [12], the
paper covers the following level of knowledge: apply,
understand and remember.[11]
Normally, academicians would categorise a
question according to the Bloom’s cognitive level manually.
This may lead to miscategorizing of the exam questions
and subsequently may fail to meet the examination
standard required for the subject. In addition, some
academicians also show no significant agreement on how
to use Bloom's taxonomy in educating students (Johnson
& Fuller, 2006). The aim of this paper is to propose a rule-
based approach in determining the Bloom’s taxonomy
cognitive level of examination questions through natural
language processing. natural language processing [NLP] is
an area of research that explore how computers can be
used to understand and manipulate natural language text
or speech to do useful things.[10] Exam questions will be
analyzed and each and every questions will be categorized
based on the Bloom’s taxonomy cognitive level. The scope
of the work is limited to computer programming domain.
This will assist the academicians in setting up suitable
exam questions according to the requirements.[11]
1.1 Motivation of the Project
Lower Level Thinking:
1) Knowledge: - Recalling the data .
2) Comprehension: - Grasping the meaning of information
3) Application: - applying the concept to a certain scenario.
Higher Level Thinking:
1) Analysis: - This level required students to breakdown
information into simpler parts.
2) Synthesis: - The students should be able to integrate
and combine ideas or concepts by rearranging
components into a new whole.
3) Evaluation: - This is a final level where judging,
criticism, supporting or defending own stand involves.
2. Literature Survey
Automatic Cloze-Questions Generation by Annamaneni
Narendra, Manish Agarwal and Rakshit shah LTRC, IIIT-
Hyderabad, India.
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 01 | Jan-2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 1176
Review The paper presents a concept for Manually
Evaluation is done 1). Evaluation of the selected sentence
2). Evaluation of selected keyword and 3). Evaluation of
selected distractor.
Automatic question generation in multimedia based
learning. Review A large-scale experiment investigating
the productivity of generating questions (time taken to
post-edit questions vs. time taken to generate questions
from scratch) is planned
3. System Architecture:
3.1Study Bloom's Cognitive Domain:
Cognitive domain of Bloom's Taxonomy (Bloom, 1956) is
one of the three domains that were introduced by
Benjamin Bloom in 1950s. This domain is designed to
verified a student's cognitive quality during written
examination. The famous Bloom's taxonomy subtype of six
levels i.e. knowledge, comprehension, application, analysis,
synthesis and evaluation (Bloom, 1956). The following
describe each levels of Bloom's Taxonomy:
a) Knowledge-level:
Also known as recalling of data (Bloom, 1956). Scott
(2003) refers it as 'rote learning' or 'memorization'. This
level serves as the beginning level of the hierarchy. It is a
level where students memorize facts or recall the
knowledge they learn before. The questions for
programming in this category have the criteria of recalling
specific input from previous lessons, explaining describing
computing terms, methodology and process, stating
relevancy description for a subject area, concept or term
and listing explicitly information from questions (Scott,
2003). Examples:
a) List all the nodes in the left
subtree of node J.
b) Describe the key properties of a
binary tree.
c) Define method in JAVA.
b) Comprehension-level:
Bloom (1956) describes this level as grasping the meaning
of information. The capacity to interpret, translating,
extrapolating, classifying, explaining are the concepts of
these levels. The questions for programming (Thompson
et al., 2003) in this category could be translating algorithm
(e.g.; write output of a program), explaining the processes
and flows of program and providing examples to illustrate
a concept or an algorithm. Examples: a) What is the output
of the following code segment? b) Define in words what
happens in the following C++ code.
c) Application-level:
Application is defined by applying the concept to a certain
scenario (Starr et al., 2008). The questions for
programming in this category have the following criteria:
understand the concept and use it to a new algorithm and
modifying controls. Examples: a) Declare a variable,
employees to represent the records of 120 employees. b)
Modify the given ‘for’ loop into ‘while’ loop.
d) Evaluation-level:
This is a final level where judging, criticism, supporting or
defending own stand involves. Thompson et al. (2008)
discuss this level in Bloom's Taxonomy for CS
Assessment.[11] According to them, programming
question is interpreted by checking codes if the code fits
the requirement for testing strategy. This level also
includes discussion quality of codes based on standards or
execution criteria. Example: a) Explain the concept of
inheritance and give the sample of code to illustrate your
answer.
e) Synthesis-level:
If a student achieves this level, the student should be able
to synthesis and associate ideas or concepts by
rearranging components into a new whole (a product,
plan, pattern or proposal) (Bloom, 1956). Scott (2003)
suggests programming questions for this level should
instruct student to write codes based on previous level by
writing a complete program or create new alternative
methods or algorithm to solve a problem. Examples: a)
Write the definition of the function Output Time if the
statements from the lines 22 to 34 were to be performed
in a function. b) Write a program that helps the user to
input the masses of the bodies and the distance between
the bodies. The program then outputs the force between
the bodies.
f) Analysis-level:
This level requires students to breakdown information
into simpler parts and analyse each of it. This may signify
drawing a relationship, assumptions, distinguish or
classifying the parts. According to Thompson et al. (2008),
programming questions should contain the following:
subdivide programming algorithm into classes, methods;
establish elements to achieve objective; recognize
components of a development and distinguish non related
factor or needs. In addition, it should be able to explain
what absolutely happens to memory when the codes are
executed line by line. Examples: a) Outline how class Book
List could be implemented using an array. b) Given the
following postfix notation: 12 9 + 9 *5 3 / - = Using the
stack implementation to evaluate the above postfix
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 01 | Jan-2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 1177
notation, show the memory configuration of the stack after
the token at number 2 is read by the compiler.[11]
Figure: System Architecture
3.2ALGORITHM STEPS:
3.2.1.Pre-Processing :
Text pre-processing is a method in natural language
processing to prepare the computer understand the
structure or content of the text. It will allow us to prepare
the text more precise and easy to use for later process.
Text preprocessing associate processes such as stopwords
removal, stemming, lemmatization and POS tagging. In this
work, stopwords removal is applied to the question in
order to make the text more readable for later process.
Following this, each and every word will then be tagged
using a tagger. In this research, NLTK tagger (Bird et al.,
2009) is used to tag the exam questions. To emphasize the
tagging process, consider the following sentence:
“Outline how class ArrayList could be implemented using
an array.”,
The tagged output is: Outline/VB how/WRB class/NN
ArrayList/NN could/MD be/VB implemented/VBN
using/VBG an/DT array/NN./.
The tagger will assist to identify important nouns and
verbs, which may be important in determining the
question’s category. In addition, the sentence pattern may
help in the correct recognition of the question’s category.
After tagging, some rules will be applies according to
question's structure.
3.2.2.Rules Development:
Through this research, a rule-based approach is accepted
in determining the category of an examination question
based on the Bloom’s taxonomy. The rules are developed
from a training set which consists of 70 examination
questions in the various subjects. There are two conditions
where the rules will be applied:
 The rules will distinguish the suitable keyword
for each question depending on its category.
 Help to choose the appropriate category if the
keyword shares more than one category. For
example, Abstract may fall under Comprehension
or Synthesis category.
After analysing all the questions in the training set, the
questions’ patterns show that most of them start with a
verb. However, only some of it begins with Wh-pronoun, a
determiner, preposition or subordinating conjunction, a
noun and an adverb. Before rules can be applied, specific
patterns should be identified from the questions item. The
following will determine how the pattern and rules are
developed after POS tagging is applied. Question: Write
down the output of the following program:
Question with tag: Write/VB down/RB output/VB
following/JJ program/NN :/:
Pattern: /VB (1st word), /VB (3rd word)
Each verb in the question will be captured. The verb
'Write' appears as its keyword. Based on Bloom’s
Taxonomy, Write can be categorised into two categories:
Knowledge and Synthesis.
FOR each sentence, read into an array. Split into words.
IF pattern is found
If the keyword “Write” is found
IF found:
Apply Rule1: Assign weight to
Knowledge
Apply Rule2 : Assign weight to
Synthesis Choose the greater value
or positive value Assign question
category Store in database
FOR EACH_match in pattern :
print join (keyword, category,
question)
Based on the algorithm, the question can be applied to two
different rules i.e. Rule 1 and Rule 2. Rule 1 states that the
questions fall under the ‘Knowledge’ category meanwhile
Rule 2 states that it can be categorized under the
‘Synthesis’ category. This raises a conflict as to which
category the question should fall into. When this situation
occurs, there is a need to introduce ‘category weighting’ to
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 05 Issue: 01 | Jan-2018 www.irjet.net p-ISSN: 2395-0072
© 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 1178
assist in the decision. The next subsection explains the
assignment of weights for the category.[11].
3.2.3. Randomization Algorithm
For N questions available in database
Step 1: Create a List ‘L’ of N elements
Step 2: Generate a random number ‘n’ such that 1<=n
Step 3: If n ∈ L
Go to Step 2
else Store n in the List L
Step 4: Select a question from database corresponding to
n, whose flag==true
Step 5: For the question, set flag=false
4. CONCLUSIONS
Bloom’s Taxonomy is a classification of learning objectives
within education that educators set for students. To
automate the process of categorizing examination
question according to Bloom’s Taxonomy based on its
cognitive levels. The formation of rules may improve the
accurateness of the result.
REFERENCES
[1] Automatic Cloze-Questions Generation by
AnnamaneniNarendra, Manish Agarwal and Rakshit
shah LTRC, IIIT-Hyderabad, India. D.
[2] Automatic Question Generation using Discourse Cues
by Manish Agarwal , RakshitShah and Prashanth
Mannem
[3] Automatic Multiple Choice Question Generation
System for Semantic Attributes Using String Similarity
Measures by Ibrahim EldesokyFattoh in journal
Computer Engineering and Intelligent Systems
www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-
2863 (Online) Vol.5, No.8, 2014.
[4] Automatic Generation of Question Bank Based on Pre-
defined Templates by Ahmed EzzAwad and Mohamed
YehiaDahab in International Journal of Innovations
Advancement in Computer Science IJIACS ISSN 2347
8616 Volume 3, Issue 1 April 2014.
[5] Automatic Multiple Choice Question Generation
System for Semantic Attributes Using String Similarity
Measures by Ibrahim EldesokyFattoh in journal
Computer Engineering and Intelligent Systems
www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-
2863 (Online) Vol.5, No.8, 2014.
[6] Semantic Based Automatic Question Generation using
Artificial Immune System by Ibrahim EldesokyFattoh
in Computer Engineering and Intelligent Systems
www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-
2863 (Online) Vol.5, No.8, 2014
[7] Automatic question generation in multimedia-based
learning by Yvonne SKALBAN , Le An HA , Lucia
SPECIA , Ruslan MITKOV.
[8] Automatic Generation of Question Bank Based on Pre-
defined Templates by Ahmed EzzAwad and Mohamed
YehiaDahab in International Journal of Innovations &
Advancement in Computer Science IJIACS ISSN 2347 –
8616 Volume 3, Issue 1 April 2014.
[9] AUTOMATIC GENERATION OF MULTIPLE CHOICE
QUESTIONS FROM DOMAIN ONTOLOGIES by Andreas
Papasalouros, Konstantinos Kanaris, Konstantinos
Kotis.
[10] An Overview of Automatic Question Generation
Systems by Ms. Deepshree S. Vibhandik, Prof. Mrs.
Rucha C. Samant in International Journal of Science,
Engineering and Technology Research (IJSETR),
Volume 3, Issue 10, October 2014 .
[11] Automated analysis of exam questions according to
bloom’s taxonomy. Nazlia Omar, Syahidah Sufi Haris,
Rosilah Hassan, Harlina Arshad, Masura Rahmat, Noor
Faridatul Ainun Zainal, Rozli Zulkifli.
[12] Bloom, B. S.; Engelhart, M. D.; Furst, E. J.; Hill, W.
H.;Krathwohl, D. R.,Taxonomy of educational
objectives: The classification of educational goals.
Handbook I: Cognitive domain. New York: David
McKay Company, 1956.

An Automatic Question Paper Generation : Using Bloom's Taxonomy

  • 1.
    International Research Journalof Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 01 | Jan-2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 1175 An Automatic Question Paper Generation: Using Bloom's Taxonomy Shubham Dhainje1,Renuka Chatur2,Komal Borse3 ,Vishal Bhamare4 1,2,3,4Student,Dept. of Computer Engineering, R. H. Sapat College of Engineering Management Studies and Research, Nashik, Maharashtra, India -----------------------------------------------------------------------***-------------------------------------------------------------------- Abstract - Blooms Taxonomy is a classification of learning object within education that for students. The cognitive domain within Bloom Taxonomy is designed to verified a student’s cognitive level during a written examination. Educators may sometimes face the challenge in analyzing whether their examination questions observe within the requirements of the Blooms taxonomy at different cognitive levels. This paper proposed an automated analysis of the exam questions to determine the appropriate category based on Bloom Taxonomy. This rule-based approach applied Natural Language Processing (NLP) techniques to identify important keywords and verbs, which may assist in the identification of the category of a various questions. This work focuses on the various subject domain. At present, a set of 100 questions (70 training set and 30 testing set) is used in the research. Preliminary results denote that the rules may successfully assist in the identification of the Blooms taxonomy category correctly in the exam questions. We introduce a new algorithm which ensures total randomization of questions and avoids repetitions or Duplication. The proposed system can be helpful to many educational institutes and Universities , NGO based institutes , Bank sector , Government Exams. Key Words: Data Mining, Blooms Taxonomy, Natural Language Processing (NLP), Randomized Algorithm . 1.INTRODUCTION There are many types of valuation or “testing” to access student's learning curves. However, written examination is the most common and important approach used by any higher education institutions for students' assessment. Question is an element that is associate with the examination. Questions raised in the paper plays an important role in efforts to test the students' overall cognitive levels held each semester. Effective style of questioning as described by Swart (2010) is always an issue to help students attend to the desired learning outcome. Furthermore, to make it effective, balancing between lower and higher-level question is a must Swart (2010). Bloom's Taxonomy, created by Bloom (1956), has been widely accepted as a guideline in designing reasonable examination questions belonging to various cognitive levels. The hierarchical models of Bloom’s are widely used in education fields (Chang & Chung, 2009) creating questions (Lister & Leaney, 2003), to ensure balancing and student cognitive mastery (Oliver et al., 2004). From the computer science domain itself, the taxonomy improves curricular design and assessments (Scoot, 2003).According to Bloom taxonomy [12], the paper covers the following level of knowledge: apply, understand and remember.[11] Normally, academicians would categorise a question according to the Bloom’s cognitive level manually. This may lead to miscategorizing of the exam questions and subsequently may fail to meet the examination standard required for the subject. In addition, some academicians also show no significant agreement on how to use Bloom's taxonomy in educating students (Johnson & Fuller, 2006). The aim of this paper is to propose a rule- based approach in determining the Bloom’s taxonomy cognitive level of examination questions through natural language processing. natural language processing [NLP] is an area of research that explore how computers can be used to understand and manipulate natural language text or speech to do useful things.[10] Exam questions will be analyzed and each and every questions will be categorized based on the Bloom’s taxonomy cognitive level. The scope of the work is limited to computer programming domain. This will assist the academicians in setting up suitable exam questions according to the requirements.[11] 1.1 Motivation of the Project Lower Level Thinking: 1) Knowledge: - Recalling the data . 2) Comprehension: - Grasping the meaning of information 3) Application: - applying the concept to a certain scenario. Higher Level Thinking: 1) Analysis: - This level required students to breakdown information into simpler parts. 2) Synthesis: - The students should be able to integrate and combine ideas or concepts by rearranging components into a new whole. 3) Evaluation: - This is a final level where judging, criticism, supporting or defending own stand involves. 2. Literature Survey Automatic Cloze-Questions Generation by Annamaneni Narendra, Manish Agarwal and Rakshit shah LTRC, IIIT- Hyderabad, India.
  • 2.
    International Research Journalof Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 01 | Jan-2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 1176 Review The paper presents a concept for Manually Evaluation is done 1). Evaluation of the selected sentence 2). Evaluation of selected keyword and 3). Evaluation of selected distractor. Automatic question generation in multimedia based learning. Review A large-scale experiment investigating the productivity of generating questions (time taken to post-edit questions vs. time taken to generate questions from scratch) is planned 3. System Architecture: 3.1Study Bloom's Cognitive Domain: Cognitive domain of Bloom's Taxonomy (Bloom, 1956) is one of the three domains that were introduced by Benjamin Bloom in 1950s. This domain is designed to verified a student's cognitive quality during written examination. The famous Bloom's taxonomy subtype of six levels i.e. knowledge, comprehension, application, analysis, synthesis and evaluation (Bloom, 1956). The following describe each levels of Bloom's Taxonomy: a) Knowledge-level: Also known as recalling of data (Bloom, 1956). Scott (2003) refers it as 'rote learning' or 'memorization'. This level serves as the beginning level of the hierarchy. It is a level where students memorize facts or recall the knowledge they learn before. The questions for programming in this category have the criteria of recalling specific input from previous lessons, explaining describing computing terms, methodology and process, stating relevancy description for a subject area, concept or term and listing explicitly information from questions (Scott, 2003). Examples: a) List all the nodes in the left subtree of node J. b) Describe the key properties of a binary tree. c) Define method in JAVA. b) Comprehension-level: Bloom (1956) describes this level as grasping the meaning of information. The capacity to interpret, translating, extrapolating, classifying, explaining are the concepts of these levels. The questions for programming (Thompson et al., 2003) in this category could be translating algorithm (e.g.; write output of a program), explaining the processes and flows of program and providing examples to illustrate a concept or an algorithm. Examples: a) What is the output of the following code segment? b) Define in words what happens in the following C++ code. c) Application-level: Application is defined by applying the concept to a certain scenario (Starr et al., 2008). The questions for programming in this category have the following criteria: understand the concept and use it to a new algorithm and modifying controls. Examples: a) Declare a variable, employees to represent the records of 120 employees. b) Modify the given ‘for’ loop into ‘while’ loop. d) Evaluation-level: This is a final level where judging, criticism, supporting or defending own stand involves. Thompson et al. (2008) discuss this level in Bloom's Taxonomy for CS Assessment.[11] According to them, programming question is interpreted by checking codes if the code fits the requirement for testing strategy. This level also includes discussion quality of codes based on standards or execution criteria. Example: a) Explain the concept of inheritance and give the sample of code to illustrate your answer. e) Synthesis-level: If a student achieves this level, the student should be able to synthesis and associate ideas or concepts by rearranging components into a new whole (a product, plan, pattern or proposal) (Bloom, 1956). Scott (2003) suggests programming questions for this level should instruct student to write codes based on previous level by writing a complete program or create new alternative methods or algorithm to solve a problem. Examples: a) Write the definition of the function Output Time if the statements from the lines 22 to 34 were to be performed in a function. b) Write a program that helps the user to input the masses of the bodies and the distance between the bodies. The program then outputs the force between the bodies. f) Analysis-level: This level requires students to breakdown information into simpler parts and analyse each of it. This may signify drawing a relationship, assumptions, distinguish or classifying the parts. According to Thompson et al. (2008), programming questions should contain the following: subdivide programming algorithm into classes, methods; establish elements to achieve objective; recognize components of a development and distinguish non related factor or needs. In addition, it should be able to explain what absolutely happens to memory when the codes are executed line by line. Examples: a) Outline how class Book List could be implemented using an array. b) Given the following postfix notation: 12 9 + 9 *5 3 / - = Using the stack implementation to evaluate the above postfix
  • 3.
    International Research Journalof Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 01 | Jan-2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 1177 notation, show the memory configuration of the stack after the token at number 2 is read by the compiler.[11] Figure: System Architecture 3.2ALGORITHM STEPS: 3.2.1.Pre-Processing : Text pre-processing is a method in natural language processing to prepare the computer understand the structure or content of the text. It will allow us to prepare the text more precise and easy to use for later process. Text preprocessing associate processes such as stopwords removal, stemming, lemmatization and POS tagging. In this work, stopwords removal is applied to the question in order to make the text more readable for later process. Following this, each and every word will then be tagged using a tagger. In this research, NLTK tagger (Bird et al., 2009) is used to tag the exam questions. To emphasize the tagging process, consider the following sentence: “Outline how class ArrayList could be implemented using an array.”, The tagged output is: Outline/VB how/WRB class/NN ArrayList/NN could/MD be/VB implemented/VBN using/VBG an/DT array/NN./. The tagger will assist to identify important nouns and verbs, which may be important in determining the question’s category. In addition, the sentence pattern may help in the correct recognition of the question’s category. After tagging, some rules will be applies according to question's structure. 3.2.2.Rules Development: Through this research, a rule-based approach is accepted in determining the category of an examination question based on the Bloom’s taxonomy. The rules are developed from a training set which consists of 70 examination questions in the various subjects. There are two conditions where the rules will be applied:  The rules will distinguish the suitable keyword for each question depending on its category.  Help to choose the appropriate category if the keyword shares more than one category. For example, Abstract may fall under Comprehension or Synthesis category. After analysing all the questions in the training set, the questions’ patterns show that most of them start with a verb. However, only some of it begins with Wh-pronoun, a determiner, preposition or subordinating conjunction, a noun and an adverb. Before rules can be applied, specific patterns should be identified from the questions item. The following will determine how the pattern and rules are developed after POS tagging is applied. Question: Write down the output of the following program: Question with tag: Write/VB down/RB output/VB following/JJ program/NN :/: Pattern: /VB (1st word), /VB (3rd word) Each verb in the question will be captured. The verb 'Write' appears as its keyword. Based on Bloom’s Taxonomy, Write can be categorised into two categories: Knowledge and Synthesis. FOR each sentence, read into an array. Split into words. IF pattern is found If the keyword “Write” is found IF found: Apply Rule1: Assign weight to Knowledge Apply Rule2 : Assign weight to Synthesis Choose the greater value or positive value Assign question category Store in database FOR EACH_match in pattern : print join (keyword, category, question) Based on the algorithm, the question can be applied to two different rules i.e. Rule 1 and Rule 2. Rule 1 states that the questions fall under the ‘Knowledge’ category meanwhile Rule 2 states that it can be categorized under the ‘Synthesis’ category. This raises a conflict as to which category the question should fall into. When this situation occurs, there is a need to introduce ‘category weighting’ to
  • 4.
    International Research Journalof Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 05 Issue: 01 | Jan-2018 www.irjet.net p-ISSN: 2395-0072 © 2018, IRJET | Impact Factor value: 6.171 | ISO 9001:2008 Certified Journal | Page 1178 assist in the decision. The next subsection explains the assignment of weights for the category.[11]. 3.2.3. Randomization Algorithm For N questions available in database Step 1: Create a List ‘L’ of N elements Step 2: Generate a random number ‘n’ such that 1<=n Step 3: If n ∈ L Go to Step 2 else Store n in the List L Step 4: Select a question from database corresponding to n, whose flag==true Step 5: For the question, set flag=false 4. CONCLUSIONS Bloom’s Taxonomy is a classification of learning objectives within education that educators set for students. To automate the process of categorizing examination question according to Bloom’s Taxonomy based on its cognitive levels. The formation of rules may improve the accurateness of the result. REFERENCES [1] Automatic Cloze-Questions Generation by AnnamaneniNarendra, Manish Agarwal and Rakshit shah LTRC, IIIT-Hyderabad, India. D. [2] Automatic Question Generation using Discourse Cues by Manish Agarwal , RakshitShah and Prashanth Mannem [3] Automatic Multiple Choice Question Generation System for Semantic Attributes Using String Similarity Measures by Ibrahim EldesokyFattoh in journal Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222- 2863 (Online) Vol.5, No.8, 2014. [4] Automatic Generation of Question Bank Based on Pre- defined Templates by Ahmed EzzAwad and Mohamed YehiaDahab in International Journal of Innovations Advancement in Computer Science IJIACS ISSN 2347 8616 Volume 3, Issue 1 April 2014. [5] Automatic Multiple Choice Question Generation System for Semantic Attributes Using String Similarity Measures by Ibrahim EldesokyFattoh in journal Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222- 2863 (Online) Vol.5, No.8, 2014. [6] Semantic Based Automatic Question Generation using Artificial Immune System by Ibrahim EldesokyFattoh in Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222- 2863 (Online) Vol.5, No.8, 2014 [7] Automatic question generation in multimedia-based learning by Yvonne SKALBAN , Le An HA , Lucia SPECIA , Ruslan MITKOV. [8] Automatic Generation of Question Bank Based on Pre- defined Templates by Ahmed EzzAwad and Mohamed YehiaDahab in International Journal of Innovations & Advancement in Computer Science IJIACS ISSN 2347 – 8616 Volume 3, Issue 1 April 2014. [9] AUTOMATIC GENERATION OF MULTIPLE CHOICE QUESTIONS FROM DOMAIN ONTOLOGIES by Andreas Papasalouros, Konstantinos Kanaris, Konstantinos Kotis. [10] An Overview of Automatic Question Generation Systems by Ms. Deepshree S. Vibhandik, Prof. Mrs. Rucha C. Samant in International Journal of Science, Engineering and Technology Research (IJSETR), Volume 3, Issue 10, October 2014 . [11] Automated analysis of exam questions according to bloom’s taxonomy. Nazlia Omar, Syahidah Sufi Haris, Rosilah Hassan, Harlina Arshad, Masura Rahmat, Noor Faridatul Ainun Zainal, Rozli Zulkifli. [12] Bloom, B. S.; Engelhart, M. D.; Furst, E. J.; Hill, W. H.;Krathwohl, D. R.,Taxonomy of educational objectives: The classification of educational goals. Handbook I: Cognitive domain. New York: David McKay Company, 1956.