SlideShare a Scribd company logo
1 of 34
PROFESSIONAL EDUCATION - Assessment of Learning
Assessment –refersto the process of gathering, describing or quantifying information about the student
performance. It includes paper and pencil test, extended responses (example essays) and performance
assessment are usually referred to as”authentic assessment” task (example presentation of research work)
Measurement-is a process of obtaining a numerical description of the degree to which an individual
possesses a particular characteristic. Measurements answers the questions”how much?
Evaluation- it refers to the process of examining the performance of student. It also determines whether or
not the student has met the lesson instructional objectives.
Test–isaninstrument or systematic proceduresdesigned to measure the quality, ability, skill or knowledge
of students by giving a set of question in a uniform manner. Since test is a form of assessment,tests also
answer the question”how does individual student perform?
Testing-is a method used to measure the level of achievement or performance of the learners. It also refers
to the administration, scoring and interpretation of an instrument (procedure) designed to elicit information
about performance in a simple of a particular area of behavior.
Types of Measurement
There are two ways of interpreting the student performance in relation to classroom instruction. These are
the Norm-reference tests and Criterion-referenced tests.
Norm-reference test is a test designed to measure the performance of a student compared with other
students. Each individual is compared with other examinees and assigned a score-usually expressed as
percentile, a grade equivalent score or a stanine. The achievement of student is reported for broad skill
areas, although some norm referenced tests do report student achievement for individual.
The purpose is to rank each student with respect to the achievement of others in broad areas of knowledge
and to discriminate high and low achievers.
Criterion- referenced test is a test designed to measure the performance of students with respect to some
particular criterion or standard. Each individual is compared with a pre determined set of standard for
acceptable achievement. The performance of the other examinees are irrelevant. A student’s score is usually
expressed as a percentage and student achievement is reported for individual skills,
The purpose is to determine whether each student has achieved specific skills or concepts. And to find out
how mush students know before instruction begins and after it has finished.
Other terms less often used for criterion-referenced are objective referenced,domain referenced,content
referenced and universe referenced.
According to Robert L. Linn and Norma E. gronlund (1995) pointed out the common characteristics and
differences of Norm-Referenced Tests and Criterion-Referenced Tests
Common Characteristics of Norm-Referenced Test and Criterion-Referenced Tests
1. Both require specification of the achievement domain to be measured
2. Both require a relevant and representative sample of test items
3. Both use the same types of test items
4. Both used the same rules for item writing (except for item difficulty)
5. Both are judge with the same qualities of goodness (validity and reliability)
6. Both are useful in educational assessment
Differences between Norm-Referenced Tests and Criterion Referenced Tests
Norm –Referenced Tests Criterion-Referenced Tests
1. Typically covers a large domain of
learning tasks, with just few items
measuring each specific task.
1.Typically focuses on a delimited domain of
learning tasks, with a relative large number of
items measuring each specific task.
2. Emphasizes discrimination among
individuals in terms of relative of level of
learning.
2.Emphasizes among individuals can and
cannot perform.
3. Favors items of large difficulty and
typically omits very easy and very hard
items
3.Matches item difficulty to learning tasks,
without altering item difficulty or omitting
easy or hard times
4. Interpretation requires clearly defined
group
4.Interpretation requires a clearly defined and
delimited achievement domain
TYPES OF ASSESSMENT
There are four type of assessmentin termsof their functional role in relation to classroom instruction. These
are the placement assessment, diagnostic assessment, formative assessment and summative assessment.
A. Placement Assessment is concerned with the entry performance of student, the purpose of
placement evaluation is to determine the prerequisite skills, degree of mastery of the course
objectives and the best mode of learning.
B. Diagnostic Assessment is a type of assessment given before instruction. It aims to identify the
strengths and weaknesses of the students regarding the topics to be discussed. The purpose of
diagnostic assessment:
1. To determine the level of competence of the students
2. To identify the students who have already knowledge about the lesson;
3. To determine the causes of learning problems and formulate a plane for remedial action.
C. Formative Assessment is a type of assessment used to monitor the learning progress of the students
during or after instruction. Purpose of formative assessment:
1. To provide feed back immediately to both student and teacherregarding the successandfailure
of learning.
2. To identify the learning errors that is need of correction
3. To provide information to the teacher for modifying instruction and used for improving
learning and instruction
D. Summative Assessmentis a type of assessmentusually given at the end of a course or unit. Purpose
of summative assessment:
1. To determine the extent to which the instructional objectives have been met;
2. To certify student mastery of the intended outcome and used for assigning grades;
3. To provide information for judging appropriateness of the instructional objectives
4. To determine the effectiveness of instruction
MODE OF ASSESSMENT
A. Traditional Assessment
1. Assessment in which students typically select an answer or recall information to complete the
assessment. Test may be standardized or teacher made test,these tests may be multiple-choice,
fill-in-the-blanks, true-false, matching type.
2. Indirect measures of assessment since the test items are designed to represent competence by
extracting knowledge and skills from their real life context.
3. Items on standardized instrument tends to test only the domain of knowledge and skill to avoid
ambiguity to the test takers.
4. One-time measures to rely on a single correct answer to each item. There is a limited potential
for traditional test to measure higher order thinking skills.
B. Performance assessment
1. Assessment in which students are asked to perform real-world tasks that demonstrate
meaningful application of essential knowledge and skills
2. Direct measures of students performance because task are design to incorporate contexts,
problems, and solutions strategies that students would use in real life.
3. Designed ill-structured challenges since the goal is to help students prepare for the complex
ambiguities in life.
4. Focus on processes and rationales. There is no single correct answer,instead students are led
to craft polished, thorough and justifiable responses, performances and products.
5. Involve long-range projects, exhibits, and performances are linked to the curriculum
6. Teacher is an important collaborator in creating tasks, as well as in developing guidelines for
scoring and interpretation
C. Portfolio Assessment
1. Portfolio is a collection of student’s work specifically to tell a particular story about the student.
2. A portfolio is not a pie of student work that accumulates over a semester or year
3. A portfolio contains a purposefully selected subset of student work
4. It measures the growth and development of students.
The Key to Effective Testing
Objectives; The specific statements of the aim of the instruction; it should express what the students should
be able to do or know as a result of taking the course; the objectives should indicate the cognitive level,
affective level and psychomotor level of expected performance.
Instruction: It consist all the elements of the curriculum designed to teach the subject, including the lesson
plans, study guide, and reading and homework assignment; the instruction should corresponds directly to
the objectives
Assessment:The process of gathering , describing or quantifying information about the performance of the
learner; testing components of the subject; the weight given to different subject matter areas on the test
should match with objectives as well as the emphasis given to each subject area during instruction.
Evaluation: Examining the performance of students and comparing and judging its quality. Determining
whether or not the learner has met the objectives of the lesson and the extent of understanding.
INSTRUCTIONAL OBJECTIVES
Instructional objectives play a very important role in the instructional process and the evaluation process.
It serves as guides for teaching and learning, communicate the intent of the instruction to others and it
provide a guidelines for assessing the learning of the students. Instructional objectives also known as
behavioral objectives or learning objectives are statement which clearly describe an anticipated learning
outcome.
Characteristics of well-written and useful instructional objectives
1. Describe a learning outcome
2. Be student oriented-focus on the learner not on the teacher
3. Be observable or describe an observable product
4. Be sequentially appropriate
5. Be attainable within a reasonable amount of time
6. Be developmental appropriate
Factors to Consider when Constructing Good Test Items
A. VALIDITYis the degree to which the testmeasureswhat is intended to measure.It is the usefulness
of the test for a given purpose. A valid test us always reliable.
B. RELIABILITY refers to the consistency of score obtained by the same person when retested using
the same instrument or one that is parallel to it.
C. ADMINISTRABILITY the test should be administered uniformly to all students so that the scores
obtained will not vary due to factors other than differences of the students knowledge and skills.
There should be a clear provision for instruction for the students, proctors and even the who will
check the test or the scorer
D. SCORABILITY the test should be easy to score,directions for scoring is clear,provide the answer
sheet and the answer key
E. APPROPRIATENESS the test item that the teacher construct must assess the exact performances
called for in the learning objectives. The test item should require the same performance of the
student as specified in the learning objectives.
F. ADEQUACY the test should contain a wide sampling if items to determine the educational
outcomes or abilities so that resulting scores are representatives of the total performance in the
areas measured.
G. FAIRNESS the test should bit be biased to the examinees, it should not be offensive to any
examinees subgroups. A test can only be good if it is also fair to all test takers.
H. OBJECTIVITY represents the agreement of two or more raters or a test administrators concerning
the score of a student. If the two raters who assess the same student on the same test cannot agree
in score, the test lacks objectivity and the score of neither judge is valid, thus, lack of objectivity
reduces test validity in the same way that lack reliability influence validity.
TABLE OF SPECIFICATIONS
Table of specification is a device for describing test items in terms of the content and the process
dimensions. That is, what a student is expected to know and what he or she is expected to do with that
knowledge. It is described by combination of content and process in the table of specification.
Sample of One way table of specification in Linear Function
Content Number of Class
Sessions
Number of Items Test Item
Distribution
1. Definition of linear function 2 4 1-4
2. Slope of a line 2 4 5-8
3. Graph of linear function 2 4 9-12
4. Equation of linear function 2 4 13-16
5. Standard Forms of a line 3 6 17-22
6. Parallel and perpendicular lines 4 8 23-30
7. Application of linear functions 5 10 31-40
TOTAL 20 40 40
Number of items= Number of class sessions x desired total number of itens
Total number of class sessions
Example :
Number of items for the topic” definition of linear function”
Number of class session= 2
Desired number of items= 40
Total number of class sessions=20
Number of items= Number of class sessions x desired total number of itens
Total number of class sessions
=2x40
20
Number of items= 4
Sample of two way table of specification in Linear Function
content Class
hours
Knowledge Comprehension Application Analysis Synthesis Evaluation Total
1.Definitionoflinear function 2 1 1 1 1 4
2.Slope of a line 2 1 1 1 1
3.Graph of linear function 2 1 1 1 1 4
4.Equation of linear function 2 1 1 1 1 4
5.Standard Forms of a line 3 1 1 1 1 1 1 6
6.Parallel and perpendicular
line
4 1 2 1 2 8
7.Application of linear
functions
5 1 1 3 1 3 10
TOTAL 20 4 6 8 8 7 7 40
ITEM ANALYSIS
Item analysis refers to the process of examining the student’s responses to each item in the test. According
to Abubakar S. Asaad and William M. Hailaya (Measurement and Evaluation Concepts & Principles) Rexr
Bookstore (2004 Edition), there are two characteristics of an item. These are desirable and undesirable
characteristics. An item that has desirable characteristics can be retained for subsequent use and that with
undesirable characteristics is either be revised or rejected.
These criteria in determining the desirability and undesirability of an item.
a. Difficulty if an item
b. Discriminating power of an item
c. Measures of attractiveness
Difficulty index refers to the proportion of the number of students in the upper and lower groups who
answeredan item correctly. In a classroom achievement test,the desired indices of difficulty not lower than
0.20 nor higher than 0.80. the average index difficulty form 0.30 or 0.40 to maximum of 0.60.
DF = PUG + PLG
2
PUG = proportion of the upper group who got an item right
PLG = proportion of the lower group who get an item right
Level of Difficulty of an Item
Index Range Difficulty Level
0.00-0.20 Very difficult
0.21-0.40 Difficult
0.41-0.60 Moderately Difficult
0.61-0.80 Easy
0.81-1.00 Very Easy
Index of Discrimination
Discrimination Index is the differences between the proportion of high performing students who got the
item and the proportion of low performing students who got an item right. The high and low performing
students usually defined as the upper 27% of the students based on the total examination score and the
lower 27% of the students based on total examination score. Discrimination are classified into positive
Discrimination if the proportion of students who got an item right in the upper performing group is greater
than the students in the upper performing group. And Zero Discrimination if the proportion of the students
who got an item right in the upper performing group and low performing group are equal.
Discrimination Index Item Evaluation
0.40 and up Very good item
0.30-0.39 Reasonably good item but possibly subject to improvement
0.20-0.29 Marginal, usually needing and being subject to improvement
Below 0.19 Poor Item, to be rejected or improved by version
Maximum Discrimination is the sum of the proportion of the upper and lower groups who answered the
item correctly. Possible maximum discrimination will occur if the half or less of the sum of the upper and
lower groups answered an item correctly.
Discriminating Efficiency is the index of discrimination divided by the maximum discrimination.
PUG = proportion of the upper group who got an item right
PLG= proportion of the lower group who got an item right
Di = discrimination index
DM – Maximum discrimination
DE = Discriminating Efficiency
Formula:
Di = PUG – PLG
DE = Di
DM
DM= PUG + PLG
Example: Eighty students took an examination in Algebra, 6 students in the upper group got the correct
answer and 4 students in the lower group got the correct answer for item number 6. Find the Discriminating
efficiency
Given:
Number of students took the exam = 80
27% of 80 = 21.6 or 22, which means that there are 22 students in the upper performing group and 22
students in the lower performing group.
PUG = 6/22 = 27%
PLG = 4/22 = 18%
Di = PUG- PLG
= 27%- 18%
Di= 9%
DM = PUG +PLG
= 27% + 18%
DM= 45%
DE = Di/DM
= .09/.45
DE = 0.20 or 20%
This can be interpreted as on the average, the item is discriminating at 20% of the potential of an item of
its difficulty.
Measures of Attractiveness
To measure the attractiveness of the incorrect option ( distracters) in multiple-choice tests, we count the
number if students who selected the incorrect option in both upperand lower groups. The incorrect option
is said to be effective distracter if there are more students in the lower group chose that incorrect option
than those students in the upper group.
Steps of Item Analysis
1. Rank the scores of the students from highest score to lowest score.
2. Select 27% of the paperswithin the upper performing group and 27% of the paperswithin the lower
performing group.
3. Set aside the 46% of papers because they will not be used for item analysis.
4. Tabulate the number of students in the upper group and lower group who selected each alternative.
5. Compute the difficulty of each item
6. Compute the discriminating powers of each item
7. Evaluate the effectiveness of the distracters
VALIDITY OF A TEST
Validity refers to the appropriateness of score-based inferences; or decisions made based on the students
test results. The extent to which a test measures what is supposed to measure.
Important Things to Remember About Validity
1. Validity refers to the decisions we make, and not to the test itself or to the measurement
2. Like reliability, validity is not all or nothing concept; it is never totally absent or absolutely perfect.
3. Avalidity estimate, called a validity coefficient, referstospecific type of validity. It rangesbetween
0 to 1.
4. Validity can never be finally determined; it is specific to each administration of the test
TYPES OF VALIDITY
1. Content Validity- a type of validation that refers to the relationship between a test and the
instructional objectives, establishes content so thatthe testmeasureswhatit is supposed to measure.
Things to remember about validity:
a. The evidence of the content validity of your test is found in the Table of Specification.
b. This is the most important type of validity to you, as a classroom teacher.
c. There is no coefficient for content validity. It is determined judgmentally, not empirically.
2. Criterion-related Validity- a type of validation that refers to the extent to which scores form a test
relate to theoretically similar measures. It is a measure of how accurately a student’s current test
score can be used to estimate a score on criterion measure,like performance in courses,classes or
another measurementinstrument. Example, classroom reading gradesshould indicate similar levels
of performance as Standardized Reading Test scores.
a. Construct Validity- a type of validation that refers to a measure of the extent to which a test
measures a hypothetical and unobservable variable or quality such as intelligence, math
achievement, performance anxiety, etc. it established through intensive study of the test or
measurement instrument.
b. Predictive Validity- a type of validation that refers to a measure of the extent to which a
person’s current test results can used to estimate accurately what that persons performance or
other criterion, such as test scores will be at the later time.
3. Concurrent Validity- a type of validation that require the correlation of the predictor or concurrent
measure with the criterion measure. Using this, we can determine whether a test is useful to us as
predictor or as substitute ( concurrent) measure. The higher the validity coefficient, the better the
validity evidence of the test. In establishing the concurrent validity evidence no time interval is
involved between the administration of the new test and the criterion or established test.
Factors Affecting the Validity of a Test Item
1. The test itself
2. The administration and scoring of a test
3. Personal factors influencing how students response to the test
4. Validity is always specific to a particular group
Ways to Reduce the Validity of the Test Item
1. Poorly constructed test items
2. Unclear directions
3. Ambiguous items
4. Reading vocabulary too difficult
5. Complicated syntax
6. Inadequate time limit
7. Inappropriate level of difficulty
8. Unintended clues
9. Improper arrangement of items
Test Design to Improve validity
1. What is the purpose of the test?
2. How well do the instructional objectives selected for the test represent the instructional goals.
3. Which test item format will best measure achievement of each objective?
4. How many test item will be required to measure the performance adequately on each objective?
5. When and how will the test be administered?
6. How many test items will required to measure the performance adequately on each objective?
7. When and how will the test be administered?
Reliability of a Test
Reliability refers to the consistency of measurement; that is, how consistent test results of other assessment
results from one measurement to another. We can say that at test is reliable when it can be used to predict
practically the same scoreswhen test administered twice to the same group of students and with a reliability
index of 0.50 or above. The reliability of a test can be determined by means of Pearson Product Correlation
Coeffficient, Spearman-Brown Formula and Kuder-Richardon Formula.
Factors Affecting the Reliability of a Test
1. Length of the test
2. Moderate item difficulty
3. Objective scorung
4. Heterogeneity of the student group
5. Limited time
Four Methods of Establishing Reliability
1. Test-retest Method. A type of reliability determined by administering the same test twice to the
same group of students with any time interval between tests. The result of the test scores are
correlated using the Pearson Product Correlation Coefficient and this correlation coefficient
provides a measure of stability. This indicates how stable the test result over a period of time.
2. Equivalent –Form Method. A type of reliability determined by administering two different but
equivalent forms of the test ( also called parallel or alternate forms) to the same group of students
in close succession. The equivalent forms are constructed to the same set of specifications that is
similar in content, type of item and difficulty. The result of the test score are correlated using the
Pearson Product Correlation Coefficient and this correlation coefficient provides a measure of the
degree to which generalization about the performance of students from one assessment to another
assessment is justified. It measures the equivalence of the tests.
3. Split –Half method- administer test once. Score two equivalent halves of the test. To split the test
into halves that are equivalent, the usual procedure is to score the even-numbered and the odd-
numbered separately. This provides a measure of internal consistency. It indicates the degree to
which consistent results are obtained from two halves of the test
4. Kuder- Richardon Formula. Administer the test score one. Score total teal and apply the Kuder-
Richardon Formula. The Kuder-Richardon formula is applicable only in situation where students
responses are scored dichotomously and therefore is most useful with traditional test items that are
scored as right or wrong. KR-20 estimates of reliability that provide information about the degree
to which the items are of equal difficulty. ( A statistical procedure used to estimate coefficient
alpha, or a correlation coefficient is given)
Descriptive Statistics of Test Scores
Statistics play a very important role in describing the test scores of students. Teachers should have a
background on the statistical techniques in orderfor them to analyze and describe the result of measurement
obtained in their own classroom; understand the statistics uses in the test and research reports; interpret the
types of scores used in testing.
Descriptive Statistics- is concerned with collecting, describing, and analyzing a set of data without drawing
conclusions or inferences about a large group of data in terms of tables, graphs, or single number (example
average score of the class in a particular test)
Inferential Statistics- is concerned with the analysis of a subset of data leading to prediction or inferences
about the entire set if data or population
We shall discussed different statistical technique used in describing and analyzing test results.
1. Measures of Central Tendency (Averages)
2. Measures of Variability ( Spread of Scores
3. Measures of Relationship (Correlation)
4. Skewness
Measures of Central Tendency it is a single value that is used to identify the center of the data, it is taught
as the typical value in a set of scores. It tends to lie within the center if it is arranged form lowest to highest
or vice versa. There are three measures of central tendency commonly used; the mean, median and mode.
The Mean
The Mean is the common measures of center and it also know as the arithmetic average.
Sample Mean = ∑x
n
∑= sum of the scores
X= individual scores
n = number of scores
Steps in solving the mean value using raw scores
1. Get the sum of all the scores in the distribution
2. Identify the number of scores (n)
3. Substitute to the given formula and solve the mean value
Example: Find the mean of the scores of students in algebra quiz
(x) scores in algebra
45
35
48
60
44
39
47
55
58
54
∑x = 485
n= 10
Mean = ∑x
n
= 485÷ 10
Mean = 48.5
Properties of Mean
1. Easy to compute
2. It may be an actual observation in the data set
3. It can be subjected to numerous mathematical computation
4. Most widely used
5. Each data affected by the extremes values
6. It is easily affected by the extremes values
7. Applied to interval level data
The Median
The median is a point that divides the scores in a distribution into two equal parts when the scores are
arranged according to magnitude, that is from lowest score to highest score or highest score to lowest score.
If the number of score is an odd number, the value of the median is the middle score. When the number of
scores is even number, the median values is the average of the two middle scores.
Example: 1. Find the median of the scores of 10 students in algebra quiz.
(x) scores of students in algebra
45
35
38
60
44
39
47
55
58
54
First , arrange the scores from lowest to highest and find the average of two middle most scores since the
number of cases in an even.
35
39
44
45
47
48
54
55
58
60
Mean = 47 + 48
2
= 47.5 is the median score
50% of the scores in the distribution fall below 47.5
Example 2. Find the median of the scores of 9 students in algebra quiz
(x) scores of students in algebra
35
39
44
45
47
48
54
55
58
The median value is the 5th
score which is 47. Which means that 50% of the scores fall below 47.
Properties of Median
1. It is not affected by extremes values
2. It is applied to ordinal level of data
3. The middle most score in the distribution
4. Most appropriate when there are extremes scores
The Mode
The mode refers to the score or scores that occurred most in the distribution. There are classification of
mode: a) unimodal is a distribution that consist of only one mode. B) bimodal is a distribution of scores
that consist of two modes, c) multimodal is a score distribution that consist of more than two modes.
Properties of Mode
1. It is the score/s occurred most frequently
2. Nominal average
3. It can be used for qualitative and quantitative data
4. Not affected by extreme values
5. It may not exist
Example 1. Find the mode of the scores of students in algebra quiz: 34,36,45,65,34,45,55,61,34,46
Mode= 34 , because it appeared three times. The distribution is called unimodal.
Example 2. Find the mode of the scores of students in algebra quiz: 34,36,45,61,34,45,55,61,34,45
Mode
= 34 and 45, because both appeared three times. The distribution is called bimodal
Measures of Variability
Measures of Variability is a single value that is used to describe the spread out of the scores in distribution,
that is above or below the measures of central tendency. There are three commonly used measures
variability, the range, quartile deviation and standard deviation
The Range
Range is the difference between highest and lowest score in the data set.
R=HS-LS
Properties of Range
1. Simplest and crudest measure
2. A rough measure of variation
3. The smaller the value, the closer the score to each other or the higher the value, the more scattered
the scores are.
4. The value easily fluctuate, meaning if there is a changes in either the highest score or lowest score
the value of range easily changes.
Example: scores of 10 students in Mathematics and Science. Find the range and what subject has a greater
variability?
Mathematics Science
35 35
33 40
45 25
55 47
62 55
34 35
54 45
36 57
47 39
40 52
Mathematics Science
HS = 62 HS =57
LS= 33 LS= 25
R = HS-LS R= HS-LS
R= 62-33 R= 57-25
R= 29 R= 32
Based form the computed value of the range, the scores in Science has greater variability. Meaning, scores
in Science are more scattered than in the scores in Mathematics
The Quartile Deviation
Quartile Deviation is the half of the differences the third quartile (Q3) and the first quartile (Q1). It is based
on the middle 50% of the range, instead the range of the entire set
Of distribution. In symbol QD = Q3-Q1
2
QD= quartile deviation
Q3= third quartile value
Q1= first quartile value
Example : In a score of 50 students, the Q3 = 50.25 and Q1 = 25.45, Find the QD
QD = Q3-Q1
2
=50.25 – 25.4
2
QD= 12.4
The value of QD =12.4 which indicates the distance we need to go above or below the median to include
approximately the middle 50% of the scores.
The standard deviation
The standard deviation is the most important and useful measures of variation, it is the square root of the
variance. It is an average of the degree to which each set of scores in the distribution deviates from the
mean value. It is more stable measures of variation because it involves all the scores in a distribution rather
than range and quartile deviation.
SD = √∑( x-mean)2
n-1
where ,x = individual score
n= number of score in a distribution
Example: 1. Find the standard deviation of scores of 10 students in algebra quiz. Using the given data
below.
x (x-mean)2
45 12.25
35 182.25
48 0.25
60 132.25
44 20.5
39 90.25
47 2.25
55 42.25
58 90.25
54 30.25
∑x= 485
N= 10
Mean = ∑x
N
∑(x-mean)2 = 602.25
= 485
10
Mean= 48.5
SD= √∑(x-mean)2
n-1
SD= √ 602.5
10-1
SD= √ 66.944444
SD= 8.18, this means that on the average the
amount that deviates from the mean value= 48.5 is
8.18
Example 2: Find the standard deviation of the score of 10 students below. In what subject has greater
variability
Mathematics Science
35 35
33 40
45 25
55 47
62 55
34 35
54 45
36 57
47 39
40 52
Solve for the standard deviation of the scores in mathematics
Mathematics (x) (x-mean)2
35 82.81
33 123.21
45 0.81
55 118.81
62 320.41
34 102.01
54 98.01
36 65.61
47 8.41
40 16.81
∑x = 441 ∑(x-mean)2 = 936.9
Mean = 44.1 ∑(x-mean)2= 918
SD= √∑(x-mean)2
n-1
= √ 936.9
10-1
=√104.1
SD = 10.20 for the mathematics subject
Solve for the standard deviation of the score in science
Science (x) (x-mean)2
36 64
40 9
25 324
47 16
55 144
35 64
45 4
57 196
39 16
52 81
∑x= 430 ∑(x-mean)2= 918
Mean =430
10
Mean= 43
SD= √∑(x-mean)2
n-1
= √ 918
10-1
=√ 102
SD= 10.10 for science subject
The standard deviation for mathematics subject is 10.20 and the standard deviation foe science subject is
10.10, which means that mathematics scores has a greater variability than science scores. In other words,
the scores in mathematics are more scattered than in science.
Interpretation of Standard Deviation
When the value of standard deviation is large, on the average,the scores will be far form the mean. On the
other hand. If the value of standard deviation is small, on the average,the score will be close form the mean.
Coefficient of Variation
Coefficient of variation is a measure of relative variation expressed as percentage of the arithmetic mean.
It is used to compare the variability of two or more sets of data even when the observations are expressed
in different units of measurement. Coefficient of variation can be solve using the formula.
CV = (SD)x 100%
Mean
The lower the value of coefficient of variation, the more the overall data approximate to the mean or more
the homogeneous the performance of the group
Group Mean Standard deviation
A 87 8.5
B 90 10.25
CV Group A= standard deviation x 100%
Mean
= 8.5 x 100%
87
CV Group A=9.77%
CV GroupB= standard deviation x 100%
Mean
= 10.25 x 100%
90
CV Group B=11.39%
The CV of Group A is 9.77% and CB of Group B is 11/39%, which means that group A has homogenous
performance.
Percentile Rank
The Percentile rank of a score is the percentage of the scoresin the frequency distribution which are lower.
This means that the percentage of the examinees in the norm group who scored below the score of interest.
Percentile rank are commonly used to clarify the interpretation of scores on standardized tests.
Z- SCORE
Z- score (also known as standard score) measures how many standard deviations an observations is above
or below the mean.A positive z-score measuresthe number of standard deviation a score is above the mean,
and a negative z-negative z-score gives the number of standard deviation a score is below the mean.
The z-score can be computed using the formula
Z= x-µ for population
o
Z= x-mean for sample
SD
Where
X= is a raw score
0= is the standard deviation of the population
µ= is the mean of the population
SD= is the standard deviation of the sample
EXAMPLE:
James Mark’s examination results in the three subjects are as follows:
Subject Mean Standard deviation James Mark’s Grade
Math Analysis 88 10 95
Natural Science 85 5 80
Labor Management 92 7.5 94
EXAMPLE:A study showed the performance of two Groups A and B in a certain testgiven by a researcher.
Group A obtained a mean score of 87 points with standard deviation of 8.5 points, Group B obtained a
mean score of 90 points with standard deviation of 10.25 points. Which of the two group has a more
homogeneous performance?
In what subject did James Mark performed best? Very Poor?
Z math analysis = 95-88
10
Z math analysis = 0.70
Z natural science= 80-85
5
Z natural Science= -1
Z labor management = 94-92
7.5
Z labor management = 0.27
James Mark had a grade in Math Analysis that was 0.70 standard deviation above the mean of the Math
Analysis grade,while in Natural Science he was -1.0 standard deviation below the mean of Natural Science
grade. He also had a grade in Labor Management that was 0.27 standard deviation above the mean of the
Labor Management grades. Comparing the z scores,James Mark performed best in Mathematics Analysis
while he performed very poor in Natural Science in relation to the group performance.
T-score
T-score can be obtained by multiplying the z-score by 10 and adding the product to 50. In symbol, T-score
= 10z +50
Using the same exercise,compute the T-score of James Mark in Math Analysis, NaturalScience and Labor
Management
T- score (math analysis) = 10 (.7) +50
= 57
T- score (natural science) = 10(-1)+50
= 40
T-score (labor management) = 10(0.27) +50
=52.7
Since the highest T-score us in math analysis = 57, we can conclude that James Mark performed best in
Math analysis than in natural science and labor management.
Stanine
Stanine also known asstandard nine, is a simple type of normalized standard score that illustrate the process
of normalization. Stanines are single digit scores ranging form 1 to 9.
The distribution of new scores is divided into nine parts
Percent
in
Stanines
4% 7% 12% 17% 20% 17% 12% 7% 4%
Stanines 1
2 3 4 5 6 7 8 9
Skewness
Describes the degree of departures of the distribution of the data from symmetry.
The degree of skewness is measured by the coefficient of lsewness, denoted as SK and computed as,
SK= 3(mean-media)
SD
Normal curve is a symmetrical bell shaped curve, the end tails are continuous and asymptotic. The mean,
median and mode are equal. The scores are normally distributed if the computed value of SK=0
Areas Under the Normal Curve
Positively skewed when the curve is skewed to the right, it has a long tail extending off to the right but a
short tail to the left. It increases the presence of a small proportion of relatively large extreme value SK˃0
When the computed value of SK is positive most of the scores of students are very low, meaning to say that
they performed poor in the said examination
Negatively skewed when a distribution is skewed to the left. It has a long tail extending off to the left but a
short tail to the right. It indicates the presence of a high proportion of relatively large extreme values SK˂0.
When the computed value of SK is negative most of the students got a very high score,meaning to say that
they performed very well in the said examination
Rubrics
Rubrics is a scoring scale and instructional tool to assess the performance of student using a task-specific
set of criteria. It contains two essential parts: the criteria for the task and levels of performance for each
criterion. It provides teachers an effective means of students-centered feedback and evaluation of the work
of students. It also enables teachers to provide a detailed and informative evaluations of their performance.
Rubrics is very important most especially if you are measuring the performance of students against a set of
standard or pre-determined set of criteria. Through the use of scoring rubrics or rubrics the teachers can
determine the strengthens and weaknesses of the students, hence it enables the students to develop their
skills.
Steps in developing a Rubrics
1. Identify your standards, objectives and goals for your students. Standard is a statement of what the
students should be able to know or be able to perform. It should indicate that your students should
be able to know or be able to perform. It should indicate that your students should met these
standards. Know also the goals for instruction, what are the learning outcomes.
2. Identify the characteristics of a good performance on the task, the criteria, when the students
perform or present their work, it should indicate that they performed well in the task given to them;
hence they met that particular standards.
3. Identify the levels of performance for each criterion. There is no guidelines with regards to the
number of levels of performance,it vary according to the task and needs. It can have as few as two
levels of performance or as many as the teacher can develop. In this case,the rater can sufficiently
discriminate the performance of the students in each criteria. Through this levels of performance,
the teacher or the rater can provide more detailed feedback about the performance of the students.
It is easier also for the teacher and students to identify the areas needed for improvement.
Types of Rubrics
1. Holistic Rubrics
In holistic rubrics does not list a separate levels of performance for each criterion. Rather , holistic,
rubrics assigns a level of performance along with a multiple criteria as a whole, in other words you
put all the component together.
Advantage: quick scoring, provide overview of students achievement.
Disadvantage: does not provide detailed information about the student performance in specific
areas of the content and skills. May be difficult to provide one overall score.
2. Analytic Rubrics
In analytic rubrics the teacher or the rater identify and assess components of a finished product.
Breaks down the final product into component parts and each part is scored independently. The
total score is the sum of all the rating for all the partsthat are to be assessedorevaluated. In analytic
scoring, it is very important for the raterto treateachpart asseparate to avoid bias toward the whole
product.
Advantage: more detailed feedback, scoring more consistent across students and graders.
Disadvantage: time consuming to score.
Example of Holistic Rubric
3-Excellent Researcher
 Included 10-12 sources
 No apparent historical inaccuracies
 Can easily tell which sources information was drawn from
 All relevant information is included
2- Good Researcher
 Included 5-9 sources
 Few historical inaccuracies
 Can tell with difficulty where information came from
 Bibliography contains most relevant information
1-Poor Researcher
 Included 1-4 sources
 Lots of historical inaccuracies
 Cannot tell from which source information came from
 Bibliography contains very little information
Example of Analytic Rubric
Criteria Limited
1
Acceptable
2
Proficient
1
Made good observations Observations are
absent or vague
Most observations
are clear and detailed
All observations are
clear and detailed
Made good predictions Predictions are
absent or irrelevant
Most predictions are
reasonable
All predictions are
reasonable
Appropriate conclusion Conclusion is
absent or
inconsistent with
observation
Conclusion is
consistent with most
observations
Conclusion is
consistent with
observations
Advantages of Using Rubrics
When assessing the performance of the students using performance based assessment it is very important
to use scoring rubrics. The advantages of using rubrics in assessing student’s performance are:
1. Rubrics allow assessment to become more objective and consistent
2. Rubrics clarify the criteria in specific terms
3. Rubrics clearly show the student how work will be evaluated and what is expected
4. Rubrics promote student awareness of the criteria to use in assessing peer performance
5. Rubrics provide useful feedbacks regarding the effectiveness of the instruction: and
6. Rubrics provide benchmarks against which to measure and document progress
PERFORMANCE BASED ASSESSMENT
Performance based assessment is a direct and systematic observation of actualperformancesof the students
based from a pre-determined performance criteria as cited by (Gabuyo, 2011). It is an alternative form of
assessing the performance of the students that represent a set of strategies for the application of knowledge,
skills and work habits through the performance of tasks that are meaningful and engaging to students”
Framework of Assessment Approaches
Selection Type Supply Type Product Performance
True-false Completion Essay, story or poem Oral presentation of
report
Multiple-choice Label a diagram Writing portfolio Musical, dance or
dramatic performance
Matching type Short answer Research report Typing test
Concept man Portfolio exhibit, Art
exhibit
Diving
Writing journal Laboratory
demonstration
Cooperation in group
works
Forms of Performance Based Assessment
1. Extended response task
a. Activities for single assessment may be multiple and varied
b. Activities may be extended over a period of time
c. Products from different students may be different in focus
2. Restricted-response tasks
a. Intended performances more narrowly defined than extended-response tasks.
b. Questions may begin like a multiple-choice or short answer stem,but then ask for explanation,
or justification.
c. May have introductory material like an interpretative exercise,but then asks for an explanation
of the answer, not just the answer itself
3. Portfolio is a purposeful collection of student work that exhibits the student’s efforts, progress and
achievements in one or more areas.
Uses of Performance Based Assessment
1. Assessing the cognitive complex outcomes such as analysis, synthesis and evaluation
2. Assessing non-writing performances and products
3. Must carefully specify the learning outcomes and constructactivity or taskthat actually called forth.
Focus of Performance Bases Assessment
Performance based assessment can assess the process, or product or both (process and product) depending
on the learning outcomes. It also involves doing rather that just knowing about the activity or task. The
teacher will assess the effectiveness of the process or procedures and the product used in carrying out the
instruction. The question is when to use the process and the product?
Use the process when:
1. There is no product
2. The process is orderly and directly observable;
3. Correct procedures/steps in crucial to later success;
4. Analysis of procedural steps can help in improving the product,
5. Learning is at the early age.
Use the product when:
1. Different procedures result in an equally good product;
2. Procedures not available for observation;
3. The procedures have been mastered already;
4. Products have qualities that can be identified and judge
The final step in performance assessment is to assess and score the student’s performance. To assess the
performance of the students the evaluator can used checklist approach , narrative or anecdotal approach,
rating scale approach, and memory approach. The evaluator can give feedback on a student’s performance
in the form of narrative report or grade. There are different was to record the results of performance-based
assessments.
1. Checklist Approach are observation instruments that divide performance whether it is certain or
not certain. The teacher has to indicate only whether or not certain elements are present in the
performances
2. Narrative/AnecdotalApproach is continuous description of student behavior as it occurs, recorded
without judgment or interpretation. The teacherwill write narrative reports of whatwasdone during
each of the performances. Form these reports teachers can determine how well their students met
their standards.
3. Rating Scale Approach is a checklist that allows the evaluator to record information on a scale,
noting the finer distinction that just presence or absence of a behavior. The teacher they indicate to
what degree the standards were met. Usually, teachers will use a numerical scale. For instance, one
teacher may arte each criterion on a scale of one to five with one meaning “ skills barely present”
and five meaning “skill extremely well executed.”
4. Memory Approach the teacher observes the students when performing the tasks without taking any
notes. They use the information from memory to determine whether or not the students were
successful.This approach is not recommended to use for assessing the performance of the students.
PORTFOLIO ASSESSMENT
Portfolio assessment is the systematic, longitudinal collection of student work created in response to
specific, know instructional objectives and evaluated in relation to the same criteria. Student Portfolio is a
purposeful collection of student work that exhibits the students efforts, progress and achievements in one
or more areas. The collection must include student participation in selecting contents, the criteria for
selection, the criteria for judging merit and evidence of student self-reflection.
Comparison of Portfolio and Traditional Forms of Assessment
Traditional Assessment Portfolio Assessment
Measures student’s ability at one time Measures student’s ability over time
Done by the teacher alone, students are not aware
of the criteria
Done by the teacher and the students, the students
are aware of the criteria
Conducted outside instruction Embedded in instruction
Assigns student a grade Involves student in own assessment
Does not capture the students language ability Capture many facets if language learning
performance
Does not include the teacher’s knowledge of
student as a learner
Allows for expression of teacher’s knowledge of
student as learner
Does not gives student responsibility Student learns how to take responsibility
Three Types of Portfolio
There are three basic typesof portfolio to consider for classroom use.These are working portfolio, showcase
portfolio and progress portfolio
1. Working Portfolio
The first type of portfolio is working portfolio also known as “teacher-student portfolio”. As the
name implies that it is a project “in work” it contains the work in progress as well as the finished
samples of work use to reflect in process by the students and teachers. It documents the stages of
learning and provides a progressive record of student growth. This is interactive teacher-student
portfolio that aids in communication between teacher and student.
The working portfolio may be used to diagnose student needs. In both student and teacher have
evidence of student strengths and weakness in achieving learning objectives, information
extremely useful in designing future instruction.
2. Showcase Portfolio
Showcase portfolio is the second type of portfolio and also know as best works portfolio or display
portfolio. In this kind of portfolio, it focuses on the student’s best and most representative work. It
exhibit the best performance of the student. Best works portfolio may document student activities
beyond school for example a story written athome. It is just like an artist’s portfolio where a variety
of work is selected to reflect breadth of talent, painters can exhibits the best paintings. Hence,in
this portfolio the student selects what he or she thinks is representative work. This folder is most
often seen at open houses and parent visitations.
The most rewarding use of student portfolios is the display of student’s best work, the work that
makes them proud. In this case, it encourages self-assessment and build self-esteem to students.
The pride and sense of accomplishment that students feel make the effort well worthwhile and
contribute to a culture for learning in the classroom
3. Progress Portfolio
This third type of portfolio is progress portfolio and it is also known as Teacher Alternative
Assessment Portfolio. It contains examples of student’s work with the same types done over a
period of time and they are utilized to assess their progress
All the works of the students in this type of portfolio are scored, rated, ranked, or evaluated.
Teacherscan keep individual student portfolios that are solely for the teacher’suse asanassessment
tool. This a focused type of portfolio and is a model approach to assessment.
Assessmentportfolios used to document student learning on specific curriculum outcomes and used
to demonstrate the extent of mastery in any curricular area,
Uses of Portfolios
1. It can provide both formative and summative opportunities for monitoring progress toward
reaching identified outcomes
2. Portfolios can communicate concrete information about what us expected of students in terms
of the content and quality of performance in specific curriculum areas.
3. A portfolio is that they allow students to document aspects of their learning that do not show
up well in traditional assessments
4. Portfolios are useful to showcase periodic or end of the year accomplishment of students such
as in poetry, reflections on growth, samples of best works, etc.
5. Portfolios may also be used to facilitate communication betweenteachersandparentsregarding
their child’s achievement and progress in a certain period of time.
6. The administrator may use portfolios for national competency testing to grant high school
credit, to evaluate education programs.
7. Portfolios may be assembled for combination of purposes such as instructional enhancement
and progress documentation. Ateacherreviewsstudents portfolios periodically and make notes
for revising instruction for next year used.
According to Mueller (2010) there are seven steps in developing portfolios of students.
Below are the discussions of each step.
1. Purpose: What is the purposes of the portfolio?
2. Audience: For what audience will the portfolio be created?
3. Content: What samples of student work will be included?
4. Process:What processes (e.g. selection of work to be included, reflection in work, conferencing)
will be engaged in during the development of the portfolio?
5. Management: How will time and materials be managed in the development of the portfolio?
6. Communication: How and when will the portfolio be shared with pertinent audiences?
7. Evaluation: If the portfolio is to be used for evaluation, when and how should it be evaluated?
Guidelines for Assessing Portfolios
1. Include enough documents (items) on which to base judgment
2. Structure the contents to provide scorable information
3. Develop judging criteria and a scoring scheme fir raters to use in assessing the portfolios
4. Use observation instruments such as checklists and rating when possible to facilitate scoring.
5. Use trained evaluators or assessors
Guidance and Counseling
Guidance and Counseling are both process to solve problems of life, they differ only on the approach used.
In guidance the client’s problems are listened carefully and readymade solutions are provided by the
experts. While in counseling the client’s problem are discussed and relevant information are provided in-
between. Through these information, the client will gain an insight to the problem and become empowered
to take his own decision.
Guidance Counselor assist each student to benefit from the school experience through attention to their
personal, social and academic needs.
Guidance (Downing) as pointed out by Lao (2006) is an organized set of specialized services established
as an integral part of the school environment designed to promote the development of students and assist
them toward a realization of sound, wholesome adjustment and maximum accomplishment commensurate
with their potentialities.
Guidance (Good) is a process id dynamic interpersonal relationship designed to influence the attitude and
subsequent behavior of the person.
Counseling is both process and relationship. It is a process by which concentrated attention is given by both
counselor and counselee to the problems and concerns of the students in a setting of privacy, warmth,
mutual acceptance and confidentiality. As a process it utilizes appropriate tools and procedure which
contribute to experience. Counseling is also a relationship characterized by trust, confidence and intimacy
in which the students gains intellectual and emotional stability from which he can resolve difficulties, make
plans and realize greatest self-fulfillment.
Villar (2207) pointed out the different guidance services basedfrom Rules and Regulations of Republic Act
9258, Rule 1, Section 3 Manila standard, 2007) and other services not mentioned in Rules and Regulations
1. Individual inventory/ analysis
2. Information
3. Counseling
4. Research
5. Placement
6. Referral
7. Follow-up
8. Evaluation
9. Consultation
10. Program development
11. Public relations
Roles of the Guidance Counselor
There are 5 roles of the guidance counselor are discussed by Dr. Imelda V.G. Villar in her book
“implementing a comprehensive Guidance and Counseling Programs in the Philippines (2007)
1. As Counselor
2. As Coordinator
3. As Consultant
4. As Conductor of Activities
5. As Change Agent
Essential Elements of Counseling Process
1. Anticipating the interview
2. Developing a positive working relationship
3. Exploring feelings and attitudes
4. Reviewing and determining present status
5. Exploring alternatives
6. Reading decision
7. Post counseling contact
Techniques and Methodologies used in the Guidance Process
1. Autobiography
2. Anecdotal record
3. Case study
4. Cumulative record
5. Interview
6. Observation
7. Projective techniques
8. Rating scale
9. Sociometry
Ethical Consideration of the Counselor
1. Counselor’s responsibility to the client and to his family
2. Recognize the boundaries of their competence and their own personal and professional limitations
3. Confidentiality
4. Imposition of one’s values and philosophy of life on the client is considered unethical.
Four Important Functions of Guidance Services
1. Counseling
 Individual counseling
 Small group counseling
 Crisis counseling
 Career counseling
 Referrals
 Peer helping programs
2. Prevention
 Primary, secondary, tertiary plans and programs
 Individual assessments coordinated student support team activities
 Students activities
 Transitional planning

More Related Content

What's hot

Grading and reporting
Grading and reportingGrading and reporting
Grading and reportingReynel Dan
 
Alternative-Response Test
Alternative-Response TestAlternative-Response Test
Alternative-Response TestMD Pits
 
K to12 ASSESSMENT AND RATING OF LEARNING OUTCOMES
K to12 ASSESSMENT AND RATING OF LEARNING OUTCOMESK to12 ASSESSMENT AND RATING OF LEARNING OUTCOMES
K to12 ASSESSMENT AND RATING OF LEARNING OUTCOMESYanne Evangelista
 
authentic vs. traditional assessment
authentic vs. traditional assessmentauthentic vs. traditional assessment
authentic vs. traditional assessmentfreshious
 
Rubrics (Analytic and Holistic)
Rubrics (Analytic and Holistic)Rubrics (Analytic and Holistic)
Rubrics (Analytic and Holistic)Jeziel Dela Cruz
 
Assessment of Learning Outcomes in the k to 12 program
Assessment of Learning Outcomes in the k to 12 programAssessment of Learning Outcomes in the k to 12 program
Assessment of Learning Outcomes in the k to 12 programKerwin Palpal
 
Shift to educational focus from content to learning outcome
Shift to educational focus from content to learning outcomeShift to educational focus from content to learning outcome
Shift to educational focus from content to learning outcomeThreecia Flores
 
K 12 basic education program
K 12 basic education programK 12 basic education program
K 12 basic education program19710802
 
Theories of instructional materials
Theories of instructional materialsTheories of instructional materials
Theories of instructional materialsMarkElnar1
 
PORTFOLIO/PRODUCT/PROCESS ORIENTED PERFORMANCE-BASED ASSESSMENT, Ed8
PORTFOLIO/PRODUCT/PROCESS ORIENTED PERFORMANCE-BASED ASSESSMENT, Ed8PORTFOLIO/PRODUCT/PROCESS ORIENTED PERFORMANCE-BASED ASSESSMENT, Ed8
PORTFOLIO/PRODUCT/PROCESS ORIENTED PERFORMANCE-BASED ASSESSMENT, Ed8Eddie Abug
 
General principles of testing to Different qualities of high quality assessment
General principles of testing to Different qualities of high quality assessmentGeneral principles of testing to Different qualities of high quality assessment
General principles of testing to Different qualities of high quality assessmentHannah Borja
 
Outcomes-Based Education (OBE)
Outcomes-Based Education (OBE)Outcomes-Based Education (OBE)
Outcomes-Based Education (OBE)Sheila Lavapie
 
Properties of Assessment Methods
Properties of Assessment MethodsProperties of Assessment Methods
Properties of Assessment MethodsMonica Angeles
 
Steps in developing performance based assessment
Steps in developing performance based assessmentSteps in developing performance based assessment
Steps in developing performance based assessmentmavs morales
 
Measurement, Assessment and Evaluation
Measurement, Assessment and EvaluationMeasurement, Assessment and Evaluation
Measurement, Assessment and EvaluationMelanio Florino
 
Ed8 Assessment of Learning 2
Ed8 Assessment of Learning 2 Ed8 Assessment of Learning 2
Ed8 Assessment of Learning 2 Eddie Abug
 
Basic Concept in Assessment
Basic Concept in AssessmentBasic Concept in Assessment
Basic Concept in AssessmentJarry Fuentes
 
K to 12 classroom assessment ppt
K to 12 classroom assessment pptK to 12 classroom assessment ppt
K to 12 classroom assessment pptCarlo Magno
 

What's hot (20)

Grading and reporting
Grading and reportingGrading and reporting
Grading and reporting
 
Alternative-Response Test
Alternative-Response TestAlternative-Response Test
Alternative-Response Test
 
K to12 ASSESSMENT AND RATING OF LEARNING OUTCOMES
K to12 ASSESSMENT AND RATING OF LEARNING OUTCOMESK to12 ASSESSMENT AND RATING OF LEARNING OUTCOMES
K to12 ASSESSMENT AND RATING OF LEARNING OUTCOMES
 
authentic vs. traditional assessment
authentic vs. traditional assessmentauthentic vs. traditional assessment
authentic vs. traditional assessment
 
Rubrics (Analytic and Holistic)
Rubrics (Analytic and Holistic)Rubrics (Analytic and Holistic)
Rubrics (Analytic and Holistic)
 
Assessment of Learning Outcomes in the k to 12 program
Assessment of Learning Outcomes in the k to 12 programAssessment of Learning Outcomes in the k to 12 program
Assessment of Learning Outcomes in the k to 12 program
 
Shift to educational focus from content to learning outcome
Shift to educational focus from content to learning outcomeShift to educational focus from content to learning outcome
Shift to educational focus from content to learning outcome
 
K 12 basic education program
K 12 basic education programK 12 basic education program
K 12 basic education program
 
Theories of instructional materials
Theories of instructional materialsTheories of instructional materials
Theories of instructional materials
 
Test Construction
Test ConstructionTest Construction
Test Construction
 
PORTFOLIO/PRODUCT/PROCESS ORIENTED PERFORMANCE-BASED ASSESSMENT, Ed8
PORTFOLIO/PRODUCT/PROCESS ORIENTED PERFORMANCE-BASED ASSESSMENT, Ed8PORTFOLIO/PRODUCT/PROCESS ORIENTED PERFORMANCE-BASED ASSESSMENT, Ed8
PORTFOLIO/PRODUCT/PROCESS ORIENTED PERFORMANCE-BASED ASSESSMENT, Ed8
 
General principles of testing to Different qualities of high quality assessment
General principles of testing to Different qualities of high quality assessmentGeneral principles of testing to Different qualities of high quality assessment
General principles of testing to Different qualities of high quality assessment
 
Outcomes-Based Education (OBE)
Outcomes-Based Education (OBE)Outcomes-Based Education (OBE)
Outcomes-Based Education (OBE)
 
Properties of Assessment Methods
Properties of Assessment MethodsProperties of Assessment Methods
Properties of Assessment Methods
 
Item analysis
Item analysisItem analysis
Item analysis
 
Steps in developing performance based assessment
Steps in developing performance based assessmentSteps in developing performance based assessment
Steps in developing performance based assessment
 
Measurement, Assessment and Evaluation
Measurement, Assessment and EvaluationMeasurement, Assessment and Evaluation
Measurement, Assessment and Evaluation
 
Ed8 Assessment of Learning 2
Ed8 Assessment of Learning 2 Ed8 Assessment of Learning 2
Ed8 Assessment of Learning 2
 
Basic Concept in Assessment
Basic Concept in AssessmentBasic Concept in Assessment
Basic Concept in Assessment
 
K to 12 classroom assessment ppt
K to 12 classroom assessment pptK to 12 classroom assessment ppt
K to 12 classroom assessment ppt
 

Similar to Assessment Types & Modes

Learning Assessmentm PPT.pptx
Learning Assessmentm PPT.pptxLearning Assessmentm PPT.pptx
Learning Assessmentm PPT.pptxMJSanchez8
 
ASSESSMENT IN LEARNING 1-LESSONS 1-4 (1).ppt
ASSESSMENT IN LEARNING 1-LESSONS 1-4 (1).pptASSESSMENT IN LEARNING 1-LESSONS 1-4 (1).ppt
ASSESSMENT IN LEARNING 1-LESSONS 1-4 (1).pptOscarAncheta
 
Roles of Assessment in Classroom Instruction
Roles of Assessment in Classroom InstructionRoles of Assessment in Classroom Instruction
Roles of Assessment in Classroom InstructionJames Robert Villacorteza
 
assessment of student learning in assessment in learning 1
assessment of student learning in assessment in learning 1assessment of student learning in assessment in learning 1
assessment of student learning in assessment in learning 1Rai Blanquera
 
What-is-Educational-Assessment.pptx
What-is-Educational-Assessment.pptxWhat-is-Educational-Assessment.pptx
What-is-Educational-Assessment.pptxANIOAYRochelleDaoaya
 
Karen villanueva
Karen villanuevaKaren villanueva
Karen villanuevaHazel Anog
 
Assessment and Evaluation 2011.pptx
Assessment and Evaluation 2011.pptxAssessment and Evaluation 2011.pptx
Assessment and Evaluation 2011.pptxSendafa edget
 
M1_Outcomes-Based Assessment in Mathematics_1.pdf
M1_Outcomes-Based Assessment in Mathematics_1.pdfM1_Outcomes-Based Assessment in Mathematics_1.pdf
M1_Outcomes-Based Assessment in Mathematics_1.pdfMartin Nobis
 
Chapter 1 Basic Concept in Assessment
Chapter 1 Basic Concept in AssessmentChapter 1 Basic Concept in Assessment
Chapter 1 Basic Concept in AssessmentKaenah Faye Padongao
 
Test Development and Evaluation
Test Development and Evaluation Test Development and Evaluation
Test Development and Evaluation HennaAnsari
 
NED 203 Criterion Referenced Test & Rubrics
NED 203 Criterion Referenced Test & RubricsNED 203 Criterion Referenced Test & Rubrics
NED 203 Criterion Referenced Test & RubricsCarmina Gurrea
 
Types of Assessment in Classroom
Types of Assessment in ClassroomTypes of Assessment in Classroom
Types of Assessment in ClassroomS. Raj Kumar
 
Evaluation of educational programs in nursing
Evaluation of educational programs in nursingEvaluation of educational programs in nursing
Evaluation of educational programs in nursingNavjyot Singh
 
teaching material
teaching material teaching material
teaching material Kadek Astiti
 
501 assignment 9 chapter 12 revision
501 assignment 9 chapter 12 revision501 assignment 9 chapter 12 revision
501 assignment 9 chapter 12 revisionKaWright1961
 
Learning_activity1_Abarca_Garzon_Scarlet_Paulette.pptx
Learning_activity1_Abarca_Garzon_Scarlet_Paulette.pptxLearning_activity1_Abarca_Garzon_Scarlet_Paulette.pptx
Learning_activity1_Abarca_Garzon_Scarlet_Paulette.pptxSCARLETPAULETTEABARC
 
Keys to Effective Testing
Keys to Effective TestingKeys to Effective Testing
Keys to Effective TestingVanz Justine
 

Similar to Assessment Types & Modes (20)

Learning Assessmentm PPT.pptx
Learning Assessmentm PPT.pptxLearning Assessmentm PPT.pptx
Learning Assessmentm PPT.pptx
 
ASSESSMENT IN LEARNING 1-LESSONS 1-4 (1).ppt
ASSESSMENT IN LEARNING 1-LESSONS 1-4 (1).pptASSESSMENT IN LEARNING 1-LESSONS 1-4 (1).ppt
ASSESSMENT IN LEARNING 1-LESSONS 1-4 (1).ppt
 
Roles of Assessment in Classroom Instruction
Roles of Assessment in Classroom InstructionRoles of Assessment in Classroom Instruction
Roles of Assessment in Classroom Instruction
 
assessment of student learning in assessment in learning 1
assessment of student learning in assessment in learning 1assessment of student learning in assessment in learning 1
assessment of student learning in assessment in learning 1
 
Ed103
Ed103Ed103
Ed103
 
Ed103
Ed103Ed103
Ed103
 
Chapter 11
Chapter 11Chapter 11
Chapter 11
 
What-is-Educational-Assessment.pptx
What-is-Educational-Assessment.pptxWhat-is-Educational-Assessment.pptx
What-is-Educational-Assessment.pptx
 
Karen villanueva
Karen villanuevaKaren villanueva
Karen villanueva
 
Assessment and Evaluation 2011.pptx
Assessment and Evaluation 2011.pptxAssessment and Evaluation 2011.pptx
Assessment and Evaluation 2011.pptx
 
M1_Outcomes-Based Assessment in Mathematics_1.pdf
M1_Outcomes-Based Assessment in Mathematics_1.pdfM1_Outcomes-Based Assessment in Mathematics_1.pdf
M1_Outcomes-Based Assessment in Mathematics_1.pdf
 
Chapter 1 Basic Concept in Assessment
Chapter 1 Basic Concept in AssessmentChapter 1 Basic Concept in Assessment
Chapter 1 Basic Concept in Assessment
 
Test Development and Evaluation
Test Development and Evaluation Test Development and Evaluation
Test Development and Evaluation
 
NED 203 Criterion Referenced Test & Rubrics
NED 203 Criterion Referenced Test & RubricsNED 203 Criterion Referenced Test & Rubrics
NED 203 Criterion Referenced Test & Rubrics
 
Types of Assessment in Classroom
Types of Assessment in ClassroomTypes of Assessment in Classroom
Types of Assessment in Classroom
 
Evaluation of educational programs in nursing
Evaluation of educational programs in nursingEvaluation of educational programs in nursing
Evaluation of educational programs in nursing
 
teaching material
teaching material teaching material
teaching material
 
501 assignment 9 chapter 12 revision
501 assignment 9 chapter 12 revision501 assignment 9 chapter 12 revision
501 assignment 9 chapter 12 revision
 
Learning_activity1_Abarca_Garzon_Scarlet_Paulette.pptx
Learning_activity1_Abarca_Garzon_Scarlet_Paulette.pptxLearning_activity1_Abarca_Garzon_Scarlet_Paulette.pptx
Learning_activity1_Abarca_Garzon_Scarlet_Paulette.pptx
 
Keys to Effective Testing
Keys to Effective TestingKeys to Effective Testing
Keys to Effective Testing
 

More from Kendral Flores

More from Kendral Flores (17)

Biological science Comprehensive LET Reviewer
Biological science Comprehensive LET ReviewerBiological science Comprehensive LET Reviewer
Biological science Comprehensive LET Reviewer
 
Curriculum development
Curriculum developmentCurriculum development
Curriculum development
 
Final coaching review
Final coaching reviewFinal coaching review
Final coaching review
 
General knowledge
General knowledge General knowledge
General knowledge
 
Gen edu set b
Gen edu   set bGen edu   set b
Gen edu set b
 
Foundations of-education-prof.ed
Foundations of-education-prof.edFoundations of-education-prof.ed
Foundations of-education-prof.ed
 
Educational technology-prof.ed
Educational technology-prof.edEducational technology-prof.ed
Educational technology-prof.ed
 
LET REVIEWER
LET REVIEWERLET REVIEWER
LET REVIEWER
 
Reviewer 3
Reviewer 3Reviewer 3
Reviewer 3
 
LET REVIEWER
LET REVIEWERLET REVIEWER
LET REVIEWER
 
Dost SEI-Scholarship Exam Reviewer
Dost SEI-Scholarship Exam ReviewerDost SEI-Scholarship Exam Reviewer
Dost SEI-Scholarship Exam Reviewer
 
Invertibrates
InvertibratesInvertibrates
Invertibrates
 
Vertebrate systematics
Vertebrate systematicsVertebrate systematics
Vertebrate systematics
 
Philippine Literature: Region 1 – Ilocos Region
Philippine Literature: Region 1 – Ilocos RegionPhilippine Literature: Region 1 – Ilocos Region
Philippine Literature: Region 1 – Ilocos Region
 
Taxation
TaxationTaxation
Taxation
 
Man: Sexual being
Man: Sexual beingMan: Sexual being
Man: Sexual being
 
Pingpong
PingpongPingpong
Pingpong
 

Recently uploaded

EPANDING THE CONTENT OF AN OUTLINE using notes.pptx
EPANDING THE CONTENT OF AN OUTLINE using notes.pptxEPANDING THE CONTENT OF AN OUTLINE using notes.pptx
EPANDING THE CONTENT OF AN OUTLINE using notes.pptxRaymartEstabillo3
 
Planning a health career 4th Quarter.pptx
Planning a health career 4th Quarter.pptxPlanning a health career 4th Quarter.pptx
Planning a health career 4th Quarter.pptxLigayaBacuel1
 
Gas measurement O2,Co2,& ph) 04/2024.pptx
Gas measurement O2,Co2,& ph) 04/2024.pptxGas measurement O2,Co2,& ph) 04/2024.pptx
Gas measurement O2,Co2,& ph) 04/2024.pptxDr.Ibrahim Hassaan
 
DATA STRUCTURE AND ALGORITHM for beginners
DATA STRUCTURE AND ALGORITHM for beginnersDATA STRUCTURE AND ALGORITHM for beginners
DATA STRUCTURE AND ALGORITHM for beginnersSabitha Banu
 
How to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERPHow to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERPCeline George
 
Grade 9 Q4-MELC1-Active and Passive Voice.pptx
Grade 9 Q4-MELC1-Active and Passive Voice.pptxGrade 9 Q4-MELC1-Active and Passive Voice.pptx
Grade 9 Q4-MELC1-Active and Passive Voice.pptxChelloAnnAsuncion2
 
ACC 2024 Chronicles. Cardiology. Exam.pdf
ACC 2024 Chronicles. Cardiology. Exam.pdfACC 2024 Chronicles. Cardiology. Exam.pdf
ACC 2024 Chronicles. Cardiology. Exam.pdfSpandanaRallapalli
 
Alper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentAlper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentInMediaRes1
 
ENGLISH6-Q4-W3.pptxqurter our high choom
ENGLISH6-Q4-W3.pptxqurter our high choomENGLISH6-Q4-W3.pptxqurter our high choom
ENGLISH6-Q4-W3.pptxqurter our high choomnelietumpap1
 
Hierarchy of management that covers different levels of management
Hierarchy of management that covers different levels of managementHierarchy of management that covers different levels of management
Hierarchy of management that covers different levels of managementmkooblal
 
Proudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxProudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxthorishapillay1
 
Like-prefer-love -hate+verb+ing & silent letters & citizenship text.pdf
Like-prefer-love -hate+verb+ing & silent letters & citizenship text.pdfLike-prefer-love -hate+verb+ing & silent letters & citizenship text.pdf
Like-prefer-love -hate+verb+ing & silent letters & citizenship text.pdfMr Bounab Samir
 
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️9953056974 Low Rate Call Girls In Saket, Delhi NCR
 
Full Stack Web Development Course for Beginners
Full Stack Web Development Course  for BeginnersFull Stack Web Development Course  for Beginners
Full Stack Web Development Course for BeginnersSabitha Banu
 
Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17Celine George
 
What is Model Inheritance in Odoo 17 ERP
What is Model Inheritance in Odoo 17 ERPWhat is Model Inheritance in Odoo 17 ERP
What is Model Inheritance in Odoo 17 ERPCeline George
 
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptx
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptxECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptx
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptxiammrhaywood
 
Earth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice greatEarth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice greatYousafMalik24
 

Recently uploaded (20)

EPANDING THE CONTENT OF AN OUTLINE using notes.pptx
EPANDING THE CONTENT OF AN OUTLINE using notes.pptxEPANDING THE CONTENT OF AN OUTLINE using notes.pptx
EPANDING THE CONTENT OF AN OUTLINE using notes.pptx
 
Planning a health career 4th Quarter.pptx
Planning a health career 4th Quarter.pptxPlanning a health career 4th Quarter.pptx
Planning a health career 4th Quarter.pptx
 
9953330565 Low Rate Call Girls In Rohini Delhi NCR
9953330565 Low Rate Call Girls In Rohini  Delhi NCR9953330565 Low Rate Call Girls In Rohini  Delhi NCR
9953330565 Low Rate Call Girls In Rohini Delhi NCR
 
Gas measurement O2,Co2,& ph) 04/2024.pptx
Gas measurement O2,Co2,& ph) 04/2024.pptxGas measurement O2,Co2,& ph) 04/2024.pptx
Gas measurement O2,Co2,& ph) 04/2024.pptx
 
DATA STRUCTURE AND ALGORITHM for beginners
DATA STRUCTURE AND ALGORITHM for beginnersDATA STRUCTURE AND ALGORITHM for beginners
DATA STRUCTURE AND ALGORITHM for beginners
 
How to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERPHow to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERP
 
Grade 9 Q4-MELC1-Active and Passive Voice.pptx
Grade 9 Q4-MELC1-Active and Passive Voice.pptxGrade 9 Q4-MELC1-Active and Passive Voice.pptx
Grade 9 Q4-MELC1-Active and Passive Voice.pptx
 
ACC 2024 Chronicles. Cardiology. Exam.pdf
ACC 2024 Chronicles. Cardiology. Exam.pdfACC 2024 Chronicles. Cardiology. Exam.pdf
ACC 2024 Chronicles. Cardiology. Exam.pdf
 
Alper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentAlper Gobel In Media Res Media Component
Alper Gobel In Media Res Media Component
 
ENGLISH6-Q4-W3.pptxqurter our high choom
ENGLISH6-Q4-W3.pptxqurter our high choomENGLISH6-Q4-W3.pptxqurter our high choom
ENGLISH6-Q4-W3.pptxqurter our high choom
 
Hierarchy of management that covers different levels of management
Hierarchy of management that covers different levels of managementHierarchy of management that covers different levels of management
Hierarchy of management that covers different levels of management
 
Proudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxProudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptx
 
Like-prefer-love -hate+verb+ing & silent letters & citizenship text.pdf
Like-prefer-love -hate+verb+ing & silent letters & citizenship text.pdfLike-prefer-love -hate+verb+ing & silent letters & citizenship text.pdf
Like-prefer-love -hate+verb+ing & silent letters & citizenship text.pdf
 
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
 
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdfTataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
 
Full Stack Web Development Course for Beginners
Full Stack Web Development Course  for BeginnersFull Stack Web Development Course  for Beginners
Full Stack Web Development Course for Beginners
 
Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17
 
What is Model Inheritance in Odoo 17 ERP
What is Model Inheritance in Odoo 17 ERPWhat is Model Inheritance in Odoo 17 ERP
What is Model Inheritance in Odoo 17 ERP
 
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptx
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptxECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptx
ECONOMIC CONTEXT - PAPER 1 Q3: NEWSPAPERS.pptx
 
Earth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice greatEarth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice great
 

Assessment Types & Modes

  • 1. PROFESSIONAL EDUCATION - Assessment of Learning Assessment –refersto the process of gathering, describing or quantifying information about the student performance. It includes paper and pencil test, extended responses (example essays) and performance assessment are usually referred to as”authentic assessment” task (example presentation of research work) Measurement-is a process of obtaining a numerical description of the degree to which an individual possesses a particular characteristic. Measurements answers the questions”how much? Evaluation- it refers to the process of examining the performance of student. It also determines whether or not the student has met the lesson instructional objectives. Test–isaninstrument or systematic proceduresdesigned to measure the quality, ability, skill or knowledge of students by giving a set of question in a uniform manner. Since test is a form of assessment,tests also answer the question”how does individual student perform? Testing-is a method used to measure the level of achievement or performance of the learners. It also refers to the administration, scoring and interpretation of an instrument (procedure) designed to elicit information about performance in a simple of a particular area of behavior. Types of Measurement There are two ways of interpreting the student performance in relation to classroom instruction. These are the Norm-reference tests and Criterion-referenced tests. Norm-reference test is a test designed to measure the performance of a student compared with other students. Each individual is compared with other examinees and assigned a score-usually expressed as percentile, a grade equivalent score or a stanine. The achievement of student is reported for broad skill areas, although some norm referenced tests do report student achievement for individual. The purpose is to rank each student with respect to the achievement of others in broad areas of knowledge and to discriminate high and low achievers. Criterion- referenced test is a test designed to measure the performance of students with respect to some particular criterion or standard. Each individual is compared with a pre determined set of standard for acceptable achievement. The performance of the other examinees are irrelevant. A student’s score is usually expressed as a percentage and student achievement is reported for individual skills, The purpose is to determine whether each student has achieved specific skills or concepts. And to find out how mush students know before instruction begins and after it has finished. Other terms less often used for criterion-referenced are objective referenced,domain referenced,content referenced and universe referenced. According to Robert L. Linn and Norma E. gronlund (1995) pointed out the common characteristics and differences of Norm-Referenced Tests and Criterion-Referenced Tests
  • 2. Common Characteristics of Norm-Referenced Test and Criterion-Referenced Tests 1. Both require specification of the achievement domain to be measured 2. Both require a relevant and representative sample of test items 3. Both use the same types of test items 4. Both used the same rules for item writing (except for item difficulty) 5. Both are judge with the same qualities of goodness (validity and reliability) 6. Both are useful in educational assessment Differences between Norm-Referenced Tests and Criterion Referenced Tests Norm –Referenced Tests Criterion-Referenced Tests 1. Typically covers a large domain of learning tasks, with just few items measuring each specific task. 1.Typically focuses on a delimited domain of learning tasks, with a relative large number of items measuring each specific task. 2. Emphasizes discrimination among individuals in terms of relative of level of learning. 2.Emphasizes among individuals can and cannot perform. 3. Favors items of large difficulty and typically omits very easy and very hard items 3.Matches item difficulty to learning tasks, without altering item difficulty or omitting easy or hard times 4. Interpretation requires clearly defined group 4.Interpretation requires a clearly defined and delimited achievement domain TYPES OF ASSESSMENT There are four type of assessmentin termsof their functional role in relation to classroom instruction. These are the placement assessment, diagnostic assessment, formative assessment and summative assessment. A. Placement Assessment is concerned with the entry performance of student, the purpose of placement evaluation is to determine the prerequisite skills, degree of mastery of the course objectives and the best mode of learning. B. Diagnostic Assessment is a type of assessment given before instruction. It aims to identify the strengths and weaknesses of the students regarding the topics to be discussed. The purpose of diagnostic assessment: 1. To determine the level of competence of the students 2. To identify the students who have already knowledge about the lesson; 3. To determine the causes of learning problems and formulate a plane for remedial action. C. Formative Assessment is a type of assessment used to monitor the learning progress of the students during or after instruction. Purpose of formative assessment: 1. To provide feed back immediately to both student and teacherregarding the successandfailure of learning.
  • 3. 2. To identify the learning errors that is need of correction 3. To provide information to the teacher for modifying instruction and used for improving learning and instruction D. Summative Assessmentis a type of assessmentusually given at the end of a course or unit. Purpose of summative assessment: 1. To determine the extent to which the instructional objectives have been met; 2. To certify student mastery of the intended outcome and used for assigning grades; 3. To provide information for judging appropriateness of the instructional objectives 4. To determine the effectiveness of instruction MODE OF ASSESSMENT A. Traditional Assessment 1. Assessment in which students typically select an answer or recall information to complete the assessment. Test may be standardized or teacher made test,these tests may be multiple-choice, fill-in-the-blanks, true-false, matching type. 2. Indirect measures of assessment since the test items are designed to represent competence by extracting knowledge and skills from their real life context. 3. Items on standardized instrument tends to test only the domain of knowledge and skill to avoid ambiguity to the test takers. 4. One-time measures to rely on a single correct answer to each item. There is a limited potential for traditional test to measure higher order thinking skills. B. Performance assessment 1. Assessment in which students are asked to perform real-world tasks that demonstrate meaningful application of essential knowledge and skills 2. Direct measures of students performance because task are design to incorporate contexts, problems, and solutions strategies that students would use in real life. 3. Designed ill-structured challenges since the goal is to help students prepare for the complex ambiguities in life. 4. Focus on processes and rationales. There is no single correct answer,instead students are led to craft polished, thorough and justifiable responses, performances and products. 5. Involve long-range projects, exhibits, and performances are linked to the curriculum 6. Teacher is an important collaborator in creating tasks, as well as in developing guidelines for scoring and interpretation C. Portfolio Assessment 1. Portfolio is a collection of student’s work specifically to tell a particular story about the student. 2. A portfolio is not a pie of student work that accumulates over a semester or year 3. A portfolio contains a purposefully selected subset of student work 4. It measures the growth and development of students.
  • 4. The Key to Effective Testing Objectives; The specific statements of the aim of the instruction; it should express what the students should be able to do or know as a result of taking the course; the objectives should indicate the cognitive level, affective level and psychomotor level of expected performance. Instruction: It consist all the elements of the curriculum designed to teach the subject, including the lesson plans, study guide, and reading and homework assignment; the instruction should corresponds directly to the objectives Assessment:The process of gathering , describing or quantifying information about the performance of the learner; testing components of the subject; the weight given to different subject matter areas on the test should match with objectives as well as the emphasis given to each subject area during instruction. Evaluation: Examining the performance of students and comparing and judging its quality. Determining whether or not the learner has met the objectives of the lesson and the extent of understanding. INSTRUCTIONAL OBJECTIVES Instructional objectives play a very important role in the instructional process and the evaluation process. It serves as guides for teaching and learning, communicate the intent of the instruction to others and it provide a guidelines for assessing the learning of the students. Instructional objectives also known as behavioral objectives or learning objectives are statement which clearly describe an anticipated learning outcome. Characteristics of well-written and useful instructional objectives 1. Describe a learning outcome 2. Be student oriented-focus on the learner not on the teacher 3. Be observable or describe an observable product 4. Be sequentially appropriate 5. Be attainable within a reasonable amount of time 6. Be developmental appropriate Factors to Consider when Constructing Good Test Items A. VALIDITYis the degree to which the testmeasureswhat is intended to measure.It is the usefulness of the test for a given purpose. A valid test us always reliable. B. RELIABILITY refers to the consistency of score obtained by the same person when retested using the same instrument or one that is parallel to it. C. ADMINISTRABILITY the test should be administered uniformly to all students so that the scores obtained will not vary due to factors other than differences of the students knowledge and skills. There should be a clear provision for instruction for the students, proctors and even the who will check the test or the scorer
  • 5. D. SCORABILITY the test should be easy to score,directions for scoring is clear,provide the answer sheet and the answer key E. APPROPRIATENESS the test item that the teacher construct must assess the exact performances called for in the learning objectives. The test item should require the same performance of the student as specified in the learning objectives. F. ADEQUACY the test should contain a wide sampling if items to determine the educational outcomes or abilities so that resulting scores are representatives of the total performance in the areas measured. G. FAIRNESS the test should bit be biased to the examinees, it should not be offensive to any examinees subgroups. A test can only be good if it is also fair to all test takers. H. OBJECTIVITY represents the agreement of two or more raters or a test administrators concerning the score of a student. If the two raters who assess the same student on the same test cannot agree in score, the test lacks objectivity and the score of neither judge is valid, thus, lack of objectivity reduces test validity in the same way that lack reliability influence validity. TABLE OF SPECIFICATIONS Table of specification is a device for describing test items in terms of the content and the process dimensions. That is, what a student is expected to know and what he or she is expected to do with that knowledge. It is described by combination of content and process in the table of specification. Sample of One way table of specification in Linear Function Content Number of Class Sessions Number of Items Test Item Distribution 1. Definition of linear function 2 4 1-4 2. Slope of a line 2 4 5-8 3. Graph of linear function 2 4 9-12 4. Equation of linear function 2 4 13-16 5. Standard Forms of a line 3 6 17-22 6. Parallel and perpendicular lines 4 8 23-30 7. Application of linear functions 5 10 31-40 TOTAL 20 40 40
  • 6. Number of items= Number of class sessions x desired total number of itens Total number of class sessions Example : Number of items for the topic” definition of linear function” Number of class session= 2 Desired number of items= 40 Total number of class sessions=20 Number of items= Number of class sessions x desired total number of itens Total number of class sessions =2x40 20 Number of items= 4 Sample of two way table of specification in Linear Function content Class hours Knowledge Comprehension Application Analysis Synthesis Evaluation Total 1.Definitionoflinear function 2 1 1 1 1 4 2.Slope of a line 2 1 1 1 1 3.Graph of linear function 2 1 1 1 1 4 4.Equation of linear function 2 1 1 1 1 4 5.Standard Forms of a line 3 1 1 1 1 1 1 6 6.Parallel and perpendicular line 4 1 2 1 2 8 7.Application of linear functions 5 1 1 3 1 3 10 TOTAL 20 4 6 8 8 7 7 40
  • 7. ITEM ANALYSIS Item analysis refers to the process of examining the student’s responses to each item in the test. According to Abubakar S. Asaad and William M. Hailaya (Measurement and Evaluation Concepts & Principles) Rexr Bookstore (2004 Edition), there are two characteristics of an item. These are desirable and undesirable characteristics. An item that has desirable characteristics can be retained for subsequent use and that with undesirable characteristics is either be revised or rejected. These criteria in determining the desirability and undesirability of an item. a. Difficulty if an item b. Discriminating power of an item c. Measures of attractiveness Difficulty index refers to the proportion of the number of students in the upper and lower groups who answeredan item correctly. In a classroom achievement test,the desired indices of difficulty not lower than 0.20 nor higher than 0.80. the average index difficulty form 0.30 or 0.40 to maximum of 0.60. DF = PUG + PLG 2 PUG = proportion of the upper group who got an item right PLG = proportion of the lower group who get an item right Level of Difficulty of an Item Index Range Difficulty Level 0.00-0.20 Very difficult 0.21-0.40 Difficult 0.41-0.60 Moderately Difficult 0.61-0.80 Easy 0.81-1.00 Very Easy Index of Discrimination Discrimination Index is the differences between the proportion of high performing students who got the item and the proportion of low performing students who got an item right. The high and low performing students usually defined as the upper 27% of the students based on the total examination score and the
  • 8. lower 27% of the students based on total examination score. Discrimination are classified into positive Discrimination if the proportion of students who got an item right in the upper performing group is greater than the students in the upper performing group. And Zero Discrimination if the proportion of the students who got an item right in the upper performing group and low performing group are equal. Discrimination Index Item Evaluation 0.40 and up Very good item 0.30-0.39 Reasonably good item but possibly subject to improvement 0.20-0.29 Marginal, usually needing and being subject to improvement Below 0.19 Poor Item, to be rejected or improved by version Maximum Discrimination is the sum of the proportion of the upper and lower groups who answered the item correctly. Possible maximum discrimination will occur if the half or less of the sum of the upper and lower groups answered an item correctly. Discriminating Efficiency is the index of discrimination divided by the maximum discrimination. PUG = proportion of the upper group who got an item right PLG= proportion of the lower group who got an item right Di = discrimination index DM – Maximum discrimination DE = Discriminating Efficiency Formula: Di = PUG – PLG DE = Di DM DM= PUG + PLG Example: Eighty students took an examination in Algebra, 6 students in the upper group got the correct answer and 4 students in the lower group got the correct answer for item number 6. Find the Discriminating efficiency
  • 9. Given: Number of students took the exam = 80 27% of 80 = 21.6 or 22, which means that there are 22 students in the upper performing group and 22 students in the lower performing group. PUG = 6/22 = 27% PLG = 4/22 = 18% Di = PUG- PLG = 27%- 18% Di= 9% DM = PUG +PLG = 27% + 18% DM= 45% DE = Di/DM = .09/.45 DE = 0.20 or 20% This can be interpreted as on the average, the item is discriminating at 20% of the potential of an item of its difficulty. Measures of Attractiveness To measure the attractiveness of the incorrect option ( distracters) in multiple-choice tests, we count the number if students who selected the incorrect option in both upperand lower groups. The incorrect option is said to be effective distracter if there are more students in the lower group chose that incorrect option than those students in the upper group. Steps of Item Analysis
  • 10. 1. Rank the scores of the students from highest score to lowest score. 2. Select 27% of the paperswithin the upper performing group and 27% of the paperswithin the lower performing group. 3. Set aside the 46% of papers because they will not be used for item analysis. 4. Tabulate the number of students in the upper group and lower group who selected each alternative. 5. Compute the difficulty of each item 6. Compute the discriminating powers of each item 7. Evaluate the effectiveness of the distracters VALIDITY OF A TEST Validity refers to the appropriateness of score-based inferences; or decisions made based on the students test results. The extent to which a test measures what is supposed to measure. Important Things to Remember About Validity 1. Validity refers to the decisions we make, and not to the test itself or to the measurement 2. Like reliability, validity is not all or nothing concept; it is never totally absent or absolutely perfect. 3. Avalidity estimate, called a validity coefficient, referstospecific type of validity. It rangesbetween 0 to 1. 4. Validity can never be finally determined; it is specific to each administration of the test TYPES OF VALIDITY 1. Content Validity- a type of validation that refers to the relationship between a test and the instructional objectives, establishes content so thatthe testmeasureswhatit is supposed to measure. Things to remember about validity: a. The evidence of the content validity of your test is found in the Table of Specification. b. This is the most important type of validity to you, as a classroom teacher. c. There is no coefficient for content validity. It is determined judgmentally, not empirically. 2. Criterion-related Validity- a type of validation that refers to the extent to which scores form a test relate to theoretically similar measures. It is a measure of how accurately a student’s current test score can be used to estimate a score on criterion measure,like performance in courses,classes or another measurementinstrument. Example, classroom reading gradesshould indicate similar levels of performance as Standardized Reading Test scores. a. Construct Validity- a type of validation that refers to a measure of the extent to which a test measures a hypothetical and unobservable variable or quality such as intelligence, math achievement, performance anxiety, etc. it established through intensive study of the test or measurement instrument. b. Predictive Validity- a type of validation that refers to a measure of the extent to which a person’s current test results can used to estimate accurately what that persons performance or other criterion, such as test scores will be at the later time. 3. Concurrent Validity- a type of validation that require the correlation of the predictor or concurrent measure with the criterion measure. Using this, we can determine whether a test is useful to us as
  • 11. predictor or as substitute ( concurrent) measure. The higher the validity coefficient, the better the validity evidence of the test. In establishing the concurrent validity evidence no time interval is involved between the administration of the new test and the criterion or established test. Factors Affecting the Validity of a Test Item 1. The test itself 2. The administration and scoring of a test 3. Personal factors influencing how students response to the test 4. Validity is always specific to a particular group Ways to Reduce the Validity of the Test Item 1. Poorly constructed test items 2. Unclear directions 3. Ambiguous items 4. Reading vocabulary too difficult 5. Complicated syntax 6. Inadequate time limit 7. Inappropriate level of difficulty 8. Unintended clues 9. Improper arrangement of items Test Design to Improve validity 1. What is the purpose of the test? 2. How well do the instructional objectives selected for the test represent the instructional goals. 3. Which test item format will best measure achievement of each objective? 4. How many test item will be required to measure the performance adequately on each objective? 5. When and how will the test be administered? 6. How many test items will required to measure the performance adequately on each objective? 7. When and how will the test be administered? Reliability of a Test Reliability refers to the consistency of measurement; that is, how consistent test results of other assessment results from one measurement to another. We can say that at test is reliable when it can be used to predict practically the same scoreswhen test administered twice to the same group of students and with a reliability index of 0.50 or above. The reliability of a test can be determined by means of Pearson Product Correlation Coeffficient, Spearman-Brown Formula and Kuder-Richardon Formula. Factors Affecting the Reliability of a Test 1. Length of the test 2. Moderate item difficulty 3. Objective scorung
  • 12. 4. Heterogeneity of the student group 5. Limited time Four Methods of Establishing Reliability 1. Test-retest Method. A type of reliability determined by administering the same test twice to the same group of students with any time interval between tests. The result of the test scores are correlated using the Pearson Product Correlation Coefficient and this correlation coefficient provides a measure of stability. This indicates how stable the test result over a period of time. 2. Equivalent –Form Method. A type of reliability determined by administering two different but equivalent forms of the test ( also called parallel or alternate forms) to the same group of students in close succession. The equivalent forms are constructed to the same set of specifications that is similar in content, type of item and difficulty. The result of the test score are correlated using the Pearson Product Correlation Coefficient and this correlation coefficient provides a measure of the degree to which generalization about the performance of students from one assessment to another assessment is justified. It measures the equivalence of the tests. 3. Split –Half method- administer test once. Score two equivalent halves of the test. To split the test into halves that are equivalent, the usual procedure is to score the even-numbered and the odd- numbered separately. This provides a measure of internal consistency. It indicates the degree to which consistent results are obtained from two halves of the test 4. Kuder- Richardon Formula. Administer the test score one. Score total teal and apply the Kuder- Richardon Formula. The Kuder-Richardon formula is applicable only in situation where students responses are scored dichotomously and therefore is most useful with traditional test items that are scored as right or wrong. KR-20 estimates of reliability that provide information about the degree to which the items are of equal difficulty. ( A statistical procedure used to estimate coefficient alpha, or a correlation coefficient is given) Descriptive Statistics of Test Scores Statistics play a very important role in describing the test scores of students. Teachers should have a background on the statistical techniques in orderfor them to analyze and describe the result of measurement obtained in their own classroom; understand the statistics uses in the test and research reports; interpret the types of scores used in testing. Descriptive Statistics- is concerned with collecting, describing, and analyzing a set of data without drawing conclusions or inferences about a large group of data in terms of tables, graphs, or single number (example average score of the class in a particular test) Inferential Statistics- is concerned with the analysis of a subset of data leading to prediction or inferences about the entire set if data or population We shall discussed different statistical technique used in describing and analyzing test results. 1. Measures of Central Tendency (Averages) 2. Measures of Variability ( Spread of Scores 3. Measures of Relationship (Correlation) 4. Skewness
  • 13. Measures of Central Tendency it is a single value that is used to identify the center of the data, it is taught as the typical value in a set of scores. It tends to lie within the center if it is arranged form lowest to highest or vice versa. There are three measures of central tendency commonly used; the mean, median and mode. The Mean The Mean is the common measures of center and it also know as the arithmetic average. Sample Mean = ∑x n ∑= sum of the scores X= individual scores n = number of scores Steps in solving the mean value using raw scores 1. Get the sum of all the scores in the distribution 2. Identify the number of scores (n) 3. Substitute to the given formula and solve the mean value Example: Find the mean of the scores of students in algebra quiz (x) scores in algebra 45 35 48 60 44 39 47 55 58 54 ∑x = 485 n= 10 Mean = ∑x n = 485÷ 10 Mean = 48.5 Properties of Mean 1. Easy to compute 2. It may be an actual observation in the data set 3. It can be subjected to numerous mathematical computation
  • 14. 4. Most widely used 5. Each data affected by the extremes values 6. It is easily affected by the extremes values 7. Applied to interval level data The Median The median is a point that divides the scores in a distribution into two equal parts when the scores are arranged according to magnitude, that is from lowest score to highest score or highest score to lowest score. If the number of score is an odd number, the value of the median is the middle score. When the number of scores is even number, the median values is the average of the two middle scores. Example: 1. Find the median of the scores of 10 students in algebra quiz. (x) scores of students in algebra 45 35 38 60 44 39 47 55 58 54 First , arrange the scores from lowest to highest and find the average of two middle most scores since the number of cases in an even. 35 39 44 45 47 48 54 55 58 60 Mean = 47 + 48 2 = 47.5 is the median score 50% of the scores in the distribution fall below 47.5
  • 15. Example 2. Find the median of the scores of 9 students in algebra quiz (x) scores of students in algebra 35 39 44 45 47 48 54 55 58 The median value is the 5th score which is 47. Which means that 50% of the scores fall below 47. Properties of Median 1. It is not affected by extremes values 2. It is applied to ordinal level of data 3. The middle most score in the distribution 4. Most appropriate when there are extremes scores The Mode The mode refers to the score or scores that occurred most in the distribution. There are classification of mode: a) unimodal is a distribution that consist of only one mode. B) bimodal is a distribution of scores that consist of two modes, c) multimodal is a score distribution that consist of more than two modes. Properties of Mode 1. It is the score/s occurred most frequently 2. Nominal average 3. It can be used for qualitative and quantitative data 4. Not affected by extreme values 5. It may not exist Example 1. Find the mode of the scores of students in algebra quiz: 34,36,45,65,34,45,55,61,34,46 Mode= 34 , because it appeared three times. The distribution is called unimodal. Example 2. Find the mode of the scores of students in algebra quiz: 34,36,45,61,34,45,55,61,34,45 Mode = 34 and 45, because both appeared three times. The distribution is called bimodal
  • 16. Measures of Variability Measures of Variability is a single value that is used to describe the spread out of the scores in distribution, that is above or below the measures of central tendency. There are three commonly used measures variability, the range, quartile deviation and standard deviation The Range Range is the difference between highest and lowest score in the data set. R=HS-LS Properties of Range 1. Simplest and crudest measure 2. A rough measure of variation 3. The smaller the value, the closer the score to each other or the higher the value, the more scattered the scores are. 4. The value easily fluctuate, meaning if there is a changes in either the highest score or lowest score the value of range easily changes. Example: scores of 10 students in Mathematics and Science. Find the range and what subject has a greater variability? Mathematics Science 35 35 33 40 45 25 55 47 62 55 34 35 54 45 36 57 47 39 40 52
  • 17. Mathematics Science HS = 62 HS =57 LS= 33 LS= 25 R = HS-LS R= HS-LS R= 62-33 R= 57-25 R= 29 R= 32 Based form the computed value of the range, the scores in Science has greater variability. Meaning, scores in Science are more scattered than in the scores in Mathematics The Quartile Deviation Quartile Deviation is the half of the differences the third quartile (Q3) and the first quartile (Q1). It is based on the middle 50% of the range, instead the range of the entire set Of distribution. In symbol QD = Q3-Q1 2 QD= quartile deviation Q3= third quartile value Q1= first quartile value Example : In a score of 50 students, the Q3 = 50.25 and Q1 = 25.45, Find the QD QD = Q3-Q1 2 =50.25 – 25.4 2 QD= 12.4 The value of QD =12.4 which indicates the distance we need to go above or below the median to include approximately the middle 50% of the scores. The standard deviation The standard deviation is the most important and useful measures of variation, it is the square root of the variance. It is an average of the degree to which each set of scores in the distribution deviates from the
  • 18. mean value. It is more stable measures of variation because it involves all the scores in a distribution rather than range and quartile deviation. SD = √∑( x-mean)2 n-1 where ,x = individual score n= number of score in a distribution Example: 1. Find the standard deviation of scores of 10 students in algebra quiz. Using the given data below. x (x-mean)2 45 12.25 35 182.25 48 0.25 60 132.25 44 20.5 39 90.25 47 2.25 55 42.25 58 90.25 54 30.25 ∑x= 485 N= 10 Mean = ∑x N ∑(x-mean)2 = 602.25
  • 19. = 485 10 Mean= 48.5 SD= √∑(x-mean)2 n-1 SD= √ 602.5 10-1 SD= √ 66.944444 SD= 8.18, this means that on the average the amount that deviates from the mean value= 48.5 is 8.18 Example 2: Find the standard deviation of the score of 10 students below. In what subject has greater variability Mathematics Science 35 35 33 40 45 25 55 47 62 55 34 35 54 45 36 57
  • 20. 47 39 40 52 Solve for the standard deviation of the scores in mathematics Mathematics (x) (x-mean)2 35 82.81 33 123.21 45 0.81 55 118.81 62 320.41 34 102.01 54 98.01 36 65.61 47 8.41 40 16.81 ∑x = 441 ∑(x-mean)2 = 936.9 Mean = 44.1 ∑(x-mean)2= 918 SD= √∑(x-mean)2 n-1 = √ 936.9 10-1 =√104.1 SD = 10.20 for the mathematics subject
  • 21. Solve for the standard deviation of the score in science Science (x) (x-mean)2 36 64 40 9 25 324 47 16 55 144 35 64 45 4 57 196 39 16 52 81 ∑x= 430 ∑(x-mean)2= 918 Mean =430 10 Mean= 43 SD= √∑(x-mean)2 n-1 = √ 918 10-1 =√ 102 SD= 10.10 for science subject
  • 22. The standard deviation for mathematics subject is 10.20 and the standard deviation foe science subject is 10.10, which means that mathematics scores has a greater variability than science scores. In other words, the scores in mathematics are more scattered than in science. Interpretation of Standard Deviation When the value of standard deviation is large, on the average,the scores will be far form the mean. On the other hand. If the value of standard deviation is small, on the average,the score will be close form the mean. Coefficient of Variation Coefficient of variation is a measure of relative variation expressed as percentage of the arithmetic mean. It is used to compare the variability of two or more sets of data even when the observations are expressed in different units of measurement. Coefficient of variation can be solve using the formula. CV = (SD)x 100% Mean The lower the value of coefficient of variation, the more the overall data approximate to the mean or more the homogeneous the performance of the group Group Mean Standard deviation A 87 8.5 B 90 10.25 CV Group A= standard deviation x 100% Mean = 8.5 x 100% 87 CV Group A=9.77% CV GroupB= standard deviation x 100% Mean = 10.25 x 100% 90 CV Group B=11.39%
  • 23. The CV of Group A is 9.77% and CB of Group B is 11/39%, which means that group A has homogenous performance. Percentile Rank The Percentile rank of a score is the percentage of the scoresin the frequency distribution which are lower. This means that the percentage of the examinees in the norm group who scored below the score of interest. Percentile rank are commonly used to clarify the interpretation of scores on standardized tests. Z- SCORE Z- score (also known as standard score) measures how many standard deviations an observations is above or below the mean.A positive z-score measuresthe number of standard deviation a score is above the mean, and a negative z-negative z-score gives the number of standard deviation a score is below the mean. The z-score can be computed using the formula Z= x-µ for population o Z= x-mean for sample SD Where X= is a raw score 0= is the standard deviation of the population µ= is the mean of the population SD= is the standard deviation of the sample EXAMPLE: James Mark’s examination results in the three subjects are as follows: Subject Mean Standard deviation James Mark’s Grade Math Analysis 88 10 95 Natural Science 85 5 80 Labor Management 92 7.5 94 EXAMPLE:A study showed the performance of two Groups A and B in a certain testgiven by a researcher. Group A obtained a mean score of 87 points with standard deviation of 8.5 points, Group B obtained a
  • 24. mean score of 90 points with standard deviation of 10.25 points. Which of the two group has a more homogeneous performance? In what subject did James Mark performed best? Very Poor? Z math analysis = 95-88 10 Z math analysis = 0.70 Z natural science= 80-85 5 Z natural Science= -1 Z labor management = 94-92 7.5 Z labor management = 0.27 James Mark had a grade in Math Analysis that was 0.70 standard deviation above the mean of the Math Analysis grade,while in Natural Science he was -1.0 standard deviation below the mean of Natural Science grade. He also had a grade in Labor Management that was 0.27 standard deviation above the mean of the Labor Management grades. Comparing the z scores,James Mark performed best in Mathematics Analysis while he performed very poor in Natural Science in relation to the group performance. T-score T-score can be obtained by multiplying the z-score by 10 and adding the product to 50. In symbol, T-score = 10z +50 Using the same exercise,compute the T-score of James Mark in Math Analysis, NaturalScience and Labor Management T- score (math analysis) = 10 (.7) +50 = 57 T- score (natural science) = 10(-1)+50
  • 25. = 40 T-score (labor management) = 10(0.27) +50 =52.7 Since the highest T-score us in math analysis = 57, we can conclude that James Mark performed best in Math analysis than in natural science and labor management. Stanine Stanine also known asstandard nine, is a simple type of normalized standard score that illustrate the process of normalization. Stanines are single digit scores ranging form 1 to 9. The distribution of new scores is divided into nine parts Percent in Stanines 4% 7% 12% 17% 20% 17% 12% 7% 4% Stanines 1 2 3 4 5 6 7 8 9 Skewness Describes the degree of departures of the distribution of the data from symmetry. The degree of skewness is measured by the coefficient of lsewness, denoted as SK and computed as, SK= 3(mean-media) SD Normal curve is a symmetrical bell shaped curve, the end tails are continuous and asymptotic. The mean, median and mode are equal. The scores are normally distributed if the computed value of SK=0 Areas Under the Normal Curve Positively skewed when the curve is skewed to the right, it has a long tail extending off to the right but a short tail to the left. It increases the presence of a small proportion of relatively large extreme value SK˃0 When the computed value of SK is positive most of the scores of students are very low, meaning to say that they performed poor in the said examination Negatively skewed when a distribution is skewed to the left. It has a long tail extending off to the left but a short tail to the right. It indicates the presence of a high proportion of relatively large extreme values SK˂0.
  • 26. When the computed value of SK is negative most of the students got a very high score,meaning to say that they performed very well in the said examination Rubrics Rubrics is a scoring scale and instructional tool to assess the performance of student using a task-specific set of criteria. It contains two essential parts: the criteria for the task and levels of performance for each criterion. It provides teachers an effective means of students-centered feedback and evaluation of the work of students. It also enables teachers to provide a detailed and informative evaluations of their performance. Rubrics is very important most especially if you are measuring the performance of students against a set of standard or pre-determined set of criteria. Through the use of scoring rubrics or rubrics the teachers can determine the strengthens and weaknesses of the students, hence it enables the students to develop their skills. Steps in developing a Rubrics 1. Identify your standards, objectives and goals for your students. Standard is a statement of what the students should be able to know or be able to perform. It should indicate that your students should be able to know or be able to perform. It should indicate that your students should met these standards. Know also the goals for instruction, what are the learning outcomes. 2. Identify the characteristics of a good performance on the task, the criteria, when the students perform or present their work, it should indicate that they performed well in the task given to them; hence they met that particular standards. 3. Identify the levels of performance for each criterion. There is no guidelines with regards to the number of levels of performance,it vary according to the task and needs. It can have as few as two levels of performance or as many as the teacher can develop. In this case,the rater can sufficiently discriminate the performance of the students in each criteria. Through this levels of performance, the teacher or the rater can provide more detailed feedback about the performance of the students. It is easier also for the teacher and students to identify the areas needed for improvement. Types of Rubrics 1. Holistic Rubrics In holistic rubrics does not list a separate levels of performance for each criterion. Rather , holistic, rubrics assigns a level of performance along with a multiple criteria as a whole, in other words you put all the component together. Advantage: quick scoring, provide overview of students achievement. Disadvantage: does not provide detailed information about the student performance in specific areas of the content and skills. May be difficult to provide one overall score. 2. Analytic Rubrics In analytic rubrics the teacher or the rater identify and assess components of a finished product. Breaks down the final product into component parts and each part is scored independently. The
  • 27. total score is the sum of all the rating for all the partsthat are to be assessedorevaluated. In analytic scoring, it is very important for the raterto treateachpart asseparate to avoid bias toward the whole product. Advantage: more detailed feedback, scoring more consistent across students and graders. Disadvantage: time consuming to score. Example of Holistic Rubric 3-Excellent Researcher  Included 10-12 sources  No apparent historical inaccuracies  Can easily tell which sources information was drawn from  All relevant information is included 2- Good Researcher  Included 5-9 sources  Few historical inaccuracies  Can tell with difficulty where information came from  Bibliography contains most relevant information 1-Poor Researcher  Included 1-4 sources  Lots of historical inaccuracies  Cannot tell from which source information came from  Bibliography contains very little information Example of Analytic Rubric Criteria Limited 1 Acceptable 2 Proficient 1 Made good observations Observations are absent or vague Most observations are clear and detailed All observations are clear and detailed Made good predictions Predictions are absent or irrelevant Most predictions are reasonable All predictions are reasonable
  • 28. Appropriate conclusion Conclusion is absent or inconsistent with observation Conclusion is consistent with most observations Conclusion is consistent with observations Advantages of Using Rubrics When assessing the performance of the students using performance based assessment it is very important to use scoring rubrics. The advantages of using rubrics in assessing student’s performance are: 1. Rubrics allow assessment to become more objective and consistent 2. Rubrics clarify the criteria in specific terms 3. Rubrics clearly show the student how work will be evaluated and what is expected 4. Rubrics promote student awareness of the criteria to use in assessing peer performance 5. Rubrics provide useful feedbacks regarding the effectiveness of the instruction: and 6. Rubrics provide benchmarks against which to measure and document progress PERFORMANCE BASED ASSESSMENT Performance based assessment is a direct and systematic observation of actualperformancesof the students based from a pre-determined performance criteria as cited by (Gabuyo, 2011). It is an alternative form of assessing the performance of the students that represent a set of strategies for the application of knowledge, skills and work habits through the performance of tasks that are meaningful and engaging to students” Framework of Assessment Approaches Selection Type Supply Type Product Performance True-false Completion Essay, story or poem Oral presentation of report Multiple-choice Label a diagram Writing portfolio Musical, dance or dramatic performance Matching type Short answer Research report Typing test Concept man Portfolio exhibit, Art exhibit Diving Writing journal Laboratory demonstration Cooperation in group works Forms of Performance Based Assessment
  • 29. 1. Extended response task a. Activities for single assessment may be multiple and varied b. Activities may be extended over a period of time c. Products from different students may be different in focus 2. Restricted-response tasks a. Intended performances more narrowly defined than extended-response tasks. b. Questions may begin like a multiple-choice or short answer stem,but then ask for explanation, or justification. c. May have introductory material like an interpretative exercise,but then asks for an explanation of the answer, not just the answer itself 3. Portfolio is a purposeful collection of student work that exhibits the student’s efforts, progress and achievements in one or more areas. Uses of Performance Based Assessment 1. Assessing the cognitive complex outcomes such as analysis, synthesis and evaluation 2. Assessing non-writing performances and products 3. Must carefully specify the learning outcomes and constructactivity or taskthat actually called forth. Focus of Performance Bases Assessment Performance based assessment can assess the process, or product or both (process and product) depending on the learning outcomes. It also involves doing rather that just knowing about the activity or task. The teacher will assess the effectiveness of the process or procedures and the product used in carrying out the instruction. The question is when to use the process and the product? Use the process when: 1. There is no product 2. The process is orderly and directly observable; 3. Correct procedures/steps in crucial to later success; 4. Analysis of procedural steps can help in improving the product, 5. Learning is at the early age. Use the product when: 1. Different procedures result in an equally good product; 2. Procedures not available for observation; 3. The procedures have been mastered already; 4. Products have qualities that can be identified and judge The final step in performance assessment is to assess and score the student’s performance. To assess the performance of the students the evaluator can used checklist approach , narrative or anecdotal approach, rating scale approach, and memory approach. The evaluator can give feedback on a student’s performance in the form of narrative report or grade. There are different was to record the results of performance-based assessments.
  • 30. 1. Checklist Approach are observation instruments that divide performance whether it is certain or not certain. The teacher has to indicate only whether or not certain elements are present in the performances 2. Narrative/AnecdotalApproach is continuous description of student behavior as it occurs, recorded without judgment or interpretation. The teacherwill write narrative reports of whatwasdone during each of the performances. Form these reports teachers can determine how well their students met their standards. 3. Rating Scale Approach is a checklist that allows the evaluator to record information on a scale, noting the finer distinction that just presence or absence of a behavior. The teacher they indicate to what degree the standards were met. Usually, teachers will use a numerical scale. For instance, one teacher may arte each criterion on a scale of one to five with one meaning “ skills barely present” and five meaning “skill extremely well executed.” 4. Memory Approach the teacher observes the students when performing the tasks without taking any notes. They use the information from memory to determine whether or not the students were successful.This approach is not recommended to use for assessing the performance of the students. PORTFOLIO ASSESSMENT Portfolio assessment is the systematic, longitudinal collection of student work created in response to specific, know instructional objectives and evaluated in relation to the same criteria. Student Portfolio is a purposeful collection of student work that exhibits the students efforts, progress and achievements in one or more areas. The collection must include student participation in selecting contents, the criteria for selection, the criteria for judging merit and evidence of student self-reflection. Comparison of Portfolio and Traditional Forms of Assessment Traditional Assessment Portfolio Assessment Measures student’s ability at one time Measures student’s ability over time Done by the teacher alone, students are not aware of the criteria Done by the teacher and the students, the students are aware of the criteria Conducted outside instruction Embedded in instruction Assigns student a grade Involves student in own assessment Does not capture the students language ability Capture many facets if language learning performance Does not include the teacher’s knowledge of student as a learner Allows for expression of teacher’s knowledge of student as learner Does not gives student responsibility Student learns how to take responsibility
  • 31. Three Types of Portfolio There are three basic typesof portfolio to consider for classroom use.These are working portfolio, showcase portfolio and progress portfolio 1. Working Portfolio The first type of portfolio is working portfolio also known as “teacher-student portfolio”. As the name implies that it is a project “in work” it contains the work in progress as well as the finished samples of work use to reflect in process by the students and teachers. It documents the stages of learning and provides a progressive record of student growth. This is interactive teacher-student portfolio that aids in communication between teacher and student. The working portfolio may be used to diagnose student needs. In both student and teacher have evidence of student strengths and weakness in achieving learning objectives, information extremely useful in designing future instruction. 2. Showcase Portfolio Showcase portfolio is the second type of portfolio and also know as best works portfolio or display portfolio. In this kind of portfolio, it focuses on the student’s best and most representative work. It exhibit the best performance of the student. Best works portfolio may document student activities beyond school for example a story written athome. It is just like an artist’s portfolio where a variety of work is selected to reflect breadth of talent, painters can exhibits the best paintings. Hence,in this portfolio the student selects what he or she thinks is representative work. This folder is most often seen at open houses and parent visitations. The most rewarding use of student portfolios is the display of student’s best work, the work that makes them proud. In this case, it encourages self-assessment and build self-esteem to students. The pride and sense of accomplishment that students feel make the effort well worthwhile and contribute to a culture for learning in the classroom 3. Progress Portfolio This third type of portfolio is progress portfolio and it is also known as Teacher Alternative Assessment Portfolio. It contains examples of student’s work with the same types done over a period of time and they are utilized to assess their progress All the works of the students in this type of portfolio are scored, rated, ranked, or evaluated. Teacherscan keep individual student portfolios that are solely for the teacher’suse asanassessment tool. This a focused type of portfolio and is a model approach to assessment. Assessmentportfolios used to document student learning on specific curriculum outcomes and used to demonstrate the extent of mastery in any curricular area,
  • 32. Uses of Portfolios 1. It can provide both formative and summative opportunities for monitoring progress toward reaching identified outcomes 2. Portfolios can communicate concrete information about what us expected of students in terms of the content and quality of performance in specific curriculum areas. 3. A portfolio is that they allow students to document aspects of their learning that do not show up well in traditional assessments 4. Portfolios are useful to showcase periodic or end of the year accomplishment of students such as in poetry, reflections on growth, samples of best works, etc. 5. Portfolios may also be used to facilitate communication betweenteachersandparentsregarding their child’s achievement and progress in a certain period of time. 6. The administrator may use portfolios for national competency testing to grant high school credit, to evaluate education programs. 7. Portfolios may be assembled for combination of purposes such as instructional enhancement and progress documentation. Ateacherreviewsstudents portfolios periodically and make notes for revising instruction for next year used. According to Mueller (2010) there are seven steps in developing portfolios of students. Below are the discussions of each step. 1. Purpose: What is the purposes of the portfolio? 2. Audience: For what audience will the portfolio be created? 3. Content: What samples of student work will be included? 4. Process:What processes (e.g. selection of work to be included, reflection in work, conferencing) will be engaged in during the development of the portfolio? 5. Management: How will time and materials be managed in the development of the portfolio? 6. Communication: How and when will the portfolio be shared with pertinent audiences? 7. Evaluation: If the portfolio is to be used for evaluation, when and how should it be evaluated? Guidelines for Assessing Portfolios 1. Include enough documents (items) on which to base judgment 2. Structure the contents to provide scorable information 3. Develop judging criteria and a scoring scheme fir raters to use in assessing the portfolios 4. Use observation instruments such as checklists and rating when possible to facilitate scoring. 5. Use trained evaluators or assessors Guidance and Counseling Guidance and Counseling are both process to solve problems of life, they differ only on the approach used. In guidance the client’s problems are listened carefully and readymade solutions are provided by the experts. While in counseling the client’s problem are discussed and relevant information are provided in-
  • 33. between. Through these information, the client will gain an insight to the problem and become empowered to take his own decision. Guidance Counselor assist each student to benefit from the school experience through attention to their personal, social and academic needs. Guidance (Downing) as pointed out by Lao (2006) is an organized set of specialized services established as an integral part of the school environment designed to promote the development of students and assist them toward a realization of sound, wholesome adjustment and maximum accomplishment commensurate with their potentialities. Guidance (Good) is a process id dynamic interpersonal relationship designed to influence the attitude and subsequent behavior of the person. Counseling is both process and relationship. It is a process by which concentrated attention is given by both counselor and counselee to the problems and concerns of the students in a setting of privacy, warmth, mutual acceptance and confidentiality. As a process it utilizes appropriate tools and procedure which contribute to experience. Counseling is also a relationship characterized by trust, confidence and intimacy in which the students gains intellectual and emotional stability from which he can resolve difficulties, make plans and realize greatest self-fulfillment. Villar (2207) pointed out the different guidance services basedfrom Rules and Regulations of Republic Act 9258, Rule 1, Section 3 Manila standard, 2007) and other services not mentioned in Rules and Regulations 1. Individual inventory/ analysis 2. Information 3. Counseling 4. Research 5. Placement 6. Referral 7. Follow-up 8. Evaluation 9. Consultation 10. Program development 11. Public relations Roles of the Guidance Counselor There are 5 roles of the guidance counselor are discussed by Dr. Imelda V.G. Villar in her book “implementing a comprehensive Guidance and Counseling Programs in the Philippines (2007) 1. As Counselor 2. As Coordinator 3. As Consultant 4. As Conductor of Activities 5. As Change Agent
  • 34. Essential Elements of Counseling Process 1. Anticipating the interview 2. Developing a positive working relationship 3. Exploring feelings and attitudes 4. Reviewing and determining present status 5. Exploring alternatives 6. Reading decision 7. Post counseling contact Techniques and Methodologies used in the Guidance Process 1. Autobiography 2. Anecdotal record 3. Case study 4. Cumulative record 5. Interview 6. Observation 7. Projective techniques 8. Rating scale 9. Sociometry Ethical Consideration of the Counselor 1. Counselor’s responsibility to the client and to his family 2. Recognize the boundaries of their competence and their own personal and professional limitations 3. Confidentiality 4. Imposition of one’s values and philosophy of life on the client is considered unethical. Four Important Functions of Guidance Services 1. Counseling  Individual counseling  Small group counseling  Crisis counseling  Career counseling  Referrals  Peer helping programs 2. Prevention  Primary, secondary, tertiary plans and programs  Individual assessments coordinated student support team activities  Students activities  Transitional planning