The document discusses the concepts of validity and reliability in testing. It defines different types of validity including content validity, face validity, criterion-oriented validity, concurrent validity, and construct validity. It also defines internal validity and external validity in research studies. The document then defines reliability and lists different types of reliability such as test-retest reliability, parallel forms reliability, inter-rater reliability, and internal consistency reliability.
It talks about the different types of validity in assessment.
* Face Validity
* Content Validity
* Predictive Validity
* Concurrent Validity
* Construct Validity
It talks about the different types of validity in assessment.
* Face Validity
* Content Validity
* Predictive Validity
* Concurrent Validity
* Construct Validity
This short SlideShare presentation explores a basic overview of test reliability and test validity. Validity is the degree to which a test measures what it is supposed to measure. Reliability is the degree to which a test consistently measures whatever it measures. Examples are given as well as a slide on considerations for writing test questions that demand higher-order thinking.
Topic: What is Reliability and its Types?
Student Name: Kanwal Naz
Class: B.Ed 1.5
Project Name: “Young Teachers' Professional Development (TPD)"
"Project Founder: Prof. Dr. Amjad Ali Arain
Faculty of Education, University of Sindh, Pakistan
What does ‘Reliability’ mean?
Types of Reliability.
Factors which can affect the scores of test papers(reliability).
What does ‘Validity’ mean?
Understanding the differences between reliability and validity.
Observation is one of the important and basic methods or techniques or tools for collecting data in almost all types of research studies.(experimental or laboratory research, descriptive research or qualitative research.)
Observation as a research tool is used in understanding, measuring, prediction and modification of human behavior.
The observer can opt for either participant or non participant observation. This technique can be used to observe the behavior or attitude of the individual or a group. The observer has to plan carefully and execute the plans to collect accurate information. The observation could be supported with interaction.
This method gives an opportunity to know the insights of the group and naturally, the observer will be able to get the factual data compared to other methods of data collection.
Validity:
Validity refers to how well a test measures what it is purported to measure.
Types of Validity:
1. Logic valididty:
Validity which is in the form of theory, statements. It has 2 types.
I. Face Validity:
It is the extent to which the measurement method appears “on its face” to measure the construct of interest.
• Example:
• suppose you were taking an instrument reportedly measuring your attractiveness, but the questions were asking you to identify the correctly spelled word in each list
II. Content Validity:
Measuring all the aspects contributing to the variable of the interest.
Example:
For physical fitness temperature, height and stamina are supposed to be assess then a test of fitness must include content about temperatures, height and stamina.
2. Criterion
It is the extent to which people’s scores are correlated with other variables or criteria that reflect the same construct
Example:
An IQ test should correlate positively with school performance.
An occupational aptitude test should correlate positively with work performance.
Types of Criterion Validity
Concurrent validity:
• When the criterion is something that is happening or being assessed at the same time as the construct of interest, it is called concurrent validity.
• Example:
Beef test.
Predictive validity:
• A new measure of self-esteem should correlate positively with an old established measure. When the criterion is something that will happen or be assessed in the future, this is called predictive validity.
• Example:
GAT, SAT
Other types of validity
Internal Validity:
It is basically the extent to which a study is free from flaws and that any differences in a measurement are due to an independent variable and nothing else
External Validity
• It is the extent to which the results of a research study can be generalized to different situations, different groups of people, different settings, different conditions, etc.
This short SlideShare presentation explores a basic overview of test reliability and test validity. Validity is the degree to which a test measures what it is supposed to measure. Reliability is the degree to which a test consistently measures whatever it measures. Examples are given as well as a slide on considerations for writing test questions that demand higher-order thinking.
Topic: What is Reliability and its Types?
Student Name: Kanwal Naz
Class: B.Ed 1.5
Project Name: “Young Teachers' Professional Development (TPD)"
"Project Founder: Prof. Dr. Amjad Ali Arain
Faculty of Education, University of Sindh, Pakistan
What does ‘Reliability’ mean?
Types of Reliability.
Factors which can affect the scores of test papers(reliability).
What does ‘Validity’ mean?
Understanding the differences between reliability and validity.
Observation is one of the important and basic methods or techniques or tools for collecting data in almost all types of research studies.(experimental or laboratory research, descriptive research or qualitative research.)
Observation as a research tool is used in understanding, measuring, prediction and modification of human behavior.
The observer can opt for either participant or non participant observation. This technique can be used to observe the behavior or attitude of the individual or a group. The observer has to plan carefully and execute the plans to collect accurate information. The observation could be supported with interaction.
This method gives an opportunity to know the insights of the group and naturally, the observer will be able to get the factual data compared to other methods of data collection.
Validity:
Validity refers to how well a test measures what it is purported to measure.
Types of Validity:
1. Logic valididty:
Validity which is in the form of theory, statements. It has 2 types.
I. Face Validity:
It is the extent to which the measurement method appears “on its face” to measure the construct of interest.
• Example:
• suppose you were taking an instrument reportedly measuring your attractiveness, but the questions were asking you to identify the correctly spelled word in each list
II. Content Validity:
Measuring all the aspects contributing to the variable of the interest.
Example:
For physical fitness temperature, height and stamina are supposed to be assess then a test of fitness must include content about temperatures, height and stamina.
2. Criterion
It is the extent to which people’s scores are correlated with other variables or criteria that reflect the same construct
Example:
An IQ test should correlate positively with school performance.
An occupational aptitude test should correlate positively with work performance.
Types of Criterion Validity
Concurrent validity:
• When the criterion is something that is happening or being assessed at the same time as the construct of interest, it is called concurrent validity.
• Example:
Beef test.
Predictive validity:
• A new measure of self-esteem should correlate positively with an old established measure. When the criterion is something that will happen or be assessed in the future, this is called predictive validity.
• Example:
GAT, SAT
Other types of validity
Internal Validity:
It is basically the extent to which a study is free from flaws and that any differences in a measurement are due to an independent variable and nothing else
External Validity
• It is the extent to which the results of a research study can be generalized to different situations, different groups of people, different settings, different conditions, etc.
Characteristics Of A Good Test, Measuring Instrument (Test)
Validity, Nature/Characteristics Of Validity
Types/Approaches To Test Validation
Validity: Advantages And Disadvantages
Reliability, Nature/Characteristics
Types Of Reliability
Methods Of Estimating Reliability
Practicality/Usability
Objectivity
Norms
The research paper has developed over the past three centuries into a tool to communicate the results of scientific inquiry.
The ability to accurately describe ideas, protocols/procedures, and outcomes are the pillars of scientific writing.
Macroeconomics- Movie Location
This will be used as part of your Personal Professional Portfolio once graded.
Objective:
Prepare a presentation or a paper using research, basic comparative analysis, data organization and application of economic information. You will make an informed assessment of an economic climate outside of the United States to accomplish an entertainment industry objective.
Operation “Blue Star” is the only event in the history of Independent India where the state went into war with its own people. Even after about 40 years it is not clear if it was culmination of states anger over people of the region, a political game of power or start of dictatorial chapter in the democratic setup.
The people of Punjab felt alienated from main stream due to denial of their just demands during a long democratic struggle since independence. As it happen all over the word, it led to militant struggle with great loss of lives of military, police and civilian personnel. Killing of Indira Gandhi and massacre of innocent Sikhs in Delhi and other India cities was also associated with this movement.
Embracing GenAI - A Strategic ImperativePeter Windle
Artificial Intelligence (AI) technologies such as Generative AI, Image Generators and Large Language Models have had a dramatic impact on teaching, learning and assessment over the past 18 months. The most immediate threat AI posed was to Academic Integrity with Higher Education Institutes (HEIs) focusing their efforts on combating the use of GenAI in assessment. Guidelines were developed for staff and students, policies put in place too. Innovative educators have forged paths in the use of Generative AI for teaching, learning and assessments leading to pockets of transformation springing up across HEIs, often with little or no top-down guidance, support or direction.
This Gasta posits a strategic approach to integrating AI into HEIs to prepare staff, students and the curriculum for an evolving world and workplace. We will highlight the advantages of working with these technologies beyond the realm of teaching, learning and assessment by considering prompt engineering skills, industry impact, curriculum changes, and the need for staff upskilling. In contrast, not engaging strategically with Generative AI poses risks, including falling behind peers, missed opportunities and failing to ensure our graduates remain employable. The rapid evolution of AI technologies necessitates a proactive and strategic approach if we are to remain relevant.
Biological screening of herbal drugs: Introduction and Need for
Phyto-Pharmacological Screening, New Strategies for evaluating
Natural Products, In vitro evaluation techniques for Antioxidants, Antimicrobial and Anticancer drugs. In vivo evaluation techniques
for Anti-inflammatory, Antiulcer, Anticancer, Wound healing, Antidiabetic, Hepatoprotective, Cardio protective, Diuretics and
Antifertility, Toxicity studies as per OECD guidelines
How to Make a Field invisible in Odoo 17Celine George
It is possible to hide or invisible some fields in odoo. Commonly using “invisible” attribute in the field definition to invisible the fields. This slide will show how to make a field invisible in odoo 17.
Model Attribute Check Company Auto PropertyCeline George
In Odoo, the multi-company feature allows you to manage multiple companies within a single Odoo database instance. Each company can have its own configurations while still sharing common resources such as products, customers, and suppliers.
2024.06.01 Introducing a competency framework for languag learning materials ...Sandy Millin
http://sandymillin.wordpress.com/iateflwebinar2024
Published classroom materials form the basis of syllabuses, drive teacher professional development, and have a potentially huge influence on learners, teachers and education systems. All teachers also create their own materials, whether a few sentences on a blackboard, a highly-structured fully-realised online course, or anything in between. Despite this, the knowledge and skills needed to create effective language learning materials are rarely part of teacher training, and are mostly learnt by trial and error.
Knowledge and skills frameworks, generally called competency frameworks, for ELT teachers, trainers and managers have existed for a few years now. However, until I created one for my MA dissertation, there wasn’t one drawing together what we need to know and do to be able to effectively produce language learning materials.
This webinar will introduce you to my framework, highlighting the key competencies I identified from my research. It will also show how anybody involved in language teaching (any language, not just English!), teacher training, managing schools or developing language learning materials can benefit from using the framework.
The Roman Empire A Historical Colossus.pdfkaushalkr1407
The Roman Empire, a vast and enduring power, stands as one of history's most remarkable civilizations, leaving an indelible imprint on the world. It emerged from the Roman Republic, transitioning into an imperial powerhouse under the leadership of Augustus Caesar in 27 BCE. This transformation marked the beginning of an era defined by unprecedented territorial expansion, architectural marvels, and profound cultural influence.
The empire's roots lie in the city of Rome, founded, according to legend, by Romulus in 753 BCE. Over centuries, Rome evolved from a small settlement to a formidable republic, characterized by a complex political system with elected officials and checks on power. However, internal strife, class conflicts, and military ambitions paved the way for the end of the Republic. Julius Caesar’s dictatorship and subsequent assassination in 44 BCE created a power vacuum, leading to a civil war. Octavian, later Augustus, emerged victorious, heralding the Roman Empire’s birth.
Under Augustus, the empire experienced the Pax Romana, a 200-year period of relative peace and stability. Augustus reformed the military, established efficient administrative systems, and initiated grand construction projects. The empire's borders expanded, encompassing territories from Britain to Egypt and from Spain to the Euphrates. Roman legions, renowned for their discipline and engineering prowess, secured and maintained these vast territories, building roads, fortifications, and cities that facilitated control and integration.
The Roman Empire’s society was hierarchical, with a rigid class system. At the top were the patricians, wealthy elites who held significant political power. Below them were the plebeians, free citizens with limited political influence, and the vast numbers of slaves who formed the backbone of the economy. The family unit was central, governed by the paterfamilias, the male head who held absolute authority.
Culturally, the Romans were eclectic, absorbing and adapting elements from the civilizations they encountered, particularly the Greeks. Roman art, literature, and philosophy reflected this synthesis, creating a rich cultural tapestry. Latin, the Roman language, became the lingua franca of the Western world, influencing numerous modern languages.
Roman architecture and engineering achievements were monumental. They perfected the arch, vault, and dome, constructing enduring structures like the Colosseum, Pantheon, and aqueducts. These engineering marvels not only showcased Roman ingenuity but also served practical purposes, from public entertainment to water supply.
Instructions for Submissions thorugh G- Classroom.pptxJheel Barad
This presentation provides a briefing on how to upload submissions and documents in Google Classroom. It was prepared as part of an orientation for new Sainik School in-service teacher trainees. As a training officer, my goal is to ensure that you are comfortable and proficient with this essential tool for managing assignments and fostering student engagement.
Welcome to TechSoup New Member Orientation and Q&A (May 2024).pdfTechSoup
In this webinar you will learn how your organization can access TechSoup's wide variety of product discount and donation programs. From hardware to software, we'll give you a tour of the tools available to help your nonprofit with productivity, collaboration, financial management, donor tracking, security, and more.
3. Submitted to:
DR.AYAZ KHAN
Submittes by: (gropu members)
Abdul Rehman 47
Affera Mujahid 01
Iqra ilyas 16
Ishrat Fatima 17
Waqas Ahmad 59
4.
5. Very simply, validity is the extent to which a test
measures what it is supposed to measure. We can
divide the types of validity into logical and
empirical.
7. When we want to find out if the entire
content of the behavior/construct/area is
represented in the test we compare the test
task with the content of the behavior. This is
a logical method, not an empirical one.
8. Basically face validity refers to the degree to
which a test appears to measure what it
purports to measure.
Face validity is a measure of how
representative a research project is ‘at face
value,' and whether it appears to be a good
project.
9. When you are expecting a future
performance based on the scores obtained
currently by the measure, correlate the
scores obtained with the performance. The
later performance is called the criterion and
the current score is the prediction.
10. Concurrent validity is the degree to which
the scores on a test are related to the scores
on another, already established, test
administered at the same time, or to some
other valid criterion available at the same
time.
11. Construct validity is the degree to which a
test measures an intended hypothetical
construct.
12.
13.
14. Internal validity
Extent to which an observed outcome
can be attributed to a planned intervention
15.
16. Refers to the occurrence of events that could
alter the outcome or the results of a study.
Concurrent history – occurs during the study
Eg. Studying the effectiveness of using
musical activities to teach mathematics
concepts. While one teacher uses the
standard curriculum, another teacher is using
the musical activities curriculum
17. Pertains to any changes that occur in the subjects during
the course of the study that are not part of the study
and that might affect the results of the study.
Biological ( growth processes)
Eg. Weight gain or increase in height due to
breakfast or lunch program
Psychological (learning or development)
Eg. Effects of certain instructional techniques on
concept learning of sixth graders, attainment of
certain operational thought during that period has to
be considered
18. Is concerned with the effects on the outcome
of a study of the inconsistent use of a
measurement instrument (what the
instrument is measuring changes during the
duration of the study).
Eg. The effects of fatigue on an achievement
test
19. Relates to the possible effects of a pretest on
the performance of participants in a study on
the posttest.
May alert subjects to the fact that they are being
studied
May affect performance on later administrations
20. Refers to the tendency of extreme scores to
move (or regress) toward the mean score on
subsequent retesting.
Eg. Students scoring below 25% (lowest extreme)
on an IQ test are given a posttest. A higher
posttest score is expected.
21. Refers to the loss of subjects from a study
due to their initial nonavailability or
subsequent withdrawal from the study.
Eg. More high-scoring people drop out from the
experimental group than from the control group
22. Refers to the loss of subjects from a study
due to their initial nonavailability or
subsequent withdrawal from the study.
Eg. More high-scoring people drop out from the
experimental group than from the control group
23. Pertains to the possibility that groups in a
study may possess different characteristics
and that those differences may affect the
results.
Differences in age, ability, gender or
racial/ethnic composition, or any of an almost
unlimited number of ways.
24.
25.
26. Refers to the extent to which the results of a
research study are able to be generalized
confidently to a group larger than the group
that participated in the study. (Bracht &
Glass)
27. Given that there is probably a causal
relationship from construct A to construct B,
how generalizable is this relationship across
persons, settings, and times?
Interaction of selection and treatment (sample to
population)
Interaction of setting and treatment
Interaction of history and treatment (time)
28.
29. Reliability is the degree to which a test
consistently measures whatever it measures.
Errors of measurement that affect reliability
are random errors and errors of measurement
30. Test-retest reliability (same people different
time)
Parallel forms reliability (different people,
same time, different test)
Inter-rater reliability (different people,
same test)
internal consistency reliability (different
question, same construct)
A. Average inter-item correlation
B. Split-half reliability
31. Test-retest
An assessment or test of a person should give
the same results whenever you apply the test
Test-retest reliability evaluates reliability
across time.
This method is particularly used in
experiments
32. Various questions for a personality test are
tried out with a class of students over
several years. This helps the researcher
determine those questions and combinations
that have better reliability
33. When multiple people are giving assessments of some
kind or are the subjects of some test, then similar
people should lead to the same resulting scores. It can
be used to calibrate people, for example those being
used as observers in an experiment.
Inter-rater reliability thus evaluates reliability across
different people.
Two major ways in which inter-rater reliability is used
are (a) testing how similarly people categorize items,
and (b) how similarly people score items.
34. This is the best way of assessing
reliability when you are using
observation, as observer bias very
easily creeps in. It does, however,
assume you have multiple observers,
which is not always the case.
Inter-rater reliability is also known
as inter-observer reliability
35. Two people may be asked to categorize
pictures of animals as being dogs or cats. A
perfectly reliable result would be that they
both classify the same pictures in the same
way.
36.
37. Used to assess the consistency of the results
of two tests constructed in the same way
from the same content domain.
38. Used to assess the consistency of results
across items within a test.
Average inter-item correlation:
is a subtype of internal consistency
reliability. It is obtained by taking all of the
items on a test that probe the same construct
(e.g., reading comprehension), determining
the correlation coefficient for each pair of
items, and finally taking the average of all of
these correlation coefficients.
39. is another subtype of internal consistency
reliability. The process of obtaining split-half
reliability is begun by “splitting in half” all
items of a test that are intended to probe
the same area of knowledge (e.g., World War
II) in order to form two “sets” of items.
40. Test length
Test-retest interval
Variability of scores
Guessing
Variation within the test situation