The concept of teaching is the fundamental study of student's cognitive action. In this paper a Trapezoidal Fuzzy Assessment model (TFAM) is developed for teaching assessment. The TFAM is a new variation of a special form, which is used in Fuzzy mathematics centre of gravity(COG) defuzzification technique. The TFAM's new idea is the replacement of the rectangles appearing in the graph of the COG method by isosceles trapezoids sharing common parts, thus covering the ambiguous cases of teachers' scores being at the limits between two successive grades (e.g between A and B). A classroom application is also presented in which the outcomes of the COG and TFAM methods are compared with those of other traditional assessment methods (calculation of means and GPA index) and explanations are provided for the differences appeared among these outcomes .
Universidad Técnica Particular de Loja
Ciclo Académico Abril Agosto 2011
Carrera: Inglés
Docente: Mgs. Orlando Lizaldes E.
Ciclo: Sexto
Bimestre: Segundo
Decision-Making Model for Student Assessment by Unifying Numerical and Lingui...IJECEIAES
Learning assessment deals with the process of making a decision on the quality or performance of student achievement in a number of competency standards. In the process, teacher’s preferences are provided through both test and non-test, generally in a numeric value, from which the final results are then converted into letters or linguistic value. In the proposed model, linguistic variables are exploited as a form of teacher’s preferences in nontest techniques. Consequently, the assessment data set will consist of numerical and linguistic information, so it requires a method to unify them to obtain the final value. A model that uses the 2-tuple linguistic approach and based on matrix operations is proposed to solve the problem. This study proposed a new procedure that consists of four stages: preprocessing, transformation, aggregation and exploitation. The final result is presented in 2-tuple linguistic representation and its equivalent number, accompanied by a description of the achievement of each competency. The α value of 2-tuple linguistic in the final result and in the description of each competency becomes meaningful information that can be interpreted as a comparative ability one student has related to other students, and shows how much potential is achieved to reach higher ranks. The proposed model contributes to enrich the learning assessment techniques, since the exploitation of linguistic variable as representation preferences provides flexible space for teachers in their assessments. Moreover, using the result with respect to students’ levels of each competency, students’ mastery of each attribute can be diagnosed and their progress of learning can be estimated.
The concept of teaching is the fundamental study of student's cognitive action. In this paper a Trapezoidal Fuzzy Assessment model (TFAM) is developed for teaching assessment. The TFAM is a new variation of a special form, which is used in Fuzzy mathematics centre of gravity(COG) defuzzification technique. The TFAM's new idea is the replacement of the rectangles appearing in the graph of the COG method by isosceles trapezoids sharing common parts, thus covering the ambiguous cases of teachers' scores being at the limits between two successive grades (e.g between A and B). A classroom application is also presented in which the outcomes of the COG and TFAM methods are compared with those of other traditional assessment methods (calculation of means and GPA index) and explanations are provided for the differences appeared among these outcomes .
Universidad Técnica Particular de Loja
Ciclo Académico Abril Agosto 2011
Carrera: Inglés
Docente: Mgs. Orlando Lizaldes E.
Ciclo: Sexto
Bimestre: Segundo
Decision-Making Model for Student Assessment by Unifying Numerical and Lingui...IJECEIAES
Learning assessment deals with the process of making a decision on the quality or performance of student achievement in a number of competency standards. In the process, teacher’s preferences are provided through both test and non-test, generally in a numeric value, from which the final results are then converted into letters or linguistic value. In the proposed model, linguistic variables are exploited as a form of teacher’s preferences in nontest techniques. Consequently, the assessment data set will consist of numerical and linguistic information, so it requires a method to unify them to obtain the final value. A model that uses the 2-tuple linguistic approach and based on matrix operations is proposed to solve the problem. This study proposed a new procedure that consists of four stages: preprocessing, transformation, aggregation and exploitation. The final result is presented in 2-tuple linguistic representation and its equivalent number, accompanied by a description of the achievement of each competency. The α value of 2-tuple linguistic in the final result and in the description of each competency becomes meaningful information that can be interpreted as a comparative ability one student has related to other students, and shows how much potential is achieved to reach higher ranks. The proposed model contributes to enrich the learning assessment techniques, since the exploitation of linguistic variable as representation preferences provides flexible space for teachers in their assessments. Moreover, using the result with respect to students’ levels of each competency, students’ mastery of each attribute can be diagnosed and their progress of learning can be estimated.
Finding and Quantifying Temporal-Aware Contradiction in ReviewsIsmail BADACHE
Opinions (reviews) on web resources (e.g., courses, movies), generated by users, become increasingly exploited in text analysis tasks, the detection of contradictory opinions being one of them. This paper focuses on the quantification of sentiment-based contradictions around specific aspects in reviews. However, it is necessary to study the contradictions with respect to the temporal dimension of reviews (their sessions). In general, for web resources such as online courses (e.g. coursera or edX), reviews are often generated during the course sessions. Between sessions, users stop reviewing courses, and there are chances that courses will be updated. So, in order to avoid the confusion of contradictory reviews coming from two or more different sessions, the reviews related to a given resource should be firstly grouped according to their corresponding session. Secondly, aspects are identified according to the distributions of the emotional terms in the vicinity of the most frequent nouns in the reviews collection. Thirdly, the polarity of each review segment containing an aspect is estimated. Then, only resources containing these aspects with opposite polarities are considered. Finally, the contradiction intensity is estimated based on the joint dispersion of polarities and ratings of the reviews containing aspects. The experiments are conducted on the Massive Open Online Courses data set containing 2244 courses and their 73,873 reviews, collected from \textit{coursera.org}. The results confirm the effectiveness of our approach to find and quantify contradiction intensity.
CLASSIFICATION OF QUESTIONS AND LEARNING OUTCOME STATEMENTS (LOS) INTO BLOOM’...IJMIT JOURNAL
Bloom’s Taxonomy (BT) have been used to classify the objectives of learning outcome by dividing the learning into three different domains; the cognitive domain, the effective domain and the psychomotor domain. In this paper, we are introducing a new approach to classify the questions and learning outcome
statements (LOS) into Blooms taxonomy (BT) and to verify BT verb lists, which are being cited and used by academicians to write questions and (LOS). An experiment was designed to investigate the semantic relationship between the action verbs used in both questions and LOS to obtain more accurate
classification of the levels of BT. A sample of 775 different action verbs collected from different universities allows us to measure an accurate and clear-cut cognitive level for the action verb. It is worth mentioning that natural language processing techniques were used to develop our rules as to induce the questions into
chunks in order to extract the action verbs. Our proposed solution was able to classify the action verb into a precise level of the cognitive domain. We, on our side, have tested and evaluated our proposed solution using confusion matrix. The results of evaluation tests yielded 97% for the macro average of precision and 90% for F1. Thus, the outcome of the research suggests that it is crucial to analyse and verify the action
verbs cited and used by academicians to write LOS and classify their questions based on blooms taxonomy in order to obtain a definite and more accurate classification.
CLASSIFICATION OF QUESTIONS AND LEARNING OUTCOME STATEMENTS (LOS) INTO BLOOM’...IJMIT JOURNAL
Bloom’s Taxonomy (BT) have been used to classify the objectives of learning outcome by dividing the
learning into three different domains; the cognitive domain, the effective domain and the psychomotor
domain. In this paper, we are introducing a new approach to classify the questions and learning outcome
statements (LOS) into Blooms taxonomy (BT) and to verify BT verb lists, which are being cited and used by
academicians to write questions and (LOS). An experiment was designed to investigate the semantic
relationship between the action verbs used in both questions and LOS to obtain more accurate
classification of the levels of BT. A sample of 775 different action verbs collected from different universities
allows us to measure an accurate and clear-cut cognitive level for the action verb. It is worth mentioning
that natural language processing techniques were used to develop our rules as to induce the questions into
chunks in order to extract the action verbs. Our proposed solution was able to classify the action verb into
a precise level of the cognitive domain. We, on our side, have tested and evaluated our proposed solution
using confusion matrix. The results of evaluation tests yielded 97% for the macro average of precision and
90% for F1. Thus, the outcome of the research suggests that it is crucial to analyse and verify the action
verbs cited and used by academicians to write LOS and classify their questions based on blooms taxonomy
in order to obtain a definite and more accurate classification.
he Comparative Study between Grade Level and Spelling Proficiency of Selected...Mariz Pascua
This is an informal research practice using a statistical treatment for the comparative data. Study requires further research and necessary treatment for reliable information.
Harnessing Ratings and Aspect-Sentiment to Estimate Contradiction Intensity i...Ismail BADACHE
Analysis of opinions (reviews) generated by users becomes increasingly exploited by a variety of applications. It allows to follow the evolution of the opinions or to carry out investigations on products. The detection of contradictory opinions about a web resource (e.g., courses, movies, products, etc.) is an important task to evaluate the latter. This paper focuses on the problem of detecting contradictions in reviews based on the sentiment analysis around specific aspects of a resource (document). In general, for web resources such as online courses (e.g. on Coursera or edX), reviews are often generated during course sessions. Between each session users stop reviewing on the course, and this course may have updates. So, in order to avoid the confusion of contradictory reviews coming from two or more different sessions, the reviews related to a given resource should be firstly grouped according to their session. Secondly, certain aspects are extracted according to the distributions of the emotional terms in the vicinity of the most frequent names in the reviews collection. Thirdly, the polarity of each review segment containing an aspect is identified. Then taking only the resources containing these aspects with opposite polarities (positive, negative). Finally, we propose a measure of contradiction intensity based on the joint dispersion of the polarity and the rating of the reviews containing the aspects within each resource. The evaluation of our approach is conducted on the Massive Open Online Courses (MOOC) collection containing 2244 courses and their 73,873 reviews, collected from Coursera. The results of experiments revealed the effectiveness of the proposed approach to capture and quantify contradiction intensity.
Using Class Frequency for Improving Centroid-based Text ClassificationIDES Editor
Most previous works on text classification,
represented importance of terms by term occurrence frequency
(tf) and inverse document frequency (idf). This paper presents
the ways to apply class frequency in centroid-based text
categorization. Three approaches are taken into account. The
first one is to explore the effectiveness of inverse class
frequency on the popular term weighting, i.e., TFIDF, as a
replacement of idf and an addition to TFIDF. The second
approach is to evaluate some functions, which are used to
adjust the power of inverse class frequency. The other approach
is to apply terms, which are found in only one class or few
classes, to improve classification performance, using two-step
classification. From the results, class frequency expresses its
usefulness on text classification, especially the two-step
classification.
How Anchoring Concepts Influence Essay Conceptual Structure And Test PerformanceRoy Clariana
Presented October 21 at CELDA 2023 in Madeira Portugal, https://www.celda-conf.org/
Abstract: This quasi-experimental study seeks to improve the conceptual quality of summary essays by comparing two conditions, essay prompts with or without a list of 13 broad concepts, the concepts were selected across a continuum of the 100 most frequent words in the lesson materials. It is anticipated that only the most central concepts will be used as “anchors” when writing. Participants (n = 90) in an Architectural Engineering undergraduate course read the assigned lesson textbook chapter and attended lectures and labs, then in a final lab session were asked to write a 300-word summary of the lesson content. Data consists of the essays converted to networks and the end-of-unit multiple choice test. Compared to the expert network benchmark, the essay networks of those receiving the broad concepts in the writing prompt were not significantly different from those who did not receive these concepts. However those receiving the broad concepts were significantly more like peer essay networks (mental model convergence) and like the networks of the two PowerPoint lectures but neither were like the textbook chapter. Further, those receiving the broad concepts performed significantly better on the end-of-unit test than those not receiving the concepts. Term frequency analysis of the essays indicates as expected that the most network-central concepts had a greater frequency in essays, the other terms frequencies were remarkably the same for both the terms and no terms groups, suggesting a similar underlying conceptual mental model of this lesson content. To further explore the influence of anchoring concepts in summary writing prompts, essays were generated with the same two summary writing prompts using OpenAI (ChatGPT) and Google Bard, plus a new prompt that used the 13 most central concepts from the expert’s network. The quality of the essay networks for both AI systems were equivalent to the students' essay networks for the broad concepts and for the no concept treatments. However the AI essays derived with the 13 most central concepts were significantly better (more like the expert network) than the students and AI essays derived with broad concepts or no concepts treatments. In addition, Bard and OpenAI used several of the same concepts at a higher frequency than the students suggesting that the two AI systems have more similar knowledge graphs of this content. In sum, adding 13 broad conceptual terms to a summary writing prompt improved both structural and declarative knowledge outcomes, but adding 13 most central concepts may be even better. More research is needed to understand how including concepts and other terms in a writing prompt influences students’ essay conceptual structure and subsequent test performance.
Finding and Quantifying Temporal-Aware Contradiction in ReviewsIsmail BADACHE
Opinions (reviews) on web resources (e.g., courses, movies), generated by users, become increasingly exploited in text analysis tasks, the detection of contradictory opinions being one of them. This paper focuses on the quantification of sentiment-based contradictions around specific aspects in reviews. However, it is necessary to study the contradictions with respect to the temporal dimension of reviews (their sessions). In general, for web resources such as online courses (e.g. coursera or edX), reviews are often generated during the course sessions. Between sessions, users stop reviewing courses, and there are chances that courses will be updated. So, in order to avoid the confusion of contradictory reviews coming from two or more different sessions, the reviews related to a given resource should be firstly grouped according to their corresponding session. Secondly, aspects are identified according to the distributions of the emotional terms in the vicinity of the most frequent nouns in the reviews collection. Thirdly, the polarity of each review segment containing an aspect is estimated. Then, only resources containing these aspects with opposite polarities are considered. Finally, the contradiction intensity is estimated based on the joint dispersion of polarities and ratings of the reviews containing aspects. The experiments are conducted on the Massive Open Online Courses data set containing 2244 courses and their 73,873 reviews, collected from \textit{coursera.org}. The results confirm the effectiveness of our approach to find and quantify contradiction intensity.
CLASSIFICATION OF QUESTIONS AND LEARNING OUTCOME STATEMENTS (LOS) INTO BLOOM’...IJMIT JOURNAL
Bloom’s Taxonomy (BT) have been used to classify the objectives of learning outcome by dividing the learning into three different domains; the cognitive domain, the effective domain and the psychomotor domain. In this paper, we are introducing a new approach to classify the questions and learning outcome
statements (LOS) into Blooms taxonomy (BT) and to verify BT verb lists, which are being cited and used by academicians to write questions and (LOS). An experiment was designed to investigate the semantic relationship between the action verbs used in both questions and LOS to obtain more accurate
classification of the levels of BT. A sample of 775 different action verbs collected from different universities allows us to measure an accurate and clear-cut cognitive level for the action verb. It is worth mentioning that natural language processing techniques were used to develop our rules as to induce the questions into
chunks in order to extract the action verbs. Our proposed solution was able to classify the action verb into a precise level of the cognitive domain. We, on our side, have tested and evaluated our proposed solution using confusion matrix. The results of evaluation tests yielded 97% for the macro average of precision and 90% for F1. Thus, the outcome of the research suggests that it is crucial to analyse and verify the action
verbs cited and used by academicians to write LOS and classify their questions based on blooms taxonomy in order to obtain a definite and more accurate classification.
CLASSIFICATION OF QUESTIONS AND LEARNING OUTCOME STATEMENTS (LOS) INTO BLOOM’...IJMIT JOURNAL
Bloom’s Taxonomy (BT) have been used to classify the objectives of learning outcome by dividing the
learning into three different domains; the cognitive domain, the effective domain and the psychomotor
domain. In this paper, we are introducing a new approach to classify the questions and learning outcome
statements (LOS) into Blooms taxonomy (BT) and to verify BT verb lists, which are being cited and used by
academicians to write questions and (LOS). An experiment was designed to investigate the semantic
relationship between the action verbs used in both questions and LOS to obtain more accurate
classification of the levels of BT. A sample of 775 different action verbs collected from different universities
allows us to measure an accurate and clear-cut cognitive level for the action verb. It is worth mentioning
that natural language processing techniques were used to develop our rules as to induce the questions into
chunks in order to extract the action verbs. Our proposed solution was able to classify the action verb into
a precise level of the cognitive domain. We, on our side, have tested and evaluated our proposed solution
using confusion matrix. The results of evaluation tests yielded 97% for the macro average of precision and
90% for F1. Thus, the outcome of the research suggests that it is crucial to analyse and verify the action
verbs cited and used by academicians to write LOS and classify their questions based on blooms taxonomy
in order to obtain a definite and more accurate classification.
he Comparative Study between Grade Level and Spelling Proficiency of Selected...Mariz Pascua
This is an informal research practice using a statistical treatment for the comparative data. Study requires further research and necessary treatment for reliable information.
Harnessing Ratings and Aspect-Sentiment to Estimate Contradiction Intensity i...Ismail BADACHE
Analysis of opinions (reviews) generated by users becomes increasingly exploited by a variety of applications. It allows to follow the evolution of the opinions or to carry out investigations on products. The detection of contradictory opinions about a web resource (e.g., courses, movies, products, etc.) is an important task to evaluate the latter. This paper focuses on the problem of detecting contradictions in reviews based on the sentiment analysis around specific aspects of a resource (document). In general, for web resources such as online courses (e.g. on Coursera or edX), reviews are often generated during course sessions. Between each session users stop reviewing on the course, and this course may have updates. So, in order to avoid the confusion of contradictory reviews coming from two or more different sessions, the reviews related to a given resource should be firstly grouped according to their session. Secondly, certain aspects are extracted according to the distributions of the emotional terms in the vicinity of the most frequent names in the reviews collection. Thirdly, the polarity of each review segment containing an aspect is identified. Then taking only the resources containing these aspects with opposite polarities (positive, negative). Finally, we propose a measure of contradiction intensity based on the joint dispersion of the polarity and the rating of the reviews containing the aspects within each resource. The evaluation of our approach is conducted on the Massive Open Online Courses (MOOC) collection containing 2244 courses and their 73,873 reviews, collected from Coursera. The results of experiments revealed the effectiveness of the proposed approach to capture and quantify contradiction intensity.
Using Class Frequency for Improving Centroid-based Text ClassificationIDES Editor
Most previous works on text classification,
represented importance of terms by term occurrence frequency
(tf) and inverse document frequency (idf). This paper presents
the ways to apply class frequency in centroid-based text
categorization. Three approaches are taken into account. The
first one is to explore the effectiveness of inverse class
frequency on the popular term weighting, i.e., TFIDF, as a
replacement of idf and an addition to TFIDF. The second
approach is to evaluate some functions, which are used to
adjust the power of inverse class frequency. The other approach
is to apply terms, which are found in only one class or few
classes, to improve classification performance, using two-step
classification. From the results, class frequency expresses its
usefulness on text classification, especially the two-step
classification.
How Anchoring Concepts Influence Essay Conceptual Structure And Test PerformanceRoy Clariana
Presented October 21 at CELDA 2023 in Madeira Portugal, https://www.celda-conf.org/
Abstract: This quasi-experimental study seeks to improve the conceptual quality of summary essays by comparing two conditions, essay prompts with or without a list of 13 broad concepts, the concepts were selected across a continuum of the 100 most frequent words in the lesson materials. It is anticipated that only the most central concepts will be used as “anchors” when writing. Participants (n = 90) in an Architectural Engineering undergraduate course read the assigned lesson textbook chapter and attended lectures and labs, then in a final lab session were asked to write a 300-word summary of the lesson content. Data consists of the essays converted to networks and the end-of-unit multiple choice test. Compared to the expert network benchmark, the essay networks of those receiving the broad concepts in the writing prompt were not significantly different from those who did not receive these concepts. However those receiving the broad concepts were significantly more like peer essay networks (mental model convergence) and like the networks of the two PowerPoint lectures but neither were like the textbook chapter. Further, those receiving the broad concepts performed significantly better on the end-of-unit test than those not receiving the concepts. Term frequency analysis of the essays indicates as expected that the most network-central concepts had a greater frequency in essays, the other terms frequencies were remarkably the same for both the terms and no terms groups, suggesting a similar underlying conceptual mental model of this lesson content. To further explore the influence of anchoring concepts in summary writing prompts, essays were generated with the same two summary writing prompts using OpenAI (ChatGPT) and Google Bard, plus a new prompt that used the 13 most central concepts from the expert’s network. The quality of the essay networks for both AI systems were equivalent to the students' essay networks for the broad concepts and for the no concept treatments. However the AI essays derived with the 13 most central concepts were significantly better (more like the expert network) than the students and AI essays derived with broad concepts or no concepts treatments. In addition, Bard and OpenAI used several of the same concepts at a higher frequency than the students suggesting that the two AI systems have more similar knowledge graphs of this content. In sum, adding 13 broad conceptual terms to a summary writing prompt improved both structural and declarative knowledge outcomes, but adding 13 most central concepts may be even better. More research is needed to understand how including concepts and other terms in a writing prompt influences students’ essay conceptual structure and subsequent test performance.
A CRITICAL REVIEW ON THE OPTIMIZATION METHODS IN SOLVING EXAM TIMETABLING AND...IAEME Publication
The Examination Timetabling problem regards the scheduling for the exams of a set
of university courses, avoiding the overlapping of exams having students in common,
fairly spreading the exams for the students, and satisfying room capacity constraints. This
paper review different optimization techniques used to solve a general time tabling
problems. The basic approach can readily handle a wide variety of exam timetabling
problem constraints, and is therefore likely to be of great practical usefulness. The
approach relies for its success on the use of specially designed mutation operators which
greatly improve upon the performance.
Opendatabay - Open Data Marketplace.pptxOpendatabay
Opendatabay.com unlocks the power of data for everyone. Open Data Marketplace fosters a collaborative hub for data enthusiasts to explore, share, and contribute to a vast collection of datasets.
First ever open hub for data enthusiasts to collaborate and innovate. A platform to explore, share, and contribute to a vast collection of datasets. Through robust quality control and innovative technologies like blockchain verification, opendatabay ensures the authenticity and reliability of datasets, empowering users to make data-driven decisions with confidence. Leverage cutting-edge AI technologies to enhance the data exploration, analysis, and discovery experience.
From intelligent search and recommendations to automated data productisation and quotation, Opendatabay AI-driven features streamline the data workflow. Finding the data you need shouldn't be a complex. Opendatabay simplifies the data acquisition process with an intuitive interface and robust search tools. Effortlessly explore, discover, and access the data you need, allowing you to focus on extracting valuable insights. Opendatabay breaks new ground with a dedicated, AI-generated, synthetic datasets.
Leverage these privacy-preserving datasets for training and testing AI models without compromising sensitive information. Opendatabay prioritizes transparency by providing detailed metadata, provenance information, and usage guidelines for each dataset, ensuring users have a comprehensive understanding of the data they're working with. By leveraging a powerful combination of distributed ledger technology and rigorous third-party audits Opendatabay ensures the authenticity and reliability of every dataset. Security is at the core of Opendatabay. Marketplace implements stringent security measures, including encryption, access controls, and regular vulnerability assessments, to safeguard your data and protect your privacy.
Levelwise PageRank with Loop-Based Dead End Handling Strategy : SHORT REPORT ...Subhajit Sahu
Abstract — Levelwise PageRank is an alternative method of PageRank computation which decomposes the input graph into a directed acyclic block-graph of strongly connected components, and processes them in topological order, one level at a time. This enables calculation for ranks in a distributed fashion without per-iteration communication, unlike the standard method where all vertices are processed in each iteration. It however comes with a precondition of the absence of dead ends in the input graph. Here, the native non-distributed performance of Levelwise PageRank was compared against Monolithic PageRank on a CPU as well as a GPU. To ensure a fair comparison, Monolithic PageRank was also performed on a graph where vertices were split by components. Results indicate that Levelwise PageRank is about as fast as Monolithic PageRank on the CPU, but quite a bit slower on the GPU. Slowdown on the GPU is likely caused by a large submission of small workloads, and expected to be non-issue when the computation is performed on massive graphs.
Adjusting primitives for graph : SHORT REPORT / NOTESSubhajit Sahu
Graph algorithms, like PageRank Compressed Sparse Row (CSR) is an adjacency-list based graph representation that is
Multiply with different modes (map)
1. Performance of sequential execution based vs OpenMP based vector multiply.
2. Comparing various launch configs for CUDA based vector multiply.
Sum with different storage types (reduce)
1. Performance of vector element sum using float vs bfloat16 as the storage type.
Sum with different modes (reduce)
1. Performance of sequential execution based vs OpenMP based vector element sum.
2. Performance of memcpy vs in-place based CUDA based vector element sum.
3. Comparing various launch configs for CUDA based vector element sum (memcpy).
4. Comparing various launch configs for CUDA based vector element sum (in-place).
Sum with in-place strategies of CUDA mode (reduce)
1. Comparing various launch configs for CUDA based vector element sum (in-place).
Data Centers - Striving Within A Narrow Range - Research Report - MCG - May 2...pchutichetpong
M Capital Group (“MCG”) expects to see demand and the changing evolution of supply, facilitated through institutional investment rotation out of offices and into work from home (“WFH”), while the ever-expanding need for data storage as global internet usage expands, with experts predicting 5.3 billion users by 2023. These market factors will be underpinned by technological changes, such as progressing cloud services and edge sites, allowing the industry to see strong expected annual growth of 13% over the next 4 years.
Whilst competitive headwinds remain, represented through the recent second bankruptcy filing of Sungard, which blames “COVID-19 and other macroeconomic trends including delayed customer spending decisions, insourcing and reductions in IT spending, energy inflation and reduction in demand for certain services”, the industry has seen key adjustments, where MCG believes that engineering cost management and technological innovation will be paramount to success.
MCG reports that the more favorable market conditions expected over the next few years, helped by the winding down of pandemic restrictions and a hybrid working environment will be driving market momentum forward. The continuous injection of capital by alternative investment firms, as well as the growing infrastructural investment from cloud service providers and social media companies, whose revenues are expected to grow over 3.6x larger by value in 2026, will likely help propel center provision and innovation. These factors paint a promising picture for the industry players that offset rising input costs and adapt to new technologies.
According to M Capital Group: “Specifically, the long-term cost-saving opportunities available from the rise of remote managing will likely aid value growth for the industry. Through margin optimization and further availability of capital for reinvestment, strong players will maintain their competitive foothold, while weaker players exit the market to balance supply and demand.”
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
1. Evaluating validity of criterion-
referenced test score
interpretations and uses
Takaaki Kumazawa
Kanto Gakuin University
(ktakaaki@kanto-gakuin.ac.jp)
Kintai Bridge, Japan (wiki
2. Purpose
ß The purpose of my talk is to evaluate
validity of criterion-referenced placement
test score interpretations and uses using
Kane’s (2006) argument-based validity
framework
ß This presentation is based on a paper I
published in the JALT Journal
(http://jalt-publications.org/jj/issues/2013-05_35.1)
3. Classical view of validity
ß Validity: the extent to which a test is
supposed to measure
ß Three types of validity
Þ Criterion-related validity
Correlation between a valid measure and a test
developing
Þ Content validity
Experts’ judgment on whether items are measuring
what is supposed to measure
Þ Construct validity
Statistical examination on whether items are measuring
what is supposed to measure
4. Current view of Validity
ß Validity is “the degree to which evidence
and theory support the interpretations of
test scores entailed by proposed uses of
tests” (American Educational Research
Association, American Psychological
Association, & National Council on
Measurement in Education [AERA, APA, &
NCME], 1999, p. 9).
5. Argument-based validity framework
Interpretive argument: proving argument that the inferences are
going to make is theoretically valid
Validity argument: evaluating the interpretive argument by providing
warrant
Observatio
n
Observed
score
Universe
score
Target
score
Use
Scoring generalization extrapolation
decision
6. Interpretive argument
ß Scoring inference
Þ to what extent do examinees get placement items correct
and high-scoring examinees get more placement items
correct
ß Generalization inference
Þ to what extent are placement items consistently sampled
from a domain and sufficient in number so as to reduce the
measurement error
ß Extrapolation inference
Þ to what extent do the difficulty of placement items match to
the objectives of a reading course
ß Decision inference
Þ to what extent do placement decisions made to place
examinees in their proper level of the course have an
impact on washback in the course
7. Participants
Þ 428 Japanese 1st year university students majoring in
law
Þ TOEIC score of about 250-450
Þ Three courses in the English program
Reading
Listening
TOEIC skills
ß Proficiency based program
Þ Three levels
Level 1: 60 high scoring students
Major objective of the reading course: improve their reading skills
such as fast reading
Level 2: about 300 students
Level 3: 50 low scoring students
Major objective of the reading class: re-learn Jr High and High
school grammar
8. Criterion-referenced placement test
ß Grammar (k = 40)
Þ Items are taken from textbooks used in junior and high schools
Þ Grammar: present, past, & future tenses, continuous, relative pronoun,
gerund, participle, etc…
Þ Sample: Hi, I ( ) Ken.
1. am 2. are 3. is 4. be
ß Vocabulary (k = 40)
Þ Items are taken from high frequent 1000-3000 words based on the
JACET 8000 corpus
Þ Sample: Bring
1. 送る (send) 2. 持ってくる (bring) 3. 鳴る (ring) 4. 購入する (buy)
ß Reading (k = 10)
Þ Two passages are taken from two textbooks used in Level 1 and Level
3 reading classes
Þ Sample: How do they travel?
1. by plane 2. by bus 3. by car 4. by train
9. Procedures
ß On the first day of semester, the placement
test was given in 45 minutes
ß A grammar pretest (k = 55, α = .85) was
given on the first day of class in Level 2
classes (n = 51) and Level 3 classes (n = 49)
ß 30 90-minute lessons in two semesters
ß The same grammar posttest (α = .92) was
given on the last day of class to the same
students (n = 51, 49)
ß A course evaluation survey was given to the
same students (n = 51, 49)
13. Backing for decision inference
Level 2 and Level 3 students’ (n = 51, 49) grammar pretest and posttest
scores (k = 55)
11 points down
6 points up
Level 2
students
scored
higher
Level 3
students
scored
higher
Grammarpretest(α=.85) Grammarposttest(α=.92)
ClassLevel n M SD n M SD
Level2a 26 30.38 6.34 21 12.14 2.50
Level2b 25 32.36 8.47 24 28.63 7.93
Level2 51 31.35 7.45 45 20.93 10.24
Level3c 25 20.80 5.09 22 26.82 5.21
Level3d 24 19.88 4.29 23 26.78 5.95
Level3 49 20.35 4.69 45 26.80 5.53
14. Validity argument
Interpretive argument
ß Scoring inference
Þ to what extent do examinees get
placement items correct and high-
scoring examinees get more
placement items correct
ß Generalization inference
Þ to what extent are placement
items consistently sampled from a
domain and sufficient in number
so as to reduce the measurement
error
ß Extrapolation inference
Þ to what extent do the difficulty of
placement items match to the
objectives of a reading course
ß Decision inference
Þ to what extent do placement
decisions made to place
examinees in their proper level of
the course have an impact on
washback in the course
Validity argument
ß Scoring inference
Þ Because most items were working well,
the inference from observation to the
observed score was valid
ß Generalization inference
Þ Because of high dependability with the
small amount of measurement error, the
inference from the observed score to
universe score was valid
ß Extrapolation inference
Þ Because the difficulty of the items were
adequate to the objectives of the program,
the inference from the universe score to
target score was valid
ß Decision inference
Þ Because Level 3 students were placed in
the right level and were able to improve
their grammar test scores, the inference
from the target score to test use was valid.
15. Conclusion
ß “Validation is simple in principle, but
difficult in practice. The argument-based
framework provides a relatively pragmatic
approach to validation” (Kane, 2012, p. 15).
William Jolly Bridge, Brisbane
16. References
ß Kane, M. (2006). Validation. In R. Brennan
(Ed.), Educational measurement (4th ed.). (pp.
17-64). Westport, CT: Greenwood Publishing.
ß Kane, M. (2012). Validating score
interpretations and uses. Language Testing,
29, 3-17. doi: 10.1177/0265532211417210
ß Kumazawa, T. (2013). Evaluating validity for
in-house placement test score interpretations
and uses. JALT Journal, 35, 73-100.