Successfully reported this slideshow.
Your SlideShare is downloading. ×

Quality Evaluation of the Higher Education Programmes

Ad
Ad
Ad
Ad
Ad
Ad
Ad
Ad
Ad
Ad
Ad
Upcoming SlideShare
UC Dual Pilot Study
UC Dual Pilot Study
Loading in …3
×

Check these out next

1 of 3 Ad

Quality Evaluation of the Higher Education Programmes

Download to read offline

Quality Evaluation of the Higher Education Programmes

David Bañeres, Laura Porta, Teresa Romeu, Montse Serra

1st International Workshop on Technology-Enhanced Assessment, Analytics and Feedback (TEAAF2014)

Quality Evaluation of the Higher Education Programmes

David Bañeres, Laura Porta, Teresa Romeu, Montse Serra

1st International Workshop on Technology-Enhanced Assessment, Analytics and Feedback (TEAAF2014)

Advertisement
Advertisement

More Related Content

Slideshows for you (20)

Advertisement

Similar to Quality Evaluation of the Higher Education Programmes (20)

More from César Pablo Córcoles Briongos (11)

Advertisement

Recently uploaded (20)

Quality Evaluation of the Higher Education Programmes

  1. 1. Quality Evaluation of the Higher Education Programmes David Bañeres, Laura Porta, Teresa Romeu, Montse Serra IT, Multimedia and Telecomunications Department Universitat Oberta de Catalunya The universities in the European Higher Education Area (EHEA) must fulfil several directives in order to be able to provide a high education degree. A degree is evaluated in all its life cycle: from its design, when it is approved, during its monitoring, until its verification. Thus, each country has a quality agency that articulates all these processes of evaluation following the policies recommended by the European Association for Quality Assurance in Higher Education (ENQA) in order to guarantee the quality required for a specific degree [1]. Related to the monitoring process of a degree, an annual report evaluating several indicators has to be delivered to the agency. More precisely, the report describes information like statistics about the enrolment, the student performance, the satisfaction results, or the type of learning resources. Moreover, there is also a summary of the year composed of: detected issues, a set of enhancement proposals and planned corrective measures. Subsequently, this report is analyzed by the agency and a report's feedback is returned to the university together with the issues that should be corrected and improved during the next year. Based on this annual report, the monitoring process consists of the steps described below:  The degree's coordinator analyzes the remarks performed by the agency in the previous report in order to make the suitable adjustments within the degree.  The issues related to a subject that should be corrected are delivered to the teacher in charge of the subject. The issues are shared within teaching staff of the same knowledge area in order to find solutions and implement the corresponding corrections and improvements.  The teachers collect evidences throughout the course in order to point out the weakness and the strength of it for the next annual report.  At the end of the course, the evidences and the analysis of the course are delivered to the coordinator's degree and teaching staff of the same knowledge area.  The degree's coordinator collects all the evidences and, simultaneously, he analyzes and synthesizes the results of the evaluation in her annual report in a continuous process of improvement. As we can observe in the previous description, the monitoring process involves the degree's coordinator and all the teaching staff of the degree. This fact tends to be time- consuming since all the teachers should collaboratively work to meet the quality standards. Currently, in our university this collaborative work is manually performed, that is, the information is transferred by arranging a meeting between the coordinator and the teachers. Additionally, the information concluded from this meeting is delivered
  2. 2. to the coordinator with unformatted documents increasing the analyzing time. This methodology is not efficient and is even more impractical within a degree with a large teaching staff. The present proposal focuses on the evaluation of the quality of a concrete degree by using techniques commonly used on e-assessment [2], since some of these techniques of assessment used to evaluate the students can be extrapolated to other areas. Specifically, we propose to use rubric-based assessment to evaluate the quality of a degree. The rubric helps to self-evaluate a specific subject by measuring some indicators. Indicators with regard to the learning resources, tools, evaluation, satisfaction, quality of the teaching plan, and even the performance of the teaching staff are evaluated. Notice that, this rubric approach guarantees the same criteria for each subject. The gradation of the quality ranges from deficient, acceptable to excellent. Moreover, each indicator also accepts information in natural language describing issues and enhancements tasks scheduled for the next semester. In order to improve the efficiency of the evaluation process, the technique has been implemented within a tool called AVALA. The tool collects all the evidences related to the operative of a concrete degree in the course of a semester. Furthermore, it has been designed as a collaborative framework where all the teaching staff and the coordinator can share all the information related to quality assurance. The tool has a twofold objective. In one hand, as we described, it helps a teacher to collect and analyze all the evidences related to the performance of the teaching process during a semester. At the same time, the teacher can analyze the progress of these evidences from the current semester in comparison with the previous ones. On the other hand, the coordinator degree can consult all the collected evidences on a single environment, it means, the coordinator has a view of the underlying information of the current semester and at a glance she knows what issues need to be addressed. Additionally, the tool has other features to help in some steps of the monitoring process:  The comments of the quality agency are stored by subjects. This feature helps the instructors to automatically know the issues that should be corrected during the current semester.  The system synthesizes the subjects’ evaluations for delivering automatically reports about the critical issues of the degree. This information will help the coordinator to perform the annual reports. The AVALA tool has been used in the IT, Telecommunication and Multimedia department of the Universitat Oberta de Catalunya during the courses 2012-13 and 2013-14 (divided in two semesters both courses). Three degrees and six masters have been managed using the tool and around sixty professors have been involved in the process of evaluation. The experimental results have been satisfactory. The coordinators agree that the tool is useful to review and centralize the information previously cited. Besides, the critical problems can be easily detected. The system misses a control panel with alerts such as subjects with low performance or satisfaction. Indeed, professors agree on the improvement of the efficiency related to the transfer of the information and the
  3. 3. structured way to store the data. However, they complain about the time invested in the evaluation of all the subjects. The tool has some deficiencies that should be corrected in the immediate future. Some users' demands will be taken into account in order to improve the framework. Nevertheless, the main goals of the tool have been achieved reducing the time invested in the guidance's quality of the degree. As a final conclusion, AVALA tool provides an added value, not only with regard to the efficiency of the teacher’s work, but also when establishing a better way to manage all the monitoring processes that belong to a degree contributing the quality’s guarantee. REFERENCES [1] Standards and Guidelines for Quality Assurance in the European Higher Education Area. http://www.enqa.eu/wp-content/uploads/2013/06/ESG_3edition-2.pdf. Last accessed: 21/10/2014 [2] Crisp, G. T. (2012). Integrative assessment: reframing assessment practice for current and future learning. Assessment & Evaluation in Higher Education, 37(1), 33- 43.

×