SlideShare a Scribd company logo
1 of 14
Download to read offline
1
External Evaluation of the Peer
Evaluation Process undertaken by
the University of Greenwich and
Linking London Credit
Accumulation and Transfer Scheme
Pilot Projects
Viki Faulkner
July 2013
2
Table of Contents
1. Introduction .................................................................................................................... 3
2. Overview of the Peer Evaluation Process.......................................................................... 3
3. Strengths of the Peer Evaluation Process.......................................................................... 4
4. Challenges Within the Peer Evaluation Process................................................................. 5
5. Additional Monitoring and Evaluation Activities ............................................................... 6
6. Learning Points................................................................................................................ 6
7. Best Practice and Recommendations................................................................................ 8
8. Conclusions ..................................................................................................................... 9
Appendix 1: CATS Pilot Project Peer Evaluation Process........................................................... 10
3
External Evaluation of the Peer Evaluation Process undertaken by the University
of Greenwich and Linking London Credit Accumulation and Transfer Scheme Pilot
Projects
1. Introduction
The University of Greenwich and Linking London each led one of the six, Association of Colleges
(AoC) Department for Business Innovation and Skills (BIS) and Skills Funding Agency (SFA) funded
credit accumulation and transfer scheme (CATS) pilot projects. The projects arose as a response to
‘New Challenges New Chances’1
, which set out the Government vision to reform the further
education and skills system. This recognised that further and higher education institutions would
need to find new ways of working together, providing clear and transparent progression routes that
would support workforce development and help learners to reach their full potential.
The University of Greenwich project, ‘Building on Competence in the Workplace’ focused on work
between the university and five members of its partner college network. The Linking London ‘CATS
Project in Management’ focused on work between two universities, Birkbeck and Middlesex and
four further education colleges. Both projects also had a strong level of engagement from
professional bodies and provided mechanisms to capture and integrate the input from learners and
employer representatives. The projects both commenced on 1 July 2012 and ran for one year,
concluding 30 June 2013.
Both projects have undertaken extensive monitoring and evaluation activities and these are the
subject of individual submissions to AoC. Additionally, they have taken part in the overarching
evaluation of the CATS projects undertaken by AoC. This report, commissioned by the University of
Greenwich, does not therefore cover ground already well reported, but instead focuses on providing
an external evaluation of the effectiveness and impact of an added quality enhancement aspect: the
peer evaluation process.
2. Overview of the Peer Evaluation Process
The peer evaluation process was identified at the time of project submission as an activity that both
teams felt would add value and it was built within the project plans from the start. The model
selected was based on one previously developed by the Kent and Medway Lifelong Learning
Network2
and had already been trialled by the University of Greenwich to measure the impact of
their LLN funded projects.
The peer evaluation process consisted to two elements: an impact evaluation model – to ensure
consistency and provide a common framework, and a monitoring and evaluation plan. It was
1
New Challenges, New Chances: Further Education and Skills Reform Plan: Building a World-Class Skills System
(BIS, 1 December 2011)
2
Perrin, L Time to Make a Difference: achieving sustainability through projects (2010) KMLLN
4
described in detail in a paper that was attached to the University of Greenwich CATS project interim
report (Feb 2013) and this is attached as appendix 1 to minimise repetition.
The University of Greenwich and Linking London were mutually selecting peer evaluation ‘buddies’.
They were logical partners in the process as both had previous experience of peer evaluation and
were partnerships with similar roots – each born out of the HEFCE funded lifelong learning networks
initiative. This proved to be a very effective factor in ensuring common purpose and a significant
level of shared understanding.
3. Strengths of the Peer Evaluation Process
The Impact Evaluation Model adopted as a mechanism to guide the peer evaluation process, set a
clear focus and structure for the activity. This was a real strength of the process that was adopted,
ensuring that peer evaluation was a constructive and productive activity and did not become an
unfocused ‘talking shop’. The project manager of the Linking London project, who was introduced to
the model for the first time through this activity, felt it was an example of good practice and found it
very useful in helping to shape her initial thinking around evidence gathering.
The project teams worked together at an early stage to identify inputs, outputs, intermediate
outcomes and final outcomes for each of their projects. They then used this information to populate
the model, identifying evidence needs and evidence sources. This meant that from a very early
phase, they were already thinking about what impact they wanted their project to achieve – and not
just focusing on what they were planning to do in the short term. The model stresses that final
outcomes are things which will extend beyond the lifespan of a one year funded project and so
consideration needs to be given at the stage to how longer term outcomes can be captured and
could be evidenced. It is a strength of the model that it forces a mindset shift into a consideration of
longer term behavioural changes rather than a focus on delivery of short term targets.
Both partners clearly articulated the benefits that they feel they gained from having a ‘critical friend’
that helped them to keep focused and added another layer under the formal reporting structure and
network set up by the AoC. Although both projects had numerous contacts with other CATS projects,
they each felt that the relationship that they established with their peer evaluation partner was
much stronger and more constructive. Comments from members of the project teams included
statements such as, ”it was nice sometimes to not feel you are alone”, “ we were more prepared to
share issues and problems, it was more constructive, helpful, not just about reporting on progress”,
“the peer evaluation was more process based – it was useful”. These comments demonstrate that
they had been able to establish an open trusting relationship with each other and this is when peer
support can be most effective. One team member described how much they valued the fact that
discussions with the peer evaluation partner did not focus only on the achievement of project aims
and objectives, but how they were also able to “talk around the edges of the project”. This was a
wonderful phrase and was illustrated by several examples where project team members described
how they had shared experiences of working with partners – and strategies for managing the
challenges this presents and working within the complexities of a steering group ‘management’
structure.
The use of Dropbox has provided an innovative tool that enhanced communication between
multiple partners. Project teams at Linking London and the University of Greenwich have all been
5
able to access Dropbox and deposit up to date versions of documentation there. It is a cloud based
storage facility which means it is accessible wherever you are and you therefore always have access
to the most up to date versions of any document – no more relying on someone to have sent you an
email or risking files ending up blocked by over cautious spam filters. The ease of access is a real
advantage and has also enabled a very effective mechanism for on-going external evaluation. The
external evaluator also had access to Dropbox and so was therefore able to regularly ‘dip in’ to
sample the evidence available – effectively moving external evaluation from being a twice in a
project snapshot activity, into a continuous monitoring process which proves much more rigorous as
an exercise.
A further strength of the Dropbox facility was the way in which the folders had been set up. It was
structured from the start with four folders that aligned with the elements of the Impact Evaluation
Model: inputs, outputs, intermediate outcomes and final outcomes. A great benefit of this structure
was that, from the very beginning, this meant that the project teams had to engage critically with
every piece of evidence they collected; having to differentiate clearly between inputs, outputs and
outcomes.
4. Challenges within the Peer Evaluation Process
Although both projects focused on credit accumulation and transfer schemes, there were substantial
differences between the two with the projects having very different outputs. At times this made it
difficult for each partner to see where lessons could be directly transferrable. This was quite
noticeable at the interim project phase where the University of Greenwich had been able to provide
clear support in providing the impact evaluation model and sharing their experience of setting up
clear reporting schedules and monitoring cycles for managing projects with partners, yet they were
less able, at that stage, to articulate tangible benefits or transferrable learning which they had got
out of the peer evaluation process. However, the value of the activity as a whole was clear and both
project teams valued the reflective process which they said kept them focused and helped them
each to clarify their own thoughts about their project.
The peer evaluation process was not a requirement of the original AoC project call. It was built into
the proposal of both projects because the teams recognised the importance of assessing impact to
justify value for money and it was felt that this would be an effective and efficient way to achieve it.
According to Appendix 1 of the University of Greenwich Interim Report (Feb 2013), ‘It was agreed
that the process should be simple, clear and sufficiently robust but not so onerous as to divert time
and resources away from other important project activities.’ In this case, the peer evaluation process
has achieved its objectives. The process was clear and simple, it had a strong focus on impact,
reinforced by the methodology, and the Impact Evaluation Model adopted provided a robust
mechanism for planning and capturing evidence on which to base sound judgements. The peer
evaluation also provided a cost effective mechanism which did minimise funding drain. However this
was, to a not insubstantial degree, mainly down to a smart selection of partnering organisations.
Linking London and the University of Greenwich are geographically close and this means that the
four face to face meetings planned within the monitoring and evaluation plan were quite achievable.
As mentioned earlier, Linking London and the University of Greenwich project team have previous
experience of each others’ work as they shared common features and a common language, working
closely together as lifelong learning networks. This enabled the project teams to move quickly past
6
the ‘getting to know you’ phase and move directly into the ‘getting down to business’ phase. Where
project teams do not already share a common bond, establishing a mutually supportive, trusting
relationship may take a little longer than one meeting. Where partners are geographically quite
distant also, this could put even more strain on developing a relationship – thereby requiring a
greater investment of time to get it established. In such a case, the peer evaluation process may
quickly begin to look less cost effective and could easily begin to look like an extra layer of
monitoring and evaluation, also beginning to divert resources from delivery of core activities. It is
therefore suggested that the smart selection of a peer evaluation partner is one of the critical
success factors for this aspect of activity and further adoption of this as a process should strongly
consider providing clear guidance on partner selection criteria.
Dropbox, a cloud-based document repository, provided a very effective mechanism for multiple
partners to share documentation and to organise evidence collection. Its use was one of the
distinctive features of the peer evaluation model but unfortunately it was not in place from the start
of the project. This was disappointing and it would have been beneficial for it to be available from
the beginning of the project. With hindsight, the team at the University of Greenwich who were
responsible for setting up the Dropbox facility agree with this and feel that in future, it would be one
of the first things put in place as part of a peer evaluation information strategy.
Timescales provided a further area of challenge. Although both projects were working towards an
overarching timetable set by the AoC, within that there were minor variations in each project, such
as the dates set for their final dissemination events. This did not cause any major challenges but an
early deadline for the Linking London dissemination event meant that the final peer evaluation
meeting had to be held before the University of Greenwich had been able to complete all of its final
evidence collection. Hence evaluation reports were written before case studies were collected and
so could not be included in the evidence base available.
5. Additional Monitoring and Evaluation Activities within the University of Greenwich Project
The Peer evaluation process within the University of Greenwich project was supplemented with
monitoring and evaluation activities undertaken by a specially devised steering group. This group
comprised senior managers from the colleges, the project team and employers representing SMEs
and large companies in SE London. It was an appropriately constructed group that enabled effective
oversight of activities and a useful forum for raising areas of concern.
Steering group meetings were hosted by partners in turn. This built the collective awareness of
others’ campuses and facilities and was an effective way in which the project was able to strengthen
relationships between partners. Steering group meetings not only provided a good vehicle for
monitoring project progress but also a forum for staff development activities, enabling all partners
to gain a shared understanding of credit and accreditation of prior learning. Greenwich has identified
these as ‘active meetings’ – ones that blend together a business agenda with a development session,
encouraging interaction and participation of all members.
6. Learning Points
Both projects agreed, as part of the peer evaluation process, that it was important to maintain a
flexible approach towards project aims and outcomes and to recognise that plans will need to be
7
adapted as a project progresses. Beginning in July as both projects did, they shared a common
concern in the early phase of trying to make steady progress at a time when most academic
institutions seem to be run by a skeleton staffing model. The University of Greenwich project was in
a more advantageous position here at it was set out with a tightly focused proposal based on a pre-
existing relationship. It was also well timed as it built on the colleges’ validation to offer the Applied
Professional Studies programme. The tight bid with a clear focus certainly helped the University of
Greenwich to be able to achieve its aims and objectives in a timely fashion. The Linking London
project found the summer break more problematic as they had to establish a number of sub-
projects and, in a number of cases, build new relationships between colleges and universities. Both
projects had to respond flexibly and adapt to staff changes in what seems to have been an especially
turbulent time for FE college partners. The Project Manager of the Linking London project found the
first meeting of the peer evaluation group especially useful, not only helping her to situate their
project, but also helping her to realise that they were not alone in the situation in which they found
themselves with regard to concerns over timescales. As a project manager new to this type of
working, the peer evaluation process was especially beneficial, enabling her to work closely with the
experienced project manager on the University of Greenwich project.
Although both projects had different challenges and very different outcomes, as well as taking on
the role of ‘critical friend’ they were able to provide mutual support and this is an aspect that is not
to be under-estimated. Although there were a number of meetings of all CATS pilots, facilitated by
AoC, both projects described the relationship that formed between them as “deeper” and “richer”
than that they had with the other projects. Neither project was afraid to ask their peer evaluation
partner for advice and support and neither felt they were being judged in any way. There are some
indications here that relationships which foster outside of the formal oversight of funding agencies
can enable a greater level of openness and are therefore more likely to be mutually beneficial.
The concept of ‘active meetings’ at the University of Greenwich, that blended together a business
agenda for the steering group, with a development session was one which appeared to enhance the
level of engagement that partners had with the project collectively. All college partners in their
interim reports commented on the “closer working relationships”, the “group collaborations” and
“stronger relationships” that had been built in this way and cited it as a real strength of the project.
This was an aspect that was picked up on through the peer evaluation process by Linking London and
is a methodology that they have already identified as one worth trialling as a future way of working.
The University of Greenwich noted that Linking London were very good at disseminating the findings
of their project work and took good advantage of every opportunity available to do so, based on a
number of well established practices they had already put in place. Greenwich felt that most of their
dissemination work had been internally focused, ensuring that they maximised the spread of
information amongst staff groups within the partner organisations. As a result of the peer evaluation
process they are now looking at putting on a number of road shows – adopting a model used by
Linking London for wider dissemination, and creating a toolkit as additional outputs to those
identified in the original project plan.
8
7. Best Practice Identified and Recommendations
Consistency: The peer evaluation process was based on both projects using the Impact Evaluation
Model that had been created originally by the University of Greenwich. Both projects worked
together to identify their inputs, outputs and outcomes and used this information to populate the
model. This exercise provided the projects with a shared language and a common level of
understanding. Although the outputs and outcomes varied between the two projects, and so by
necessity the evidence needs and evidence sources they would require, they were each clear about
the process that they would be undertaking. This gave a commonality that enabled them to evaluate
each others’ project against identified negotiated areas. Without the use of a common model the
differences between the projects may have begun to be too distracting making it hard to bring any
aspects of consistency to the peer evaluation process. It is therefore recommended that, for a peer
evaluation to be effective it is important to adopt a consistent model against which both projects
can be measured.
Clear focus on impact: The peer evaluation process had a clear focus on impact and this was led by
the use of the model adopted. Impact measures, if they are to be effective, need to be embedded
into a project from the outset. The University of Greenwich had clearly considered the long term
impact of their project and had built within the evaluation model final outcomes that stretched way
beyond the one year funded project. This is important. With a project such as the CATS pilot, within
the space of a year we can truly only realistically measure project outputs, such as the publishing of
the prospectus and some intermediate outcomes such as focus group feedback. The true success of
the project will only be revealed through increasing progression rates and this will take a number of
years before we can see a clear picture. It is recommended that funded projects should not feel
constrained to set evaluation measures which are limited only to the timeframe of the funding
attached – but should also consider longer term impact measures and look at how this may be
captured in the future.
Dropbox: Digital, cloud-based storage facilities can provide a cost efficient and very effective
mechanism for sharing large amounts of documentation amongst multiple users who may or may
not share a common employing organisation and are likely to based in multiple locations. In this
case, Dropbox was used very effectively to improve ease of access to documentation and improve
lines of communication between the project teams and the External Evaluator. It is recommended
that any multi-partner, complex project seriously consider using digital cloud based document
storage as an integral part of an effective communication strategy.
Cost effective: The peer evaluation process in this project provided a well integrated and cost
effective mechanism for monitoring and evaluation. To ensure robustness of the project oversight,
this was also supplemented (at the University of Greenwich) by a rigorous system which included the
regular submission and review of partner progress reports, monitoring of progress at steering group
meetings and a formal reporting structure set in place by the AoC. This is also supplemented by an
overarching evaluation of the CATS projects commissioned by AoC. Where extensive formal
monitoring and evaluation processes are already established, it is recommended that projects
consider peer evaluation as a useful quality enhancement mechanism.
9
8. Conclusions
The peer evaluation process is a real strength of this project. It has been an integral component from
project conception through to completion and has provided a genuine added value to both projects
whilst placing very limited demands on project resources. Subject to the caveats raised above
regarding partner selection and cost effectiveness, it is a model worthy of consideration for adoption
in future national project initiatives.
The University of Greenwich Impact Evaluation Model selected to form the basis for the peer
evaluation activity is now one that has been tested in a range of situations. It has been used
effectively by both CATS projects as well as the smaller scale LLN projects, and has demonstrated its
transferability and flexibility. It has a strong emphasis on impact – which is critically important within
any project, and ensures that longer term outcomes that stretch beyond the funded term of the
project are built in from the start. Both of these aspects build sustainability and maximise the long
term impact gained from pump prime funding. The model is worthy of further dissemination on a
wider stage.
10
Appendix – Building on Competence in the Workplace
Interim Report 1 February 2013 University of Greenwich
CATS Pilot Project Peer Evaluation Process
Background
The University of Greenwich and Linking London CATS Pilot Project teams recognise the
importance of assessing impact to justify value for money, to determine the direction and
consequences of projects and to show outcomes in achievements, products and processes.
They have therefore developed and agreed a Peer Evaluation Process by which both project
teams will monitor the process and outcomes of the other’s project through a shared ‘virtual’
portfolio of evidence and produce an evaluative commentary to accompany the Project Final
Report.
It was agreed that the process should be simple, clear and sufficiently robust but not so
onerous as to divert time and resources away from other important project activities.
The Peer Evaluation Process
The Peer Evaluation Process consists of 2 elements:
1. An Impact Evaluation model which ensures consistency and provides a common
framework to enable each project team to monitor each others’ progress against their
work plan
2. A monitoring and evaluation plan which underpins the IEM by confirming the process
by which each project team will monitor and evaluate
Impact Evaluation Model (IEM)
The Impact Evaluation Model is based on a model previously developed by the Kent and
Medway Lifelong Learning Network3
, led by the University of Greenwich to help its partner
institutions measure the impact of their LLN-funded projects. In turn based on a model
developed by the Training and Development Agency for Schools (TDA)4
it did not claim to
offer a brand new approach to evaluation, but attempted to provide an accessible and user-
friendly framework to assist their activity.
The fact that it brings together qualitative and quantitative evidence and allows us to look not
only at data but also the real people involved and the impact the project could have on
communities (work-based learners, employers and learning providers) makes it particularly
applicable to the CATS project. The model is designed to help the project team build up a
picture of how they expect the project to work and by being clear about the outcomes of the
project.
Initially, the project teams worked together to identify the inputs, outputs, intermediate
3
Perrin, L. Time to Make a Difference: achieving sustainability through projects (2010) KMLLN
4
The impact evaluation model (IEM) (2009) Training and Development Agency for Schools
11
outcomes and final outcomes that their project was trying to achieve and to populate the
model with these. This enabled them to articulate how they expected their project to have
an impact. The model was then used to identify the evidence needed and from there, to
identify the potential sources of evidence. (See page 4 for Linking London and UoG IEMs)
By working through the model each project team was able to demonstrate the links between
the various stages of project delivery, from planning all the way through to the impact on
individual users of the activities and products being delivered and on the overall project
aims.
From past experience we were aware that the model is most effective when put in place
early in the project. By completing the model at this stage the project teams knew what data
and evidence we needed to collect at the outset and were able to clarify for each other any
uncertainties about the aims of our projects. Articulating the underlying assumptions of the
projects and identifying the potential evidence sources at an early stage have informed the
project work plans and are being explored as delivery progresses.
Monitoring and evaluation plan
The following methods are being used to monitor the progress of each others’ project:
Method Purpose Frequency
Face-to-Face
Meetings
1. To update and question
each other on progress
2. To agree strengths and
weaknesses
3. To share knowledge and
best practice
4. To discuss issues and
propose solutions
Beginning (Sep 12)
Mid-term (Jan 13) and
End (May 13)
Dropbox 1. To store and share
records, documents and
other evidence and make
available to each other
Ongoing – as records
and documents are
produced
Emails 1. To update each other on
events, dropbox updates,
developments
2. To share ideas and deal
with issues arising
between meetings
As and when required
between meetings
Evaluative
Commentary
1. To provide an evaluative
commentary to inform final
report
Once – end of project
12
Face to Face Meetings
The principle method to support the monitoring and evaluation process, three meetings of
Project Directors and Managers will be held over the course of the project as follows:
 27 September 2012 (Birkbeck) – to agree impact evaluation model and agree process
 14 January 2013 (UoG) – to report progress to date and agree outcomes for draft interim
report
 17 May 2013 (Birkbeck) – to report progress to date and agree outcomes for final report
Dropbox
Dropbox will be used as a virtual repository for all documents, publications, records and
other evidence identified on the Impact Evaluation Model. Each project will have its own
folder, each containing four sub folders - Inputs, Outputs, Intermediate Outcomes and Final
Outcomes (as per IEM). Evidence will be filed in the appropriate sub-folder.
Folders and sub-folders will be shared and accessible to each other. In addition, the UoG
folder will be accessible to the Viki Faulkner of the Sussex Lifelong Learning Network who
has been appointed to undertake an external evaluation of the Peer Evaluation Process.
Emails
Probably the easiest and most effective way of keeping in touch between meetings, these
will be used to maintain contact on an ongoing basis.
Evaluative Commentary
Each project team will provide a brief evaluative commentary (approx. 500 words) on the
others’ project to support the final report and confirming:
1. To what extent did the project meet its aims and objectives?
2. What, if any, are the additional benefits delivered by the project?
3. What lessons did we learn that we could apply to future projects?
CATS Pilot Project Teams
Linking London
Project Director: Sue Betts
Project Manager: Pamela Calabro
University of Greenwich
Project Director: Hugh Joslin
Project Manager: Lindsay Perrin
13
Impact Evaluation Models – Linking London and University of Greenwich
14

More Related Content

What's hot

"Hybrid" KT & Project Planning Template
"Hybrid" KT & Project Planning Template "Hybrid" KT & Project Planning Template
"Hybrid" KT & Project Planning Template KBHN KT
 
Techniques and measurement for engaging your audience
Techniques and measurement for engaging your audienceTechniques and measurement for engaging your audience
Techniques and measurement for engaging your audienceJ'ette Novakovich
 
"Hybrid" KT & Project Planning GUIDE
"Hybrid" KT & Project Planning GUIDE"Hybrid" KT & Project Planning GUIDE
"Hybrid" KT & Project Planning GUIDEKBHN KT
 
Rethinking Lessons Learned in the PMBoK Process Groups: A Model based on Peop...
Rethinking Lessons Learned in the PMBoK Process Groups: A Model based on Peop...Rethinking Lessons Learned in the PMBoK Process Groups: A Model based on Peop...
Rethinking Lessons Learned in the PMBoK Process Groups: A Model based on Peop...Marcirio Chaves
 
ESOQ_final results
ESOQ_final resultsESOQ_final results
ESOQ_final resultsDiana-Adela
 
11.the association between project success and project initiation phase
11.the association between project success and project initiation phase11.the association between project success and project initiation phase
11.the association between project success and project initiation phaseAlexander Decker
 
8.[60 68]the association between project success and project initiation phase
8.[60 68]the association between project success and project initiation phase8.[60 68]the association between project success and project initiation phase
8.[60 68]the association between project success and project initiation phaseAlexander Decker
 
synthesis report_.strengthening_collaboration_for_co-operative_research.cca_c...
synthesis report_.strengthening_collaboration_for_co-operative_research.cca_c...synthesis report_.strengthening_collaboration_for_co-operative_research.cca_c...
synthesis report_.strengthening_collaboration_for_co-operative_research.cca_c...Ahmad Maruf
 

What's hot (9)

"Hybrid" KT & Project Planning Template
"Hybrid" KT & Project Planning Template "Hybrid" KT & Project Planning Template
"Hybrid" KT & Project Planning Template
 
B08 B4pc 142 Diapo Miedes En
B08 B4pc 142 Diapo Miedes EnB08 B4pc 142 Diapo Miedes En
B08 B4pc 142 Diapo Miedes En
 
Techniques and measurement for engaging your audience
Techniques and measurement for engaging your audienceTechniques and measurement for engaging your audience
Techniques and measurement for engaging your audience
 
"Hybrid" KT & Project Planning GUIDE
"Hybrid" KT & Project Planning GUIDE"Hybrid" KT & Project Planning GUIDE
"Hybrid" KT & Project Planning GUIDE
 
Rethinking Lessons Learned in the PMBoK Process Groups: A Model based on Peop...
Rethinking Lessons Learned in the PMBoK Process Groups: A Model based on Peop...Rethinking Lessons Learned in the PMBoK Process Groups: A Model based on Peop...
Rethinking Lessons Learned in the PMBoK Process Groups: A Model based on Peop...
 
ESOQ_final results
ESOQ_final resultsESOQ_final results
ESOQ_final results
 
11.the association between project success and project initiation phase
11.the association between project success and project initiation phase11.the association between project success and project initiation phase
11.the association between project success and project initiation phase
 
8.[60 68]the association between project success and project initiation phase
8.[60 68]the association between project success and project initiation phase8.[60 68]the association between project success and project initiation phase
8.[60 68]the association between project success and project initiation phase
 
synthesis report_.strengthening_collaboration_for_co-operative_research.cca_c...
synthesis report_.strengthening_collaboration_for_co-operative_research.cca_c...synthesis report_.strengthening_collaboration_for_co-operative_research.cca_c...
synthesis report_.strengthening_collaboration_for_co-operative_research.cca_c...
 

Viewers also liked

Greenwich Peninsula – Where riverside living starts
Greenwich Peninsula – Where riverside living startsGreenwich Peninsula – Where riverside living starts
Greenwich Peninsula – Where riverside living startsNewsworks
 
Eric Girod Project Resume
Eric Girod Project ResumeEric Girod Project Resume
Eric Girod Project ResumeEric Girod
 
Slide Share 1
Slide Share 1Slide Share 1
Slide Share 1mmaksymiak
 
MarkSFong_Resume-Nov2016
MarkSFong_Resume-Nov2016MarkSFong_Resume-Nov2016
MarkSFong_Resume-Nov2016Mark Steve Fong
 
Resume-Master Educator
Resume-Master EducatorResume-Master Educator
Resume-Master EducatorRoger McLain
 
Teferi_Seyoum_Resume
Teferi_Seyoum_ResumeTeferi_Seyoum_Resume
Teferi_Seyoum_ResumeTeferi Seyoum
 
CV(new) site-civil engineer
CV(new) site-civil engineerCV(new) site-civil engineer
CV(new) site-civil engineerManolis Taliadouros
 
Alex Imholz Resume (July 5, 2015)
Alex Imholz Resume (July 5, 2015)Alex Imholz Resume (July 5, 2015)
Alex Imholz Resume (July 5, 2015)alex imholz
 
Senior Voice engineer
Senior Voice engineerSenior Voice engineer
Senior Voice engineerJustin Cho
 
Online student management system
Online student management systemOnline student management system
Online student management systemMumbai Academisc
 
Harpreet Singh Resume
Harpreet Singh ResumeHarpreet Singh Resume
Harpreet Singh Resumeharpreet singh
 
2014-11-28b BAT_TAFOYA_Bruce_Updated_Resume
2014-11-28b BAT_TAFOYA_Bruce_Updated_Resume2014-11-28b BAT_TAFOYA_Bruce_Updated_Resume
2014-11-28b BAT_TAFOYA_Bruce_Updated_ResumeBruce Tafoya
 
English syllabus 2011
English syllabus 2011English syllabus 2011
English syllabus 2011Nila Shah
 
Oracle R12 SCM Functional Interview Questions - Order Management,
Oracle R12 SCM Functional Interview Questions - Order Management, Oracle R12 SCM Functional Interview Questions - Order Management,
Oracle R12 SCM Functional Interview Questions - Order Management, Boopathy CS
 

Viewers also liked (19)

CV_ADJeng2015may_prof2page
CV_ADJeng2015may_prof2pageCV_ADJeng2015may_prof2page
CV_ADJeng2015may_prof2page
 
Greenwich Peninsula – Where riverside living starts
Greenwich Peninsula – Where riverside living startsGreenwich Peninsula – Where riverside living starts
Greenwich Peninsula – Where riverside living starts
 
Eric Girod Project Resume
Eric Girod Project ResumeEric Girod Project Resume
Eric Girod Project Resume
 
Slide Share 1
Slide Share 1Slide Share 1
Slide Share 1
 
MarkSFong_Resume-Nov2016
MarkSFong_Resume-Nov2016MarkSFong_Resume-Nov2016
MarkSFong_Resume-Nov2016
 
Resume-Master Educator
Resume-Master EducatorResume-Master Educator
Resume-Master Educator
 
resume 2017
resume 2017resume 2017
resume 2017
 
Teferi_Seyoum_Resume
Teferi_Seyoum_ResumeTeferi_Seyoum_Resume
Teferi_Seyoum_Resume
 
YOLANDA AGUILAR
YOLANDA AGUILARYOLANDA AGUILAR
YOLANDA AGUILAR
 
Objective
ObjectiveObjective
Objective
 
CV(new) site-civil engineer
CV(new) site-civil engineerCV(new) site-civil engineer
CV(new) site-civil engineer
 
Alex Imholz Resume (July 5, 2015)
Alex Imholz Resume (July 5, 2015)Alex Imholz Resume (July 5, 2015)
Alex Imholz Resume (July 5, 2015)
 
Senior Voice engineer
Senior Voice engineerSenior Voice engineer
Senior Voice engineer
 
cvADJmay2015
cvADJmay2015cvADJmay2015
cvADJmay2015
 
Online student management system
Online student management systemOnline student management system
Online student management system
 
Harpreet Singh Resume
Harpreet Singh ResumeHarpreet Singh Resume
Harpreet Singh Resume
 
2014-11-28b BAT_TAFOYA_Bruce_Updated_Resume
2014-11-28b BAT_TAFOYA_Bruce_Updated_Resume2014-11-28b BAT_TAFOYA_Bruce_Updated_Resume
2014-11-28b BAT_TAFOYA_Bruce_Updated_Resume
 
English syllabus 2011
English syllabus 2011English syllabus 2011
English syllabus 2011
 
Oracle R12 SCM Functional Interview Questions - Order Management,
Oracle R12 SCM Functional Interview Questions - Order Management, Oracle R12 SCM Functional Interview Questions - Order Management,
Oracle R12 SCM Functional Interview Questions - Order Management,
 

Similar to External evaluation of the peer evaluation process final1

A Step By Step Guide To Monitoring And Evaluation
A Step By Step Guide To Monitoring And EvaluationA Step By Step Guide To Monitoring And Evaluation
A Step By Step Guide To Monitoring And EvaluationHeather Strinden
 
Degree Works Project Final as of Nov 16 2016-Final
Degree Works Project Final as of Nov 16 2016-FinalDegree Works Project Final as of Nov 16 2016-Final
Degree Works Project Final as of Nov 16 2016-FinalLisa Threlkeld
 
Using Theories of Change to evaluate Information Literacy initiatives ECIL 20...
Using Theories of Change to evaluate Information Literacy initiatives ECIL 20...Using Theories of Change to evaluate Information Literacy initiatives ECIL 20...
Using Theories of Change to evaluate Information Literacy initiatives ECIL 20...Pamela McKinney
 
Using theories of change to evaluate information literacy initiatives
Using theories of change to evaluate information literacy initiatives Using theories of change to evaluate information literacy initiatives
Using theories of change to evaluate information literacy initiatives Sheila Webber
 
Participatory monitoring and evaluation.pptx
Participatory monitoring and evaluation.pptxParticipatory monitoring and evaluation.pptx
Participatory monitoring and evaluation.pptxJohnAutoroza1
 
Running head CCCU SIGNAGE1CCCU SIGNAGE2Project manage.docx
Running head CCCU SIGNAGE1CCCU SIGNAGE2Project manage.docxRunning head CCCU SIGNAGE1CCCU SIGNAGE2Project manage.docx
Running head CCCU SIGNAGE1CCCU SIGNAGE2Project manage.docxsusanschei
 
Different methods of co design- how can different decisions in co-design affe...
Different methods of co design- how can different decisions in co-design affe...Different methods of co design- how can different decisions in co-design affe...
Different methods of co design- how can different decisions in co-design affe...Michael Solaymantash
 
Writing evaluation report of a project
Writing evaluation report of a projectWriting evaluation report of a project
Writing evaluation report of a project03363635718
 
PMC8
PMC8PMC8
PMC8PMI2011
 
Ngo’s project management
Ngo’s project managementNgo’s project management
Ngo’s project managementMohamed Ahmed Said
 
THESIS RESEARCH REPORT NOTESProject relationship managemen.docx
THESIS RESEARCH REPORT NOTESProject relationship managemen.docxTHESIS RESEARCH REPORT NOTESProject relationship managemen.docx
THESIS RESEARCH REPORT NOTESProject relationship managemen.docxchristalgrieg
 
Collaborative 2 ingrid margarita and sandra
Collaborative 2 ingrid margarita and sandraCollaborative 2 ingrid margarita and sandra
Collaborative 2 ingrid margarita and sandraSandra Guevara
 
Project management from simple to complex
Project management   from simple to complexProject management   from simple to complex
Project management from simple to complexBui Huong
 
Change Agent Network - Viewpoints cards - evaluation impact and sustainability
Change Agent Network - Viewpoints cards - evaluation impact and sustainabilityChange Agent Network - Viewpoints cards - evaluation impact and sustainability
Change Agent Network - Viewpoints cards - evaluation impact and sustainabilitybalham
 
EES2014_full paper_Benjamin Aisya Aug 30 2014
EES2014_full paper_Benjamin Aisya Aug 30 2014EES2014_full paper_Benjamin Aisya Aug 30 2014
EES2014_full paper_Benjamin Aisya Aug 30 2014Benjamin Aisya
 
Using theories of change to evaluate information literacy initiatives: LILAC ...
Using theories of change to evaluate information literacy initiatives: LILAC ...Using theories of change to evaluate information literacy initiatives: LILAC ...
Using theories of change to evaluate information literacy initiatives: LILAC ...Pamela McKinney
 
Using theory of change to evaluate information literacy initiatives
Using theory of change to evaluate information literacy initiativesUsing theory of change to evaluate information literacy initiatives
Using theory of change to evaluate information literacy initiativesSheila Webber
 

Similar to External evaluation of the peer evaluation process final1 (20)

A Step By Step Guide To Monitoring And Evaluation
A Step By Step Guide To Monitoring And EvaluationA Step By Step Guide To Monitoring And Evaluation
A Step By Step Guide To Monitoring And Evaluation
 
Stakeholder analysis tool
Stakeholder analysis toolStakeholder analysis tool
Stakeholder analysis tool
 
Degree Works Project Final as of Nov 16 2016-Final
Degree Works Project Final as of Nov 16 2016-FinalDegree Works Project Final as of Nov 16 2016-Final
Degree Works Project Final as of Nov 16 2016-Final
 
Using Theories of Change to evaluate Information Literacy initiatives ECIL 20...
Using Theories of Change to evaluate Information Literacy initiatives ECIL 20...Using Theories of Change to evaluate Information Literacy initiatives ECIL 20...
Using Theories of Change to evaluate Information Literacy initiatives ECIL 20...
 
Using theories of change to evaluate information literacy initiatives
Using theories of change to evaluate information literacy initiatives Using theories of change to evaluate information literacy initiatives
Using theories of change to evaluate information literacy initiatives
 
Participatory monitoring and evaluation.pptx
Participatory monitoring and evaluation.pptxParticipatory monitoring and evaluation.pptx
Participatory monitoring and evaluation.pptx
 
Running head CCCU SIGNAGE1CCCU SIGNAGE2Project manage.docx
Running head CCCU SIGNAGE1CCCU SIGNAGE2Project manage.docxRunning head CCCU SIGNAGE1CCCU SIGNAGE2Project manage.docx
Running head CCCU SIGNAGE1CCCU SIGNAGE2Project manage.docx
 
Different methods of co design- how can different decisions in co-design affe...
Different methods of co design- how can different decisions in co-design affe...Different methods of co design- how can different decisions in co-design affe...
Different methods of co design- how can different decisions in co-design affe...
 
Writing evaluation report of a project
Writing evaluation report of a projectWriting evaluation report of a project
Writing evaluation report of a project
 
PMC8
PMC8PMC8
PMC8
 
Ngo’s project management
Ngo’s project managementNgo’s project management
Ngo’s project management
 
THESIS RESEARCH REPORT NOTESProject relationship managemen.docx
THESIS RESEARCH REPORT NOTESProject relationship managemen.docxTHESIS RESEARCH REPORT NOTESProject relationship managemen.docx
THESIS RESEARCH REPORT NOTESProject relationship managemen.docx
 
Collaborative 2 ingrid margarita and sandra
Collaborative 2 ingrid margarita and sandraCollaborative 2 ingrid margarita and sandra
Collaborative 2 ingrid margarita and sandra
 
Project management from simple to complex
Project management   from simple to complexProject management   from simple to complex
Project management from simple to complex
 
Change Agent Network - Viewpoints cards - evaluation impact and sustainability
Change Agent Network - Viewpoints cards - evaluation impact and sustainabilityChange Agent Network - Viewpoints cards - evaluation impact and sustainability
Change Agent Network - Viewpoints cards - evaluation impact and sustainability
 
EES2014_full paper_Benjamin Aisya Aug 30 2014
EES2014_full paper_Benjamin Aisya Aug 30 2014EES2014_full paper_Benjamin Aisya Aug 30 2014
EES2014_full paper_Benjamin Aisya Aug 30 2014
 
Using theories of change to evaluate information literacy initiatives: LILAC ...
Using theories of change to evaluate information literacy initiatives: LILAC ...Using theories of change to evaluate information literacy initiatives: LILAC ...
Using theories of change to evaluate information literacy initiatives: LILAC ...
 
Using theory of change to evaluate information literacy initiatives
Using theory of change to evaluate information literacy initiativesUsing theory of change to evaluate information literacy initiatives
Using theory of change to evaluate information literacy initiatives
 
CIB8942.pdf
CIB8942.pdfCIB8942.pdf
CIB8942.pdf
 
Final intropm
Final intropmFinal intropm
Final intropm
 

More from Association of Colleges

Aoc prospectus online_2014-15_vict
Aoc prospectus online_2014-15_victAoc prospectus online_2014-15_vict
Aoc prospectus online_2014-15_victAssociation of Colleges
 
AoC Beacon Awards 2014-15 - How to Apply
AoC Beacon Awards 2014-15 - How to ApplyAoC Beacon Awards 2014-15 - How to Apply
AoC Beacon Awards 2014-15 - How to ApplyAssociation of Colleges
 
AoC Beacon Awards 2014-15 - Prospectus
AoC Beacon Awards 2014-15 - ProspectusAoC Beacon Awards 2014-15 - Prospectus
AoC Beacon Awards 2014-15 - ProspectusAssociation of Colleges
 
AoC Beacon Awards 2014-15 - Programme
AoC Beacon Awards 2014-15 - ProgrammeAoC Beacon Awards 2014-15 - Programme
AoC Beacon Awards 2014-15 - ProgrammeAssociation of Colleges
 
AoC Beacon Awards 2014-15 - OCR Award for Innovation in FE
AoC Beacon Awards 2014-15 - OCR Award for Innovation in FEAoC Beacon Awards 2014-15 - OCR Award for Innovation in FE
AoC Beacon Awards 2014-15 - OCR Award for Innovation in FEAssociation of Colleges
 
AoC Beacon Awards 2014-15 - NAMSS Award for Student Support
AoC Beacon Awards 2014-15 - NAMSS Award for Student SupportAoC Beacon Awards 2014-15 - NAMSS Award for Student Support
AoC Beacon Awards 2014-15 - NAMSS Award for Student SupportAssociation of Colleges
 
AoC Beacon Awards 2014-15 - Microlink, AoC Charitable Trust and Achievement f...
AoC Beacon Awards 2014-15 - Microlink, AoC Charitable Trust and Achievement f...AoC Beacon Awards 2014-15 - Microlink, AoC Charitable Trust and Achievement f...
AoC Beacon Awards 2014-15 - Microlink, AoC Charitable Trust and Achievement f...Association of Colleges
 
AoC Beacon Awards 2014-15 - Learning Consortium Award for Improvement in Teac...
AoC Beacon Awards 2014-15 - Learning Consortium Award for Improvement in Teac...AoC Beacon Awards 2014-15 - Learning Consortium Award for Improvement in Teac...
AoC Beacon Awards 2014-15 - Learning Consortium Award for Improvement in Teac...Association of Colleges
 
AoC Beacon Awards 2014-15 prospectus - JLT Employee Benefits Award for Health...
AoC Beacon Awards 2014-15 prospectus - JLT Employee Benefits Award for Health...AoC Beacon Awards 2014-15 prospectus - JLT Employee Benefits Award for Health...
AoC Beacon Awards 2014-15 prospectus - JLT Employee Benefits Award for Health...Association of Colleges
 
AoC Beacon Awards 2014-15 - How to Ppply
AoC Beacon Awards 2014-15 - How to PpplyAoC Beacon Awards 2014-15 - How to Ppply
AoC Beacon Awards 2014-15 - How to PpplyAssociation of Colleges
 
AoC Beacon Awards 2014-15 - Jisc Award for the Effective Use of Technology in FE
AoC Beacon Awards 2014-15 - Jisc Award for the Effective Use of Technology in FEAoC Beacon Awards 2014-15 - Jisc Award for the Effective Use of Technology in FE
AoC Beacon Awards 2014-15 - Jisc Award for the Effective Use of Technology in FEAssociation of Colleges
 
AoC Beacon Awards 2014-15 - Guidance on How to Apply
AoC Beacon Awards 2014-15 - Guidance on How to ApplyAoC Beacon Awards 2014-15 - Guidance on How to Apply
AoC Beacon Awards 2014-15 - Guidance on How to ApplyAssociation of Colleges
 
AoC Beacon Awards 2014-15 - Education and Training Foundation Award for Trans...
AoC Beacon Awards 2014-15 - Education and Training Foundation Award for Trans...AoC Beacon Awards 2014-15 - Education and Training Foundation Award for Trans...
AoC Beacon Awards 2014-15 - Education and Training Foundation Award for Trans...Association of Colleges
 
AoC Beacon Awards 2014-15 - edge award for practical teaching and practical l...
AoC Beacon Awards 2014-15 - edge award for practical teaching and practical l...AoC Beacon Awards 2014-15 - edge award for practical teaching and practical l...
AoC Beacon Awards 2014-15 - edge award for practical teaching and practical l...Association of Colleges
 
AoC Beacon Awards 2014-15 - Calendar
AoC Beacon Awards 2014-15 - CalendarAoC Beacon Awards 2014-15 - Calendar
AoC Beacon Awards 2014-15 - CalendarAssociation of Colleges
 
AoC Beacon Awards 2014-15 - Association of Colleges Award for College Engagem...
AoC Beacon Awards 2014-15 - Association of Colleges Award for College Engagem...AoC Beacon Awards 2014-15 - Association of Colleges Award for College Engagem...
AoC Beacon Awards 2014-15 - Association of Colleges Award for College Engagem...Association of Colleges
 
AoC Beacon Awards 2014-15 - Assessment
AoC Beacon Awards 2014-15 - AssessmentAoC Beacon Awards 2014-15 - Assessment
AoC Beacon Awards 2014-15 - AssessmentAssociation of Colleges
 
AoC Beacon Awards 2014-15 - AQA Award for Continued Engagement in Education a...
AoC Beacon Awards 2014-15 - AQA Award for Continued Engagement in Education a...AoC Beacon Awards 2014-15 - AQA Award for Continued Engagement in Education a...
AoC Beacon Awards 2014-15 - AQA Award for Continued Engagement in Education a...Association of Colleges
 
AoC Beacon Awards 2014-15 - Application Form
AoC Beacon Awards 2014-15 - Application FormAoC Beacon Awards 2014-15 - Application Form
AoC Beacon Awards 2014-15 - Application FormAssociation of Colleges
 

More from Association of Colleges (20)

Aoc prospectus online_2014-15_vict
Aoc prospectus online_2014-15_victAoc prospectus online_2014-15_vict
Aoc prospectus online_2014-15_vict
 
AoC Prospectus 2014-15
AoC Prospectus 2014-15AoC Prospectus 2014-15
AoC Prospectus 2014-15
 
AoC Beacon Awards 2014-15 - How to Apply
AoC Beacon Awards 2014-15 - How to ApplyAoC Beacon Awards 2014-15 - How to Apply
AoC Beacon Awards 2014-15 - How to Apply
 
AoC Beacon Awards 2014-15 - Prospectus
AoC Beacon Awards 2014-15 - ProspectusAoC Beacon Awards 2014-15 - Prospectus
AoC Beacon Awards 2014-15 - Prospectus
 
AoC Beacon Awards 2014-15 - Programme
AoC Beacon Awards 2014-15 - ProgrammeAoC Beacon Awards 2014-15 - Programme
AoC Beacon Awards 2014-15 - Programme
 
AoC Beacon Awards 2014-15 - OCR Award for Innovation in FE
AoC Beacon Awards 2014-15 - OCR Award for Innovation in FEAoC Beacon Awards 2014-15 - OCR Award for Innovation in FE
AoC Beacon Awards 2014-15 - OCR Award for Innovation in FE
 
AoC Beacon Awards 2014-15 - NAMSS Award for Student Support
AoC Beacon Awards 2014-15 - NAMSS Award for Student SupportAoC Beacon Awards 2014-15 - NAMSS Award for Student Support
AoC Beacon Awards 2014-15 - NAMSS Award for Student Support
 
AoC Beacon Awards 2014-15 - Microlink, AoC Charitable Trust and Achievement f...
AoC Beacon Awards 2014-15 - Microlink, AoC Charitable Trust and Achievement f...AoC Beacon Awards 2014-15 - Microlink, AoC Charitable Trust and Achievement f...
AoC Beacon Awards 2014-15 - Microlink, AoC Charitable Trust and Achievement f...
 
AoC Beacon Awards 2014-15 - Learning Consortium Award for Improvement in Teac...
AoC Beacon Awards 2014-15 - Learning Consortium Award for Improvement in Teac...AoC Beacon Awards 2014-15 - Learning Consortium Award for Improvement in Teac...
AoC Beacon Awards 2014-15 - Learning Consortium Award for Improvement in Teac...
 
AoC Beacon Awards 2014-15 prospectus - JLT Employee Benefits Award for Health...
AoC Beacon Awards 2014-15 prospectus - JLT Employee Benefits Award for Health...AoC Beacon Awards 2014-15 prospectus - JLT Employee Benefits Award for Health...
AoC Beacon Awards 2014-15 prospectus - JLT Employee Benefits Award for Health...
 
AoC Beacon Awards 2014-15 - How to Ppply
AoC Beacon Awards 2014-15 - How to PpplyAoC Beacon Awards 2014-15 - How to Ppply
AoC Beacon Awards 2014-15 - How to Ppply
 
AoC Beacon Awards 2014-15 - Jisc Award for the Effective Use of Technology in FE
AoC Beacon Awards 2014-15 - Jisc Award for the Effective Use of Technology in FEAoC Beacon Awards 2014-15 - Jisc Award for the Effective Use of Technology in FE
AoC Beacon Awards 2014-15 - Jisc Award for the Effective Use of Technology in FE
 
AoC Beacon Awards 2014-15 - Guidance on How to Apply
AoC Beacon Awards 2014-15 - Guidance on How to ApplyAoC Beacon Awards 2014-15 - Guidance on How to Apply
AoC Beacon Awards 2014-15 - Guidance on How to Apply
 
AoC Beacon Awards 2014-15 - Education and Training Foundation Award for Trans...
AoC Beacon Awards 2014-15 - Education and Training Foundation Award for Trans...AoC Beacon Awards 2014-15 - Education and Training Foundation Award for Trans...
AoC Beacon Awards 2014-15 - Education and Training Foundation Award for Trans...
 
AoC Beacon Awards 2014-15 - edge award for practical teaching and practical l...
AoC Beacon Awards 2014-15 - edge award for practical teaching and practical l...AoC Beacon Awards 2014-15 - edge award for practical teaching and practical l...
AoC Beacon Awards 2014-15 - edge award for practical teaching and practical l...
 
AoC Beacon Awards 2014-15 - Calendar
AoC Beacon Awards 2014-15 - CalendarAoC Beacon Awards 2014-15 - Calendar
AoC Beacon Awards 2014-15 - Calendar
 
AoC Beacon Awards 2014-15 - Association of Colleges Award for College Engagem...
AoC Beacon Awards 2014-15 - Association of Colleges Award for College Engagem...AoC Beacon Awards 2014-15 - Association of Colleges Award for College Engagem...
AoC Beacon Awards 2014-15 - Association of Colleges Award for College Engagem...
 
AoC Beacon Awards 2014-15 - Assessment
AoC Beacon Awards 2014-15 - AssessmentAoC Beacon Awards 2014-15 - Assessment
AoC Beacon Awards 2014-15 - Assessment
 
AoC Beacon Awards 2014-15 - AQA Award for Continued Engagement in Education a...
AoC Beacon Awards 2014-15 - AQA Award for Continued Engagement in Education a...AoC Beacon Awards 2014-15 - AQA Award for Continued Engagement in Education a...
AoC Beacon Awards 2014-15 - AQA Award for Continued Engagement in Education a...
 
AoC Beacon Awards 2014-15 - Application Form
AoC Beacon Awards 2014-15 - Application FormAoC Beacon Awards 2014-15 - Application Form
AoC Beacon Awards 2014-15 - Application Form
 

Recently uploaded

How to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERPHow to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERPCeline George
 
Gas measurement O2,Co2,& ph) 04/2024.pptx
Gas measurement O2,Co2,& ph) 04/2024.pptxGas measurement O2,Co2,& ph) 04/2024.pptx
Gas measurement O2,Co2,& ph) 04/2024.pptxDr.Ibrahim Hassaan
 
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptxMULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptxAnupkumar Sharma
 
Crayon Activity Handout For the Crayon A
Crayon Activity Handout For the Crayon ACrayon Activity Handout For the Crayon A
Crayon Activity Handout For the Crayon AUnboundStockton
 
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️9953056974 Low Rate Call Girls In Saket, Delhi NCR
 
Like-prefer-love -hate+verb+ing & silent letters & citizenship text.pdf
Like-prefer-love -hate+verb+ing & silent letters & citizenship text.pdfLike-prefer-love -hate+verb+ing & silent letters & citizenship text.pdf
Like-prefer-love -hate+verb+ing & silent letters & citizenship text.pdfMr Bounab Samir
 
Quarter 4 Peace-education.pptx Catch Up Friday
Quarter 4 Peace-education.pptx Catch Up FridayQuarter 4 Peace-education.pptx Catch Up Friday
Quarter 4 Peace-education.pptx Catch Up FridayMakMakNepo
 
Introduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher EducationIntroduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher Educationpboyjonauth
 
Romantic Opera MUSIC FOR GRADE NINE pptx
Romantic Opera MUSIC FOR GRADE NINE pptxRomantic Opera MUSIC FOR GRADE NINE pptx
Romantic Opera MUSIC FOR GRADE NINE pptxsqpmdrvczh
 
ROOT CAUSE ANALYSIS PowerPoint Presentation
ROOT CAUSE ANALYSIS PowerPoint PresentationROOT CAUSE ANALYSIS PowerPoint Presentation
ROOT CAUSE ANALYSIS PowerPoint PresentationAadityaSharma884161
 
Field Attribute Index Feature in Odoo 17
Field Attribute Index Feature in Odoo 17Field Attribute Index Feature in Odoo 17
Field Attribute Index Feature in Odoo 17Celine George
 
Judging the Relevance and worth of ideas part 2.pptx
Judging the Relevance  and worth of ideas part 2.pptxJudging the Relevance  and worth of ideas part 2.pptx
Judging the Relevance and worth of ideas part 2.pptxSherlyMaeNeri
 
Procuring digital preservation CAN be quick and painless with our new dynamic...
Procuring digital preservation CAN be quick and painless with our new dynamic...Procuring digital preservation CAN be quick and painless with our new dynamic...
Procuring digital preservation CAN be quick and painless with our new dynamic...Jisc
 
ENGLISH 7_Q4_LESSON 2_ Employing a Variety of Strategies for Effective Interp...
ENGLISH 7_Q4_LESSON 2_ Employing a Variety of Strategies for Effective Interp...ENGLISH 7_Q4_LESSON 2_ Employing a Variety of Strategies for Effective Interp...
ENGLISH 7_Q4_LESSON 2_ Employing a Variety of Strategies for Effective Interp...JhezDiaz1
 
How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17Celine George
 
Proudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxProudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxthorishapillay1
 

Recently uploaded (20)

How to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERPHow to do quick user assign in kanban in Odoo 17 ERP
How to do quick user assign in kanban in Odoo 17 ERP
 
Gas measurement O2,Co2,& ph) 04/2024.pptx
Gas measurement O2,Co2,& ph) 04/2024.pptxGas measurement O2,Co2,& ph) 04/2024.pptx
Gas measurement O2,Co2,& ph) 04/2024.pptx
 
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptxMULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
 
Crayon Activity Handout For the Crayon A
Crayon Activity Handout For the Crayon ACrayon Activity Handout For the Crayon A
Crayon Activity Handout For the Crayon A
 
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
 
Raw materials used in Herbal Cosmetics.pptx
Raw materials used in Herbal Cosmetics.pptxRaw materials used in Herbal Cosmetics.pptx
Raw materials used in Herbal Cosmetics.pptx
 
Like-prefer-love -hate+verb+ing & silent letters & citizenship text.pdf
Like-prefer-love -hate+verb+ing & silent letters & citizenship text.pdfLike-prefer-love -hate+verb+ing & silent letters & citizenship text.pdf
Like-prefer-love -hate+verb+ing & silent letters & citizenship text.pdf
 
Quarter 4 Peace-education.pptx Catch Up Friday
Quarter 4 Peace-education.pptx Catch Up FridayQuarter 4 Peace-education.pptx Catch Up Friday
Quarter 4 Peace-education.pptx Catch Up Friday
 
Model Call Girl in Bikash Puri Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Bikash Puri  Delhi reach out to us at 🔝9953056974🔝Model Call Girl in Bikash Puri  Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Bikash Puri Delhi reach out to us at 🔝9953056974🔝
 
Introduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher EducationIntroduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher Education
 
Romantic Opera MUSIC FOR GRADE NINE pptx
Romantic Opera MUSIC FOR GRADE NINE pptxRomantic Opera MUSIC FOR GRADE NINE pptx
Romantic Opera MUSIC FOR GRADE NINE pptx
 
ROOT CAUSE ANALYSIS PowerPoint Presentation
ROOT CAUSE ANALYSIS PowerPoint PresentationROOT CAUSE ANALYSIS PowerPoint Presentation
ROOT CAUSE ANALYSIS PowerPoint Presentation
 
Field Attribute Index Feature in Odoo 17
Field Attribute Index Feature in Odoo 17Field Attribute Index Feature in Odoo 17
Field Attribute Index Feature in Odoo 17
 
Judging the Relevance and worth of ideas part 2.pptx
Judging the Relevance  and worth of ideas part 2.pptxJudging the Relevance  and worth of ideas part 2.pptx
Judging the Relevance and worth of ideas part 2.pptx
 
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdfTataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
 
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Tilak Nagar Delhi reach out to us at 🔝9953056974🔝
 
Procuring digital preservation CAN be quick and painless with our new dynamic...
Procuring digital preservation CAN be quick and painless with our new dynamic...Procuring digital preservation CAN be quick and painless with our new dynamic...
Procuring digital preservation CAN be quick and painless with our new dynamic...
 
ENGLISH 7_Q4_LESSON 2_ Employing a Variety of Strategies for Effective Interp...
ENGLISH 7_Q4_LESSON 2_ Employing a Variety of Strategies for Effective Interp...ENGLISH 7_Q4_LESSON 2_ Employing a Variety of Strategies for Effective Interp...
ENGLISH 7_Q4_LESSON 2_ Employing a Variety of Strategies for Effective Interp...
 
How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17
 
Proudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxProudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptx
 

External evaluation of the peer evaluation process final1

  • 1. 1 External Evaluation of the Peer Evaluation Process undertaken by the University of Greenwich and Linking London Credit Accumulation and Transfer Scheme Pilot Projects Viki Faulkner July 2013
  • 2. 2 Table of Contents 1. Introduction .................................................................................................................... 3 2. Overview of the Peer Evaluation Process.......................................................................... 3 3. Strengths of the Peer Evaluation Process.......................................................................... 4 4. Challenges Within the Peer Evaluation Process................................................................. 5 5. Additional Monitoring and Evaluation Activities ............................................................... 6 6. Learning Points................................................................................................................ 6 7. Best Practice and Recommendations................................................................................ 8 8. Conclusions ..................................................................................................................... 9 Appendix 1: CATS Pilot Project Peer Evaluation Process........................................................... 10
  • 3. 3 External Evaluation of the Peer Evaluation Process undertaken by the University of Greenwich and Linking London Credit Accumulation and Transfer Scheme Pilot Projects 1. Introduction The University of Greenwich and Linking London each led one of the six, Association of Colleges (AoC) Department for Business Innovation and Skills (BIS) and Skills Funding Agency (SFA) funded credit accumulation and transfer scheme (CATS) pilot projects. The projects arose as a response to ‘New Challenges New Chances’1 , which set out the Government vision to reform the further education and skills system. This recognised that further and higher education institutions would need to find new ways of working together, providing clear and transparent progression routes that would support workforce development and help learners to reach their full potential. The University of Greenwich project, ‘Building on Competence in the Workplace’ focused on work between the university and five members of its partner college network. The Linking London ‘CATS Project in Management’ focused on work between two universities, Birkbeck and Middlesex and four further education colleges. Both projects also had a strong level of engagement from professional bodies and provided mechanisms to capture and integrate the input from learners and employer representatives. The projects both commenced on 1 July 2012 and ran for one year, concluding 30 June 2013. Both projects have undertaken extensive monitoring and evaluation activities and these are the subject of individual submissions to AoC. Additionally, they have taken part in the overarching evaluation of the CATS projects undertaken by AoC. This report, commissioned by the University of Greenwich, does not therefore cover ground already well reported, but instead focuses on providing an external evaluation of the effectiveness and impact of an added quality enhancement aspect: the peer evaluation process. 2. Overview of the Peer Evaluation Process The peer evaluation process was identified at the time of project submission as an activity that both teams felt would add value and it was built within the project plans from the start. The model selected was based on one previously developed by the Kent and Medway Lifelong Learning Network2 and had already been trialled by the University of Greenwich to measure the impact of their LLN funded projects. The peer evaluation process consisted to two elements: an impact evaluation model – to ensure consistency and provide a common framework, and a monitoring and evaluation plan. It was 1 New Challenges, New Chances: Further Education and Skills Reform Plan: Building a World-Class Skills System (BIS, 1 December 2011) 2 Perrin, L Time to Make a Difference: achieving sustainability through projects (2010) KMLLN
  • 4. 4 described in detail in a paper that was attached to the University of Greenwich CATS project interim report (Feb 2013) and this is attached as appendix 1 to minimise repetition. The University of Greenwich and Linking London were mutually selecting peer evaluation ‘buddies’. They were logical partners in the process as both had previous experience of peer evaluation and were partnerships with similar roots – each born out of the HEFCE funded lifelong learning networks initiative. This proved to be a very effective factor in ensuring common purpose and a significant level of shared understanding. 3. Strengths of the Peer Evaluation Process The Impact Evaluation Model adopted as a mechanism to guide the peer evaluation process, set a clear focus and structure for the activity. This was a real strength of the process that was adopted, ensuring that peer evaluation was a constructive and productive activity and did not become an unfocused ‘talking shop’. The project manager of the Linking London project, who was introduced to the model for the first time through this activity, felt it was an example of good practice and found it very useful in helping to shape her initial thinking around evidence gathering. The project teams worked together at an early stage to identify inputs, outputs, intermediate outcomes and final outcomes for each of their projects. They then used this information to populate the model, identifying evidence needs and evidence sources. This meant that from a very early phase, they were already thinking about what impact they wanted their project to achieve – and not just focusing on what they were planning to do in the short term. The model stresses that final outcomes are things which will extend beyond the lifespan of a one year funded project and so consideration needs to be given at the stage to how longer term outcomes can be captured and could be evidenced. It is a strength of the model that it forces a mindset shift into a consideration of longer term behavioural changes rather than a focus on delivery of short term targets. Both partners clearly articulated the benefits that they feel they gained from having a ‘critical friend’ that helped them to keep focused and added another layer under the formal reporting structure and network set up by the AoC. Although both projects had numerous contacts with other CATS projects, they each felt that the relationship that they established with their peer evaluation partner was much stronger and more constructive. Comments from members of the project teams included statements such as, ”it was nice sometimes to not feel you are alone”, “ we were more prepared to share issues and problems, it was more constructive, helpful, not just about reporting on progress”, “the peer evaluation was more process based – it was useful”. These comments demonstrate that they had been able to establish an open trusting relationship with each other and this is when peer support can be most effective. One team member described how much they valued the fact that discussions with the peer evaluation partner did not focus only on the achievement of project aims and objectives, but how they were also able to “talk around the edges of the project”. This was a wonderful phrase and was illustrated by several examples where project team members described how they had shared experiences of working with partners – and strategies for managing the challenges this presents and working within the complexities of a steering group ‘management’ structure. The use of Dropbox has provided an innovative tool that enhanced communication between multiple partners. Project teams at Linking London and the University of Greenwich have all been
  • 5. 5 able to access Dropbox and deposit up to date versions of documentation there. It is a cloud based storage facility which means it is accessible wherever you are and you therefore always have access to the most up to date versions of any document – no more relying on someone to have sent you an email or risking files ending up blocked by over cautious spam filters. The ease of access is a real advantage and has also enabled a very effective mechanism for on-going external evaluation. The external evaluator also had access to Dropbox and so was therefore able to regularly ‘dip in’ to sample the evidence available – effectively moving external evaluation from being a twice in a project snapshot activity, into a continuous monitoring process which proves much more rigorous as an exercise. A further strength of the Dropbox facility was the way in which the folders had been set up. It was structured from the start with four folders that aligned with the elements of the Impact Evaluation Model: inputs, outputs, intermediate outcomes and final outcomes. A great benefit of this structure was that, from the very beginning, this meant that the project teams had to engage critically with every piece of evidence they collected; having to differentiate clearly between inputs, outputs and outcomes. 4. Challenges within the Peer Evaluation Process Although both projects focused on credit accumulation and transfer schemes, there were substantial differences between the two with the projects having very different outputs. At times this made it difficult for each partner to see where lessons could be directly transferrable. This was quite noticeable at the interim project phase where the University of Greenwich had been able to provide clear support in providing the impact evaluation model and sharing their experience of setting up clear reporting schedules and monitoring cycles for managing projects with partners, yet they were less able, at that stage, to articulate tangible benefits or transferrable learning which they had got out of the peer evaluation process. However, the value of the activity as a whole was clear and both project teams valued the reflective process which they said kept them focused and helped them each to clarify their own thoughts about their project. The peer evaluation process was not a requirement of the original AoC project call. It was built into the proposal of both projects because the teams recognised the importance of assessing impact to justify value for money and it was felt that this would be an effective and efficient way to achieve it. According to Appendix 1 of the University of Greenwich Interim Report (Feb 2013), ‘It was agreed that the process should be simple, clear and sufficiently robust but not so onerous as to divert time and resources away from other important project activities.’ In this case, the peer evaluation process has achieved its objectives. The process was clear and simple, it had a strong focus on impact, reinforced by the methodology, and the Impact Evaluation Model adopted provided a robust mechanism for planning and capturing evidence on which to base sound judgements. The peer evaluation also provided a cost effective mechanism which did minimise funding drain. However this was, to a not insubstantial degree, mainly down to a smart selection of partnering organisations. Linking London and the University of Greenwich are geographically close and this means that the four face to face meetings planned within the monitoring and evaluation plan were quite achievable. As mentioned earlier, Linking London and the University of Greenwich project team have previous experience of each others’ work as they shared common features and a common language, working closely together as lifelong learning networks. This enabled the project teams to move quickly past
  • 6. 6 the ‘getting to know you’ phase and move directly into the ‘getting down to business’ phase. Where project teams do not already share a common bond, establishing a mutually supportive, trusting relationship may take a little longer than one meeting. Where partners are geographically quite distant also, this could put even more strain on developing a relationship – thereby requiring a greater investment of time to get it established. In such a case, the peer evaluation process may quickly begin to look less cost effective and could easily begin to look like an extra layer of monitoring and evaluation, also beginning to divert resources from delivery of core activities. It is therefore suggested that the smart selection of a peer evaluation partner is one of the critical success factors for this aspect of activity and further adoption of this as a process should strongly consider providing clear guidance on partner selection criteria. Dropbox, a cloud-based document repository, provided a very effective mechanism for multiple partners to share documentation and to organise evidence collection. Its use was one of the distinctive features of the peer evaluation model but unfortunately it was not in place from the start of the project. This was disappointing and it would have been beneficial for it to be available from the beginning of the project. With hindsight, the team at the University of Greenwich who were responsible for setting up the Dropbox facility agree with this and feel that in future, it would be one of the first things put in place as part of a peer evaluation information strategy. Timescales provided a further area of challenge. Although both projects were working towards an overarching timetable set by the AoC, within that there were minor variations in each project, such as the dates set for their final dissemination events. This did not cause any major challenges but an early deadline for the Linking London dissemination event meant that the final peer evaluation meeting had to be held before the University of Greenwich had been able to complete all of its final evidence collection. Hence evaluation reports were written before case studies were collected and so could not be included in the evidence base available. 5. Additional Monitoring and Evaluation Activities within the University of Greenwich Project The Peer evaluation process within the University of Greenwich project was supplemented with monitoring and evaluation activities undertaken by a specially devised steering group. This group comprised senior managers from the colleges, the project team and employers representing SMEs and large companies in SE London. It was an appropriately constructed group that enabled effective oversight of activities and a useful forum for raising areas of concern. Steering group meetings were hosted by partners in turn. This built the collective awareness of others’ campuses and facilities and was an effective way in which the project was able to strengthen relationships between partners. Steering group meetings not only provided a good vehicle for monitoring project progress but also a forum for staff development activities, enabling all partners to gain a shared understanding of credit and accreditation of prior learning. Greenwich has identified these as ‘active meetings’ – ones that blend together a business agenda with a development session, encouraging interaction and participation of all members. 6. Learning Points Both projects agreed, as part of the peer evaluation process, that it was important to maintain a flexible approach towards project aims and outcomes and to recognise that plans will need to be
  • 7. 7 adapted as a project progresses. Beginning in July as both projects did, they shared a common concern in the early phase of trying to make steady progress at a time when most academic institutions seem to be run by a skeleton staffing model. The University of Greenwich project was in a more advantageous position here at it was set out with a tightly focused proposal based on a pre- existing relationship. It was also well timed as it built on the colleges’ validation to offer the Applied Professional Studies programme. The tight bid with a clear focus certainly helped the University of Greenwich to be able to achieve its aims and objectives in a timely fashion. The Linking London project found the summer break more problematic as they had to establish a number of sub- projects and, in a number of cases, build new relationships between colleges and universities. Both projects had to respond flexibly and adapt to staff changes in what seems to have been an especially turbulent time for FE college partners. The Project Manager of the Linking London project found the first meeting of the peer evaluation group especially useful, not only helping her to situate their project, but also helping her to realise that they were not alone in the situation in which they found themselves with regard to concerns over timescales. As a project manager new to this type of working, the peer evaluation process was especially beneficial, enabling her to work closely with the experienced project manager on the University of Greenwich project. Although both projects had different challenges and very different outcomes, as well as taking on the role of ‘critical friend’ they were able to provide mutual support and this is an aspect that is not to be under-estimated. Although there were a number of meetings of all CATS pilots, facilitated by AoC, both projects described the relationship that formed between them as “deeper” and “richer” than that they had with the other projects. Neither project was afraid to ask their peer evaluation partner for advice and support and neither felt they were being judged in any way. There are some indications here that relationships which foster outside of the formal oversight of funding agencies can enable a greater level of openness and are therefore more likely to be mutually beneficial. The concept of ‘active meetings’ at the University of Greenwich, that blended together a business agenda for the steering group, with a development session was one which appeared to enhance the level of engagement that partners had with the project collectively. All college partners in their interim reports commented on the “closer working relationships”, the “group collaborations” and “stronger relationships” that had been built in this way and cited it as a real strength of the project. This was an aspect that was picked up on through the peer evaluation process by Linking London and is a methodology that they have already identified as one worth trialling as a future way of working. The University of Greenwich noted that Linking London were very good at disseminating the findings of their project work and took good advantage of every opportunity available to do so, based on a number of well established practices they had already put in place. Greenwich felt that most of their dissemination work had been internally focused, ensuring that they maximised the spread of information amongst staff groups within the partner organisations. As a result of the peer evaluation process they are now looking at putting on a number of road shows – adopting a model used by Linking London for wider dissemination, and creating a toolkit as additional outputs to those identified in the original project plan.
  • 8. 8 7. Best Practice Identified and Recommendations Consistency: The peer evaluation process was based on both projects using the Impact Evaluation Model that had been created originally by the University of Greenwich. Both projects worked together to identify their inputs, outputs and outcomes and used this information to populate the model. This exercise provided the projects with a shared language and a common level of understanding. Although the outputs and outcomes varied between the two projects, and so by necessity the evidence needs and evidence sources they would require, they were each clear about the process that they would be undertaking. This gave a commonality that enabled them to evaluate each others’ project against identified negotiated areas. Without the use of a common model the differences between the projects may have begun to be too distracting making it hard to bring any aspects of consistency to the peer evaluation process. It is therefore recommended that, for a peer evaluation to be effective it is important to adopt a consistent model against which both projects can be measured. Clear focus on impact: The peer evaluation process had a clear focus on impact and this was led by the use of the model adopted. Impact measures, if they are to be effective, need to be embedded into a project from the outset. The University of Greenwich had clearly considered the long term impact of their project and had built within the evaluation model final outcomes that stretched way beyond the one year funded project. This is important. With a project such as the CATS pilot, within the space of a year we can truly only realistically measure project outputs, such as the publishing of the prospectus and some intermediate outcomes such as focus group feedback. The true success of the project will only be revealed through increasing progression rates and this will take a number of years before we can see a clear picture. It is recommended that funded projects should not feel constrained to set evaluation measures which are limited only to the timeframe of the funding attached – but should also consider longer term impact measures and look at how this may be captured in the future. Dropbox: Digital, cloud-based storage facilities can provide a cost efficient and very effective mechanism for sharing large amounts of documentation amongst multiple users who may or may not share a common employing organisation and are likely to based in multiple locations. In this case, Dropbox was used very effectively to improve ease of access to documentation and improve lines of communication between the project teams and the External Evaluator. It is recommended that any multi-partner, complex project seriously consider using digital cloud based document storage as an integral part of an effective communication strategy. Cost effective: The peer evaluation process in this project provided a well integrated and cost effective mechanism for monitoring and evaluation. To ensure robustness of the project oversight, this was also supplemented (at the University of Greenwich) by a rigorous system which included the regular submission and review of partner progress reports, monitoring of progress at steering group meetings and a formal reporting structure set in place by the AoC. This is also supplemented by an overarching evaluation of the CATS projects commissioned by AoC. Where extensive formal monitoring and evaluation processes are already established, it is recommended that projects consider peer evaluation as a useful quality enhancement mechanism.
  • 9. 9 8. Conclusions The peer evaluation process is a real strength of this project. It has been an integral component from project conception through to completion and has provided a genuine added value to both projects whilst placing very limited demands on project resources. Subject to the caveats raised above regarding partner selection and cost effectiveness, it is a model worthy of consideration for adoption in future national project initiatives. The University of Greenwich Impact Evaluation Model selected to form the basis for the peer evaluation activity is now one that has been tested in a range of situations. It has been used effectively by both CATS projects as well as the smaller scale LLN projects, and has demonstrated its transferability and flexibility. It has a strong emphasis on impact – which is critically important within any project, and ensures that longer term outcomes that stretch beyond the funded term of the project are built in from the start. Both of these aspects build sustainability and maximise the long term impact gained from pump prime funding. The model is worthy of further dissemination on a wider stage.
  • 10. 10 Appendix – Building on Competence in the Workplace Interim Report 1 February 2013 University of Greenwich CATS Pilot Project Peer Evaluation Process Background The University of Greenwich and Linking London CATS Pilot Project teams recognise the importance of assessing impact to justify value for money, to determine the direction and consequences of projects and to show outcomes in achievements, products and processes. They have therefore developed and agreed a Peer Evaluation Process by which both project teams will monitor the process and outcomes of the other’s project through a shared ‘virtual’ portfolio of evidence and produce an evaluative commentary to accompany the Project Final Report. It was agreed that the process should be simple, clear and sufficiently robust but not so onerous as to divert time and resources away from other important project activities. The Peer Evaluation Process The Peer Evaluation Process consists of 2 elements: 1. An Impact Evaluation model which ensures consistency and provides a common framework to enable each project team to monitor each others’ progress against their work plan 2. A monitoring and evaluation plan which underpins the IEM by confirming the process by which each project team will monitor and evaluate Impact Evaluation Model (IEM) The Impact Evaluation Model is based on a model previously developed by the Kent and Medway Lifelong Learning Network3 , led by the University of Greenwich to help its partner institutions measure the impact of their LLN-funded projects. In turn based on a model developed by the Training and Development Agency for Schools (TDA)4 it did not claim to offer a brand new approach to evaluation, but attempted to provide an accessible and user- friendly framework to assist their activity. The fact that it brings together qualitative and quantitative evidence and allows us to look not only at data but also the real people involved and the impact the project could have on communities (work-based learners, employers and learning providers) makes it particularly applicable to the CATS project. The model is designed to help the project team build up a picture of how they expect the project to work and by being clear about the outcomes of the project. Initially, the project teams worked together to identify the inputs, outputs, intermediate 3 Perrin, L. Time to Make a Difference: achieving sustainability through projects (2010) KMLLN 4 The impact evaluation model (IEM) (2009) Training and Development Agency for Schools
  • 11. 11 outcomes and final outcomes that their project was trying to achieve and to populate the model with these. This enabled them to articulate how they expected their project to have an impact. The model was then used to identify the evidence needed and from there, to identify the potential sources of evidence. (See page 4 for Linking London and UoG IEMs) By working through the model each project team was able to demonstrate the links between the various stages of project delivery, from planning all the way through to the impact on individual users of the activities and products being delivered and on the overall project aims. From past experience we were aware that the model is most effective when put in place early in the project. By completing the model at this stage the project teams knew what data and evidence we needed to collect at the outset and were able to clarify for each other any uncertainties about the aims of our projects. Articulating the underlying assumptions of the projects and identifying the potential evidence sources at an early stage have informed the project work plans and are being explored as delivery progresses. Monitoring and evaluation plan The following methods are being used to monitor the progress of each others’ project: Method Purpose Frequency Face-to-Face Meetings 1. To update and question each other on progress 2. To agree strengths and weaknesses 3. To share knowledge and best practice 4. To discuss issues and propose solutions Beginning (Sep 12) Mid-term (Jan 13) and End (May 13) Dropbox 1. To store and share records, documents and other evidence and make available to each other Ongoing – as records and documents are produced Emails 1. To update each other on events, dropbox updates, developments 2. To share ideas and deal with issues arising between meetings As and when required between meetings Evaluative Commentary 1. To provide an evaluative commentary to inform final report Once – end of project
  • 12. 12 Face to Face Meetings The principle method to support the monitoring and evaluation process, three meetings of Project Directors and Managers will be held over the course of the project as follows:  27 September 2012 (Birkbeck) – to agree impact evaluation model and agree process  14 January 2013 (UoG) – to report progress to date and agree outcomes for draft interim report  17 May 2013 (Birkbeck) – to report progress to date and agree outcomes for final report Dropbox Dropbox will be used as a virtual repository for all documents, publications, records and other evidence identified on the Impact Evaluation Model. Each project will have its own folder, each containing four sub folders - Inputs, Outputs, Intermediate Outcomes and Final Outcomes (as per IEM). Evidence will be filed in the appropriate sub-folder. Folders and sub-folders will be shared and accessible to each other. In addition, the UoG folder will be accessible to the Viki Faulkner of the Sussex Lifelong Learning Network who has been appointed to undertake an external evaluation of the Peer Evaluation Process. Emails Probably the easiest and most effective way of keeping in touch between meetings, these will be used to maintain contact on an ongoing basis. Evaluative Commentary Each project team will provide a brief evaluative commentary (approx. 500 words) on the others’ project to support the final report and confirming: 1. To what extent did the project meet its aims and objectives? 2. What, if any, are the additional benefits delivered by the project? 3. What lessons did we learn that we could apply to future projects? CATS Pilot Project Teams Linking London Project Director: Sue Betts Project Manager: Pamela Calabro University of Greenwich Project Director: Hugh Joslin Project Manager: Lindsay Perrin
  • 13. 13 Impact Evaluation Models – Linking London and University of Greenwich
  • 14. 14