SlideShare a Scribd company logo
1 of 50
A comparison of the practices used by human resource
development
professionals to evaluate web-based and classroom-based
training
programmes within seven Korean companies
Younghee Jessie Kong
a
* and Ronald Lynn Jacobs
b
*
a
College of Business, Franklin University, Columbus, OH, USA;
b
College of Education,
University of Illinois, Champaign, IL, USA
(Received 28 September 2010; final version received 25
October 2011)
The purpose of the study was to compare the practices used by
human resource
development (HRD) professionals to evaluate web-based and
classroom-based
training (CBT) programmes within seven Korean companies.
This study used
four components of evaluation and three factors of evaluation
barriers to
compare the differences between these two training approaches.
This study also
explored the key decision factors for determining how HRD
professionals
evaluated their web-based and CBT programmes. Two data sets
were used for the
study; one set of data was gathered from a survey questionnaire
distributed to
HRD professionals and the other was gathered from interviews
with HR/HRD
directors within the seven companies. The results showed that
web-based and
CBT programmes were not meaningfully different on the most
components of
evaluation and evaluation barriers. The results also found six
key decision factors
determining evaluation for web-based and CBT programmes.
Keywords: training evaluation; training evaluation barriers;
primary factors
determining evaluation
Introduction
Since many organizations consider training as a way of
improving their performance
in today’s competitive environment (Yanmil and McLean 2001),
they want to select
the best training programmes for their employees among a wide
variety of training
approaches (Bartley and Golek 2004). Classroom-based training
(CBT) is one of the
most frequently used training approaches in organizations, since
learners can directly
communicate with an instructor and peers and directly share
information with each
other (Kapp and McKeague 2002). In addition to the use of
CBT, the development
of web-based training (WBT) has been increased in the human
resource development
(HRD) field due to rapid advancement in the capabilities and
distribution of
technologies (Bassi and Van Buren 1998).
Since investment in these training programmes has been rapidly
increasing in
organizations, senior management asks HRD professionals to
evaluate their WBT
*Corresponding author. Email: [email protected];
[email protected]
Human Resource Development International
Vol. 15, No. 1, February 2012, 79–98
ISSN 1367-8868 print/ISSN 1469-8374 online
� 2012 Taylor & Francis
http://dx.doi.org/10.1080/13678868.2012.658632
http://www.tandfonline.com
and classroom-based training (CBT) programmes in order to
determine the best
training programmes for their employees (Olson and Wisher
2002).
Many previous researchers (Curtain 2002; Jung and Rha 2003;
Rumble 2001;
Whalen and Wright 1999) have stressed that evaluating web-
based and classroom-
based training should be considered and performed differently,
because their
learning environments are entirely different from each other.
For instance, WBT is
delivered via the Internet or intranet using a computer, while
CBT is delivered face-
to-face via an instructor. In addition, WBT allows trainees to
learn at their own
pace at a convenient time and location. On the other hand, CBT
is typically group-
based requiring that trainees meet their instructor and peers at a
fixed time and
location.
Following these differences, researchers (Curtain 2002; Jung
and Rha 2000;
Rumble 2001; Trentin 2000; Whalen and Wright 1999) have
suggested that
evaluating WBT programmes also needs to be approached
differently from the
evaluation of CBT programmes. They have found that there are
more difficulties and
challenges to evaluating WBT programmes than classroom-
based training, due to
the wide variation in WBT forms. The evaluation of WBT is
influenced by a number
of factors, such as IT infrastructure used, the number of
learners, flexibility of
learning times, frequency of course revision, type and amount
of media used, type
and amount of learner support, type and amount of interaction,
learners’ experience
level with technology, or unexpected technical issues.
For example, WBT structures change at an exceptional pace,
since
technologies have been changing so quickly. Consequently, it
can be difficult to
thoroughly understand the range of different types and
organizational settings in
which WBT is delivered among organizations (Bradsford,
Brown, and Cocking
1999). In addition, when an organization provides different
types of online
courses that require different preparation times and levels of
online interaction, it
is more difficult to evaluate the level of workload and extra
time spent by
administrative staff and instructors to support online courses
(Curtain 2002;
Rumble 2001).
However, despite calls for the evaluation of WBT programmes
to be conducted
differently, there is limited information available on whether
HRD professionals
actually follow this suggestion in practice. Strother (2002)
suggested that useful and
sound comparative study about evaluating web-based and CBT
has been extremely
limited over the years, since evaluating WBT programmes has
not been treated
differently. Wisher and Olson (2003) also pointed out the lack
of any empirical
studies comparing the evaluation of WBT and CBT programmes.
They reviewed
over 500 articles published between 1996 and 2002, and found
that most of
these studies dealt with design issues or technology concerns
rather than with
strategies or tools for successful evaluation. Hence, due to an
insufficient number of
previous evaluation studies, they had difficulty assessing
evaluation practices in their
survey.
If WBT and CBT approaches differ from each other in important
ways, which
have implications on how these training approaches should be
evaluated, and, if
there is limited information on whether organizations actually
differ in their
evaluation practices related to these training approaches, then
more should be
known about how HRD professionals actually evaluate their
web-based and CBT
programmes. Understanding the practices used by HRD
professionals to evaluate
different training approaches will help us identify elements to
evaluate that are
80 Y.J. Kong and R.L. Jacobs
missing from current practice, examine enablers and disablers
of the evaluation of
these elements, and conduct appropriate evaluations of them.
The purpose of the study is to compare the practices used by
HRD professionals
to evaluate web-based and CBT programmes within seven
Korean companies. More
specifically, this study (1) examines how HRD professionals
evaluate web-based and
CBT programmes in their organizations, and (2) compares the
differences between
these training approaches. Also, this study (3) identifies the
barriers that might
prevent HRD professionals from evaluating web-based and CBT
programmes, and
(4) compares the differences between these training approaches
with the evaluation
barriers. Finally, this study (5) explores the important decision
factors that HRD
professionals use to determine how to evaluate web-based and
CBT programmes in
their organizations.
Literature review
The purpose of evaluation
Kraiger (2002) suggests three primary reasons to evaluate
training programmes:
decision-making, feedback, and marketing. The first purpose for
conducting
evaluation is to provide input for making decisions about
training programmes,
such as course retention, course revision, or personnel decisions
(e.g. quality of
instructor). The second purpose for evaluation is to provide
feedback to course
designers, trainers, or the trainees themselves that would allow
them to design or
engage with the course more effectively. The final purpose for
collecting training
evaluation data is for the purpose of marketing the training
programmes (e.g.
demonstrating the value of training to upper management or
helping future sponsors
or trainees understand the beneficial changes resulting from
training).
Training evaluation models
On reviewing the work on training evaluation in HRD literature,
Kirkpatrick’s
four-level framework and the CIPP model emerge as
particularly common in
organizations.
Kirkpatrick’s four-level framework
Kirkpatrick’s evaluation is the oldest and most widely used
evaluation criteria
in organizations. Many of the other evaluation models found in
the literature
are designed and revised based on Kirkpatrick’s framework
(Russ-Eft and Preskill
2001).
The Kirkpatrick framework consists of a series of four levels:
reaction, learning,
behaviour, and results (1998). In this model, reactions refer to
trainees’ levels of
satisfaction, measured via surveys or interviews designed to
identify their perceptions
of the training programme. Reaction is measured during or
immediately after
training. The next level of evaluation, learning, refers to the
extent of trainees’
changes in knowledge, skills, and attitudes. Learning is
measured before and after
training. The next level of evaluation is behaviour. Behaviour
refers to the extent of
trainees’ behaviour changes in transferring knowledge, skills,
and attitudes to their
jobs. Behaviour is measured after training. The final level of
evaluation is results.
A result refers to organizational results or changes because of
enhanced behaviours.
Human Resource Development International 81
In Kirkpatrick’s framework, reactions to training are related to
learning, learning is
related to behaviour, and behaviour is related to results.
However, Kirkpatrick’s framework has been routinely criticized
or revised in the
HRD literature. For example, Alliger and Janak (1989) question
its hierarchical
evaluation approach. According to their critiques, reactions to
training should be
viewed as unrelated to learning, and learning can have a direct
influence on results
criteria as opposed to only indirect influence through
behavioural change. In
addition, Holton (1996) argues that Kirkpatrick’s evaluation is
nothing more than a
taxonomy of outcomes. He removes reactions from the level of
primary outcomes of
training and defines it as a moderating variable between
trainees’ motivation to learn
and their actual learning. In addition, he states that learning is
related to transfer and
transfer is related to results.
The CIPP model
Another of the most frequently used evaluation models for
training in organizations
is the systems-based model (Phillips 1997). Under the systems
approach,
Stufflebeam’s (1971) CIPP model is the best known
(Fitzpatirck, Sanders, and
Worthen 2004). He points out that, since evaluation should be
informative to
decision makers, the systems approach can be effectively
applied to evaluation. The
CIPP model acronym stands for the four components of training
systems: context,
input, process, and output.
The following are the main characteristics of these four types of
evaluation:
(1) Context evaluation involves obtaining information about the
organizational
context to determine training needs and to identify objectives
for the
programme.
(2) Input evaluation involves determining the availability of
resources, alter-
native programme strategies, and plans for instructional design
process.
(3) Process evaluation is assessing the plans and barriers for
implementing the
programme and monitoring procedural events and activities.
(4) Product evaluation includes making judgments about the
outcomes of the
programme related to their objectives, context, input, and
process informa-
tion, interpreting their worth and merit.
Evaluating web-based and CBT based on a systems approach
In practice, Jacobs (2003) provides an interactive process of
evaluation from the
systems view for evaluating structured on-the-job training
(SOJT). In his book, he
points out that this systems view of training evaluation enables
HRD professionals
to see all the relevant elements separately, allowing them to
evaluate training
programmes, and how the elements bring together each
component. In addition, it
enables HRD professionals to examine the relationship among
the four components
within a system (e.g. SOJT) and with other systems in an entire
system or an
organization, allowing them to determine the value of each
component and of the
entire system.
As a result, Jacobs suggests that the evaluation questions for
SOJT should be
asked based on system inputs, processes, and outputs. The
questions also need to be
asked about the organizational context in which the training
takes place. For
82 Y.J. Kong and R.L. Jacobs
example, the training input questions include units of material
to be learned, the
training locations, and the characteristics of trainers and
trainees at the time of the
training. The training process questions include the amount of
time required to
complete the training, the availability of training resources, and
the behaviours of
the trainer and trainee during the training. The training output
questions include
whether the training goals have been achieved, whether the
training met the needs of
trainees and whether the training is more effective or efficient
than other training
approaches in terms of financial benefits after the training.
Lastly, the organizational
context questions include the extent of management
commitment to the use of the
training and the interaction of the training with other systems in
the organization.
The evaluation of SOJT from a systems view can be applied to
the evaluation of
other training approaches such as web-based and CBT, because
they can also be
viewed as systems themselves and, at the same time, as one of
many subsystems in a
larger system (i.e. the organization). Following the systems
approach proposed by
Jacobs (2003), Figure 1 explains how to evaluate WBT
programmes; Figure 2
illustrates how to evaluate CBT programmes.
Barriers to training evaluation
Previous research found that organizations do not actually
conduct evaluation well
(Curtain 2002; Holton and Naquin 2005; Jung and Rha 2000;
Kirkpatrick 1998;
Figure 1. The evaluation of WBT programmes from a systems
view.
Human Resource Development International 83
Kraiger 2002; Rumble 2001; Whalen and Wright 1999). The
primary barriers
inhibiting evaluation proposed by these authors are:
(1) Lack of expertise in evaluation: lack of knowledge and
skills or lack of
experience in evaluation
(2) Lack of organizational support for evaluation: unavailability
of resources,
organizational confidentiality practices or policies, fear of
negative financial
return, and limited budget and time
(3) Lack of evaluation methods and tools: evaluation activities
limited to
reaction sheets and statements of learning outcome, lack of
common cost
framework, and methodological limitations of financial returns
measurement
Three barriers affecting evaluation
Previous studies have endeavoured to find appropriate ways of
conducting effective
evaluation (Hofstetter and Alkin 2003; Johnson 1998; Leviton
2003; Shulha and
Cousins 1997). Taut and Alkin (2003) propose that this body of
research can be
addressed to oppose the barriers that inhibit evaluation. In their
study, they
examined the perception of project staff on barriers that
inhibited evaluation
implementation and its eventual use, based on Alkin’s (1985)
evaluation framework.
In his evaluation framework, Alkin (1985) divides factors
affecting the conduct of
Figure 2. The evaluation of CBT programmes from a systems
view.
84 Y.J. Kong and R.L. Jacobs
evaluation and its use into three categories: human factors,
context factors, and
evaluation factors.
First, ‘human factors describe evaluator and user
characteristics’ (Taut and Alkin
2003, 251). For example, human factors include evaluators’
commitment to facilitate
and enhance evaluation, evaluator’s willingness to actively
involve stakeholders
in the evaluation process, evaluator’s competence and reliability
in evaluation,
stakeholders’ organizational positions, stakeholders’ experience
levels to help the
evaluator and their interest in the evaluation, stakeholders’
commitment to use
evaluation results, and so on.
Second, ‘context factors refer to the context in which the
evaluated program
exists’ (Taut and Alkin 2003, 251). For example, context factors
include written
requirements (e.g. federal/state requirements and organizational
policies), fiscal
constraints (e.g. the amount of money and time available for the
evaluation), the
context outside of the actual training programme and the
organization
(e.g. external funding agencies), and programme characteristics
(e.g. age/maturity
of the programme, innovativeness of the programme, and
overlap with other
projects).
Third, evaluation factors involve the evaluation itself. For
example, evaluation
factors include evaluation procedures (e.g. the appropriateness
and accuracy of the
methods used and use of a general model to guide evaluation),
the information
dialogue (e.g. amount and quality of interaction between the
evaluator and
stakeholders), substance of evaluation information (e.g.
information relevancy and
specificity), and evaluation-reporting factor (e.g. timing of
information, style of oral
presentations, and format of reports).
Methods
This study employed a mix of quantitative and qualitative non-
experimental
research. Two data sets were used for the study. One set of data
was gathered from a
survey questionnaire distributed to Korean HRD professionals
via online in October
of 2008 to compare the differences between web-based and CBT
programmes in
evaluation practices.
The other data set was gathered from interviews with HR/HRD
directors in
February of 2009 to explore the research questions in more
depth. The researcher
adapted Jacobs’ evaluation framework (2003) that was based on
the systems
approach to four components of evaluation and Alkin’s
framework (1985) for
evaluation barriers. Then, she developed new construct scales
for a survey
questionnaire and interview questions based on these
established constructs to fit
this study’s purpose.
Organization selection
The researcher selected seven companies in Korea where web-
based and CBT
programmes were implemented and evaluated. The seven
companies selected
represent a variety of industries in Korea, including food,
chemical manufacturing,
automobile manufacturing, financial services, IT services, and
educational services
industries.
The total number of employees in each organization varies from
400 to 75,000,
and the total revenue of the companies’ ranges, as of 2007, from
$1.5 million to
Human Resource Development International 85
$30 billion. All the companies are headquartered in Seoul,
Korea, with branch offices
in many regional cities across the nation and around the world.
Survey to HRD professionals
Participant selection
The population for this study was all HRD professionals
(N¼147) who currently
evaluated web-based and/or CBT programmes within the seven
companies in Korea,
and the total number of respondents was 73. Out of the 73
respondents, 69
respondents had experience in the evaluation of CBT
programmes, whereas 43 had
experience in the evaluation of WBT programmes.
Instrument design
For the survey questionnaire, the researcher designated one
independent variable
with two levels: web-based and CBT programmes. Two
dependent variables were
measured in this study: the four components of evaluation and
three factors of
evaluation barriers. Six demographic characteristics were also
measured in this
study: type of industry, size of organization, an organization’s
years of conducting
training programmes, employee’s job level, employee’s
educational level, and
percentage of employee’s work that is related to training
evaluation.
Respondents were asked to identify the extent to which input,
process, output,
and organizational context components of evaluation were used
by HRD
professionals to evaluate their web-based and CBT programmes,
using a five-point
Likert-type scale ranging from 1 (never) to 5 (always), with
higher scores indicating
more frequent use of evaluation questions. Respondents were
also asked to assess the
degree of barriers to evaluating web-based and CBT
programmes based on three
factors affecting evaluation (i.e. human, context, and evaluation
factors), using a
five-point Likert-type scale ranging from 1 (strongly disagree)
to 5 (strongly agree),
with lower scores indicating more barriers perceived by HRD
professionals.
Validity and reliability
To ensure the instrument’s content validity, the panel of five
experts reviewed and
commented on the items in the instrument to ensure that they
accurately represented
the content area intended. Then, the English version of the
instrument was translated
to the Korean version by the researcher, and the Korean
versions were reviewed by
Korean Ph.D. students and HRD professionals to gain their
consensus on the items.
To ensure the reliability of the survey questionnaire,
Cronbach’s alpha coefficient
was employed to measure internal consistency of the items and
the alpha values of
the instrument were between .8 and .9.
Interviews with HR or HRD directors
Participant selection
To obtain qualitative data, the researcher conducted individual
interviews with HR
or HRD directors. The participants were all current directors in
the departments of
HR or HRD in the seven companies selected. Five out of seven
HR/HRD directors
agreed to participate in this study and were interviewed.
Interviews were conducted
86 Y.J. Kong and R.L. Jacobs
either phone, email, or face to face, whatever was most
convenient for them, and
supplementary follow-up phone calls were made, when
necessary.
Interview questions
The interview was composed of four main questions including
demographic
information. Respondents were asked about how web-based and
CBT programmes
were implemented and evaluated differently in their
organizations based on the four
components of evaluation. Respondents were also asked about
the factors that
inhibited the evaluation of web-based and CBT programmes and
whether there were
differences between these training approaches based on the
three factors of
evaluation barriers. Finally, they were asked what kind of
information they looked
for from the evaluations and how they used the results of the
evaluation in order to
explore the key decision factors determining evaluation for their
web-based and CBT
programmes.
Credibility and dependability
Maxwell (2005) states that, in qualitative research, credibility
refers to internal
validity, and the credibility in this study was enhanced based on
rich data and
control of bias. The researcher audio-taped and transcribed all
interviews word by
word and took field notes recording the interviewees’ actions
during interview. To
ensure consensus and reduce bias in this study, the verbatim
transcripts were shared
with interviewees or another researcher to ensure that the
researcher’s interpretations
of the interviewees’ statements were consistent with the
transcripts.
Lincoln and Guba (1985) state that, in qualitative research,
dependability closely
corresponds to the notion of reliability in quantitative research,
and ‘since there can
be no validity without reliability, a demonstration of validity is
sufficient to establish
reliability’ (p. 316).
Data analysis
First, the data from the survey questionnaire were analysed
using the Statistical
Package for the Social Science (SPSS). Descriptive statistics
such as frequency,
percentage, mean for population (m), and sigma for population
(s) were used to
describe the evaluation practices used by Korean HRD
professionals and to
compare the differences between web-based and CBT
programmes in the evaluation
practices. In addition, multiple regression analysis was
conducted to examine the
relationship between demographic variables of HRD
professionals and the evalua-
tion practices used for web-based and CBT programmes.
Second, the data from the
interview was read, coded, and analysed by the researcher to
seek a deeper
understanding of these relationships in the natural setting.
Results
Four components of evaluation
As shown in Tables 1 and 2, the results of the survey showed
that web-based and
CBT programmes were most frequently evaluated for the
process evaluation
component (mw¼4.02, mc¼4.29) followed by the input
evaluation component
Human Resource Development International 87
(mw¼3.70, mc¼3.78). On the other hand, web-based and CBT
programmes were
least frequently evaluated for the organizational context
evaluation component
(mw¼3.03, mc¼2.93), followed by the output evaluation
component (mw¼3.12,
mc¼3.46).
The results also showed that, the input evaluation component
produced a higher
mean score for CBT (m¼3.78, s¼0.59) than WBT programmes
(m¼3.70, s¼0.86).
The process evaluation component produced a higher mean
score for CBT (m¼4.29,
s¼0.56) than WBT programmes (m¼4.02, s¼0.75). The output
evaluation compo-
nent produced a higher mean score for CBT (m¼3.46, s¼0.74)
than WBT
programmes (m¼3.12, s¼0.86). On the other hand, the
organizational context
evaluation component produced a lower mean score for CBT
(m¼2.93, s¼0.93)
than WBT programmes (m¼3.03, s¼0.85).
The results of the interview (Table 3) produced the same results
as the survey;
web-based and CBT programmes were most frequently
evaluated for the process
evaluation component, whereas they were least frequently
evaluated for the
organizational context evaluation component.
These results also showed that, in the input and process
components of
evaluation, the five interviewees all said that their HRD
professionals used existing
survey forms to evaluate trainees’ satisfaction with training.
However, they said that
they included different questions for web-based and CBT
programmes on the survey
questionnaire, although the main categories for the survey were
the same. All five
interviewees stated that the online survey had more detailed
questions included on
the questionnaire than the survey for the classroom. In addition,
they mentioned that
WBT was more complicated to evaluate than CBT, because
web-based courses were
specialized and individualized.
Table 2. Descriptive statistics on CBT programs for four
components of evaluation (N¼69).
Cases
Included Excluded Total
Mean Std. deviation N Per cent N Per cent N Per cent
Input 3.78 .59 69 94.5 4 5.5 73 100.0
Process 4.29 .56 69 94.5 4 5.5 73 100.0
Output 3.46 .74 68 93.2 5 6.8 73 100.0
Context 2.93 .93 69 94.5 4 5.5 73 100.0
Table 1. Descriptive statistics on WBT programs for four
components of evaluation
(N¼43).
Cases
Included Excluded Total
Mean Std. deviation N Per cent N Per cent N Per cent
Input 3.70 .86 43 58.9 30 41.1 73 100.0
Process 4.02 .75 43 58.9 30 41.1 73 100.0
Output 3.12 .86 43 58.9 30 41.1 73 100.0
Context 3.03 .85 43 58.9 30 41.1 73 100.0
88 Y.J. Kong and R.L. Jacobs
Meanwhile, the HRD professionals at the five companies rarely
evaluated their
web-based and CBT programmes for the output evaluation
component, but they
more frequently evaluated WBT programmes than CBT
programmes for the
organizational context evaluation component in order to receive
reimbursement of
tuition money from the Korean government.
In summary, the results of both the survey and the interviews
showed that web-
based and CBT programmes were meaningfully different on the
process evaluation
component. However, web-based and CBT programmes were not
meaningfully
different on the input, output, and organizational context
components of evalua-
tion. The results of both the survey and the interviews also
showed that the
demographic characteristics of HRD professionals were not
meaningfully related to
the practices used to evaluate web-based and CBT programmes
for each evaluation
component.
Three factors of evaluation barriers
As shown in Tables 4 and 5, the results of the survey showed
that the respondents
perceived more barriers in the context (mw¼3.17, mc¼3.11) and
evaluation factors
Table 3. The differences between web-based and CBT programs
for the four components of
evaluation.
Categories Web-based Classroom-based
Input evaluation
1. Survey questionnaire Different questions: adding
more questions
Different questions
Process evaluation
1. Survey questionnaire Different questions: adding
more questions
Different questions
2. Trainees’ participation
and learning progress
Checking through
online:
Checking through
direct observation:
. Number of clicking slides . Class
. Login times . CC-TV
. Discussion board . Interview
. Pop-up window
. Homework
. Homework or
discussion in class
Output evaluation
Self-check Self-check
1. Learning tests Multiple choices Multiple choices
Pass/fail Pass/fail
Simple essay Essay
Reflection paper Reflection/project paper
2. Transfer of training Action plan Projects
Supervisor’s observation Action plan
Number of sales Supervisor’s observation
360 degree feedback Number of sales
Statistics 360 degree feedback
Statistics
Organizational context evaluation
1. Survey questionnaire Little difference Little difference
2. Korean government
support
Report evaluation results to
the government
N/A
Human Resource Development International 89
(mw¼3.14, mc¼3.22) than in the human factors (mw¼3.68,
mc¼3.81) to evaluating
their web-based and CBT programmes. The results also showed
that the human
factors produced a lower mean score for WBT (m¼3.69,
s¼0.58) than CBT
programmes (m¼3.81, s¼0.55). When comparing web-based and
CBT programmes
for the context and evaluation factors of evaluation barriers, the
context factors
produced a higher mean score for WBT (m¼3.17, s¼0.74) than
CBT programmes
(m¼3.11, s¼0.73). The evaluation factors produced a lower
mean score for WBT
programmes (m¼3.14, s¼0.80) than CBT programmes (m¼3.22,
s¼0.71).
The results of the interviews (Table 6) produced the same
results as the survey;
the context and evaluation factors were the primary barriers
preventing HRD
professionals from evaluating web-based and CBT programmes.
Table 6 presents the
barriers that inhibited the evaluation of these training
approaches in each area –
human, context, and evaluation factors.
The interview results also showed that evaluation barriers were
slightly different
between web-based and CBT programmes in the area of human
factors due to the
lack of credibility for an external evaluator or of responsibility
of vendors for WBT
programmes. Most companies being studied here purchased
online courses from
professional vendors. Three interviewees stated that the vendors
did not take much
responsibility for further evaluation after the training was
completed. One
interviewee also stated that his senior management did not trust
the evaluation
results provided by an external evaluator, because they did not
believe that an
external evaluator knew their organization’s characteristics and
circumstances better
than their own staff.
In addition, evaluation barriers were slightly different between
web-based and
CBT programmes in the area of context factors due to the
unique structure of web-
based courses. Based on interviews with HR/HRD directors,
most of their HRD
Table 4. Descriptive statistics on three factors of evaluation
barriers for WBT programs
(N¼40).
Cases
Included Excluded Total
Mean Std. deviation N Per cent N Per cent N Per cent
Human 3.69 .58 40 54.8 33 45.2 73 100.0
Context 3.17 .74 40 54.8 33 45.2 73 100.0
Evaluation 3.14 .80 40 54.8 33 45.2 73 100.0
Table 5. Descriptive Statistics on three factors of evaluation
barriers for CBT programs
(N¼67).
Cases
Included Excluded Total
Mean Std. deviation N Per cent N Percent N Per cent
Human 3.81 .55 67 91.8 6 8.2 73 100.0
Context 3.11 .73 67 91.8 6 8.2 73 100.0
Evaluation 3.22 .71 67 91.8 6 8.2 73 100.0
90 Y.J. Kong and R.L. Jacobs
professionals needed to add more questions to the survey for
their WBT
programmes. Finally, all interviewees pointed out that there was
little difference
between web-based and CBT programmes in the evaluation
factors of evaluation
barriers. HRD professionals perceived that barriers in current
evaluation methods
and tools, especially for evaluating trainees’ performance and
financial benefits, as
apply to both web-based and CBT programmes.
In summary, the survey and interview results both showed that
web-based and
CBT programmes were not meaningfully different on the three
factors of evaluation
barriers. The results of both the survey and interviews also
showed that the
demographic characteristics of HRD professionals were not
meaningfully related to
the barriers to evaluating web-based and CBT programmes.
Key decision factors for determining evaluation
After interviewing five HR/HRD directors, the primary decision
factors determining
evaluation for web-based and CBT programmes were identified
as follows: senior
management’s need, development of current or new training
programmes, type of
training programme, political issues for online courses,
personnel benefits, and
budget for training programmes.
Senior management’s needs
The first decision factor determining evaluation was the needs
or requests of senior
management for the evaluation of web-based and CBT
programmes. All interviewees
Table 6. The differences between web-based and CBT programs
in the three factors of
evaluation barriers.
Categories Web-based Classroom-based
Human factor
1. Lack of professional knowledge and skills No difference No
difference
2. Lack of credibility & responsibility of an
external evaluator or vendors
More barriers
perceived
N/A
3. Lack of HRD professionals’ commitment to
evaluation
No difference No difference
4. Lack of stakeholders’ commitment to
evaluation
No difference No difference
Context factor
1. Lack of senior management’s interest in
training evaluation
No difference No difference
2. Lack of support from organizations
(a) Limited budget for evaluation No difference No difference
(b) Limited time for evaluation No difference No difference
3. The unique structure of web-based course Barriers perceived
Not perceived
4. Organizational characteristics and
circumstances
No difference No difference
Evaluation factor
1. Lack of evaluation methods and tools
(a) Outdated, unreliable survey tools No difference No
difference
(b) Few reliable evaluation methods and
tools available to evaluate trainees’
performance and organizational results
No difference No difference
Human Resource Development International 91
said that this factor was the most powerful factor in how HRD
professionals evaluated
their web-based and CBT programmes. The senior management
mostly asked HRD
professionals to use surveys to measure trainees’ satisfaction
with training and their
participation rates in classes.
However, sometimes, members of senior management asked
HRD professionals
to evaluate trainees’ knowledge and skills, especially when
trainees needed to acquire
certifications. In addition, they occasionally requested
information on trainees’ job
performance as a result of training.
Development of current or new training programmes
The second decision factor determining decisions about training
evaluation was
deciding whether to improve current training programmes or
develop new training
programmes. Companies interviewed identified and analysed
problems with
current training programmes based on satisfaction rates of
trainees primarily
through surveys. If training satisfaction was low, then they
revised current training
programmes or developed new training programmes.
Type of training programme
The third decision factor for training evaluation depended on
what type of training
programme was offered. Generally, WBT was purchased from
vendors for awareness
or managerial training, whereas CBT was provided from within
the organization for
the purpose of technical training. All interviewees said that they
generally evaluated
knowledge tests, supervisor’s observations, and trainees’
satisfaction, and checked
attendance rates of trainees for technical training programmes.
On the other hand,
they generally provided self-assessment or evaluated trainees’
satisfaction for
awareness and managerial training programmes.
Political issues for online courses
All interviewees talked about political issues related to their
online courses. The politics
related to online courses were mainly determined by the
relationships between the
Korean government and organizations and between
organizations and their HRD
professionals. First, interviewees discussed the relationship
between the government
and their organizations. The Korean government has encouraged
companies to provide
their employees with WBT programmes, reimbursing companies
for training costs.
Since companies can reduce the cost of training through going
online, they endeavour to
meet the government’s evaluation criteria for WBT by
reporting, for example, the
completion rates and satisfaction rates of trainees.
Second, the relationship between an organization and its HRD
professionals
worked to accomplish the intended goals of HRD professionals.
For example, one of
the interviewees stated that HRD professionals sometimes
needed to purchase new
equipment to enable the delivery of multimedia course content
or replace old
equipment or learning management systems with new ones.
Then, when they
evaluated trainees’ satisfaction with current online courses, they
included evaluation
of their IT infrastructure, online servers, or other equipments
that they needed to
buy or replace, and reported the results to management to
achieve their intended
goals in delivering training programmes.
92 Y.J. Kong and R.L. Jacobs
Personnel benefits
The fifth decision factor involved how to evaluate employees’
performance related to
training. Three interviewees stated that they evaluated their
employees’ performance
to measure personnel benefits. The interviewees stated that they
evaluated HRD
professionals regarding how they provided their employees with
training pro-
grammes and whether they increased the purchase of online
courses from out-
sourcing companies within their organization’s budgets. In
addition, the interviewees
mentioned that they evaluated trainees regarding what training
programmes they
selected, how they took their training programmes, and whether
they completed the
training hours required for their job promotions.
Budget for training programmes
The last decision factor for training evaluation involved setting
up the budget for
implementing future training programmes. Two interviewees
said that they
conducted surveys to set up budgets for upcoming web-based
and CBT programmes.
Discussion
The premise of the study was that the evaluation of different
training
approaches, such as classroom-based and WBT, should be
planned and performed
differently because of different learning environments (Curtain
2002; Jung and Rha
2003; Rumble 2001; Whalen and Wright 1999). In general, the
results of the study
did not support the premise of the study, except for the process
component of
evaluation. The results did show some differences between web-
based and CBT
programmes in evaluation practices, but they were not judged to
be meaningfully
different.
Six points might be considered in explanation of the results of
the study: (1) lack
of management interest in evaluation, (2) uncertain workforce
development policy,
(3) possible confusion between the input and process evaluation
components,
(4) lack of knowledge and skills among HRD professionals, and
(5) uncertainty
about the original premise.
Lack of interest of senior management in evaluation
The first point suggests that senior management may not have
expressed sufficient
interest in all of the evaluation components. All five
interviewees reported that, when
evaluating web-based and CBT programmes, senior managers
were most interested
in trainees’ satisfaction with training and their participation
rates in classes. These
areas are most often categorized as part of the process
component of training
evaluation. Managers may have assumed that training has its
own effectiveness and,
thus, the most important goal of training is to have employees
take classes. They also
appeared to believe that trainees’ satisfaction equalled
successful training and, thus,
evaluating trainees’ satisfaction was the most important.
As a result, they asked HRD professionals in effect to focus on
the process
evaluation component, when evaluating their web-based and
CBT programmes and,
thus, the differences between these training approaches may be
clearly revealed in the
process evaluation component, an area of evaluation performed
directly to satisfy
their senior management’s goals.
Human Resource Development International 93
Uncertain workforce development policy
Another possible explanation for the results involves the Korean
government’s
policies regarding how organizations are reimbursed for the
costs of WBT
programmes. The interviewees stated that, in order to receive
this funding, their
organizations must conduct evaluation of WBT programmes,
according to
evaluation criteria set by the government, and report the results.
The criteria
required for evaluating WBT programmes include primarily
completion rates and
satisfaction rates of trainees.
Thus, to receive government support, organizations might
request their HRD
professionals to frequently evaluate their WBT programmes for
the process
evaluation component. Accordingly, HRD professionals may be
more familiar
with how to appropriately evaluate the training approaches for
this component than
for the other three components and, thus, they might evaluate
web-based and CBT
programmes differently for this component.
Possible confusion between input and process evaluation
components
The third point suggests that there might be some
misunderstanding about the input
and process evaluation components. The results showed that
web-based and CBT
programmes were more frequently evaluated for the input
evaluation component
than for the output and organizational context components of
evaluation. However,
it can hardly be concluded that the input component was as
frequently evaluated by
HRD professionals.
The reason for the observed frequency in conducting input
evaluation, based
on the interview results, is that the input evaluation component
was often
considered to be a part of the process evaluation component,
which means that
it involved evaluating trainees’ satisfaction with training rather
than intentionally
conducting separate input evaluation. Thus, the results suggest
that web-based
and CBT programmes were basically evaluated for the process
component,
and that there are some differences between these training
approaches in this
component.
Lack of HRD professionals’ knowledge and skills
The results might also be explained by the fundamental lack of
HRD professionals’
knowledge and skills in training evaluation. The results of both
the survey and the
interviews showed that HRD professionals perceived fewer
barriers in the human
factors than in the context and evaluation factors. The interview
results further
explained that since HRD professionals mostly used pre-
existing survey forms for
20–30 years to evaluate their training programmes, which were
not difficult to use,
they may not be aware of their own lack of knowledge and
expertise in the
evaluation of both web-based and CBT programmes. Thus, they
may not perceive
any differences between web-based and CBT programmes in the
human factors of
evaluation barriers.
In addition, current HRD professionals can have difficulty of
performing higher-
level evaluation with currently available evaluation methods
and tools, thus, those in
the study may rarely conduct this kind of evaluation for either
training approach.
Accordingly, they may not be able to recognize any differences
between web-based
and CBT programmes in evaluation barriers.
94 Y.J. Kong and R.L. Jacobs
Uncertain original premise
Finally, the original premise – that the evaluation of different
training approaches
should be performed differently – might be not as strong as
initially assumed. Many
researchers (Jung and Rha 2000; Rumble 2001; Strother 2002;
Trentin 2000; Whalen
and Wright 1999) have emphasized that different training
approaches should be
evaluated differently, and, following this premise, that the
evaluation of web-based
and CBT should be conducted differently.
However, other researchers (Clark 2000; Garrison and Vaughan
2008) argue that
the differences between training approaches such as web-based
and CBT might be
due to quality of course design, quality of the instructor, or
potential individual
differences among instructors rather than the delivery modes
themselves. Garrison
and Vaughan (2008) stated that sound learning theories such as
constructivism and
social cognition for learning may be more important than
delivery modes for
ensuring effective learning. Thus, as they suggested, effective
teaching and learning
principles should be considered more than the unique elements
of delivery mode,
when comparing the evaluation of different training approaches.
Contribution to HRD
This study contributes to both the theory and practice of
evaluation for training
programmes in the HRD field. Theoretically, employing a
methodology that
combines quantitative and qualitative elements, this study
contributes to empirical
research that has been conducted rarely in the HRD field. That
is, the study provides
deeper understanding of how organizations actually evaluate
web-based and CBT,
identifying evaluation barriers to the different training
approaches, and how unique
programme features and organizational and social circumstances
result in different
evaluation methods for web-based and CBT programmes.
In addition, this study contributes to our understanding of the
dynamic
interactions among people who have different roles and
responsibilities during an
evaluation process, examining how evaluators respond
differently to evaluate their
web-based and CBT programmes.
This study also provides some recommendations for further
research. First, this
study found that the lack of reliable evaluation methods and
tools was one of the
most prevalent reasons why organizations were not committed
to evaluating their
training programmes and avoided higher levels of training
evaluation. Consequently,
more active research is needed to develop evaluation methods
and tools that are easy
for practitioners to use and guide them to use these tools for
evaluating different
training approaches in appropriate ways.
Second, unlike evidence from existing literature, this study
showed that WBT
programmes were more frequently evaluated than CBT
programmes for the
organizational context component of evaluation due to the
Korean government’s
support for online learning. Since this study is limited to
examining seven companies
in one country, Korea, more comparison studies among various
countries and
companies will be needed to determine to what degree a
particular country or
organization’s characteristics, culture, and circumstances
influence different training
evaluation approaches.
Practically, the results of this study have three important
implications for HRD
professionals to improve current training evaluation in their
organizations. First,
this study found that the primary barrier was the context factors
due to lack of
Human Resource Development International 95
interest in and commitment to training evaluation on the part of
senior management.
Thus, senior management must deliberate thoughtfully on how
to manage and
maintain current training evaluation effectively.
For example, HRD professionals need to revise current
personnel policy and
benefits related to training evaluation, focusing on ways of
improving the quality of
training programmes rather than increasing the quantity of
online courses. Since
organizations evaluate their HRD professionals’ performance
based on the number
of online courses purchased from vendors with the cheapest
prices, HRD
professionals concentrate more on increasing the quantity of
courses rather than
on increasing the quality of those courses. This approach can be
hardly guaranteed
to deliver good quality training courses to employees.
Second, this study found that most participants were unaware of
evaluation
barriers that came from the lack of HRD professional’s
knowledge and skills in
training evaluation, since they mostly used and reused existing
survey templates to
evaluate trainees’ satisfaction for both web-based and CBT
programmes. Thus,
HRD professionals should improve their job competency in the
area of current
training evaluation in the workplace.
For example, HRD professionals need to reflect on the adequacy
of their current
evaluation practices. They should take stock of what they
currently do in the
evaluation of their web-based or CBT programmes and examine
whether they meet
the basic requirements to successfully implement these training
approaches. They
also need to identify different requisite elements and barriers to
evaluate web-based
and CBT programmes and implement an appropriate evaluation
for the different
training approaches. In addition, they should examine and
evaluate how various
stakeholders correspond to multiple goals of a specific training
programme to help
select the best training approaches for their organizations.
Professional development
of training evaluation is a great opportunity for HRD
professionals to improve these
job competencies.
Finally, national evaluation policy makers in Korea should
provide organiza-
tions with effective guidance and standards to conduct
appropriate evaluation for
web-based and CBT programmes. Especially, they should be
aware of the negative
results related to their support for online courses. They should
accurately investigate
actual prices of online courses in the market and re-establish
current evaluation
criteria in order to reduce the negative practice of focusing on
the quantity of
online courses with cheapest prices rather than the quality of
the courses. They
should also request diverse evaluation reports that go beyond
trainees’ satisfaction
with training to help organizations determine the best training
approaches for their
employees.
References
Alkin, M. 1985. A guide for evaluation decision makers.
Thousand Oaks, CA: Sage.
Alliger, G., and E. Janak. 1989. Kirkpatirck’s levels of training
criteria: Thirty years later.
Personnel Psychology 42, no. 2: 331–42.
Bartley, S., and J. Golek. 2004. Evaluating the cost
effectiveness of online and face-to-face
instruction. Educational Technology & Society 7, no. 4: 167–75.
Bassi, L., G. Benson, and S. Cheney. 1996. The top ten trends.
Training and Development 50,
no. 11: 28–42.
Bradsford, J., A. Brown, and R. Cocking. 1999. How people
learn: Brain, mind experience and
school. Washington, DC: National Academy Press.
96 Y.J. Kong and R.L. Jacobs
Clark, R. 2000. Evaluating distance education: Strategies and
cautions. Quarterly Review of
Distance Education 1, no. 1: 3–16.
Curtain, R. 2002. Online learning: How cost-effective? In
Online delivery in the vocational
education and training sector: Improving cost effectiveness, ed.
Richard Curtain, 125–43.
Leabrook, SA, Australia: Australian National Training
Authority, NCVER.
Fitzpatrick, J., J. Sanders, and B. Worthen. 2004. Program
evaluation: Alternative approaches
and practical guidelines. Boston, MA: Pearson Education, Inc.
Garrison, R., and N. Vaughan. 2008. Part one: Community of
inquiry framework. In Blended
learning in higher education, ed. Garrison R. and V. Vaughan,
1–68. San Francisco, CA:
Jossey-Bass.
Hofstetter, C., and M. Alkin. 2003. Evaluation use revisited. In
International Handbook of
Educational Evaluation, ed. T. Kellaghan and D.L. Stufflebeam,
197–222. London, Great
Britain: Kluwer Academic Publishers.
Holton, E. 1996. The flawed four-level evaluation model.
Human Resource Development
Quarterly 7, no. 1: 5–21.
Holton, E., and S. Naquin. 2005. A critical analysis of HRD
evaluation models from
a decision-making perspective. Human Resource Development
Quarterly 16, no. 2: 257–
80.
Jacobs, R. 2003. Structured on-the-job training: Unleashing
employee expertise in the workplace.
San Francisco, CA: Berrett-Koehler.
Johnson, R. 1998. Toward a theoretical model of evaluation
utilization. Evaluation and
Program Planning 21, no. 1: 93–110.
Jung, I.S., and I. Rha. 2000. Effectiveness and cost-
effectiveness of online education: A review
of literature. Education Technology 40, no. 4: 57–60.
Kapp, K., and C. McKeague. 2002. Blend two proven training
methods to improve results.
Chemical Engineering 109, no. 8: 191–4.
Kirkpatrick, D.L. 1998. Evaluating training programs: The four
levels. San Francisco, CA:
Berrett-Koehler, 1998.
Kraiger, K. 2002. Decision-based evaluation. In Creating,
implementing, and managing
effective training and development systems in organizations:
State-of-the-art lessons for
practice, ed. Kurt Kraiger, 331–75. San Francisco, CA: Jossey-
Bass.
Leviton, L. 2003. Evaluation use: Advances, challenges and
applications. American Journal of
Evaluation 24, no. 4: 525–35.
Lincoln, Y.S., and E.G. Guba. 1985. Naturalistic Inquiry.
Beverly Hills, CA: Sage
Publications, Inc.
Maxwell, J. 2005. Qualitative research design: An interactive
approach. Thousand Oaks, CA:
Sage Publications, 2005.
Olson T., and R. Wisher. 2002. The effectiveness of web-based
instruction: An initial inquiry.
International Review of Research in Open and Distance
Learning 3, no. 2.
Phillips, J. 1997. Handbook of training evaluation and
measurement methods. Houston, TX:
Gulf Publishing.
Rumble, G. 2001. The costs and costing of networked learning.
Journal of Asynchronous
Learning Network 5, no. 2: 75–96.
Russ-Eft, D., and H. Preskill. 2011. Evaluation in organizations.
Cambridge, Massachusetts:
Perseus Publishing.
Shulha, L., and J. Cousins. 1997. Evaluation use: Theory,
research, and practice since 1986.
Evaluation Practice 18, no. 3: 195–208.
Strother, J. 2002. An assessment of the effectiveness of e-
learning in corporate training
programs. The International Review of Research in Open and
Distance Learning 3, no. 10.
Stufflebeam, D., W. Foley, W. Gephart, L. Hammond, H.
Merriman, and M. Provus. 1971.
Educational evaluation and decision-making in education.
Itasca, IL: Peacock.
Taut, S., and M. Alkin. 2003. Program staff perceptions of
barriers to evaluation
implementation. American Journal of Evaluation 24, no. 2: 213–
26.
Trentin, V. 2000. The evaluation of online courses. Journal of
Computer Assisted Learning 16,
no. 3: 259–70.
Whalen, T., and D. Wright. 1999. Methodology for cost-benefit
analysis of web-based tele-
learning program: Case study of the bell institute. The
American Journal of Distance
Education 13, no. 1: 24–44.
Human Resource Development International 97
Wisher, R.A., and T.M. Olson. 2003. The effectiveness of web-
based instruction. Alexandria,
VA: United States Army Research Institute for the Behavioural
and Social Sciences
Publication.
Yanmil, S., and G. McLean. 2001. Theories supporting transfer
of training. Human Resource
Development Quarterly 12, no. 2: 195–208.
98 Y.J. Kong and R.L. Jacobs
Copyright of Human Resource Development International is the
property of Routledge and its content may not
be copied or emailed to multiple sites or posted to a listserv
without the copyright holder's express written
permission. However, users may print, download, or email
articles for individual use.
Task 1: half Page, times new roman, 12, double spaced.
What is the difference between Research and Evaluation?
How do you see the role of evaluation in organizations?
What are the ethical issues in evaluation practice?
Task 2: Reflection of the attached article(Times, new roman,
12, double spaced). The reflection should be 400 words and
should include the following:
1- Section 1: Rationale for choosing the article
2- Section 2: Critique and connection to the course(course
name: Evaluation models in Human Resource development)
3- Section 3: Article Reference (and other references if used
in the Critique)

More Related Content

Similar to A comparison of the practices used by human resource developme.docx

Similarities Between Structured Situational And....pdfSimilarities Between St...
Similarities Between Structured Situational And....pdfSimilarities Between St...Similarities Between Structured Situational And....pdfSimilarities Between St...
Similarities Between Structured Situational And....pdfSimilarities Between St...Kimberly Brooks
 
Apt chapter 12 summative evaluation
Apt chapter 12 summative evaluationApt chapter 12 summative evaluation
Apt chapter 12 summative evaluationReginald Smith
 
Opportunities and Challenges Inherent In Technical and Vocational Schools; Co...
Opportunities and Challenges Inherent In Technical and Vocational Schools; Co...Opportunities and Challenges Inherent In Technical and Vocational Schools; Co...
Opportunities and Challenges Inherent In Technical and Vocational Schools; Co...inventionjournals
 
PEER RESPONSESWeek 3 - Discussion 1Kirkpatricks Taxo.docx
PEER RESPONSESWeek 3 - Discussion 1Kirkpatricks Taxo.docxPEER RESPONSESWeek 3 - Discussion 1Kirkpatricks Taxo.docx
PEER RESPONSESWeek 3 - Discussion 1Kirkpatricks Taxo.docxdanhaley45372
 
An Organizational Development Framework For Assessing Readiness And Capacity ...
An Organizational Development Framework For Assessing Readiness And Capacity ...An Organizational Development Framework For Assessing Readiness And Capacity ...
An Organizational Development Framework For Assessing Readiness And Capacity ...Angie Miller
 
Evaluation Of Evaluation And Performance Measurement
Evaluation Of Evaluation And Performance MeasurementEvaluation Of Evaluation And Performance Measurement
Evaluation Of Evaluation And Performance MeasurementAmanda Burkett
 
Evaluation Of A Employee Evaluation System
Evaluation Of A Employee Evaluation SystemEvaluation Of A Employee Evaluation System
Evaluation Of A Employee Evaluation SystemJenny Richardson
 
Performance appraisal research design
Performance appraisal research designPerformance appraisal research design
Performance appraisal research designGT Imsr
 
THE IMPACT OF TRAINING AND DEVELOPMENT ON EMPLOYEES PERFORMANCE IN AN ORGAN...
THE IMPACT OF TRAINING AND  DEVELOPMENT ON EMPLOYEES  PERFORMANCE IN AN ORGAN...THE IMPACT OF TRAINING AND  DEVELOPMENT ON EMPLOYEES  PERFORMANCE IN AN ORGAN...
THE IMPACT OF TRAINING AND DEVELOPMENT ON EMPLOYEES PERFORMANCE IN AN ORGAN...Research Publish Journals (Publisher)
 
· Assignment 3 Creating a Compelling VisionLeaders today must be .docx
· Assignment 3 Creating a Compelling VisionLeaders today must be .docx· Assignment 3 Creating a Compelling VisionLeaders today must be .docx
· Assignment 3 Creating a Compelling VisionLeaders today must be .docxgerardkortney
 
Fundamentals of human resource management
Fundamentals of human resource managementFundamentals of human resource management
Fundamentals of human resource managementwww.Jobbazzar.com
 
Performance Appraisal of Human Resources in Anglo-Saxon Higher Education Inst...
Performance Appraisal of Human Resources in Anglo-Saxon Higher Education Inst...Performance Appraisal of Human Resources in Anglo-Saxon Higher Education Inst...
Performance Appraisal of Human Resources in Anglo-Saxon Higher Education Inst...ijtsrd
 
Independent evaluation of the united nations internship program results & pro...
Independent evaluation of the united nations internship program results & pro...Independent evaluation of the united nations internship program results & pro...
Independent evaluation of the united nations internship program results & pro...Cristobal Mingo
 
MODELLING TRAINING EFFECTIVENESS AND ITS IMPACT ON DEVELOPMENT OF MANAGERS IN...
MODELLING TRAINING EFFECTIVENESS AND ITS IMPACT ON DEVELOPMENT OF MANAGERS IN...MODELLING TRAINING EFFECTIVENESS AND ITS IMPACT ON DEVELOPMENT OF MANAGERS IN...
MODELLING TRAINING EFFECTIVENESS AND ITS IMPACT ON DEVELOPMENT OF MANAGERS IN...IAEME Publication
 
“The Effectiveness and Efficiency of the Various Graduate Recruitment Methods”
“The Effectiveness and Efficiency of the Various Graduate Recruitment Methods”“The Effectiveness and Efficiency of the Various Graduate Recruitment Methods”
“The Effectiveness and Efficiency of the Various Graduate Recruitment Methods”IOSRJBM
 
Module 4 Reflective DiscussionPreviousNextIn the Module 4 .docx
Module 4 Reflective DiscussionPreviousNextIn the Module 4 .docxModule 4 Reflective DiscussionPreviousNextIn the Module 4 .docx
Module 4 Reflective DiscussionPreviousNextIn the Module 4 .docxroushhsiu
 
Training need assessment in a 5star ho
Training need assessment in a 5star hoTraining need assessment in a 5star ho
Training need assessment in a 5star hoShamimansary
 
Running Head DATA SOURCE EVALUATION .docx
Running Head DATA SOURCE EVALUATION                              .docxRunning Head DATA SOURCE EVALUATION                              .docx
Running Head DATA SOURCE EVALUATION .docxtodd271
 

Similar to A comparison of the practices used by human resource developme.docx (20)

Similarities Between Structured Situational And....pdfSimilarities Between St...
Similarities Between Structured Situational And....pdfSimilarities Between St...Similarities Between Structured Situational And....pdfSimilarities Between St...
Similarities Between Structured Situational And....pdfSimilarities Between St...
 
Apt chapter 12 summative evaluation
Apt chapter 12 summative evaluationApt chapter 12 summative evaluation
Apt chapter 12 summative evaluation
 
Opportunities and Challenges Inherent In Technical and Vocational Schools; Co...
Opportunities and Challenges Inherent In Technical and Vocational Schools; Co...Opportunities and Challenges Inherent In Technical and Vocational Schools; Co...
Opportunities and Challenges Inherent In Technical and Vocational Schools; Co...
 
PEER RESPONSESWeek 3 - Discussion 1Kirkpatricks Taxo.docx
PEER RESPONSESWeek 3 - Discussion 1Kirkpatricks Taxo.docxPEER RESPONSESWeek 3 - Discussion 1Kirkpatricks Taxo.docx
PEER RESPONSESWeek 3 - Discussion 1Kirkpatricks Taxo.docx
 
An Organizational Development Framework For Assessing Readiness And Capacity ...
An Organizational Development Framework For Assessing Readiness And Capacity ...An Organizational Development Framework For Assessing Readiness And Capacity ...
An Organizational Development Framework For Assessing Readiness And Capacity ...
 
Evaluation Of Evaluation And Performance Measurement
Evaluation Of Evaluation And Performance MeasurementEvaluation Of Evaluation And Performance Measurement
Evaluation Of Evaluation And Performance Measurement
 
Evaluation Of A Employee Evaluation System
Evaluation Of A Employee Evaluation SystemEvaluation Of A Employee Evaluation System
Evaluation Of A Employee Evaluation System
 
Performance appraisal research design
Performance appraisal research designPerformance appraisal research design
Performance appraisal research design
 
THE IMPACT OF TRAINING AND DEVELOPMENT ON EMPLOYEES PERFORMANCE IN AN ORGAN...
THE IMPACT OF TRAINING AND  DEVELOPMENT ON EMPLOYEES  PERFORMANCE IN AN ORGAN...THE IMPACT OF TRAINING AND  DEVELOPMENT ON EMPLOYEES  PERFORMANCE IN AN ORGAN...
THE IMPACT OF TRAINING AND DEVELOPMENT ON EMPLOYEES PERFORMANCE IN AN ORGAN...
 
· Assignment 3 Creating a Compelling VisionLeaders today must be .docx
· Assignment 3 Creating a Compelling VisionLeaders today must be .docx· Assignment 3 Creating a Compelling VisionLeaders today must be .docx
· Assignment 3 Creating a Compelling VisionLeaders today must be .docx
 
Fundamentals of human resource management
Fundamentals of human resource managementFundamentals of human resource management
Fundamentals of human resource management
 
Performance Appraisal of Human Resources in Anglo-Saxon Higher Education Inst...
Performance Appraisal of Human Resources in Anglo-Saxon Higher Education Inst...Performance Appraisal of Human Resources in Anglo-Saxon Higher Education Inst...
Performance Appraisal of Human Resources in Anglo-Saxon Higher Education Inst...
 
Independent evaluation of the united nations internship program results & pro...
Independent evaluation of the united nations internship program results & pro...Independent evaluation of the united nations internship program results & pro...
Independent evaluation of the united nations internship program results & pro...
 
MODELLING TRAINING EFFECTIVENESS AND ITS IMPACT ON DEVELOPMENT OF MANAGERS IN...
MODELLING TRAINING EFFECTIVENESS AND ITS IMPACT ON DEVELOPMENT OF MANAGERS IN...MODELLING TRAINING EFFECTIVENESS AND ITS IMPACT ON DEVELOPMENT OF MANAGERS IN...
MODELLING TRAINING EFFECTIVENESS AND ITS IMPACT ON DEVELOPMENT OF MANAGERS IN...
 
A
AA
A
 
“The Effectiveness and Efficiency of the Various Graduate Recruitment Methods”
“The Effectiveness and Efficiency of the Various Graduate Recruitment Methods”“The Effectiveness and Efficiency of the Various Graduate Recruitment Methods”
“The Effectiveness and Efficiency of the Various Graduate Recruitment Methods”
 
Module 4 Reflective DiscussionPreviousNextIn the Module 4 .docx
Module 4 Reflective DiscussionPreviousNextIn the Module 4 .docxModule 4 Reflective DiscussionPreviousNextIn the Module 4 .docx
Module 4 Reflective DiscussionPreviousNextIn the Module 4 .docx
 
Training need assessment in a 5star ho
Training need assessment in a 5star hoTraining need assessment in a 5star ho
Training need assessment in a 5star ho
 
Running Head DATA SOURCE EVALUATION .docx
Running Head DATA SOURCE EVALUATION                              .docxRunning Head DATA SOURCE EVALUATION                              .docx
Running Head DATA SOURCE EVALUATION .docx
 
Research In Action #2
Research In Action #2Research In Action #2
Research In Action #2
 

More from sleeperharwell

For this assignment, review the articleAbomhara, M., & Koie.docx
For this assignment, review the articleAbomhara, M., & Koie.docxFor this assignment, review the articleAbomhara, M., & Koie.docx
For this assignment, review the articleAbomhara, M., & Koie.docxsleeperharwell
 
For this assignment, provide your perspective about Privacy versus N.docx
For this assignment, provide your perspective about Privacy versus N.docxFor this assignment, provide your perspective about Privacy versus N.docx
For this assignment, provide your perspective about Privacy versus N.docxsleeperharwell
 
For this assignment, provide your perspective about Privacy vers.docx
For this assignment, provide your perspective about Privacy vers.docxFor this assignment, provide your perspective about Privacy vers.docx
For this assignment, provide your perspective about Privacy vers.docxsleeperharwell
 
For this Assignment, read the case study for Claudia and find two to.docx
For this Assignment, read the case study for Claudia and find two to.docxFor this Assignment, read the case study for Claudia and find two to.docx
For this Assignment, read the case study for Claudia and find two to.docxsleeperharwell
 
For this assignment, please start by doing research regarding the se.docx
For this assignment, please start by doing research regarding the se.docxFor this assignment, please start by doing research regarding the se.docx
For this assignment, please start by doing research regarding the se.docxsleeperharwell
 
For this assignment, please discuss the following questionsWh.docx
For this assignment, please discuss the following questionsWh.docxFor this assignment, please discuss the following questionsWh.docx
For this assignment, please discuss the following questionsWh.docxsleeperharwell
 
For this assignment, locate a news article about an organization.docx
For this assignment, locate a news article about an organization.docxFor this assignment, locate a news article about an organization.docx
For this assignment, locate a news article about an organization.docxsleeperharwell
 
For this assignment, it requires you Identifies the historic conte.docx
For this assignment, it requires you Identifies the historic conte.docxFor this assignment, it requires you Identifies the historic conte.docx
For this assignment, it requires you Identifies the historic conte.docxsleeperharwell
 
For this assignment, create a framework from which an international .docx
For this assignment, create a framework from which an international .docxFor this assignment, create a framework from which an international .docx
For this assignment, create a framework from which an international .docxsleeperharwell
 
For this assignment, create a 15-20 slide digital presentation in tw.docx
For this assignment, create a 15-20 slide digital presentation in tw.docxFor this assignment, create a 15-20 slide digital presentation in tw.docx
For this assignment, create a 15-20 slide digital presentation in tw.docxsleeperharwell
 
For this assignment, you are to complete aclinical case - narrat.docx
For this assignment, you are to complete aclinical case - narrat.docxFor this assignment, you are to complete aclinical case - narrat.docx
For this assignment, you are to complete aclinical case - narrat.docxsleeperharwell
 
For this assignment, you are to complete aclinical case - narr.docx
For this assignment, you are to complete aclinical case - narr.docxFor this assignment, you are to complete aclinical case - narr.docx
For this assignment, you are to complete aclinical case - narr.docxsleeperharwell
 
For this assignment, you are provided with four video case studies (.docx
For this assignment, you are provided with four video case studies (.docxFor this assignment, you are provided with four video case studies (.docx
For this assignment, you are provided with four video case studies (.docxsleeperharwell
 
For this assignment, you are going to tell a story, but not just.docx
For this assignment, you are going to tell a story, but not just.docxFor this assignment, you are going to tell a story, but not just.docx
For this assignment, you are going to tell a story, but not just.docxsleeperharwell
 
For this assignment, you are asked to prepare a Reflection Paper. Af.docx
For this assignment, you are asked to prepare a Reflection Paper. Af.docxFor this assignment, you are asked to prepare a Reflection Paper. Af.docx
For this assignment, you are asked to prepare a Reflection Paper. Af.docxsleeperharwell
 
For this assignment, you are asked to prepare a Reflection Paper. .docx
For this assignment, you are asked to prepare a Reflection Paper. .docxFor this assignment, you are asked to prepare a Reflection Paper. .docx
For this assignment, you are asked to prepare a Reflection Paper. .docxsleeperharwell
 
For this assignment, you are asked to conduct some Internet research.docx
For this assignment, you are asked to conduct some Internet research.docxFor this assignment, you are asked to conduct some Internet research.docx
For this assignment, you are asked to conduct some Internet research.docxsleeperharwell
 
For this assignment, you are a professor teaching a graduate-level p.docx
For this assignment, you are a professor teaching a graduate-level p.docxFor this assignment, you are a professor teaching a graduate-level p.docx
For this assignment, you are a professor teaching a graduate-level p.docxsleeperharwell
 
For this assignment, we will be visiting the PBS website,Race  .docx
For this assignment, we will be visiting the PBS website,Race  .docxFor this assignment, we will be visiting the PBS website,Race  .docx
For this assignment, we will be visiting the PBS website,Race  .docxsleeperharwell
 
For this assignment, the student starts the project by identifying a.docx
For this assignment, the student starts the project by identifying a.docxFor this assignment, the student starts the project by identifying a.docx
For this assignment, the student starts the project by identifying a.docxsleeperharwell
 

More from sleeperharwell (20)

For this assignment, review the articleAbomhara, M., & Koie.docx
For this assignment, review the articleAbomhara, M., & Koie.docxFor this assignment, review the articleAbomhara, M., & Koie.docx
For this assignment, review the articleAbomhara, M., & Koie.docx
 
For this assignment, provide your perspective about Privacy versus N.docx
For this assignment, provide your perspective about Privacy versus N.docxFor this assignment, provide your perspective about Privacy versus N.docx
For this assignment, provide your perspective about Privacy versus N.docx
 
For this assignment, provide your perspective about Privacy vers.docx
For this assignment, provide your perspective about Privacy vers.docxFor this assignment, provide your perspective about Privacy vers.docx
For this assignment, provide your perspective about Privacy vers.docx
 
For this Assignment, read the case study for Claudia and find two to.docx
For this Assignment, read the case study for Claudia and find two to.docxFor this Assignment, read the case study for Claudia and find two to.docx
For this Assignment, read the case study for Claudia and find two to.docx
 
For this assignment, please start by doing research regarding the se.docx
For this assignment, please start by doing research regarding the se.docxFor this assignment, please start by doing research regarding the se.docx
For this assignment, please start by doing research regarding the se.docx
 
For this assignment, please discuss the following questionsWh.docx
For this assignment, please discuss the following questionsWh.docxFor this assignment, please discuss the following questionsWh.docx
For this assignment, please discuss the following questionsWh.docx
 
For this assignment, locate a news article about an organization.docx
For this assignment, locate a news article about an organization.docxFor this assignment, locate a news article about an organization.docx
For this assignment, locate a news article about an organization.docx
 
For this assignment, it requires you Identifies the historic conte.docx
For this assignment, it requires you Identifies the historic conte.docxFor this assignment, it requires you Identifies the historic conte.docx
For this assignment, it requires you Identifies the historic conte.docx
 
For this assignment, create a framework from which an international .docx
For this assignment, create a framework from which an international .docxFor this assignment, create a framework from which an international .docx
For this assignment, create a framework from which an international .docx
 
For this assignment, create a 15-20 slide digital presentation in tw.docx
For this assignment, create a 15-20 slide digital presentation in tw.docxFor this assignment, create a 15-20 slide digital presentation in tw.docx
For this assignment, create a 15-20 slide digital presentation in tw.docx
 
For this assignment, you are to complete aclinical case - narrat.docx
For this assignment, you are to complete aclinical case - narrat.docxFor this assignment, you are to complete aclinical case - narrat.docx
For this assignment, you are to complete aclinical case - narrat.docx
 
For this assignment, you are to complete aclinical case - narr.docx
For this assignment, you are to complete aclinical case - narr.docxFor this assignment, you are to complete aclinical case - narr.docx
For this assignment, you are to complete aclinical case - narr.docx
 
For this assignment, you are provided with four video case studies (.docx
For this assignment, you are provided with four video case studies (.docxFor this assignment, you are provided with four video case studies (.docx
For this assignment, you are provided with four video case studies (.docx
 
For this assignment, you are going to tell a story, but not just.docx
For this assignment, you are going to tell a story, but not just.docxFor this assignment, you are going to tell a story, but not just.docx
For this assignment, you are going to tell a story, but not just.docx
 
For this assignment, you are asked to prepare a Reflection Paper. Af.docx
For this assignment, you are asked to prepare a Reflection Paper. Af.docxFor this assignment, you are asked to prepare a Reflection Paper. Af.docx
For this assignment, you are asked to prepare a Reflection Paper. Af.docx
 
For this assignment, you are asked to prepare a Reflection Paper. .docx
For this assignment, you are asked to prepare a Reflection Paper. .docxFor this assignment, you are asked to prepare a Reflection Paper. .docx
For this assignment, you are asked to prepare a Reflection Paper. .docx
 
For this assignment, you are asked to conduct some Internet research.docx
For this assignment, you are asked to conduct some Internet research.docxFor this assignment, you are asked to conduct some Internet research.docx
For this assignment, you are asked to conduct some Internet research.docx
 
For this assignment, you are a professor teaching a graduate-level p.docx
For this assignment, you are a professor teaching a graduate-level p.docxFor this assignment, you are a professor teaching a graduate-level p.docx
For this assignment, you are a professor teaching a graduate-level p.docx
 
For this assignment, we will be visiting the PBS website,Race  .docx
For this assignment, we will be visiting the PBS website,Race  .docxFor this assignment, we will be visiting the PBS website,Race  .docx
For this assignment, we will be visiting the PBS website,Race  .docx
 
For this assignment, the student starts the project by identifying a.docx
For this assignment, the student starts the project by identifying a.docxFor this assignment, the student starts the project by identifying a.docx
For this assignment, the student starts the project by identifying a.docx
 

Recently uploaded

IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...
IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...
IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...PsychoTech Services
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxheathfieldcps1
 
The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13Steve Thomason
 
Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..Disha Kariya
 
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Sapana Sha
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsTechSoup
 
Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactdawncurless
 
Nutritional Needs Presentation - HLTH 104
Nutritional Needs Presentation - HLTH 104Nutritional Needs Presentation - HLTH 104
Nutritional Needs Presentation - HLTH 104misteraugie
 
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...fonyou31
 
Z Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphZ Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphThiyagu K
 
Measures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeMeasures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeThiyagu K
 
Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxVishalSingh1417
 
9548086042 for call girls in Indira Nagar with room service
9548086042  for call girls in Indira Nagar  with room service9548086042  for call girls in Indira Nagar  with room service
9548086042 for call girls in Indira Nagar with room servicediscovermytutordmt
 
Beyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactBeyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactPECB
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introductionMaksud Ahmed
 
Class 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfClass 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfAyushMahapatra5
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)eniolaolutunde
 
A Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformA Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformChameera Dedduwage
 

Recently uploaded (20)

IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...
IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...
IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptx
 
The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13The Most Excellent Way | 1 Corinthians 13
The Most Excellent Way | 1 Corinthians 13
 
Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..
 
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
 
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptxINDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impact
 
Nutritional Needs Presentation - HLTH 104
Nutritional Needs Presentation - HLTH 104Nutritional Needs Presentation - HLTH 104
Nutritional Needs Presentation - HLTH 104
 
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
 
Z Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphZ Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot Graph
 
Measures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeMeasures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and Mode
 
Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptx
 
9548086042 for call girls in Indira Nagar with room service
9548086042  for call girls in Indira Nagar  with room service9548086042  for call girls in Indira Nagar  with room service
9548086042 for call girls in Indira Nagar with room service
 
Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1
 
Beyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactBeyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global Impact
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introduction
 
Class 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfClass 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdf
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)
 
A Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformA Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy Reform
 

A comparison of the practices used by human resource developme.docx

  • 1. A comparison of the practices used by human resource development professionals to evaluate web-based and classroom-based training programmes within seven Korean companies Younghee Jessie Kong a * and Ronald Lynn Jacobs b * a College of Business, Franklin University, Columbus, OH, USA; b College of Education, University of Illinois, Champaign, IL, USA (Received 28 September 2010; final version received 25 October 2011) The purpose of the study was to compare the practices used by human resource development (HRD) professionals to evaluate web-based and classroom-based training (CBT) programmes within seven Korean companies. This study used
  • 2. four components of evaluation and three factors of evaluation barriers to compare the differences between these two training approaches. This study also explored the key decision factors for determining how HRD professionals evaluated their web-based and CBT programmes. Two data sets were used for the study; one set of data was gathered from a survey questionnaire distributed to HRD professionals and the other was gathered from interviews with HR/HRD directors within the seven companies. The results showed that web-based and CBT programmes were not meaningfully different on the most components of evaluation and evaluation barriers. The results also found six key decision factors determining evaluation for web-based and CBT programmes. Keywords: training evaluation; training evaluation barriers; primary factors determining evaluation Introduction Since many organizations consider training as a way of improving their performance in today’s competitive environment (Yanmil and McLean 2001), they want to select the best training programmes for their employees among a wide variety of training approaches (Bartley and Golek 2004). Classroom-based training (CBT) is one of the most frequently used training approaches in organizations, since learners can directly
  • 3. communicate with an instructor and peers and directly share information with each other (Kapp and McKeague 2002). In addition to the use of CBT, the development of web-based training (WBT) has been increased in the human resource development (HRD) field due to rapid advancement in the capabilities and distribution of technologies (Bassi and Van Buren 1998). Since investment in these training programmes has been rapidly increasing in organizations, senior management asks HRD professionals to evaluate their WBT *Corresponding author. Email: [email protected]; [email protected] Human Resource Development International Vol. 15, No. 1, February 2012, 79–98 ISSN 1367-8868 print/ISSN 1469-8374 online � 2012 Taylor & Francis http://dx.doi.org/10.1080/13678868.2012.658632 http://www.tandfonline.com and classroom-based training (CBT) programmes in order to determine the best training programmes for their employees (Olson and Wisher 2002). Many previous researchers (Curtain 2002; Jung and Rha 2003;
  • 4. Rumble 2001; Whalen and Wright 1999) have stressed that evaluating web- based and classroom- based training should be considered and performed differently, because their learning environments are entirely different from each other. For instance, WBT is delivered via the Internet or intranet using a computer, while CBT is delivered face- to-face via an instructor. In addition, WBT allows trainees to learn at their own pace at a convenient time and location. On the other hand, CBT is typically group- based requiring that trainees meet their instructor and peers at a fixed time and location. Following these differences, researchers (Curtain 2002; Jung and Rha 2000; Rumble 2001; Trentin 2000; Whalen and Wright 1999) have suggested that evaluating WBT programmes also needs to be approached differently from the evaluation of CBT programmes. They have found that there are more difficulties and challenges to evaluating WBT programmes than classroom- based training, due to the wide variation in WBT forms. The evaluation of WBT is influenced by a number of factors, such as IT infrastructure used, the number of learners, flexibility of learning times, frequency of course revision, type and amount of media used, type and amount of learner support, type and amount of interaction, learners’ experience level with technology, or unexpected technical issues.
  • 5. For example, WBT structures change at an exceptional pace, since technologies have been changing so quickly. Consequently, it can be difficult to thoroughly understand the range of different types and organizational settings in which WBT is delivered among organizations (Bradsford, Brown, and Cocking 1999). In addition, when an organization provides different types of online courses that require different preparation times and levels of online interaction, it is more difficult to evaluate the level of workload and extra time spent by administrative staff and instructors to support online courses (Curtain 2002; Rumble 2001). However, despite calls for the evaluation of WBT programmes to be conducted differently, there is limited information available on whether HRD professionals actually follow this suggestion in practice. Strother (2002) suggested that useful and sound comparative study about evaluating web-based and CBT has been extremely limited over the years, since evaluating WBT programmes has not been treated differently. Wisher and Olson (2003) also pointed out the lack of any empirical studies comparing the evaluation of WBT and CBT programmes. They reviewed over 500 articles published between 1996 and 2002, and found that most of these studies dealt with design issues or technology concerns
  • 6. rather than with strategies or tools for successful evaluation. Hence, due to an insufficient number of previous evaluation studies, they had difficulty assessing evaluation practices in their survey. If WBT and CBT approaches differ from each other in important ways, which have implications on how these training approaches should be evaluated, and, if there is limited information on whether organizations actually differ in their evaluation practices related to these training approaches, then more should be known about how HRD professionals actually evaluate their web-based and CBT programmes. Understanding the practices used by HRD professionals to evaluate different training approaches will help us identify elements to evaluate that are 80 Y.J. Kong and R.L. Jacobs missing from current practice, examine enablers and disablers of the evaluation of these elements, and conduct appropriate evaluations of them. The purpose of the study is to compare the practices used by HRD professionals to evaluate web-based and CBT programmes within seven Korean companies. More specifically, this study (1) examines how HRD professionals evaluate web-based and
  • 7. CBT programmes in their organizations, and (2) compares the differences between these training approaches. Also, this study (3) identifies the barriers that might prevent HRD professionals from evaluating web-based and CBT programmes, and (4) compares the differences between these training approaches with the evaluation barriers. Finally, this study (5) explores the important decision factors that HRD professionals use to determine how to evaluate web-based and CBT programmes in their organizations. Literature review The purpose of evaluation Kraiger (2002) suggests three primary reasons to evaluate training programmes: decision-making, feedback, and marketing. The first purpose for conducting evaluation is to provide input for making decisions about training programmes, such as course retention, course revision, or personnel decisions (e.g. quality of instructor). The second purpose for evaluation is to provide feedback to course designers, trainers, or the trainees themselves that would allow them to design or engage with the course more effectively. The final purpose for collecting training evaluation data is for the purpose of marketing the training programmes (e.g. demonstrating the value of training to upper management or helping future sponsors
  • 8. or trainees understand the beneficial changes resulting from training). Training evaluation models On reviewing the work on training evaluation in HRD literature, Kirkpatrick’s four-level framework and the CIPP model emerge as particularly common in organizations. Kirkpatrick’s four-level framework Kirkpatrick’s evaluation is the oldest and most widely used evaluation criteria in organizations. Many of the other evaluation models found in the literature are designed and revised based on Kirkpatrick’s framework (Russ-Eft and Preskill 2001). The Kirkpatrick framework consists of a series of four levels: reaction, learning, behaviour, and results (1998). In this model, reactions refer to trainees’ levels of satisfaction, measured via surveys or interviews designed to identify their perceptions of the training programme. Reaction is measured during or immediately after training. The next level of evaluation, learning, refers to the extent of trainees’ changes in knowledge, skills, and attitudes. Learning is measured before and after training. The next level of evaluation is behaviour. Behaviour refers to the extent of trainees’ behaviour changes in transferring knowledge, skills,
  • 9. and attitudes to their jobs. Behaviour is measured after training. The final level of evaluation is results. A result refers to organizational results or changes because of enhanced behaviours. Human Resource Development International 81 In Kirkpatrick’s framework, reactions to training are related to learning, learning is related to behaviour, and behaviour is related to results. However, Kirkpatrick’s framework has been routinely criticized or revised in the HRD literature. For example, Alliger and Janak (1989) question its hierarchical evaluation approach. According to their critiques, reactions to training should be viewed as unrelated to learning, and learning can have a direct influence on results criteria as opposed to only indirect influence through behavioural change. In addition, Holton (1996) argues that Kirkpatrick’s evaluation is nothing more than a taxonomy of outcomes. He removes reactions from the level of primary outcomes of training and defines it as a moderating variable between trainees’ motivation to learn and their actual learning. In addition, he states that learning is related to transfer and transfer is related to results. The CIPP model
  • 10. Another of the most frequently used evaluation models for training in organizations is the systems-based model (Phillips 1997). Under the systems approach, Stufflebeam’s (1971) CIPP model is the best known (Fitzpatirck, Sanders, and Worthen 2004). He points out that, since evaluation should be informative to decision makers, the systems approach can be effectively applied to evaluation. The CIPP model acronym stands for the four components of training systems: context, input, process, and output. The following are the main characteristics of these four types of evaluation: (1) Context evaluation involves obtaining information about the organizational context to determine training needs and to identify objectives for the programme. (2) Input evaluation involves determining the availability of resources, alter- native programme strategies, and plans for instructional design process. (3) Process evaluation is assessing the plans and barriers for implementing the programme and monitoring procedural events and activities. (4) Product evaluation includes making judgments about the outcomes of the programme related to their objectives, context, input, and process informa-
  • 11. tion, interpreting their worth and merit. Evaluating web-based and CBT based on a systems approach In practice, Jacobs (2003) provides an interactive process of evaluation from the systems view for evaluating structured on-the-job training (SOJT). In his book, he points out that this systems view of training evaluation enables HRD professionals to see all the relevant elements separately, allowing them to evaluate training programmes, and how the elements bring together each component. In addition, it enables HRD professionals to examine the relationship among the four components within a system (e.g. SOJT) and with other systems in an entire system or an organization, allowing them to determine the value of each component and of the entire system. As a result, Jacobs suggests that the evaluation questions for SOJT should be asked based on system inputs, processes, and outputs. The questions also need to be asked about the organizational context in which the training takes place. For 82 Y.J. Kong and R.L. Jacobs example, the training input questions include units of material to be learned, the training locations, and the characteristics of trainers and
  • 12. trainees at the time of the training. The training process questions include the amount of time required to complete the training, the availability of training resources, and the behaviours of the trainer and trainee during the training. The training output questions include whether the training goals have been achieved, whether the training met the needs of trainees and whether the training is more effective or efficient than other training approaches in terms of financial benefits after the training. Lastly, the organizational context questions include the extent of management commitment to the use of the training and the interaction of the training with other systems in the organization. The evaluation of SOJT from a systems view can be applied to the evaluation of other training approaches such as web-based and CBT, because they can also be viewed as systems themselves and, at the same time, as one of many subsystems in a larger system (i.e. the organization). Following the systems approach proposed by Jacobs (2003), Figure 1 explains how to evaluate WBT programmes; Figure 2 illustrates how to evaluate CBT programmes. Barriers to training evaluation Previous research found that organizations do not actually conduct evaluation well (Curtain 2002; Holton and Naquin 2005; Jung and Rha 2000; Kirkpatrick 1998;
  • 13. Figure 1. The evaluation of WBT programmes from a systems view. Human Resource Development International 83 Kraiger 2002; Rumble 2001; Whalen and Wright 1999). The primary barriers inhibiting evaluation proposed by these authors are: (1) Lack of expertise in evaluation: lack of knowledge and skills or lack of experience in evaluation (2) Lack of organizational support for evaluation: unavailability of resources, organizational confidentiality practices or policies, fear of negative financial return, and limited budget and time (3) Lack of evaluation methods and tools: evaluation activities limited to reaction sheets and statements of learning outcome, lack of common cost framework, and methodological limitations of financial returns measurement Three barriers affecting evaluation Previous studies have endeavoured to find appropriate ways of conducting effective evaluation (Hofstetter and Alkin 2003; Johnson 1998; Leviton 2003; Shulha and Cousins 1997). Taut and Alkin (2003) propose that this body of
  • 14. research can be addressed to oppose the barriers that inhibit evaluation. In their study, they examined the perception of project staff on barriers that inhibited evaluation implementation and its eventual use, based on Alkin’s (1985) evaluation framework. In his evaluation framework, Alkin (1985) divides factors affecting the conduct of Figure 2. The evaluation of CBT programmes from a systems view. 84 Y.J. Kong and R.L. Jacobs evaluation and its use into three categories: human factors, context factors, and evaluation factors. First, ‘human factors describe evaluator and user characteristics’ (Taut and Alkin 2003, 251). For example, human factors include evaluators’ commitment to facilitate and enhance evaluation, evaluator’s willingness to actively involve stakeholders in the evaluation process, evaluator’s competence and reliability in evaluation, stakeholders’ organizational positions, stakeholders’ experience levels to help the evaluator and their interest in the evaluation, stakeholders’ commitment to use evaluation results, and so on. Second, ‘context factors refer to the context in which the
  • 15. evaluated program exists’ (Taut and Alkin 2003, 251). For example, context factors include written requirements (e.g. federal/state requirements and organizational policies), fiscal constraints (e.g. the amount of money and time available for the evaluation), the context outside of the actual training programme and the organization (e.g. external funding agencies), and programme characteristics (e.g. age/maturity of the programme, innovativeness of the programme, and overlap with other projects). Third, evaluation factors involve the evaluation itself. For example, evaluation factors include evaluation procedures (e.g. the appropriateness and accuracy of the methods used and use of a general model to guide evaluation), the information dialogue (e.g. amount and quality of interaction between the evaluator and stakeholders), substance of evaluation information (e.g. information relevancy and specificity), and evaluation-reporting factor (e.g. timing of information, style of oral presentations, and format of reports). Methods This study employed a mix of quantitative and qualitative non- experimental research. Two data sets were used for the study. One set of data was gathered from a survey questionnaire distributed to Korean HRD professionals
  • 16. via online in October of 2008 to compare the differences between web-based and CBT programmes in evaluation practices. The other data set was gathered from interviews with HR/HRD directors in February of 2009 to explore the research questions in more depth. The researcher adapted Jacobs’ evaluation framework (2003) that was based on the systems approach to four components of evaluation and Alkin’s framework (1985) for evaluation barriers. Then, she developed new construct scales for a survey questionnaire and interview questions based on these established constructs to fit this study’s purpose. Organization selection The researcher selected seven companies in Korea where web- based and CBT programmes were implemented and evaluated. The seven companies selected represent a variety of industries in Korea, including food, chemical manufacturing, automobile manufacturing, financial services, IT services, and educational services industries. The total number of employees in each organization varies from 400 to 75,000, and the total revenue of the companies’ ranges, as of 2007, from $1.5 million to
  • 17. Human Resource Development International 85 $30 billion. All the companies are headquartered in Seoul, Korea, with branch offices in many regional cities across the nation and around the world. Survey to HRD professionals Participant selection The population for this study was all HRD professionals (N¼147) who currently evaluated web-based and/or CBT programmes within the seven companies in Korea, and the total number of respondents was 73. Out of the 73 respondents, 69 respondents had experience in the evaluation of CBT programmes, whereas 43 had experience in the evaluation of WBT programmes. Instrument design For the survey questionnaire, the researcher designated one independent variable with two levels: web-based and CBT programmes. Two dependent variables were measured in this study: the four components of evaluation and three factors of evaluation barriers. Six demographic characteristics were also measured in this study: type of industry, size of organization, an organization’s years of conducting training programmes, employee’s job level, employee’s educational level, and
  • 18. percentage of employee’s work that is related to training evaluation. Respondents were asked to identify the extent to which input, process, output, and organizational context components of evaluation were used by HRD professionals to evaluate their web-based and CBT programmes, using a five-point Likert-type scale ranging from 1 (never) to 5 (always), with higher scores indicating more frequent use of evaluation questions. Respondents were also asked to assess the degree of barriers to evaluating web-based and CBT programmes based on three factors affecting evaluation (i.e. human, context, and evaluation factors), using a five-point Likert-type scale ranging from 1 (strongly disagree) to 5 (strongly agree), with lower scores indicating more barriers perceived by HRD professionals. Validity and reliability To ensure the instrument’s content validity, the panel of five experts reviewed and commented on the items in the instrument to ensure that they accurately represented the content area intended. Then, the English version of the instrument was translated to the Korean version by the researcher, and the Korean versions were reviewed by Korean Ph.D. students and HRD professionals to gain their consensus on the items. To ensure the reliability of the survey questionnaire, Cronbach’s alpha coefficient
  • 19. was employed to measure internal consistency of the items and the alpha values of the instrument were between .8 and .9. Interviews with HR or HRD directors Participant selection To obtain qualitative data, the researcher conducted individual interviews with HR or HRD directors. The participants were all current directors in the departments of HR or HRD in the seven companies selected. Five out of seven HR/HRD directors agreed to participate in this study and were interviewed. Interviews were conducted 86 Y.J. Kong and R.L. Jacobs either phone, email, or face to face, whatever was most convenient for them, and supplementary follow-up phone calls were made, when necessary. Interview questions The interview was composed of four main questions including demographic information. Respondents were asked about how web-based and CBT programmes were implemented and evaluated differently in their organizations based on the four components of evaluation. Respondents were also asked about the factors that
  • 20. inhibited the evaluation of web-based and CBT programmes and whether there were differences between these training approaches based on the three factors of evaluation barriers. Finally, they were asked what kind of information they looked for from the evaluations and how they used the results of the evaluation in order to explore the key decision factors determining evaluation for their web-based and CBT programmes. Credibility and dependability Maxwell (2005) states that, in qualitative research, credibility refers to internal validity, and the credibility in this study was enhanced based on rich data and control of bias. The researcher audio-taped and transcribed all interviews word by word and took field notes recording the interviewees’ actions during interview. To ensure consensus and reduce bias in this study, the verbatim transcripts were shared with interviewees or another researcher to ensure that the researcher’s interpretations of the interviewees’ statements were consistent with the transcripts. Lincoln and Guba (1985) state that, in qualitative research, dependability closely corresponds to the notion of reliability in quantitative research, and ‘since there can be no validity without reliability, a demonstration of validity is sufficient to establish reliability’ (p. 316).
  • 21. Data analysis First, the data from the survey questionnaire were analysed using the Statistical Package for the Social Science (SPSS). Descriptive statistics such as frequency, percentage, mean for population (m), and sigma for population (s) were used to describe the evaluation practices used by Korean HRD professionals and to compare the differences between web-based and CBT programmes in the evaluation practices. In addition, multiple regression analysis was conducted to examine the relationship between demographic variables of HRD professionals and the evalua- tion practices used for web-based and CBT programmes. Second, the data from the interview was read, coded, and analysed by the researcher to seek a deeper understanding of these relationships in the natural setting. Results Four components of evaluation As shown in Tables 1 and 2, the results of the survey showed that web-based and CBT programmes were most frequently evaluated for the process evaluation component (mw¼4.02, mc¼4.29) followed by the input evaluation component Human Resource Development International 87
  • 22. (mw¼3.70, mc¼3.78). On the other hand, web-based and CBT programmes were least frequently evaluated for the organizational context evaluation component (mw¼3.03, mc¼2.93), followed by the output evaluation component (mw¼3.12, mc¼3.46). The results also showed that, the input evaluation component produced a higher mean score for CBT (m¼3.78, s¼0.59) than WBT programmes (m¼3.70, s¼0.86). The process evaluation component produced a higher mean score for CBT (m¼4.29, s¼0.56) than WBT programmes (m¼4.02, s¼0.75). The output evaluation compo- nent produced a higher mean score for CBT (m¼3.46, s¼0.74) than WBT programmes (m¼3.12, s¼0.86). On the other hand, the organizational context evaluation component produced a lower mean score for CBT (m¼2.93, s¼0.93) than WBT programmes (m¼3.03, s¼0.85). The results of the interview (Table 3) produced the same results as the survey; web-based and CBT programmes were most frequently evaluated for the process evaluation component, whereas they were least frequently evaluated for the organizational context evaluation component. These results also showed that, in the input and process components of
  • 23. evaluation, the five interviewees all said that their HRD professionals used existing survey forms to evaluate trainees’ satisfaction with training. However, they said that they included different questions for web-based and CBT programmes on the survey questionnaire, although the main categories for the survey were the same. All five interviewees stated that the online survey had more detailed questions included on the questionnaire than the survey for the classroom. In addition, they mentioned that WBT was more complicated to evaluate than CBT, because web-based courses were specialized and individualized. Table 2. Descriptive statistics on CBT programs for four components of evaluation (N¼69). Cases Included Excluded Total Mean Std. deviation N Per cent N Per cent N Per cent Input 3.78 .59 69 94.5 4 5.5 73 100.0 Process 4.29 .56 69 94.5 4 5.5 73 100.0 Output 3.46 .74 68 93.2 5 6.8 73 100.0 Context 2.93 .93 69 94.5 4 5.5 73 100.0 Table 1. Descriptive statistics on WBT programs for four components of evaluation (N¼43). Cases
  • 24. Included Excluded Total Mean Std. deviation N Per cent N Per cent N Per cent Input 3.70 .86 43 58.9 30 41.1 73 100.0 Process 4.02 .75 43 58.9 30 41.1 73 100.0 Output 3.12 .86 43 58.9 30 41.1 73 100.0 Context 3.03 .85 43 58.9 30 41.1 73 100.0 88 Y.J. Kong and R.L. Jacobs Meanwhile, the HRD professionals at the five companies rarely evaluated their web-based and CBT programmes for the output evaluation component, but they more frequently evaluated WBT programmes than CBT programmes for the organizational context evaluation component in order to receive reimbursement of tuition money from the Korean government. In summary, the results of both the survey and the interviews showed that web- based and CBT programmes were meaningfully different on the process evaluation component. However, web-based and CBT programmes were not meaningfully different on the input, output, and organizational context components of evalua- tion. The results of both the survey and the interviews also showed that the demographic characteristics of HRD professionals were not meaningfully related to the practices used to evaluate web-based and CBT programmes
  • 25. for each evaluation component. Three factors of evaluation barriers As shown in Tables 4 and 5, the results of the survey showed that the respondents perceived more barriers in the context (mw¼3.17, mc¼3.11) and evaluation factors Table 3. The differences between web-based and CBT programs for the four components of evaluation. Categories Web-based Classroom-based Input evaluation 1. Survey questionnaire Different questions: adding more questions Different questions Process evaluation 1. Survey questionnaire Different questions: adding more questions Different questions 2. Trainees’ participation and learning progress Checking through online: Checking through direct observation:
  • 26. . Number of clicking slides . Class . Login times . CC-TV . Discussion board . Interview . Pop-up window . Homework . Homework or discussion in class Output evaluation Self-check Self-check 1. Learning tests Multiple choices Multiple choices Pass/fail Pass/fail Simple essay Essay Reflection paper Reflection/project paper 2. Transfer of training Action plan Projects Supervisor’s observation Action plan Number of sales Supervisor’s observation 360 degree feedback Number of sales Statistics 360 degree feedback Statistics Organizational context evaluation 1. Survey questionnaire Little difference Little difference 2. Korean government support Report evaluation results to
  • 27. the government N/A Human Resource Development International 89 (mw¼3.14, mc¼3.22) than in the human factors (mw¼3.68, mc¼3.81) to evaluating their web-based and CBT programmes. The results also showed that the human factors produced a lower mean score for WBT (m¼3.69, s¼0.58) than CBT programmes (m¼3.81, s¼0.55). When comparing web-based and CBT programmes for the context and evaluation factors of evaluation barriers, the context factors produced a higher mean score for WBT (m¼3.17, s¼0.74) than CBT programmes (m¼3.11, s¼0.73). The evaluation factors produced a lower mean score for WBT programmes (m¼3.14, s¼0.80) than CBT programmes (m¼3.22, s¼0.71). The results of the interviews (Table 6) produced the same results as the survey; the context and evaluation factors were the primary barriers preventing HRD professionals from evaluating web-based and CBT programmes. Table 6 presents the barriers that inhibited the evaluation of these training approaches in each area – human, context, and evaluation factors. The interview results also showed that evaluation barriers were slightly different
  • 28. between web-based and CBT programmes in the area of human factors due to the lack of credibility for an external evaluator or of responsibility of vendors for WBT programmes. Most companies being studied here purchased online courses from professional vendors. Three interviewees stated that the vendors did not take much responsibility for further evaluation after the training was completed. One interviewee also stated that his senior management did not trust the evaluation results provided by an external evaluator, because they did not believe that an external evaluator knew their organization’s characteristics and circumstances better than their own staff. In addition, evaluation barriers were slightly different between web-based and CBT programmes in the area of context factors due to the unique structure of web- based courses. Based on interviews with HR/HRD directors, most of their HRD Table 4. Descriptive statistics on three factors of evaluation barriers for WBT programs (N¼40). Cases Included Excluded Total Mean Std. deviation N Per cent N Per cent N Per cent Human 3.69 .58 40 54.8 33 45.2 73 100.0
  • 29. Context 3.17 .74 40 54.8 33 45.2 73 100.0 Evaluation 3.14 .80 40 54.8 33 45.2 73 100.0 Table 5. Descriptive Statistics on three factors of evaluation barriers for CBT programs (N¼67). Cases Included Excluded Total Mean Std. deviation N Per cent N Percent N Per cent Human 3.81 .55 67 91.8 6 8.2 73 100.0 Context 3.11 .73 67 91.8 6 8.2 73 100.0 Evaluation 3.22 .71 67 91.8 6 8.2 73 100.0 90 Y.J. Kong and R.L. Jacobs professionals needed to add more questions to the survey for their WBT programmes. Finally, all interviewees pointed out that there was little difference between web-based and CBT programmes in the evaluation factors of evaluation barriers. HRD professionals perceived that barriers in current evaluation methods and tools, especially for evaluating trainees’ performance and financial benefits, as apply to both web-based and CBT programmes. In summary, the survey and interview results both showed that web-based and CBT programmes were not meaningfully different on the three
  • 30. factors of evaluation barriers. The results of both the survey and interviews also showed that the demographic characteristics of HRD professionals were not meaningfully related to the barriers to evaluating web-based and CBT programmes. Key decision factors for determining evaluation After interviewing five HR/HRD directors, the primary decision factors determining evaluation for web-based and CBT programmes were identified as follows: senior management’s need, development of current or new training programmes, type of training programme, political issues for online courses, personnel benefits, and budget for training programmes. Senior management’s needs The first decision factor determining evaluation was the needs or requests of senior management for the evaluation of web-based and CBT programmes. All interviewees Table 6. The differences between web-based and CBT programs in the three factors of evaluation barriers. Categories Web-based Classroom-based Human factor 1. Lack of professional knowledge and skills No difference No difference 2. Lack of credibility & responsibility of an
  • 31. external evaluator or vendors More barriers perceived N/A 3. Lack of HRD professionals’ commitment to evaluation No difference No difference 4. Lack of stakeholders’ commitment to evaluation No difference No difference Context factor 1. Lack of senior management’s interest in training evaluation No difference No difference 2. Lack of support from organizations (a) Limited budget for evaluation No difference No difference (b) Limited time for evaluation No difference No difference 3. The unique structure of web-based course Barriers perceived Not perceived 4. Organizational characteristics and circumstances No difference No difference Evaluation factor 1. Lack of evaluation methods and tools
  • 32. (a) Outdated, unreliable survey tools No difference No difference (b) Few reliable evaluation methods and tools available to evaluate trainees’ performance and organizational results No difference No difference Human Resource Development International 91 said that this factor was the most powerful factor in how HRD professionals evaluated their web-based and CBT programmes. The senior management mostly asked HRD professionals to use surveys to measure trainees’ satisfaction with training and their participation rates in classes. However, sometimes, members of senior management asked HRD professionals to evaluate trainees’ knowledge and skills, especially when trainees needed to acquire certifications. In addition, they occasionally requested information on trainees’ job performance as a result of training. Development of current or new training programmes The second decision factor determining decisions about training evaluation was deciding whether to improve current training programmes or develop new training programmes. Companies interviewed identified and analysed
  • 33. problems with current training programmes based on satisfaction rates of trainees primarily through surveys. If training satisfaction was low, then they revised current training programmes or developed new training programmes. Type of training programme The third decision factor for training evaluation depended on what type of training programme was offered. Generally, WBT was purchased from vendors for awareness or managerial training, whereas CBT was provided from within the organization for the purpose of technical training. All interviewees said that they generally evaluated knowledge tests, supervisor’s observations, and trainees’ satisfaction, and checked attendance rates of trainees for technical training programmes. On the other hand, they generally provided self-assessment or evaluated trainees’ satisfaction for awareness and managerial training programmes. Political issues for online courses All interviewees talked about political issues related to their online courses. The politics related to online courses were mainly determined by the relationships between the Korean government and organizations and between organizations and their HRD professionals. First, interviewees discussed the relationship between the government and their organizations. The Korean government has encouraged
  • 34. companies to provide their employees with WBT programmes, reimbursing companies for training costs. Since companies can reduce the cost of training through going online, they endeavour to meet the government’s evaluation criteria for WBT by reporting, for example, the completion rates and satisfaction rates of trainees. Second, the relationship between an organization and its HRD professionals worked to accomplish the intended goals of HRD professionals. For example, one of the interviewees stated that HRD professionals sometimes needed to purchase new equipment to enable the delivery of multimedia course content or replace old equipment or learning management systems with new ones. Then, when they evaluated trainees’ satisfaction with current online courses, they included evaluation of their IT infrastructure, online servers, or other equipments that they needed to buy or replace, and reported the results to management to achieve their intended goals in delivering training programmes. 92 Y.J. Kong and R.L. Jacobs Personnel benefits The fifth decision factor involved how to evaluate employees’ performance related to training. Three interviewees stated that they evaluated their
  • 35. employees’ performance to measure personnel benefits. The interviewees stated that they evaluated HRD professionals regarding how they provided their employees with training pro- grammes and whether they increased the purchase of online courses from out- sourcing companies within their organization’s budgets. In addition, the interviewees mentioned that they evaluated trainees regarding what training programmes they selected, how they took their training programmes, and whether they completed the training hours required for their job promotions. Budget for training programmes The last decision factor for training evaluation involved setting up the budget for implementing future training programmes. Two interviewees said that they conducted surveys to set up budgets for upcoming web-based and CBT programmes. Discussion The premise of the study was that the evaluation of different training approaches, such as classroom-based and WBT, should be planned and performed differently because of different learning environments (Curtain 2002; Jung and Rha 2003; Rumble 2001; Whalen and Wright 1999). In general, the results of the study did not support the premise of the study, except for the process component of
  • 36. evaluation. The results did show some differences between web- based and CBT programmes in evaluation practices, but they were not judged to be meaningfully different. Six points might be considered in explanation of the results of the study: (1) lack of management interest in evaluation, (2) uncertain workforce development policy, (3) possible confusion between the input and process evaluation components, (4) lack of knowledge and skills among HRD professionals, and (5) uncertainty about the original premise. Lack of interest of senior management in evaluation The first point suggests that senior management may not have expressed sufficient interest in all of the evaluation components. All five interviewees reported that, when evaluating web-based and CBT programmes, senior managers were most interested in trainees’ satisfaction with training and their participation rates in classes. These areas are most often categorized as part of the process component of training evaluation. Managers may have assumed that training has its own effectiveness and, thus, the most important goal of training is to have employees take classes. They also appeared to believe that trainees’ satisfaction equalled successful training and, thus, evaluating trainees’ satisfaction was the most important.
  • 37. As a result, they asked HRD professionals in effect to focus on the process evaluation component, when evaluating their web-based and CBT programmes and, thus, the differences between these training approaches may be clearly revealed in the process evaluation component, an area of evaluation performed directly to satisfy their senior management’s goals. Human Resource Development International 93 Uncertain workforce development policy Another possible explanation for the results involves the Korean government’s policies regarding how organizations are reimbursed for the costs of WBT programmes. The interviewees stated that, in order to receive this funding, their organizations must conduct evaluation of WBT programmes, according to evaluation criteria set by the government, and report the results. The criteria required for evaluating WBT programmes include primarily completion rates and satisfaction rates of trainees. Thus, to receive government support, organizations might request their HRD professionals to frequently evaluate their WBT programmes for the process evaluation component. Accordingly, HRD professionals may be more familiar
  • 38. with how to appropriately evaluate the training approaches for this component than for the other three components and, thus, they might evaluate web-based and CBT programmes differently for this component. Possible confusion between input and process evaluation components The third point suggests that there might be some misunderstanding about the input and process evaluation components. The results showed that web-based and CBT programmes were more frequently evaluated for the input evaluation component than for the output and organizational context components of evaluation. However, it can hardly be concluded that the input component was as frequently evaluated by HRD professionals. The reason for the observed frequency in conducting input evaluation, based on the interview results, is that the input evaluation component was often considered to be a part of the process evaluation component, which means that it involved evaluating trainees’ satisfaction with training rather than intentionally conducting separate input evaluation. Thus, the results suggest that web-based and CBT programmes were basically evaluated for the process component, and that there are some differences between these training approaches in this component.
  • 39. Lack of HRD professionals’ knowledge and skills The results might also be explained by the fundamental lack of HRD professionals’ knowledge and skills in training evaluation. The results of both the survey and the interviews showed that HRD professionals perceived fewer barriers in the human factors than in the context and evaluation factors. The interview results further explained that since HRD professionals mostly used pre- existing survey forms for 20–30 years to evaluate their training programmes, which were not difficult to use, they may not be aware of their own lack of knowledge and expertise in the evaluation of both web-based and CBT programmes. Thus, they may not perceive any differences between web-based and CBT programmes in the human factors of evaluation barriers. In addition, current HRD professionals can have difficulty of performing higher- level evaluation with currently available evaluation methods and tools, thus, those in the study may rarely conduct this kind of evaluation for either training approach. Accordingly, they may not be able to recognize any differences between web-based and CBT programmes in evaluation barriers. 94 Y.J. Kong and R.L. Jacobs
  • 40. Uncertain original premise Finally, the original premise – that the evaluation of different training approaches should be performed differently – might be not as strong as initially assumed. Many researchers (Jung and Rha 2000; Rumble 2001; Strother 2002; Trentin 2000; Whalen and Wright 1999) have emphasized that different training approaches should be evaluated differently, and, following this premise, that the evaluation of web-based and CBT should be conducted differently. However, other researchers (Clark 2000; Garrison and Vaughan 2008) argue that the differences between training approaches such as web-based and CBT might be due to quality of course design, quality of the instructor, or potential individual differences among instructors rather than the delivery modes themselves. Garrison and Vaughan (2008) stated that sound learning theories such as constructivism and social cognition for learning may be more important than delivery modes for ensuring effective learning. Thus, as they suggested, effective teaching and learning principles should be considered more than the unique elements of delivery mode, when comparing the evaluation of different training approaches. Contribution to HRD This study contributes to both the theory and practice of
  • 41. evaluation for training programmes in the HRD field. Theoretically, employing a methodology that combines quantitative and qualitative elements, this study contributes to empirical research that has been conducted rarely in the HRD field. That is, the study provides deeper understanding of how organizations actually evaluate web-based and CBT, identifying evaluation barriers to the different training approaches, and how unique programme features and organizational and social circumstances result in different evaluation methods for web-based and CBT programmes. In addition, this study contributes to our understanding of the dynamic interactions among people who have different roles and responsibilities during an evaluation process, examining how evaluators respond differently to evaluate their web-based and CBT programmes. This study also provides some recommendations for further research. First, this study found that the lack of reliable evaluation methods and tools was one of the most prevalent reasons why organizations were not committed to evaluating their training programmes and avoided higher levels of training evaluation. Consequently, more active research is needed to develop evaluation methods and tools that are easy for practitioners to use and guide them to use these tools for evaluating different training approaches in appropriate ways.
  • 42. Second, unlike evidence from existing literature, this study showed that WBT programmes were more frequently evaluated than CBT programmes for the organizational context component of evaluation due to the Korean government’s support for online learning. Since this study is limited to examining seven companies in one country, Korea, more comparison studies among various countries and companies will be needed to determine to what degree a particular country or organization’s characteristics, culture, and circumstances influence different training evaluation approaches. Practically, the results of this study have three important implications for HRD professionals to improve current training evaluation in their organizations. First, this study found that the primary barrier was the context factors due to lack of Human Resource Development International 95 interest in and commitment to training evaluation on the part of senior management. Thus, senior management must deliberate thoughtfully on how to manage and maintain current training evaluation effectively. For example, HRD professionals need to revise current personnel policy and
  • 43. benefits related to training evaluation, focusing on ways of improving the quality of training programmes rather than increasing the quantity of online courses. Since organizations evaluate their HRD professionals’ performance based on the number of online courses purchased from vendors with the cheapest prices, HRD professionals concentrate more on increasing the quantity of courses rather than on increasing the quality of those courses. This approach can be hardly guaranteed to deliver good quality training courses to employees. Second, this study found that most participants were unaware of evaluation barriers that came from the lack of HRD professional’s knowledge and skills in training evaluation, since they mostly used and reused existing survey templates to evaluate trainees’ satisfaction for both web-based and CBT programmes. Thus, HRD professionals should improve their job competency in the area of current training evaluation in the workplace. For example, HRD professionals need to reflect on the adequacy of their current evaluation practices. They should take stock of what they currently do in the evaluation of their web-based or CBT programmes and examine whether they meet the basic requirements to successfully implement these training approaches. They also need to identify different requisite elements and barriers to evaluate web-based
  • 44. and CBT programmes and implement an appropriate evaluation for the different training approaches. In addition, they should examine and evaluate how various stakeholders correspond to multiple goals of a specific training programme to help select the best training approaches for their organizations. Professional development of training evaluation is a great opportunity for HRD professionals to improve these job competencies. Finally, national evaluation policy makers in Korea should provide organiza- tions with effective guidance and standards to conduct appropriate evaluation for web-based and CBT programmes. Especially, they should be aware of the negative results related to their support for online courses. They should accurately investigate actual prices of online courses in the market and re-establish current evaluation criteria in order to reduce the negative practice of focusing on the quantity of online courses with cheapest prices rather than the quality of the courses. They should also request diverse evaluation reports that go beyond trainees’ satisfaction with training to help organizations determine the best training approaches for their employees. References Alkin, M. 1985. A guide for evaluation decision makers. Thousand Oaks, CA: Sage.
  • 45. Alliger, G., and E. Janak. 1989. Kirkpatirck’s levels of training criteria: Thirty years later. Personnel Psychology 42, no. 2: 331–42. Bartley, S., and J. Golek. 2004. Evaluating the cost effectiveness of online and face-to-face instruction. Educational Technology & Society 7, no. 4: 167–75. Bassi, L., G. Benson, and S. Cheney. 1996. The top ten trends. Training and Development 50, no. 11: 28–42. Bradsford, J., A. Brown, and R. Cocking. 1999. How people learn: Brain, mind experience and school. Washington, DC: National Academy Press. 96 Y.J. Kong and R.L. Jacobs Clark, R. 2000. Evaluating distance education: Strategies and cautions. Quarterly Review of Distance Education 1, no. 1: 3–16. Curtain, R. 2002. Online learning: How cost-effective? In Online delivery in the vocational education and training sector: Improving cost effectiveness, ed. Richard Curtain, 125–43. Leabrook, SA, Australia: Australian National Training Authority, NCVER. Fitzpatrick, J., J. Sanders, and B. Worthen. 2004. Program evaluation: Alternative approaches and practical guidelines. Boston, MA: Pearson Education, Inc.
  • 46. Garrison, R., and N. Vaughan. 2008. Part one: Community of inquiry framework. In Blended learning in higher education, ed. Garrison R. and V. Vaughan, 1–68. San Francisco, CA: Jossey-Bass. Hofstetter, C., and M. Alkin. 2003. Evaluation use revisited. In International Handbook of Educational Evaluation, ed. T. Kellaghan and D.L. Stufflebeam, 197–222. London, Great Britain: Kluwer Academic Publishers. Holton, E. 1996. The flawed four-level evaluation model. Human Resource Development Quarterly 7, no. 1: 5–21. Holton, E., and S. Naquin. 2005. A critical analysis of HRD evaluation models from a decision-making perspective. Human Resource Development Quarterly 16, no. 2: 257– 80. Jacobs, R. 2003. Structured on-the-job training: Unleashing employee expertise in the workplace. San Francisco, CA: Berrett-Koehler. Johnson, R. 1998. Toward a theoretical model of evaluation utilization. Evaluation and Program Planning 21, no. 1: 93–110. Jung, I.S., and I. Rha. 2000. Effectiveness and cost- effectiveness of online education: A review of literature. Education Technology 40, no. 4: 57–60. Kapp, K., and C. McKeague. 2002. Blend two proven training methods to improve results.
  • 47. Chemical Engineering 109, no. 8: 191–4. Kirkpatrick, D.L. 1998. Evaluating training programs: The four levels. San Francisco, CA: Berrett-Koehler, 1998. Kraiger, K. 2002. Decision-based evaluation. In Creating, implementing, and managing effective training and development systems in organizations: State-of-the-art lessons for practice, ed. Kurt Kraiger, 331–75. San Francisco, CA: Jossey- Bass. Leviton, L. 2003. Evaluation use: Advances, challenges and applications. American Journal of Evaluation 24, no. 4: 525–35. Lincoln, Y.S., and E.G. Guba. 1985. Naturalistic Inquiry. Beverly Hills, CA: Sage Publications, Inc. Maxwell, J. 2005. Qualitative research design: An interactive approach. Thousand Oaks, CA: Sage Publications, 2005. Olson T., and R. Wisher. 2002. The effectiveness of web-based instruction: An initial inquiry. International Review of Research in Open and Distance Learning 3, no. 2. Phillips, J. 1997. Handbook of training evaluation and measurement methods. Houston, TX: Gulf Publishing. Rumble, G. 2001. The costs and costing of networked learning. Journal of Asynchronous
  • 48. Learning Network 5, no. 2: 75–96. Russ-Eft, D., and H. Preskill. 2011. Evaluation in organizations. Cambridge, Massachusetts: Perseus Publishing. Shulha, L., and J. Cousins. 1997. Evaluation use: Theory, research, and practice since 1986. Evaluation Practice 18, no. 3: 195–208. Strother, J. 2002. An assessment of the effectiveness of e- learning in corporate training programs. The International Review of Research in Open and Distance Learning 3, no. 10. Stufflebeam, D., W. Foley, W. Gephart, L. Hammond, H. Merriman, and M. Provus. 1971. Educational evaluation and decision-making in education. Itasca, IL: Peacock. Taut, S., and M. Alkin. 2003. Program staff perceptions of barriers to evaluation implementation. American Journal of Evaluation 24, no. 2: 213– 26. Trentin, V. 2000. The evaluation of online courses. Journal of Computer Assisted Learning 16, no. 3: 259–70. Whalen, T., and D. Wright. 1999. Methodology for cost-benefit analysis of web-based tele- learning program: Case study of the bell institute. The American Journal of Distance Education 13, no. 1: 24–44. Human Resource Development International 97
  • 49. Wisher, R.A., and T.M. Olson. 2003. The effectiveness of web- based instruction. Alexandria, VA: United States Army Research Institute for the Behavioural and Social Sciences Publication. Yanmil, S., and G. McLean. 2001. Theories supporting transfer of training. Human Resource Development Quarterly 12, no. 2: 195–208. 98 Y.J. Kong and R.L. Jacobs Copyright of Human Resource Development International is the property of Routledge and its content may not be copied or emailed to multiple sites or posted to a listserv without the copyright holder's express written permission. However, users may print, download, or email articles for individual use. Task 1: half Page, times new roman, 12, double spaced. What is the difference between Research and Evaluation? How do you see the role of evaluation in organizations? What are the ethical issues in evaluation practice? Task 2: Reflection of the attached article(Times, new roman, 12, double spaced). The reflection should be 400 words and should include the following: 1- Section 1: Rationale for choosing the article
  • 50. 2- Section 2: Critique and connection to the course(course name: Evaluation models in Human Resource development) 3- Section 3: Article Reference (and other references if used in the Critique)