SlideShare a Scribd company logo
The new engl and jour nal o f medicine
n engl j med 356;4 www.nejm.org january 25, 2007 387
Medical Education
Malcolm Cox, M.D., and David M. Irby, Ph.D., Editors
Review article
Medical Education
Malcolm Cox, M.D., and David M. Irby, Ph.D., Editors
Assessment in Medical Education
Ronald M. Epstein, M.D.
FromtheDepartmentsofFamilyMedicine,
Psychiatry, and Oncology and the Roch-
ester Center to Improve Communication
in Health Care, University of Rochester
School of Medicine and Dentistry, Roch-
ester, NY. Address reprint requests to Dr.
Epstein at 1381 South Ave., Rochester,
NY 14620, or at ronald_epstein@urmc.
rochester.edu.
N Engl J Med 2007;356:387-96.
Copyright © 2007 Massachusetts Medical Society.
As an attending physician working with a student for a week, you receive a form that
asks you to evaluate the student’s fund of knowledge, procedural skills, professional-
ism, interest in learning, and “systems-based practice.” You wonder which of these
attributes you can reliably assess and how the data you provide will be used to further
the student’s education. You also wonder whether other tests of knowledge and com-
petence that students must undergo before they enter practice are equally problematic.
I
n one way or another, most practicing physicians are involved in
assessing the competence of trainees, peers, and other health professionals. As
the example above suggests, however, they may not be as comfortable using
educational assessment tools as they are using more clinically focused diagnostic
tests. This article provides a conceptual framework for and a brief update on com-
monly used and emerging methods of assessment, discusses the strengths and
limitations of each method, and identifies several challenges in the assessment of
physicians’ professional competence and performance.
Competence and Performance
Elsewhere, Hundert and I have defined competence in medicine as “the habitual and
judicious use of communication, knowledge, technical skills, clinical reasoning,
emotions, values, and reflection in daily practice for the benefit of the individuals
and communities being served.”1 In the United States, the assessment of medical
residents, and increasingly of medical students, is largely based on a model that
was developed by the Accreditation Council for Graduate Medical Education (ACGME).
This model uses six interrelated domains of competence: medical knowledge, patient
care, professionalism, communication and interpersonal skills, practice-based learn-
ing and improvement, and systems-based practice.2
Competence is not an achievement but rather a habit of lifelong learning3; as-
sessment plays an integral role in helping physicians identify and respond to their
own learning needs. Ideally, the assessment of competence (what the student or
physician is able to do) should provide insight into actual performance (what he or
she does habitually when not observed), as well as the capacity to adapt to change,
find and generate new knowledge, and improve overall performance.4
Competence is contextual, reflecting the relationship between a person’s abilities
and the tasks he or she is required to perform in a particular situation in the real
world.5 Common contextual factors include the practice setting, the local prevalence
of disease, the nature of the patient’s presenting symptoms, the patient’s education-
al level, and other demographic characteristics of the patient and of the physician.
Many aspects of competence, such as history taking and clinical reasoning, are
also content-specific and not necessarily generalizable to all situations. A student’s
Copyright © 2007 Massachusetts Medical Society. All rights reserved.
Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .
The new engl and jour nal o f medicine
n engl j med 356;4 www.nejm.org january 25, 2007
388
clinical reasoning may appear to be competent in
areas in which his or her base of knowledge is
well organized and accessible6 but may appear to
be much less competent in unfamiliar territory.7
However, some important skills (e.g., the ability
to form therapeutic relationships) may be less
dependent on content.8
Competence is also developmental. Habits of
mind and behavior and practical wisdom are
gained through deliberate practice9 and reflection
on experience.10-14 Students begin their training
at a novice level, using abstract, rule-based formu-
las that are removed from actual practice. At high-
er levels, students apply these rules differentially
to specific situations. During residency, trainees
make judgments that reflect a holistic view of a
situation and eventually take diagnostic shortcuts
based on a deeper understanding of underlying
principles. Experts are able to make rapid, context-
based judgments in ambiguous real-life situations
and have sufficient awareness of their own cogni-
tive processes to articulate and explain how they
recognize situations in which deliberation is es-
sential. Development of competence in different
contexts and content areas may proceed at differ-
ent rates. Context and developmental level also
interact. Although all clinicians may perform at
a lower level of competence when they are tired,
distracted, or annoyed, the competence of less
experienced clinicians may be particularly suscep-
tible to the influence of stress.15,16
Goals of Assessment
Over the past decade, medical schools, postgrad-
uate training programs, and licensing bodies have
made new efforts to provide accurate, reliable, and
timely assessments of the competence of trainees
and practicing physicians.1,2,17 Such assessments
have three main goals: to optimize the capabili-
ties of all learners and practitioners by providing
motivation and direction for future learning, to
protect the public by identifying incompetent phy-
sicians, and to provide a basis for choosing ap-
plicants for advanced training.
Assessment can be formative (guiding future
learning,providingreassurance,promotingreflec-
tion, and shaping values) or summative (making
an overall judgment about competence, fitness to
practice, or qualification for advancement to high-
er levels of responsibility). Formative assessments
provide benchmarks to orient the learner who is
approaching a relatively unstructured body of
knowledge. They can reinforce students’ intrinsic
motivation to learn and inspire them to set higher
standards for themselves.18 Although summative
assessments are intended to provide professional
self-regulation and accountability, they may also
act as a barrier to further practice or training.19
A distinction should be made between assess-
ments that are suitable only for formative use
and those that have sufficient psychometric rigor
for summative use. This distinction is especially
important in selecting a method of evaluating
competence for high-stakes assessments (i.e.,
licensing and certification examinations). Corre-
spondingly, summative assessments may not pro-
vide sufficient feedback to drive learning.20 How-
ever, because students tend to study that which
they expect to be tested on, summative assess-
ment may influence learning even in the absence
of feedback.
Assessment Methods
All methods of assessment have strengths and
intrinsic flaws (Table 1). The use of multiple ob-
servations and several different assessment meth-
ods over time can partially compensate for flaws
in any one method.1,21 Van der Vleuten22 describes
five criteria for determining the usefulness of a
particular method of assessment: reliability (the
degree to which the measurement is accurate and
reproducible), validity (whether the assessment
measures what it claims to measure), impact on
future learning and practice, acceptability to learn-
ers and faculty, and costs (to the individual trainee,
the institution, and society at large).
Written Examinations
Written examination questions are typically classi-
fied according to whether they are open-ended or
multiple choice. In addition, questions can be
“context rich” or “context poor.”23 Questions with
rich descriptions of the clinical context invite the
more complex cognitive processes that are char-
acteristic of clinical practice.24 Conversely, context-
poor questions can test basic factual knowledge
but not its transferability to real clinical problems.
Multiple-choice questions are commonly used
for assessment because they can provide a large
number of examination items that encompass
many content areas, can be administered in a rel-
atively short period, and can be graded by com-
Copyright © 2007 Massachusetts Medical Society. All rights reserved.
Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .
Medical Education
n engl j med 356;4 www.nejm.org january 25, 2007 389
puter. These factors make the administration of
the examination to large numbers of trainees
straightforward and standardized.25 Formats that
ask the student to choose the best answer from a
list of possible answers are most commonly used.
However, newer formats may better assess pro-
cesses of diagnostic reasoning. Key-feature items
focus on critical decisions in particular clinical
cases.26 Script-concordance items present a situ-
ation (e.g., vaginal discharge in a patient), add a
piece of information (dysuria), and ask the exam-
inee to assess the degree to which this new in-
formation increases or decreases the probability
of a particular outcome (acute salpingitis due to
Chlamydiatrachomatis).27 Because the situations por-
trayed are ambiguous, script-concordance items
may provide insight into clinical judgment in the
real world. Answers to such items have been
shown to correlate with the examinee’s level of
training and to predict future performance on
oral examinations of clinical reasoning.28
Multiple-choice questions that are rich in con-
text are difficult to write, and those who write
them tend to avoid topics — such as ethical dilem-
mas or cultural ambiguities — that cannot be
asked about easily.29 Multiple-choice questions
may also create situations in which an examinee
can answer a question by recognizing the correct
option, but could not have answered it in the
absence of options.23,30 This effect, called cueing,
is especially problematic when diagnostic reason-
ing is being assessed, because premature closure
— arriving at a decision before the correct diag-
nosis has been considered — is a common reason
for diagnostic errors in clinical practice.31,32 Ex-
tended matching items (several questions, all with
the same long list of possible answers), as well
as open-ended short-answer questions, can mini-
mize cueing.23 Structured essays also preclude
cueing. In addition, they involve more complex
cognitive processes and allow for more contex-
tualized answers than do multiple-choice ques-
tions. When clear grading guidelines are in place,
structured essays can be psychometrically robust.
Assessments by Supervising Clinicians
Supervising clinicians’ observations and impres-
sions of students over a specific period remain the
most common tool used to evaluate performance
with patients. Students and residents most com-
monly receive global ratings at the end of a rota-
tion, with comments from a variety of supervis-
ing physicians. Although subjectivity can be a
problem in the absence of clearly articulated
standards, a more important issue is that direct
observation of trainees while they are interacting
with patients is too infrequent.33
Direct Observation or Video Review
The “long case”34 and the “mini–clinical-evalua-
tion exercise” (mini-CEX)35 have been developed
so that learners will be directly observed more fre-
quently. In these assessments, a supervising phy-
sician observes while a trainee performs a focused
history taking and physical examination over a
period of 10 to 20 minutes. The trainee then pres-
ents a diagnosis and a treatment plan, and the
faculty member rates the resident and may pro-
vide educational feedback. Structured exercises
with actual patients under the observation of the
supervising physician can have the same level of
reliability as structured examinations using stan-
dardized patients34,36 yet encompass a wider range
of problems, physical findings, and clinical set-
tings. Direct observation of trainees in clinical
settings can be coupled with exercises that train-
ees perform after their encounters with patients,
such as oral case presentations, written exercises
that assess clinical reasoning, and literature search-
es.8,37 In addition, review of videos of encounters
with patients offers a powerful means of evaluat-
ing and providing feedback on trainees’ skills in
clinical interactions.8,38
Clinical Simulations
Standardized patients — actors who are trained
to portray patients consistently on repeated occa-
sions — are often incorporated into objective
structured clinical examinations (OSCEs), which
consist of a series of timed “stations,” each one
focused on a different task. Since 2004, these ex-
aminations have been part of the U.S. Medical
Licensing Examination that all senior medical stu-
dents take.39 The observing faculty member or
the standardized patient uses either a checklist
of specific behaviors or a global rating form to
evaluate the student’s performance.40 The check-
list might include items such as “asked if the pa-
tient smoked” and “checked ankle reflexes.” The
global rating form might ask for a rating of how
well the visit was organized and whether the stu-
dent was appropriately empathetic. A minimum of
10 stations, which the student usually visits over
the course of 3 to 4 hours, is necessary to achieve
Copyright © 2007 Massachusetts Medical Society. All rights reserved.
Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .
T
h
e
n
e
w
e
ng
l
a
n
d
j
o
u
r
na
l
o
f
m
e
dic
i
n
e
n
engl
j
med
356;4
www.nejm.org
january
25,
2007
390
Table 1. Commonly Used Methods of Assessment.
Method Domain Type of Use Limitations Strengths
Written exercises
Multiple-choice questions in
either single-best-answer or
extended matching format
Knowledge, ability to solve
problems
Summative assessments within
courses or clerkships; nation-
al in-service, licensing, and
certification examinations
Difficult to write, especially in cer-
tain content areas; can result in
cueing; can seem artificial and
removed from real situations
Can assess many content areas in rela-
tively little time, have high reliability,
can be graded by computer
Key-feature and script-concor-
dance questions
Clinical reasoning, problem-
solving ability, ability to
apply knowledge
National licensing and certifica-
tion examinations
Not yet proven to transfer to real-
life situations that require clini-
cal reasoning
Assess clinical problem-solving ability,
avoid cueing, can be graded by com-
puter
Short-answer questions Ability to interpret diagnostic
tests, problem-solving
ability, clinical reasoning
skills
Summative and formative as-
sessments in courses and
clerkships
Reliability dependent on training of
graders
Avoid cueing, assess interpretation and
problem-solving ability
Structured essays Synthesis of information, in-
terpretation of medical
literature
Preclinical courses, limited use
in clerkships
Time-consuming to grade, must work
to establish interrater reliability,
long testing time required to en-
compass a variety of domains
Avoid cueing, use higher-order cogni-
tive processes
Assessments by supervising clinicians
Global ratings with comments
at end of rotation
Clinical skills, communication,
teamwork, presentation
skills, organization, work
habits
Global summative and some-
times formative assess-
ments in clinical rotations
Often based on second-hand re-
ports and case presentations
rather than on direct observa-
tion, subjective
Use of multiple independent raters can
overcome some variability due to
subjectivity
Structured direct observation
with checklists for ratings
(e.g., mini–clinical-evaluation
exercise or video review)
Communication skills, clinical
skills
Limited use in clerkships and
residencies, a few board-
certification examinations
Selective rather than habitual be-
haviors observed, relatively
time-consuming
Feedback provided by credible experts
Oral examinations Knowledge, clinical
reasoning
Limited use in clerkships and com-
prehensive medical school
assessments, some board-
certification examinations
Subjective, sex and race bias has
been reported, time-consuming,
require training of examiners,
summative assessments need
two or more examiners
Feedback provided by credible experts
Copyright
©
2007
Massachusetts
Medical
Society.
All
rights
reserved.
Downloaded
from
www.nejm.org
by
ALEXANDRE
A.
HOLANDA
MD
on
February
13,
2007
.
Medical
Education
n
engl
j
med
356;4
www.nejm.org
january
25,
2007
391
Clinical simulations
Standardized patients and ob-
jective structured clinical
examinations
Some clinical skills, inter-
personal behavior,
communication skills
Formative and summative as-
sessments in courses, clerk-
ships, medical schools, na-
tional licensure examinations,
board certification in Canada
Timing and setting may seem artifi-
cial, require suspension of dis-
belief, checklists may penalize
examinees who use shortcuts,
expensive
Tailored to educational goals; reliable,
consistent case presentation and
ratings; can be observed by faculty
or standardized patients; realistic
Incognito standardized patients Actual practice habits Primarily used in research;
some courses, clerkships,
and residencies use for for-
mative feedback
Requires prior consent, logistically
challenging, expensive
Very realistic, most accurate way of
assessing clinician’s behavior
High-technology simulations Procedural skills, teamwork,
simulated clinical dilem-
mas
Formative and some summative
assessment
Timing and setting may seem artifi-
cial, require suspension of dis-
belief, checklists may penalize
examinees who use shortcuts,
expensive
Tailored to educational goals, can be
observed by faculty, often realistic
and credible
Multisource (“360-degree”) assessments
Peer assessments Professional demeanor,
work habits, interpersonal
behavior, teamwork
Formative feedback in courses
and comprehensive medical
school assessments, forma-
tive assessment for board
recertification
Confidentiality, anonymity, and
trainee buy-in essential
Ratings encompass habitual behaviors,
credible source, correlates with future
academic and clinical performance
Patient assessments Ability to gain patients’ trust;
patient satisfaction, com-
munication skills
Formative and summative, board
recertification, use by insur-
ers to determine bonuses
Provide global impressions rather
than analysis of specific behav-
iors, ratings generally high with
little variability
Credible source of assessment
Self-assessments Knowledge, skills, attitudes,
beliefs, behaviors
Formative Do not accurately describe actual
behavior unless training and
feedback provided
Foster reflection and development of
learning plans
Portfolios All aspects of competence,
especially appropriate for
practice-based learning
and improvement and
systems-based practice
Formative and summative uses
across curriculum and with-
in clerkships and residency
programs, used by some
U.K. medical schools and
specialty boards
Learner selects best case material,
time-consuming to prepare and
review
Display projects for review, foster reflec-
tion and development of learning
plans
Copyright
©
2007
Massachusetts
Medical
Society.
All
rights
reserved.
Downloaded
from
www.nejm.org
by
ALEXANDRE
A.
HOLANDA
MD
on
February
13,
2007
.
The new engl and jour nal o f medicine
n engl j med 356;4 www.nejm.org january 25, 2007
392
a reliability of 0.85 to 0.90.41 Under these condi-
tions, structured assessments with the use of
standardized patients are as reliable as ratings of
directly observed encounters with real patients
and take about the same amount of time.42
Interactions with standardized patients can be
tailored to meet specific educational goals, and
the actors who portray the patients can reliably
rate students’ performance with respect to history
taking and physical examinations. Faculty mem-
bers who observe encounters with standardized
patients can offer additional insights on trainees’
clinical judgment and the overall coherence of
the history taking or physical examination. Unan-
nounced standardized patients, who with the ex-
aminees’ prior approval present incognito in ac-
tual clinical settings, have been used in health
services research to evaluate examinees’ diagnos-
tic reasoning, treatment decisions, and communi-
cation skills.43-46 The use of unannounced stan-
dardized patients may prove to be particularly
valuable in the assessment of higher-level trainees
and physicians in practice.
The use of simulation to assess trainees’ clini-
cal skills in intensive care and surgical settings
is on the rise.47 Simulations involving sophisticat-
ed mannequins with heart sounds, respirations,
oximeter readings, and pulses that respond to a
variety of interventions can be used to assess how
individuals or teams manage unstable vital signs.
Surgical simulation centers now routinely use
high-fidelity computer graphics and hands-on
manipulation of surgical instruments to create
a multisensory environment. High-technology
simulation is seen increasingly as an important
learning aid and may prove to be useful in the
assessment of knowledge, clinical reasoning, and
teamwork.
Multisource (“360-Degree”) Assessments
Assessments by peers, other members of the clin-
ical team, and patients can provide insight into
trainees’ work habits, capacity for teamwork, and
interpersonal sensitivity.48-50 Although there are
few published data on outcomes of multisource
feedback in medical settings, several large pro-
grams are being developed, including one for all
first- and second-year house officers in the United
Kingdom and another for all physicians under-
going recertification in internal medicine in the
United States. Multisource feedback is most effec-
tive when it includes narrative comments as well
as statistical data, when the sources are recog-
nized as credible, when the feedback is framed
constructively, and when the entire process is ac-
companied by good mentoring and follow-up.51
Recent studies of peer assessments suggest
that when trainees receive thoughtful ratings and
comments by peers in a timely and confidential
manner, along with support from advisers to help
them reflect on the reports, they find the process
powerful, insightful, and instructive.51,52 Peer as-
sessments have been shown to be consistent re-
gardless of the way the raters are selected. Such
assessments are stable from year to year53 and
predict subsequent class rankings as well as sub-
sequent ratings by supervisors.54 Peer assessments
depend on trust and require scrupulous attention
to confidentiality. Otherwise they can be under-
mining, destructive, and divisive.
Although patients’ ratings of clinical perfor-
mance are valuable in principle, they pose several
problems. As many as 50 patient surveys may be
necessary to achieve satisfactory reliability.55 Pa-
tients who are seriously ill often do not complete
surveys; those who do tend to rate physicians less
favorably than do patients who have milder con-
ditions.56 Furthermore, patients are not always
able to discriminate among the elements of clin-
ical practice,57 and their ratings are typically
high. These limitations make it difficult to use
patient reports as the only tool for assessing
clinical performance. However, ratings by nurses
can be valuable. Such ratings have been found to
be reliable with as few as 6 to 10 reports,58 and
they correlate with both patients’ and faculty
members’ ratings of the interpersonal aspects of
trainees’ performance.59
Fundamental cognitive limitations in the abil-
ity of humans to know themselves as others see
them restrict the usefulness of self-assessment.
Furthermore, rating oneself on prior clinical per-
formance may not achieve another important goal
of self-assessment: the ability to monitor oneself
from moment to moment during clinical prac-
tice.10,60 A physician must possess this ability in
order to meet patients’ changing needs, to recog-
nize the limits of his or her own competence,
and to manage unexpected situations.
Portfolios
Portfolios include documentation of and reflection
about specific areas of a trainee’s competence.
This evidence is combined with self-reflection.61
Copyright © 2007 Massachusetts Medical Society. All rights reserved.
Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .
Medical Education
n engl j med 356;4 www.nejm.org january 25, 2007 393
In medicine, just as in the visual arts, portfolios
demonstrate a trainee’s development and techni-
cal capacity. They can include chart notes, refer-
ral letters, procedure logs, videotaped consulta-
tions,peerassessments,patientsurveys,literature
searches, quality-improvement projects, and any
other type of learning material. Portfolios also
frequently include self-assessments, learning
plans, and reflective essays. For portfolios to be
maximally effective, close mentoring is required
in the assembly and interpretation of the contents;
considerable time can be expended in this effort.
Portfolios are most commonly used in formative
assessments, but their use for summative evalua-
tions and high-stakes decisions about advance-
ment is increasing.20
Challenges in Assessment
New Domains of Assessment
There are several domains in which assessment is
in its infancy and remains problematic. Quality
of care and patient safety depend on effective
teamwork,62 and teamwork training is empha-
sized as an essential element of several areas of
competence specified by the ACGME, yet there is
no validated method of assessing teamwork. Ex-
perts do not agree on how to define professional-
ism — let alone how best to measure it.63 Dozens
of scales that rate communication are used in
medical education and research,64 yet there is
little evidence that any one scale is better than
another; furthermore, the experiences that patients
report often differ considerably from ratings given
by experts.65
Multimethod and Longitudinal Assessment
The use of multiple methods of assessment can
overcome many of the limitations of individual
assessment formats.8,22,36,66 Variation of the clin-
ical context allows for broader insights into com-
petence, the use of multiple formats provides
greater variety in the areas of content that are
evaluated, and input from multiple observers pro-
vides information on distinct aspects of a trainee’s
performance. Longitudinal assessment avoids ex-
cessive testing at any one point in time and serves
as the foundation for monitoring ongoing profes-
sional development.
In the example at the beginning of this article,
a multimethod assessment might include direct
observation of the student interacting with several
patients at different points during the rotation,
a multiple-choice examination with both “key
features” and “script-concordance” items to as-
sess clinical reasoning, an encounter with a stan-
dardized patient followed by an oral examination
to assess clinical skills in a standardized setting,
written essays that would require literature search-
es and synthesis of the medical literature on the
basic science or clinical aspects of one or more
of the diseases the student encountered, and peer
assessments to provide insights into interperson-
al skills and work habits.
The combination of all these results into a
portfolio resembles the art of diagnosis; it de-
mands that the student synthesize various bits
and types of information in order to come up
with an overall picture. Although a few medical
schools have begun to institute longitudinal as-
sessments that use multiple methods,8 the best
way to deal with the quantity and the qualitatively
different types of data that the process generates
is not yet clear. New ways of combining qualita-
tive and quantitative data will be required if port-
folio assessments are to find widespread applica-
tion and withstand the test of time.
Standardization of Assessment
Although accrediting organizations specify broad
areas that the curriculum should cover and assess,
for the most part individual medical schools make
their own decisions about methods and standards
of assessment. This model may have the advan-
tage of ensuring consistency between the curricu-
lum and assessment, but it also makes it difficult
to compare students across medical schools for
the purpose of subsequent training.67 The ideal
balance between nationally standardized and
school-specific assessment remains to be deter-
mined.Furthermore,withinagivenmedicalschool,
all students may not require the same package of
assessments — for example, initial screening ex-
aminations may be followed by more extensive
testing for those who have difficulties.
Assessment and Learning
It is generally acknowledged that assessment drives
learning; however, assessment can have both in-
tended and unintended consequences.22 Students
study more thoughtfully when they anticipate cer-
tain examination formats,68 and changes in the
format can shift their focus to clinical rather than
theoretical issues.69 Assessment by peers seems
Copyright © 2007 Massachusetts Medical Society. All rights reserved.
Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .
The new engl and jour nal o f medicine
n engl j med 356;4 www.nejm.org january 25, 2007
394
to promote professionalism, teamwork, and com-
munication.52 The unintended effects of assess-
ment include the tendency for students to cram
for examinations and to substitute superficial
knowledge for reflective learning.
Assessment of Expertise
The assessment of trainees and physicians who
have higher levels of expertise presents particular
challenges. Expertise is characterized by unique,
elaborated, and well-organized bodies of knowl-
edge that are often revealed only when they are
triggered by characteristic clinical patterns.70,71
Thus, experts who are unable to access their
knowledge in artificial testing situations but who
make sound judgments in practice may do poorly
on some tests that are designed to assess commu-
nication skills, knowledge, or reasoning. Further-
more, clinical expertise implies the practical wis-
dom to manage ambiguous and unstructured
problems, balance competing explanations, avoid
premature closure, note exceptions to rules and
principles, and — even when under stress —
choose one of the several courses of action that
are acceptable but imperfect. Testing either induc-
tive thinking (the organization of data to gener-
ate possible interpretations) or deductive thinking
(the analysis of data to discern among possibili-
ties) in situations in which there is no consensus
on a single correct answer presents formidable
psychometric challenges.
Assessment and Future Performance
The evidence that assessment protects the public
from poor-quality care is both indirect and scarce;
it consists of a few studies that show correlations
between assessment programs that use multiple
methods and relatively crude estimates of quality
such as diagnostic testing, prescribing, and refer-
ral patterns.72 Correlating assessment with future
performance is difficult not only because of in-
adequacies in the assessment process itself but
also because relevant, robust measures of out-
come that can be directly attributed to the effects
of training have not been defined. Current efforts
to measure the overall quality of care include pa-
tient surveys and analyses of institutional and
practice databases. When these new tools are re-
fined, they may provide a more solid foundation
for research on educational outcomes.
Conclusions
Considering all these challenges, current assess-
ment practices would be enhanced if the principles
summarized in Table 2 were kept clearly in mind.
The content, format, and frequency of assessment,
as well as the timing and format of feedback,
should follow from the specific goals of the med-
ical education program. The various domains of
competence should be assessed in an integrated,
coherent, and longitudinal fashion with the use
of multiple methods and provision of frequent
and constructive feedback. Educators should be
mindful of the impact of assessment on learning,
the potential unintended effects of assessment,
the limitations of each method (including cost),
and the prevailing culture of the program or in-
stitution in which the assessment is occurring.
Assessment is entering every phase of profes-
sional development. It is now used during the
medical school application process,73 at the start
of residency training,74 and as part of the “main-
Table 2. Principles of Assessment.
Goals of assessment
Provide direction and motivation for future learning, including knowledge,
skills, and professionalism
Protect the public by upholding high professional standards and screening
out trainees and physicians who are incompetent
Meet public expectations of self-regulation
Choose among applicants for advanced training
What to assess
Habits of mind and behavior
Acquisition and application of knowledge and skills
Communication
Professionalism
Clinical reasoning and judgment in uncertain situations
Teamwork
Practice-based learning and improvement
Systems-based practice
How to assess
Use multiple methods and a variety of environments and contexts to capture
different aspects of performance
Organize assessments into repeated, ongoing, contextual, and developmen-
tal programs
Balance the use of complex, ambiguous real-life situations requiring reason-
ing and judgment with structured, simplified, and focused assessments
of knowledge, skills, and behavior
Include directly observed behavior
Use experts to test expert judgment
Use pass–fail standards that reflect appropriate developmental levels
Provide timely feedback and mentoring
Cautions
Be aware of the unintended effects of testing
Avoid punishing expert physicians who use shortcuts
Do not assume that quantitative data are more reliable, valid, or useful than
qualitative data
Copyright © 2007 Massachusetts Medical Society. All rights reserved.
Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .
Medical Education
n engl j med 356;4 www.nejm.org january 25, 2007 395
tenance of certification” requirements that several
medical boards have adopted.75 Multiple methods
of assessment implemented longitudinally can
provide the data that are needed to assess train-
ees’ learning needs and to identify and remediate
suboptimal performance by clinicians. Decisions
about whether to use formative or summative
assessment formats, how frequently assessments
should be made, and what standards should be
in place remain challenging. Educators also face
the challenge of developing tools for the assess-
ment of qualities such as professionalism, team-
work, and expertise that have been difficult to
define and quantify.
No potential conflict of interest relevant to this article was
reported.
I thank Tana Grady-Weliky, M.D., Stephen Lurie, M.D., Ph.D.,
John McCarthy, M.S., Anne Nofziger, M.D., and Denham Ward,
M.D., Ph.D., for helpful suggestions.
References
Epstein RM, Hundert EM. Defining
and assessing professional competence.
JAMA 2002;287:226-35.
Batalden P, Leach D, Swing S, Dreyfus
H, Dreyfus S. General competencies and
accreditation in graduate medical educa-
tion. Health Aff (Millwood) 2002;21(5):
103-11.
Leach DC. Competence is a habit.
JAMA 2002;287:243-4.
Fraser SW, Greenhalgh T. Coping with
complexity: educating for capability. BMJ
2001;323:799-803.
Klass D. Reevaluation of clinical com-
petency. Am J Phys Med Rehabil 2000;79:
481-6.
Bordage G, Zacks R. The structure of
medical knowledge in the memories of
medical students and general practitioners:
categories and prototypes. Med Educ 1984;
18:406-16.
Gruppen LD, Frohna AZ. Clinical rea-
soning. In: Norman GR, Van Der Vleuten
CP, Newble DI, eds. International hand-
book of research in medical education.
Part 1. Dordrecht, the Netherlands: Kluwer
Academic, 2002:205-30.
Epstein RM, Dannefer EF, Nofziger
AC, et al. Comprehensive assessment of
professional competence: the Rochester ex-
periment. Teach Learn Med 2004;16:186-
96.
Ericsson KA. Deliberate practice and
the acquisition and maintenance of expert
performance in medicine and related do-
mains. Acad Med 2004;79:Suppl:S70-S81.
Epstein RM. Mindful practice. JAMA
1999;282:833-9.
Schon DA. Educating the reflective
practitioner. San Francisco: Jossey-Bass,
1987.
Epstein RM. Mindful practice in ac-
tion. II. Cultivating habits of mind. Fam
Syst Health 2003;21:11-7.
Dreyfus HL. On the Internet (thinking
in action). New York: Routledge, 2001.
Eraut M. Learning professional pro-
cesses: public knowledge and personal
experience. In: Eraut M, ed. Developing
professional knowledge and competence.
London: Falmer Press, 1994:100-22.
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
11.
12.
13.
14.
Shanafelt TD, Bradley KA, Wipf JE,
Back AL. Burnout and self-reported patient
care in an internal medicine residency
program. Ann Intern Med 2002;136:358-
67.
Borrell-Carrio F, Epstein RM. Prevent-
ing errors in clinical practice: a call for self-
awareness. Ann Fam Med 2004;2:310-6.
Leung WC. Competency based medical
training: review. BMJ 2002;325:693-6.
Friedman Ben-David M. The role of as-
sessment in expanding professional hori-
zons. Med Teach 2000;22:472-7.
Sullivan W. Work and integrity: the
crisis and promise of professionalism in
America. 2nd ed. San Francisco: Jossey-
Bass, 2005.
Schuwirth L, van der Vleuten C. Merg-
ing views on assessment. Med Educ 2004;
38:1208-10.
Wass V, Van der Vleuten C, Shatzer J,
Jones R. Assessment of clinical compe-
tence. Lancet 2004;357:945-9.
Van Der Vleuten CPM. The assessment
of professional competence: developments,
research and practical implications. Adv
Health Sci Educ 1996;1:41-67.
Schuwirth LW, van der Vleuten CP.
Different written assessment methods:
what can be said about their strengths
and weaknesses? Med Educ 2004;38:974-9.
Schuwirth LW, Verheggen MM, van
der Vleuten CP, Boshuizen HP, Dinant GJ.
Do short cases elicit different thinking
processes than factual knowledge ques-
tions do? Med Educ 2001;35:348-56.
Case S, Swanson D. Constructing writ-
ten test questions for the basic and clinical
sciences. 3rd ed. Philadelphia: National
Board of Medical Examiners, 2000.
Farmer EA, Page G. A practical guide
to assessing clinical decision-making skills
using the key features approach. Med
Educ 2005;39:1188-94.
Charlin B, Roy L, Brailovsky C, Goulet
F, van der Vleuten C. The Script Concor-
dance test: a tool to assess the reflective
clinician. Teach Learn Med 2000;12:189-
95.
Brailovsky C, Charlin B, Beausoleil S,
Cote S, Van der Vleuten C. Measurement
15.
16.
17.
18.
19.
20.
21.
22.
23.
24.
25.
26.
27.
28.
of clinical reflective capacity early in train-
ing as a predictor of clinical reasoning
performance at the end of residency: an
experimental study on the script concor-
dance test. Med Educ 2001;35:430-6.
Frederiksen N. The real test bias: influ-
ences of testing on teaching and learning.
Am Psychol 1984;39:193-202.
Schuwirth LW, van der Vleuten CP,
Donkers HH. A closer look at cueing ef-
fects in multiple-choice questions. Med
Educ 1996;30:44-9.
Friedman MH, Connell KJ, Olthoff AJ,
Sinacore JM, Bordage G. Medical student
errors in making a diagnosis. Acad Med
1998;73:Suppl:S19-S21.
Graber ML, Franklin N, Gordon R.
Diagnostic error in internal medicine.
Arch Intern Med 2005;165:1493-9.
Pulito AR, Donnelly MB, Plymale M,
Mentzer RM Jr. What do faculty observe
of medical students’ clinical performance?
Teach Learn Med 2006;18:99-104.
Norman G. The long case versus ob-
jective structured clinical examinations.
BMJ 2002;324:748-9.
Norcini JJ, Blank LL, Duffy FD, Fortna
GS. The mini-CEX: a method for assess-
ing clinical skills. Ann Intern Med 2003;
138:476-81.
Van der Vleuten CP, Norman GR, De
Graaff E. Pitfalls in the pursuit of objec-
tivity: issues of reliability. Med Educ 1991;
25:110-8.
Anastakis DJ, Cohen R, Reznick RK.
The structured oral examination as a
method for assessing surgical residents.
Am J Surg 1991;162:67-70.
Ram P, Grol R, Rethans JJ, Schouten
B, Van der Vleuten C, Kester A. Assess-
ment of general practitioners by video ob-
servation of communicative and medical
performance in daily practice: issues of
validity, reliability and feasibility. Med
Educ 1999;33:447-54.
Papadakis MA. The Step 2 clinical-
skills examination. N Engl J Med 2004;350:
1703-5.
Hodges B, McIlroy JH. Analytic global
OSCE ratings are sensitive to level of
training. Med Educ 2003;37:1012-6.
29.
30.
31.
32.
33.
34.
35.
36.
37.
38.
39.
40.
A video showing an exami-
nation involving a standard-
ized patient is available with
the full text of this article at
www.nejm.org.
Copyright © 2007 Massachusetts Medical Society. All rights reserved.
Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .
n engl j med 356;4 www.nejm.org january 25, 2007
396
Medical Education
Reznick RK, Blackmore D, Cohen R,
et al. An objective structured clinical ex-
amination for the licentiate of the Medi-
cal Council of Canada: from research to
reality. Acad Med 1993;68:Suppl:S4-S6.
Wass V, Jones R, Van der Vleuten C.
Standardized or real patients to test clini-
cal competence? The long case revisited.
Med Educ 2001;35:321-5.
Carney PA, Eliassen MS, Wolford GL,
Owen M, Badger LW, Dietrich AJ. How
physician communication influences rec-
ognition of depression in primary care.
J Fam Pract 1999;48:958-64.
Epstein RM, Franks P, Shields CG, et
al. Patient-centered communication and
diagnostic testing. Ann Fam Med 2005;3:
415-21.
Tamblyn RM. Use of standardized pa-
tients in the assessment of medical prac-
tice. CMAJ 1998;158:205-7.
Kravitz RL, Epstein RM, Feldman MD,
et al. Influence of patients’ requests for
direct-to-consumer advertised antidepres-
sants: a randomized controlled trial. JAMA
2005;293:1995-2002. [Erratum, JAMA 2005;
294:2436.]
Reznick RK, MacRae H. Teaching sur-
gical skills — changes in the wind. N Engl
J Med 2006;355:2664-9.
Ramsey PG, Wenrich MD, Carline JD,
Inui TS, Larson EB, LoGerfo JP. Use of
peer ratings to evaluate physician perfor-
mance. JAMA 1993;269:1655-60.
Dannefer EF, Henson LC, Bierer SB, et
al. Peer assessment of professional com-
petence. Med Educ 2005;39:713-22.
Violato C, Marini A, Toews J, Lockyer
J, Fidler H. Feasibility and psychometric
properties of using peers, consulting phy-
sicians, co-workers, and patients to assess
physicians. Acad Med 1997;72:Suppl 1:S82-
S84.
Norcini JJ. Peer assessment of compe-
tence. Med Educ 2003;37:539-43.
Nofziger AC, Davis B, Naumburg EH,
Epstein RM. The impact of peer assess-
ment on professional development. Pre-
sented at the Ottawa Conference on Medi-
cal Education and Assessment, Ottawa,
July 15, 2002. abstract.
Lurie SJ, Nofziger AC, Meldrum S,
41.
42.
43.
44.
45.
46.
47.
48.
49.
50.
51.
52.
53.
Mooney C, Epstein RM. Temporal and
group-related trends in peer assessment
amongst medical students. Med Educ 2006;
40:840-7.
Lurie S, Lambert D, Grady-Weliky TA.
Relationship between dean’s letter rank-
ings, peer assessment during medical
school, and ratings by internship directors.
[Submitted for Publication] (in press).
Calhoun JG, Woolliscroft JO, Hockman
EM, Wolf FM, Davis WK. Evaluating med-
ical student clinical skill performance:
relationships among self, peer, and expert
ratings. Proc Annu Conf Res Med Educ
1984;23:205-10.
Hall JA, Milburn MA, Roter DL, Dal-
troy LH. Why are sicker patients less satis-
fied with their medical care? Tests of two
explanatory models. Health Psychol 1998;
17:70-5.
Chang JT, Hays RD, Shekelle PG, et al.
Patients’ global ratings of their health
care are not associated with the technical
quality of their care. Ann Intern Med 2006;
144:665-72.
Butterfield PS, Mazzaferri EL. A new
rating form for use by nurses in assessing
residents’ humanistic behavior. J Gen In-
tern Med 1991;6:155-61.
Klessig J, Robbins AS, Wieland D, Ru-
benstein L. Evaluating humanistic attri-
butes of internal medicine residents. J Gen
Intern Med 1989;4:514-21.
Eva KW, Regehr G. Self-assessment in
the health professions: a reformulation
and research agenda. Acad Med 2005;80:
Suppl:S46-S54.
Carraccio C, Englander R. Evaluating
competence using a portfolio: a literature
review and Web-based application to the
ACGME competencies. Teach Learn Med
2004;16:381-7.
Committee on Quality of Health Care
in America. Crossing the quality chasm:
a new health system for the 21st century.
Washington, DC: National Academy Press,
2001.
Ginsburg S, Regehr G, Lingard L.
Basing the evaluation of professionalism
on observable behaviors: a cautionary tale.
Acad Med 2004;79:Suppl:S1-S4.
Schirmer JM, Mauksch L, Lang F, et al.
54.
55.
56.
57.
58.
59.
60.
61.
62.
63.
64.
Assessing communication competence:
a review of current tools. Fam Med 2005;
37:184-92.
Epstein RM, Franks P, Fiscella K, et
al. Measuring patient-centered communi-
cation in patient-physician consultations:
theoretical and practical issues. Soc Sci
Med 2005;61:1516-28.
Norman GR, Van der Vleuten CP, De
Graaff E. Pitfalls in the pursuit of objec-
tivity: issues of validity, efficiency and ac-
ceptability. Med Educ 1991;25:119-26.
Colliver JA, Vu NV, Barrows HS.
Screening test length for sequential test-
ing with a standardized-patient examina-
tion: a receiver operating characteristic
(ROC) analysis. Acad Med 1992;67:592-5.
Hakstian RA. The effects of type of
examination anticipated on test prepara-
tion and performance. J Educ Res 1971;64:
319-24.
Newble DI, Jaeger K. The effect of as-
sessments and examinations on the learn-
ing of medical students. Med Educ 1983;
17:165-71.
Schmidt HG, Norman GR, Boshuizen
HP. A cognitive perspective on medical
expertise: theory and implication. Acad
Med 1990;65:611-21. [Erratum, Acad Med
1992;67:287.]
Bowen JL. Educational strategies to
promote clinical diagnostic reasoning.
N Engl J Med 2006;355:2217-25.
Tamblyn R, Abrahamowicz M, Dau-
phinee WD, et al. Association between li-
censure examination scores and practice
in primary care. JAMA 2002;288:3019-26.
Eva KW, Reiter HI, Rosenfeld J, Nor-
man GR. The ability of the multiple mini-
interview to predict preclerkship perfor-
mance in medical school. Acad Med 2004;
79:Suppl:S40-S42.
Lypson ML, Frohna JG, Gruppen LD,
Woolliscroft JO. Assessing residents’ com-
petencies at baseline: identifying the gaps.
Acad Med 2004;79:564-70.
Batmangelich S, Adamowski S. Main-
tenance of certification in the United
States: a progress report. J Contin Educ
Health Prof 2004;24:134-8.
Copyright © 2007 Massachusetts Medical Society.
65.
66.
67.
68.
69.
70.
71.
72.
73.
74.
75.
EARLY JOB ALERT SERVICE AVAILABLE AT THE NEJM CAREERCENTER
Register to receive weekly e-mail messages with the latest job openings
that match your specialty, as well as preferred geographic region,
practice setting, call schedule, and more. Visit the NEJM CareerCenter
at www.nejmjobs.org for more information.
Copyright © 2007 Massachusetts Medical Society. All rights reserved.
Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .

More Related Content

Similar to Assessment in Medical Education.pdf

Assessment of-clinical-competence-wass-van-der-vleuten-shatzer-jones
Assessment of-clinical-competence-wass-van-der-vleuten-shatzer-jonesAssessment of-clinical-competence-wass-van-der-vleuten-shatzer-jones
Assessment of-clinical-competence-wass-van-der-vleuten-shatzer-jonesPROIDDBahiana
 
Essential Skills: Critical Thinking For College Students
Essential Skills: Critical Thinking For College StudentsEssential Skills: Critical Thinking For College Students
Essential Skills: Critical Thinking For College Students
noblex1
 
Teaching pathology what students learn
Teaching pathology what students learnTeaching pathology what students learn
Teaching pathology what students learn
Hidee Cyd
 
A Format For Case Conceptualization
A Format For Case ConceptualizationA Format For Case Conceptualization
A Format For Case Conceptualization
Steven Wallach
 
Beyond Competencies and Milestones: Adding Meaning Through Context
Beyond Competencies and Milestones: Adding Meaning Through ContextBeyond Competencies and Milestones: Adding Meaning Through Context
Beyond Competencies and Milestones: Adding Meaning Through Context
Jibran Mohsin
 
Mindful practice. ronald epstein[1]
Mindful practice. ronald epstein[1]Mindful practice. ronald epstein[1]
Mindful practice. ronald epstein[1]
Institut Català de la Salut
 
Client patient relationship
Client patient relationshipClient patient relationship
Client patient relationship
HadeyaQ
 
Reflective Journal Week 5Topic Philosophies and Theories for Ad.docx
Reflective Journal Week 5Topic Philosophies and Theories for Ad.docxReflective Journal Week 5Topic Philosophies and Theories for Ad.docx
Reflective Journal Week 5Topic Philosophies and Theories for Ad.docx
sodhi3
 
Evidence based practice
Evidence based practiceEvidence based practice
Evidence based practiceWayan Ardhana
 
NURS 6531 Practice Care of Adults Practicum Experience Journal.docx
NURS 6531 Practice Care of Adults Practicum Experience Journal.docxNURS 6531 Practice Care of Adults Practicum Experience Journal.docx
NURS 6531 Practice Care of Adults Practicum Experience Journal.docx
stirlingvwriters
 
Can Revalidation Deliver What the Public Expects?
Can Revalidation Deliver What the Public Expects?Can Revalidation Deliver What the Public Expects?
Can Revalidation Deliver What the Public Expects?
IAMRAreval2015
 
Bridging gap between theory & practice in pharmacology
Bridging gap between theory & practice in pharmacologyBridging gap between theory & practice in pharmacology
Bridging gap between theory & practice in pharmacology
gundu333pappu
 
Bridging gap between theory & practice in pharmacology
Bridging gap between theory & practice in pharmacologyBridging gap between theory & practice in pharmacology
Bridging gap between theory & practice in pharmacology
Dr Debasish Pradhan
 
Group 1.pdf
Group 1.pdfGroup 1.pdf
Group 1.pdf
KeirelEdrin
 
Final study guide tahun 3 tahun 2013,2014 (1)
Final study guide tahun 3 tahun 2013,2014 (1)Final study guide tahun 3 tahun 2013,2014 (1)
Final study guide tahun 3 tahun 2013,2014 (1)
Christopher C.K. Ho
 
Each response should be between 25 – 50 words and in some cases, i.docx
Each response should be between 25 – 50 words and in some cases, i.docxEach response should be between 25 – 50 words and in some cases, i.docx
Each response should be between 25 – 50 words and in some cases, i.docx
joellemurphey
 
Advaned_ROLE_of_NP_Management_ON_Common_Acute_Pediatric_Problems.pdf
Advaned_ROLE_of_NP_Management_ON_Common_Acute_Pediatric_Problems.pdfAdvaned_ROLE_of_NP_Management_ON_Common_Acute_Pediatric_Problems.pdf
Advaned_ROLE_of_NP_Management_ON_Common_Acute_Pediatric_Problems.pdf
johnadamu3
 
New microsoft office power point presentation
New microsoft office power point presentationNew microsoft office power point presentation
New microsoft office power point presentation
Emani Aparna
 
The Ethics Of Nursing Practice
The Ethics Of Nursing PracticeThe Ethics Of Nursing Practice
The Ethics Of Nursing Practice
Kendra Cote
 
A win-win deal: Critical thinking in Health Sciences education
A win-win deal: Critical thinking in Health Sciences educationA win-win deal: Critical thinking in Health Sciences education
A win-win deal: Critical thinking in Health Sciences education
CRITHINKEDU - Critical Thinking Across the European Higher Education Curricula
 

Similar to Assessment in Medical Education.pdf (20)

Assessment of-clinical-competence-wass-van-der-vleuten-shatzer-jones
Assessment of-clinical-competence-wass-van-der-vleuten-shatzer-jonesAssessment of-clinical-competence-wass-van-der-vleuten-shatzer-jones
Assessment of-clinical-competence-wass-van-der-vleuten-shatzer-jones
 
Essential Skills: Critical Thinking For College Students
Essential Skills: Critical Thinking For College StudentsEssential Skills: Critical Thinking For College Students
Essential Skills: Critical Thinking For College Students
 
Teaching pathology what students learn
Teaching pathology what students learnTeaching pathology what students learn
Teaching pathology what students learn
 
A Format For Case Conceptualization
A Format For Case ConceptualizationA Format For Case Conceptualization
A Format For Case Conceptualization
 
Beyond Competencies and Milestones: Adding Meaning Through Context
Beyond Competencies and Milestones: Adding Meaning Through ContextBeyond Competencies and Milestones: Adding Meaning Through Context
Beyond Competencies and Milestones: Adding Meaning Through Context
 
Mindful practice. ronald epstein[1]
Mindful practice. ronald epstein[1]Mindful practice. ronald epstein[1]
Mindful practice. ronald epstein[1]
 
Client patient relationship
Client patient relationshipClient patient relationship
Client patient relationship
 
Reflective Journal Week 5Topic Philosophies and Theories for Ad.docx
Reflective Journal Week 5Topic Philosophies and Theories for Ad.docxReflective Journal Week 5Topic Philosophies and Theories for Ad.docx
Reflective Journal Week 5Topic Philosophies and Theories for Ad.docx
 
Evidence based practice
Evidence based practiceEvidence based practice
Evidence based practice
 
NURS 6531 Practice Care of Adults Practicum Experience Journal.docx
NURS 6531 Practice Care of Adults Practicum Experience Journal.docxNURS 6531 Practice Care of Adults Practicum Experience Journal.docx
NURS 6531 Practice Care of Adults Practicum Experience Journal.docx
 
Can Revalidation Deliver What the Public Expects?
Can Revalidation Deliver What the Public Expects?Can Revalidation Deliver What the Public Expects?
Can Revalidation Deliver What the Public Expects?
 
Bridging gap between theory & practice in pharmacology
Bridging gap between theory & practice in pharmacologyBridging gap between theory & practice in pharmacology
Bridging gap between theory & practice in pharmacology
 
Bridging gap between theory & practice in pharmacology
Bridging gap between theory & practice in pharmacologyBridging gap between theory & practice in pharmacology
Bridging gap between theory & practice in pharmacology
 
Group 1.pdf
Group 1.pdfGroup 1.pdf
Group 1.pdf
 
Final study guide tahun 3 tahun 2013,2014 (1)
Final study guide tahun 3 tahun 2013,2014 (1)Final study guide tahun 3 tahun 2013,2014 (1)
Final study guide tahun 3 tahun 2013,2014 (1)
 
Each response should be between 25 – 50 words and in some cases, i.docx
Each response should be between 25 – 50 words and in some cases, i.docxEach response should be between 25 – 50 words and in some cases, i.docx
Each response should be between 25 – 50 words and in some cases, i.docx
 
Advaned_ROLE_of_NP_Management_ON_Common_Acute_Pediatric_Problems.pdf
Advaned_ROLE_of_NP_Management_ON_Common_Acute_Pediatric_Problems.pdfAdvaned_ROLE_of_NP_Management_ON_Common_Acute_Pediatric_Problems.pdf
Advaned_ROLE_of_NP_Management_ON_Common_Acute_Pediatric_Problems.pdf
 
New microsoft office power point presentation
New microsoft office power point presentationNew microsoft office power point presentation
New microsoft office power point presentation
 
The Ethics Of Nursing Practice
The Ethics Of Nursing PracticeThe Ethics Of Nursing Practice
The Ethics Of Nursing Practice
 
A win-win deal: Critical thinking in Health Sciences education
A win-win deal: Critical thinking in Health Sciences educationA win-win deal: Critical thinking in Health Sciences education
A win-win deal: Critical thinking in Health Sciences education
 

More from Rachel Doty

What Is Important In Academic Writing Key Ess
What Is Important In Academic Writing Key EssWhat Is Important In Academic Writing Key Ess
What Is Important In Academic Writing Key Ess
Rachel Doty
 
Top 20 College Of Charleston, Charleston House Rentals Vrbo
Top 20 College Of Charleston, Charleston House Rentals VrboTop 20 College Of Charleston, Charleston House Rentals Vrbo
Top 20 College Of Charleston, Charleston House Rentals Vrbo
Rachel Doty
 
Nursing Essay Sample On Pantone Canvas Gallery
Nursing Essay Sample On Pantone Canvas GalleryNursing Essay Sample On Pantone Canvas Gallery
Nursing Essay Sample On Pantone Canvas Gallery
Rachel Doty
 
Academic Essay Format Example
Academic Essay Format ExampleAcademic Essay Format Example
Academic Essay Format Example
Rachel Doty
 
Custom Writing Essay Service Best Research P
Custom Writing Essay Service Best Research PCustom Writing Essay Service Best Research P
Custom Writing Essay Service Best Research P
Rachel Doty
 
Cheap Essay Writers Essay Writing Service Ess
Cheap Essay Writers Essay Writing Service EssCheap Essay Writers Essay Writing Service Ess
Cheap Essay Writers Essay Writing Service Ess
Rachel Doty
 
PPYY NEW 48PCS Writing Stationery Paper , Letter
PPYY NEW 48PCS Writing Stationery Paper , LetterPPYY NEW 48PCS Writing Stationery Paper , Letter
PPYY NEW 48PCS Writing Stationery Paper , Letter
Rachel Doty
 
Reddit An Analysis Of The Reason For Participating In
Reddit An Analysis Of The Reason For Participating InReddit An Analysis Of The Reason For Participating In
Reddit An Analysis Of The Reason For Participating In
Rachel Doty
 
Writing Paper - Notebook Paper Latest Price, Manufacturers
Writing Paper - Notebook Paper Latest Price, ManufacturersWriting Paper - Notebook Paper Latest Price, Manufacturers
Writing Paper - Notebook Paper Latest Price, Manufacturers
Rachel Doty
 
How To Write My Essay - Iran Bi
How To Write My Essay - Iran BiHow To Write My Essay - Iran Bi
How To Write My Essay - Iran Bi
Rachel Doty
 
IELTS Writing. Task 2 - Online Presentation Ielts Writing, Online
IELTS Writing. Task 2 - Online Presentation Ielts Writing, OnlineIELTS Writing. Task 2 - Online Presentation Ielts Writing, Online
IELTS Writing. Task 2 - Online Presentation Ielts Writing, Online
Rachel Doty
 
How To Write A Synthesis Essay Telegraph
How To Write A Synthesis Essay TelegraphHow To Write A Synthesis Essay Telegraph
How To Write A Synthesis Essay Telegraph
Rachel Doty
 
April Blog I Hate Writing David Keane
April Blog I Hate Writing David KeaneApril Blog I Hate Writing David Keane
April Blog I Hate Writing David Keane
Rachel Doty
 
Read 6 Transfer Essays That Worked Blog Assignment
Read 6 Transfer Essays That Worked Blog AssignmentRead 6 Transfer Essays That Worked Blog Assignment
Read 6 Transfer Essays That Worked Blog Assignment
Rachel Doty
 
Narrative Essay Scholarships No Essay 2021
Narrative Essay Scholarships No Essay 2021Narrative Essay Scholarships No Essay 2021
Narrative Essay Scholarships No Essay 2021
Rachel Doty
 
How To Write Essay First Paragraph - Adermann Script
How To Write Essay First Paragraph - Adermann ScriptHow To Write Essay First Paragraph - Adermann Script
How To Write Essay First Paragraph - Adermann Script
Rachel Doty
 
011 Argumentative Essay On School Uniforms P1 That
011 Argumentative Essay On School Uniforms P1 That011 Argumentative Essay On School Uniforms P1 That
011 Argumentative Essay On School Uniforms P1 That
Rachel Doty
 
Analytical Essay - 6 Examples, Format, Pdf Exa
Analytical Essay - 6 Examples, Format, Pdf ExaAnalytical Essay - 6 Examples, Format, Pdf Exa
Analytical Essay - 6 Examples, Format, Pdf Exa
Rachel Doty
 
Top Best Australian Essay Writing Service Tips - Aress
Top Best Australian Essay Writing Service Tips - AressTop Best Australian Essay Writing Service Tips - Aress
Top Best Australian Essay Writing Service Tips - Aress
Rachel Doty
 
Synthesis EssayDefintion, Writing Tips Examples
Synthesis EssayDefintion, Writing Tips ExamplesSynthesis EssayDefintion, Writing Tips Examples
Synthesis EssayDefintion, Writing Tips Examples
Rachel Doty
 

More from Rachel Doty (20)

What Is Important In Academic Writing Key Ess
What Is Important In Academic Writing Key EssWhat Is Important In Academic Writing Key Ess
What Is Important In Academic Writing Key Ess
 
Top 20 College Of Charleston, Charleston House Rentals Vrbo
Top 20 College Of Charleston, Charleston House Rentals VrboTop 20 College Of Charleston, Charleston House Rentals Vrbo
Top 20 College Of Charleston, Charleston House Rentals Vrbo
 
Nursing Essay Sample On Pantone Canvas Gallery
Nursing Essay Sample On Pantone Canvas GalleryNursing Essay Sample On Pantone Canvas Gallery
Nursing Essay Sample On Pantone Canvas Gallery
 
Academic Essay Format Example
Academic Essay Format ExampleAcademic Essay Format Example
Academic Essay Format Example
 
Custom Writing Essay Service Best Research P
Custom Writing Essay Service Best Research PCustom Writing Essay Service Best Research P
Custom Writing Essay Service Best Research P
 
Cheap Essay Writers Essay Writing Service Ess
Cheap Essay Writers Essay Writing Service EssCheap Essay Writers Essay Writing Service Ess
Cheap Essay Writers Essay Writing Service Ess
 
PPYY NEW 48PCS Writing Stationery Paper , Letter
PPYY NEW 48PCS Writing Stationery Paper , LetterPPYY NEW 48PCS Writing Stationery Paper , Letter
PPYY NEW 48PCS Writing Stationery Paper , Letter
 
Reddit An Analysis Of The Reason For Participating In
Reddit An Analysis Of The Reason For Participating InReddit An Analysis Of The Reason For Participating In
Reddit An Analysis Of The Reason For Participating In
 
Writing Paper - Notebook Paper Latest Price, Manufacturers
Writing Paper - Notebook Paper Latest Price, ManufacturersWriting Paper - Notebook Paper Latest Price, Manufacturers
Writing Paper - Notebook Paper Latest Price, Manufacturers
 
How To Write My Essay - Iran Bi
How To Write My Essay - Iran BiHow To Write My Essay - Iran Bi
How To Write My Essay - Iran Bi
 
IELTS Writing. Task 2 - Online Presentation Ielts Writing, Online
IELTS Writing. Task 2 - Online Presentation Ielts Writing, OnlineIELTS Writing. Task 2 - Online Presentation Ielts Writing, Online
IELTS Writing. Task 2 - Online Presentation Ielts Writing, Online
 
How To Write A Synthesis Essay Telegraph
How To Write A Synthesis Essay TelegraphHow To Write A Synthesis Essay Telegraph
How To Write A Synthesis Essay Telegraph
 
April Blog I Hate Writing David Keane
April Blog I Hate Writing David KeaneApril Blog I Hate Writing David Keane
April Blog I Hate Writing David Keane
 
Read 6 Transfer Essays That Worked Blog Assignment
Read 6 Transfer Essays That Worked Blog AssignmentRead 6 Transfer Essays That Worked Blog Assignment
Read 6 Transfer Essays That Worked Blog Assignment
 
Narrative Essay Scholarships No Essay 2021
Narrative Essay Scholarships No Essay 2021Narrative Essay Scholarships No Essay 2021
Narrative Essay Scholarships No Essay 2021
 
How To Write Essay First Paragraph - Adermann Script
How To Write Essay First Paragraph - Adermann ScriptHow To Write Essay First Paragraph - Adermann Script
How To Write Essay First Paragraph - Adermann Script
 
011 Argumentative Essay On School Uniforms P1 That
011 Argumentative Essay On School Uniforms P1 That011 Argumentative Essay On School Uniforms P1 That
011 Argumentative Essay On School Uniforms P1 That
 
Analytical Essay - 6 Examples, Format, Pdf Exa
Analytical Essay - 6 Examples, Format, Pdf ExaAnalytical Essay - 6 Examples, Format, Pdf Exa
Analytical Essay - 6 Examples, Format, Pdf Exa
 
Top Best Australian Essay Writing Service Tips - Aress
Top Best Australian Essay Writing Service Tips - AressTop Best Australian Essay Writing Service Tips - Aress
Top Best Australian Essay Writing Service Tips - Aress
 
Synthesis EssayDefintion, Writing Tips Examples
Synthesis EssayDefintion, Writing Tips ExamplesSynthesis EssayDefintion, Writing Tips Examples
Synthesis EssayDefintion, Writing Tips Examples
 

Recently uploaded

2024.06.01 Introducing a competency framework for languag learning materials ...
2024.06.01 Introducing a competency framework for languag learning materials ...2024.06.01 Introducing a competency framework for languag learning materials ...
2024.06.01 Introducing a competency framework for languag learning materials ...
Sandy Millin
 
Digital Tools and AI for Teaching Learning and Research
Digital Tools and AI for Teaching Learning and ResearchDigital Tools and AI for Teaching Learning and Research
Digital Tools and AI for Teaching Learning and Research
Vikramjit Singh
 
S1-Introduction-Biopesticides in ICM.pptx
S1-Introduction-Biopesticides in ICM.pptxS1-Introduction-Biopesticides in ICM.pptx
S1-Introduction-Biopesticides in ICM.pptx
tarandeep35
 
Acetabularia Information For Class 9 .docx
Acetabularia Information For Class 9  .docxAcetabularia Information For Class 9  .docx
Acetabularia Information For Class 9 .docx
vaibhavrinwa19
 
Synthetic Fiber Construction in lab .pptx
Synthetic Fiber Construction in lab .pptxSynthetic Fiber Construction in lab .pptx
Synthetic Fiber Construction in lab .pptx
Pavel ( NSTU)
 
Welcome to TechSoup New Member Orientation and Q&A (May 2024).pdf
Welcome to TechSoup   New Member Orientation and Q&A (May 2024).pdfWelcome to TechSoup   New Member Orientation and Q&A (May 2024).pdf
Welcome to TechSoup New Member Orientation and Q&A (May 2024).pdf
TechSoup
 
The French Revolution Class 9 Study Material pdf free download
The French Revolution Class 9 Study Material pdf free downloadThe French Revolution Class 9 Study Material pdf free download
The French Revolution Class 9 Study Material pdf free download
Vivekanand Anglo Vedic Academy
 
Group Presentation 2 Economics.Ariana Buscigliopptx
Group Presentation 2 Economics.Ariana BuscigliopptxGroup Presentation 2 Economics.Ariana Buscigliopptx
Group Presentation 2 Economics.Ariana Buscigliopptx
ArianaBusciglio
 
Model Attribute Check Company Auto Property
Model Attribute  Check Company Auto PropertyModel Attribute  Check Company Auto Property
Model Attribute Check Company Auto Property
Celine George
 
June 3, 2024 Anti-Semitism Letter Sent to MIT President Kornbluth and MIT Cor...
June 3, 2024 Anti-Semitism Letter Sent to MIT President Kornbluth and MIT Cor...June 3, 2024 Anti-Semitism Letter Sent to MIT President Kornbluth and MIT Cor...
June 3, 2024 Anti-Semitism Letter Sent to MIT President Kornbluth and MIT Cor...
Levi Shapiro
 
special B.ed 2nd year old paper_20240531.pdf
special B.ed 2nd year old paper_20240531.pdfspecial B.ed 2nd year old paper_20240531.pdf
special B.ed 2nd year old paper_20240531.pdf
Special education needs
 
Normal Labour/ Stages of Labour/ Mechanism of Labour
Normal Labour/ Stages of Labour/ Mechanism of LabourNormal Labour/ Stages of Labour/ Mechanism of Labour
Normal Labour/ Stages of Labour/ Mechanism of Labour
Wasim Ak
 
Mule 4.6 & Java 17 Upgrade | MuleSoft Mysore Meetup #46
Mule 4.6 & Java 17 Upgrade | MuleSoft Mysore Meetup #46Mule 4.6 & Java 17 Upgrade | MuleSoft Mysore Meetup #46
Mule 4.6 & Java 17 Upgrade | MuleSoft Mysore Meetup #46
MysoreMuleSoftMeetup
 
Supporting (UKRI) OA monographs at Salford.pptx
Supporting (UKRI) OA monographs at Salford.pptxSupporting (UKRI) OA monographs at Salford.pptx
Supporting (UKRI) OA monographs at Salford.pptx
Jisc
 
CACJapan - GROUP Presentation 1- Wk 4.pdf
CACJapan - GROUP Presentation 1- Wk 4.pdfCACJapan - GROUP Presentation 1- Wk 4.pdf
CACJapan - GROUP Presentation 1- Wk 4.pdf
camakaiclarkmusic
 
Guidance_and_Counselling.pdf B.Ed. 4th Semester
Guidance_and_Counselling.pdf B.Ed. 4th SemesterGuidance_and_Counselling.pdf B.Ed. 4th Semester
Guidance_and_Counselling.pdf B.Ed. 4th Semester
Atul Kumar Singh
 
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
siemaillard
 
1.4 modern child centered education - mahatma gandhi-2.pptx
1.4 modern child centered education - mahatma gandhi-2.pptx1.4 modern child centered education - mahatma gandhi-2.pptx
1.4 modern child centered education - mahatma gandhi-2.pptx
JosvitaDsouza2
 
A Strategic Approach: GenAI in Education
A Strategic Approach: GenAI in EducationA Strategic Approach: GenAI in Education
A Strategic Approach: GenAI in Education
Peter Windle
 
BÀI TẬP BỔ TRỢ TIẾNG ANH GLOBAL SUCCESS LỚP 3 - CẢ NĂM (CÓ FILE NGHE VÀ ĐÁP Á...
BÀI TẬP BỔ TRỢ TIẾNG ANH GLOBAL SUCCESS LỚP 3 - CẢ NĂM (CÓ FILE NGHE VÀ ĐÁP Á...BÀI TẬP BỔ TRỢ TIẾNG ANH GLOBAL SUCCESS LỚP 3 - CẢ NĂM (CÓ FILE NGHE VÀ ĐÁP Á...
BÀI TẬP BỔ TRỢ TIẾNG ANH GLOBAL SUCCESS LỚP 3 - CẢ NĂM (CÓ FILE NGHE VÀ ĐÁP Á...
Nguyen Thanh Tu Collection
 

Recently uploaded (20)

2024.06.01 Introducing a competency framework for languag learning materials ...
2024.06.01 Introducing a competency framework for languag learning materials ...2024.06.01 Introducing a competency framework for languag learning materials ...
2024.06.01 Introducing a competency framework for languag learning materials ...
 
Digital Tools and AI for Teaching Learning and Research
Digital Tools and AI for Teaching Learning and ResearchDigital Tools and AI for Teaching Learning and Research
Digital Tools and AI for Teaching Learning and Research
 
S1-Introduction-Biopesticides in ICM.pptx
S1-Introduction-Biopesticides in ICM.pptxS1-Introduction-Biopesticides in ICM.pptx
S1-Introduction-Biopesticides in ICM.pptx
 
Acetabularia Information For Class 9 .docx
Acetabularia Information For Class 9  .docxAcetabularia Information For Class 9  .docx
Acetabularia Information For Class 9 .docx
 
Synthetic Fiber Construction in lab .pptx
Synthetic Fiber Construction in lab .pptxSynthetic Fiber Construction in lab .pptx
Synthetic Fiber Construction in lab .pptx
 
Welcome to TechSoup New Member Orientation and Q&A (May 2024).pdf
Welcome to TechSoup   New Member Orientation and Q&A (May 2024).pdfWelcome to TechSoup   New Member Orientation and Q&A (May 2024).pdf
Welcome to TechSoup New Member Orientation and Q&A (May 2024).pdf
 
The French Revolution Class 9 Study Material pdf free download
The French Revolution Class 9 Study Material pdf free downloadThe French Revolution Class 9 Study Material pdf free download
The French Revolution Class 9 Study Material pdf free download
 
Group Presentation 2 Economics.Ariana Buscigliopptx
Group Presentation 2 Economics.Ariana BuscigliopptxGroup Presentation 2 Economics.Ariana Buscigliopptx
Group Presentation 2 Economics.Ariana Buscigliopptx
 
Model Attribute Check Company Auto Property
Model Attribute  Check Company Auto PropertyModel Attribute  Check Company Auto Property
Model Attribute Check Company Auto Property
 
June 3, 2024 Anti-Semitism Letter Sent to MIT President Kornbluth and MIT Cor...
June 3, 2024 Anti-Semitism Letter Sent to MIT President Kornbluth and MIT Cor...June 3, 2024 Anti-Semitism Letter Sent to MIT President Kornbluth and MIT Cor...
June 3, 2024 Anti-Semitism Letter Sent to MIT President Kornbluth and MIT Cor...
 
special B.ed 2nd year old paper_20240531.pdf
special B.ed 2nd year old paper_20240531.pdfspecial B.ed 2nd year old paper_20240531.pdf
special B.ed 2nd year old paper_20240531.pdf
 
Normal Labour/ Stages of Labour/ Mechanism of Labour
Normal Labour/ Stages of Labour/ Mechanism of LabourNormal Labour/ Stages of Labour/ Mechanism of Labour
Normal Labour/ Stages of Labour/ Mechanism of Labour
 
Mule 4.6 & Java 17 Upgrade | MuleSoft Mysore Meetup #46
Mule 4.6 & Java 17 Upgrade | MuleSoft Mysore Meetup #46Mule 4.6 & Java 17 Upgrade | MuleSoft Mysore Meetup #46
Mule 4.6 & Java 17 Upgrade | MuleSoft Mysore Meetup #46
 
Supporting (UKRI) OA monographs at Salford.pptx
Supporting (UKRI) OA monographs at Salford.pptxSupporting (UKRI) OA monographs at Salford.pptx
Supporting (UKRI) OA monographs at Salford.pptx
 
CACJapan - GROUP Presentation 1- Wk 4.pdf
CACJapan - GROUP Presentation 1- Wk 4.pdfCACJapan - GROUP Presentation 1- Wk 4.pdf
CACJapan - GROUP Presentation 1- Wk 4.pdf
 
Guidance_and_Counselling.pdf B.Ed. 4th Semester
Guidance_and_Counselling.pdf B.Ed. 4th SemesterGuidance_and_Counselling.pdf B.Ed. 4th Semester
Guidance_and_Counselling.pdf B.Ed. 4th Semester
 
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
 
1.4 modern child centered education - mahatma gandhi-2.pptx
1.4 modern child centered education - mahatma gandhi-2.pptx1.4 modern child centered education - mahatma gandhi-2.pptx
1.4 modern child centered education - mahatma gandhi-2.pptx
 
A Strategic Approach: GenAI in Education
A Strategic Approach: GenAI in EducationA Strategic Approach: GenAI in Education
A Strategic Approach: GenAI in Education
 
BÀI TẬP BỔ TRỢ TIẾNG ANH GLOBAL SUCCESS LỚP 3 - CẢ NĂM (CÓ FILE NGHE VÀ ĐÁP Á...
BÀI TẬP BỔ TRỢ TIẾNG ANH GLOBAL SUCCESS LỚP 3 - CẢ NĂM (CÓ FILE NGHE VÀ ĐÁP Á...BÀI TẬP BỔ TRỢ TIẾNG ANH GLOBAL SUCCESS LỚP 3 - CẢ NĂM (CÓ FILE NGHE VÀ ĐÁP Á...
BÀI TẬP BỔ TRỢ TIẾNG ANH GLOBAL SUCCESS LỚP 3 - CẢ NĂM (CÓ FILE NGHE VÀ ĐÁP Á...
 

Assessment in Medical Education.pdf

  • 1. The new engl and jour nal o f medicine n engl j med 356;4 www.nejm.org january 25, 2007 387 Medical Education Malcolm Cox, M.D., and David M. Irby, Ph.D., Editors Review article Medical Education Malcolm Cox, M.D., and David M. Irby, Ph.D., Editors Assessment in Medical Education Ronald M. Epstein, M.D. FromtheDepartmentsofFamilyMedicine, Psychiatry, and Oncology and the Roch- ester Center to Improve Communication in Health Care, University of Rochester School of Medicine and Dentistry, Roch- ester, NY. Address reprint requests to Dr. Epstein at 1381 South Ave., Rochester, NY 14620, or at ronald_epstein@urmc. rochester.edu. N Engl J Med 2007;356:387-96. Copyright © 2007 Massachusetts Medical Society. As an attending physician working with a student for a week, you receive a form that asks you to evaluate the student’s fund of knowledge, procedural skills, professional- ism, interest in learning, and “systems-based practice.” You wonder which of these attributes you can reliably assess and how the data you provide will be used to further the student’s education. You also wonder whether other tests of knowledge and com- petence that students must undergo before they enter practice are equally problematic. I n one way or another, most practicing physicians are involved in assessing the competence of trainees, peers, and other health professionals. As the example above suggests, however, they may not be as comfortable using educational assessment tools as they are using more clinically focused diagnostic tests. This article provides a conceptual framework for and a brief update on com- monly used and emerging methods of assessment, discusses the strengths and limitations of each method, and identifies several challenges in the assessment of physicians’ professional competence and performance. Competence and Performance Elsewhere, Hundert and I have defined competence in medicine as “the habitual and judicious use of communication, knowledge, technical skills, clinical reasoning, emotions, values, and reflection in daily practice for the benefit of the individuals and communities being served.”1 In the United States, the assessment of medical residents, and increasingly of medical students, is largely based on a model that was developed by the Accreditation Council for Graduate Medical Education (ACGME). This model uses six interrelated domains of competence: medical knowledge, patient care, professionalism, communication and interpersonal skills, practice-based learn- ing and improvement, and systems-based practice.2 Competence is not an achievement but rather a habit of lifelong learning3; as- sessment plays an integral role in helping physicians identify and respond to their own learning needs. Ideally, the assessment of competence (what the student or physician is able to do) should provide insight into actual performance (what he or she does habitually when not observed), as well as the capacity to adapt to change, find and generate new knowledge, and improve overall performance.4 Competence is contextual, reflecting the relationship between a person’s abilities and the tasks he or she is required to perform in a particular situation in the real world.5 Common contextual factors include the practice setting, the local prevalence of disease, the nature of the patient’s presenting symptoms, the patient’s education- al level, and other demographic characteristics of the patient and of the physician. Many aspects of competence, such as history taking and clinical reasoning, are also content-specific and not necessarily generalizable to all situations. A student’s Copyright © 2007 Massachusetts Medical Society. All rights reserved. Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .
  • 2. The new engl and jour nal o f medicine n engl j med 356;4 www.nejm.org january 25, 2007 388 clinical reasoning may appear to be competent in areas in which his or her base of knowledge is well organized and accessible6 but may appear to be much less competent in unfamiliar territory.7 However, some important skills (e.g., the ability to form therapeutic relationships) may be less dependent on content.8 Competence is also developmental. Habits of mind and behavior and practical wisdom are gained through deliberate practice9 and reflection on experience.10-14 Students begin their training at a novice level, using abstract, rule-based formu- las that are removed from actual practice. At high- er levels, students apply these rules differentially to specific situations. During residency, trainees make judgments that reflect a holistic view of a situation and eventually take diagnostic shortcuts based on a deeper understanding of underlying principles. Experts are able to make rapid, context- based judgments in ambiguous real-life situations and have sufficient awareness of their own cogni- tive processes to articulate and explain how they recognize situations in which deliberation is es- sential. Development of competence in different contexts and content areas may proceed at differ- ent rates. Context and developmental level also interact. Although all clinicians may perform at a lower level of competence when they are tired, distracted, or annoyed, the competence of less experienced clinicians may be particularly suscep- tible to the influence of stress.15,16 Goals of Assessment Over the past decade, medical schools, postgrad- uate training programs, and licensing bodies have made new efforts to provide accurate, reliable, and timely assessments of the competence of trainees and practicing physicians.1,2,17 Such assessments have three main goals: to optimize the capabili- ties of all learners and practitioners by providing motivation and direction for future learning, to protect the public by identifying incompetent phy- sicians, and to provide a basis for choosing ap- plicants for advanced training. Assessment can be formative (guiding future learning,providingreassurance,promotingreflec- tion, and shaping values) or summative (making an overall judgment about competence, fitness to practice, or qualification for advancement to high- er levels of responsibility). Formative assessments provide benchmarks to orient the learner who is approaching a relatively unstructured body of knowledge. They can reinforce students’ intrinsic motivation to learn and inspire them to set higher standards for themselves.18 Although summative assessments are intended to provide professional self-regulation and accountability, they may also act as a barrier to further practice or training.19 A distinction should be made between assess- ments that are suitable only for formative use and those that have sufficient psychometric rigor for summative use. This distinction is especially important in selecting a method of evaluating competence for high-stakes assessments (i.e., licensing and certification examinations). Corre- spondingly, summative assessments may not pro- vide sufficient feedback to drive learning.20 How- ever, because students tend to study that which they expect to be tested on, summative assess- ment may influence learning even in the absence of feedback. Assessment Methods All methods of assessment have strengths and intrinsic flaws (Table 1). The use of multiple ob- servations and several different assessment meth- ods over time can partially compensate for flaws in any one method.1,21 Van der Vleuten22 describes five criteria for determining the usefulness of a particular method of assessment: reliability (the degree to which the measurement is accurate and reproducible), validity (whether the assessment measures what it claims to measure), impact on future learning and practice, acceptability to learn- ers and faculty, and costs (to the individual trainee, the institution, and society at large). Written Examinations Written examination questions are typically classi- fied according to whether they are open-ended or multiple choice. In addition, questions can be “context rich” or “context poor.”23 Questions with rich descriptions of the clinical context invite the more complex cognitive processes that are char- acteristic of clinical practice.24 Conversely, context- poor questions can test basic factual knowledge but not its transferability to real clinical problems. Multiple-choice questions are commonly used for assessment because they can provide a large number of examination items that encompass many content areas, can be administered in a rel- atively short period, and can be graded by com- Copyright © 2007 Massachusetts Medical Society. All rights reserved. Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .
  • 3. Medical Education n engl j med 356;4 www.nejm.org january 25, 2007 389 puter. These factors make the administration of the examination to large numbers of trainees straightforward and standardized.25 Formats that ask the student to choose the best answer from a list of possible answers are most commonly used. However, newer formats may better assess pro- cesses of diagnostic reasoning. Key-feature items focus on critical decisions in particular clinical cases.26 Script-concordance items present a situ- ation (e.g., vaginal discharge in a patient), add a piece of information (dysuria), and ask the exam- inee to assess the degree to which this new in- formation increases or decreases the probability of a particular outcome (acute salpingitis due to Chlamydiatrachomatis).27 Because the situations por- trayed are ambiguous, script-concordance items may provide insight into clinical judgment in the real world. Answers to such items have been shown to correlate with the examinee’s level of training and to predict future performance on oral examinations of clinical reasoning.28 Multiple-choice questions that are rich in con- text are difficult to write, and those who write them tend to avoid topics — such as ethical dilem- mas or cultural ambiguities — that cannot be asked about easily.29 Multiple-choice questions may also create situations in which an examinee can answer a question by recognizing the correct option, but could not have answered it in the absence of options.23,30 This effect, called cueing, is especially problematic when diagnostic reason- ing is being assessed, because premature closure — arriving at a decision before the correct diag- nosis has been considered — is a common reason for diagnostic errors in clinical practice.31,32 Ex- tended matching items (several questions, all with the same long list of possible answers), as well as open-ended short-answer questions, can mini- mize cueing.23 Structured essays also preclude cueing. In addition, they involve more complex cognitive processes and allow for more contex- tualized answers than do multiple-choice ques- tions. When clear grading guidelines are in place, structured essays can be psychometrically robust. Assessments by Supervising Clinicians Supervising clinicians’ observations and impres- sions of students over a specific period remain the most common tool used to evaluate performance with patients. Students and residents most com- monly receive global ratings at the end of a rota- tion, with comments from a variety of supervis- ing physicians. Although subjectivity can be a problem in the absence of clearly articulated standards, a more important issue is that direct observation of trainees while they are interacting with patients is too infrequent.33 Direct Observation or Video Review The “long case”34 and the “mini–clinical-evalua- tion exercise” (mini-CEX)35 have been developed so that learners will be directly observed more fre- quently. In these assessments, a supervising phy- sician observes while a trainee performs a focused history taking and physical examination over a period of 10 to 20 minutes. The trainee then pres- ents a diagnosis and a treatment plan, and the faculty member rates the resident and may pro- vide educational feedback. Structured exercises with actual patients under the observation of the supervising physician can have the same level of reliability as structured examinations using stan- dardized patients34,36 yet encompass a wider range of problems, physical findings, and clinical set- tings. Direct observation of trainees in clinical settings can be coupled with exercises that train- ees perform after their encounters with patients, such as oral case presentations, written exercises that assess clinical reasoning, and literature search- es.8,37 In addition, review of videos of encounters with patients offers a powerful means of evaluat- ing and providing feedback on trainees’ skills in clinical interactions.8,38 Clinical Simulations Standardized patients — actors who are trained to portray patients consistently on repeated occa- sions — are often incorporated into objective structured clinical examinations (OSCEs), which consist of a series of timed “stations,” each one focused on a different task. Since 2004, these ex- aminations have been part of the U.S. Medical Licensing Examination that all senior medical stu- dents take.39 The observing faculty member or the standardized patient uses either a checklist of specific behaviors or a global rating form to evaluate the student’s performance.40 The check- list might include items such as “asked if the pa- tient smoked” and “checked ankle reflexes.” The global rating form might ask for a rating of how well the visit was organized and whether the stu- dent was appropriately empathetic. A minimum of 10 stations, which the student usually visits over the course of 3 to 4 hours, is necessary to achieve Copyright © 2007 Massachusetts Medical Society. All rights reserved. Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .
  • 4. T h e n e w e ng l a n d j o u r na l o f m e dic i n e n engl j med 356;4 www.nejm.org january 25, 2007 390 Table 1. Commonly Used Methods of Assessment. Method Domain Type of Use Limitations Strengths Written exercises Multiple-choice questions in either single-best-answer or extended matching format Knowledge, ability to solve problems Summative assessments within courses or clerkships; nation- al in-service, licensing, and certification examinations Difficult to write, especially in cer- tain content areas; can result in cueing; can seem artificial and removed from real situations Can assess many content areas in rela- tively little time, have high reliability, can be graded by computer Key-feature and script-concor- dance questions Clinical reasoning, problem- solving ability, ability to apply knowledge National licensing and certifica- tion examinations Not yet proven to transfer to real- life situations that require clini- cal reasoning Assess clinical problem-solving ability, avoid cueing, can be graded by com- puter Short-answer questions Ability to interpret diagnostic tests, problem-solving ability, clinical reasoning skills Summative and formative as- sessments in courses and clerkships Reliability dependent on training of graders Avoid cueing, assess interpretation and problem-solving ability Structured essays Synthesis of information, in- terpretation of medical literature Preclinical courses, limited use in clerkships Time-consuming to grade, must work to establish interrater reliability, long testing time required to en- compass a variety of domains Avoid cueing, use higher-order cogni- tive processes Assessments by supervising clinicians Global ratings with comments at end of rotation Clinical skills, communication, teamwork, presentation skills, organization, work habits Global summative and some- times formative assess- ments in clinical rotations Often based on second-hand re- ports and case presentations rather than on direct observa- tion, subjective Use of multiple independent raters can overcome some variability due to subjectivity Structured direct observation with checklists for ratings (e.g., mini–clinical-evaluation exercise or video review) Communication skills, clinical skills Limited use in clerkships and residencies, a few board- certification examinations Selective rather than habitual be- haviors observed, relatively time-consuming Feedback provided by credible experts Oral examinations Knowledge, clinical reasoning Limited use in clerkships and com- prehensive medical school assessments, some board- certification examinations Subjective, sex and race bias has been reported, time-consuming, require training of examiners, summative assessments need two or more examiners Feedback provided by credible experts Copyright © 2007 Massachusetts Medical Society. All rights reserved. Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .
  • 5. Medical Education n engl j med 356;4 www.nejm.org january 25, 2007 391 Clinical simulations Standardized patients and ob- jective structured clinical examinations Some clinical skills, inter- personal behavior, communication skills Formative and summative as- sessments in courses, clerk- ships, medical schools, na- tional licensure examinations, board certification in Canada Timing and setting may seem artifi- cial, require suspension of dis- belief, checklists may penalize examinees who use shortcuts, expensive Tailored to educational goals; reliable, consistent case presentation and ratings; can be observed by faculty or standardized patients; realistic Incognito standardized patients Actual practice habits Primarily used in research; some courses, clerkships, and residencies use for for- mative feedback Requires prior consent, logistically challenging, expensive Very realistic, most accurate way of assessing clinician’s behavior High-technology simulations Procedural skills, teamwork, simulated clinical dilem- mas Formative and some summative assessment Timing and setting may seem artifi- cial, require suspension of dis- belief, checklists may penalize examinees who use shortcuts, expensive Tailored to educational goals, can be observed by faculty, often realistic and credible Multisource (“360-degree”) assessments Peer assessments Professional demeanor, work habits, interpersonal behavior, teamwork Formative feedback in courses and comprehensive medical school assessments, forma- tive assessment for board recertification Confidentiality, anonymity, and trainee buy-in essential Ratings encompass habitual behaviors, credible source, correlates with future academic and clinical performance Patient assessments Ability to gain patients’ trust; patient satisfaction, com- munication skills Formative and summative, board recertification, use by insur- ers to determine bonuses Provide global impressions rather than analysis of specific behav- iors, ratings generally high with little variability Credible source of assessment Self-assessments Knowledge, skills, attitudes, beliefs, behaviors Formative Do not accurately describe actual behavior unless training and feedback provided Foster reflection and development of learning plans Portfolios All aspects of competence, especially appropriate for practice-based learning and improvement and systems-based practice Formative and summative uses across curriculum and with- in clerkships and residency programs, used by some U.K. medical schools and specialty boards Learner selects best case material, time-consuming to prepare and review Display projects for review, foster reflec- tion and development of learning plans Copyright © 2007 Massachusetts Medical Society. All rights reserved. Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .
  • 6. The new engl and jour nal o f medicine n engl j med 356;4 www.nejm.org january 25, 2007 392 a reliability of 0.85 to 0.90.41 Under these condi- tions, structured assessments with the use of standardized patients are as reliable as ratings of directly observed encounters with real patients and take about the same amount of time.42 Interactions with standardized patients can be tailored to meet specific educational goals, and the actors who portray the patients can reliably rate students’ performance with respect to history taking and physical examinations. Faculty mem- bers who observe encounters with standardized patients can offer additional insights on trainees’ clinical judgment and the overall coherence of the history taking or physical examination. Unan- nounced standardized patients, who with the ex- aminees’ prior approval present incognito in ac- tual clinical settings, have been used in health services research to evaluate examinees’ diagnos- tic reasoning, treatment decisions, and communi- cation skills.43-46 The use of unannounced stan- dardized patients may prove to be particularly valuable in the assessment of higher-level trainees and physicians in practice. The use of simulation to assess trainees’ clini- cal skills in intensive care and surgical settings is on the rise.47 Simulations involving sophisticat- ed mannequins with heart sounds, respirations, oximeter readings, and pulses that respond to a variety of interventions can be used to assess how individuals or teams manage unstable vital signs. Surgical simulation centers now routinely use high-fidelity computer graphics and hands-on manipulation of surgical instruments to create a multisensory environment. High-technology simulation is seen increasingly as an important learning aid and may prove to be useful in the assessment of knowledge, clinical reasoning, and teamwork. Multisource (“360-Degree”) Assessments Assessments by peers, other members of the clin- ical team, and patients can provide insight into trainees’ work habits, capacity for teamwork, and interpersonal sensitivity.48-50 Although there are few published data on outcomes of multisource feedback in medical settings, several large pro- grams are being developed, including one for all first- and second-year house officers in the United Kingdom and another for all physicians under- going recertification in internal medicine in the United States. Multisource feedback is most effec- tive when it includes narrative comments as well as statistical data, when the sources are recog- nized as credible, when the feedback is framed constructively, and when the entire process is ac- companied by good mentoring and follow-up.51 Recent studies of peer assessments suggest that when trainees receive thoughtful ratings and comments by peers in a timely and confidential manner, along with support from advisers to help them reflect on the reports, they find the process powerful, insightful, and instructive.51,52 Peer as- sessments have been shown to be consistent re- gardless of the way the raters are selected. Such assessments are stable from year to year53 and predict subsequent class rankings as well as sub- sequent ratings by supervisors.54 Peer assessments depend on trust and require scrupulous attention to confidentiality. Otherwise they can be under- mining, destructive, and divisive. Although patients’ ratings of clinical perfor- mance are valuable in principle, they pose several problems. As many as 50 patient surveys may be necessary to achieve satisfactory reliability.55 Pa- tients who are seriously ill often do not complete surveys; those who do tend to rate physicians less favorably than do patients who have milder con- ditions.56 Furthermore, patients are not always able to discriminate among the elements of clin- ical practice,57 and their ratings are typically high. These limitations make it difficult to use patient reports as the only tool for assessing clinical performance. However, ratings by nurses can be valuable. Such ratings have been found to be reliable with as few as 6 to 10 reports,58 and they correlate with both patients’ and faculty members’ ratings of the interpersonal aspects of trainees’ performance.59 Fundamental cognitive limitations in the abil- ity of humans to know themselves as others see them restrict the usefulness of self-assessment. Furthermore, rating oneself on prior clinical per- formance may not achieve another important goal of self-assessment: the ability to monitor oneself from moment to moment during clinical prac- tice.10,60 A physician must possess this ability in order to meet patients’ changing needs, to recog- nize the limits of his or her own competence, and to manage unexpected situations. Portfolios Portfolios include documentation of and reflection about specific areas of a trainee’s competence. This evidence is combined with self-reflection.61 Copyright © 2007 Massachusetts Medical Society. All rights reserved. Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .
  • 7. Medical Education n engl j med 356;4 www.nejm.org january 25, 2007 393 In medicine, just as in the visual arts, portfolios demonstrate a trainee’s development and techni- cal capacity. They can include chart notes, refer- ral letters, procedure logs, videotaped consulta- tions,peerassessments,patientsurveys,literature searches, quality-improvement projects, and any other type of learning material. Portfolios also frequently include self-assessments, learning plans, and reflective essays. For portfolios to be maximally effective, close mentoring is required in the assembly and interpretation of the contents; considerable time can be expended in this effort. Portfolios are most commonly used in formative assessments, but their use for summative evalua- tions and high-stakes decisions about advance- ment is increasing.20 Challenges in Assessment New Domains of Assessment There are several domains in which assessment is in its infancy and remains problematic. Quality of care and patient safety depend on effective teamwork,62 and teamwork training is empha- sized as an essential element of several areas of competence specified by the ACGME, yet there is no validated method of assessing teamwork. Ex- perts do not agree on how to define professional- ism — let alone how best to measure it.63 Dozens of scales that rate communication are used in medical education and research,64 yet there is little evidence that any one scale is better than another; furthermore, the experiences that patients report often differ considerably from ratings given by experts.65 Multimethod and Longitudinal Assessment The use of multiple methods of assessment can overcome many of the limitations of individual assessment formats.8,22,36,66 Variation of the clin- ical context allows for broader insights into com- petence, the use of multiple formats provides greater variety in the areas of content that are evaluated, and input from multiple observers pro- vides information on distinct aspects of a trainee’s performance. Longitudinal assessment avoids ex- cessive testing at any one point in time and serves as the foundation for monitoring ongoing profes- sional development. In the example at the beginning of this article, a multimethod assessment might include direct observation of the student interacting with several patients at different points during the rotation, a multiple-choice examination with both “key features” and “script-concordance” items to as- sess clinical reasoning, an encounter with a stan- dardized patient followed by an oral examination to assess clinical skills in a standardized setting, written essays that would require literature search- es and synthesis of the medical literature on the basic science or clinical aspects of one or more of the diseases the student encountered, and peer assessments to provide insights into interperson- al skills and work habits. The combination of all these results into a portfolio resembles the art of diagnosis; it de- mands that the student synthesize various bits and types of information in order to come up with an overall picture. Although a few medical schools have begun to institute longitudinal as- sessments that use multiple methods,8 the best way to deal with the quantity and the qualitatively different types of data that the process generates is not yet clear. New ways of combining qualita- tive and quantitative data will be required if port- folio assessments are to find widespread applica- tion and withstand the test of time. Standardization of Assessment Although accrediting organizations specify broad areas that the curriculum should cover and assess, for the most part individual medical schools make their own decisions about methods and standards of assessment. This model may have the advan- tage of ensuring consistency between the curricu- lum and assessment, but it also makes it difficult to compare students across medical schools for the purpose of subsequent training.67 The ideal balance between nationally standardized and school-specific assessment remains to be deter- mined.Furthermore,withinagivenmedicalschool, all students may not require the same package of assessments — for example, initial screening ex- aminations may be followed by more extensive testing for those who have difficulties. Assessment and Learning It is generally acknowledged that assessment drives learning; however, assessment can have both in- tended and unintended consequences.22 Students study more thoughtfully when they anticipate cer- tain examination formats,68 and changes in the format can shift their focus to clinical rather than theoretical issues.69 Assessment by peers seems Copyright © 2007 Massachusetts Medical Society. All rights reserved. Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .
  • 8. The new engl and jour nal o f medicine n engl j med 356;4 www.nejm.org january 25, 2007 394 to promote professionalism, teamwork, and com- munication.52 The unintended effects of assess- ment include the tendency for students to cram for examinations and to substitute superficial knowledge for reflective learning. Assessment of Expertise The assessment of trainees and physicians who have higher levels of expertise presents particular challenges. Expertise is characterized by unique, elaborated, and well-organized bodies of knowl- edge that are often revealed only when they are triggered by characteristic clinical patterns.70,71 Thus, experts who are unable to access their knowledge in artificial testing situations but who make sound judgments in practice may do poorly on some tests that are designed to assess commu- nication skills, knowledge, or reasoning. Further- more, clinical expertise implies the practical wis- dom to manage ambiguous and unstructured problems, balance competing explanations, avoid premature closure, note exceptions to rules and principles, and — even when under stress — choose one of the several courses of action that are acceptable but imperfect. Testing either induc- tive thinking (the organization of data to gener- ate possible interpretations) or deductive thinking (the analysis of data to discern among possibili- ties) in situations in which there is no consensus on a single correct answer presents formidable psychometric challenges. Assessment and Future Performance The evidence that assessment protects the public from poor-quality care is both indirect and scarce; it consists of a few studies that show correlations between assessment programs that use multiple methods and relatively crude estimates of quality such as diagnostic testing, prescribing, and refer- ral patterns.72 Correlating assessment with future performance is difficult not only because of in- adequacies in the assessment process itself but also because relevant, robust measures of out- come that can be directly attributed to the effects of training have not been defined. Current efforts to measure the overall quality of care include pa- tient surveys and analyses of institutional and practice databases. When these new tools are re- fined, they may provide a more solid foundation for research on educational outcomes. Conclusions Considering all these challenges, current assess- ment practices would be enhanced if the principles summarized in Table 2 were kept clearly in mind. The content, format, and frequency of assessment, as well as the timing and format of feedback, should follow from the specific goals of the med- ical education program. The various domains of competence should be assessed in an integrated, coherent, and longitudinal fashion with the use of multiple methods and provision of frequent and constructive feedback. Educators should be mindful of the impact of assessment on learning, the potential unintended effects of assessment, the limitations of each method (including cost), and the prevailing culture of the program or in- stitution in which the assessment is occurring. Assessment is entering every phase of profes- sional development. It is now used during the medical school application process,73 at the start of residency training,74 and as part of the “main- Table 2. Principles of Assessment. Goals of assessment Provide direction and motivation for future learning, including knowledge, skills, and professionalism Protect the public by upholding high professional standards and screening out trainees and physicians who are incompetent Meet public expectations of self-regulation Choose among applicants for advanced training What to assess Habits of mind and behavior Acquisition and application of knowledge and skills Communication Professionalism Clinical reasoning and judgment in uncertain situations Teamwork Practice-based learning and improvement Systems-based practice How to assess Use multiple methods and a variety of environments and contexts to capture different aspects of performance Organize assessments into repeated, ongoing, contextual, and developmen- tal programs Balance the use of complex, ambiguous real-life situations requiring reason- ing and judgment with structured, simplified, and focused assessments of knowledge, skills, and behavior Include directly observed behavior Use experts to test expert judgment Use pass–fail standards that reflect appropriate developmental levels Provide timely feedback and mentoring Cautions Be aware of the unintended effects of testing Avoid punishing expert physicians who use shortcuts Do not assume that quantitative data are more reliable, valid, or useful than qualitative data Copyright © 2007 Massachusetts Medical Society. All rights reserved. Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .
  • 9. Medical Education n engl j med 356;4 www.nejm.org january 25, 2007 395 tenance of certification” requirements that several medical boards have adopted.75 Multiple methods of assessment implemented longitudinally can provide the data that are needed to assess train- ees’ learning needs and to identify and remediate suboptimal performance by clinicians. Decisions about whether to use formative or summative assessment formats, how frequently assessments should be made, and what standards should be in place remain challenging. Educators also face the challenge of developing tools for the assess- ment of qualities such as professionalism, team- work, and expertise that have been difficult to define and quantify. No potential conflict of interest relevant to this article was reported. I thank Tana Grady-Weliky, M.D., Stephen Lurie, M.D., Ph.D., John McCarthy, M.S., Anne Nofziger, M.D., and Denham Ward, M.D., Ph.D., for helpful suggestions. References Epstein RM, Hundert EM. Defining and assessing professional competence. JAMA 2002;287:226-35. Batalden P, Leach D, Swing S, Dreyfus H, Dreyfus S. General competencies and accreditation in graduate medical educa- tion. Health Aff (Millwood) 2002;21(5): 103-11. Leach DC. Competence is a habit. JAMA 2002;287:243-4. Fraser SW, Greenhalgh T. Coping with complexity: educating for capability. BMJ 2001;323:799-803. Klass D. Reevaluation of clinical com- petency. Am J Phys Med Rehabil 2000;79: 481-6. Bordage G, Zacks R. The structure of medical knowledge in the memories of medical students and general practitioners: categories and prototypes. Med Educ 1984; 18:406-16. Gruppen LD, Frohna AZ. Clinical rea- soning. In: Norman GR, Van Der Vleuten CP, Newble DI, eds. International hand- book of research in medical education. Part 1. Dordrecht, the Netherlands: Kluwer Academic, 2002:205-30. Epstein RM, Dannefer EF, Nofziger AC, et al. Comprehensive assessment of professional competence: the Rochester ex- periment. Teach Learn Med 2004;16:186- 96. Ericsson KA. Deliberate practice and the acquisition and maintenance of expert performance in medicine and related do- mains. Acad Med 2004;79:Suppl:S70-S81. Epstein RM. Mindful practice. JAMA 1999;282:833-9. Schon DA. Educating the reflective practitioner. San Francisco: Jossey-Bass, 1987. Epstein RM. Mindful practice in ac- tion. II. Cultivating habits of mind. Fam Syst Health 2003;21:11-7. Dreyfus HL. On the Internet (thinking in action). New York: Routledge, 2001. Eraut M. Learning professional pro- cesses: public knowledge and personal experience. In: Eraut M, ed. Developing professional knowledge and competence. London: Falmer Press, 1994:100-22. 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. 13. 14. Shanafelt TD, Bradley KA, Wipf JE, Back AL. Burnout and self-reported patient care in an internal medicine residency program. Ann Intern Med 2002;136:358- 67. Borrell-Carrio F, Epstein RM. Prevent- ing errors in clinical practice: a call for self- awareness. Ann Fam Med 2004;2:310-6. Leung WC. Competency based medical training: review. BMJ 2002;325:693-6. Friedman Ben-David M. The role of as- sessment in expanding professional hori- zons. Med Teach 2000;22:472-7. Sullivan W. Work and integrity: the crisis and promise of professionalism in America. 2nd ed. San Francisco: Jossey- Bass, 2005. Schuwirth L, van der Vleuten C. Merg- ing views on assessment. Med Educ 2004; 38:1208-10. Wass V, Van der Vleuten C, Shatzer J, Jones R. Assessment of clinical compe- tence. Lancet 2004;357:945-9. Van Der Vleuten CPM. The assessment of professional competence: developments, research and practical implications. Adv Health Sci Educ 1996;1:41-67. Schuwirth LW, van der Vleuten CP. Different written assessment methods: what can be said about their strengths and weaknesses? Med Educ 2004;38:974-9. Schuwirth LW, Verheggen MM, van der Vleuten CP, Boshuizen HP, Dinant GJ. Do short cases elicit different thinking processes than factual knowledge ques- tions do? Med Educ 2001;35:348-56. Case S, Swanson D. Constructing writ- ten test questions for the basic and clinical sciences. 3rd ed. Philadelphia: National Board of Medical Examiners, 2000. Farmer EA, Page G. A practical guide to assessing clinical decision-making skills using the key features approach. Med Educ 2005;39:1188-94. Charlin B, Roy L, Brailovsky C, Goulet F, van der Vleuten C. The Script Concor- dance test: a tool to assess the reflective clinician. Teach Learn Med 2000;12:189- 95. Brailovsky C, Charlin B, Beausoleil S, Cote S, Van der Vleuten C. Measurement 15. 16. 17. 18. 19. 20. 21. 22. 23. 24. 25. 26. 27. 28. of clinical reflective capacity early in train- ing as a predictor of clinical reasoning performance at the end of residency: an experimental study on the script concor- dance test. Med Educ 2001;35:430-6. Frederiksen N. The real test bias: influ- ences of testing on teaching and learning. Am Psychol 1984;39:193-202. Schuwirth LW, van der Vleuten CP, Donkers HH. A closer look at cueing ef- fects in multiple-choice questions. Med Educ 1996;30:44-9. Friedman MH, Connell KJ, Olthoff AJ, Sinacore JM, Bordage G. Medical student errors in making a diagnosis. Acad Med 1998;73:Suppl:S19-S21. Graber ML, Franklin N, Gordon R. Diagnostic error in internal medicine. Arch Intern Med 2005;165:1493-9. Pulito AR, Donnelly MB, Plymale M, Mentzer RM Jr. What do faculty observe of medical students’ clinical performance? Teach Learn Med 2006;18:99-104. Norman G. The long case versus ob- jective structured clinical examinations. BMJ 2002;324:748-9. Norcini JJ, Blank LL, Duffy FD, Fortna GS. The mini-CEX: a method for assess- ing clinical skills. Ann Intern Med 2003; 138:476-81. Van der Vleuten CP, Norman GR, De Graaff E. Pitfalls in the pursuit of objec- tivity: issues of reliability. Med Educ 1991; 25:110-8. Anastakis DJ, Cohen R, Reznick RK. The structured oral examination as a method for assessing surgical residents. Am J Surg 1991;162:67-70. Ram P, Grol R, Rethans JJ, Schouten B, Van der Vleuten C, Kester A. Assess- ment of general practitioners by video ob- servation of communicative and medical performance in daily practice: issues of validity, reliability and feasibility. Med Educ 1999;33:447-54. Papadakis MA. The Step 2 clinical- skills examination. N Engl J Med 2004;350: 1703-5. Hodges B, McIlroy JH. Analytic global OSCE ratings are sensitive to level of training. Med Educ 2003;37:1012-6. 29. 30. 31. 32. 33. 34. 35. 36. 37. 38. 39. 40. A video showing an exami- nation involving a standard- ized patient is available with the full text of this article at www.nejm.org. Copyright © 2007 Massachusetts Medical Society. All rights reserved. Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .
  • 10. n engl j med 356;4 www.nejm.org january 25, 2007 396 Medical Education Reznick RK, Blackmore D, Cohen R, et al. An objective structured clinical ex- amination for the licentiate of the Medi- cal Council of Canada: from research to reality. Acad Med 1993;68:Suppl:S4-S6. Wass V, Jones R, Van der Vleuten C. Standardized or real patients to test clini- cal competence? The long case revisited. Med Educ 2001;35:321-5. Carney PA, Eliassen MS, Wolford GL, Owen M, Badger LW, Dietrich AJ. How physician communication influences rec- ognition of depression in primary care. J Fam Pract 1999;48:958-64. Epstein RM, Franks P, Shields CG, et al. Patient-centered communication and diagnostic testing. Ann Fam Med 2005;3: 415-21. Tamblyn RM. Use of standardized pa- tients in the assessment of medical prac- tice. CMAJ 1998;158:205-7. Kravitz RL, Epstein RM, Feldman MD, et al. Influence of patients’ requests for direct-to-consumer advertised antidepres- sants: a randomized controlled trial. JAMA 2005;293:1995-2002. [Erratum, JAMA 2005; 294:2436.] Reznick RK, MacRae H. Teaching sur- gical skills — changes in the wind. N Engl J Med 2006;355:2664-9. Ramsey PG, Wenrich MD, Carline JD, Inui TS, Larson EB, LoGerfo JP. Use of peer ratings to evaluate physician perfor- mance. JAMA 1993;269:1655-60. Dannefer EF, Henson LC, Bierer SB, et al. Peer assessment of professional com- petence. Med Educ 2005;39:713-22. Violato C, Marini A, Toews J, Lockyer J, Fidler H. Feasibility and psychometric properties of using peers, consulting phy- sicians, co-workers, and patients to assess physicians. Acad Med 1997;72:Suppl 1:S82- S84. Norcini JJ. Peer assessment of compe- tence. Med Educ 2003;37:539-43. Nofziger AC, Davis B, Naumburg EH, Epstein RM. The impact of peer assess- ment on professional development. Pre- sented at the Ottawa Conference on Medi- cal Education and Assessment, Ottawa, July 15, 2002. abstract. Lurie SJ, Nofziger AC, Meldrum S, 41. 42. 43. 44. 45. 46. 47. 48. 49. 50. 51. 52. 53. Mooney C, Epstein RM. Temporal and group-related trends in peer assessment amongst medical students. Med Educ 2006; 40:840-7. Lurie S, Lambert D, Grady-Weliky TA. Relationship between dean’s letter rank- ings, peer assessment during medical school, and ratings by internship directors. [Submitted for Publication] (in press). Calhoun JG, Woolliscroft JO, Hockman EM, Wolf FM, Davis WK. Evaluating med- ical student clinical skill performance: relationships among self, peer, and expert ratings. Proc Annu Conf Res Med Educ 1984;23:205-10. Hall JA, Milburn MA, Roter DL, Dal- troy LH. Why are sicker patients less satis- fied with their medical care? Tests of two explanatory models. Health Psychol 1998; 17:70-5. Chang JT, Hays RD, Shekelle PG, et al. Patients’ global ratings of their health care are not associated with the technical quality of their care. Ann Intern Med 2006; 144:665-72. Butterfield PS, Mazzaferri EL. A new rating form for use by nurses in assessing residents’ humanistic behavior. J Gen In- tern Med 1991;6:155-61. Klessig J, Robbins AS, Wieland D, Ru- benstein L. Evaluating humanistic attri- butes of internal medicine residents. J Gen Intern Med 1989;4:514-21. Eva KW, Regehr G. Self-assessment in the health professions: a reformulation and research agenda. Acad Med 2005;80: Suppl:S46-S54. Carraccio C, Englander R. Evaluating competence using a portfolio: a literature review and Web-based application to the ACGME competencies. Teach Learn Med 2004;16:381-7. Committee on Quality of Health Care in America. Crossing the quality chasm: a new health system for the 21st century. Washington, DC: National Academy Press, 2001. Ginsburg S, Regehr G, Lingard L. Basing the evaluation of professionalism on observable behaviors: a cautionary tale. Acad Med 2004;79:Suppl:S1-S4. Schirmer JM, Mauksch L, Lang F, et al. 54. 55. 56. 57. 58. 59. 60. 61. 62. 63. 64. Assessing communication competence: a review of current tools. Fam Med 2005; 37:184-92. Epstein RM, Franks P, Fiscella K, et al. Measuring patient-centered communi- cation in patient-physician consultations: theoretical and practical issues. Soc Sci Med 2005;61:1516-28. Norman GR, Van der Vleuten CP, De Graaff E. Pitfalls in the pursuit of objec- tivity: issues of validity, efficiency and ac- ceptability. Med Educ 1991;25:119-26. Colliver JA, Vu NV, Barrows HS. Screening test length for sequential test- ing with a standardized-patient examina- tion: a receiver operating characteristic (ROC) analysis. Acad Med 1992;67:592-5. Hakstian RA. The effects of type of examination anticipated on test prepara- tion and performance. J Educ Res 1971;64: 319-24. Newble DI, Jaeger K. The effect of as- sessments and examinations on the learn- ing of medical students. Med Educ 1983; 17:165-71. Schmidt HG, Norman GR, Boshuizen HP. A cognitive perspective on medical expertise: theory and implication. Acad Med 1990;65:611-21. [Erratum, Acad Med 1992;67:287.] Bowen JL. Educational strategies to promote clinical diagnostic reasoning. N Engl J Med 2006;355:2217-25. Tamblyn R, Abrahamowicz M, Dau- phinee WD, et al. Association between li- censure examination scores and practice in primary care. JAMA 2002;288:3019-26. Eva KW, Reiter HI, Rosenfeld J, Nor- man GR. The ability of the multiple mini- interview to predict preclerkship perfor- mance in medical school. Acad Med 2004; 79:Suppl:S40-S42. Lypson ML, Frohna JG, Gruppen LD, Woolliscroft JO. Assessing residents’ com- petencies at baseline: identifying the gaps. Acad Med 2004;79:564-70. Batmangelich S, Adamowski S. Main- tenance of certification in the United States: a progress report. J Contin Educ Health Prof 2004;24:134-8. Copyright © 2007 Massachusetts Medical Society. 65. 66. 67. 68. 69. 70. 71. 72. 73. 74. 75. EARLY JOB ALERT SERVICE AVAILABLE AT THE NEJM CAREERCENTER Register to receive weekly e-mail messages with the latest job openings that match your specialty, as well as preferred geographic region, practice setting, call schedule, and more. Visit the NEJM CareerCenter at www.nejmjobs.org for more information. Copyright © 2007 Massachusetts Medical Society. All rights reserved. Downloaded from www.nejm.org by ALEXANDRE A. HOLANDA MD on February 13, 2007 .