A Conceptual Framework for Healthy Eating Behavior inEcuador.docx
Â
QURE_2006
1. Characteristics of health-related self-report measures for children aged three
to eight years: A review of the literature
Joanne Cremeens1
, Christine Eiser2
& Mark Blades2
1
St. Jude Childrenâs Research Hospital, Division of Behavioral Medicine, Memphis, TN, 38105-2794, USA
(E-mail: Joanne.Cremeens@stjude.org); 2
Department of Psychology, University of SheďŹeld, SheďŹeld, UK
Accepted in revised form 19 October 2005
Abstract
Aims: To review and make recommendations about the format and quality of health-related self-report
measures for children aged 3â8 years. Methods: Literature searches used to identify measures of QOL, self-
esteem, self-concept and mental health. The format (i.e., scale type, presentation style) and quality (i.e.,
item generation, reliability, validity, responsiveness) of measures were compared and evaluated. Results:
Fifty three measures were identiďŹed: QOL (n=25, 47%), self-esteem/concept (n=15, 28%), mental health
(n=13, 25%). Likert scales were used most frequently to represent response choices (n=34, 64%). The
authors of 11 (21%) measures provided justiďŹcation for their scale choice. Items were most commonly
presented in written format (n=24, 45%). Item content was generated from the respondent population in
only 21 (40%) measures. Twenty-seven (51%) measures reported internal reliability between 0.70 and 0.90,
and 12 (23%) reported reproducibility in this range. Although validity was reported for 48 (91%) measures,
evidence for three or more aspects occurred for only 9 (17%). Eleven (21%) measures evidenced respon-
siveness to change. Conclusion: Authors should provide clearer evidence for reliability and responsiveness.
Newly developed instruments need to meet established standards, and further studies should assess the
impact of scale and presentation types on the psychometrics of measures.
Key words: Children, Outcome measurement, Quality of Life, Review
Abbreviations: QOL â quality of life; SAC â ScientiďŹc Advisory Committee
Introduction
Traditionally self-report measures of subjective
concepts such as quality of life (QOL) or mental
health have been aimed at children above 8 years,
and proxy reports have been used to gain infor-
mation for children below this age [1, 2]. However
the value of obtaining childrenâs self-reports about
their health, functioning, abilities, and emotions is
increasingly recognized within both medical care
and child health research.
Childrenâs legal status has changed over the last
four decades, with several high proďŹle abuse cases
in the 1970s highlighting the value of childrenâs
own views in legal situations [3]. Gaining
information from children themselves for treat-
ment decisions is becoming necessary from a legal
standpoint [4, 5]. There is evidence that children
can: be competent reporters in legal contexts (for
example recalling events over long time periods);
resist misleading suggestions [e.g., 6, 7]; commu-
nicate their health needs and make decisions about
care if given age-appropriate information [e.g.,
8â10]. In turn, researchers and health professionals
are acknowledging that young children should be
more involved in decisions about their own lives,
families, and health care [11, 12].
Clinical trials designed to evaluate existing, new,
or alternative treatments have traditionally
focused on objective indices such as survival rates
Quality of Life Research (2006) 15: 739â754 Ă Springer 2006
DOI 10.1007/s11136-005-4184-x
2. or reduced morbidity [13]. The inclusion of
subjective outcome measures such as QOL can
help distinguish between treatment programs with
similar physical health outcomes, or between drugs
with apparently equal eďŹcacy and safety [14, 15].
Self-reports on subjective states from patients
themselves can provide information on conse-
quences of treatment plans (such as behavioral or
psychological problems) that may not be captured
by traditional outcome indices [16]. Self-report
measures can help pediatricians and parents make
decisions about the care and treatment of sick
children by providing information about the
quality of childrenâs lives alongside survival time.
Researchers have criticized trials designers for
neglecting to incorporate the patientâs perspective
of outcomes in clinical trials [17â19]. There is
considerable interest and need for inclusion of
subjective child measures in the evaluation of
treatment outcomes for chronic childhood condi-
tions such as cancer [20].
There is evidence that the level of parentâchild
agreement about childrenâs lives and functioning
can be quite poor [e.g., 21â24]. Discrepancies
between parent and child reports could reďŹect real
diďŹerences in perspectives, or parentâs lack of
knowledge or insight into their childrenâs lives and
opinions [25]. Lack of concordance between par-
ent and child reports emphasizes the need to
obtain both perspectives whenever possible [13].
Partly in response to this interest in, and need to
gain information about young childrenâs lives,
some researchers have developed self-report mea-
sures for children below 8 years [e.g., 26â30].
However there is no âgold standardâ regarding
format and content. A variety of response scales
and presentation styles varying in attractiveness
and appropriateness for children have been used
by authors. Guidelines regarding the most appro-
priate and helpful formats for child self-report
measures are needed.
The purpose of this paper is to begin this process
by reviewing the format and quality of health-
related self-report measures for children aged 3 and
8 years. This is a pragmatic review based on the
need to review existing child self-report instruments
(due to the proliferation of papers published in the
last 20 years on the development of new measures).
We focused on QOL, self-esteem, self-concept, and
mental health measures. Literature reviews focus-
ing separately on QOL [e.g., 31â36], self-esteem and
self-concept [e.g., 37â39], and mental health [e.g.,
40, 41] measures are available. Despite existence of
these reviews, there were two reasons for this
review. First, no current review focuses speciďŹcally
on self-report measures for children aged between 3
and 8 years. Second, no current review had collec-
tively evaluated, or attempted to integrate, the
methodologies used in developing and validating
measures from all of these four conceptual areas.
Methods
Rationale for review
The principal rationale for inclusion of QOL, self-
esteem, self-concept, and mental health measures
was the overlap in deďŹnition and conceptualiza-
tion of these assessment areas. One of the most
widely cited deďŹnition of QOL is provided by the
World Health Organization (WHO) who deďŹned
QOL as an individualâs physical health, psycho-
logical states, level of independence, social rela-
tionships, and their relationship to salient features
of the environment [42, 43]. Most deďŹnitions
emphasize also the subjective nature of QOL. For
example, Calman [44] argued that QOL is related
to the perceived gap between an individualâs hopes
and expectations and their present experience. It is
in this deďŹnition of QOL being related to satis-
faction with abilities and functioning, that we see
the link to self-esteem and self-concept measures.
Satisfaction with life has become an important
criterion of individual QOL [35].
Self-esteem has been deďŹned as positive or neg-
ative attitudes about the self, degree of satisfaction
with self, and oneâs feeling of perceived worth as
compared to others (i.e., an evaluative concept)
[45]. Self-concept has been conceptualized as the
way in which our view of self changes over time
due to experiences and social interactions (i.e., a
descriptive concept) [38]. In many studies, self-
concept and self-esteem have been used inter-
changeably, with no distinction made between
them [46]. It is also diďŹcult to create a measure
that taps into only either the descriptive or the
evaluative aspects of self [38].
Assessment of mental health focuses on symp-
tomatology, but must also take into account
740
3. medical, familial, and social factors which may
have contributed to the referral [41]. Therefore,
measures of child mental health have included
items asking children about behavior, self-esteem,
family and peer relationships, and physical health
[41]. Due to these overlaps in deďŹnition, QOL, self-
esteem, self-concept, and mental health measures
have included similar domains and items. There-
fore, we included measures of these four assess-
ment areas in our review.
Search strategy and inclusion/exclusion criteria
The following databases were searched from 1980
to 2004: MEDLINE, ISI Web of Science, Psy-
cINFO, PubMed, Cochrane Central Register of
Controlled Trials, Embase, and ERIC-AT Test
Locator. Databases were searched from 1980 to
2004 to provide information on how the science of
self-report measurement has developed over time.
Text word and thesaurus searches were used to
minimize the chance of missing relevant articles.
The following keywords were searched: child,
childhood, children, pediatric, paediatric; quality
of life, QOL, HRQOL, health status, functional
status, well-being; self-esteem, self-concept, self-
competence, self-image; mental health, depression,
anxiety, fear; measure/s, instrument/s, tool/s.
Internet search engines were used to identify
additional references, and authors were contacted
to identify work in preparation or submitted.
Inclusion criteria were self-report measures of
QOL, self-esteem, self-concept, and mental health,
for children aged between 3 and 8 years. Papers
were excluded if they were targeted only at chil-
dren over 8 years or developed solely for proxy
report of the childâs functioning.
Data extraction and quality assessment
The format of measures was compared across the:
response scales (i.e., Likert, graphic, facial, or
visual analogue); justiďŹcation/evidence reported
for choice of response scale; presentation styles
(i.e., written, pictorial, verbal, computerized or
props); target age ranges.
The quality of measures was assessed by com-
paring: the item generation methods (i.e., how the
content of items was developed); and the reliabil-
ity, validity, and responsiveness data. The psy-
chometric properties of measures have been the
most frequently cited criteria for evaluation of self-
report measures [e.g., 32, 33, 38], i.e., whether
scores on a given measure are reliable, valid, and
responsive.
Reliability is concerned with how well the items
measure an underlying construct, and whether
scores on a measure will remain consistent over
time. The internal consistency of items can be
tested at one time point by calculating the corre-
lations between items. Testâretest reliability can be
assessed by calculating the correlations between
scores at two time points. Criteria outlined by the
ScientiďŹc Advisory Committee (SAC) of the
Medical Outcomes Trust [47] state that internal
consistency and testâretest correlation values
above 0.70 are necessary for health assessment
measures. However, a coeďŹcient value of $0.70
while acceptable, indicates an explained variance
of less than 50% in the variables measured. Some
authors have argued that high internal consistency
(i.e., up to 0.90) should be expected and desired
when validating self-report measures [e.g., 48, 49].
We evaluated measures using a coeďŹcient of 0.70
or above as a minimum cut-oďŹ criteria for evi-
dence of reliability, with ideal coeďŹcient values
being between 0.70 and 0.90 (Table 1).
Validity is concerned with whether measures
assess the construct intended. The guidelines out-
lined by the SAC [47] and used by existing reviews
of child measures [e.g., 32, 36] follow the tradi-
tional conceptualization of validity: content, cri-
terion-related and construct. Evidence for content
validity has been considered by the use of either an
expert panel to rate the appropriateness of items or
statistical procedures to identify how strongly the
item content deďŹnes each domain. Evidence of
criterion validity has been judged by the relation-
ship between scores on a measure with those of a
criterion measure. Evidence of construct validity
has been considered if scores on multiple measures
of the same or related constructs are correlated
(convergent validity), or alternatively scores on
measures assessing diďŹerent concepts are not cor-
related (discriminant validity). Evidence for dis-
criminant validity has also been judged as present
when an instrument discriminates between groups
of individuals known to diďŹer on an accepted
741
7. criterion. Ideally authors should report evidence
for multiple aspects of validity [47].
We acknowledge that a more comprehensive
conception of validity has been deďŹned, where all
three types of validity are taken to be diďŹerent
facets of a single uniďŹed form of construct validity
[50]. However as many of the existing measures of
child self-report were developed and validated
using the traditional conception of validity, we
chose to evaluate included measures using this
conception. Implications of modern test theory for
future work are made in the discussion.
Responsiveness is related to whether scores on an
instrument are sensitive to changes over time. Based
on the SAC [47] guidelines, we considered evidence
of sensitivity where scores on a measure could de-
tect diďŹerences in outcome or real changes experi-
enced over time. For example, this might include
situations where scores on a QOL measure were
sensitive to changes occurring following a targeted
psychosocial intervention. We considered respon-
siveness to be one of the most important aspects of a
measure, following the establishment of reliability.
Constructs such as QOL, self-esteem/concept and
mental health are dynamic processes that do not
remain static over time. Developing (child) mea-
sures which are both sensitive to change over time
and show stability over practical short time periods
is a major challenge for researchers [48].
An additional impact on the validity of
measures is the source used to generate the content
of items. The item generation techniques used can
impact on the content validity of measures [51, 52].
We argue that use of the respondent population (in
this case, children) to generate the content of items
should be a classical principle for any self-report
instrument (especially where the respondentâs
perspective may be substantially diďŹerent from the
developerâs perspective, such as with children, or
cognitively compromised and vulnerable groups).
Results
Search results
Fifty three measures met the inclusion criteria
(Table 1). More QOL (n=25) measures were
identiďŹed, compared to self-esteem/concept
(n=15) and mental health (n=13). The number of
domains in measures ranged from one (various
measures) to 20 (the Nordic Quality of Life
Questionnaire for Children). The total number of
items ranged from 1 [99] to 137 [92]. The majority
of measures were developed in the U.S. or Canada
(n=34). The remaining originated from Europe
(n=15), Australia or New Zealand (n=3), or
Russia (n=1).
Measures
53
Likert 34
(verbal/numerical)
QOL 16
Self-esteem/
self-concept 11
Mental health 7
Graphic 11
(pictorial/3D)
QOL 3
Self-esteem/
self-concept 3
Mental health 5
Facial 7
(cartoon/photo)
QOL 5
Self-esteem/
self-concept 1
Mental health 1
Visual analog 1
(visual/numerical)
QOL 1
Self-esteem/
self-concept 0
Mental health 0
Figure 1. Breakdown of the number of measures by response scale type.
745
8. Format of measures
Response scales
Four response scales were identiďŹed. These
included: Likert (i.e., written linear scale anchored
in sections with numbers and/or words); Graphic
(i.e., three-dimensional or pictorial scale with/
without word/number anchors); Facial expression
(i.e., pictorial linear scale anchored in sections with
cartoon or photographic faces); or Visual analogue
(i.e., visual linear scale anchored at ends with
numbers and/or words). As shown in Figure 1, the
most commonly used response scale was Likert
(n=34, 64%), in the context of either QOL (n=16)
or self-esteem/concept (n=11). The least com-
monly used was visual analogue scales (n=1, 2%).
(Figure 2 gives examples of scale types).
JustiďŹcation/evidence for choice of response scale
Of the 53 measures included, the authors of 11
(21%) provided justiďŹcation for their choice of
response scale. These were classiďŹed into three
areas: testing childrenâs understanding of a chosen
scale (n=7); comparing the psychometric proper-
ties of measures across diďŹerent scales (n=2); or
basing their scale choice on childrenâs own pref-
erences for diďŹerent types (n=2). The authors of
seven measures [27, 28, 59, 70, 82, 88, 92] assessed
childrenâs understanding of their chosen scale.
This was achieved either by asking children:
hypothetical questions (e.g., asking a question
where the appropriate point on the scale can be
predicted to a greater extent); or rating their an-
swer twice (e.g., rating answer on the chosen re-
sponse scale and then on an equivalent scale). The
authors of the CHIP-CE [29] and the TedQL.4 [59]
compared the psychometric properties of the same
measure across diďŹerent response scales (i.e.,
which scale produced the most consistent and
reproducible responses from children).
Presentation styles
Presentation styles were classiďŹed into ďŹve cate-
gories: Written (i.e., child reads items from a
written questionnaire); Pictorial (i.e., items pre-
sented to child using visual aids such as cartoons,
drawings or photographs); Verbal (i.e., written
items read aloud to child); Computerized (i.e.,
items presented to child with the aid of a com-
puter); or Props (i.e., items presented by inter-
viewer verbally using three-dimensional aids, such
as puppets, teddy bears, dolls). As shown in
Figure 2. Example of the graphic scale from the Pictorial Instrument for Child and Adolescent Psychiatry [92] mental health
measure and the facial expression scale from the CAQ [27] QOL measure.
746
9. Figure 3, the most common presentation style was
written format (n=24, 45%), and 12 of these as-
sessed QOL. The least commonly used methods
were computers (n=5) and props (n=4).
Age ranges
As shown in Table 1, most measures targeted
children aged 6 years and above. The widest age
range (5â18 years) was found for the PedsQLTM
[30], however this measure contained age-speciďŹc
versions with parallel items with minor simpliďŹ-
cation of language and adapted response scales.
Half of the QOL measures (13 out of 25) were
directed at children aged 6 years and older. Just
over one third (6/15) of the self-esteem/concept
measures were targeted at children as young as 3
and 4 years (Table 1).
Quality of measures
Item generation and development methods
The item generation methods were unclear for 18
(34%) measures. This omission did not necessarily
mean there was no rationale or justiďŹcation for
content; however it was not reported. The authors
of 21 (40%) measures reported children themselves
informed the content of items (Table 1). Some
researchers interviewed children and asked them to
list, or talk about, relevant topics [e.g., 29, 59],
while others used childrenâs spontaneous com-
ments from pilot work to develop the vocabulary
and wording of their items [e.g., 86]. Authors of an
additional 14 (26%) measures used other tech-
niques to develop item content (namely adaptation
of wording or language of items from existing
child/adolescent measures, or generated from an
expert panel consisting of health professionals,
clinicians and/or psychologists).
The authors of 25 (47%) measures reported
further pilot testing of their items on children, and
resulting adaptation (e.g., alteration of wording, or
deletion/addition of items). Such pilot testing was
undertaken for over half (14/25) of the QOL mea-
sures, half of the self-esteem/concept measures (7/
15), and one third of mental health measures (4/13).
Reliability data
As shown in Table 1, the authors of 42 (79%)
measures reported internal reliability. The most
commonly used statistic for assessing internal
consistency was Cronbachâs alpha reliability coef-
ďŹcient (n=40). Of these measures, 13 (26%) re-
ported values of $0.70, and 27 (51%) evidenced
values of between 0.70 and 0.90. Two measures
[QOL: 64; self-esteem/concept: 87] assessed inter-
nal reliability, but reported only values below the
minimum 0.70 criteria.
Reproducibility, using correlation coeďŹcients
between two time points, was reported for 29
(55%) measures. Ten (19%) of these reported
values of $0.70, and 12 (23%) evidenced values
between 0.70 and 0.90. Seven (13%) measures
assessed reproducibility but reported only values
below the minimum 0.70 criteria (Table 1).
Measures
53
Written 24
QOL 12
Self-esteem/
self-concept 6
Mental health 6
Verbal 6
(interviewer-
administered)
QOL 3
Self-esteem/
self-concept 1
Mental health 2
Pictorial 14
QOL 5
Self-esteem/
self-concept 5
Mental health 4
Props 4
QOL 1
Self-esteem/
self-concept 3
Mental health 0
Computerized 5
QOL 4
Self-esteem/
self-concept 0
Mental health 1
Figure 3. Breakdown of the number of measures by presentation style type.
747
10. Validity data
As shown in Table 1, the authors of 48 (91%)
measures reported evidence for at least one form
of validity. Of these, 16 (30%) measures showed
evidence of one aspect of validity, 23 (43%) evi-
dence for two aspects, and 9 (17%) for three or
more aspects.
The authors of 23 (43%) measures assessed
content validity using one (or more) of three tech-
niques. Items were either rated for relevance,
depth, or breadth using an expert panel (n=4); or
content validity was supported by information
obtained from children themselves (n=6), or by
statistical methods such as factor analysis or prin-
cipal component analysis (n=15). It was noted that
the authors of two measures [29, 86] used more
than one technique to establish content validity.
The authors of 17 (32%) measures reported
assessment of criterion validity. Of the three as-
pects of criterion validity, concurrent validity was
most frequently assessed (n=15). Authors assessed
this using either observed behavior/pain/anxiety/
function scores, achievement tests or diary symp-
tom scores for comparisons. The authors of two
measures [84, 93] assessed predictive validity (i.e.,
investigated the relationship between scores on
their measures with criterion scores at a later time
point). No authors provided evidence for retro-
spective validity.
The authors of 23 (43%) measures assessed
convergent validity by assessing the relationship
between scores on their measure to scores on a
measure of a similar construct. Discriminant
validity was investigated in 22 (42%) measures by
assessing either: whether a given measure can dis-
tinguish between groups known to diďŹer (n=20);
or whether scores on a given measures were
unrelated to other measures assessing diďŹerent
concepts (n=2).
Responsiveness data
Responsiveness was reported for 11 (21%)
measures, shown by evidence that scores on these
measures were sensitive to real changes in QOL,
self-esteem/concept or mental health over time
(Table 1). For example, Le Coq et al. [71] exam-
ined whether scores on the HAY QOL measure
were sensitive to a change in childrenâs asthma
status between two time points (where a change in
symptoms was reported by parents).
Independent testing undertaken
As shown in Table 1, 21 (39%) measures were
validated using independent testing beyond that
performed by the developers of the measure. Such
further testing can help provide increased evidence
for the psychometric properties of measures. It
was noted when this review was performed that
various newly developed measures, namely QOL
instruments [e.g., 55, 63, 73], were still under
evaluation by developers or had not been available
long enough to be tested (and used) by other
researchers.
Discussion
Evaluating the format of existing child self-report
measures
Fifty three child self-report measures have been
described in Table 1. They represent the range of
techniques and formats currently used in child
outcome measures. Likert response scales have
been most commonly employed to represent re-
sponse choices to children, mostly for measures to
assess QOL or self-esteem/concept. It has been
assumed that ratings produced by children on
Likert scales are valid. However Chambers and
Johnston [102] have shown evidence that 5â6 year
old children tend to select responses at the ex-
tremes of Likert-type scales when rating subjective
states (such as emotions).
We identiďŹed a small selection of measures that
used facial and graphic response scales to repre-
sent choices to children, speciďŹcally more recently
developed QOL instruments [e.g., 29, 61, 58]. It
may be that graphic scales (e.g., diďŹering in size or
color) oďŹer children more information on how to
grade their answers to items. However there is
some evidence that young childrenâs ratings may
not actually be equivalent across diďŹerent types of
response scales, i.e., their responses may vary
depending on the type of scale used [103â105], and
the validity of diďŹerent scale types has not been
determined [106]. Therefore we argue that
researchers need empirical evidence to support
scale choice. We identiďŹed 11 measures where au-
thors provided such justiďŹcation, most commonly
by testing childrenâs understanding of chosen scale
types. We argue that direct comparisons of the
748
11. psychometric properties across scale types should
be incorporated in the development stages of new
child measures. Some work has been conducted
into the predictors of childrenâs ability to use
response scales [e.g., 102, 104, 106]. Shields et al.
[106] reported that cognitive ability and chrono-
logical age were the best predictors of kindergarten
childrenâs accurate use of a visual analogue scale.
Further research is needed to establish the cogni-
tive ability and age needed for successful use of
other scale types.
Items were most commonly presented to chil-
dren in a written format. Pictorial presentation
was identiďŹed in a smaller selection of measures.
Using pictures or props to present items to young
children could potentially beneďŹt researchers in
gaining reliable responses. Visual aids can also be
useful in facilitating childrenâs understanding of
items, reducing demands on their memory,
engaging and maintaining their attention, and
avoiding reliance on verbal or reading skills that
may be lacking in younger children [107â108]. In
addition, computers have been utilized by devel-
opers of ďŹve measures [62â64, 66, 89]. The value of
this format becomes apparent for busy clinic set-
tings, and in theory many existing measures using
written formats could be converted to computer-
ized presentation. Studies are needed to determine
how data collected electronically correlates to the
original mode of administration, and also the
value of this new medium for enhancing the
quality of child self-reports.
The measures we identiďŹed were targeted at a
range of ages, with some being aimed at wide age
ranges (e.g., 6â14 years), and others focusing on
narrower age ranges (e.g., 6â8 years). Our review
included any measures for children aged between 3
and 8 years, which meant that measures met the
inclusion criteria if they were targeted at for
example children aged between 7 and 13 years.
Measures targeted at a wide age range may be less
appropriate for the youngest children, as the
content and format may not have been developed
speciďŹcally with these children in mind. There are
diďŹerences in the cognitive capabilities and lin-
guistic skills of children under and over 8 years
[109â111], and this raises the issue as to whether
any one measure will be appropriate for a large
range of ages.
One solution to the age range problem is to use
a similar instrument across a wide age range while
screening for childrenâs ability to understand and
rate items (e.g., using a calibration task). McGrath
et al. [112] developed a calibration task to assess
whether children could make proportional judg-
ments using scales (using ratings of diďŹerent sized
circles). Cummins [113] used a similar task to test
cognitively compromised adultsâ rating abilities
before completion of a QOL instrument (using
wooden cubes of varying sizes). Another solution
is to keep items parallel but make minor simpliďŹ-
cations of language and in length/type of response
scale used. Authors of measures such as the CAQ
[27], CHIPTM
[29], KINDLR
[56], and the Peds-
QLTM
[30] have developed age-speciďŹc versions of
their instruments.
It is also relevant here to consider the impact of
developmental level on child measures. Develop-
mental issues have not received suďŹcient attention
in relation to child outcome measurement [35].
There is a need to include a developmental
framework in the development and validation of
child measures, which takes into account that
constructs such as QOL and self-esteem can have
diďŹerent meanings to children of diďŹerent ages, or
to the same children over time [36]. Factors that
are important to children lives may vary with
diďŹerent stages of development, as well as the
personal relevance of diďŹerent domains [36]. Re-
search is needed to document changes in percep-
tions of health, QOL, self and mental health that
occur as a function of speciďŹc developmental
stages.
Evaluating the quality of existing child self-report
measures
The item generation techniques for over one third
(n=18) of measures were not reported, and this
omission is a serious one. The methods used to
generate information for items can impact directly
on the validity of a measure [51, 52]. Of those
measures where authors did report how items were
generated, only 40% (n=21) reported using chil-
dren themselves to inform content. We urge
developers of new measures to consider using
information from children (respondents) as a
complement to researcher/expert panel generated
749
12. item content, to avoid inclusion of items that are
essentially meaningless to young children [114].
More attention needs to be given to the content of
items, and the language used. Estimates of reading
age could be used to assess the appropriateness of
items within new child measures.
Although nearly 80% (n=42) of authors
assessed the internal reliability of childrenâs scores
on their measures, only 51% (n=27) of measures
evidenced values between 0.70 and 0.90. We argue
that high internal consistency (>0.70â0.90) should
be desired when developing new measures. Issues
concerning âbandwidthâ versus âďŹdelityâ warrant
discussion here [48]. Small sets of items (i.e., 5â10
items) focusing on similar aspects of a given con-
struct will almost always give high internal con-
sistency values, however assessment of constructs
such as QOL often require items spanning a wide
range of diďŹerent aspects [35]. This can mean that
high correlations between items may be harder to
evidence, or if achieved may reďŹect a tendency to
respond similarly to diďŹerent items (termed
âresponse setâ syndrome [48]).
Testâretest reliability was reported for half the
measures (n=29), but only 23% (n=12) of mea-
sures evidenced values between 0.70 and 0.90.
Although the SAC guidelines [47] advocated high
reproducibility values for new measures, for dy-
namic constructs such as QOL low values may
actually represent a real change in levels rather
than a measurement artifact [48]. Therefore sen-
sitivity to change may be a more important crite-
rion than reproducibility in the context of
measuring constructs such as QOL, self-esteem, or
anxiety/depression.
Modern test theory [48] is favored at least within
the adult QOL literature [115â118], as an alterna-
tive to classic psychometric principles. Researchers
such as Schwartz and Rapkin [48] and Bjorner
et al. [119] have discussed alternatives to tradi-
tional psychometrics that will better apply to
measurement of subjective states such as QOL.
Application of modern test theory to child self-
report measurement needs consideration, as the
psychometric evidence for currently available
measures has been couched within classical test
theory and the true score tradition [48].
Based on the traditional concept of validity,
authors of 91% (n=48) of measures reported
evidence for at least one form of validity. Content
and construct validity were most frequently
assessed. The guidelines set out by the SAC [47]
state that it is necessary to provide evidence for all
three types of validity for outcome measures.
Evidence for three or more aspects of validity was
reported for only 17% (n=9) measures. The lack
of attention to establishing validity can seriously
negate the quality of available measures for future
research.
It is relevant to discuss the date of development
of measures, to consider whether our quality
judgments were confounded by recency. Newer
instruments may be of higher quality than older
instruments (due in part to the development of the
science of measurement over the last two decades).
However for many of these newly developed
measures the authors may not have had the
opportunity to establish and report psychometric
concepts (especially those aspects which require
some passage of time such as predictive validity),
or had time for measures to be tested (and used) by
others. Indeed the last decade has shown a large
increase in the development of health-related QOL
measures [120]. Two newly developed computer-
ized QOL instruments evidence this recency eďŹect,
the CAT-SCREEN [62] and the DUX-25 [63],
showing low levels of psychometric validation.
However these computerized measures have not
been as widely available for use in child research as
say the PedsQLTM
[30] or the PAQLQ [68].
The authors of only 11 measures considered the
responsiveness of scores on their measures over
time. Many authors reported evidence for the
reproducibility of childrenâs scores over a short
time period, but did not consider whether their
measures were sensitive to change over longer
periods of time. While it is important that a mea-
sure can produce reproducible scores over a short
time period, a measure also needs to be sensitive to
important changes over time [121]. Again this lack
of attention within the literature needs to be
redressed; either by assessing the sensitivity of
existing measures or ensuring assessment in the
validation stages of new child measures. Estab-
lishing the responsiveness of subjective outcome
measures is a pertinent issue considering calls for
the inclusion of these measures within longitudinal
clinical trials.
750
13. Final comment
In conclusion, there are a variety of measures
currently available for assessing subjective out-
comes through child self-report. However there are
shortcomings with some of the currently available
instruments, speciďŹcally in relation to internal
reliability, validity and responsiveness. In addi-
tion, a variety of diďŹerent response scales and
presentation styles have been employed to repre-
sent items in child measures. There is a need for a
set of minimum standards for child self-report
measures, and further research is needed to reach a
consensus as to the most appropriate formats for
child-centered instruments. Despite the limitations
with current measures and the diďŹculties in
developing new instruments, it is important that
children have the opportunity to describe their
own subjective experience. Children can provide
important information about their health, self-
image and mental state, information that is often
substantially diďŹerent from parentsâ or health care
professionalsâ perspectives.
Acknowledgement
This review was funded in part by a University
of ShefďŹeld grant awarded to the second author.
References
1. OâDonoghue GM, Archbold S. Measuring the immeasur-
able? Introduction. Quality of Life in Deaf Children.
Nottingham, UK: The Ear Foundation, 2002.
2. Bullinger M, Ravens-Sieberer U. Health-related QOL
assessment in children: A review of the literature. Eur Rev
Appl Psychol 1995; 45(4): 245â254.
3. Peterson CC, Siegal M. Cognitive development and the
competence to consent to medical and psychotherapeutic
treatment. In: Siegal M, Peterson CC (eds), Childrenâs
Understanding of Biology and Health. Cambridge, UK:
Cambridge University Press, 1999: 257â279.
4. Department of Health. Health Secretary Announces Mem-
bership of Modernization Actions Teams to Draw up
National Plans for Health. London: Department of
Health, 2000 (Press release 2000/02081).
5. National Institutes of Health. Inclusion of Children as
Participants in Research Involving Human Subjects. (Avail-
able at http://grants.nih.gov/grants/guide/notice-ďŹles/
not98â024.html).
6. Fivush R, Haden C, Adam S. Structure and coherence of
preschoolersâ personal narratives over time: Implications
for childhood amnesia. J Exp Child Psychol 1995; 60: 32â
56.
7. Newcombe PA, Siegal M. Explicitly questioning the
nature of suggestibility in preschoolersâ memory and
retention. J Exp Child Psychol 1997; 67(2): 185â203.
8. Lewis CE, Mewis MA, Lorimer A, Palmer BB. Child-
initiated care: The use of school nursing services by
children in an ââadult-freeââ system. Pediatrics 1977; 60(4):
499â507.
9. Brewster A. Chronically ill childrenâs concepts of their
illness. Pediatrics 1982; 69(3): 355â362.
10. Pantell RH, Lewis CC. Measuring the impact of medical
care on children. J Chron Dis 1987; 40(S1): 99â115.
11. Alderson P, Montgomery J. Health Choices: Making
Decisions with Children. London: Institute for Policy
Research, 1996.
12. Hart C, Chesson R. Children as consumers. BMJ 1998;
316(7144): 1600â1603.
13. Eiser C. Use of quality of life measures in clinical trials.
Ambul Pediatr 2004; 4(S4): 395â399.
14. Varni JW, Seid M, Kurtin PS. Pediatric health-related
quality of life measurement technology: A guide for health
care decision makers. J Clin Outcomes Manag 1999; 6: 33â
40.
15. Testa MA, Lenderking WR. Quality of life considerations
in AIDS clinical trials. In: Finkelstein DM, Schoenfeld
DA (eds), AIDS clinical trials. New York: Wiley-Liss Inc,
1995: 231â41.
16. Hawkins MM, Stevens MC. The long term survivors.
Br Med Bull 1996; 52(4): 898â923.
17. Fitzpatrick R, Davey C, Buxton MJ, Jones DR. Evalu-
ating patient-based outcome measures for use in clinical
trials. Health Technol Assess 1998; 2(14): 1â74.
18. Baker GA, Hesdon B, Marson AG. Quality of life and
behavioral outcome measures in randomized controlled
trials of antiepileptic drugs: A systematic review of
methodology and reporting standards. Epilepsia 2000;
41(11): 1357â1363.
19. Duncan PW, Jorgensen HS, Wade DT. Outcome mea-
sures in acute stroke trials: A systematic review and some
recommendations to improve practice. Stroke 2000;
31(16): 1429â1438.
20. Craft A. Childhood cancer â mainly curable so what next?
Acta Paediatr 2000; 89(4): 386â392.
21. Ennett S, Devellis BM, Earp JA, Kredich D, Warren RW,
Wilhelm CL. Disease experience and psychosocial adjust-
ment in children with juvenile rheumatoid arthritis:
Childrenâs versus mothersâ reports. J Pediatr Psychol
1991; 16(5): 557â568.
22. Glaser AW, Davies K, Walker D, Brazier D. InďŹuence of
proxy respondents and mode of administration on health
status assessment following central nervous system tumors
in childhood. Qual Life Res 1997; 6(1): 43â53.
23. Le Coq EM, Boeke AJP, Bezemer PD, Colland VT, Eijk
JHM. Which source should we use to measure quality of
life in children with asthma: The children themselves or
their parents?. Qual Life Res 2000; 9(6): 625â36.
751
14. 24. Vance Y, Morse R, Jenney M, Eiser C. Methodological
issues in measuring quality of life in childhood cancer:
Measures, proxies and parental mental health. J Child
Psychol Psychiatry 2001; 42(5): 661â667.
25. Jovokic A, Locker D, Guyatt G. How well do parents
know their children? Implications for proxy reporting of
child health-related quality of life. Qual Life Res 2004;
13(7): 1297â1307.
26. Chapman JW, Tunmer WE. Development of young
childrenâs reading self-concepts: An examination of
emerging subcomponents and their relationship with
reading achievement. J Educ Psychol 1995; 87(1): 154â167.
27. Christie MJ, French D, Sowden A, West A. Development
of child-centred disease-speciďŹc questionnaires for living
with asthma. Psychosom Med 1993; 55(6): 514â518.
28. Collier J, Mackinlay D, Phillips D. Norm values for the
Generic Childrenâs Quality of Life measure (GCQ) from a
large school-based sample. Qual Life Res 2000; 9(6):
617â623.
29. Riley AW, Forrest CB, Rebok GW, et al. The child report
form of the CHIP-Child Edition: Reliability and validity.
Med Care 2004; 42(3): 221â231.
30. Varni JW, Seid M, Kurtin PS. Reliability and validity of
the pediatric quality of life inventory version 4.0. Generic
core scales in healthy and patient populations. Med Care
2001; 39(8): 800â812.
31. Connolly MA, Johnson JA. Measuring quality of life in
paediatric patients. Pharmacoeconomics 1999; 16(6):
605â625.
32. Eiser C, Morse R. Quality-of-life measures in chronic
diseases of childhood. Health Technol Assess 2001; 5(4):
1â157.
33. Schmidt LJ, Garratt AM, Fitzpatrick R. Child/parent-
assessed health outcome measures: A structured review.
Child Care Health Dev 2002; 28(3): 227â237.
34. Spieth LE, Harris CV. Assessment of health-related
quality of life in children and adolescents â an integrative
review. J Pediatr Psychol 1996; 21(2): 175â193.
35. Koot HM, Wallander JL. Challenges in child and
adolescent quality of life research. In: Koot HM,
Wallander JL (eds), Quality of Life in Child and Adoles-
cent Illness: Concepts, Methods and Findings. Brunner-
Routledge: Hove, Sussex, 2001: 431â56.
36. Levi R, Drotar D. Critical issues and needs in health-
related quality of life assessment of children and adoles-
cents with chronic health conditions. In: Drotar D (ed),
Measuring Health-Related Quality of Life in Children and
Adolescents: Implications for Research and Practice.
Mahwah, NJ: Lawrence Erlbaum Associates, 1998: 3â24.
37. Beitchman JH, Corradini A. Self-report measures for use
with children: A review and comment. J Clin Psychol
1988; 44(4): 477â490.
38. Davis-Kean PE, Sandler HM. A meta-analysis of mea-
sures of self-esteem for young children: A framework or
future measures. Child Dev 2001; 72(3): 887â906.
39. McDowell AL, Lindholm BW. Measures of self-esteem by
preschool children. Psychol Rep 1986; 59(2): 615â621.
40. Aartman IHA, van Everdingen T, Hoogstraten J, Schuurs
AHB. Appraisal of behavioral measurement techniques
for assessing dental anxiety and fear in children; a review.
J Psychopathol Behav Assess 1996; 18: 153â171.
41. Hunter J, Garralda E, Higginson IJ. Systematic literature
review: outcome measures for child and adolescent mental
health services. J Public Health Med 1996; 18(2): 197â206.
42. World Health Organization. World Health Organization
Constitution. Geneva: World Health Organization, 1994.
43. World Health Group. World Health Organization Quality
of Life Assessment (WHOQOL): Position paper from the
World Health Organization. Soc Sci Med 1995; 41: 1403.
44. Calman KC. DeďŹnitions and dimensions of quality of life.
In: Aaronson NK, Beckmann J (eds), The Quality of Life
of Cancer Patients. New York: Raven Press, 1987: 1â9.
45. Brinthaupt TM, Erwin LJ. Reporting about the self:
issues and implications. In: Brinthaupt TM, Lipka RP
(eds), The Self: DeďŹnitional and methodological Issues.
New York: State University of New York Press, 1992:
137â71.
46. Hughes HM. Measures of Self-Concept and Self-Esteem
for Children Ages 3â12 years: A Review and Recommen-
dations. Clin Psychol Rev 1984; 4: 657â692.
47. ScientiďŹc Advisory Committee of the Medical Outcomes
Trust. Assessing health status and quality-of-life instru-
ments: Attributes and review. Qual Life Res 2002; 11:
193â205.
48. Schwartz CE, Rapkin BD. Reconsidering the psychomet-
rics of quality of life assessment in light of response shift
and appraisal. Health Qual Life Outcomes 2004; 2(16).
49. Varni JW, Burwinkle TM, Lane MM. Health-related
quality of life measurement in pediatric clinical practice:
an appraisal and precept for future research and applica-
tion. Health Qual Life Outcomes 2005; 3(34).
50. Messick S. The once and future issues of validity:
Assessing the meaning and consequences of measurement.
In: Wainer H, Braun HI (eds), Test Validity. Hillsdale,
NJ: Lawrence Erlbaum Associates, 1988: 33â45.
51. McLaughlin JF, Bjornson K. Quality of life and develop-
mental disabilities, editorial. Dev Med Child Neurol 1998;
40(7): 435â436.
52. Bradlyn AS, Ritchey AK, Harris CV, et al. Quality of life
research in pediatric oncology: Research methods and
barriers. Cancer 1995; 78(8): 1333â1339.
53. Jirojankul P, Skevington S. Developing a quality of life
measure for children aged ďŹve to eight years. Br J Health
Psychol 2000; 5: 299â321.
54. Kaplan SH, Barlow S, Spetter D, et al. Assessing
Functional Status and Health-related Quality of Life
among School Age Children with Chronic Disease:
Reliability and Validity of a New Self-report Measure.
Meeting of the Association for Health Services Research,
Chicago, 1995.
55. Parsons SK, Shih MC, Mayer DK, et al. Preliminary
psychometric evaluation of the child health ratings inven-
tories (CHRIs) and disease-speciďŹc impairment inventory-
hematopoietic stem cell transplantation (DSII-HSCT) in
parents and children. Qual Life Res 2005; 14(6): 1613â
1625.
56. Ravens-Sieberer U, Thomas C, Kluth W, Teschke L,
Bullinger M, Lilientahl S. A disease-speciďŹc quality of life
752
15. module for children with cancer â news from the KINDL-
questionnaire. Psychooncology 2001; 10(4): 18â19.
57. Hester N. Child health self-concept scale: Its development
and psychometric properties. Adv Nurs Sci 1990; 7: 45â55.
58. Theunissen NC, Kamp GA, Koopman HM, Zwinderman
KA, Vogels T, Wit JM. Quality of life and self-esteem in
children treated for idiopathic short stature. J Pediatr
2002; 140(5): 507â515.
59. Lawford J, Volavka N, Eiser C. A generic measure of
quality of life for children aged three to eight years:
Results of two preliminary studies. Pediatr Rehab 2001;
4(4): 197â207.
60. Walker DK, Stein REK, Perrin EC, Jessop DJ. Assessing
the psychological adjustment of children with chronic
physical conditions: A review of the technical properties of
PARS III. J Dev Behav Pediatr 1990; 11: 116â121.
61. ManiďŹcat S, Dazord S. Assessing adolescentsâ quality of
life (QOL): Validation of a new questionnaire. QOL News
Lett 2002; 28: 3â4.
62. Ravens-Sieberer U, Heilmann M, Walleser S. Assessment
of quality of life in young children with a computer assisted
touch screen program (CAT-SCREEN) â reliability,
validity, and feasibility. Qual Life Res 2000; 9(3): 298â299.
63. Kolsteren MM, Koopman HM, Schalekamp G, Mearin
ML. Health-related quality of life in children with celiac
disease. J Pediatr 2001; 138(4): 593â595.
64. Eiser C, Vance Y, Seamark D. The development of a
theoretically driven generic measure of quality of life for
children aged 6â12 years: A preliminary report. Child
Care Health Dev 2000; 26(6): 445â456.
65. NeďŹ EJ, Dale JC. Assessment of quality of life in school-
aged children: A method-phase 1. J Matern Child Nurs
1990; 19(4): 313â320.
66. Buller H. Assessment of quality of life in the younger
child: The use of an animated computer program. J Pediatr
Gastroenterol Nutr 1999; 28: S53âS55.
67. Lindstrom B, Eriksson B. Quality of life among children
in Nordic countries. Qual Life Res 1993; 2(1): 23â32.
68. Juniper EF, Guyatt GH, Epstein RS, Ferrie PJ, Jaeschke
R, Hiller TK. Evaluation of impairment of health related
quality of life in asthma: development of a questionnaire
for use in clinical trials. Thorax 1992; 47: 76â83.
69. Juniper EF, Howland WC, Roberts NB, Thompson AK,
King DR. Measuring quality of life in children with
rhinoconjunctivitis. J Allergy Clin Immunol 1998; 101:
163â170.
70. Modi AC, Quittner AL. Validation of a disease-speciďŹc
measure of health-related quality of life for children with
cystic ďŹbrosis. J Pediatr Psychol 2003; 28(7): 535â546.
71. Le Coq EM, Colland VT, Boeke JA, Bezemer DP, Van
Ejik J. Reproducibility, construct validity, and respon-
siveness of the âHow Are You?â (HAY), a self-report
quality of life questionnaire for children with asthma. J
Asthma 2000; 37(1): 43â58.
72. Mishoe SC, Baker RR, Poole S, Harrell LM, Arant CB,
Rupp NT. Development of an instrument to assess stress
levels and quality of life in children with asthma. J Asthma
1998; 35(7): 553â563.
73. Holme SA, Man I, Sharpe JL, Dykes PJ, Lewis-Jones MS,
Finlay AY. The childrenâs dermatology life quality index:
Validation of the cartoon version. Br J Dermatol 2003;
148(2): 285â290.
74. Creer TL, Wigal JK, Kotses H, Hatala JC, McConnaughy
K, Winder JA. A life activities questionnaire for childhood
asthma. J Asthma 1993; 30(6): 467â473.
75. Harter S, Pike K. The pictorial scale of perceived
competence and social acceptance for young children.
Child Dev 1984; 48(5): 80â87.
76. Marsh HW. Physical self description questionnaire: Sta-
bility and discriminant validity. Res Q Exerc Sport 1996;
67(3): 249â264.
77. Perez JR. Perez Self-Concept Inventory. Test manual.
Dallas, TX: Evaluation, Dissemination and Assessment
Center, 1982.
78. Hughes HM, Leatherman MK. ReďŹnement of the Mary-
land Preschool Self-Concept Scale. University of Arkan-
sas, AR: Annual Meeting of American Psychological
Association, 1982.
79. Jensen MA. Development of a preschool self-concept
scale. Early Child Dev Care 1985; 22: 89â107.
80. Stager S, Young RD. A self-concept measure for pre-
school and early primary grade children. J Pers Assess
1982; 46(5): 536â543.
81. Marsh HW, Craven R, Debus R. Structure, stability, and
development of young childrenâs self-concepts: A multi-
cohortâmultioccasion study. Child Dev 1998; 69(4): 1030â
53.
82. Eccles J, WigďŹeld A, Harold RD, Blumenfeld P. Age and
gender diďŹerences in childrenâs self- and task perceptions
during elementary school. Child Dev 1993; 64: 830â847.
83. Measelle JR, Ablow JC, Cowan PA, Cowan CP. Assessing
young childrenâs views of their academic, social, and
emotional lives: an evaluation of the self-perception scales
of the Berkeley Puppet Interview. Child Dev 1998; 69(6):
1556â1576.
84. Verschueren K, Buyck P, Marcoen A. Self-representations
and socioemotional competence in young children: A
three year longitudinal study. Dev Psychol 2001; 37(1):
126â134.
85. Reichenberg K, Broberg AG. Global Self-worth in Chil-
dren with Asthma. Quality of Life Research in Medicine
Conference, St Petersburg, Russia, 2002.
86. Eder RA. The emergent personologist: the structure and
content of three and a half, ďŹve and a half, and seven and
a half year oldsâ concepts of themselves and other persons.
Child Dev 1989; 60: 1218â1228.
87. Davis TM, Johnston JM. On the stability and internal
consistency of the self-concept and motivation inventory:
Preschool/kindergarten form. Psychol Rep 1987; 61: 871â
874.
88. Stein RJ, Bracken BA, Haddock CK, Shadish WR.
Preliminary development of the childrenâs physical self-
concept scale. J Dev Behav Pediatr 1998; 19(1): 1â8.
89. Valla J. The Dominic-R: A pictorial interview for six to
eleven year old children. J Am Acad Child Adolesc
Psychiatr 2000; 39(1): 85â93.
753
16. 90. Martini DR, Straythorn JM, Puig-Antich J. A symptom
self-report measure for preschool children. J Am Acad
Child Adolesc Psychiatr 1990; 29(4): 594â600.
91. Beitchman JH, Raman S, Carlson J, Clegg M, Krudenier
B. The development and validation of the childrenâs self-
report psychiatric rating scale. J Am Acad Child Adolesc
Psychiatr 1985; 24: 413â428.
92. Ernst M, Godfrey KA, Silva RR, Pouget ER, Welkowitz
J. A new pictorial instrument for child and adolescent
psychiatry: A pilot study. Psychiatry Res 1994; 51: 87â104.
93. Carey M, Faulstich M, Gresham F, Ruggiero L, Enyart P.
Childrenâs depression inventory: Construct and discrimi-
nant validity across clinical and non-referred (control)
populations. J Consult Clin Psychol 1987; 55: 755â761.
94. Carlson GA, Cantwell DP. A survey of depressive
symptoms, syndrome and disorder in a child psychiatric
population. J Child Psychol Psychiatry 1980; 21: 19â25.
95. La Greca AM, Stone WL. Social anxiety scale for
children-revised: factor structure and concurrent validity.
J Clin Child Psychol 1993; 22(1): 17â27.
96. Broome ME. The relationship between childrenâs fears
and behavior during a painful event. Child Health Care
1986; 14: 142â145.
97. Melamed BG, Lumley MA. Hospital fears rating scale.
In: Hersen M, Bellack A (eds), Directory of Behavioral
Assessment Techniques. New York: Pergamon Press,
1988: 558â559.
98. Richters JE, Martinez P, Valla JP. Levonn: A Cartoon
based Structured Interview for Assessing Young Chil-
drenâs Distress Symptoms. Washington: National Institute
of Mental Health, 1990.
99. Beyer JE, Aradine CR. Convergent and discriminant
validity of a self-report measure of pain intensity for
children. Child Health Care 1988; 16(4): 274â82.
100. Katz ER, Kellerman J, Sharp B, Siegal SE, Marsten AR,
Hirshman JE. B-endorphin and acute behavioral distress
in children with leukemia. J Nerv Ment Dis 1982; 170:
72â77.
101. Roberts MC, Wurtele SK, Boone RR, Ginther LJ, Elkins
PD. Reduction of medical fears by use of modeling: A
preventive application in a general population of children.
J Pediatr Psychol 1981; 6: 293â300.
102. Chambers CT, Johnston C. Developmental diďŹerences in
childrenâs use of rating scales. J Pediatr Psychol 2002;
27(1): 27â36.
103. Chambers CT, Craig KD. An intrusive impact of anchors
in childrenâs faces pain scales. Pain 1998; 78(1): 27â37.
104. Rebok G, Riley A, Forrest C, et al. Elementary school-
aged childrenâs reports of their health: A cognitive
interviewing study. Qual Life Res 2001; 10(1): 59â70.
105. Laerhoven Hvan , Zaag-Loonen HJ van der, Derkx BHF.
A comparison of Likert scale and visual analogue scales as
response options in childrenâs questionnaires. Acta Pae-
diatr 2004; 93(6): 830â835.
106. Shields BJ, Palermo TM, Powers JD, Grewe SD, Smith
GA. Predictors of a childâs ability to use a visual analogue
scale. Child Care Health Dev 2003; 29(4): 281â290.
107. Mize J, Ladd GW. Predicting preschoolers peer behav-
ior and status from their interpersonal strategies â a
comparison of verbal and enactive responses to hypo-
thetical social dilemmas. Dev Psychol 1988; 24(6): 782â
788.
108. Salmon K, Bidrose S, Pipe M-E. Providing props to
facilitate childrenâs event reports: A comparison of toys
and real items. J Exp Child Psychol 1995; 60: 174â194.
109. Flavell JH.. Cognitive development: childrenâs knowledge
about the mind. Annu Rev Psychol 1999; 50: 21â45.
110. Kalish CW, Gelman SA. On wooden pillows: Multiple
classiďŹcation and childrenâs category-based inductions.
Child Dev 1992; 63: 1536â1557.
111. Siegal M. Knowing Children: Experiments in Conversa-
tion and Cognition, 2nd ed., Hove, Sussex: Psychology
Press, 1997.
112. McGrath PA, Seifert CE, Speechley KN, Booth JC, Stitt
L, Gibson MC. A new analogue scale for assessing
childrenâs pain: an initial validation study. Pain 1996; 64:
435â443.
113. Cummins RA. Self-rated quality of life scales for people
with an intellectual disability: A review. J Appl Res
Intellect Disabil 1997; 10: 199â216.
114. Ronen GM, Rosenbaum P, Law M, Streiner DL. Health-
related quality of life in childhood epilepsy: The results of
children participating in identifying the components. Dev
Med Child Neurol 1999; 41(8): 554â559.
115. Rapkin BD, Schwartz CE. Toward a theoretical model of
quality-of-life appraisal: Implications of ďŹndings from
studies of response shift. Health Qual Life Outcomes 2004;
2(14).
116. Stratford PW, Riddle DL. Assessing sensitivity to change:
Choosing the appropriate change coeďŹcient. Health Qual
Life Outcomes 2005; 3(1): 23.
117. Puhan MA, Bryant D, Guyatt GH, Heels-Ansdell D,
Schu¨ nemann HJ. Internal consistency reliability is a poor
predictor of responsiveness. Health Qual Life Outcomes
2005; 3(1): 33.
118. Visser MRM, Oort FJ, Sprangers MAG. Methods to
detect response shift in quality of life data: A convergent
validity study. Qual Life Res 2005; 14: 629â639.
119. Bjorner JB, Ware JE, Kosinski M. The potential synergy
between cognitive models and modern psychometric
models. Qual Life Res 2003; 12: 261â274.
120. Fayers PM, Machin D. Quality of life: Assessment,
analysis and interpretation. New York: Wiley, 2000.
121. Terwee CB, Dekker FW, Wiersinga WM, Prummel MF,
Bossuyt PMM. On assessing responsiveness of health-
related quality of life instruments: Guidelines for
instrument evaluation. Qual Life Res 2003; 12(14):
349â362.
Address for correspondence: Joanne Cremeens, St. Jude Chil-
drenâs Research Hospital, Division of Behavioral Medicine,
Memphis, TN 38105-2794, USA
Phone.: 001-901-495-2358; Fax: 001-901-495-4701
E-mail: Joanne.Cremeens@stjude.org
754