Review Article
Development and Psychometric Evaluation of Scales: A Survey of
Published Articles
Foroozan Atashzadeh-Shoorideh, PhD1; Farideh Yaghmaei, PhD2∗
PhD in Nursing, Nursing Management Department, Nursing & Midwifery School, Shahid Beheshti University of
Medical Sciences, Tehran, Iran
2
Department of Nursing, Zanjan Branch, Islamic Azad University, Zanjan, Iran.
1
Abstract
Background and purpose: Using valid and reliable instruments is an important way for collecting data
in qualitative researches. This paper is a report of a study conducted to examine the extent of
psychometric properties of the scales in research papers published in Journal of Advanced Nursing.
Methods: In this study, the Journal of Advanced Nursing was chosen for systematic review. All articles
which were published during 2007-2009 in this journal were collected and articles related to instrument
development were selected. Each article was completely reviewed to identify the methods of instrument
validation and reliability.
Results: From 980 articles published in Journal of Advanced Nursing during 2007-2009, 41 (4.18%)
articles were about research methodology. In these, 12 articles (29.27%) were related to developing an
instrument. In this study, review of 12 articles that published in Journal of Advanced Nursing, 20072009, showed that some of the articles did not measure psychometric properties properly, thus some of
the developed scales need to measure other types of necessary validity. In addition, reliability testing
needs to be performed on each instrument used in a study before other statistical analysis are performed.
From 12 articles, all of the articles measured and reported Cronbach’s alpha, but four of them did not
measure test-retest.
Conclusions: Although researchers put a great emphasis on methodology and statistical analysis, they
pay less attention to the psychometric properties of their new instruments. The authors of this article
hope to draw the attention of researcher to the importance of measuring psychometric properties of new
instruments.
Keywords: PSYCHOMETRIC, SCALES, CRITICAL REVIEW
Journal of Medical Education Summer 2015; 14(4):174-205
Introduction1
The credibility of results from a study is
totally dependent on identifying, measuring,
and collecting the right variables. Instruments
are used to measure variables directly from
subjects (1) and research instruments refer to
questionnaires or inventories on which, data
from a research project can be entered and
∗Corresponding
author:
Farideh Yaghmaei,
Associate professor, Department of Nursing,
Zanjan Branch, Islamic Azad University, Zanjan,
Iran.
Email: farideh.yaghmaei@iauz.ac.ir
174
stored for later analysis. An important part in
the process of developing a questionnaire is
to ensure its validity and reliability (2).
Using a valid and reliable instrument is an
integral part of any research. Since
interpretation of results depends on the
validity of instruments used in studies,
researchers should be sure about it (3).
Validity is a significant and complicated issue
which is considered by authors as well as
readers (4). Types of validity includes: face
validity, content validity, construct validity
(factor analysis, validity by convergent
validity, divergent validity, discriminating
analysis) criterion validity (concurrent
Journal of Medical Education
validity and predictive validity), and
successive verifications (5).
Measuring and reporting content validity of
instruments is very important (6). Some
authors in their articles have reported the
process of measuring content validity
frequently, while others did not. This type of
validity can also help to ensure construct
validity and give confidence to the readers
and researchers about instruments. Content
validity is used to measure the variables of
interest. It is also known as content related
validity, intrinsic validity, relevance validity,
representative validity and logical or
sampling validity (7-9). Therefore, content
validity measures the comprehensiveness and
representativeness of the content of a scale
(10, 11).
Construct validity of an instrument is the
theoretical frame or feature of a concept that
the instrument measures such as intelligence,
sorrow, or prejudice. Construct validity can
be calculated by different methods including
contrasted groups, convergent and divergent
analysis or discriminate and factor analysis
(12).
The criterion validity indicates to what degree
the
subject’s
performance
on
the
measurement instrument and subject’s actual
behavior are related. Two forms of criterionrelated validity are concurrent and predictive.
Concurrent validity refers to an instrument’s
ability to distinguish among people who
differ in their present status on the same
criterion (13). Predictive validity refers to an
instrument’s ability to differentiate between
people’s performances or behaviors on the
same future criterion (12).
Reliability refers to the consistency with
which participants of similar characteristics
and outlook understand and respond to the
questions (2). The most common method of
testing a scale’s reliability is Cronbach’s
Alpha coefficient (14), and to determine the
stability of the instrument, a test-retest must
be carried out (15, 16). The internal
consistency may be a necessary condition for
homogeneity or unidimensionality of a scale
Fall 2015, Vol. 14, No. 4
and Cronbach’s alpha should be 0.70 or
higher (14, 17, 18).
Test-retest can be used to determine the
stability of the instrument (15, 16). It is
accomplished
by
administrating
an
instrument, waiting a reasonable period of
time, and then re-administrating the
instrument. The best correlation coefficient
between the two sets of item scores is 0.70 or
higher (1, 16).
Since strong measurement strategy is critical
for proper research (1, 19), this study was
conducted to evaluate the process of
measuring validity and reliability of 12
development instruments papers published in
Journal of Advanced Nursing (JAN) during
2007-2009.
Methods
In this study, the "Journal of Advanced
Nursing" was chosen for review. All articles
published during 2007-2009 in this journal
were collected and articles related to
instrument development were included. Each
article was completely reviewed to identify
the methods of instrument validation and
reliability.
Results
From 980 articles published in Journal of
Advanced Nursing during 2007-2009, 41
(4.18%) articles were about research
methodology. In these research methodology
papers, 12 articles (29.27%) were related to
developing a instrument. Table 1 shows the
features of the articles. None of 12 articles
mentioned their psychometric properties
absolutely (Table 1).
Discussion
Appropriate instruments have a significant
influence on validity of a study. Invalid and
unreliable instruments may show incorrect
results and using findings is doubtful. In
175
Development and Psychometric Evaluation of Scales… / Atashzadeh-Shoorideh et al.
Table 1. Instruments' characteristics of published articles in Journal of Advanced Nursing, 2007-2009.
Author
Instrument/s
/s
Type of validity
Criticism of validity
Type
of
Criticism
reliability
reliability
of
Ushiro
The
Content validity, in this
Content validity is an
Cronbach’s alpha
The
R
psychometric
article was measured by
initial
coefficients
coefficients
(2009)
properties
revising the content and
establishing validity, but
test–retest
0.70 and above
(20)
the
wording
of
Nurse–
in
and
of
based
on
the
the best method in this
reliability
indicate
made
by
the
regard
coefficients were
these scales are
Validity Index (14), that
measured.
internally
with
didn’t measure in this
Cronbach’s alpha
consistent
exploratory factor analysis
study. In addition, the
coefficient for the
All results for
was (CFI) <0.8 and RMSEA
number of person for
physicians’
test–retest
>0.08 for the single-factor
measuring
responses to the
reliability
model, and
validity
Nurse–Physician
satisfactory,
CFI <0.9 and RMSEA <0.08
between 15-20 (9) that
Collaboration
except for the
for the three-factor model.
did not mention in this
Scale
physician
Concurrent
study.
were
Physician
responses
Collaboration
physicians and nurses.
Scale (NPCS)
Factor
analysis:
validity
was
is
Content
content
should
be
0.911
for
responses
patient
regarding
(16).
were
Factor
between nurses’ responses to
exploratory
the
analysis was measured
0.926
Collaboration Scale (NPCS)
and
participation
and the Intergroup Conflict
acceptable but cut-off
the
Scale.
value for factor loadings
decision-making
However, other
wasn’t reported.
process and 0.842
α values were
negative correlations for all
Concurrent validity was
for
0.70
three factors (r = _0.20 to
reported but the ranges
cooperativeness.
which confirms
_0.236, P < 0.01). Among
of correlations for item-
When Cronbach’s
the stability of
the
between
totals
inter-item
alpha coefficient
the scales.
physicians’ responses to the
were
The
of the item-total
The
Nurse–Physician
concurrent validity value
correlations were
correlation
Collaboration Scale (NPCS)
must be ranging from 0
compared
coefficients
and the Intergroup Conflict
to +1 (4).
those
Scale, there were statistically
Convergent validity was
when an item had
mentioned and it
significant small negative
reported but these ranges
been
is acceptable.
correlations
were
no
Nurse–Physician
There
were
significant
relationship
for
shared
analysis
(NPCS)
that
measured by relationships
statistically
176
step
alpha
factor
reported.
and
low.
low.
with
It
is
The
shared
information,
for
sharing
joint
in
cure/care
with
obtained
eliminated,
items
was
patient’s information, (r =
convergent
validity
found lower than
_0.165,
value must be ranging
coefficient value.
P
<
0.01)
and
of
patient
information
(0.629).
nurses
–
0.92,
test–retest
for
were
Journal of Medical Education
Fall 2015, Vol. 14, No. 4
cooperativeness. (r = _0.152,
from 0 to +1 (4, 5). In
The
P < 0.01).
addition;
correlation values
Convergent
done
validity
with
the
Characteristic
with
both
was
Team
Scale
the
and
nurses’
the
item-total
psychometric of the used
were
high,
scale
ranging
from
for
validity
mentioned.
convergent
did
not
0.502 to 0.801.
The
item-total
responses (r = 0.360–0.523,
correlation values
P < 0.01) and physicians’
were
high,
responses (r = 0.435–0.639,
ranging
from
P < 0.01) to the Nurse–
0.423 to 0.787.
Physician
The
Collaboration
test–retest
Scale (NPCS). The used
(The
scale
between the first
in
this
study
for
interval
convergent validity did not
and
the
second
validate or didn’t report its
test
validity and reliability.
weeks) correlation
was
coefficients
2–3
for
nurses were 0.710
(P<0.01)
for
sharing of patient
information,
0.658
(P<0.01)
for
joint
participation
the
in
cure/care
decision-making
process,
and
0.676 (P < 0.01)
for
cooperativeness.
The
test–retest
correlation
coefficients
physicians
for
were
0.624 (P < 0.01)
for sharing
patient
information,
0.798 (P < 0.01)
177
Development and Psychometric Evaluation of Scales… / Atashzadeh-Shoorideh et al.
for
joint
participation
the
in
cure/care
decision-making
process and 0.774
(P < 0.01) for
cooperativeness.
Author
Instrument/s
/s
Type of validity
Criticism of validity
Type
of
Criticism
reliability
reliability
Chang
Chinese
Content validity, in this
Content validity or face
Cronbach’s alpha
Internal
H-J
version of
article was not measured.
validity is an initial step
coefficients
consistency
Factor
in establishing validity
test–retest
based
reliability
suggested
et
al the
Positive
analysis
was
and
on
(2009)(
and Negative
examined by using both with
(6)
2)
Suicide
exploratory factor analysis
measured in this study.
coefficients were
criterion
Ideation
and
Factor
measured.
indicating
(PANSI)
analysis (CFA) and all item-
exploratory
Inventory
total
analysis
confirmatory
factor
coefficients
ranged
that
was
not
analysis
with
the
level
factor
Cronbach’s alpha
adequate internal
and
Coefficients were
consistency for a
from 0.42 to 0.71. The
confirmatory
factor
0.86 and 0.94 for
coefficient’s α of
results indicated that the two
analysis was measured
the total scores on
0.70 or above
factor oblique model had the
and
the positive and
(14).
best fit. The confirmatory
acceptable but, cut-off
negative
suicide
factor analysis using the two
value for factor loadings
ideation
positive
factor model yielded the
wasn’t reported.
ideation (PANSI-
following results: CFI =
Convergent validity was
PI)
0.950, RMSEA=0.078.
reported
these
positive
and
was
ranges were moderate
negative
suicide
demonstrated by statistically
level. The convergent
ideation-negative
significant
positive
validity value must be
suicide
correlations between total
ranging from 0 to +1. If
(PANSI-NSI)
scores on the positive and
the convergent measures
respectively.
negative suicide ideation-
are closely related, the
The
negative
validity
(The
Convergent
validity
and
suicide
ideation
(PANSI-NSI)
and
the
Children’s
Depression
Inventory (CDI) (r=0.61),
reported.
and
of
instrument
It
is
each
is
and
the
ideation
test–retest
interval
between the first
strengthened (Burns and
and
Grove 2007).
test was 4 weeks)
the
second
was carried out.
the positive and negative
positive
Divergent validity was
Intra-class
ideation (PANSI-PI) and the
reported but the ranges
correlation
Cognitive Triad for Children
of
coefficients were
suicide
178
The
of
ideation
correlations
were
Journal of Medical Education
Fall 2015, Vol. 14, No. 4
Inventory
=
moderate. The divergent
0.82 and 0.70 for
and
validity value must be
the total scores on
ideation
ranging from -1 to 0. If
the positive and
positive ideation (PANSI-PI)
the convergent measure
negative
suicide
and the self-control schedule
of
ideation
positive
(SCS) (r = 0.46).
negatively
correlated
ideation (PANSI-
with
measures,
PI) and positive
0.65),
(CTI-C)
the
negative
(r
positive
suicide
Divergent
validity
was
instrument
other
is
demonstrated by statistically
validity for each of the
and
negative
significant
instrument
suicide
ideation-
and
correlations
negative
between
strengthened (Burns and
negative
total
Grove 2007).
ideation (PANSI-
Scores on the positive and
The
negative
ideation
predictive validity and
correlations were
positive ideation (PANSI-
the score of this study is
statistically
acceptable.
significant at the
suicide
the
PI),
the
is
Children’s
Depression Inventory (CDI)
(r=-0.52),
suicide
suicide
the
process
of
suicide
NSI).
All
P<0.05 level.
negative
ideation-negative
ideation
(PANSI-
NSI), the Cognitive Triad
for Children Inventory (CTIC)
(r=-0.52),
and
the
negative suicide ideationnegative
suicide
ideation
(PANSI-NSI) and the selfcontrol schedule (SCS) (r= 0.30). All correlations were
statistically significant at the
P<0.01 level.
Predictive
Validity
was
measured one year after
first-wave study with the
Chinese
Version
Positive
and
of
the
Negative
Suicide Ideation Inventory
(PANSI-C).
Logistic regression analysis
showed that the total score
on
the
negative
suicide
179
Development and Psychometric Evaluation of Scales… / Atashzadeh-Shoorideh et al.
ideation-negative
suicide
ideation (PANSI-NSI) in the
first-wave study statistically
significantly predicted the
attempted- suicide behaviour
after 1 year (coefficient =
0.095, P<0.001; CI = 1.05–
1.15).
The
overall
classification rate was good,
at 89.4%. The total score of
the positive and negative
suicide
ideation
positive
ideation (PANSI-PI) in the
first-wave
study
statistically
also
significantly
predicted
the
attempted
suicide behaviour after 1
year (coefficient = _0.084,
P<0.05, CI = 0.86–0.99).
Author
Instrument/s
/s
Type of validity
Criticism of validity
Type
of
Criticism
reliability
reliability
The
of
Eizenbe
Moral Distress
Content validity, in this
Content validity is an
Internal
rg MM
Questionnaire
article was not measured.
initial
consistency
et
for
Factor
was
establishing validity (6,
measured
using
16), that didn’t measure
using Cronbach’s
indicate
exploratory factor analysis
in this study.
alpha.
these scales are
and
Measuring and reporting
three factors the
internally
of content validity in
internal
consistent
0.56 to 0.90. The results
questionnaire
consistency
is
16).
indicated
three
developing is necessary
above 0.79 (for
The
factors yielded. The authors
and important (16). It is
three factors are
correlation
didn’t report CFI and other
recommended
0.851, 0.791 and
coefficients were
results of factor analysis.
determine
0.804).
mentioned but it
But they mentioned cut-off
validity before construct
is low (1). It is
value.
validity.
recommended to
al
(2009)
(22)
Nurses
Clinical
analysis
examined
by
all
coefficients
ranged
that
Discriminate
addition,
180
item-total
the
from
validity:
to
In
provide
step
Factor
in
to
content
analysis
exploratory
For
was
by
the
with
factor
Stability
construct validity of the
and
examined by use
is
of
0.70 and above
that
(15,
test–retest
the
items in second
analysis was measured
It
coefficients
increase
additional evidence for the
reported.
alpha
was
version of this
questionnaire.
Journal of Medical Education
Fall 2015, Vol. 14, No. 4
test-retest
questionnaire, a comparison
necessary reporting of
of
was
their
reliability
made
between
two
results
but
the
(The
groups (hospital nurses and
authors didn’t report CFI
interval
community clinic nurses), as
and
the first and the
it
factor analysis (23).
was
assumed
that
would
be
differences
other
results
of
between
second test was 1
month).
The
observed
in
pressure
correlation
resulting
from
different
between the two
moral
dilemmas.
To
measurements
examine these differences, t-
was
tests
(P<0.001), 0.385
for
independent
samples were conducted. A
(P<0.05)
statistically
0.535
significant
difference
was
found
three
and
(P<0.01)
respectively
between means for two of
the
0.624
for
the three factors.
factors
relationships and time (For
relationship t=2.171 and for
time
t=2.208).
These
differences provide further
evidence
for
the
discriminant validity of the
questionnaire.
Author
Instrument/s
/s
Type of validity
Criticism of validity
Type
of
Criticism
reliability
reliability
of
Liu M
Competency
Content validity, in this
Content validity is an
Internal
Measuring
et
Inventory for
article and Content Validity
initial
consistency
reliability
(2009)
Registered
Index (CVI) was reported
establishing validity (6),
reliability
(24)
Nurses
based on the other studies.
and it supports construct
stability
Factor
validity (3) that didn’t
estimated
exploratory factor analysis
measure in this study.
Cronbach’s alpha
was (CFI) <0.8 and RMSEA
Measuring and reporting
and paired t-test,
The
>0.08 for the single-factor
of content validity in
respectively.
indicates a high
model, and
questionnaire
Internal
degree
CFI <0.9 and RMSEA <0.08
developing is necessary
consistency
stability over a
for the three-factor model.
and important (16). It is
Cronbach’s alpha
period of time
Confirmatory factor analysis
recommended
to
was 0.90 for the
and satisfactory
was employed to test the
determine
content
overall scale and
degree
construct validity of the
validity
Content
0.71–0.90
homogeneity (8).
al
Macao
in
analysis
with
step
and
in
and
were
is
reported and is
acceptable.
by
for
stability
of
of
181
Development and Psychometric Evaluation of Scales… / Atashzadeh-Shoorideh et al.
instrument.
loading
The
value
factor
across
55
Validity Index (CVI) in
subscales.
The best interval
every
Internal
time
questionnaire
items ranged from 0.310 to
developing (6, 16).
0.725.
Factor
A cut-off value of
analysis
0.3 for factor loadings was
exploratory
applied as this is considered
analysis
to
confirmatory
indicate
statistical
significance.
consistency
between
was
first and second
with
0.74. The interval
test in test-retest
factor
between the first
is 2-4 weeks (5,
and
and
16).
factor
test
analysis was measured
the
second
didn’t
reported.
It
is
recommended to
and reported.
report of interval
between
two
tests.
Author
Instrument/s
/s
Type of validity
Criticism of validity
Type
of
reliability
Criticism
of
reliability
Zisberg
Scale of Older
Content validity: In this
Content
A,
Adults’
study, items were generated
Measuring and reporting correlation
should
be
Young
Routine
on the basis of a literature
of content validity in coefficient
considered
as
HM &
(SOAR)
review
questionnaire developing statistics
and
then
Schepp
systematically
tested
for
K
content validity. Then, the
(2009)
(25)
necessary
and used
important (16, 19).
ICC
were
to
test
scores
reliability
indices in four
reliability at the
groups
instrument’s content validity
item level of the
estimate
was rated on the basis of the
continuous scores
high
of
>0.80),
levels:
(ICC
instrument’s item relevance
Convergent validity was
as
to older adult routine in the
reported
subscale
pilot sample. The relevance,
ranges were moderate
Across all types
<ICC<0.80),
clarity, simplicity based on
level. In this study, the
of
moderate (0.41<
Content
Validity
authors
(50%)
ICC <0.60) and
(CVI)
items
consistently
poor to fair (ICC
presented
<0.40) (26).
moderate to high
Kappa
coefficient
Index
weren’t
and
these
reported
convergent
the
validity
between
reported.
Convergent validity:
In
_1
to
0.
But
the
well
as
scores.
scores,
21
substantial (0.60
order to test the convergent
convergent
validity
test–retest
validity of SOAR, the mean
value must be ranging
reliability
deviation
scores
the
from 0 to +1. If the
>0.41). Six items
subscale
level
correlated
convergent measures are
(14.3%) presented
functional
closely
related,
poor reliability on
The
indicators (ADL and IADL).
validity
of
all
correlation
The ADL score was found to
instrument
be negatively correlated with
strengthened (5).
with
182
is
validity: Intra-class
the
on
the
each
is
four
(ICC
scores
almost
is
perfect
(27).
test–retest
(ICC <0.40).
coefficients were
These items were
mentioned but, it
the consistency of time spent
shopping, passive
is low (1).
(mean deviation score for
transportation,
Reliability
for
Journal of Medical Education
Fall 2015, Vol. 14, No. 4
duration) on each basic and
and
medical
overall
scale
rest activity (r = -0.41, -0.34;
treatment,
wasn’t
P < 0.01 respectively), as
attending
mentioned.
well as with the consistency
concerts/movies/s
of total time spent on basic
ports
and rest activities (mean
participating
deviation score for total
group
duration, r = 00.56, -0.33; P
and taking care of
< 0.01 respectively).
an older person.
events,
in
activities
On the subscale
level, over 73% of
the scores showed
high to substantial
reliability
and
none showed poor
reliability.
Kappa
coefficients
was
done for nominal
variables and it
was
over
0.75
%
of
(item
agreement
=
88.4%–100%).
Only16.6%
had
kappa coefficients
in the low range (j
< 0.40).
Test–retest
reliability
for
subscales is 0.46
to
0.85.
interval
The
between
the first and the
second test didn’t
report.
Author
Instrument/s
/s
Pelande
Child
Care
Type of validity
Criticism of validity
Content validity: In this
Content
validity:
Type
A
of
Criticism
reliability
reliability
Internal
A correlation
183
of
Development and Psychometric Evaluation of Scales… / Atashzadeh-Shoorideh et al.
r
T,
Quality
at
study, following a literature
scale-level CVI of %75
consistency
or higher is acceptable.
using
between
The reporting of content
Cronbach’s alpha
and
validity index must be
was 0.373–0.812
desirable,
Leino-
Hospital
review
Kilpi H
(CCQH)
drawings
&
instrument
children
and
interviews/
by
hospitalized
by
Coefficient
0.90
Katajist
(n=40),
were
based on percent (3, 16).
for subscales, but
0.70
o
designed and an expert panel
Factor analysis didn’t
for
acceptable
(2009)
(n=7)
the
report obviously. The
scale
(28)
instrument’s
content
process of it should be
report.
clear.
The alpha values
J
the
assessed
validity.
To
judge
the
the
overall
didn’t
showed
a
subcategories on a scale
tendency
to
from
increase
one
to
four
for
is
but
is
for
new
instruments (29).
validity of the items and
during
Correlations for
relevance and clarity; to
the course of the
item-totals
and
indicate
instrument
inter-item
were
whether
(yes/no)
a
or
not
subcategory
development
main category; whether or
categories:
not
nursing
subcategories or
characteristics
increasing more
subcategory
quality
and
whether or not there was any
from
overlap
0.557,
between
the
main
reported.
all
the
the
for
belonged to a particular
measured
in
0.383
to
nursing
certain
items, especially
in
the
activities
least relevant subcategories
0.763 to 0.809,
improve
were 0.38 and 0.67, so these
and
reliability (30).
items deleted.
environment from
The
least
from
Combining
different subcategories. The
nursing
clarity of subcategories was
0.584 to 0.761.
0.65 and 0.69, whereas the
Item-to-total
level of agreement for all
correlations were
other subcategories was over
calculated for the
0.90. Level of agreement
various
among nurses was over 0.95
subcategories
for
subcategories
nursing activities
measuring quality, except
and environment
for appearance (0.37), sense
and for the main
of
and
category of nurse
In
the
characteristics.
assessments,
the
Item-to-total
all
humour
humanity
nurses’
184
items
0.80
(0.69)
(0.93).
in
subcategories of humanity
correlations
(0.31),
ranged from 0.062
caring
and
subcategory, can
the
Journal of Medical Education
Fall 2015, Vol. 14, No. 4
communication (0.31), and
to
education (0.31) showed the
lowest
greatest overlap with other
total correlations
subcategories.
were obtained for
factor
The
analysis
0.611.
The
item-to-
of
the subcategories
was assessed by
of physical care
using principal component
and treatment, and
analysis to measure the level
entertainment.
of congruence of empirical
The items ‘takes
results
main
account of child’s
nursing
food preferences’
CCQH
with
categories
the
of
activities and environment.
and
No
component
relief
for
pain’
analysis was carried out for
were
the
most
the main category of nurse
problematic.
characteristics.
These items were,
principal
‘provides
however,
not
deleted from the
instrument
as
their contents are
crucial
in
this
context.
Author
Instrument/s
/s
Type of validity
Criticism of validity
Type
of
Criticism
reliability
reliability
of
Carlson
Carlson’s
Content validity was done
The reporting of content
Cronbach’s alpha
The
C
Prior
by reviewing literature and
validity in this study is
coefficients were
coefficients
(2008)
Conditions
theoretical
acceptable.
measured.
0.70 and above
(31)
Instruments
was
The reporting of content
instrument
indicate
validity index must be
demonstrated
these scales are
(CPCIs),
to
definition
supported
review
by
and
through
experts.
The
Each
alpha
of
that
assess the four
average of CVI
based on percent (3, 16).
internal
internally
theoretically-
scores for relevancy of all
Rattray and Jones
consistency (alpha
consistent
derived
items within each instrument
suggest that a KMO
range=
19).
conditions of
were
(The
greater than 0.5 supports
0.825).
Previous
average of CVI scores of all
a factor analysis, and
retest reliability
felt
items within each instrument
that anything less than
needs
needs/problem
were 1.0 for the Previous
0.5
confirmed
s,
Practice Instrument, 0.79 for
amenable
innovativeness
the
factor analysis. So, this
practice,
prior
0.79
Felt
to
1.0
Needs/Problems
is
probably
to
not
useful
0.731–
(16,
In addition, test–
to
assess
be
to
the
stability of the
185
Development and Psychometric Evaluation of Scales… / Atashzadeh-Shoorideh et al.
and norms of
Instrument, 0.94 for the
KMO
the
Innovativeness
acceptable.
social
system
that
Instrument,
measure
is
Inter-item
measures
correlations
and 0.98 for the Norms of
For
influence
the
accurate
nurses’
Instrument).
another type of construct
analysis
The clarity, simplicity based
validity
such
as
internal
adopt
on Content Validity Index
predictive
validity
is
consistency
evidence-
(CVI)
needed (6).
decisions
to
based
pain
Social
System
items
was
not
achieving
more
instrument,
are
time (6).
between 0.2 and
0.7.
After
item
for
reliability,
reported.
over
the
Previous Practice
management
Factor
analysis
practices.
examined through principal
reduced
components factor analysis
items,
with varimax rotation and
Needs/ Problems
reported for each factor of
Instrument to 14
instruments.
items,
was
Instrument
Factors were
was
to
the
13
Felt
the
established using the Kaiser
Innovativeness
rationale
Instrument to nine
by
retaining
eigenvalues over 1.0.
items,
and
the
To establish salient factors,
Norms
of
the
the items with correlations
Social
above 0.3 on more than one
Instrument to nine
factor were deleted, as they
items.
were
The
were 0.825, 0.76,
Kaiser–Meyer Olkin (KMO)
0.731 and 0.775
measure of sample adequacy
respectively.
repetitious.
System
Alphas
was then determined. The
KMOs of Carlson’s Prior
Conditions
Instruments
(CPCIs) ranged from 0.655
to 0.841.
Author
Instrument/s
/s
Type of validity
Content validity is an
Internal reliability
The
initial
was estimated by
consistency
establishing validity (6),
calculating
Cronbach’s
Factor analysis: exploratory
and it supports construct
Cronbach’s alpha
alpha
factor
validity (3) that didn’t
coefficient for the
The
confirmatory factor analysis
measure in this study.
scale(s)
coefficients
was done. Construct validity
Factor analysis:
from the analysis
R
Coping
validity was not measured.
Efficacy
article,
for
(2008)
Nurses
(32)
(OCSE-N)
186
this
content
In
al
Scale
Criticism
reliability
Occupational
et
of
reliability
Pisanti
Self-
Type
Criticism of validity
analysis,
and
step
in
with
the
derived
of
Internal
reported.
alpha
of
0.7 and above
Journal of Medical Education
with
Fall 2015, Vol. 14, No. 4
exploratory
factor
exploratory
analysis is (CFI) <0.75and
analysis
RMSEA >0.15 for the first
confirmatory
model, and CFI <0.92 and
RMSEA
<0.08
for
the
second model.
Concurrent
assessed
validity
by
correlations
was
estimating
between
the
factor
and
indicate
whether
these scales are
internally
analysis was measured
Cronbach’s alpha.
consistent
and
Cronbach’s alpha
18).
acceptable.
reliability
In
Criterion validity was
done
reported carefully.
subscales
It
is
item
every
that
increased
reported.
factor
and by checking
were
for
addition,
two
reliability
(For
as
such
test-retest
‘CSE to manage
needs
general
confirmed
nursing
(16,
to
be
to
Occupational Coping Self-
burden’ alpha =
assess
Efficacy for Nurses Scale
0.77;
stability of the
(OCSE-N) dimensions and
‘CSE to manage
measures
two
the
time (6).
external
criteria:
and
for
relational
Maslach Burnout Inventory
difficulties in the
(MBI)
workplace’, alpha
dimensions
coping
and
dimensions.
Pearson’s
the
over
= 0.79).
correlation
coefficients
between
the
Occupational Coping SelfEfficacy for Nurses Scale
(OCSE-N) dimensions and
both the Maslach Burnout
Inventory (MBI) variables
and Coping Inventory for
Stressful Situations – Short
Version
(CISS-SV)
dimensions
were
all
statistically significant. The
OCSE-N dimensions were
positively associated with
task coping strategies (r =
0.07 to 0.08, P < 0.05) and
negatively associated with
both emotion-focused and
avoidant
strategies
(r
=
_0.09 to _0.08, P < 0.01).
The OCSE-N Scales also
correlated with the burnout
187
Development and Psychometric Evaluation of Scales… / Atashzadeh-Shoorideh et al.
dimensions.
They
were
negatively correlated with
both emotional exhaustion
(r= _0.31 to _0.21, P < 0.01)
and depersonalization (r =
_0.25 to _0.19, P < 0.01),
and
positively
associated
with
personal
accomplishment (r = 0.21 to
0.22, P < 0.01). These
patterns
of
correlations
support
the
construct
validity of the Occupational
Coping
Self-Efficacy
for
Nurses Scale (OCSE-N).
Author
Instrument/s
/s
Type of validity
Content
study,
of
Criticism
reliability
reliability
Content validity is an
Cronbach’s alpha
The
initial
coefficient
coefficients
of
Barnes
Perceived
In
C.R.&
Maternal
validity
Adams
Parenting
reviewing
literature
and
establishing validity, but
used to calculate
0.70 and above
on-
Self-Efficacy
theoretical
definition
and
the best method in this
internal
indicate
Maced
(PMP
was
regard
consistency
these scales are
o E.N.
instrument
S-E)
this
Type
Criticism of validity
was
done
supported
by
through
step
is
in
Content
was
alpha
review by participants in a
Validity Index (14), that
reliability
internally
(2007)
pilot study.
didn’t measure in this
estimates for the
consistent
(33)
Factor
study.
Perceived
18).
measured and cut-off value
Construct validity with
Maternal
of 0.3 for factor loadings
exploratory
Parenting
was
analysis was measured
Efficacy (PMP S-
and
E)
analysis
applied
considered
statistical
as
to
was
this
is
indicate
significance.
reported.
It
is
instrument;
The
correlation
Factor 1 had an Eigen value
their
acceptable
coefficient
of 8.235 and explained 41%
authors didn’t report CFI
(0.91).
of the variance, factor 2 had
and
internal
an Eigen value of 1.496 and
factor analysis (23).
consistency
explained
In addition, cut-off point
reliability
variance, factor 3 had an
is low.
estimates for each
Eigen value of 1.314 and
Divergent validity was
of the subscales
explained
reported but the ranges
were
of
acceptable
7.48%
6.57%
of
of
the
the
other
but
the
results
correlations
of
were
(16,
test–retest
this reached an
results
that
Self-
necessary reporting of
variance, and factor 4 had an
188
factor
of
level
The
mentioned and it
is acceptable.
also
was
Journal of Medical Education
Fall 2015, Vol. 14, No. 4
Eigen
value
0.255
moderate. The divergent
[subscale 1 (0.74),
explaining 6.27% of the
validity value must be
2 (0.89), 3 (0.74)
variance.
ranging from -1 to 0. If
and 4 (0.72)]. In
Divergent Validity by using
the convergent measure
addition,
the
of
is
whole correlation
Maternal
of
Self-Report
instrument
item-
Inventory was rs = 0.4 (P <
negatively
correlated
revealed that all
0.05) and using the Maternal
with
measures,
items
Postnatal Attachment Scale
validity for each of the
statistically
was rs = 0.31, (P< 0.01).
instrument
significantly with
other
is
strengthened (4).
correlated
total
scores
(ranging
from
0.30–0.77).
test–retest
The
(The
interval
between the first
and
the
second
test was 10 days)
correlation
coefficients
was
0.96.
Author
Instrument/s
/s
Type of validity
Criticism of validity
Type
of
Criticism
reliability
reliability
The
of
Van
Work-Related
Survey of the literature and
Content validity is an
Internal
Laar, D
Quality of Life
qualitative expert reviews
initial
consistency
et
(WRQoL)
were used to assess the
establishing validity, but
using
Cronbach’s
content
the best method in this
Cronbach’s alpha
alpha
regard
Content
was 0.75–0.86 for
The
analysis,
Validity Index (14), that
subscales, and for
coefficients
exploratory factor analysis
didn’t measure in this
the overall scale
0.7 and above
and
factor
study.
was 0.96.
indicate
analysis were done. A cut-
Factor
off value of 0.5 for factor
exploratory
loadings was applied. By
analysis
using
confirmatory
al.
(2007)
scale
for
(34)
healthcare
measure.
workers
For
validity
factor
Confirmatory
Split-half
of
the
factor
step
is
in
analysis
with
by
Internal
consistency
reported.
alpha
that
these scales are
factor
internally
and
consistent
factor
of
(16,
18).
analysis for the full data, a
analysis was measured
In addition, other
first data set with 481 cases
and
type of reliability
to be used in the exploratory
criterion for establishing
such
step (hereafter referred to as
model fit via goodness
retest needs to be
data set EXPLORE), and a
of fit indices statistics
confirmed
reported.
The
as
test-
189
to
Development and Psychometric Evaluation of Scales… / Atashzadeh-Shoorideh et al.
second data set with 472
generally suggest that
assess
cases to be used in the
values around 0.90 are
stability of the
confirmatory
acceptable and values
measures
(hereafter referred to as data
>0.90
time (6).
set
A
considered good fit for
principal
the CFI, GFI and the
analysis
(CONFIRM).
preliminary
higher
are
component analysis (PCA)
NFI (35).
was carried out on the
Values < 0.05 for the
WRQoL
RMSEA indicate a close
set.
EXPLORE
Twelve
data
components
fit
whereas
values
with eigenvalues above 1.0
between 0.05 and 0.10
were generated. Using this
represent
procedure, 34 items were
mediocre fit (36).
removed, leaving 24 items,
which together represented
six factors [Factor 1: Job and
Career
Satisfaction
(JCS)
contained six items, Factor
2:
General
Well-Being
(GWB) also contained six
questions, Factor 3: Home–
Work
Interface
(HWI)
reflected three items, Factor
4: Stress at Work (SAW)
was
represented
by two
items, Factor 5: Control at
Work (CAW): Three items
loaded on component five,
Factor
6:
Conditions
Working
(WCS)
with
three items].
Confirmatory factor analysis
was
conducted
remaining
23
on
the
items
and
support was found for the
model in the CONFIRM
data set (P < 0.01, CFI =
0.93, GFI = 0.90, NFI = 0.89
and RMSEA = 0.06).
190
or
adequate
to
the
over
Journal of Medical Education
Author
Instrument/s
/s
Fall 2015, Vol. 14, No. 4
Type of validity
Type
Criticism of validity
validity
of
Criticism
of
reliability
reliability
Measuring and reporting
The reliability of
The
each
resultant
consistency
was
Cronbach’s
Otieno
An instrument
Content
O.G et
to
addressed by basing the
content
al
measures
items on previous surveys
questionnaire
factor
(2007)
nurses’
use,
and
the
developing is necessary
computed
(37)
quality
and
instrument by a panel of
and important (16, 19).
Cronbach’s alpha
The
satisfaction
nurses
In
coefficient.
coefficients
with
nursing informatics.
Electronic
Factor
Medical
was
reviewing
experienced
analysis,
in
validity
this
reporting
in
study
of
the
content
Criteria
using
were
reported.
alpha
of
0.7 and above
based
study was examined. A cut-
But CVI didn’t report.
Cronbach’s alpha
these scales are
Record (EMR)
off value of 0.4 for factor
Factor
analysis:
coefficients ≥0.7
internally
systems
loadings was applied. Factor
Exploratory
factor
within a construct
consistent
analysis
analysis was measured
and
16).
and
correlation
subscales
in
three
use
of
reported.
It
is
on
alpha
validity is acceptable.
revealed
this
in
Internal
item–total
Electronic Medical Record
acceptable.
within
(EMR) scale. Also factor
Concurrent validity was
subscales.
analysis
revealed
two
measured but the degree
Items
subscales
in
of
of correlation was not
deleted
mentioned.
necessary
‘quality
≥0.4
the
indicate
In
that
addition,
reliability needs
to be confirmed
were
where
to
assess
measures
achieve an alpha
time (6).
value of at least
In
satisfaction’ are determined
0.7.
validity
by factor analysis.
In
(EMR)’
and
three-factor
subscales
in
Concurrent
‘user
validity
this
study,
the
stability of the
to
Electronic Medical Record
(15,
this
over
study,
and
reliability of the
overall
instrument
calculating
Cronbach’s alpha
reported
coefficients
coefficient didn’t
together.
between the scales of the
mention,
recommended
instrument and the global
reported for each
reporting
measure.
Criterion-related
subscale.
Three
validity
validity was not addressed
subscales
with
explicitly
in
low
However,
the
assessed
by
correlation
this
was
study.
degree
of
but
Cronbach’s
were
scores of the two subscales
from
of EMR use (Nursing Care
instrument.
Management
between
and
It
of
reliability will be
separated.
removed
the
is
and
alpha coefficient
the
correlation
it
was
final
Order
Entry); two subscales of
quality of EMR (Information
191
Development and Psychometric Evaluation of Scales… / Atashzadeh-Shoorideh et al.
Quality and Service Quality)
and one subscale of user
satisfaction (Impact of EMR
systems on Clinical Care)
revealed in all cases.
Author
Instrument/s
/s
Type of validity
Type
Criticism of validity
validity
of
Criticism
of
reliability
reliability
the
Cronbach’s alpha
The reliability is
content
was computed to
reported
correctly.
FU
Adapted
Content
M.R.,
Symptom
Symptom
reporting
McDan
Distress scale:
Experience Index (SEI) was
validity is acceptable.
measure
iel
The Symptom
ensured
But CVI didn’t report.
consistency.
stability
R.W. &
Experience
medical-surgical
The validity of this study
Correlation
indicates a high
Rhodes
Index (SEI)
oncology patients in the
isn’t complete.
analysis for the
degree
stability over a
by
of
15
the
general
and
In
this
study,
of
internal
experience
The
of
V.A.
study who had tested the
total
(2007)
reliability and validity of the
revealed
period of time
(38)
Adapted Symptom Distress
Cronbach’s alpha
and satisfactory
Scale version 2 (ASDS-2).
0.91;
for
total
degree
In addition, content validity
occurrence
0.85;
homogeneity (8).
of the SEI is supported by
Construct validity was
for total distress
inclusion of symptoms that
measured
0.84.
have
multiple
been
identified
by
through
comparisons.
Reliability
for the subscales
For
test-retest
patients in other studies as
But, factor analysis can
was
estimated
procedures,
well as those perceived by
be
an
using
Cronbach
second
patients with cancer in a
exploratory
or
alpha
for
administration
series of the investigators’
confirmatory technique
subscale:
generally
studies.
to
respiratory (0.8),
recommended
cognitive
about 2–14 days
Construct
authors
validity:
used
The
multiple
comparisons (with KruskalWallis
test)
construct
to
estimate
validity
by
used
as
estimate
underlying
the
dimensions
each
(0.79),
is
eating/gastrointest
after
items in an instrument.
inal
(39). Because of
(0.73),
the
the
or to reduce redundant
first
pain/discomfort
the attributes of
(0.76),
the
phenomena
determining
statistically
neurological
being measured
significant
differences
(0.78),
(symptom
between pairs of contrasting
fatigue/sleep/restl
occurrence
groups.
essness
distress),
only
healthy
adult
(0.81),
eliminations
(0.74)
192
of
and
and
participants were
appearance (0.77).
asked
To measure the
complete
to
the
Journal of Medical Education
Fall 2015, Vol. 14, No. 4
stability
of
the
Symptom
SEI, a test-retest
Experience
method
Index (SEI)
(during
two
different
during
periods
of
different periods
2–4
two
hours apart) was
of
used
apart. This time
with
healthy
63
adult
2–4
hours
lapse
was
participants.
sufficient
Intra-class
avert
correlation
participants’
coefficients were
recall
calculated
to
previous
test-
response
estimate
retest
reliability.
of
absence of flu
symptoms)
were
to
correlated
total
for
symptom
their
(i.e.
Test-retest scores
strongly
to
preclude
activities
onset
and
(i.e.
of
flu
experience
symptoms after 2
(r=0.93),
weeks) that may
occurrence
have affected the
(r=0.94)
and
distress (r=0.92).
stability of the
characteristic
(symptom
experience)
being measured
(40).
addition, it affects implications of research
findings to the population under study (19).
In this study, review of 12 articles that
published in the Journal of Advanced
Nursing,
2007-2009,
showed
that
psychometric properties did not present, since
from 12 articles only 2 of the articles
documented validity completely, and 5 of the
articles reported incomplete content validity
and 5 of them did not measured it. In regard
to measuring construct validity, factor
analysis is a useful method. From 12 articles
that reviewed, 4 articles measured factor
analysis completely, 4 of them measured or
reported incomplete and 4 of the articles did
not measure it. In regard to other type of
validity, from 12 articles, only one article
measured concurrent validity, one article
measured discriminate validity, one article
measured divergent validity and one article
measured convergent validity.
As stated before, measuring 3 types of
validity for new developed instruments is
necessary. Therefore, measuring validity to
determine the appropriateness of an
instrument should be for a special group. The
193
Development and Psychometric Evaluation of Scales… / Atashzadeh-Shoorideh et al.
findings showed that some of the articles did
not
measure psychometric properties
properly, thus some of the developed scales
need to measure other types of necessary
validity.
In addition, reliability testing needs to be
performed on each instrument used in a study
before other statistical analysis are performed.
From 12 articles, all of the articles measured
and reported Cronbach’s alpha and test-retest,
but 4 of them did not measure test-retest.
Conclusion
It can be concluded that although researchers
put a great emphasis on methodology and
statistical analysis, they pay less attention to
the psychometric properties of their new
instruments. The authors of this article hope
to draw the attention of researcher to the
importance of measuring psychometric
properties of new instruments.
Acknowledgements
The authors would like to thank Dr. Zagheri
Tafreshi for commenting on a draft of this
paper. Her feedback was much appreciated
References
1. Houser J. Nursing Research: Reading, Using,
and Creating Evidence. 2nd Ed. Sudbury, Jones and
Bartlett Publishers; 2011.
2. Watson R, Hugh McK, Seamus C, John K.
Nursing Research: Designs and Methods. Elsevier
Health Sciences; 2008.
3. Yaghmaie F. Subjective computer training:
Development of a scale. Journal of Medical
Education. 2004; 5(1):33-7.
4. Burns N, Grove SK. Practice of Nursing
Research, Conduct, Critique and Utilization. 6th Ed.
Philadelphia: Saunders Co; 2009.
5. Burns N, Grove SK. Understanding Nursing
Research, Building an Evidence-Based Practice. 6th
Ed. Philadelphia: Saunders Co; 2014.
6. Rattray J, Jones MC. Essential elements of
questionnaire design and development. Journal of
Clinical Nursing. 2007; 16(2):234–43.
7. Bush CT. Nursing research. Virginia: Reston
Publishing Co; 1985.
204
8. Polit DF, Hungler BP. Nursing Research
Principles and Methods. 7th Ed. Philadelphia:
Lippincott Co; 2004.
9. Dempsy PA, Dempsy AD. Using Nursing
Research, Process, Critical Evaluation and
Utilization. 5th Ed. Philadelphia: Lippincott Co.;
2000.
10. Kerlinger FN. Foundations of behavioral
research. 3rd Ed. New York: CBS Publishing; 1986.
11. Yaghmaie F. Content validity and its
estimation. Journal of Medical Education. 2003;
3(1):25-7.
12. LoBiondo-Wood G, Haber J. Nursing
Research; Methods, Critical Appraisal, and
Utilization. 8th Ed. St. Louis: Mosby-Elsevier;
2014.
13. Polit DF, Beck CT. Essentials of Nursing
Research, Methods, Appraisal and Utilization. 6th
Ed. Philadelphia: Lippincott Williams & Wilkins;
2006.
14. Nunnally JC, Bernstein IH. Psychometric
Theory. 3rd Ed. New York: McGraw Hill Inc; 1994.
15. Yaghmaie F. Development of a scale for
measuring user computer experience. Journal of
Research Nursing. 2007; 12(2):185-90.
16. Yaghmaei F. Measuring Behavior in Research
by Valid and Reliable Instruments. 2nd Ed. Tehran:
Shahid Beheshti Medical University Publishing;
2009. (Persian)
17. Clark LA, Watson D. Constructing validity:
Basic issues in objective scale development.
Psychological Assessment. 1995; 7(9):309–19.
18. Yaghmaie F. Reliability and its measurement in
quantitative studies. Journal of Faculty of Nursing
& Midwifery, Shahid Beheshti University of
Medical Sciences. 2003; 13(42):22-7. (Persian)
19. Zagheri-Tafreshi M, Yaghmaie F. Factor
analysis of construct validity: A review of nursing
articles. Journal of Medical Education. 2007;
10(1):19-26.
20. Ushiro R. Nurse–Physician Collaboration
Scale: development and psychometric testing.
Journal of Advanced Nursing. 2009; 65(7):1497–
508.
21. Chang HJ, et al. Chinese version of the positive
and negative suicide ideation: Instrument
development. Journal of Advanced Nursing. 2009;
65(7):1485–96.
22. Eizenberg MM, et al. Moral distress
questionnaire for clinical nurses: instrument
development. Journal of Advanced Nursing. 2009;
65(4):885–92.
Journal of Medical Education
23. Munro BH. Statistical Methods for Health Care
Research. 5th edition. Philadelphia: Lippincott
Williams and Wilkins; 2005.
24. Liu M, Yin L, Ma E, Lo S, Zeng L.
Competency Inventory for Registered Nurses in
Macao: instrument validation. Journal of Advanced
Nursing. 2009; 65(4):893–900.
25. Zisberg A, Young HM, Schepp K.
Development and psychometric testing of the Scale
of Older Adults’ Routine. Journal of Advanced
Nursing. 2009; 65(3):672–83.
26. Woods-Dauphinee S, Berg K, Daley K.
Monitoring status and evaluating outcomes: An
overview of rating scales for the use with patients
who have sustained a stroke. Topics in Geriatric
Rehabilitation. 1994; 10(2):22–41.
27. Cyr L, Francis K. Measures of clinical for
nominal and categorical data: The kappa
coefficient. Computers in Biology and Medicine.
1992; 22(4):239-46.
28. Pelander T, Leino-Kilpi H, Katajisto J. The
quality of pediatric nursing care: developing the
Child Care Quality at Hospital instrument for
children. Journal of Advanced Nursing. 2009;
65(2):443–53.
29. Stotts NA, Aldrich KM. How to try this defines
your terms, Evaluating instruments for use in
nursing practice. Advanced Journal of Nursing.
2007; 107(10):71-2.
30. Ferketich S. Focus on psychometrics: Aspects
of item analysis. Research in Nursing & Health.
1991; 14(2); 165–8.
31. Carlson C. Development and testing of four
instruments to assess prior conditions that influence
nurses’ adoption of evidence-based pain
management practices. Journal of Advanced
Nursing. 2008; 64(6):632–43.
Fall 2015, Vol. 14, No. 4
32. Pisanti R, Lombardo C, Lucidi F, Lazzari D,
Bertini M. Development and validation of a brief
Occupational Coping Self-Efficacy Questionnaire
for Nurses. Journal of Advanced Nursing. 2008;
62(2):238–47.
33. Barnes CR, Adamson-Macedo EN. Perceived
Maternal Parenting Self-Efficacy (PMP S-E)
instrument: development and validation with
mothers of hospitalized preterm neonates. Journal
of Advanced Nursing. 2007; 60(5):550–60.
34. Van Laar D, Edwards J A, Easton S. The Workrelated quality of life scale for healthcare workers.
Journal of Advanced Nursing. 2007; 60(3):325–33.
35. Bentler PM, Bonett DG. Significance tests and
goodness of fit in the analysis of covariance
structures.
Psychological
Bulletin.
1980;
88(3):588–606.
36. Browne MW, Cudeck R. Alternative Ways of
Assessing Model Fit in Testing Structural Equation
Models. Sage, Newbury Park, CA, USA; 1993.
37. Otieno OG, Toyama H, Asonuma M. Nurses’
views on the use, quality and user satisfaction with
electronic
medical
records:
questionnaire
development. Journal of Advanced Nursing. 2007;
60(2):209–19.
38. Fu M, McDaniel RW, Rhodes VA. Measuring
symptom occurrence and symptom distress:
development of the symptom experience index.
Journal of Advanced Nursing. 2007; 59(6):623–34.
39. Streiner DL, Norman GR. Health Measurement
Scales: A Practical Guide to Their Development
and Use. 5th edition. Oxford University Press Inc.:
New York; 2015.
40. Fu MR, Rhodes VA, Xu B. The Chinese
translation: The index of nausea, vomiting, and
retching (INVR). Cancer Nursing. 2002;
25(2):134–40.
205