SlideShare a Scribd company logo
1 of 37
Download to read offline
Validation Studies in Simulation-based
Education
April 15, 2015
Deb Rooney, PhD
Professor of Learning Health Sciences
All Rights Reserved.
1
•  Validity in the current framework used to evaluate
evidence
•  How we gather /evaluate validity evidence from a
simulator and it’s associated measures
•  Context of academic product (manuscripts)
•  Final considerations
Objectives
2
A few definitions to consider….
Validity: What is it?
2. the degree to which evidence and
theory support the interpretations of test
scores as entailed by proposed uses of
tests –Standards (1999)
1. the degree to which the tool measures
what it claims to measure
American Educational Research Association, American Psychological Association, & National Council on Measurement in Education.
(1999) Standards for educational and psychological testing. Washington, DC: American Educational Research Association. 3
•  Evidence relevant to relationships to other variables (e.g. novice
versus expert discriminant validity) is over-represented
•  Evidence relevant to response process and consequences of
testing and are infrequently reported
Cook, D. A., Brydges, R., Zendejas, B., Hamstra, S. J., & Hatala, R. (2013). Technology-enhanced simulation
to assess health professionals: A systematic review of validity evidence, research methods, and reporting
quality. Academic Medicine. Jun;88(6):872-83.
•  Apply current Standards to ensure rigorous research and
reporting
Simulator Validation:
The Framework
4
Simulator Validation:
The Evidence
Current AERA Standards*
o  Not new/novel
o  Unitary Construct (All evidence falls under “Construct” validity)
o  Test content
o  Internal structure
o  Response Processes
o  Relationships to other variables
o  Consequences of testing
(psychometric properties of measures)
(comparison w/ previously-validated measures)
(standard setting, rater/ings quality, fidelity vs. stakes)
(reliability, dimensionality, function across groups)
(face validity, subjective measures, construct alignment)
*Standards for Educational and Psychological Testing (AERA, APA, & NCME, 2014)
5
Validity evidence;
Validity: What is it NOT?
•  Does not allow us to make inferences about a
curriculum
•  Does not allow us to make inferences about different
applications, settings, or learners
•  Is not a terminal quality determination
(about the quality of your measures, application)
•  “The scale was valid” ! “evidence supports the use
of a scale to measure X in a particular setting/
application” 6
We have a much more complex
environment to evaluate!
Simulator
test content
relationships to other variables
consequences of testing
Instrument (measures)
test content
internal structure
relationships to other variables
response processes
consequences of testing
Simulator Validation:
How does this evidence apply to us?
7
CREATIONDESIGN &
PLANNING
IMPLEMENTATION &
EVALUATION
Test content (measures/simulator)
Internal structure (m)
Response processes
(m)
Relationship to other
variables (m/s)
Consequences of testing (m/s)
Validity Evidence in Simulation
How/when do we gather evidence?
8
PAPER 2PAPER 1 PAPER 3
Test content
(measures &
simulator)
Before
implementation
Validity Evidence in Simulation
How disseminate findings?
Quality of
performance
measures
Before or After
Implementation
Impact on
performance and/
or patient
outcomes
After full
implementation
9
Most Recent Example:
Neurosurgery Sim
a Tai B, Rooney D Stephenson F, Liao P, Sagher O, Shih A, Savastano LE. Development of 3D-
printing built ventriculostomy placement simulator, Journal of Neurosurgery, (in press)
bRooney DM, Tai BL, Sagher O, Shih AJ, Wilkinson DA, Savastano LE. A Simulator and two
tools: Validation of performance measures from a novel neurosurgery simulator using the
current Standards framework. Surgery, submitted 3/15.
Paper 1a: Preliminary evaluation of
quality of simulator/ measures
Paper 2b: Evaluation of performance
measures from simulator
Paper 3: Evaluation of Impact on
performance measures / patient
outcomes
10
Paper 1: Simulator validation process
n=7 n=5 n=5
4 months
11
The Content Validity Form: 5 domains
•  Physical
Attributes
•  Realism-
experience
•  Value
•  Relevance
•  Overall (global)
12
Paper 1: The preliminary validation process
(Sim)*
•  Using Rasch model, analyzed data for;
•  Looked at domain rating differences across 3 sites
•  Mean averages by item
•  Looked at Rasch variability indices to identify possible
inconsistency in ratings
•  Ensured psychometric quality of survey
•  Using traditional methods, estimated;
•  Inter-item consistency, Cronbach alpha
•  Inter-rater agreement, ICC(2,k)
•  Using a Rasch model ensured rating scales’ function
*performance checklist is separate/ different process
13
Results: Domain mean averages by site
0	
  
0.5	
  
1	
  
1.5	
  
2	
  
2.5	
  
3	
  
3.5	
  
4	
  
4.5	
  
UM	
  
HF	
  
WS	
  
3.4 3.3 3.9 3.3 2.4
Combined mean averages
“This simulator requires minor adjustments before it can be considered for use in
ventriculostomy placement training.”
14
Results: Mean averages by item
15
Paper 1*: Test Content-Checklist
*shoulda, coulda, woulda
16
Proposed items Definitely do not
include this task
(1)
Not sure if this
task should be
included
(2)
Pretty sure this
task should be
included
(3)
Definitely include
this task
(4)
Position head and mark
midline
Locate Kocher's point
(10.5 cm posterior to the
nasion and 3 cm lateral to
midline)
Mark incision (approximately
2cm long in a parasagittal
location)
Incise, clear tissue off
cranium, retracted scalp
…
Suture wound (Staples or a
3-0 running nylon or prolene
suture)
Paper 1*:
Test Content-Checklist
17
Paper 1*:
Test Content-Checklist
Ask expert instructors about the value of included steps (items) for
measuring X at doing Y.
•  Reasonable number of experts ~ 3
What else do you ask about?
•  Clarity of item
•  Appropriateness of qualifiers (use X instrument, at x location)
•  Rating scale
•  Missing steps
•  Objective measures to include (eg. time to)
*shoulda, coulda, woulda
18
PAPER 2PAPER 1 PAPER 3
Before
implementation
Test content
(measures &
simulator)
Before or After
Implementation*
Quality of
performance
measures
After full
implementation
Impact on
performance and/or
patient outcomes
Next Step: Deeper Evaluation (Paper 2)
19
Next Step: Deeper Evaluation (Paper 2)
•  Evaluation of all validity evidence of performance
measures [ala Standards]
•  Capture broader (regional/national) sample of
performance data via videotaped performances
•  ideal N (+++) and range of experience
•  Compare measures from the novel performance
checklist and gold standard (eg-OSATS] (relationship
with other variables*)
•  Set/test performance standards (if appropriate)
Rooney DM, Tai BL, Sagher O, Shih A, Wilkinson DA, Savastano, L. A Simulator and two tools: Validation of
performance measures from a novel neurosurgery simulator using the current Standards framework. Surgery,
submitted 3/15)
20
•  Nationally-recognized training
program sponsored by Society
of Neurological Surgeons
•  Total n=14 (11 trainees*, 3
attendings) performed
ventriculostomy on simulator
•  All performances were video-
captured, scored by 3 raters
using novel checklist and
modified version of OSATS
Paper 2: study design
*first year neurosurgery fellows 21
Checklist: Ventriculostomy Procedural
Assessment Tool (V-PAT)
22
Martin JA, Regehr G, Reznick R, MacRae H, Murnaghan J, Hutchison C, et al. (1997). Objective structured assessment of technical skill
(OSATS) for surgical residents. Br J Surg 84:273-278.
Modified Objective Structured
Assessment of Technical Skills:
m-OSATS
23
Measures adequately reflect ventriculostomy performance “quality”
•  Relationships to other variables
•  Trainee v expert ratings
•  Correlation of summed V-PAT with summed OSATS scores
Measures are psychometrically sound (Adequate “Quality Control”)
•  Psychometric function of V-PAT & OSATS measures;
•  Response processes: Rasch indices ! rating scale function
•  Test Content: Rasch item point-measure correlations, item fit (variability)
•  Internal structure: Inter-item consistency-Cα, Inter-rater agreement- ICC(2,k)
Measures are free from rater bias
•  Consequences of testing
•  Evaluated Rasch bias indices to identify potential rating differences at rater level
Paper 2: evidence examined
Examined evidence from 5 sources, but packaged a bit differently;
24
Response processes
•  Rasch indices: (Avg meas., Fit statistics, and RA
thresholds) indicated all rating scales for both V-PAT &
OSATS were well-functioning
Test content
Point-measure correlations: all positive, [.39, .81]
Rasch item Outfit MS: all < 2.0
Internal structure
•  Cronbach α: both α = 0.95
•  Intraclass correlation: V-PAT= [-0.33, 0.93]
OSATS = [0.80, 0.93]
Results: “Quality control”
25
Do V-PAT measures adequately differentiate trainee and
expert performances?
	
  Instrument Indices
V-PAT
Resident
Observed
Average
(SE)
Attending
Observed
Average
(SE)
P-
value
ICC
(2,k)
Value
1. Position head and mark midline 3.75 (.20) 4.00 (.47) 0.54 *
2. Locate Kocher's point 3.84 (.18) 4.11 (.34) 0.52 .86
3. Mark an incision 3.61 (.16) 4.00 (.32) 0.42 .10
4. Select drain exit site from the scalp 2.94 (.17) 3.71 (.50) 0.22 *
5. Incise, clear tissue off cranium, retract
scalp 3.56 (.14) 4.11 (.34) 0.20 .83
6. Set drill stop and drill trephine 2.80 (.16) 3.78 (.37) 0.08 .70
7. Confirm dura and pierce 2.98 (.17) 3.67 (.35) 0.25 .76
8. Confirm landmarks and place
catheter 2.88 (.17) 3.33 (.35) 0.24 .91
9. Confirm CSF flow 3.41 (.13) 3.67 (.30) 0.39 .73
10. Remove trocar cover, tunnel trocar
to exit site and recap trocar 2.78 (.14) 3.00 (.47) 0.62 -.33
11. Place purse string suture to anchor
the catheter at scalp exit site 3.00 (.36) 4.25 (.69) 0.26 *
Overall average 3.30 (.06) 3.80 (.11) 0.01 –
	
  Instrument Indices
V-PAT
Resident
Observed
Average
(SE)
Attending
Observed
Average
(SE)
P-
value
ICC
(2,k)
Value
1. Position head and mark midline 3.75 (.20) 4.00 (.47) 0.54 *
2. Locate Kocher's point 3.84 (.18) 4.11 (.34) 0.52 .86
3. Mark an incision 3.61 (.16) 4.00 (.32) 0.42 .10
4. Select drain exit site from the scalp 2.94 (.17) 3.71 (.50) 0.22 *
5. Incise, clear tissue off cranium, retract
scalp 3.56 (.14) 4.11 (.34) 0.20 .83
6. Set drill stop and drill trephine 2.80 (.16) 3.78 (.37) 0.08 .70
7. Confirm dura and pierce 2.98 (.17) 3.67 (.35) 0.25 .76
8. Confirm landmarks and place
catheter 2.88 (.17) 3.33 (.35) 0.24 .91
9. Confirm CSF flow 3.41 (.13) 3.67 (.30) 0.39 .73
10. Remove trocar cover, tunnel trocar
to exit site and recap trocar 2.78 (.14) 3.00 (.47) 0.62 -.33
11. Place purse string suture to anchor
the catheter at scalp exit site 3.00 (.36) 4.25 (.69) 0.26 *
Overall average 3.30 (.06) 3.80 (.11) 0.01 –
Results: Relationship to Other
variables (V-PAT)
	
  Instrument Indices
V-PAT
Resident
Observed
Average
(SE)
Attending
Observed
Average
(SE)
P-
value
ICC
(2,k)
Value
1. Position head and mark midline 3.75 (.20) 4.00 (.47) 0.54 *
2. Locate Kocher's point 3.84 (.18) 4.11 (.34) 0.52 .86
3. Mark an incision 3.61 (.16) 4.00 (.32) 0.42 .10
4. Select drain exit site from the scalp 2.94 (.17) 3.71 (.50) 0.22 *
5. Incise, clear tissue off cranium, retract
scalp 3.56 (.14) 4.11 (.34) 0.20 .83
6. Set drill stop and drill trephine 2.80 (.16) 3.78 (.37) 0.08 .70
7. Confirm dura and pierce 2.98 (.17) 3.67 (.35) 0.25 .76
8. Confirm landmarks and place
catheter 2.88 (.17) 3.33 (.35) 0.24 .91
9. Confirm CSF flow 3.41 (.13) 3.67 (.30) 0.39 .73
10. Remove trocar cover, tunnel trocar
to exit site and recap trocar 2.78 (.14) 3.00 (.47) 0.62 -.33
11. Place purse string suture to anchor
the catheter at scalp exit site 3.00 (.36) 4.25 (.69) 0.26 *
Overall average 3.30 (.06) 3.80 (.11) 0.01 –
26
Do m-OSATS measures adequately differentiate trainee
and expert performances?
•  Correlation of summed V-PAT scores with summed
m-OSATS
•  Pearson’s r = 0.72, p = 0.001
V-PAT
Resident
Observed
Average
(SE)
Attending
Observed
Average
(SE)
P-
value
ICC
(2,k)
Value
1. Position head and mark midline 3.75 (.20) 4.00 (.47) 0.54 *
2. Locate Kocher's point 3.84 (.18) 4.11 (.34) 0.52 .86
3. Mark an incision approximately 2 cm
long in a parasagittal location 3.61 (.16) 4.00 (.32) 0.42 .10
4. Select drain exit site from the scalp 2.94 (.17) 3.71 (.50) 0.22 *
5. Incise, clear tissue off cranium, retract
scalp 3.56 (.14) 4.11 (.34) 0.20 .83
6. Set drill stop and drill trephine 2.80 (.16) 3.78 (.37) 0.08 .70
7. Confirm dura and pierced with 18g
spinal needle or 11 blade scalpel 2.98 (.17) 3.67 (.35) 0.25 .76
8. Confirm landmarks and place
catheter to 6-7cm from outer table of
skull 2.88 (.17) 3.33 (.35) 0.24 .91
9. Confirm CSF flow 3.41 (.13) 3.67 (.30) 0.39 .73
10. Remove trocar cover, tunnel trocar
to exit site and recap trocar 2.78 (.14) 3.00 (.47) 0.62 -.33
11. Place purse string suture at the
scalp exit site to anchor the catheter 3.00 (.36) 4.25 (.69) 0.26 *
Overall average 3.30 (.06) 3.80 (.11) 0.01 –
OSATS
1. Respect for Tissue 2.66 (.21) 4.11 (.34) 0.004 .93
2. Time and Motion 2.42 (.22) 4.00 (.43) 0.005 .85
3. Instrument Handling 2.51 (.22) 4.00 (.46) 0.007 .86
4. Knowledge of Instruments 2.36 (.21) 4.33 (.43) 0.001 .84
5. Flow of Operation 2.36 (.21) 4.22 (.45) 0.001 .85
6. Knowledge of specific procedure 2.33 (.23) 4.33 (.46) 0.001 .80
Overall average 2.32 (.08) 3.73 (.15) 0.001 –
* too few cases to estimate
	
   	
   	
   	
   	
   	
  
	
  Instrument Indices
V-PAT
Resident
Observed
Average
(SE)
Attending
Observed
Average
(SE)
P-
value
ICC
(2,k)
Value
1. Position head and mark midline 3.75 (.20) 4.00 (.47) 0.54 *
2. Locate Kocher's point 3.84 (.18) 4.11 (.34) 0.52 .86
3. Mark an incision approximately 2 cm
long in a parasagittal location 3.61 (.16) 4.00 (.32) 0.42 .10
4. Select drain exit site from the scalp 2.94 (.17) 3.71 (.50) 0.22 *
5. Incise, clear tissue off cranium, retract
scalp 3.56 (.14) 4.11 (.34) 0.20 .83
6. Set drill stop and drill trephine 2.80 (.16) 3.78 (.37) 0.08 .70
7. Confirm dura and pierced with 18g
spinal needle or 11 blade scalpel 2.98 (.17) 3.67 (.35) 0.25 .76
8. Confirm landmarks and place
catheter to 6-7cm from outer table of
skull 2.88 (.17) 3.33 (.35) 0.24 .91
9. Confirm CSF flow 3.41 (.13) 3.67 (.30) 0.39 .73
10. Remove trocar cover, tunnel trocar
to exit site and recap trocar 2.78 (.14) 3.00 (.47) 0.62 -.33
11. Place purse string suture at the
OSATS
Results: Relationship to Other
variables (m-OSATS)
27
Are measures free from rater bias?
Results:
Consequences of Testing
V-PAT
Observed Average
(SE)
1. Rater 1 (LS) 3.60 (.11)
2. Rater 2 (OS) 3.60 (.10)
3. Rater 3 (DW) 3.60 (.11)
4. Participants (Variable)† 2.90 (.11)
Overall average 3.60 ( – )
OSATS
1. Rater 1 (LS)* 2.10 (.17)
2. Rater 2 (OS) 3.30 (.15)
3. Rater 3 (DW) 3.00 (.15)
Overall average 2.80 (–)
†Comparison with 3 expert raters, p = 0.01
* Comparison with 2 expert raters, p = 0.01
V-PAT
Observed Average
(SE)
1. Rater 1 (LS) 3.60 (.11)
2. Rater 2 (OS) 3.60 (.10)
3. Rater 3 (DW) 3.60 (.11)
4. Participants (Variable)† 2.90 (.11)
Overall average 3.60 ( – )
OSATS
1. Rater 1 (LS)* 2.10 (.17)
2. Rater 2 (OS) 3.30 (.15)
3. Rater 3 (DW) 3.00 (.15)
Overall average 2.80 (–)
†Comparison with 3 expert raters, p = 0.01
* Comparison with 2 expert raters, p = 0.01
V-PAT
Observed Average
(SE)
1. Rater 1 (LS) 3.60 (.11)
2. Rater 2 (OS) 3.60 (.10)
3. Rater 3 (DW) 3.60 (.11)
4. Participants (Variable)† 2.90 (.11)
Overall average 3.60 ( – )
OSATS
1. Rater 1 (LS)* 2.10 (.17)
2. Rater 2 (OS) 3.30 (.15)
3. Rater 3 (DW) 3.00 (.15)
Overall average 2.80 (–)
†Comparison with 3 expert raters, p = 0.01
* Comparison with 2 expert raters, p = 0.01
28
Are OSATS measures free from rater bias?
0
0.5
1
1.5
2
2.5
3
3.5
4
4.5
5
Q1 Q2 Q3 Q4 Q5 Q6
Observedaverage
OSATS item
Bias interaction: Rater by item
Rater1-LS
Rater2-OS
Rater3-DW
Results:
Consequences of Testing
29
Summary of Results
Evidence Inferences V-PAT OSATS
Quality Control
Response Processes Adequate rating scale function √ √
Test Content Items align with construct √ √
Internal Structure Inter-item consistency, inter-rater
agreement*
X* √
Test of Assumptions
Rel. Other Variables Measures dif. bet. N/E performances X √
Rel. Other Variables V-PAT/OSATS summed scores correlate √ √
Cons. of Testing V-PAT/OSATS measures are bias free √ X
X = challenges that require resolution 30
Standards for educational and psychological testing. American Educational Research Association, American Psychological Association, &
National Council on Measurement in Education. (2014). Washington, DC: American Educational Research.
Available for purchase via http://teststandards.org/
Validity: ala 2015
•  Evidence is important, but interpretive argument is critical
•  Content of interpretive argument determines the kinds of
evidence that are most relevant (most important) in validation
•  Strategy of developing interpretive argument based on
•  Validity evidence relevant to inferences
•  Assumptions
•  Challenges (alternative interpretations)
31
Problematic inter-rater agreement (ICC) for 5 items
should be resolved;
•  Item 3 (Mark an incision approximately 2 cm long in a
parasagittal location), ICC= .10
•  Item 10 (Remove trocar cover, tunnel trocar to exit site
and recap trocar), ICC= -.33
Examine, refine items to ensure
alignment with simulator capabilities/ are
mutually exclusive
Challenges:
potential threats to validity (V-PAT)
•  Item 1 (Position head and mark midline)*
•  Item 4 (Select drain exit site from the scalp)*
•  Item 1 (Place purse string suture at the scalp exit site to
anchor the catheter) *
* ICC incalculable
32
“Hawkish” OSATS ratings by one expert rater
requires follow-up
Refine items, add rater training on scoring
rubric and administration standards
Challenges: potential threats to
validity (m-OSATS)
33
PAPER 2PAPER 1 PAPER 3
Before
implementation
Test content
(measures &
simulator)
Before or After
Implementation*
Quality of
performance
measures
After full
implementation
Impact on
performance and/or
patient outcomes
Next Step: Evaluation of Impact
(Paper 3)
34
•  Evaluation of impact on trainee’s clinical performance
or patient outcomes [relationship with other
variables]
•  Examine;
•  Change in trainees’ clinical performance (checklist
ratings, objective measures (“time to”, LOS, adverse
events)
•  Impact on hospital costs
Barsuk JH, Cohen ER, Feinglass J, Kozmic SE, McGaghie WC, Ganger D, Wayne DB.
Cost savings of performing paracentesis procedures at the bedside after simulation-
based education. Simul Healthc. 2014 Oct;9(5):312-8.
Next Step: Evaluation of Impact
(Paper 3)
35
•  Validation process is fluid/reiterative/on-
going
•  It takes a team;
•  Development (clinicians, instructors,
engineers, researchers assistants)!
•  Outcomes (+QI, hospital info)
•  There is funding;
•  AHRQ
•  P-CORi
•  Michigan Blue Cross Blue Shield
Considerations
36
Thank you
Questions?
	
  
	
  
Deborah	
  Rooney,	
  PhD	
  
dmrooney@med.umich.edu	
  
	
  
	
  

More Related Content

What's hot

Experimental Design | Statistics
Experimental Design | StatisticsExperimental Design | Statistics
Experimental Design | StatisticsTransweb Global Inc
 
Experimental design version 4.3
Experimental design version 4.3Experimental design version 4.3
Experimental design version 4.3jschmied
 
Research Methods in Psychology Sampling and Experimental Design
Research Methods in Psychology Sampling and Experimental DesignResearch Methods in Psychology Sampling and Experimental Design
Research Methods in Psychology Sampling and Experimental DesignCrystal Delosa
 
What is an experimental research (1)
What is an experimental research (1)What is an experimental research (1)
What is an experimental research (1)M Usama Sehgal
 
Experimental research
Experimental researchExperimental research
Experimental researchsadia ahmad
 
Principles of design of experiments (doe)20 5-2014
Principles of  design of experiments (doe)20 5-2014Principles of  design of experiments (doe)20 5-2014
Principles of design of experiments (doe)20 5-2014Awad Albalwi
 
Types of experimental design
Types of experimental designTypes of experimental design
Types of experimental designDr-Jitendra Patel
 
Research Design Slide Show
Research Design Slide Show Research Design Slide Show
Research Design Slide Show Myla German
 
Experimental design techniques- why experimental design needed?
Experimental design techniques- why experimental design needed?Experimental design techniques- why experimental design needed?
Experimental design techniques- why experimental design needed?Dr-Jitendra Patel
 
Multivariate Variate Techniques
Multivariate Variate TechniquesMultivariate Variate Techniques
Multivariate Variate TechniquesDr. Keerti Jain
 
Experimental research
Experimental researchExperimental research
Experimental researchGul Shair
 
Causal design & research
Causal design & researchCausal design & research
Causal design & researchgadiabinit
 
Statistics design
Statistics designStatistics design
Statistics designAhmed Taha
 
Experimental and statistical designs
Experimental and statistical designsExperimental and statistical designs
Experimental and statistical designsAudrey Antee
 
Experimental research
Experimental researchExperimental research
Experimental researchizzajalil
 
Scientific Method and Experimental Design
Scientific Method and Experimental DesignScientific Method and Experimental Design
Scientific Method and Experimental Designcgales
 
Casual research design experimentation
Casual research design experimentationCasual research design experimentation
Casual research design experimentationRohit Kumar
 

What's hot (20)

Experimental Design | Statistics
Experimental Design | StatisticsExperimental Design | Statistics
Experimental Design | Statistics
 
Experimental design version 4.3
Experimental design version 4.3Experimental design version 4.3
Experimental design version 4.3
 
Research Methods in Psychology Sampling and Experimental Design
Research Methods in Psychology Sampling and Experimental DesignResearch Methods in Psychology Sampling and Experimental Design
Research Methods in Psychology Sampling and Experimental Design
 
What is an experimental research (1)
What is an experimental research (1)What is an experimental research (1)
What is an experimental research (1)
 
Experimental research
Experimental researchExperimental research
Experimental research
 
Principles of design of experiments (doe)20 5-2014
Principles of  design of experiments (doe)20 5-2014Principles of  design of experiments (doe)20 5-2014
Principles of design of experiments (doe)20 5-2014
 
Types of experimental design
Types of experimental designTypes of experimental design
Types of experimental design
 
Research Design Slide Show
Research Design Slide Show Research Design Slide Show
Research Design Slide Show
 
Experimental design
Experimental designExperimental design
Experimental design
 
Experimental design techniques- why experimental design needed?
Experimental design techniques- why experimental design needed?Experimental design techniques- why experimental design needed?
Experimental design techniques- why experimental design needed?
 
Multivariate Variate Techniques
Multivariate Variate TechniquesMultivariate Variate Techniques
Multivariate Variate Techniques
 
Experimental research
Experimental researchExperimental research
Experimental research
 
Causal design & research
Causal design & researchCausal design & research
Causal design & research
 
Statistics design
Statistics designStatistics design
Statistics design
 
Experimental and statistical designs
Experimental and statistical designsExperimental and statistical designs
Experimental and statistical designs
 
Causal Research
Causal ResearchCausal Research
Causal Research
 
Experimental research
Experimental researchExperimental research
Experimental research
 
Scientific Method and Experimental Design
Scientific Method and Experimental DesignScientific Method and Experimental Design
Scientific Method and Experimental Design
 
Quantitative research present
Quantitative research   presentQuantitative research   present
Quantitative research present
 
Casual research design experimentation
Casual research design experimentationCasual research design experimentation
Casual research design experimentation
 

Similar to Validation Studies in Simulation-based Education - Deb Rooney

Psychometrics for Clinical Skills Assessment
Psychometrics for Clinical Skills AssessmentPsychometrics for Clinical Skills Assessment
Psychometrics for Clinical Skills AssessmentINSPIRE_Network
 
ASC Model: A Process Model for the Evaluation of Simulated Field Exercises in...
ASC Model: A Process Model for the Evaluation of Simulated Field Exercises in...ASC Model: A Process Model for the Evaluation of Simulated Field Exercises in...
ASC Model: A Process Model for the Evaluation of Simulated Field Exercises in...streamspotter
 
NIHR Complex Reviews Support Unit (CRSU) - An Introduction
NIHR Complex Reviews Support Unit (CRSU) - An IntroductionNIHR Complex Reviews Support Unit (CRSU) - An Introduction
NIHR Complex Reviews Support Unit (CRSU) - An IntroductionHEHTAslides
 
Systematic Review & Meta Analysis.pptx
Systematic Review & Meta Analysis.pptxSystematic Review & Meta Analysis.pptx
Systematic Review & Meta Analysis.pptxDr. Anik Chakraborty
 
Software-Testing-Chapgdgdgsghshshshshshshs
Software-Testing-ChapgdgdgsghshshshshshshsSoftware-Testing-Chapgdgdgsghshshshshshshs
Software-Testing-Chapgdgdgsghshshshshshshsshaikbab
 
Ag Extn.504 :- RESEARCH METHODS IN BEHAVIOURAL SCIENCE
Ag Extn.504 :-  RESEARCH METHODS IN BEHAVIOURAL SCIENCE  Ag Extn.504 :-  RESEARCH METHODS IN BEHAVIOURAL SCIENCE
Ag Extn.504 :- RESEARCH METHODS IN BEHAVIOURAL SCIENCE Pradip Limbani
 
Measurement Systems Analysis - Variable Gage R&R Study Metrics, Applications ...
Measurement Systems Analysis - Variable Gage R&R Study Metrics, Applications ...Measurement Systems Analysis - Variable Gage R&R Study Metrics, Applications ...
Measurement Systems Analysis - Variable Gage R&R Study Metrics, Applications ...Gabor Szabo, CQE
 
TCI in general pracice - reliability (2006)
TCI in general pracice - reliability (2006)TCI in general pracice - reliability (2006)
TCI in general pracice - reliability (2006)Evangelos Kontopantelis
 
Analytical procedures life cycle management
Analytical procedures life cycle managementAnalytical procedures life cycle management
Analytical procedures life cycle managementChandra Prakash Singh
 
IA details for IBDP Biology teachers and students
IA details for IBDP Biology teachers and studentsIA details for IBDP Biology teachers and students
IA details for IBDP Biology teachers and studentsRawda Eada
 
Planning of experiment in industrial research
Planning of experiment in industrial researchPlanning of experiment in industrial research
Planning of experiment in industrial researchpbbharate
 
chapter-no-4-test-management fudhg ddh j
chapter-no-4-test-management fudhg ddh jchapter-no-4-test-management fudhg ddh j
chapter-no-4-test-management fudhg ddh jAmitDeshai
 
Feature selection for classification
Feature selection for classificationFeature selection for classification
Feature selection for classificationefcastillo744
 

Similar to Validation Studies in Simulation-based Education - Deb Rooney (20)

research_paper_acm_1.6
research_paper_acm_1.6research_paper_acm_1.6
research_paper_acm_1.6
 
Chapter 3
Chapter 3Chapter 3
Chapter 3
 
Psychometrics for Clinical Skills Assessment
Psychometrics for Clinical Skills AssessmentPsychometrics for Clinical Skills Assessment
Psychometrics for Clinical Skills Assessment
 
Chapter Eight Quantitative Methods
Chapter Eight Quantitative MethodsChapter Eight Quantitative Methods
Chapter Eight Quantitative Methods
 
ASC Model: A Process Model for the Evaluation of Simulated Field Exercises in...
ASC Model: A Process Model for the Evaluation of Simulated Field Exercises in...ASC Model: A Process Model for the Evaluation of Simulated Field Exercises in...
ASC Model: A Process Model for the Evaluation of Simulated Field Exercises in...
 
NIHR Complex Reviews Support Unit (CRSU) - An Introduction
NIHR Complex Reviews Support Unit (CRSU) - An IntroductionNIHR Complex Reviews Support Unit (CRSU) - An Introduction
NIHR Complex Reviews Support Unit (CRSU) - An Introduction
 
Testing Process
Testing ProcessTesting Process
Testing Process
 
Systematic Review & Meta Analysis.pptx
Systematic Review & Meta Analysis.pptxSystematic Review & Meta Analysis.pptx
Systematic Review & Meta Analysis.pptx
 
Software-Testing-Chapgdgdgsghshshshshshshs
Software-Testing-ChapgdgdgsghshshshshshshsSoftware-Testing-Chapgdgdgsghshshshshshshs
Software-Testing-Chapgdgdgsghshshshshshshs
 
Ag Extn.504 :- RESEARCH METHODS IN BEHAVIOURAL SCIENCE
Ag Extn.504 :-  RESEARCH METHODS IN BEHAVIOURAL SCIENCE  Ag Extn.504 :-  RESEARCH METHODS IN BEHAVIOURAL SCIENCE
Ag Extn.504 :- RESEARCH METHODS IN BEHAVIOURAL SCIENCE
 
TCI in primary care - SEM (2006)
TCI in primary care - SEM (2006)TCI in primary care - SEM (2006)
TCI in primary care - SEM (2006)
 
Measurement Systems Analysis - Variable Gage R&R Study Metrics, Applications ...
Measurement Systems Analysis - Variable Gage R&R Study Metrics, Applications ...Measurement Systems Analysis - Variable Gage R&R Study Metrics, Applications ...
Measurement Systems Analysis - Variable Gage R&R Study Metrics, Applications ...
 
TCI in general pracice - reliability (2006)
TCI in general pracice - reliability (2006)TCI in general pracice - reliability (2006)
TCI in general pracice - reliability (2006)
 
Analytical procedures life cycle management
Analytical procedures life cycle managementAnalytical procedures life cycle management
Analytical procedures life cycle management
 
Qc and qa
Qc and qaQc and qa
Qc and qa
 
IA details for IBDP Biology teachers and students
IA details for IBDP Biology teachers and studentsIA details for IBDP Biology teachers and students
IA details for IBDP Biology teachers and students
 
Planning of experiment in industrial research
Planning of experiment in industrial researchPlanning of experiment in industrial research
Planning of experiment in industrial research
 
chapter-no-4-test-management fudhg ddh j
chapter-no-4-test-management fudhg ddh jchapter-no-4-test-management fudhg ddh j
chapter-no-4-test-management fudhg ddh j
 
Feature selection for classification
Feature selection for classificationFeature selection for classification
Feature selection for classification
 
ISTQB Test Process
ISTQB Test ProcessISTQB Test Process
ISTQB Test Process
 

More from Department of Learning Health Sciences, University of Michigan Medical School

More from Department of Learning Health Sciences, University of Michigan Medical School (20)

The Missing Piece? Understanding Provider Organization Capabilities to Engage...
The Missing Piece? Understanding Provider Organization Capabilities to Engage...The Missing Piece? Understanding Provider Organization Capabilities to Engage...
The Missing Piece? Understanding Provider Organization Capabilities to Engage...
 
LHS Collaboratory March 14, 2017 Welcome and Intro Slides
LHS Collaboratory March 14, 2017 Welcome and Intro SlidesLHS Collaboratory March 14, 2017 Welcome and Intro Slides
LHS Collaboratory March 14, 2017 Welcome and Intro Slides
 
University of Michigan Learning Health System Collaboratory
University of Michigan Learning Health System CollaboratoryUniversity of Michigan Learning Health System Collaboratory
University of Michigan Learning Health System Collaboratory
 
Mobile Health Technologies in Cardiovascular Disease
Mobile Health Technologies in Cardiovascular DiseaseMobile Health Technologies in Cardiovascular Disease
Mobile Health Technologies in Cardiovascular Disease
 
Michigan Surgical Quality Collaborative Presentation
Michigan Surgical Quality Collaborative PresentationMichigan Surgical Quality Collaborative Presentation
Michigan Surgical Quality Collaborative Presentation
 
Hunt for Collaboration: Realizing a Learning Health System Together to Democr...
Hunt for Collaboration: Realizing a Learning Health System Together to Democr...Hunt for Collaboration: Realizing a Learning Health System Together to Democr...
Hunt for Collaboration: Realizing a Learning Health System Together to Democr...
 
Realizing a Learning Health System: A Vision for Education to Transform the F...
Realizing a Learning Health System: A Vision for Education to Transform the F...Realizing a Learning Health System: A Vision for Education to Transform the F...
Realizing a Learning Health System: A Vision for Education to Transform the F...
 
The Learning Health System: Democratizing Health Together
The Learning Health System: Democratizing Health TogetherThe Learning Health System: Democratizing Health Together
The Learning Health System: Democratizing Health Together
 
Breast Density Notification Law Information For Primary Care Providers
Breast Density Notification Law Information For Primary Care ProvidersBreast Density Notification Law Information For Primary Care Providers
Breast Density Notification Law Information For Primary Care Providers
 
Implementation science and learning health systems: Connecting the dots
Implementation science and learning health systems:  Connecting the dotsImplementation science and learning health systems:  Connecting the dots
Implementation science and learning health systems: Connecting the dots
 
Simulation –based Education in Turkey: Insights from a recent trip
Simulation –based Education in Turkey: Insights from a recent tripSimulation –based Education in Turkey: Insights from a recent trip
Simulation –based Education in Turkey: Insights from a recent trip
 
Simulation and the Family Medicine Physician: Current and future applications...
Simulation and the Family Medicine Physician: Current and future applications...Simulation and the Family Medicine Physician: Current and future applications...
Simulation and the Family Medicine Physician: Current and future applications...
 
Simulation-based education: How could simulation be used for technical skills...
Simulation-based education: How could simulation be used for technical skills...Simulation-based education: How could simulation be used for technical skills...
Simulation-based education: How could simulation be used for technical skills...
 
Simulation-based education: How could simulation be used for technical skills...
Simulation-based education: How could simulation be used for technical skills...Simulation-based education: How could simulation be used for technical skills...
Simulation-based education: How could simulation be used for technical skills...
 
Learning Health System Briefing
Learning Health System BriefingLearning Health System Briefing
Learning Health System Briefing
 
Implementation science and learning health systems: Pieces of a puzzle
Implementation science and learning health systems: Pieces of a puzzle Implementation science and learning health systems: Pieces of a puzzle
Implementation science and learning health systems: Pieces of a puzzle
 
Transforming Your Educational Efforts into Scholarship
Transforming Your Educational Efforts into ScholarshipTransforming Your Educational Efforts into Scholarship
Transforming Your Educational Efforts into Scholarship
 
Overview of National Learning Health Community & Learning Health for Michigan...
Overview of National Learning Health Community & Learning Health for Michigan...Overview of National Learning Health Community & Learning Health for Michigan...
Overview of National Learning Health Community & Learning Health for Michigan...
 
National Learning Health System Landscape
National Learning Health System LandscapeNational Learning Health System Landscape
National Learning Health System Landscape
 
CHEPS Poster: Human Infrastructure for Patient Safety: PASQUAL Scholars Program
CHEPS Poster: Human Infrastructure for Patient Safety: PASQUAL Scholars ProgramCHEPS Poster: Human Infrastructure for Patient Safety: PASQUAL Scholars Program
CHEPS Poster: Human Infrastructure for Patient Safety: PASQUAL Scholars Program
 

Recently uploaded

UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdfUGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdfNirmal Dwivedi
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfciinovamais
 
ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.MaryamAhmad92
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxAmanpreet Kaur
 
PROCESS RECORDING FORMAT.docx
PROCESS      RECORDING        FORMAT.docxPROCESS      RECORDING        FORMAT.docx
PROCESS RECORDING FORMAT.docxPoojaSen20
 
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxBasic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxDenish Jangid
 
Grant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingGrant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingTechSoup
 
Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Jisc
 
How to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POSHow to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POSCeline George
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfPoh-Sun Goh
 
How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17Celine George
 
Sociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning ExhibitSociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning Exhibitjbellavia9
 
Unit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptxUnit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptxVishalSingh1417
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhikauryashika82
 
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...ZurliaSoop
 
General Principles of Intellectual Property: Concepts of Intellectual Proper...
General Principles of Intellectual Property: Concepts of Intellectual  Proper...General Principles of Intellectual Property: Concepts of Intellectual  Proper...
General Principles of Intellectual Property: Concepts of Intellectual Proper...Poonam Aher Patil
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsMebane Rash
 
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...christianmathematics
 

Recently uploaded (20)

UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdfUGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
 
ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.ICT role in 21st century education and it's challenges.
ICT role in 21st century education and it's challenges.
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
 
PROCESS RECORDING FORMAT.docx
PROCESS      RECORDING        FORMAT.docxPROCESS      RECORDING        FORMAT.docx
PROCESS RECORDING FORMAT.docx
 
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxBasic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
 
Grant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingGrant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy Consulting
 
Asian American Pacific Islander Month DDSD 2024.pptx
Asian American Pacific Islander Month DDSD 2024.pptxAsian American Pacific Islander Month DDSD 2024.pptx
Asian American Pacific Islander Month DDSD 2024.pptx
 
Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)
 
How to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POSHow to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POS
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdf
 
How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17
 
Sociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning ExhibitSociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning Exhibit
 
Unit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptxUnit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptx
 
Spatium Project Simulation student brief
Spatium Project Simulation student briefSpatium Project Simulation student brief
Spatium Project Simulation student brief
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
 
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
 
General Principles of Intellectual Property: Concepts of Intellectual Proper...
General Principles of Intellectual Property: Concepts of Intellectual  Proper...General Principles of Intellectual Property: Concepts of Intellectual  Proper...
General Principles of Intellectual Property: Concepts of Intellectual Proper...
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan Fellows
 
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
 

Validation Studies in Simulation-based Education - Deb Rooney

  • 1. Validation Studies in Simulation-based Education April 15, 2015 Deb Rooney, PhD Professor of Learning Health Sciences All Rights Reserved. 1
  • 2. •  Validity in the current framework used to evaluate evidence •  How we gather /evaluate validity evidence from a simulator and it’s associated measures •  Context of academic product (manuscripts) •  Final considerations Objectives 2
  • 3. A few definitions to consider…. Validity: What is it? 2. the degree to which evidence and theory support the interpretations of test scores as entailed by proposed uses of tests –Standards (1999) 1. the degree to which the tool measures what it claims to measure American Educational Research Association, American Psychological Association, & National Council on Measurement in Education. (1999) Standards for educational and psychological testing. Washington, DC: American Educational Research Association. 3
  • 4. •  Evidence relevant to relationships to other variables (e.g. novice versus expert discriminant validity) is over-represented •  Evidence relevant to response process and consequences of testing and are infrequently reported Cook, D. A., Brydges, R., Zendejas, B., Hamstra, S. J., & Hatala, R. (2013). Technology-enhanced simulation to assess health professionals: A systematic review of validity evidence, research methods, and reporting quality. Academic Medicine. Jun;88(6):872-83. •  Apply current Standards to ensure rigorous research and reporting Simulator Validation: The Framework 4
  • 5. Simulator Validation: The Evidence Current AERA Standards* o  Not new/novel o  Unitary Construct (All evidence falls under “Construct” validity) o  Test content o  Internal structure o  Response Processes o  Relationships to other variables o  Consequences of testing (psychometric properties of measures) (comparison w/ previously-validated measures) (standard setting, rater/ings quality, fidelity vs. stakes) (reliability, dimensionality, function across groups) (face validity, subjective measures, construct alignment) *Standards for Educational and Psychological Testing (AERA, APA, & NCME, 2014) 5
  • 6. Validity evidence; Validity: What is it NOT? •  Does not allow us to make inferences about a curriculum •  Does not allow us to make inferences about different applications, settings, or learners •  Is not a terminal quality determination (about the quality of your measures, application) •  “The scale was valid” ! “evidence supports the use of a scale to measure X in a particular setting/ application” 6
  • 7. We have a much more complex environment to evaluate! Simulator test content relationships to other variables consequences of testing Instrument (measures) test content internal structure relationships to other variables response processes consequences of testing Simulator Validation: How does this evidence apply to us? 7
  • 8. CREATIONDESIGN & PLANNING IMPLEMENTATION & EVALUATION Test content (measures/simulator) Internal structure (m) Response processes (m) Relationship to other variables (m/s) Consequences of testing (m/s) Validity Evidence in Simulation How/when do we gather evidence? 8
  • 9. PAPER 2PAPER 1 PAPER 3 Test content (measures & simulator) Before implementation Validity Evidence in Simulation How disseminate findings? Quality of performance measures Before or After Implementation Impact on performance and/ or patient outcomes After full implementation 9
  • 10. Most Recent Example: Neurosurgery Sim a Tai B, Rooney D Stephenson F, Liao P, Sagher O, Shih A, Savastano LE. Development of 3D- printing built ventriculostomy placement simulator, Journal of Neurosurgery, (in press) bRooney DM, Tai BL, Sagher O, Shih AJ, Wilkinson DA, Savastano LE. A Simulator and two tools: Validation of performance measures from a novel neurosurgery simulator using the current Standards framework. Surgery, submitted 3/15. Paper 1a: Preliminary evaluation of quality of simulator/ measures Paper 2b: Evaluation of performance measures from simulator Paper 3: Evaluation of Impact on performance measures / patient outcomes 10
  • 11. Paper 1: Simulator validation process n=7 n=5 n=5 4 months 11
  • 12. The Content Validity Form: 5 domains •  Physical Attributes •  Realism- experience •  Value •  Relevance •  Overall (global) 12
  • 13. Paper 1: The preliminary validation process (Sim)* •  Using Rasch model, analyzed data for; •  Looked at domain rating differences across 3 sites •  Mean averages by item •  Looked at Rasch variability indices to identify possible inconsistency in ratings •  Ensured psychometric quality of survey •  Using traditional methods, estimated; •  Inter-item consistency, Cronbach alpha •  Inter-rater agreement, ICC(2,k) •  Using a Rasch model ensured rating scales’ function *performance checklist is separate/ different process 13
  • 14. Results: Domain mean averages by site 0   0.5   1   1.5   2   2.5   3   3.5   4   4.5   UM   HF   WS   3.4 3.3 3.9 3.3 2.4 Combined mean averages “This simulator requires minor adjustments before it can be considered for use in ventriculostomy placement training.” 14
  • 16. Paper 1*: Test Content-Checklist *shoulda, coulda, woulda 16
  • 17. Proposed items Definitely do not include this task (1) Not sure if this task should be included (2) Pretty sure this task should be included (3) Definitely include this task (4) Position head and mark midline Locate Kocher's point (10.5 cm posterior to the nasion and 3 cm lateral to midline) Mark incision (approximately 2cm long in a parasagittal location) Incise, clear tissue off cranium, retracted scalp … Suture wound (Staples or a 3-0 running nylon or prolene suture) Paper 1*: Test Content-Checklist 17
  • 18. Paper 1*: Test Content-Checklist Ask expert instructors about the value of included steps (items) for measuring X at doing Y. •  Reasonable number of experts ~ 3 What else do you ask about? •  Clarity of item •  Appropriateness of qualifiers (use X instrument, at x location) •  Rating scale •  Missing steps •  Objective measures to include (eg. time to) *shoulda, coulda, woulda 18
  • 19. PAPER 2PAPER 1 PAPER 3 Before implementation Test content (measures & simulator) Before or After Implementation* Quality of performance measures After full implementation Impact on performance and/or patient outcomes Next Step: Deeper Evaluation (Paper 2) 19
  • 20. Next Step: Deeper Evaluation (Paper 2) •  Evaluation of all validity evidence of performance measures [ala Standards] •  Capture broader (regional/national) sample of performance data via videotaped performances •  ideal N (+++) and range of experience •  Compare measures from the novel performance checklist and gold standard (eg-OSATS] (relationship with other variables*) •  Set/test performance standards (if appropriate) Rooney DM, Tai BL, Sagher O, Shih A, Wilkinson DA, Savastano, L. A Simulator and two tools: Validation of performance measures from a novel neurosurgery simulator using the current Standards framework. Surgery, submitted 3/15) 20
  • 21. •  Nationally-recognized training program sponsored by Society of Neurological Surgeons •  Total n=14 (11 trainees*, 3 attendings) performed ventriculostomy on simulator •  All performances were video- captured, scored by 3 raters using novel checklist and modified version of OSATS Paper 2: study design *first year neurosurgery fellows 21
  • 23. Martin JA, Regehr G, Reznick R, MacRae H, Murnaghan J, Hutchison C, et al. (1997). Objective structured assessment of technical skill (OSATS) for surgical residents. Br J Surg 84:273-278. Modified Objective Structured Assessment of Technical Skills: m-OSATS 23
  • 24. Measures adequately reflect ventriculostomy performance “quality” •  Relationships to other variables •  Trainee v expert ratings •  Correlation of summed V-PAT with summed OSATS scores Measures are psychometrically sound (Adequate “Quality Control”) •  Psychometric function of V-PAT & OSATS measures; •  Response processes: Rasch indices ! rating scale function •  Test Content: Rasch item point-measure correlations, item fit (variability) •  Internal structure: Inter-item consistency-Cα, Inter-rater agreement- ICC(2,k) Measures are free from rater bias •  Consequences of testing •  Evaluated Rasch bias indices to identify potential rating differences at rater level Paper 2: evidence examined Examined evidence from 5 sources, but packaged a bit differently; 24
  • 25. Response processes •  Rasch indices: (Avg meas., Fit statistics, and RA thresholds) indicated all rating scales for both V-PAT & OSATS were well-functioning Test content Point-measure correlations: all positive, [.39, .81] Rasch item Outfit MS: all < 2.0 Internal structure •  Cronbach α: both α = 0.95 •  Intraclass correlation: V-PAT= [-0.33, 0.93] OSATS = [0.80, 0.93] Results: “Quality control” 25
  • 26. Do V-PAT measures adequately differentiate trainee and expert performances?  Instrument Indices V-PAT Resident Observed Average (SE) Attending Observed Average (SE) P- value ICC (2,k) Value 1. Position head and mark midline 3.75 (.20) 4.00 (.47) 0.54 * 2. Locate Kocher's point 3.84 (.18) 4.11 (.34) 0.52 .86 3. Mark an incision 3.61 (.16) 4.00 (.32) 0.42 .10 4. Select drain exit site from the scalp 2.94 (.17) 3.71 (.50) 0.22 * 5. Incise, clear tissue off cranium, retract scalp 3.56 (.14) 4.11 (.34) 0.20 .83 6. Set drill stop and drill trephine 2.80 (.16) 3.78 (.37) 0.08 .70 7. Confirm dura and pierce 2.98 (.17) 3.67 (.35) 0.25 .76 8. Confirm landmarks and place catheter 2.88 (.17) 3.33 (.35) 0.24 .91 9. Confirm CSF flow 3.41 (.13) 3.67 (.30) 0.39 .73 10. Remove trocar cover, tunnel trocar to exit site and recap trocar 2.78 (.14) 3.00 (.47) 0.62 -.33 11. Place purse string suture to anchor the catheter at scalp exit site 3.00 (.36) 4.25 (.69) 0.26 * Overall average 3.30 (.06) 3.80 (.11) 0.01 –  Instrument Indices V-PAT Resident Observed Average (SE) Attending Observed Average (SE) P- value ICC (2,k) Value 1. Position head and mark midline 3.75 (.20) 4.00 (.47) 0.54 * 2. Locate Kocher's point 3.84 (.18) 4.11 (.34) 0.52 .86 3. Mark an incision 3.61 (.16) 4.00 (.32) 0.42 .10 4. Select drain exit site from the scalp 2.94 (.17) 3.71 (.50) 0.22 * 5. Incise, clear tissue off cranium, retract scalp 3.56 (.14) 4.11 (.34) 0.20 .83 6. Set drill stop and drill trephine 2.80 (.16) 3.78 (.37) 0.08 .70 7. Confirm dura and pierce 2.98 (.17) 3.67 (.35) 0.25 .76 8. Confirm landmarks and place catheter 2.88 (.17) 3.33 (.35) 0.24 .91 9. Confirm CSF flow 3.41 (.13) 3.67 (.30) 0.39 .73 10. Remove trocar cover, tunnel trocar to exit site and recap trocar 2.78 (.14) 3.00 (.47) 0.62 -.33 11. Place purse string suture to anchor the catheter at scalp exit site 3.00 (.36) 4.25 (.69) 0.26 * Overall average 3.30 (.06) 3.80 (.11) 0.01 – Results: Relationship to Other variables (V-PAT)  Instrument Indices V-PAT Resident Observed Average (SE) Attending Observed Average (SE) P- value ICC (2,k) Value 1. Position head and mark midline 3.75 (.20) 4.00 (.47) 0.54 * 2. Locate Kocher's point 3.84 (.18) 4.11 (.34) 0.52 .86 3. Mark an incision 3.61 (.16) 4.00 (.32) 0.42 .10 4. Select drain exit site from the scalp 2.94 (.17) 3.71 (.50) 0.22 * 5. Incise, clear tissue off cranium, retract scalp 3.56 (.14) 4.11 (.34) 0.20 .83 6. Set drill stop and drill trephine 2.80 (.16) 3.78 (.37) 0.08 .70 7. Confirm dura and pierce 2.98 (.17) 3.67 (.35) 0.25 .76 8. Confirm landmarks and place catheter 2.88 (.17) 3.33 (.35) 0.24 .91 9. Confirm CSF flow 3.41 (.13) 3.67 (.30) 0.39 .73 10. Remove trocar cover, tunnel trocar to exit site and recap trocar 2.78 (.14) 3.00 (.47) 0.62 -.33 11. Place purse string suture to anchor the catheter at scalp exit site 3.00 (.36) 4.25 (.69) 0.26 * Overall average 3.30 (.06) 3.80 (.11) 0.01 – 26
  • 27. Do m-OSATS measures adequately differentiate trainee and expert performances? •  Correlation of summed V-PAT scores with summed m-OSATS •  Pearson’s r = 0.72, p = 0.001 V-PAT Resident Observed Average (SE) Attending Observed Average (SE) P- value ICC (2,k) Value 1. Position head and mark midline 3.75 (.20) 4.00 (.47) 0.54 * 2. Locate Kocher's point 3.84 (.18) 4.11 (.34) 0.52 .86 3. Mark an incision approximately 2 cm long in a parasagittal location 3.61 (.16) 4.00 (.32) 0.42 .10 4. Select drain exit site from the scalp 2.94 (.17) 3.71 (.50) 0.22 * 5. Incise, clear tissue off cranium, retract scalp 3.56 (.14) 4.11 (.34) 0.20 .83 6. Set drill stop and drill trephine 2.80 (.16) 3.78 (.37) 0.08 .70 7. Confirm dura and pierced with 18g spinal needle or 11 blade scalpel 2.98 (.17) 3.67 (.35) 0.25 .76 8. Confirm landmarks and place catheter to 6-7cm from outer table of skull 2.88 (.17) 3.33 (.35) 0.24 .91 9. Confirm CSF flow 3.41 (.13) 3.67 (.30) 0.39 .73 10. Remove trocar cover, tunnel trocar to exit site and recap trocar 2.78 (.14) 3.00 (.47) 0.62 -.33 11. Place purse string suture at the scalp exit site to anchor the catheter 3.00 (.36) 4.25 (.69) 0.26 * Overall average 3.30 (.06) 3.80 (.11) 0.01 – OSATS 1. Respect for Tissue 2.66 (.21) 4.11 (.34) 0.004 .93 2. Time and Motion 2.42 (.22) 4.00 (.43) 0.005 .85 3. Instrument Handling 2.51 (.22) 4.00 (.46) 0.007 .86 4. Knowledge of Instruments 2.36 (.21) 4.33 (.43) 0.001 .84 5. Flow of Operation 2.36 (.21) 4.22 (.45) 0.001 .85 6. Knowledge of specific procedure 2.33 (.23) 4.33 (.46) 0.001 .80 Overall average 2.32 (.08) 3.73 (.15) 0.001 – * too few cases to estimate              Instrument Indices V-PAT Resident Observed Average (SE) Attending Observed Average (SE) P- value ICC (2,k) Value 1. Position head and mark midline 3.75 (.20) 4.00 (.47) 0.54 * 2. Locate Kocher's point 3.84 (.18) 4.11 (.34) 0.52 .86 3. Mark an incision approximately 2 cm long in a parasagittal location 3.61 (.16) 4.00 (.32) 0.42 .10 4. Select drain exit site from the scalp 2.94 (.17) 3.71 (.50) 0.22 * 5. Incise, clear tissue off cranium, retract scalp 3.56 (.14) 4.11 (.34) 0.20 .83 6. Set drill stop and drill trephine 2.80 (.16) 3.78 (.37) 0.08 .70 7. Confirm dura and pierced with 18g spinal needle or 11 blade scalpel 2.98 (.17) 3.67 (.35) 0.25 .76 8. Confirm landmarks and place catheter to 6-7cm from outer table of skull 2.88 (.17) 3.33 (.35) 0.24 .91 9. Confirm CSF flow 3.41 (.13) 3.67 (.30) 0.39 .73 10. Remove trocar cover, tunnel trocar to exit site and recap trocar 2.78 (.14) 3.00 (.47) 0.62 -.33 11. Place purse string suture at the OSATS Results: Relationship to Other variables (m-OSATS) 27
  • 28. Are measures free from rater bias? Results: Consequences of Testing V-PAT Observed Average (SE) 1. Rater 1 (LS) 3.60 (.11) 2. Rater 2 (OS) 3.60 (.10) 3. Rater 3 (DW) 3.60 (.11) 4. Participants (Variable)† 2.90 (.11) Overall average 3.60 ( – ) OSATS 1. Rater 1 (LS)* 2.10 (.17) 2. Rater 2 (OS) 3.30 (.15) 3. Rater 3 (DW) 3.00 (.15) Overall average 2.80 (–) †Comparison with 3 expert raters, p = 0.01 * Comparison with 2 expert raters, p = 0.01 V-PAT Observed Average (SE) 1. Rater 1 (LS) 3.60 (.11) 2. Rater 2 (OS) 3.60 (.10) 3. Rater 3 (DW) 3.60 (.11) 4. Participants (Variable)† 2.90 (.11) Overall average 3.60 ( – ) OSATS 1. Rater 1 (LS)* 2.10 (.17) 2. Rater 2 (OS) 3.30 (.15) 3. Rater 3 (DW) 3.00 (.15) Overall average 2.80 (–) †Comparison with 3 expert raters, p = 0.01 * Comparison with 2 expert raters, p = 0.01 V-PAT Observed Average (SE) 1. Rater 1 (LS) 3.60 (.11) 2. Rater 2 (OS) 3.60 (.10) 3. Rater 3 (DW) 3.60 (.11) 4. Participants (Variable)† 2.90 (.11) Overall average 3.60 ( – ) OSATS 1. Rater 1 (LS)* 2.10 (.17) 2. Rater 2 (OS) 3.30 (.15) 3. Rater 3 (DW) 3.00 (.15) Overall average 2.80 (–) †Comparison with 3 expert raters, p = 0.01 * Comparison with 2 expert raters, p = 0.01 28
  • 29. Are OSATS measures free from rater bias? 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 Q1 Q2 Q3 Q4 Q5 Q6 Observedaverage OSATS item Bias interaction: Rater by item Rater1-LS Rater2-OS Rater3-DW Results: Consequences of Testing 29
  • 30. Summary of Results Evidence Inferences V-PAT OSATS Quality Control Response Processes Adequate rating scale function √ √ Test Content Items align with construct √ √ Internal Structure Inter-item consistency, inter-rater agreement* X* √ Test of Assumptions Rel. Other Variables Measures dif. bet. N/E performances X √ Rel. Other Variables V-PAT/OSATS summed scores correlate √ √ Cons. of Testing V-PAT/OSATS measures are bias free √ X X = challenges that require resolution 30
  • 31. Standards for educational and psychological testing. American Educational Research Association, American Psychological Association, & National Council on Measurement in Education. (2014). Washington, DC: American Educational Research. Available for purchase via http://teststandards.org/ Validity: ala 2015 •  Evidence is important, but interpretive argument is critical •  Content of interpretive argument determines the kinds of evidence that are most relevant (most important) in validation •  Strategy of developing interpretive argument based on •  Validity evidence relevant to inferences •  Assumptions •  Challenges (alternative interpretations) 31
  • 32. Problematic inter-rater agreement (ICC) for 5 items should be resolved; •  Item 3 (Mark an incision approximately 2 cm long in a parasagittal location), ICC= .10 •  Item 10 (Remove trocar cover, tunnel trocar to exit site and recap trocar), ICC= -.33 Examine, refine items to ensure alignment with simulator capabilities/ are mutually exclusive Challenges: potential threats to validity (V-PAT) •  Item 1 (Position head and mark midline)* •  Item 4 (Select drain exit site from the scalp)* •  Item 1 (Place purse string suture at the scalp exit site to anchor the catheter) * * ICC incalculable 32
  • 33. “Hawkish” OSATS ratings by one expert rater requires follow-up Refine items, add rater training on scoring rubric and administration standards Challenges: potential threats to validity (m-OSATS) 33
  • 34. PAPER 2PAPER 1 PAPER 3 Before implementation Test content (measures & simulator) Before or After Implementation* Quality of performance measures After full implementation Impact on performance and/or patient outcomes Next Step: Evaluation of Impact (Paper 3) 34
  • 35. •  Evaluation of impact on trainee’s clinical performance or patient outcomes [relationship with other variables] •  Examine; •  Change in trainees’ clinical performance (checklist ratings, objective measures (“time to”, LOS, adverse events) •  Impact on hospital costs Barsuk JH, Cohen ER, Feinglass J, Kozmic SE, McGaghie WC, Ganger D, Wayne DB. Cost savings of performing paracentesis procedures at the bedside after simulation- based education. Simul Healthc. 2014 Oct;9(5):312-8. Next Step: Evaluation of Impact (Paper 3) 35
  • 36. •  Validation process is fluid/reiterative/on- going •  It takes a team; •  Development (clinicians, instructors, engineers, researchers assistants)! •  Outcomes (+QI, hospital info) •  There is funding; •  AHRQ •  P-CORi •  Michigan Blue Cross Blue Shield Considerations 36
  • 37. Thank you Questions?     Deborah  Rooney,  PhD   dmrooney@med.umich.edu