Upload
others
View
1
Download
0
Embed Size (px)
Citation preview
9/21/2012
1
R b HRob Horner, Charlie Greenwood
www.uoecs.org
Define an emerging role for single case designsDefining features of Single‐case research
Programs of research
Grant applications
Review standards for visual analysis of single caseReview standards for visual analysis of single case studies
WWC Design standards
WWC Evidence criteria
Training protocol for new scholars
Introduce Repeated Acquisition Designs
Single‐case methods developed and used within Applied Behavior Analysis
9/21/2012
2
We have a window of time to define the contribution of Single‐case research
What is Needed?
‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐‐Document how Single‐case methods constitute a rigorous form of science
Improve the training provided to new scholars
Improve the review standards used for journal and grant applications (visual and statistical analysis)
Define a standard for documenting evidence‐base practices using single‐case research methods.
IES White Paper
What Works ClearinghouseProtocol for using Single‐case Research in the identification of “promising” and “evidence‐based” practices
Real Need: Agreement within the Single‐case research community about standards for design and analysis.
Reversal/Withdrawal Designs
Multiple Baseline Designs
Alternating Treatment Designs
Others: Changing Criterion
Repeated Acquisition
Multiple Probe
9/21/2012
3
Useful in the iterative development of interventions.
Useful in design of any interventions focused on the “individual participant” (especially with low‐incidence p p ( p ypopulations)
Useful for pilot research to assess the effect size needed for RCTs.
Useful for fine‐grained analysis of “weak and non‐responders”
Designing interventions
DevelopDevelopInterventionIntervention
It ti t tiIt ti t tiIterative testing Iterative testing with Singlewith Single--case Designcase Design
Randomized Randomized Control TrialsControl Trials
Analysis of Weak and Analysis of Weak and NonNon--Responders with Responders with SingleSingle--case Designscase Designs
Goal 4:Goal 4:EffectivenessEffectiveness
Goal 3:Goal 3:EfficacyEfficacy
Goal 2:Goal 2:DevelopmentDevelopment
What Works Clearinghouse StandardsDesign Standards
Evidence Criteria
Social Validity
9/21/2012
4
Evaluate the Design
Meets Design Standards Meets with Reservations Does Not Meet Design Standards
Evaluate the Evidence
Does the Design Does the Design allow allow
Documentation of Documentation of Experimental Experimental
Control ?Control ?
Do the data Do the data document document
Experimental Experimental Control ?Control ?
StopStop
Strong Evidence Moderate Evidence No Evidence
Effect‐Size Estimation
Social Validity Assessment
Control ?Control ?
Is the effect Is the effect something we something we should care should care
about?about?
StopStop
Evaluate the Design
Meets Design Standards Meets with Reservations Does Not Meet Design Standards
Evaluate the Evidence
Strong Evidence Moderate Evidence No Evidence
Effect‐Size Estimation
Social Validity Assessment
Meets StandardIV manipulated directly
Measure is reliable (IOA = .80 percent agreement; .60 Kappa)
20% of data points in each phase
Design allows opportunity to assess basic effect at three different points in time. Controls for threat to Int Validity.
Five data points per phase (or design equivalent)
ATD (four comparison option)
Meets with ReservationAll of above, except at least three data points per phase
Non‐concurrent Multiple Baseline
Does not Meet Standard
9/21/2012
5
Design Standard
Does the design allow for the opportunity to assess experimental control?
Baseline
At least five data points per phase (3 w/reservation)
O t it t d t t l t 3 b i ff tOpportunity to document at least 3 basic effects, each at a different point in time.
First Demonstration of Basic Effect Third Demonstration of Basic Effect
Intervention X Intervention X
First Demonstration of Basic Effect
Second Demonstration of Basic Effect
Third Demonstration of Basic Effect
1. Baseline 2. Each phase has at least 5 data points (3 w/reservation)3. Design allows for assessment of “basic effect” at three different points in time.
Intervention X
Does Not MeetDoes Not MeetDoes Not MeetDoes Not MeetStandardStandard
9/21/2012
6
Intervention X Intervention Y
Does Not MeetDoes Not MeetStandardStandard
Intervention X Intervention X
Does Not MeetDoes Not MeetStandardStandard
Intervention X Intervention X
Meets StandardMeets StandardWith With
ReservationReservation
9/21/2012
7
First Demonstration of Basic Effect
Second Demonstration of Basic Effect
Third Demonstration of Basic Effect
Does Not Meet Does Not Meet StandardStandard
Meets StandardMeets Standard
9/21/2012
8
Time
Review Fisher studies:
To what extent does each graph allow documentation of experimental control?
9/21/2012
9
Evaluate the Design
Meets Design Standards Meets with Reservations Does Not Meet Design Standards
Evaluate the Evidence
Strong Evidence Moderate Evidence No Evidence
Effect‐Size Estimation
Social Validity Assessment
StrongBaseline
Documentation of research question “problem”
Documentation of predictable pattern (>5 data points)
Each Phase of the AnalysisDocumentation of predictable pattern (> 5 data points)
Basic effectsDocumentation of predicted change in the DV when IV is manipulated
Experimental ControlThree demonstrations of basic effect, each at a different point in time.
No demonstrations of intervention failure
ModerateAll of “Strong” criteria, with these exceptions:
Only 3‐4 data points per phase
Three demonstrations of effect, but with additional demonstrations of failure‐to‐document effect.
Non‐concurrent multiple baseline
No EvidenceMisnomer
Evidence does not meet Moderate level.
9/21/2012
10
BaselineDocument the “problem” requiring intervention
Typically 5 or more data points
Documentation of a pattern of responding that allows prediction into the future.
Each PhaseDocuments a clear pattern of responding
Typically 5 or more data points
Adjacent phasesDo data document a “basic effect”
Whole studyDo the phases document experimental control (e.g. at least three demonstrations of a basic effect, each at a different point in time.
Level
Trend
Variability
Within Within PhasePhase
y
+
Overlap
Immediacy of Effect
Consistency across similar phases
_________________________________________
Other: vertical analysis; intercept gap
Between Between PhasesPhases
First Demonstration of Effect Third Demonstration of EffectFirst Demonstration of Effect
Second Demonstration of Effect
Third Demonstration of Effect
Visual Analysis:
1. Change in Level
2. Change in Trend
3. Change in Variability
4. Immediacy of Effect
5. Overlap
6. Consistency of Data Patterns across similar Phases
Parsonson & Baer, 1978;
Kratochwill & Levin, 1992
9/21/2012
11
First Demonstration of Effect Third Demonstration of EffectFirst Demonstration of Effect
Second Demonstration of Effect
Third Demonstration of Effect
Comparison of actual against projected data
(Analysis of Transition States versus Analysis of Steady States)
First Demonstration of Effect Third Demonstration of EffectFirst Demonstration of Effect
Second Demonstration of Effect
Third Demonstration of Effect
Consistency in data pattern across similar phases
Level, Trend, Variability, Overlap
Immediacy of Effect
Consistency across similar phases
Stability in non‐intervened series when effect demonstrated in one series
9/21/2012
12
Magnitude of separationGreater the difference between two conditions, larger the demonstration of a functional relation
Consistency of separationConsistency of separationGreater consistency of separation between two conditions (no overlap) larger the demonstration of a functional relation
Number of data points used to establish experimental control
At least 4 comparisons. The more points documenting separation the stronger the demonstration of experimental control.
40
50
60
70
with
Scr
eam
, P
ush Alice
Tangible
Escape
0
10
20
30
Per
cent
age
of T
rials
1 2 3 4 5 6 7 8 9 10FA sessions
Escape
Attention
Control
* * *
9/21/2012
13
Increase the precision of Research QuestionsDefine conceptual logic for research question
Define research question with greater precisionIV related to change in level, trend, variability?
“Is there is a functional relation between Functional Communication Training and reduction in the level and variability of problem behavior?”
Development of visual analysis graphsBruce Wampold, Richard Freund, Rick Albin, Tom Kratochwill, Chris SwobodaLevel, trend, variability, overlapPurpose is to emphasize interaction
Examine and score each graphUse ALL data in the graphCombine assessment of:
L l t d i bilit l i di f ff t i t f d tLevel, trend, variability, overlap, immediacy of effect, consistency of data patterns in similar phases.
Use scoring metric:0= no functional relation 5 = publishable7 = strong functional relation
Compare scores with Excel fileMedian scores from five “expert” Single‐case Researchers ABCB
MBL
Excel
Is there a functional relation between the Intervention and reduction in the level of the problem behavior?
N E C t l P bli h bl St E C t lNo Exp Control Publishable Strong Exp Control
1 2 3 4 5 6 7
9/21/2012
14
6
Level of Experimental Control
No Exp Control Publishable Strong Exp Control1 2 3 4 5 6 7
1
Level of Experimental Control
No Exp Control Publishable Strong Exp Control1 2 3 4 5 6 7
Level of Experimental Control
No Exp Control Publishable Strong Exp Control1 2 3 4 5 6 7
9/21/2012
15
Level of Experimental Control
No Exp Control Publishable Strong Exp Control1 2 3 4 5 6 7
Level of Experimental Control
No Exp Control Publishable Strong Exp Control1 2 3 4 5 6 7
Level of Experimental Control
No Exp Control Publishable Strong Exp Control1 2 3 4 5 6 7
9/21/2012
16
Evaluation for LEVELEvaluation for LEVELEvaluate for TRENDEvaluate for TREND
Level of Experimental Control
No Exp Control Publishable Strong Exp Control1 2 3 4 5 6 7
Is there a functional relation between introducing the intervention and reduction in the level of the problem behavior?
6
DV:DV:ProportionProportionProportion Proportion of intervals of intervals with social with social disruptiondisruption
9/21/2012
17
Level = 2Variability = 6
Level = 7Variability = 1
Level = 3
9/21/2012
18
9/21/2012
19
Is there a difference in the percentage of 10 s intervals with social initiation under Condition B compared with Condition C?
Level of Experimental Control
No Exp Control Publishable Strong Exp Control1 2 3 4 5 6 7
Level of Experimental Control
No Exp Control Publishable Strong Exp Control1 2 3 4 5 6 7
9/21/2012
20
Level of Experimental Control
No Exp Control Publishable Strong Exp Control1 2 3 4 5 6 7
Level of Experimental Control
No Exp Control Publishable Strong Exp Control1 2 3 4 5 6 7
Level of Experimental Control
No Exp Control Publishable Strong Exp Control1 2 3 4 5 6 7
9/21/2012
21
Level of Experimental Control
No Exp Control Publishable Strong Exp Control1 2 3 4 5 6 7
Level of Experimental Control
No Exp Control Publishable Strong Exp Control1 2 3 4 5 6 7
Five studies documenting experimental control
Conducted in at least three different locations by at least three different researchersat least three different researchers
Across at least 20 different participants
_______________________________________
Each study demonstrates an effect size of ( d > 0.50)??
9/21/2012
22
Functional Communication Training to reduce the level of problem behavior in school, home and community
At least Five StudiesBird, Dores, Moniz & Robinson (1989)Brown et al., (2000)Carr & Durand (1985)D d & C (1987)Durand & Carr (1987)Hagopian, Fisher, Sullivan, Acquisto & LeBlanc (1998)Mildon, Moore, & Dixon (2004)Wacker et al., (1990)
At least three settings /scholarsSix locations, Research Groups
At least 20 participants42 different participants document effect
Evidence‐based is not enoughIn addition to the features of the practice: define whatoutcomes, when/where used, by whom, with what target populations, at what fidelity?
The innovative practice needs to not only be evidence‐based, but dramatically easier and better than what is already being used.
The practice should be defined conceptually as well as procedurally, to allow guidance for adaptation.
Documenting Experimental Control/ Statistical Significance:
No statistical model currently replicates the conceptual logic used in visual analysis
Statistical Analyses of Single‐case Research
conceptual logic used in visual analysis.
Consider statistical analysis of SCD as supplemental
Effect SizeNot a variable traditionally addressed in Single‐case Analyses.
It needs to be.
9/21/2012
23
Core criteria for an effect size measure used with single‐case research:
Comparable to Cohen’s dSo results can be used in meta‐analyses
Is this really Is this really necessary?necessary?
Assesses the “experimental effect” under analysis (e.g. examines ALL phases of a study,
not just a AB comparison
Control for:Autocorrelation (serial dependency)Weighting of level, trend, variability, overlap, immediacyOutliersType of data (e.g., counts, rates… from a non‐normal distribution)
Statistical Analysis of SCD: Current Status
Emerging ApproachesPercentage of non‐overlapping data
Parker et al.,
Wolery et al.,
Randomization Tests
K t h ill & L i
Statistical Statistical Analysis:Analysis:
1.1. Done to Done to confirm confirm visual visual
analysisanalysisKratochwill & Levin
Generalized Least Squares Regression
Swaminathan et al.,
SingSub
Hierarchical Linear Modeling
Hedges, Shadish, & Rindskopf
d‐Estimator
analysisanalysis
2.2. Done as Done as part of part of metameta--
analysesanalyses
Swaminathan et al., 2010
Control for autocorrelation
Establish regression line
Build confidence intervals
Compare observed with expected
SingSubTesting in progress
Need adaptation for dealing with non‐normal distributions.
9/21/2012
24
Baseline Acquisition Full BP‐PBS Implementationo
f Bu
llyin
g B
ehav
ior
School 1
Rob
Bruce
Cindy
School 2
Ross & Ross & Horner, Horner, JABAJABA20092009
70
Nu
mb
er o
f In
cid
ents
o
School Days
Scott
Anne
Ken
School 3
3.14 1.88 .88 72%
Conditional Probabilities of Victim Responses to Problem Behavior
20%
30%
40%
50%
bilit
y of
Res
ponse
Baseline
BP-PBS
28% increase 19% decrease
BP‐PBS, Scott Ross 71
0%
10%
"Sto
p"
"Walk
"
Posi
tive R
esp
onse
(laughin
g/c
heering)
Negativ
e
Resp
onse
(cry
ing/fi
ghtin
g
back
)
No R
esp
onse
Pro
bab
Baseline Acquisition Full BP‐PBS Implementation
of B
ully
ing
Beh
avio
r
School 1
Rob
Bruce
Cindy
School 2
SingSubSingSub--------------------------Control for Control for
AutocorrelationAutocorrelation
Build regression Build regression
72
Nu
mb
er o
f In
cid
ents
o
School Days
Scott
Anne
Ken
School 3
line with line with confidence confidence
intervalsintervals
Compare to Compare to Observed DataObserved Data
9/21/2012
25
Student Autocorrelation MSE Standardized Effect t df p
Rob -0.333 1.192 -9.961 -7.579 40 <0.001*
Table 3GLS Effect Sizes for Ross & Horner (2009), Target Students
Mean Effect Size= 5.87Mean Effect Size= 5.87
Bruce 0.006 0.927 -8.663 -4.612 40 <0.001*
Cindy -0.254 0.951 -3.220 -5.044 42 <0.001*
Scott -0.272 0.736 -5.034 -8.308 43 <0.001*
Anne 0.059 1.010 -4.450 -7.398 41 <0.001*
Ken -0.281 0.689 -3.900 -8.740 42 <0.001*
Here is the basic version of the d‐estimator
22 2 2
1223 2 2 2
4 84 5
1 1 2 1
1 1 2 1
mn
ii
B n
mD
mES
S m n n
Sn n n
Hedges, Hedges, ShadishShadish & Rindskopf& Rindskopf
D‐bar is the difference between baseline and treatment period means (averaged across people), Si
2 is the sample variance of the observations in the baseline period for the ith person, SB
2 is the sample variance between the baseline person means, is the autocorrelation, n is the number of time points, m is the number of people. The parenthetical expression in the numerator is a correction for small sample bias.
Single case methods are an effective and efficient approach for documenting experimental effects.
Need exists for more precise standards for training and using visual analysis, and combinations of visual analysis with statistical analysis.
Exp Control: Three demonstrations of basic effect, each at a different point in time.
Research Questions: Define variable of expected effect (level, trend)
We need agreement as a field on the standard(s) we will use to document evidence‐based practices via single‐case methods.
Define “Practices”: Op Description, Who can use, For whom, with What outcome
Evidence base practices: 5 studies, 3 Researcher/locations, >participants.
There are encouraging (but still emerging) approaches for statistical analysis that will improve meta‐analysis options.