Chapter 14 - Repeated Measures Designs
Chapter 14 – Repeated-Measures Designs
[As in previous chapters, there will be substantial rounding in
these answers. I have attempted to make the answers fit with the
correct values, rather than the exact results of the specific
calculations shown here. Thus I may round cell means to two
decimals, but calculation is carried out with many more
decimals.]
14.1Does taking the GRE repeatedly lead to higher scores?
a.Statistical model:
or
ijijijijijijij
XeXe
mptptmpt
=++++=+++
b.Analysis:
(
)
(
)
2
2
2
13520
7811200194933.33
24
total
X
SSX
N
=-=-=
å
å
(
)
(
)
(
)
(
)
2
22
..
.
3[566.67563.33...573.33563.33]363222.221
89,666.67
subj
i
SStXX
=S-
=-++-==
(
)
(
)
(
)
[
]
2
22
2
...
8[552.50563.33(563.75563.33)573.75563.33
]
8226.041808.33
testj
SSnXX
=S-=-+-+-
==
194,933.33189,666.671808.333458.33
errortotalsubjtest
SSSSSSSS
=--
=--=
Source
df
SS
MS
F
Subjects
7
189,666.66
Within subj
16
5266.67
Test session
2
1808.33
904.17
3.66 ns
Error
14
3458.33
247.02
Total
23
194,933.33
14.2Data on first two Test Sessions in Exercise 14.1:
a.Related-sample t test:
Subj
First
Second
Diff
1
550
570
20
2
440
440
0
3
610
630
20
4
650
670
20
5
400
460
60
6
700
680
-20
7
490
510
20
8
580
550
-30
Mean
11.25
(
)
2
90
6500
8
27.999
7
011.25
1.14
27.999
8
D
D
s
D
t
s
-
==
-
===
[t.025(7) = +2.365]
Do not reject H0
b.Repeated-measures ANOVA:
Source
df
SSMSF
Between subj
7
130,793.75
Within subj
8
3250.00
Test session
1
506.25
506.25
1.29ns
Error
7
2743.75
391.96
Total
15
134,185.94
F
=
1
.
29
=
1
.
14
=
t
from
part
a
.
14.3Teaching of self-care skills to severely retarded
children:
Cell means:
Phase
Baseline
Training
Mean
Group:
Exp
4.80
7.00
5.90
Control
4.70
6.40
5.55
Mean
4.75
6.70
5.72
Subject means:
S1
S2
S3
S4
S5
S6
S7
S8
S9
S10
Grp
Exp
8.5
6.0
2.5
6.0
5.5
6.5
6.5
5.5
5.5
6.5
Control
4.0
5.0
9.0
3.5
4.0
8.0
7.5
4.5
5.0
5.5
X2 = 1501
X = 229
N = 40
n = 10
g = 2
p = 2
(
)
2
2
2
229
1501189.975
40
total
X
SSX
N
=-=-=
å
å
(
)
(
)
2
2
2
....
2[8.55.72...(5.55.72)]106.475
subjij
SSpXX
=S-
=-++-=
(
)
(
)
(
)
(
)
2
22
.....
28[5.905.725.555.72]1.225
groupk
SSpnXX
=S-
=-+-=
(
)
(
)
(
)
(
)
2
22
.....
210[4.755.726.705.72]38.025
phasej
SSgnXX
=S-
=-+-=
(
)
(
)
(
)
2
22
....
104.805.72...6.405.7239.875
cellsjk
SSnXX
=S-
éù
=-++-=
ëû
39.87538.0251.2250.925
PGcellsphasegroup
SSSSSSSS
=--=--=
Source
df
SS
MS
F
Between Subj
19
106.475
Groups
1
1.125
1.125
0.19
Ss w/in Grps
18
105.250
5.847
Within Subj
20
83.500
Phase
1
38.025
38.025
15.26*
P x G
1
0.625
0.625
0.25
P x Ss w/in Grps
18
44.850
2.492
Total
39
189.975
*
p
<
.
05
[
F
.
05
(
1
,
18
)
=
4
.
41
]
There is a significant difference between baseline and training,
but there are no group differences nor a group x phase
interaction.
14.4Independent t test on data in Exercise 14.3:
a.Difference scores (Training - Baseline)
Exper.
1
2
-1
2
7
1
3
-1
3
5
Control
2
0
2
3
-2
4
3
1
4
0
2
2
2
2
2.2 6.1778 10
1.7 3.7889 10
2.21.7
6.17783.7889
0.50
1010
EEE
EEE
EC
C
E
EC
Xsn
Xsn
XX
t
s
s
nn
===
===
-
-
==
+=
+
[t.025(9) = +2.262]
Do not reject H0
b.
t
2
=
.
50
2
=
.
25
=
F
for
P
x
G
interaction
.
c.The t is a test on whether the baseline vs. training
difference is the same for both groups. This is a test of an
interaction, not a test of overall group differences.
14.5Adding a No Attention control group to the study in Exercise
14.3:
Cell means:
Phase
Baseline
Training
Total
Group
Exp
4.8
7.0
5.90
Att Cont
4.7
6.4
5.55
No Att Cont
5.1
4.6
4.85
Total
4.87
6.00
5.43
Subject means:
S1
S2
S3
S4
S5
S6
S7
S8
S9
S10
Group:
Exp
8.5
6.0
2.5
6.0
5.5
6.5
6.5
5.5
5.5
6.5
Att Cont
4.0
5.0
9.0
3.5
4.0
8.0
7.5
4.5
5.0
5.0
No Att Cont
3.5
5.0
7.0
5.5
4.5
6.5
6.5
4.5
2.5
3.0
2
2026 326 60 10 3 2
XXNngp
======
åå
X = 326
N = 60
n = 10
g = 3
p = 2
(
)
2
2
2
326
2026254.7333
60
total
X
SSX
N
=-=-=
å
å
(
)
(
)
2
2
2
....
2[8.55.43...(3.05.43)]159.733
subjij
SSpXX
=S-
=-++-=
(
)
(
)
(
)
(
)
(
)
2
222
.....
28[5.905.435.555.434.855.43]11.433
groupk
SSpnXX
=S-
=-+-+-=
(
)
(
)
(
)
(
)
2
22
.....
310[4.875.436.005.43]19.267
phasej
SSgnXX
=S-
=-+-=
(
)
(
)
(
)
2
22
....
104.805.43...4.605.4352.333
cellsjk
SSnXX
=S-
éù
=-++-=
ëû
51.33319.26711.43320.633
PGcellsphasegroup
SSSSSSSS
=--=--=
Source
df
SSMSF
Between subj
29
159.7333
Groups
2
11.4333
5.7166
1.04
Ss w/ Grps
27
148.300
5.4926
Within subj
30
95.0000
Phase
1
19.2667
19.2667
9.44*
P * G
2
20.6333
10.3165
5.06*
P * Ss w/Grps
27
55.1000
2.0407
Total
59
254.733
*
p
<
.
05
[
F
.
05
(
1
,
27
)
=
4
.
22
;
F
.
05
(
2
,
27
)
=
3
.
36
]
b.Plot:
B
B
J
J
H
H
Baseline
Training
4
4.5
5
5.5
6
6.5
7
Phase
B
Exp
J
Att Cont
H
No Att Cont
c.There seems to be no difference between the Experimental and
Attention groups, but both show significantly more improvement than
the No Attention group.
14.6Summarization of stories by adult and children good and poor
readers:
Cell Means for Age * Readers * Items:
Adults
Items:
Setting
Goal
Disp
Mean
Good Readers
6.20
6.0
5.0
5.73
Poor Readers
5.40
4.8
2.0
4.07
Mean
5.80
5.40
3.50
4.90
Children
Items:
Setting
Goal
Disp
Mean
Good Readers
5.80
5.60
3.00
4.80
Poor Readers
3.00
2.40
1.20
2.20
Mean
4.40
4.00
2.10
3.50
Cell Means for Age * Readers:
Adults
Children
Mean
Good Readers
5.73
4.80
5.27
Poor Readers
4.07
2.20
3.13
Mean
4.90
3.50
4.20
Cell Means for Age * Items:
Adults
Children
Mean
Setting
5.80
4.40
5.10
Goal
5.40
4.00
4.70
Disposition
3.50
2.10
2.80
Mean
4.90
3.50
4.20
Cell Means for Reader * Items:
Good Readers
Poor Readers
Mean
Setting
6.00
4.20
5.10
Goal
5.80
3.60
4.70
Disposition
4.00
1.60
2.80
Mean
5.27
3.13
4.20
Subject Means:
Good Adult Readers:
7.00
5.00
5.00
7.00
4.67
Good Children Readers:
4.00
6.33
6.00
4.33
3.33
Poor Adult Readers:
5.33
3.00
4.67
3.00
4.33
Poor Children Readers:
2.00
1.00
3.33
3.33
1.33
(
)
2
2
2
252
1312253.600
60
total
X
SSX
N
=-=-=
å
å
(
)
(
)
2
2
2
....
3[7.004.20...(1.334.20)]164.933
subjijki
SSiXX
=S-
=-++-=
(
)
(
)
(
)
(
)
(
)
2
22
.......
325[4.904.203.504.20]29.400
agej
SSirnXX
=S-
=-+-=
(
)
(
)
(
)
(
)
(
)
2
22
.......
235[5.274.203.134.20]68.267
readeri
SSainXX
=S-
=-+-=
(
)
(
)
(
)
(
)
(
)
2
22
......
355.734.20...2.204.20100.933
cellsARij
SSinXX
=S-
éù
=-++-=
ëû
100.93329.40068.2673.267
ARcellsARagereader
SSSSSSSS
=--=--=
(
)
(
)
(
)
(
)
(
)
2
22
2
.......
225[5.104.204.704.20(2.804.20)]60.400
itemj
SSarnXX
=S-
=-+-+-=
(
)
(
)
(
)
(
)
2
22
......
255.804.20...2.104.2089.800
cellsAIjk
SSrnXX
=S-
éù
=-++-=
ëû
89.80029.40060.4000.00
AIcellsAIageitem
SSSSSSSS
=--=--=
(
)
(
)
(
)
(
)
2
22
.......
256.004.20...1.604.20129.600
cellsRIik
SSanXX
=S-
éù
=-++-=
ëû
129.60068.26760.4000.933
RIcellsRIreaderitem
SSSSSSSS
=--=--=
(
)
(
)
(
)
2
22
.....
56.204.20...1.204.20170.800
cellsARIijk
SSnXX
=S-
éù
=-++-=
ëû
170.80029.40068.26760.4003.2670.0000.933
8.533
ARIcellsARIagereaderitemARAIRI
SSSSSSSSSSSSSSSS
=------
=------=
14.7From Exercise 14.6:
a.Simple effect of reading ability for children:
(
)
(
)
(
)
(
)
2
22
35[4.803.502.203.50]50.70
RatCRatCC
SSinXX
=S-
=-+-=
50.70
50.70
1
RatC
RatC
RatC
SS
MS
df
===
Because we are using only the data from Children, it would be
wise not to use a pooled error term. The following is the relevant
printout from SPSS for the Between-subject effect of Reader.
Tests of Between-Subjects Effects
a
Measure: MEASURE_1
Transformed Variable: Average
367.500
1
367.500
84.483
.000
50.700
1
50.700
11.655
.009
34.800
8
4.350
Source
Intercept
READERS
Error
Type III Sum
of Squares
df
Mean Square
F
Sig.
AGE = Children
a.
b.Simple effect of items for adult good readers:
(
)
(
)
(
)
(
)
2
222
5[6.205.736.005.735.005.73]4.133
IatAGIatAGAG
SSnXX
=S-
=-+-+-=
Again, we do not want to pool error terms. The following is the
relevant printout from SPSS for Adult Good readers. The difference
is not significant, nor would it be for any decrease in the df if
we used a correction factor.
Tests of Within-Subjects Effects
Measure: MEASURE_1
Sphericity Assumed
4.133
2
2.067
3.647
.075
4.533
8
.567
Source
ITEMS
Error(ITEMS)
Type III Sum
of Squares
df
Mean Square
F
Sig.
14.8Within-groups covariance matrices for the data in Exercise
14.10:
S
ˆ
within
AG
=
1
.
70
1
.
25
1
.
00
1
.
25
2
.
50
1
.
25
1
.
00
1
.
25
1
.
00
S
ˆ
within
CG
=
1
.
70
1
.
90
1
.
25
1
.
90
3
.
30
1
.
50
1
.
25
1
.
50
1
.
00
S
ˆ
within
AP
=
1
.
30
1
.
10
0
.
75
1
.
10
1
.
70
1
.
00
0
.
75
1
.
00
1
.
00
S
ˆ
within
CP
=
2
.
00
2
.
00
0
.
25
2
.
00
2
.
80
0
.
40
0
.
25
0
.
40
0
.
70
S
ˆ
pooled
=
1
.
6750
1
.
5625
0
.
8125
1
.
5625
2
.
5750
1
.
0375
0
.
8125
1
.
0375
0
.
9250
14.9It would certainly affect the covariances because we would
force a high level of covariance among items. As the number of
responses classified at one level of Item went up, another item
would have to go down.
14.10Cigarette smoking quitting techniques:
a.Analysis:
Cell Means for Group * Time * Place:
Pre
Post
Home
Work
Home
Work
Mean
Taper
6.80
6.00
5.80
3.60
5.55
Immediate
7.00
6.20
5.80
4.80
5.95
Aversion
7.00
6.20
4.80
2.40
5.10
Mean
6.93
6.13
5.47
3.60
5.53
Means Group * Time:
Means Group * Place:
Pre
Post
Mean
Home
Work
Mean
Taper
6.40
4.70
5.55
6.30
4.80
5.55
Immediate
6.60
5.30
5.95
6.40
5.50
5.95
Aversion
6.60
3.60
5.10
5.90
4.30
5.10
Mean
6.53
4.53
5.53
6.20
4.87
5.53
Time * Place:
Pre
Post
Total
Home
6.93
5.47
6.20
Work
6.13
3.60
4.87
Tot
6.53
4.53
5.53
Subject * Time:
Pre
6.5
4.5
7.5
8.0
5.5
7.5
5.0
6.5
7.5
6.5
8.5
4.0
7.0
6.0
7.5
Post
5.0
3.5
5.5
5.5
4.0
6.5
4.5
5.5
5.5
4.5
4.5
2.5
4.0
2.5
4.5
Mean
5.75
4.00
6.50
6.75
4.75
7.00
3.75
6.00
6.50
5.50
6.50
3.25
5.50
4.25
6.00
Subject * Place:
Home
6.5
5.0
7.5
7.0
5.5
7.5
5.0
6.5
7.0
6.0
7.0
3.5
6.0
6.0
7.0
Work
5.0
3.0
5.5
5.5
4.0
6.5
4.5
5.5
6.0
5.0
6.0
3.0
5.0
2.5
5.0
(
)
2
2
2
332
2024186.933
60
total
X
SSX
N
=-=-=
å
å
(
)
(
)
(
)
(
)
2
2
2
.......
22[5.755.53...(6.005.53)]69.433
subjl
SStpXX
=S-
=-++-=
(
)
(
)
(
)
(
)
(
)
(
)
2
222
...
....
225[5.555.535.955.535.105.53]7.233
groupi
SStpnXX
=S-
=-+-+-=
(
)
(
)
(
)
(
)
(
)
2
22
.......
325[6.535.534.535.53]60.000
timej
SSgpnXX
=S-
=-+-=
(
)
(
)
(
)
2
22
......
(2)6.505.53...4.505.53143.933
cellsTSijl
SSpnXX
=S-
éù
=-++-=
ëû
143.93360.00069.43314.500
TScellsTStimesubj
SSSSSSSS
=--=--=
(
)
(
)
(
)
(
)
2
22
......
256.405.53...3.605.5375.133
cellsGTij
SSpnXX
=S-
éù
=-++-=
ëû
75.1337.23360.0007.900
GTcellsGTgrouptime
SSSSSSSS
=--=--=
(
)
(
)
(
)
(
)
(
)
2
22
.......
325[6.205.534.875.5326.667
placek
SSgtnXX
=S-
=-+-=
(
)
(
)
(
)
2
22
......
26.505.53...5.005.53104.933
cellsPSkl
SStXX
=S-
éù
=-++-=
ëû
104.93326.66769.4338.833
PScellsPSplacesubj
SSSSSSSS
=--=--=
(
)
(
)
(
)
(
)
2
22
.......
256.305.53...4.305.5335.333
cellsGPik
SStnXX
=S-
éù
=-++-=
ëû
35.3337.23326.6671.433
GPcellsGPgroupplace
SSSSSSSS
=--=--=
(
)
(
)
(
)
(
)
2
22
.......
356.935.53...3.605.5390.933
cellsTPjk
SSgnXX
=S-
éù
=-++-=
ëû
90.93360.00026.6674.267
TPcellsTPtimeplace
SSSSSSSS
=--=--=
186.933
cellsTPStotal
SSSS
==
186.93360.00026.66769.43314.5008.8334.26
73.233
TPScellsTPStimeplacesubjTSPSTP
SSSSSSSSSSSSSSSS
=------
=------=
(
)
(
)
(
)
2
22
.....
56.805.53...2.405.53108.933
cellsGTPijk
SSnXX
=S-
éù
=-++-=
ëû
108.9337.23360.00026.6677.9001.4334.2671
.433
GTPgrouptimeplaceGTGPTP
SSSSSSSSSSSSSS
=-----
=------=
Source
df
SS
MS
F
Between subj
14
69.433
Group
2
7.233
3.617
0.70
Ss w/in grp
12
62.200
5.183
Within subj
45
116.500
Time
1
60.000
60.000
109.09*
TxS
14
14.500
GxT
2
7.900
3.950
7.18*
GxTw/in grps
12
6.600
0.550
Place
1
26.667
26.667
43.24*
PxS
14
8.833
GxP
2
1.433
0.717
1.16
GxPw/in grps
12
7.400
0.617
TxP
1
4.267
4.267
28.44*
TxPxS
14
3.233
TxPxG
2
1.433
0.717
4.78*
TxPxSw/in grp
12
1.800
0.150
Total
59
186.933
*p < .05 [F.05(1,12) = 4.75; F.05(2,12) = 3.89]
b.There is a significant decrease in desire from Pre to Post and
a significant reduction at Work relative to at Home. There is also
a Time by Place interaction, with a greater Place difference after
treatment. The Time by Group interaction is the real test of our
hypothesis, showing that the Pre-Post difference depends on the
treatment group, with the greatest difference in the Aversion
condition. The 3-way interaction shows that the Time by Group
interaction itself interacts with Place.
14.11Plot of results in Exercise 14.10:
B
B
J
J
H
H
Pre
Post
3
3.5
4
4.5
5
5.5
6
6.5
7
Time
B
Taper
J
Immediate
H
Aversion
14.12I will look at the Group × Time interaction by looking at
the simple effect of Time for each Group. For a more complex
design, we should run separate analyses for each group, to avoid
problems with sphericity. However, with only two levels of the
repeated measures variables, we have only one off-diagonal
covariance, so we don’t have a problem with constant covariances. I
will test each by the same test term as was used to test the Group
× Time interaction, MSGxTxSs w/in groups (The pattern of
significance would not change with separate analyses. The Fs if we
used separate analyses would be 44.46, 18.78, and 51.43,
respectively.)
(
)
(
)
(
)
(
)
2
11
22
.....
256.405.554.705.5514.45
timeattaperj
SSpnXX
=S=
éù
=-+-=
ëû
14.45
14.45
1
14.45
26.27*
0.55
tatt
timeattaper
tatt
tatt
timeattaper
error
SS
MS
df
MS
F
MS
===
===
(
)
(
)
(
)
(
)
2
22
22
.....
256.605.955.305.958.45
timeatimmedj
SSpnXX
=S=
éù
=-+-=
ëû
8.45
8.45
1
8.45
15.36*
0.55
tati
timeatimmed
tati
tati
timeatimmed
error
SS
MS
df
MS
F
MS
===
===
(
)
(
)
(
)
(
)
2
33
22
256.605.103.605.1045.00
...
..
timeataverj
SSpnXX
=S=
éù
=-+-=
ëû
15.00
45.00
1
45.00
81.82*
0.55
tata
timeataver
tata
tati
timeatimmed
error
SS
MS
df
MS
F
MS
===
===
*
p
<
.
05
[
F
.
05
(
1
,
12
)
=
4
.
75
]
These SSSimple Effects sum to the same total as SSTime and SSG *
T:
14.45 + 8.45 + 45.00 = 67.90 = 60.00 + 7.90
Each of the methods led to a significant reduction in desire to
smoke. If we then look at the effect of Group at Post:
(
)
(
)
(
)
(
)
(
)
2
122
222
.....
25[4.704.535.304.533.604.53]14.867
groupatpost
SSpnXX
=S-
=-+-+-=
/
//
/
/
14.867
7.433
2
66.2006.600
3.033
1212
GxTxSswingrps
groupatpost
groupatpost
groupatpost
SswingrpsGxTxSswingrps
wincell
Sswingrps
SS
MS
df
SSSS
MS
dfdf
===
+
+
===
++
/
7.433
2.45
3.033
GatP
groupatpost
wincell
MS
F
MS
===
This would not be significant even for the maximum possible
value of f ', meaning that we do not have data to allow us to
recommend one method over the other. (If we had run a separate
analysis on just the Posttest data, the corresponding F would have
been 4.33, with p = .038.)
14.13Analysis of data in Exercise 14.5 by BMDP:
a.Comparison with results obtained by hand in Exercise 14.5.
b.The F for Mean is a test on H0: = 0.
c.MSw/in Cell is the average of the cell variances.
14.14An analysis of data in Exercise 14.6 by SPSS as if it were
a factorial:
From Exercise 14.6:
/ */
/
/ */
64.00018.800
1.725
1632
SswingroupsISswngroups
wincell
SswingroupsISswngroups
MSMS
MS
dfdf
+
+
===
++
This equals the MSResidual in the SPSS printout.
14.15Source column of summary table for 4-way ANOVA with
repeated measures on A & B and independent measures on C &
D.
Source
Between Ss
C
D
CD
Ss w/in groups
Within Ss
A
AC
AD
ACD
A x Ss w/in groups
B
BC
BD
BCD
B x Ss w/in groups
AB
ABC
ABD
ABCD
AB x Ss w/in groups
Total
14.16Analysis of Foa et al. (1991) data
All three groups decreased over time, but the Supportive
Counseling group decreased the least and the interaction was
significant.
14.17Using the mixed models procedure on data from Exercise
14.16
If we assume that sphericity is a reasonable assumption, we
could run the analysis with covtype(cs). That will give us the
following, and we can see that the F’s are the same as they were in
our analysis above.
However, the correlation matrix below would make us concerned
about the reasonableness of a sphericity assumption. (This matrix
is collapsed over groups, but reflects the separate matrices well.)
Therefore we will assume an autoregressive model for our
correlations.
These F values are reasonably close, but certainly not the
same.
14.18Standard analysis with missing data:
Notice that we have lost considerable degrees of freedom and our
F for Group is no longer significant
14.19Mixed model analysis with unequal size example.
Notice that we have a substantial change in the F for Time,
though it is still large.
14.20Analysis of Stress data:
Source
df
SS
MS
F
Pillai F
Prob
Between subj
97
137.683
Gender
1
7.296
7.296
5.64*
Role
1
8.402
8.402
6.49*
G * R
1
0.298
0.298
<1
Ss w/in Grps
94
121.687
1.294
Within subj
97
87.390
Time
1
1.064
1.064
1.23*
1.23
0.2700
T*G
1
0.451
0.451
<1
0.52
0.4720
T*R
1
0.001
0.001
<1
0.00
0.9708
T*G*R
1
4.652
4.652
5.38*
5.38
0.0225
T*Ss w/in grps
94
81.222
0.864
Total
194
103.386
*p < .05
The univariate and multivariate F values agree because we have
only two levels of each independent variable.
14.21Everitt’s study of anorexia:
a.SPSS printout on gain scores:
Tests of Between-Subjects Effects
Dependent Variable: GAIN
614.644
a
2
307.322
5.422
.006
732.075
1
732.075
12.917
.001
614.644
2
307.322
5.422
.006
3910.742
69
56.677
5075.400
72
4525.386
71
Source
Corrected Model
Intercept
TREAT
Error
Total
Corrected Total
Type III Sum
of Squares
df
Mean Square
F
Sig.
R Squared = .136 (Adjusted R Squared = .111)
a.
b.SPSS printout using pretest and posttest:
Tests of Within-Subjects Effects
Measure: MEASURE_1
Sphericity Assumed
366.037
1
366.037
12.917
.001
307.322
2
153.661
5.422
.006
1955.371
69
28.339
Source
TIME
TIME * TREAT
Error(TIME)
Type III Sum
of Squares
df
Mean Square
F
Sig.
c.The F comparing groups on gain scores is exactly the same as
the F for the interaction in the repeated measures design.
d.
TREAT: 1.00 Cognitive Behavioral
PRETEST
100
90
80
70
60
POSTTEST
110
100
90
80
70
EMBED StaticEnhancedMetafile
TREAT: 2.00 Control
PRETEST
100
90
80
70
POSTTEST
100
90
80
70
TREAT: 3.00 Family Therapy
PRETEST
100
90
80
70
POSTTEST
110
100
90
80
70
The plots show that there is quite a different relationship
between the variables in the different groups.
e.Treatment Group = Control
One-Sample Statistics
a
26
-.4500
7.9887
1.5667
GAIN
N
Mean
Std. Deviation
Std. Error Mean
Treatment Group = Control
a.
One-Sample Test
a
-.287
25
.776
-.4500
-3.6767
2.7767
GAIN
t
df
Sig. (2-tailed)
Mean
Difference
Lower
Upper
95% Confidence Interval
of the Difference
Test Value = 0
Treatment Group = Control
a.
This group did not gain significantly over the course of the
study. This suggests that any gain we see in the other groups
cannot be attributed to normal gains seen as a function of age.
f.Without the control group we could not separate gains due to
therapy from gains due to maturation.
14.22When multiple respondents come from the same family, their
data are not likely to be independent. We act as if we have 98
different respondents, but in fact we do not have 98 independent
respondents, which is important. If we had complete data from each
family we could treat Spouse and Patient as a repeated measures
variable—it is a "within-family" variable. Alternatively, we could
delete data so as to have only one respondent per family. In this
situation, numerous studies have shown that there is a remarkably
small degree of dependence between members of the same family, and
many people would ignore the problem entirely.
14.23t = -0.555. There is no difference in Time 1 scores between
those who did, and did not, have a score at Time 2.
b.If there had been differences, I would worried that people did
not drop out at random.to answer.
14.24Intraclass correlation:
(
)
(
)
(
)
(
)
1/
82.574.08
82.5724.08370.124.08/20
78.49
.85
82.578.166.30
subjectsIxS
subjectsJxSjudgeJxS
MSMS
IC
MSjMSjMSMSn
-
=
+-+-
-
=
++-
==
++
The remainder of this exercise raises some questions that anyone
interested in the reliability of their data (and we all should be)
needs to be prepared
14.25Differences due to Judges play an important role.
14.26I would leave the variability due to Judge out of my
calculations entirely.
14.27If I were particularly interested in differences between
subjects, and recognized that judges probably didn’t have a good
anchoring point, and if this lack was not meaningful, I would not
be interested in considering it.
14.28The fact that the “parent” who supplies the data changes
from case to case simply adds additional variability to our data,
and this variability is confounded with differences between
children
14.29Strayer et al. (2006)
b.Contrasts on means:
Because the variances within each condition are so similar, I
have used MSerror(within) as my error term. The means are 776.95,
778.95, and 849.00 for Baseline, Alcohol, and Cell phone
conditions, respectively..
2
12
13
23
2
12
*
13
*
23
ˆ
ˆ
776.95778.952
ˆ
776.95849.0072.05
ˆ
778.95849.0070.5
216303.709
28.551
40
2/28.5510.07
72.05/28.5512.52
70.05/28.5512.45
ierror
vs
vs
vs
ierror
vs
vs
vs
t
aMS
n
aMS
den
n
t
t
t
y
y
y
y
=
S
=-=
=-=
=-=
S
´
===
==
==
==
Both Baseline and Alcohol conditions show poorer performance
than the cell phone condition, but, interestingly, the Baseline and
Alcohol conditions do not differ from each other.
14.30 Study by Teri et al. (1997):
The following are the results from SPSS. As we would expect, the
interaction is significant. In fact, the F for the interaction is
exactly equal to the F for the Group effect on Change in Exercise
11-37, and the sum of squares for both the interaction and the
error term are exactly half of what they were in Exercise
11-37.
Tests of Within-Subjects Effects
Measure:MEASURE_1
Source
Type III Sum of Squares
df
Mean Square
F
Sig.
PrePost
Sphericity Assumed
164.360
1
164.360
32.557
.000
Greenhouse-Geisser
164.360
1.000
164.360
32.557
.000
Huynh-Feldt
164.360
1.000
164.360
32.557
.000
Lower-bound
164.360
1.000
164.360
32.557
.000
PrePost * Group
Sphericity Assumed
195.695
3
65.232
12.922
.000
Greenhouse-Geisser
195.695
3.000
65.232
12.922
.000
Huynh-Feldt
195.695
3.000
65.232
12.922
.000
Lower-bound
195.695
3.000
65.232
12.922
.000
Error(PrePost)
Sphericity Assumed
343.285
68
5.048
Greenhouse-Geisser
343.285
68.000
5.048
Huynh-Feldt
343.285
68.000
5.048
Lower-bound
343.285
68.000
5.048
Tests of Between-Subjects Effects
Measure:MEASURE_1
Transformed Variable:Average
Source
Type III Sum of Squares
df
Mean Square
F
Sig.
Intercept
26031.127
1
26031.127
911.469
.000
Group
42.771
3
14.257
.499
.684
Error
1942.048
68
28.560
Chapter 15 - Multiple Regression
15.1Predicting Quality of Life:
a.All other variables held constant, a difference of +1 degree
in Temperature is associated with a difference of –.01 in perceived
Quality of Life. A difference of $1000 in median Income, again all
other variables held constant, is associated with a +.05 difference
in perceived Quality of Life. A similar interpretation applies to
b3 and b4. Since values of 0.00 cannot reasonably occur for all
predictors, the intercept has no meaningful interpretation.
b.
Y
ˆ
=
5
.
37
-
.
01
(
55
)
+
.
05
(
12
)
+
.
003
(
500
)
-
.
01
(
200
)
=
4
.
92
c.
Y
ˆ
=
5
.
37
-
.
01
(
55
)
+
.
05
(
12
)
+
.
003
(
100
)
-
.
01
(
200
)
=
3
.
72
15.2A difference of +1 standard deviation in Temperature is
associated with a difference of ‑.438 standard deviations in
perceived quality of life, while a difference of +1 standard
deviation in Income is associated with about three quarters of a
standard deviation difference in perceived Quality of Life. A
similar interpretation can be made for the other variables, but in
all cases it is assumed that all variables are held constant except
for the one in question.
15.3The F values for the four regression coefficients would be
as follows:
F
1
=
b
1
s
b
1
2
=
-
0
.
438
0
.
397
2
=
1
.
22
F
2
=
b
2
s
b
2
2
=
0
.
762
0
.
252
2
=
9
.
14
F
3
=
b
3
s
b
3
2
=
0
.
081
0
.
052
2
=
2
.
43
F
4
=
b
4
s
b
4
2
=
-
0
.
132
0
.
025
2
=
27
.
88
I would thus delete Temperature, since it has the smallest F,
and therefore the smallest semi-partial correlation with the
dependent variable.
15.4Predicting Job Satisfaction:
a.
Y
ˆ
=
.
605
Respon
-
.
334
NumSup
+
.
486
Envir
+
.
070
Yrs
+
1
.
669
b.[.624 -.311 .514 .063]
15.5a.Envir has the largest semi-partial correlation with the
criterion, because it has the largest value of t.
b.The gain in prediction (from r = .58 to R = .697) which we
obtain by using all the predictors is more than offset by the loss
of power we sustain as p became large relative to N.
15.6Adjusted R2 for the data in Exercise 15.4:
est
R
*
2
=
1
-
(
1
-
R
2
)
(
N
-
1
)
(
N
-
p
-
1
)
=
1
-
(
1
-
.
4864
)
(
14
)
(
15
-
4
-
1
)
=
.
2810
15.7As the correlation between two variables decreases, the
amount of variance in a third variable that they share decreases.
Thus the higher will be the possible squared semi-partial
correlation of each variable with the criterion. They each can
account for more previously unexplained variation.
15.8The more highly two predictor variables are intercorrelated,
the more "substitutability" there is between them. Thus for one
data set variable X1 may receive the greater weight, but for a
second data set variable X2 may happen to "substitute" for X1,
leaving X1 with only a minor role to play.
15.9The tolerance column shows us that NumSup and Respon are
fairly well correlated with the other predictors, whereas Yrs is
nearly independent of them.
15.10For the data in Exercise 15.4:
Y
ˆ
=
.
605
Respon
-
.
334
NumSup
+
.
486
Envir
+
.
070
Yrs
+
1
.
669
Satisfaction
Y
ˆ
2
3.26
2
2.85
3
5.90
3
4.63
5
4.18
5
8.15
6
4.73
6
5.59
6
6.91
7
5.99
8
7.86
8
5.62
8
5.93
9
6.86
9
8.55
µ
µ
µ
(
)
(
)
0.1234
2.426
1.693
2.864
2.846
.697
2.4261.693
Y
Y
YY
YY
s
s
s
rR
=
=
=
===
15.11Using Y and
µ
Y
from Exercise 15.10:
(
)
2
ˆ
1
42.322
4.232 (also calculated by BMDP in Exerc
ise 15.4)
1541
residual
YY
MS
Np
-
=
--
==
--
å
15.12The effect of sample size on the multiple correlation,
using random data:
For N = 15, R2 = .173
For N = 10, R2 = .402
For N = 6, R2 = .498 Notice that the correlation increases as N
- p decreases.
For N = 5, R2 = 1.000 Here N is equal to the number of
variables.
For N = 4 the matrix is singular and there is no solution.
15.13Adjusted R2 for 15 cases in Exercise 15.12:
R
2
0
.
1234
=
.
173
est
R
*
2
=
1
-
(
1
-
R
2
)
(
N
-
1
)
(
N
-
p
-
1
)
=
1
-
(
1
-
.
173
)
(
14
)
(
15
-
4
-
1
)
=
-
.
158
Since a squared value cannot be negative, we will declare it
undefined. This is all the more reasonable in light of the fact
that we cannot reject H0:R* = 0.
15.14Using the first three variables from Exercise 15.4:
a.The squared semi-partial correlation is .32546 – .31654 =
.00892.
The squared partial correlation is .00892/(1 – .32546) =
.01305.
b.Venn diagram illustrating squared partial and semi-partial
correlations for Satisfaction predicted by Number Supervised,
partialling out Responsibility.
NumSup
Respon
R
0(2.1)
2
R
2
0.12
15.15Using the first three variables from Exercise 15.4:
a.Figure comparable to Figure 15.1:
b.
µ
Y
= 0.4067Respon + 0.1845NumSup + 2.3542
The slope of the plane with respect to the Respon axis (X1) =
.4067
The slope of the plane with respect to the NumSup axis (X2) =
.1845
The plane intersects the Y axis at 2.3542
15.16Predicting percentage of low-birthweight live births from
Vermont Health Planning statistics:
a.
R
0
.
2
=
.
6215
R
0
.
25
=
.
7748
R
0
.
2
54
=
.
8181
b.
r
2
0
(
5
.
2
)
=
r
2
05
.
2
(
1
-
R
2
0
.
2
)
=
(
-
.
59063
2
)
(
1
-
.
3862
)
=
.
2141
R
2
0
.
25
=
R
2
0
.
2
+
r
2
0
(
5
.
2
)
=
.
3862
+
.
2141
=
.
6003
=
.
7748
2
15.17It has no meaning in that we have the data for the
population of interest (the 10 districts).
15.18The gain in R2 is not sufficient to offset the gain in p
relative to N.
15.19It plays a major role through its correlation with the
residual components of the other variables.
15.20For the data in Exercise 15.16:
Y
µ
Y
1
µ
Y
2
6.1
34.2
4.48
7.1
37.3
5.12
7.4
43.9
5.57
6.3
36.6
4.74
6.5
41.0
5.22
5.7
29.6
3.87
6.6
33.0
4.51
8.1
43.0
5.40
6.3
30.6
4.32
6.9
41.3
5.48
µ
Y
1 = 1X2 + 1X4 – 3X5
µ
Y
2 = .10446X2 + .189720X4 – .29372X5
Using the approximate regression coefficients the correlation
between Y and
µ
Y
would be 0.793 instead of the 0.818 which would be calculated
from the regression of Y and the
µ
Y
obtained using the actual regression equation.
15.21Within the context of a multiple-regression equation, we
cannot look at one variable alone. The slope for one variable is
only the slope for that variable when all other variables are held
constant. The percentage of mothers not seeking care until the
third trimester is correlated with a number of other variables.
15.22Create set of data illustrating leverage, distance, and
influence.
15.23Create set of data examining residuals.
15.24Modeling depression
DepressT
=
0
.
614
PVLoss
-
0
.
164
SuppTotl
-
0
.
106
AgeAtLoss
+
59
.
961
R
2
=
.
2443
F
.
05
(
3
,
131
)
=
14
.
115
Both PVLoss and SuppTotl are significant predictors, but
AgeAtLoss is not.
15.25Rerun of Exercise 15.24 adding PVTotal.
b.The value of R2 was virtually unaffected. However, the
standard error of the regression coefficient for PVLoss increased
from 0.105 to 0.178. Tolerance for PVLoss decreased from .981 to
.345, whereas VIF increased from 1.019 to 2.900. (c) PVTotal should
not be included in the model because it is redundant with the other
variables.
15.26Vulnerability as a function of social support and age at
loss.
PVLoss
=
.
041
SuppTotl
+
.
164
AgeAtLos
+
17
.
578
R
2
=
.
0203
;
F
=
1
.
42
;
F
.
05
(
2
,
137
)
=
3
.
06
The relationship is not significant.
Neither of the predictors is significant.
15.27Path diagram showing the relationships among the variables
in the model.
SuppTotl
PVLoss
AgeAtLoss
DepressT
-.2361
.0837
.4490
.1099
-.0524
15.28The direct effect of SuppTotl is represented by the arrow
that goes from SuppTotl to DepressT. This is a semi-partial
relationship because PVLoss was in the model to begin with. The
indirect effect of SuppTotl is the path that runs from SuppTotl to
PVLoss and then runs from PVLoss to DepressT. The expected change
in DepressT for a one standard deviation increase in SuppTotl (all
other things equal) is the sum of the direct and indirect effects
as measured by the standardized coefficients. In this case it is
–.2361 + (.0837)(.4490) = –.1985. (Keep in mind that the indirect
path is far from significant, and it is discussed here just to
illustrate a point.)
15.29Regression diagnostics.
Case # 104 has the largest value of Cook's D (.137) but not a
very large Studentized residual (t = –1.88). When we delete this
case the squared multiple correlation is increased slightly. More
importantly, the standard error of regression and the standard
error of one of the predictors (PVLoss) also decrease slightly.
This case is not sufficiently extreme to have a major impact on the
data.
15.30Adding error to a predictor.
As we add error to a predictor we should expect to see that the
standard error of that predictor will increase and its significance
decrease. For my particular example the standard error actually
decreased slightly. The most noticeable results in this example
were a substantial decrease in the multiple correlation
coefficient, and a corresponding increase in the residual
variance.
15.31Logistic regression using Harass.dat:
The dependent variable (Reporting) is the last variable in the
data set.
I cannot provide all possible models, so I am including just the
most complete. This is a less than optimal model, but it provides a
good starting point. This result was given by SPSS.
Block 1: Method = Enter
Omnibus Tests of Model Coefficients
35.442
5
.000
35.442
5
.000
35.442
5
.000
Step
Block
Model
Step 1
Chi-square
df
Sig.
Model Summary
439.984
.098
.131
Step
1
-2 Log
likelihood
Cox & Snell
R Square
Nagelkerke
R Square
Classification Table
a
111
63
63.8
77
92
54.4
59.2
Observed
No
Yes
REPORT
Overall Percentage
Step 1
No
Yes
REPORT
Percentage
Correct
Predicted
The cut value is .500
a.
Variables in the Equation
-.014
.013
1.126
1
.289
.986
-.072
.234
.095
1
.757
.930
.007
.015
.228
1
.633
1.007
-.046
.153
.093
1
.761
.955
.488
.095
26.431
1
.000
1.629
-1.732
1.430
1.467
1
.226
.177
AGE
MARSTAT
FEMIDEOL
FREQBEH
OFFENSIV
Constant
Step 1
a
B
S.E.
Wald
df
Sig.
Exp(B)
Variable(s) entered on step 1: AGE, MARSTAT, FEMIDEOL, FREQBEH,
OFFENSIV.
a.
From this set of predictors we see that overall LR = 35.44,
which is significant on 5 df with a p value of .0000 (to 4 decimal
places). The only predictor that contributes significantly is the
Offensiveness of the behavior, which has a Wald of 26.43. The
exponentiation of the regression coefficient yields 0.9547. This
would suggest that as the offensiveness of the behavior increases,
the likelihood of reporting decreases. That’s an odd result. But
remember that we have all variables in the model. If we simply
predicting reporting by using Offensiveness, exp(B) = 1.65, which
means that a 1 point increase in Offensiveness multiplies the odds
of reporting by 1.65. Obviously we have some work to do to make
sense of these data. I leave that to you.
15.32Predicting Reporting from Marital Status:
I requested this because both variables are a dichotomy and it
is easy to see the odds ratios. If we set this up as a contingency
table using the CrossTabs procedure, we get the following from
SPSS:
MARSTAT * REPORT Crosstabulation
Count
84
84
168
90
85
175
174
169
343
Married
Not Married
MARSTAT
Total
No
Yes
REPORT
Total
Chi-Square Tests
.070
b
1
.791
.024
1
.876
.070
1
.791
.829
.438
.070
1
.792
343
Pearson Chi-Square
Continuity Correction
a
Likelihood Ratio
Fisher's Exact Test
Linear-by-Linear Association
N of Valid Cases
Value
df
Asymp. Sig.
(2-sided)
Exact Sig.
(2-sided)
Exact Sig.
(1-sided)
Computed only for a 2x2 table
a.
0 cells (.0%) have expected count less than 5. The minimum
expected count is 82.78.
b.
For married women, the odds of reporting are 84/84 = 1.00
For unmarried women, the odds of reporting are 85/90 = .944
The odds ratio is 1.059, which means that you are 1.059 times
more likely to report the offense if you are married. Put the other
way around, you are .944/1.00 = .944 times more likely to report
the offensive behavior if you are unmarried. Since the odds are
less than 1.00, being unmarried means that you are less likely to
report.
If we run the logistic regression, we obtain:
Classification Table
a
90
84
51.7
85
84
49.7
50.7
Observed
No
Yes
REPORT
Overall Percentage
Step 1
No
Yes
REPORT
Percentage
Correct
Predicted
The cut value is .500
a.
Variables in the Equation
-.057
.216
.070
1
.791
.944
.057
.344
.028
1
.868
1.059
MARSTAT
Constant
Step 1
a
B
S.E.
Wald
df
Sig.
Exp(B)
Variable(s) entered on step 1: MARSTAT.
a.
Notice that the exp(B) = .9444, which is exactly what you
obtained above for the odds ratio given that you are unmarried.
That was the point of the exercise. Notice also how the
classification table here matches with the one in the CrossTabs
procedure, except that the rows and columns are now labeled
“Observed” and “Predicted.”
15.33It may well be that the frequency of the behavior is tied
in with its offensiveness, which is related to the likelihood of
reporting. In fact, the correlation between those two variables is
.20, which is significant at p < .000. (I think my explanation
would be more convincing if Frequency were a significant predictor
when used on its own.)
15.34Malcarne’s data on distress in cancer patients,
a.Predicting Distress2 from Distress1 and BlamPer
Both predictors play a significant role in Distress2
b.We need to include Distress1 in our prediction because it is
very reasonable to assume that initial distress would relate to
later distress, and we want to control for that effect when looking
at the effect of BlamPer.
15.35BlamPer and BlamBeh are correlated at a moderate level (r =
.52), and once we condition on BlamPer by including it in the
equation, there is little left for BlamBeh to explain.
15.36I want students to think about what it means when we speak
of “capitalizing on chance.” They also should think about the fact
that stepwise regression is a very atheoretical way of going about
things, and perhaps theory should take more of a role.
15.37Make up an example.
15.38They should see no change in the interaction term when they
center the data, but they should see important differences in the
“main effects” themselves. Have them look at the matrix of
intercorrelations of the predictors.
15.39This should cause them to pause. It is impossible to change
one of the variables without changing the interaction in which that
variable plays a role. In other words, I can’t think of a sensible
interpretation of “holding all other variables constant” in this
situation.
15.40Testing mediation with Jose’s data.
(
)
(
)
(
)
(
)
(
)
222222222222
.478.022.321.017.017.022
0.000150.0123
.478.321
12.47
.012
ab
ab
abbaab
ab
sssss
z
s
bb
bb
bb
bb
=+-=+-
==
===
15.41Analysis of results from Feinberg and Willer (2011).
The following comes from using the program by Preacher and
Leonardelli referred to in the chapter. I calculated the t values
from the regression coefficients and their standard errors and then
inserted those t values in the program. You can see that the
mediated path is statistically significant regardless of which
standard error you use for that path.
15.42Guber’s results:
This is a computer analysis question and there is no fixed set
of answers.
Chapter 16 - Analyses of Variance and Covariance as General
Linear Models
16.1Eye fixations per line of text for poor, average, and good
readers:
a.Design matrix, using only the first subject in each group:
X
=
1
0
-
1
0
1
-
1
b.Computer exercise:
R
2
=
.
608
SS
reg
=
57
.
7333
SS
residual
=
37
.
2000
c.Analysis of variance:
X
1 = 8.2000
X
2 = 5.6
X
3 = 3.4
X
. = 5.733
n1 = 5n2 = 5
n3 = 5
N = 15
X = 86X 2 = 588
(
)
(
)
(
)
(
)
22
2
2
222
.
()86
58894.933
15
5[8.20005.7335.65.7333.45.733]
57.733
94.93357.73337.200
total
groupj
errortotalgroup
X
SSX
N
SSnXX
SSSSSS
S
=S-=-=
=S-=-+-+-
=
=-=-=
Source
df
SS
MS
F
Group
2
57.733
28.867
9.312*
Error
12
37.200
3.100
Total
14
94.933
*
p
<
.
05
[
F
.
05
(
2
,
12
)
=
3
.
89
]
16.2Continuing with the data in Exercise 16.1:
a.Treatment effects:
EMBED Equation.DSMT4
123'
11.1
22.2
8.2 5.6 3.4 5.733
8.25.7332.467
5.65.7330.133
XXXX
XXb
XXb
a
a
====
=-=-==
=-=-=-=
b.
EMBED Equation.DSMT4
22
57.733
.608
94.933
treatment
total
SS
R
SS
h
====
16.3Data from Exercise 16.1, modified to make unequal ns:
R
2
=
.
624
SS
reg
=
79
.
0095
SS
residual
=
47
.
6571
Analysis of variance:
X
1 = 8.2000
X
2 = 5.8571
X
3 = 3.3333
.
X
= 5.7968
n1 = 5n2 = 7n3 = 9N = 21X = 112X 2 = 724
(
)
(
)
(
)
(
)
2
2
2
2
222
.
()
112
724126.6666
21
5[8.20005.796875.85715.796893.33335.7968
]
79.0095
126.666679.009547.6571
total
groupjj
errortotalgroup
X
SSX
N
SSnXX
SSSSSS
=-=-=
=-=-+-+-
=
=-=-=
S
å
å
Source
df
SS
MS
F
Group
2
79.0095
39.5048
14.92*
Error
18
47.6571
2.6476
Total
20
126.6666
*
p
<
.
05
[
F
.
05
(
2
,
18
)
=
3
.
55
]
16.4Continuing with the data in Exercise 16.3:
a.Treatment effects:
123.
11.1
22.12
.0
8.2 5.8571 3.3333 5.7968
8.25.79682.4032
5.85715.79680.0603
5.7968
XXXX
XXb
XXb
Xb
a
a
====
=-=-==
=-=-==
==
b.
22
79.0095
.624
126.6666
treatment
total
SS
R
SS
h
====
16.5Relationship between Gender, SES, and Locus of Control:
a.Analysis of Variance:
SES
Low
Average
High
Mean
Gender
Male
12.25
14.25
17.25
14.583
Female
8.25
12.25
16.25
12.250
Mean
10.25
13.25
16.75
13.417
X = 644X 2 = 9418n = 8N = 48
(
)
(
)
(
)
(
)
(
)
(
)
(
)
(
)
(
)
(
)
22
2
2
22
2
222
2
...
...
..
()644
9418777.6667
48
38[14.58313.41712.25013.417]
65.333
28[10.2513.41713.2513.41716.7513.417]
338.6667
8[12.251
total
genderi
SESj
cellsij
X
SSX
N
SSsnXX
SSgnXX
SSnXX
S
=S-=-=
=S-=-+-
=
=S-=-+-+-
=
=S-=-
(
)
(
)
22
3.417...16.2513.417]422.6667
422.666765.3333338.666718.6667
777.6667422.6667355.0000
GScellsgenderSES
errortotalcells
SSSSSSSS
SSSSSS
++-=
=--=--=
=-=-=
Source
df
SS
MS
F
Gender
1
65.333
65.333
7.730*
SES
2
338.667
169.333
20.034*
G x S
2
18.667
9.333
1.104
Error
42
355.000
8.452
Total
47
777.667
*
p
<
.
05
[
F
.
05
(
1
,
42
)
=
4
.
08
;
F
.
05
(
2
,
42
)
=
3
.
23
]
b.ANOVA summary table constructed from sums of squares
calculated from design matrix:
(
)
(
)
(
)
(
)
(
)
(
)
,,,
,,,
,,,
422.6667357.333365.333
422.666784.0000338.667
422.6667404.00018.667
777.667
G
regareg
S
regareg
GS
regareg
totalY
SSSSSS
SSSSSS
SSSSSS
SSSS
abbbab
abbaab
abbab
=-=-=
=-=-=
=-=-=
==
The summary table is exactly the same as in part a (above).
16.6
SSSES = SSreg
a.This is because we have equal ns, and therefore the variables
are orthogonal—that is they do not account for overlapping portions
of the variance.
b.This will not be true with unequal ns.
16.7The data from Exercise 16.5 modified to make unequal ns:
(
)
(
)
(
)
(
)
,,
(,,)
,)
(,,)
,
(,,)
,
750.1951458.7285291.467
458.7285398.713560.015
458.7285112.3392346.389
458.7285437.633821
errorY
reg
Greg
reg
reg
reg
reg
reg
SSSSSS
SSSSSS
SSSSSS
SSSSSS
abab
abab
bab
abab
aab
abab
ab
=-=-=
=-=-=
=-=-=
=-=-=
.095
Source
df
SS
MS
F
Gender
1
60.015
60.015
7.21*
SES
2
346.389
173.195
20.80*
G x S
2
21.095
10.547
1.27
Error
35
291.467
8.328
Total
40
*
p
<
.
05
[
F
.
05
(
1
,
35
)
=
4
.
12
;
F
.
05
(
2
,
35
)
=
3
.
27
]
16.8
SSSES ≠ SSreg()
346.389 ≠ 379.3325
The two values are not the same because (as pointed out in
Exercise 16.6) they will not agree when there are unequal ns. In
Exercise 16.7 some of the variation accounted for by SES was shared
with Gender and the interaction, and thus was not included in SSSES
from Exercise 16.5.
16.9Model from data in Exercise 16.5:
1
.
1667
A
1
-
3
.
1667
B
1
-
0
.
1667
B
2
+
0
.
8333
AB
11
-
0
.
1667
AB
12
+
13
.
4167
Means:
SES (B)
Low
Avg
High
Gender (A)
Male
12.25
14.25
17.25
14.583
Female
8.25
12.25
16.25
12.250
¶
¶
0
111
112
223
11
11114
12
121
ˆ
..13.4167intercept
ˆ
..14.58313.41671.1667
ˆ
..10.2513.41673.1667
ˆ
..13.2513.41670.1667
..12.2514.58310.2513.14670.8337
Xb
AXb
BXb
BXb
ABABXb
ABA
m
a
b
b
ab
ab
====
=-=-==
=-=-=-=
=-=-=-=
=--+=--+==
=-
25
..14.2514.58313.2513.14670.1667
BXb
-+=--+=-=
16.10Model from data in Exercise 16.7:
1
.
2306
A
1
-
3
.
7167
B
1
+
0
.
3500
B
2
+
0
.
4778
AB
11
+
0
.
5444
AB
12
+
13
.
6750
Means:
SES (B)
Weighted
Means
Unweighted
Means
Low
Avg
High
Gender (A)
Male
11.617
15.800
17.250
15.105
14.906
Female
8.250
12.250
16.833
12.045
12.444
Weighted Means:
9.714
13.615
17.070
13.463
Unweighted:
9.958
14.025
17.043
13.675
With unequal ns we need to use the unweighted means in order to
reproduce the values found by the regression model.
¶
¶
0
111
112
223
11
11114
12
1212
ˆ
..13.475intercept
ˆ
..14.90613.6751.231
ˆ
..9.95813.6753.717
ˆ
..14.02513.6750.350
..11.66714.9069.95813.6750.478
..
Xb
AXb
BXb
BXb
ABABXb
ABABX
m
a
b
b
ab
ab
====
=-=-==
=-=-=-=
=-=-==
=--+=--+==
=--+=
5
15.80014.90614.02513.6750.544
b
--+==
16.11Does Method III really deal with unweighted means?
Means:
B1
B2
weighted
unweighted
A1
4
10
8.5
7.0
A2
10
4
8.0
7.0
weighted
8.0
8.5
8.29
unweighted
7.0
7.0
7.0
The full model produced by Method 1:
1111
ˆ
0.00.03.07.0
YABAB
=+-+
Effects calculated on weighted means:
¶
0
111
112
11
11113
ˆ
..8.29intercept
ˆ
..8.508.290.21
ˆ
..8.008.290.29
..4.008.508.008.294.21
Xb
AXb
BXb
ABABXb
m
a
b
ab
===¹
=-=-=¹
=-=-=¹
=--+=--+=-¹
Effects calculated on unweighted means:
¶
0
111
112
11
11113
ˆ
..7.00=intercept
ˆ
..7.007.000.00
ˆ
..7.007.000.00
..4.007.007.007.003.00
Xb
AXb
BXb
ABABXb
m
a
b
ab
===
=-=-==
=-=-==
=--+=--+=-=
These coefficients found by the model clearly reflect the
effects computed on unweighted means. Alternately, carrying out the
complete analysis leads to SSA = SSB = 0.00, again reflecting
equality of unweighted means.
16.12Venn diagram representing the sums of squares in Exercise
16.5:
SS(total)
16.13Venn diagram representing the sums of squares in Exercise
16.7:
SES
Sex
SxS
SS(total)
SS(error)
16.14SAS printout for the data in Exercise 16.7:
The SAS System 13:05 Saturday, November 18, 2000
The GLM Procedure
Dependent Variable: dv
Source DF Type I SS Mean Square F Value Pr > F
A 1 95.4511028 95.4511028 11.46 0.0018
B 2 342.1827043 171.0913522 20.55 <.0001
A*B 2 21.0946481 10.5473241 1.27 0.2944
Source DF Type II SS Mean Square F Value Pr > F
A 1 58.3013226 58.3013226 7.00 0.0121
B 2 342.1827043 171.0913522 20.55 <.0001
A*B 2 21.0946481 10.5473241 1.27 0.2944
Source DF Type III SS Mean Square F Value Pr > F
A 1 60.0149847 60.0149847 7.21 0.0110
B 2 346.3892120 173.1946060 20.80 <.0001
A*B 2 21.0946481 10.5473241 1.27 0.2944
Source DF Type IV SS Mean Square F Value Pr > F
A 1 60.0149847 60.0149847 7.21 0.0110
B 2 346.3892120 173.1946060 20.80 <.0001
A*B 2 21.0946481 10.5473241 1.27 0.2944
16.15Energy consumption of families:
a.Design matrix, using only the first entry in each group for
illustration purposes:
X
=
1
.
.
.
0
.
.
.
-
1
0
.
.
.
1
.
.
.
-
1
58
.
.
.
60
.
.
.
75
75
.
.
.
70
.
.
.
80
b.Analysis of covariance:
SS
reg
(
a
,
cov
,
a
c
)
=
2424
.
6202
SS
reg
(
a
,
cov
)
=
2369
.
2112
SS
residual
=
246
.
5221
=
SS
error
There is not a significant decrement in SSreg and thus we can
continue to assume homogeneity of regression.
SS
reg
(
a
)
=
1118
.
5333
SS
cov
=
SS
reg
(
a
,
cov
)
-
SS
reg
(
a
)
=
2369
.
2112
-
1118
.
5333
=
1250
.
6779
SS
reg
(
cov
)
=
1716
.
2884
SS
A
=
SS
reg
(
a
,
cov
)
-
SS
reg
(
cov
)
=
2369
.
2112
-
1716
.
2884
=
652
.
9228
Source
df
SS
MS
F
Covariate
1
1250.6779
1250.6779
55.81*
A (Group)
2
652.9228
326.4614
14.57*
Error
11
246.5221
22.4111
Total
14
2615.7333
*
p
<
.
05
[
F
.
05
(
1
,
11
)
=
4
.
84
;
F
.
05
(
2
,
11
)
=
3
.
98
]
16.16Exercise 16.15 expanded into a two-way analysis of
covariance.
a.Analysis of covariance:
First we will test for homogeneity of regression:
(
)
2
,,,cov,
(,,,cov,)
2
,,,cov
,,,cov
.8931
4288.5572
.8931
4283.9008
512.79916
c
regbc
reg
residual
R
SS
R
SS
SS
ababab
abaab
abab
abab
=
=
=
=
=
There is a nonsignificant decrement in attributable variation
(in fact, R2 is still the same to four decimal places!), so we will
take our second model as our full model.
(
)
(
)
(
)
(
)
(
)
(
)
(
)
,,cov
,,,cov,,cov
,,cov
,,,cov,,cov
,,cov
4111.2036
4283.90084111.2036172.6972
3163.0287
4283.90083163.02871120.8721
4275.6550
reg
B
regreg
reg
A
regreg
reg
AB
SS
SSSSSS
SS
SSSSSS
SS
SSSS
aab
ababaab
bab
ababbab
ab
=
=-=-=
=
=-=-=
=
=
(
)
(
)
(
)
(
)
(
)
,,,cov,,cov
,,
cov
,,,cov,,
4283.90084275.65508.2458
1979.1000
4283.90081979.10002304.8008
regreg
reg
regreg
SS
SS
SSSSSS
ababab
abab
abababab
-=-=
=
=-=-=
Source
df
SS
MS
F
Covariate
1
2304.8008
2304.8008
103.37*
A (Group)
2
1120.8721
560.4361
25.14*
B (Meter)
1
172.6972
172.6972
7.75*
AB
2
8.2458
4.1229
<1
Error
11
512.7992
22.2956
Total
14
4796.7000
*
p
<
.
05
[
F
.
05
(
1
,
11
)
=
4
.
84
;
F
.
05
(
2
,
11
)
=
3
.
98
]
b.Conclusions: After adjusting for last year’s usage, there are
significant differences between the time-of-day groups in terms of
usage and between those who could check on current usage (the
‘metered’ group) and those who could not. In particular, Group 3
appears to use more electricity than the other two groups, and
those with meters use less than those without meters. (A more
precise statement would require the calculation of adjusted means
as in Exercise 16.17.) There is no interaction between the two
independent variables.
16.17Adjusted means for the data in Exercise 16.16:
(The order of the means may differ depending on how you code the
group membership and how the software sets up its design matrix.
But the numerical values should agree.)
1211121
ˆ
7.90990.87862.40220.56670.13110.72606.37
40
YAABABABC
=-+-++++
Y
ˆ
11
=
-
7
.
9099
(
1
)
+
0
.
8786
(
0
)
-
2
.
4022
(
1
)
+
0
.
5667
(
1
)
+
0
.
1311
(
0
)
+
0
.
7260
(
61
.
3333
)
+
6
.
3740
=
41
.
1566
Y
ˆ
12
=
-
7
.
9099
(
1
)
+
0
.
8786
(
0
)
-
2
.
4022
(
-
1
)
+
0
.
5667
(
-
1
)
+
0
.
1311
(
0
)
+
0
.
7260
(
61
.
3333
)
+
6
.
3740
=
44
.
8276
Y
ˆ
21
=
-
7
.
9099
(
0
)
+
0
.
8786
(
1
)
-
2
.
4022
(
1
)
+
0
.
5667
(
0
)
+
0
.
1311
(
1
)
+
0
.
7260
(
61
.
3333
)
+
6
.
3740
=
49
.
5095
Y
ˆ
22
=
-
7
.
9099
(
0
)
+
0
.
8786
(
1
)
-
2
.
4022
(
-
1
)
+
0
.
5667
(
0
)
+
0
.
1311
(
-
1
)
+
0
.
7260
(
61
.
3333
)
+
6
.
3740
=
54
.
0517
Y
ˆ
31
=
-
7
.
9099
(
-
1
)
+
0
.
8786
(
-
1
)
-
2
.
4022
(
1
)
+
0
.
5667
(
-
1
)
+
0
.
1311
(
-
1
)
+
0
.
7260
(
61
.
3333
)
+
6
.
3740
=
54
.
8333
Y
ˆ
32
=
-
7
.
9099
(
-
1
)
+
0
.
8786
(
-
1
)
-
2
.
4022
(
-
1
)
+
0
.
5667
(
1
)
+
0
.
1311
(
1
)
+
0
.
7260
(
61
.
3333
)
+
6
.
3740
=
61
.
0333
(We enter 61.3333 for the covariate in each case, because we
want to estimate what the cell means would be if the observations
in those cells were always at the mean of the covariate.)
16.18Analysis of difference scores for data in Exercise
16.16:
Source
df
SS
MS
F
Group
2
1086.4667
543.2333
15.50*
Meter
1
197.6333
197.6333
5.64*
G(M
2
6.0667
3.0333
<1
Error
24
841.2000
35.0500
Total
29
2131.3667
*p < .05
16.19Klemchuk, Bond, & Howell (1990)
16.20Analysis of variance on Epinuneq.dat:
Model Includes:
SSregression
SSresidual
Dose, Interval, DxI
Dose, Interval
Dose, DxI
Interval, DxI
162.39512
150.78858
159.44821
45.29395
226.7619
Source
df
SS
MS
F
Dose
2
117.1112
58.5556
28.92*
Interval
2
2.9469
1.4734
0.73
DI
4
11.60654
2.9016
1.43
Error
112
226.7619
2.0247
Total
120
16.21Analysis of GSIT in Mireault.dat:
Tests of Between-Subjects Effects
Dependent Variable: GSIT
1216.924
a
5
243.385
2.923
.013
1094707.516
1
1094707.516
13146.193
.000
652.727
1
652.727
7.839
.005
98.343
2
49.172
.590
.555
419.722
2
209.861
2.520
.082
30727.305
369
83.272
1475553.000
375
31944.229
374
Source
Corrected Model
Intercept
GENDER
GROUP
GENDER * GROUP
Error
Total
Corrected Total
Type III Sum
of Squares
df
Mean Square
F
Sig.
R Squared = .038 (Adjusted R Squared = .025)
a.
Estimated Marginal Means
GENDER * GROUP
Dependent Variable: GSIT
62.367
1.304
59.804
64.931
64.676
1.107
62.500
66.853
63.826
1.903
60.084
67.568
62.535
.984
60.600
64.470
60.708
.858
59.020
62.396
58.528
1.521
55.537
61.518
GROUP
1
2
3
1
2
3
GENDER
Male
Female
Mean
Std. Error
Lower Bound
Upper Bound
95% Confidence Interval
16.22Analysis of covariance for Mireault’s data:
Tests of Between-Subjects Effects
Dependent Variable: GSIT
1441.155
a
6
240.192
2.844
.010
178328.902
1
178328.902
2111.844
.000
496.225
1
496.225
5.877
.016
303.986
1
303.986
3.600
.059
154.540
2
77.270
.915
.402
325.689
2
162.845
1.928
.147
24319.374
288
84.442
1163091.000
295
25760.529
294
Source
Corrected Model
Intercept
YEARCOLL
GENDER
GROUP
GENDER * GROUP
Error
Total
Corrected Total
Type III Sum
of Squares
df
Mean Square
F
Sig.
R Squared = .056 (Adjusted R Squared = .036)
a.
Estimated Marginal Means
GENDER * GROUP
Dependent Variable: GSIT
61.879
a
1.449
59.026
64.731
64.467
a
1.169
62.165
66.768
62.841
a
2.108
58.692
66.991
62.699
a
1.162
60.412
64.986
61.077
a
1.037
59.036
63.118
58.513
a
1.650
55.265
61.762
GROUP
1
2
3
1
2
3
GENDER
Male
Female
Mean
Std. Error
Lower Bound
Upper Bound
95% Confidence Interval
Evaluated at covariates appeared in the model: YEARCOLL =
2.68.
a.
16.23Analysis of variance on the covariate from Exercise
16.22.
The following is abbreviated SAS output.
General Linear Models Procedure
Dependent Variable: YEARCOLL
Sum ofMean
SourceDFSquaresSquareF ValuePr > F
Model513.34776452.66955292.150.0600
Error292363.00122881.2431549
Corrected Total297376.3489933
R-SquareC.V.Root MSE
YEARCOLL Mean
0.03546641.532581.11497
2.6845638
SourceDFType III SSMean SquareF ValuePr > F
GENDER15.950062995.950062994.790.0295
GROUP20.780704310.390352160.310.7308
GENDER*GROUP22.962723101.481361551.19 0.3052
GENDERGROUPYEARCOLL
LSMEAN
112.27906977
122.53225806
132.68421053
212.88888889
222.85000000
232.70967742
These data reveal a significant difference between males and
females in terms of YearColl. Females are slightly ahead of males.
If the first year of college is in fact more stressful than later
years, this could account for some of the difference we found in
Exercise 16.21.
16.24Analysis of Everitt’s data:
a.Analysis on Gain scores:
Dependent Variable: Gain
Sum of
Source DF Squares Mean Square F Value Pr > F
Model 2 614.643667 307.321833 5.42 0.0065
Error 69 3910.742444 56.677427
Corrected Total 71 4525.386111
R-Square Coeff Var Root MSE Gain Mean
0.135821 272.3858 7.528441 2.763889
Source DF Type III SS Mean Square F Value Pr > F
Group 2 614.6436669 307.3218334 5.42 0.0065
b.Analysis of Post scores ignoring Pre scores:
Dependent Variable: Post
Sum of
Source DF Squares Mean Square F Value Pr > F
Model 2 918.986916 459.493458 8.65 0.0004
Error 69 3665.057528 53.116776
Corrected Total 71 4584.044444
R-Square Coeff Var Root MSE Post Mean
0.200475 8.556928 7.288126 85.17222
Source DF Type III SS Mean Square F Value Pr > F
Group 2 918.9869160 459.4934580 8.65 0.0004
c.Analysis of Post scores with Pre score as covariate
Dependent Variable: Post
Sum of
Source DF Squares Mean Square F Value Pr > F
Model 3 1272.781825 424.260608 8.71 <.0001
Error 68 3311.262620 48.695039
Corrected Total 71 4584.044444
R-Square Coeff Var Root MSE Post Mean
0.277655 8.193027 6.978183 85.17222
Source DF Type III SS Mean Square F Value Pr > F
Group 2 766.2728128 383.1364064 7.87 0.0008
Pre 1 353.7949086 353.7949086 7.27 0.0089
d.The analysis of gain scores and the analysis of covariance ask
similar questions, but they would only agree if the relationship
between Pre and Post had a slope of 1. In general the analysis of
covariance will be better. The analysis of Post scores is
confounded, because we can’t discriminate between effects of the
treatments and any pre-existing group differences. In addition, the
analysis of covariance would generally be better because it would
adjust the error term.
e.Because there is variability in the pretest scores that has
little to do with weight, I would be tempted to remove that from
SStotal to get a more appropriate denominator for h2. If I did that
would be 766.27/(4584.04-353.795) = 766.27/4230.245 = .18, which
strikes me as a far amount of explained variance given all of the
factors that influence weight. (A case might well be made for
leaving in the variation due to pretest scores for the same reasons
given when talking about matched sample t.)
f.When we leave out the Control group we can run a standard
analysis of covariance between the remaining groups adjusted for
the pretest. This gives us adjusted means of 85.70 and 90.198 for
CogBehav and Family groups, respectively. The standard deviations
of the groups are close, and a weighted average of the variances is
72.63, which gives a standard deviation of 8.5225. Then, using
adjusted means,
12
85.7091.198
0.625
8.5225
XX
d
s
-
-
===-
Thus after we adjust for pretest weights, the Family Therapy
group gained about two thirds of a standard deviation more than the
Cognitive Behavior Therapy group.
16.25Everitt compared two therapy groups and a control group
treatment for anorexia. The groups differed significantly in
posttest weight when controlling for pretest weight (F = 8.71, p
< .0001, with the Control group weighing the least at posttest.
When we examine the difference between just the two treatment
groups at posttest, the F does not reach significant, F = 3.745, p
= .060, though the effect size for the difference between means
(again controlling for pretest weights) is 0.62 with the Family
Therapy group weighing about six pounds more than the
Cognitive/Behavior Therapy group. It is difficult to k