Académique Documents
Professionnel Documents
Culture Documents
Section 12-1
12-1.
223
553
10
X y = 43550.8
104736.8
b)
c)
12-5.
12-7.
171.054
Predictor
Coef
StDev
T
P
Constant
33.449
1.576
21.22
0.000
xl
-0.054349
0.006329
-8.59
0.000
x6
1.0782
0.6997
1.54
0.138
S = 2.834
R-Sq = 82.9%
R-Sq(adj) = 81.3%
Analysis of Variance
Source
DF
SS
MS
F
P
Regression
2
856.24
428.12
53.32
0.000
Error
22
176.66
8.03
Total
24
1032.90
a)
b)
c)
2 = 8.03
y = 33.4491 0.05435(300) + 1.07822(2) = 19.30 mpg.
Predictor
Constant
Xl
X2
S = 12.35
Coef
SE Coef
T
P
383.80
36.22
10.60
0.002
-3.6381
0.5665
-6.42
0.008
-0.11168
0.04338
-2.57
0.082
R-Sq = 98.5%
R-Sq(adj) = 97.5%
Analysis of Variance
Source
DF
Regression
2
Residual Error
3
Total
5
SS
29787
458
30245
MS
14894
153
F
97.59
P
0.002
d) Predictor
Coef
SE Coef
T
P
Constant
484.0
101.3
4.78
0.041
Xl
-7.656
3.846
-1.99
0.185
X2
-0.2221
0.1129
-1.97
0.188
X1*X2
0.004087
0.003871
1.06
0.402
S = 12.12
R-Sq = 99.0%
R-Sq(adj) = 97.6%
Analysis of Variance
Source
DF
SS
MS
F
Regression
3
29951.4
9983.8
67.92
Residual Error
2
294.0
147.0
Total
5
30245.3
P
0.015
2 = 147.0 , se( 0 ) = 101.3 , se( 1 ) = 3.846 , se( 2 ) = 0.113 and se( 12 ) = 0.0039
y = 484 . 0 7 . 656 ( 25 ) 0 . 222 (1000 ) 0 . 0041 ( 25 )(1000 ) = 31 . 3
e)
f)
Predictor
Constant
xl
x2
x3
Coef
47.17
-9.735
0.4283
18.237
S = 3.480
SE Coef
49.58
3.692
0.2239
1.312
R-Sq = 99.4%
Analysis of Variance
Source
DF
Regression
3
Residual Error
16
Total
19
T
0.95
-2.64
1.91
13.90
P
0.356
0.018
0.074
0.000
R-Sq(adj) = 99.3%
SS
30532
194
30725
MS
10177
12
F
840.55
P
0.000
a)
y = 47
. 174 .9 7352 x1 + .0 4283x 2 + 18
. 2375x 3
b)
2 = 12
se( 0 ) = 49.5815 , se( 1 ) = 3.6916 , se ( 2 ) = 0 . 2239 , and se( 3 ) = 1.312
c)
y = 47
. 174 .9 7352
( 14.5) + 0.4283(220) + 18.2375(5) = 9143
.
d)
Section 12-2
H 0 : 1 = 2 = ... = k = 0
H 1 : j 0 for at least one j
(1916) 2
= 4490
10
yi 1030
SST = 371595.6
1916
' X' y = [171.054 3.713 1.126] 43550.8 = 371535.9
104736.8
19162
SS R = 371535.9
= 4430.38
10
SS E = SST SS R = 4490 4430.38 = 59.62
f0 =
SS R
k
SS E
n p
4430.38 / 2
= 260.09
59.62 / 7
f 0.05, 2, 7 = 4.74
f 0 > f 0.05, 2, 7
Reject H0 and conclude that the regression model is significant at = 0.05.
b)
H 0 : 1 = 0
2 = 0
H 1 : 1 0
1
t0 =
se( )
2 0
2
se( 2 )
1.126
=
= 13.08
0.0861
t0 =
3.713
=
= 19.20
0.1934
t / 2,7 = t.025,7 = 2.365
Reject H0
Reject H0
Both regression coefficients are significant
12-17.
a)
H 0 : 1 = 6 = 0
H1 : at least one 0
f 0 = 53.3162
f , 2, 22 = f .05, 2, 22 = 3.44
f 0 > f , 2, 22
Reject H0 and conclude regression model is significant at = 0.05
b)
H 0 : 1 = 0
H 1 : 1 0
t0 = 8.59
H0 : 6 = 0
H1 : 6 0
= 0.05
t 0 = 1.5411
| t0 |>/ t / 2, 22 ,
Do not reject H0, conclude that evidence is not significant to state 6 is significant at = 0.05.
12-21. a) H 0 : 1 = 2 = 12 = 0
H1 at least one
j 0
= 0.05
f 0 = 67.92
f ,3, 2 = f.05,3, 2 = 19.16
f 0 >/ f ,3, 2
Reject H0
b) H
: 12 = 0
H 1 : 12 0
= 0.05
2 = 147.0
2 (no interaction term) = 153.0
MSE ( 2 ) was reduced in the interaction term model due to the addition of this term.
12-23.
a) H 0 : 1 = 2 = 3 = 0
H1 : j 0
for all j
f 0 = 840.55
f.05,3,16 = 3.24
f 0 > f ,3,16
Reject H0 and conclude regression is significant at = 0.05
b) = 0.05
t / 2 , n p = t. 025,16 = 2. 12
H 0 : 1 = 0
2 = 0
3 = 0
3 0
H1: 1 0
2 0
t0 = 1.91
| t 0 | >/ t / 2 ,16
t 0 = 13. 9
| t 0 | > t / 2 ,16
Reject H0
Do not reject H0
Reject H0
t 0 = 2.637
| t 0 | > t / 2 ,16
a)
b)
c)
0.00657 8 0.00122
12-29.
a) 95 % CI on coeficients
1 t a / 2,n p ( 1 )
0.0972 1 1.4174
1.9646 2 17.0026
1.7953 3 6.7613
1.7941 4 0.8319
b) Y | x = 290.44
0
se( Y | x0 ) = 7.61
t.025, 7 = 2.365
Y | x0 t / 2,n p se( Y | x0 )
290.44 (2.365)(7.61)
272.44 Y | x0 308.44
y 0 t / 2,n p 2 (1 + x 0 ( XX) 1 x 0 )
c)
290.44 2.365(14.038)
257.25 y0 323.64
12-31
0.595 2 0.535
0.229 3 0.812
0.216 4 0.013
7.982 5 2.977
b) Y | x = 8.99568
0
se( Y | x0 ) = 0.472445
t .025,14 = 2.145
Y | x0 t / 2, n p se( Y | x0 )
8.99568 (2.145)(0.472445)
7.982 Y | x0 10.009
se( y 0 ) = 1.00121
8.99568 2.145(1.00121)
6.8481 y0 11.143
c) y0 = .8 99568
12-35.
b)
c)
a)
d) The simple linear regression model has the shorter interval. Yes, the simple linear regression model in this case is
preferable.
Section 12-5
12-37.
a) r = 0.82897
b) Normality assumption appears valid.
Residual
-5
-2
-1
Normal Score
Residuals Versus xl
(response is y)
(response is y)
Residual
Residual
-5
-5
1
x6
100
200
300
400
500
xl
Residual
-5
10
20
30
Fitted Value
12-39.
a) r 2 = 0.985
b) r 2 = 0.990
r2 increases with addition of interaction term. No, adding additional regressor will always increase r2
12-41
a) There is some indication of nonconstant variance since the residuals appear to fan out with increasing
values of y.
Residuals
5
2
-1
-4
-7
0
40
80
120
160
200
240
Predicted
b)
Source
Model
Error
Total (Corr.)
Sum of Squares
30531.5
193.725
30725.2
R-squared = 0.993695
R-squared (Adj. for d.f.) = 0.992513
DF
3
16
19
Mean Square
10177.2
12.1078
F-Ratio
840.546
P-value
.0000
R 2 = 0 . 9937 or 99.37 %;
2
= 0 . 9925 or 99.25%;
R Adj
c)
y = 6.22489 016647
.
x1 0.000228x 2 + 0.157312 x 3
d)
Residual Plot for log(y)
0.1
Residuals
0.05
0
-0.05
-0.1
-0.15
4
4.3
4.6
4.9
5.2
5.5
Predicted
Residuals
0.05
0
-0.05
-0.1
-0.15
3.3
5.3
7.3
9.3
11.3
x3
Residuals
0.05
0.02
-0.01
-0.04
-0.07
3.8
4.1
4.4
4.7
5.3
Predicted
Using 1/x3
Section 12-6
2
12-47.
Residuals Versus x
(response is y)
(response is y)
1
Residual
Residual
-1
-1
0.0
0.5
1.0
1.5
2.0
2.5
(response is y)
1
-1
-1
-7
-6
-5
-4
-3
Fitted Value
Residual
Normal Score
-2
-1
12-49.
x' =
xx
Sx
285 297.125
= 1.016
11.9336
y = 759.395 90.783(1.106) 47.166(1.106) 2 = 802.943
a) At x = 285
b)
x' =
297.125
y = 759.395 90.783( x11
.9336 ) 47.166(
psi
x 297.125 2
11.9336
12-51
a) Predictor
Constant
xl
x2
x3
x1x2
x1x3
x2x3
x1^2
x2^2
x3^3
Coef
-1.769
0.4208
0.2225
-0.12800
-0.01988
0.009151
0.002576
-0.01932
-0.00745
0.000824
S = 0.06092
where
SE Coef
1.287
0.2942
0.1307
0.07025
0.01204
0.007621
0.007039
0.01680
0.01205
0.001441
R-Sq = 91.7%
y y
Sy
y' =
T
-1.37
1.43
1.70
-1.82
-1.65
1.20
0.37
-1.15
-0.62
0.57
and x' =
x x
Sx
P
0.188
0.172
0.108
0.087
0.118
0.247
0.719
0.267
0.545
0.575
R-Sq(adj) = 87.0%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
9
16
25
SS
0.655671
0.059386
0.715057
MS
0.072852
0.003712
F
19.63
P
0.000
f 0 = 19.628
f.05,9,16 = 2.54
f 0 > f ,9,16
Reject H0 and conclude that the model is significant at = 0.05
c) Model is acceptable.
d)
H 0 : 11 = 22 = 33 = 12 = 13 = 23 = 0
SS R ( 12 , 13 , 23 , 11 , 22 , 33 | 1 , 2 , 3 , 0 ) = SS R ( 1 , 2 , 3 , 12 , 13 , 23 , 11 , 22 , 33 | 0 )
SS R ( 1 2 3 | 0 )
= 0.65567068 0.619763
= 0.0359
y = 0 + 1 x1 + 2 x 2 + 3 x 3
Reduced Model:
12-55.
C p = 5.02
b)
MS E = 7.97
y = 34.434 0.048x1
MSE = 8.81
C p = 5.55
c) Same as part b.
d) y = 0.341 + 2.862 x 5 + 0.246 x 8 0.010 x10
MS E = 7.97
C p = 5.02
e) Minimum Cp and backward elimination result in the same model. Stepwise and forward selection
result in the same model. Because it is much smaller, the minimum Cp model seems preferable.
12-61.
a) Min. Cp
y = 3.517 + 0.486 x1 0.156 x9
C p = 1.67
b) Min MSE model is x1,
Supplemental Exercises
12-65.
a) H 0 : 3 = 4 = 5 = 0
for at least one j
H1 : j 0
= 0.01
f0 = 1323. 62
f.01, 3, 36 = 4 . 38
f0 >> f , 3, 36
b) = 0.01
H0 :3
H1: 3
t.005, 36 = 2 . 72
=0
0
t0 = 1. 32
| t 0 | >/ t / 2 , 36
Do not reject H0
Only regressor x4 is significant
H0:4 = 0
H1:4 0
H 0 :5 = 0
H1:5 0
t 0 = 19. 97
| t 0 | > t / 2 , 36
t 0 = 2 . 48
| t 0 | >/ t / 2 , 36
Reject H0
Do not reject H0
c) Curvature is evident in the residuals vs. regressor plots from this model.
12-67.
H1 : j 0
= 0.05
f 0 = 109.02
f.05, 4,19 = 2.90
f 0 >> f , 4,19
t. 025,19 = 2 . 093
H 0 : 1 = 0
H0 : 2 = 0
H 0 : 3 = 0
H0 : 4 = 0
H1 : 2 0
H1 : 3 0
H1 : 4 0
H 1 : 1 0
t 0 = 11.27
t0 = 14.59 t 0 = 6. 98
| t 0 | > t / 2 ,19
| t 0 | > t / 2 ,19
Reject H0
Reject H0
t 0 = 8. 11
| t 0 | > t / 2 ,19
| t 0 | > t / 2 ,19
Reject H0
Reject H0
c) The residual plots are more pleasing than those in Exercise 12-66.
12-69.
MS E ( p) = 694.93
C p = 5.62
MS E ( p) = 714.20
C p = 5.57
c) Same as model b.
d) Models from parts b. and c. are identical. Model in part a. is the same with x4 added in.
MSE model in part a. = 694.93 C p = 5.62
MSE model in parts b.&c. = 714.20 C p = 5.57
12-71.
a) VIF(3 ) = 51.86
VIF ( 4 ) = 9.11
VIF ( ) = 28.99
5
12-73.
a)
R2 =
SS R
SST
f0 =
0.47 / 6
SS R / k
=
= 18.28
SS E / n p 0.03 / 7
f .05, 6, 7 = 3.87
f 0 > f , 6 , 7
Reject H0.
R 2 = 0.92
SS R ' = R 2 ( SST ) = 0.92(0.50) = 0.46
b) k = 5 n = 14
p=6
f0 =
= 0.47 0.46
= 0.01
SS R ( j | i ,i =1, 2, , 6,i j | 0 ) / r
SS E ' /(n p )
0.01 / 1
=2
0.04 / 8
f .05,1,8 = 5.32
f 0 >/ f ,1,8
Do not reject H0 and conclude that the evidence is insufficient to claim that the removed variable is significant at
= 0.05
c)
MS E (reduced ) =
MS E ( full ) =
0.04
SS E
=
= 0.005
8
n p
0.03
= 0.00429
7
No, the MSE is larger for the reduced model, although not by much. Generally, if adding a variable to a model
reduces the MSE it is an indication that the variable may be useful in explaining the response variable. Here the
decrease in MSE is not very great because the added variable had no real explanatory power.