Professional Documents
Culture Documents
Analyzing Quantitative
Data: Multivariate
Statistics
The basic linear regression equation is The formulas for these two components are as
follows:
Y¿ a bX
xy
where Y prediced value of variable Y b a Y bX
a intercept constant x2
b regression coefficient (slope of the where b regression coefficient
line) a intercept constant
X actual value of variable X Y mean of variable Y
Regression analysis solves for a and b, so for X mean of variable X
any value of X, a prediction about Y can be made. x deviation scores from X
You may remember from high school algebra that y deviation scores from Y
the preceding equation is the algebraic equation for
a straight line. Linear regression is used to deter- Calculations for the current example are pre-
mine a straight-line fit to the data that minimizes sented in columns 3 through 7 of Table 21-1. As
deviations from the line. shown at the bottom of the table, the solution to the
As an illustration, consider the data for five regression equation is Y 1.5 .9X. Now sup-
subjects on two variables (X and Y) shown in pose for the moment that the X values in column 1
columns 1 and 2 of Table 21-1. These two vari- are the only data we have, and we want to predict
ables are strongly related (r .90). We use the five values for Y. For the first subject, X 1; we would
pairs of X and Y values to solve for a and b in a predict that Y 1.5 (.9)(1), or 2.4. Column 8
regression equation, which will allow us to predict shows Y values for each X. These numbers show
Y values for a new group of people about whom we that Y does not exactly equal the actual values
will have information on variable X only. obtained for Y (column 2). Most errors of
(1) (2) (3) (4) (5) (6) (7) (8) (9) (10)
X Y 2 y y2 y Y e e2
prediction (e) are small, as shown in column 9. the X axis, we must go up 3.6 units (.9 4) on the
Errors of prediction occur because the correlation Y axis. The line, then, embodies the regression
between X and Y is not perfect. Only when r 1.00 equation. To predict a value for Y, we would go to
or 1.00 does Y Y. The regression equation the point on the X axis for an obtained X value, go
solves for a and b in a way that minimizes such up to vertically to the point on the regression line
errors. More precisely, the solution minimizes the directly above the X score, and then read the pre-
sums of squares of prediction errors, which is why dicted Y value horizontally on the Y axis. For
standard regression analysis is said to use a least- example, for an X value of 5, we would predict a Y
squares criterion. Indeed, standard regression is of 6, indicated by the star.
sometimes referred to as ordinary least squares, The correlation coefficient, it may be recalled,
or OLS, regression. In column 10 of Table 21-1, is an index of the degree to which variables are
the error terms—referred to as the residuals—have related. From a statistical point of view, relation-
been squared and summed to yield a value of 7.60. ships are concerned with how variation in one vari-
Any values of a and b other than 1.5 and .9 would able is associated with variation in another. The
have yielded a larger sum of squared residuals. square of r (r2) tells us the proportion of variance
Figure 21-1 displays the solution to this in Y that can be accounted for by X. In our exam-
regression analysis graphically. The actual X and Y ple, r .90, so r2 .81. This means that 81% of
values are plotted on the graph with circles. The the variability in Y values can be understood in
line running through these points represents the terms of variability in X values. The remaining
regression solution. The intercept (a) is the point at 19% is variability resulting from other sources.
which the line crosses the Y axis, which in this case In the regression problem, the sum of the
is 1.5. The slope (b) is the angle of the line. With squared residuals (es) is 7.6. Column 6 of Table
b .90, the line slopes so that for every 4 units on 21-1 also tells us that the sum of the squared
deviations of Y values from the mean of Y (y2) is applicants would have made successful students. It
40, which represents variability in Y. To demon- may be possible, by using additional information,
strate that the residuals are 19% of the unexplained to make more accurate predictions through multi-
variability in Y, we compute the following ratio: ple regression. The basic multiple regression equa-
tion is
7.6
.19 1 r 2
40 Y¿ a b1X1 b2X2 p bkXk
Although the computations are not shown, the where Y predicted value for variable Y
sum of the squared deviations of the Y values from a intercept constant
the mean of Y (y2) equals 32.4. As a proportion k number of predictor (independent)
of the variability in the actual Y values, this would variables
equal b1 to bk regression coefficients for the
k variables
32.4
.81 r 2 X1 to Xk scores or values on the k independent
40 variables
These calculations reinforce a point made ear-
When the number of predictor variables exceeds
lier: The stronger the correlation, the better the pre-
two, computations to solve this equation are very
diction; the stronger the correlation, the greater the
complex and so we restrict our discussion to hypo-
percentage of variance explained.
thetical rather than worked-out examples.
In our example of predicting graduate nursing
M U LT I P L E L I N E A R students’ GPAs, suppose we hypothesized that
REGRESSION information on undergraduate GPA (GPA-U) and
scores on the quantitative portion of the GRE
Because the correlation between two variables is (GRE-Q) would improve our ability to predict
rarely perfect, researchers often try to improve pre- graduate GPA. The resulting equation might be
dictions of Y by including more than one indepen-
dent (predictor) variable. Multiple regression is Y .4 .05(GPA-U)
used for this purpose. .003(GRE-Q) .002(GRE-V)
For instance, suppose an applicant had a GRE-V
Basic Concepts for Multiple score of 600, a GRE-Q score of 550, and a GPA-U of
Regression 3.2. The predicted graduate GPA would be
Suppose we wanted to predict graduate nursing
Y .4 (.05)(3.2)
students’ grade-point averages (GPA). Only a lim-
.003(550) .002(600) 3.41
ited number of students can be accepted, so we
want to select applicants with the greatest chance We can assess the degree to which adding two
of success. Suppose we had previously found that independent variables improves our ability to pre-
students with high scores on the verbal portion of dict graduate school performance through the mul-
the Graduate Record Examination (GRE) tended to tiple correlation coefficient. In bivariate correla-
get better grades than those with lower GRE tion, the index is Pearson’s r. With two or more
scores. The correlation between GRE verbal scores independent variables, the correlation index is the
(GRE-V) and graduate GPAs has been calculated multiple correlation coefficient, or R. Unlike r, R
as .50. With only 25% (.502) of the variance of does not have negative values. R varies only from
graduate GPA accounted for, there will be many .00 to 1.00, showing the strength of the relationship
errors of prediction: Many admitted students will between several independent variables and a
not perform as well as expected, and many rejected dependent variable but not direction. Direction
CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 515
cannot be shown because X1 might be positively GRE-V (r .70). All correlations are fairly sub-
related to Y, while X2 might be negatively related. stantial, a fact that helps to explain why R is not
The R statistic, when squared (R2), indicates the much higher than the r between the dependent
proportion of variance in Y accounted for by the variable and undergraduate grades alone (.71 com-
combined, simultaneous influence of the indepen- pared with .60). This somewhat puzzling phenom-
dent variables. enon results from redundancy of information
R2 provides a way to evaluate the accuracy of among predictors. When correlations among inde-
a prediction equation. Suppose that with the three pendent variables are high, they tend to add little
predictor variables in the current example, the predictive power to each other. With low correla-
value of R .71. This means that 50% (.712) of the tions among predictor variables, each one can con-
variability in graduate students’ grades can tribute something unique to the prediction of the
be explained by their verbal and quantitative GRE dependent variable.
scores and undergraduate grades. The addition of Figure 21-2 illustrates this concept. Each circle
two predictors doubled the variance accounted for represents total variability in a variable. The circle
by the single independent variable (GRE-V), from on the left (Y) is the dependent variable (graduate
.25 to .50. GPA). The overlap between this circle and the other
The multiple correlation coefficient cannot be circles represents variability that the variables have
less than the highest bivariate correlation between in common. If the overlap were complete—if the
an independent variable and the dependent vari- entire Y circle were covered by the other circles—
able. Table 21-2 presents a correlation matrix with R2 would be 1.00. As it is, only half the circle is
the correlation coefficients for all pairs of variables covered, because R2 .50. The hatched area desig-
in this example. The independent variable most nates the independent contribution of undergraduate
strongly correlated with graduate performance is GPA in explaining graduate grades. This contribu-
(GPA-U), r .60. The value of R could not have tion amounts to 36% of Y’s variance (.602). The
been less than .60. remaining two independent variables do not con-
Another important point is that R can be more tribute as much as we would expect by considering
readily increased when the independent variables their bivariate correlation with graduate GPA (r for
have relatively low correlations among them- GRE-V .50; r for GRE-Q .55). In fact, their
selves. In the current case, the lowest correlation combined additional contribution is only 14%
coefficient is between GRE-Q and GPA-U (.50 .36 .14), designated by the dotted area on
(r .40), and the highest is between GRE-Q and the figure. The contribution is small because the two
GPA-GRAD 1.00
GPA-U .60 1.00
GRE-Q .55 .40 1.00
GRE-V .50 .50 .70 1.00
GPA, grade point average; GRE, Graduate Record Examination; GPA-GRAD, graduate GPA; GPA-
U, undergraduate GPA; GRE-Q, GRE quantitative score; GRE-V, GRE verbal score.
516 ■ PART 5 The Analysis of Research Data
Undergraduate
GPA
X1
Graduate GRE
Y X3
GPA verbal
X2
GRE
quantitative
F I G U R E 2 1 . 2 Visual representation of
multiple regression example.
In the case of multiple regression, the form is predictor increase our ability to predict Y after two
similar: predictors have been used? An F-statistic can be
computed to answer this question.
SSdue to regression/dfregression
F Let us number each independent variable in
SSof residuals/dfresiduals the current example: X1 GPA-U; X2 GRE-Q;
and X3 GRE-V. We can then symbolize various
The basic principle is the same in both cases: correlation coefficients as follows:
Variance resulting from independent variables is
contrasted with variance attributable to other fac- Ry.1 the correlation of Y with GPA-U
tors, or error. To calculate the F-statistic for Ry.12 the correlation of Y with GPA-U and
regression, we can use the following alternative GRE-Q
formula: Ry.123 the correlation of Y with all three
predictors
R2 k
F The values of these Rs are as follows:
(1 R )(N k 1)
2
were 11 independent variables measuring primarily where R2y.12...k1 squared multiple correlation
maternal factors such as maternal confidence, coefficient for Y correlated
maternal depression, and family income. The value with k1 predictor variables
of R2 was .40 for motor development and .60 for k1 the number of predictors for
mental development scores, both p .01. the larger of the two sets of
predictors
Tests for Adding Predictors R2y.12...k squared R for Y correlated
2
Another question researchers may want to answer with k2 predictor variables
is: Does adding Xk to the regression equation k2 the number of predictors for
significantly improve the prediction of Y over that the smaller of the two sets of
achieved with Xk1? For example, does a third predictors
518 ■ PART 5 The Analysis of Research Data
In the current example, the F-statistic for test- in the regression equation, it means that the vari-
ing whether adding GRE-Q scores significantly able associated with the coefficient contributes sig-
improves our prediction of Y is nificantly to the regression, even after extraneous
variables are taken into account.
(.50 .36) 1
F 27.16
.50 97
Consulting Table A-2 in Appendix A, we find Strategies for Handling Predictors in
that with df 1 and 97 and a significance level of Multiple Regression
.01, the critical value is about 6.90. Therefore, adding There are various strategies for entering predictor
GRE-Q to the regression equation significantly variables into regression equations. The three most
improved the accuracy of predicting graduate GPA, common are simultaneous, hierarchical, and step-
beyond the .01 level. wise regressions.
Example of adding predictors in multiple
regression: Simultaneous Multiple Regression
McCarter-Spaulding and Kearney (2001) studied The most basic strategy, simultaneous multiple
mothers’ perceptions of insufficient milk. They first regression, enters all predictor variables into the
examined demographic predictors of these percep- regression equation at the same time. A single
tions (age, education, and parity), which yielded a regression equation is developed, and statistical
nonsignificant R2 of .02. They then added a fourth tests indicate the significance of R and of individ-
predictor, scores on a parenting self-efficacy scale. ual regression coefficients. The study about pre-
R2 increased to .25; the .23 increment was statisti- dicting infant development cited earlier (Badr,
cally significant. 2001) used a simultaneous approach: All 11 inde-
pendent variables were put into the regression
Tests of the Regression Coefficients equation at the same time. This strategy is most
Researchers often want to know the significance appropriate when there is no basis for considering
of individual predictors in the regression equa- any particular predictor as causally prior to anoth-
tion, which can be done by calculating t statistics er, and when the predictors are of comparable
that test the unique contribution of each indepen- importance to the research problem.
dent variable. A significant t indicates that the
regression coefficient (b) is significantly different Hierarchical Multiple Regression
from zero. Many researchers use hierarchical multiple
Regression coefficients in multiple regression regression, which involves entering predictors
must be interpreted differently than coefficients in into the equation in a series of steps. Researchers
bivariate regression. In simple regression, the value control the order of entry, with the order typically
of b indicates the amount of change in predicted based on logical or theoretical considerations. For
values of Y, for a specified rate of change in X. In example, some predictors may be thought of as
multiple regression, the coefficients represent the causally or temporally prior to others, in which
number of units the dependent variable is predicted case they could be entered in an early step.
to change for each unit change in a given indepen- Another common reason for using hierarchical
dent variable when the effects of other predictors regression is to examine the effect of a key inde-
are held constant. This “holding constant” of other pendent variable after first removing the effect of
variables means that they are statistically con- extraneous variables. The study by McCarter-
trolled, a very important feature that can enhance a Spaulding and Kearney (2001) exemplifies this
study’s internal validity. If a regression coefficient strategy. Demographic characteristics were
is significant and extraneous variables are included entered in step 1, and then parenting self-efficacy
CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 519
X3
X1
Y Y Y
X2
was entered in step 2. The findings indicated that, cedure continues until no additional predictor sig-
even after taking age, parity, and education into nificantly increases the value of R2.
account, parenting self-efficacy was a significant Figure 21-3 illustrates stepwise multiple regres-
predictor of the mothers’ perceptions about insuf- sion. The first variable (X1), has a correlation of .60
ficient milk supply. with Y (r2 .36); no other predictor variable is cor-
With hierarchical regression, researchers related more strongly with the dependent variable.
determine the number of steps as well as the num- The variable X1 accounts for the portion of the vari-
ber of predictors included in each step. When sev- ability of Y represented by the hatched area in step 1
eral variables are added as a block, as in the of the figure. This hatched area is, in effect, removed
McCarter-Spaulding and Kearney example, the from further consideration, because this portion of Y’s
analysis is a simultaneous regression for those vari- variability is explained or accounted for. Thus, the
ables at that stage. Thus, hierarchical regression variable chosen in step 2 is not necessarily the X vari-
can be considered a controlled sequence of simul- able with the second largest correlation with Y. The
taneous regressions. selected predictor is the one that explains the largest
portion of what remains of Y’s variability after X1 has
Stepwise Multiple Regression been taken into account. Variable X2, in turn, removes
Stepwise multiple regression is a method of a second part of Y so that the independent variable
empirically selecting the combination of indepen- selected in step 3 is the one that accounts for the most
dent variables with the most predictive power. In variability in Y after both X1 and X2 are removed.
stepwise multiple regression, predictors enter the
regression equation in the order that produces the TIP: Stepwise regression is a somewhat con-
greatest increments to R2. The first step selects troversial procedure because variables are
the single best predictor of the dependent variable, entered into the regression equation based solely on sta-
that is, the independent variable with the highest tistical rather than theoretical criteria. Pure stepwise
bivariate correlation with Y. The second variable to regression is best suited to exploratory work, but cross-
enter the equation is the one that produces the validation is essential. If a sample is sufficiently large,
largest increase to R2 when used simultaneously we recommend that it be split in half to determine if
with the variable selected in the first step. The pro- similar equations result with both subsets of data.
520 ■ PART 5 The Analysis of Research Data
importance of independent variables. Unfortunately, where zY¿ predicted value of the standard score
the determination of the relative contributions of for Y
independent variables in predicting a dependent 1 to k standardized regression weights for k
variable is one of the thorniest issues in regression independent variables
analysis. When independent variables are correlated, zX1 to zXk standard scores for the k predictors
as they usually are, there is no ideal way to
With all the s (referred to as beta [ ]
disentangle the effects of variables in the equation.
weights) in the same measurement units, can their
It may appear that the solution is to compare
relative size shed light on how much importance to
the contributions of the Xs to R2. In our graduate
attach to predictor variables? Many researchers
school example, GPA-U accounted for 36% of Y’s
have interpreted beta weights in this fashion, but
variance; GRE-Q explained an additional 14%.
there are problems in doing so. These regression
Should we conclude that undergraduate grades are
coefficients will be the same no matter what the
more than twice as important as GRE-Q scores in
order of entry of the variables. The difficulty, how-
explaining graduate school grades? This conclu-
ever, is that regression weights are unstable. The
sion would be inaccurate because the order of entry
values of tend to fluctuate from sample to sam-
of variables in a regression equation affects their
ple. Moreover, when a variable is added to or
apparent contribution. If these two predictor vari-
ables were entered in reverse order (i.e., GRE-Q
first), R2 would remain unchanged at .50; however, *Consult a standard statistical textbook for a complete discus-
GRE-Q’s contribution would be .30 (.552), and sion of standard scores.
CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 521
subtracted from the regression equation, the beta where N estimated number of subjects needed
weights change. Because there is nothing absolute L tabled value for the desired ␣ and
about the values of the regression coefficients, it is power
difficult to attach theoretical importance to them. k number of predictors
Another method of disentangling relationships
estimated effect size
is through causal modeling, which represents an
important tool in testing theoretical expectations Table 21-3 presents values for L when ␣ .05 and
about relationships. Causal modeling is described power is .50 to .95, in situations in which there are
later in this chapter. up to 20 predictor variables.
As an example, suppose we were planning a
Sample Size for Multiple Regression study to predict functional ability in nursing home
residents using five predictor variables. We estimate
Small samples are especially problematic in multi- a moderate effect size (R2 .13) and want to achieve
ple regression and other multivariate procedures. a power of .80 and ␣ .05. With R2 .13, the esti-
Inadequate sample size can lead to Type II errors, mated population effect size (
) is .149 (.13 .87).
and can also yield erratic and misleading regres- From Table 21-3, we find that the value of
sion coefficients. L 12.83. Thus:
There are two approaches to estimating sample
size needs. One concerns the ratio of predictor vari- 12.83
N 5 1 92.1
ables to total number of cases. Tabachnick and .149
Fidell (2000) indicate that the barest minimum is to
include five times as many cases as there are pre- Thus, a sample of about 92 nursing home res-
dictors in the regression equation. They recom- idents is needed to detect a population R2 of .13
mend, however, a ratio of 20 to 1 for simultaneous with five predictors, with a 5% chance of a Type I
and hierarchical regression and a ratio of 40 to 1 for error and a 20% chance of a Type II error.
stepwise. More cases are needed for stepwise TIP: If you are testing hypotheses and ana-
regression because this procedure capitalizes on lyzing data by computer, you should defi-
the idiosyncrasies of a specific data set. nitely consider using multivariate procedures in
A better way to estimate sample size needs is lieu of bivariate ones because the commands to the
to perform a power analysis. The number of sub- computer are not appreciably more complicated.
jects needed to reject the null hypothesis that R Researchers can come to the wrong conclusions
equals zero is estimated as a function of effect size, about their hypotheses (i.e., commit Type I errors)
number of predictors, desired power, and the sig- when extraneous variables are not controlled, and
nificance criterion. For multiple regression, the multivariate procedures offer a way to control
estimated population effect size is: them.
R2
1 R2 A N A LY S I S O F
C O VA R I A N C E
Researchers must either predict the value of R2
on the basis of earlier research, or use the conven- Analysis of covariance (ANCOVA) has much in
tion that the effect size will be small (R2 .02), common with multiple regression, but it also has fea-
moderate (R2 .13), or large (R2 .30). Next, the tures of ANOVA. Like ANOVA, ANCOVA is used to
following formula is applied: compare the means of two or more groups, and the
central question for both is the same: Are mean group
L differences likely to be real, or do they reflect chance
N k1
fluctuations? Like multiple regression, however,
522 ■ PART 5 The Analysis of Research Data
TABLE 21.3 Power Analysis Table for Multiple Regression: Values of L for .05
Power (1 )
Covariate
X*
Y Y X
Step 1 Step 2
FIGURE 21.4 Visual respresentation of analysis of covariance.
scores. Covariates can either be continuous vari- of weight change than simple ANOVA. Some
ables (e.g., anxiety test scores) or dichotomous hypothetical data for such a study are displayed in
variables (male/female); the independent variable Table 21-4.
is always a nominal-level variable. Two aspects of the weight values in this table
ANCOVA tests the significance of differences are discernible. First, despite random assignment to
between group means after first adjusting scores on treatment groups, initial group means are different.
the dependent variable to remove the effect of the Subjects in diet B differ from those in diet C by an
covariate. The adjustment uses regression proce- average of 10 pounds (175 versus 185 pounds).
dures. In essence, the first step in ANCOVA is the This difference is not significant*; it reflects chance
same as the first step in hierarchical multiple sampling fluctuations. Second, post-treatment
regression. Variability in the dependent measure means are also different by a maximum of only 10
that can be explained by the covariate is removed pounds (160 to 170). However, the mean number of
from further consideration. ANOVA is performed pounds lost ranged from 10 pounds for diets A and
on what remains of Y’s variability to see whether, B to 25 pounds for diet C.
once the covariate is controlled, significant differ- A summary table for ordinary ANOVA is pre-
ences between group means exist. Figure 21-4 sented in part A of Table 21-5. According to the
illustrates this two-step process: Y is patients’ results, group differences in post-treatment means
posttest anxiety scores, the covariate X* is pretest result in a nonsignificant F-value. Based on
anxiety scores, and the independent variable X ANOVA, we would conclude that all three diets
designates whether patients were exposed to the had equal effects on weight loss.
biofeedback therapy. Part B of Table 21-5 presents the summary
Let us consider another example to explore table for ANCOVA. The first step of the analysis
further aspects of ANCOVA. Suppose we were breaks total variability in post-treatment weights
testing the effectiveness of three weight-loss diets, into two components: (1) variability explained by
and we randomly assign 30 subjects to 1 of 3
groups. An ANCOVA, using pretreatment weight *Calculations demonstrating nonsignificance are not shown.
as the covariate, permits a more sensitive analysis The calculated F .45. With 2 and 27 df, p
.05.
524 ■ PART 5 The Analysis of Research Data
X, independent variable (diet); X*, covariate (pretreatment weight); Y, dependent variable (posttreatment weight).
the covariate, pretreatment weights; and (2) residual if appropriate covariates are used. The increase in
variability. The table shows that the covariate sensitivity results from the fact that the covariate
accounted for a significant amount of variance, reduces the error term (within-group variability),
which is not surprising given the strong relationship against which treatment effects are compared. In
between pretreatment and post-treatment weights part A of Table 21-5, it can be seen that the within-
(r .91). This merely indicates that people who group term is extremely large (12,300), masking the
started out especially heavy tended to stay that way, contribution of the experimental treatments.
relative to others in the sample. In the second phase Theoretically, it is possible to use any number of
of the analysis, residual variance is broken down to covariates. It is probably unwise, however, to use
reflect between-group and within-group contribu- more than five or six. For one thing, including more
tions. With 2 and 26 degrees of freedom, the F of covariates is often unnecessary because of the typi-
17.54 is significant beyond the .01 level. The con- cally high degree of redundancy beyond the first few
clusion is that, after controlling for initial weight, variables. Moreover, it is unwise to add covariates that
there is a significant difference in weight loss attrib- do not explain a significant amount of variability in a
utable to exposure to different diets. dependent variable. Each covariate uses up a degree
This fictitious example was contrived so that the of freedom, and fewer degrees of freedom means that
ANOVA result of “no difference” would be altered a higher computed F is required for significance. For
by adding a covariate. Most actual results are less instance, with 2 and 26 df, an F of 5.53 is required for
dramatic than this example. Nonetheless, ANCOVA significance at the .01 level, but with 2 and 23 df (i.e.,
yields a more sensitive statistical test than ANOVA adding three covariates), an F of 5.66 is needed.
CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 525
SUM OF MEAN
SOURCE OF VARIATION SQUARES df SQUARE f p
TIP: ANCOVA is almost always prefer- after adjusting for pretreatment weights. It some-
able to ANOVA or t-tests, and is relatively times is useful to examine adjusted means, that is,
easy to run on the computer. Covariates are usual- group means on the dependent variable after adjust-
ly available: At a minimum, you can use key back- ing for (i.e., removing the effect of) covariates.
ground characteristics, such as age, gender, and so Means can be adjusted through a process some-
on. Covariates should be variables that you suspect times referred to as multiple classification analysis
are correlated with the dependent variable, and (MCA). Adjusted means allow researchers to deter-
demographic characteristics usually are related to mine net effects (i.e., group differences on the
many other attributes. A pretest measure (i.e., a dependent variable that are net of the effect of the
measure of the dependent variable before the covariate.) A subsequent section of this chapter,
occurrence of the independent variable) is an excel- which presents computer examples of multivariate
lent choice for a covariate. statistics, provides an illustration of adjusted means.
Example of ANCOVA:
Adjusted Means
Badger and Collins-Joyce (2000) studied
As shown in part B of Table 21-5, an ANCOVA depression in relation to functional ability in older
summary table provides information about signifi- adults. The sample was divided into two groups
cance tests. The table indicates that at least one of (depressed or not depressed). After statistically
the three groups had a post-treatment weight that is controlling for scores on a physical health impair-
significantly different from the overall grand mean, ment measure, they found that those in the
526 ■ PART 5 The Analysis of Research Data
depressed group had significantly lower functional Mathematically, a factor is a linear combina-
ability scores than those in the nondepressed group tion of variables in a data matrix, which contains
(F 7.22, df 1,76, p .01). the scores of N people on k different measures. For
instance, a factor might be defined by the following
equation:
F A C T O R A N A LY S I S
The major purpose of factor analysis is to reduce F b1X1 b2X2 b3X3 p bkXk
a large set of variables into a smaller, more man- where
ageable set. Factor analysis disentangles complex F a factor score
interrelationships among variables and identifies X1 to Xk values on the k original variables
variables that “go together” as unified concepts. b1 to bk weights
This section deals with a type of factor analysis
known as exploratory factor analysis. Another The development of such equations allows us
type—confirmatory factor analysis—uses more to reduce the Xk scores to one or more factor scores.
complex modeling and estimation procedures and
more sophisticated computer programs, as
Factor Extraction
described later.
Suppose we developed 100 Likert-type items Most factor analyses involve two phases. The first
measuring women’s attitudes toward menopause. phase (factor extraction) condenses variables in
Our goal is to compare the attitudes of urban ver- the data matrix into a smaller number of factors.
sus rural women. If we do not combine items to The general goal is to extract clusters of highly
form a scale, we would have to perform 100 sepa- interrelated variables from a correlation matrix.
rate statistical tests (such as chi-square tests) to There are various methods of performing the first
compare the two groups of women on the 100 step, each of which uses different criteria for
items. We could form a scale by adding together assigning weights to variables. The most widely
scores from several individual items, but which used factor extraction method is called principal
items should be combined? Would it be reasonable components (or principal factor or principal
to combine all 100 items? Probably not, because axes), but other methods include the image, alpha,
the 100 items are not all tapping exactly the centroid, maximum likelihood, and canonical
same thing. There are various aspects to women’s techniques.
attitude toward menopause. One aspect may relate Factor extraction results in an unrotated fac-
to aging, and another to loss of reproductive tor matrix, which contains coefficients or weights
ability. Other questions may involve sexuality, and for all original variables on each extracted factor.
yet others may concern avoidance of monthly (Because unrotated factor matrixes are difficult to
aggravation. There are, in short, multiple dimen- interpret, we postpone a detailed discussion of fac-
sions to women’s attitudes toward menopause, and tor matrixes until the second factor analysis phase
each dimension should be captured on a separate is described.) In the principal components method,
scale. Women’s attitude on one dimension may be weights for the first factor are defined such that the
independent of their attitude on another. The iden- average squared weight is maximized, permitting a
tification of dimensions can be made a priori by maximum amount of variance to be extracted by
researchers, but different researchers may read dif- the first factor. The second factor, or linear combi-
ferent concepts into the items. Factor analysis nation, is formed so that the highest possible
offers an empirical method of clarifying the under- amount of variance is extracted from what remains
lying dimensionality of a large set of measures. after the first factor has been taken into account.
Underlying dimensions thus identified are called The factors thus represent independent sources of
factors. variation in the data matrix.
CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 527
Researchers choose from two classes of rota- cating oblique rotation point out that if the
tion. Figure 21-5 illustrates orthogonal rotation, concepts are correlated, then the analysis should
in which factors are kept at right angles to one reflect this fact. In practice, similar conclusions are
another. Orthogonal rotations maintain the inde- often reached by both rotation procedures.
pendence of factors—that is, orthogonal factors are Researchers work with a rotated factor matrix
uncorrelated with one another. Oblique rotations, in interpreting the factor analysis. An example of
on the other hand, permit rotated axes to depart such a matrix is displayed in Table 21-7. To make
from a 90-degree angle. In our figure, an oblique the discussion less abstract, let us say that the 10
rotation would have put axis B1 between variables variables listed in the table are the first 10 of the 100
2 and 3 and axis B2 between variables 5 and 6. items measuring attitudes toward menopause, and
This placement strengthens the clustering of that 3 factors have been extracted and rotated. The
variables around an associated factor, but results in entries under each factor are the weights, called fac-
correlated factors. Some researchers insist that tor loadings. For orthogonally rotated factors, factor
orthogonal rotation leads to greater theoretical clar- loadings can range from 1.00 to 1.00 and can be
ity; others claim that it is unrealistic. Those advo- interpreted like correlation coefficients—they
CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 529
express the correlations between individual vari- indicate 5 underlying dimensions. By reducing 100
ables and factors (underlying dimensions). In this variables to 5 new variables, the analysis of differ-
example, variable 1 is fairly highly correlated with ences between urban and rural women using t-tests
factor 1, .75. By examining factor loadings, it is pos- or ANCOVA would be greatly simplified.
sible to find which variables “belong” to a factor. Several methods can be used to form factor
Variables 1, 3, 8, and 10 have fairly sizable loadings scores, one of which is to weight variables accord-
on factor I. (Loadings with an absolute value of ing to their factor loading on each factor. For exam-
about .40 or .30 normally are used as cutoff values.) ple, a factor score for the first factor in Table 21-7
The underlying dimensionality of the data can then could be computed for each subject by the follow-
be interpreted. By inspecting items 1, 3, 8, and 10, ing formula:
we would search for a common theme that makes
the variables go together. Perhaps these four ques- F1 .75X1 .02X2 .59X3 .05X4
tions have to do with the link between menopause .21X5 .07X6 .08X7 .68X8 .04X9
and infertility, and perhaps items 2, 5, and 6 (items .43X10 p b100X100
with high loadings on factor II) are related to sexu- where the X values* subjects’ scores on the 100
ality. The naming of factors is a process of identify- items
ing underlying constructs.
Most factor analysis computer programs can
Factor Scores directly compute this type of factor score.
A second method of obtaining factor scores is
When the main purpose of factor analysis is to to select variables to represent the factor and to
delineate the conceptual makeup of a set of mea- combine them with unit weighting. This method
sures, the analysis may end at this point. Frequently,
however, researchers want to develop factor scores *Standardized (z) scores are often used in lieu of raw scores in
to use in subsequent analyses. For instance, the computing factor scores when the means and standard devia-
factor analysis of the 100 menopause items might tions of the variables differ substantially.
530 ■ PART 5 The Analysis of Research Data
would produce the following factor score on the practitioners, whose responses were factor ana-
first factor of our example: lyzed. An eigenvalue of 1.00 was used as the
extraction criterion; 33 items with factor loadings
F1 X1 X3 X8 X10 p less than .35 were discarded. The remaining items
The variables are those with high loadings on fac- formed six factors: collegiality; autonomy; interac-
tor I (the “? . . .” indicates that some of the other 90 tions; professional growth; and benefits.
items with high loadings would be included). This pro-
cedure is computationally and conceptually simpler O T H E R L E A S T- S Q U A R E S
than the first approach, and yields similar results. M U L T I VA R I A T E
To illustrate more concretely, consider the TECHNIQUES
rotated factor matrix in Table 21-7 and assume that
factor scores are being computed on the basis of In this section, the methods known as discriminant
this 10-item analysis, omitting for the sake of sim- function analysis, canonical correlation, and multi-
plicity the remaining 90 items on the menopause variate analysis of variance are introduced. The
scale. Suppose two respondents had the following introduction is brief, and computations are entirely
scores on the first 10 items: omitted because these procedures are very complex.
The intent is to acquaint you with the types of
Subject 1: 7, 1, 1, 5, 1, 6, 7, 6, 3, 2 research situations for which these methods are
Subject 2: 2, 5, 6, 3, 7, 3, 2, 3, 4, 4 appropriate. Advanced statistical texts such as the
Factor I has high loadings on variables 1, 3, 8, ones listed in the references may be consulted for
and 10. Thus, the two subjects’ factor scores on fac- more information.
tor 1 would be
Discriminant Analysis
7 1 6 2 10
2 6 3 4 5 In multiple regression, the dependent variable is
normally a measure on either the interval or ratio
The minus signs reflect the negative loadings scale. The regression equation makes predictions
on variables 3 and 10.* The same procedure would about scores that take on a range of values, such as
be performed for all three factors, yielding the fol- heart rate, weight, or scores on a depression scale.
lowing factor scores: Discriminant analysis, in contrast, makes predic-
Subject 1: 10, 3, 9 tions about membership in categories or groups.
Subject 2: 5, 9, 1 The purpose of the analysis is to distinguish groups
from one another on the basis of available indepen-
Factor scores would be computed for all dent variables. For instance, we may wish to pre-
respondents, and these scores could be used in sub- dict membership in such groups as complying
sequent analyses as measures of different dimen- versus noncomplying cancer patients, graduating
sions of attitudes toward menopause. nursing students versus dropouts, or normal preg-
Example of factor analysis: nancies versus those terminating in a miscarriage.
Misener and Cox (2001) developed a scale Discriminant analysis develops a regression
to measure nurse practitioner job satisfaction. They equation—called a discriminant function—for a
administered 77 items to a sample of 342 nurse categorical dependent variable, with independent
variables that are either dichotomous or continu-
ous. Researchers begin with data from subjects
*Researchers forming scale scores with Likert items often reverse
the direction of the scoring on negatively loaded items before whose group membership is known, and develop
forming factor scores. Direction of an item can be reversed by an equation to predict membership when only mea-
subtracting the raw score from the sum of 1 the maximum item sures of the independent variables are available.
value. For example, in a five-point scale, a score of 2 would be
reversed to 4 [(1 5) 2]. When such reversals have been done, The discriminant function indicates to which group
all raw scores can be added to compute factor scores. each subject will probably belong.
CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 531
Discriminant analysis for predicting member- gap between multiple regression and canonical
ship into only two groups (e.g., nursing school correlation is greater than this statement suggests.
dropouts versus graduates) is relatively simple and Like other techniques described thus far,
can be interpreted in much the same way as canonical correlation uses the least-squares prin-
multiple regression. When there are more than two ciple to partition and analyze variance. Basically,
groups, the calculations and interpretations are two linear composites are developed, one of
more complex. With three or more groups (e.g., which is associated with the dependent variables,
very—low-birth-weight, low-birth-weight, and nor- the other of which is for the independent vari-
mal-birth-weight infants), the number of discrimi- ables. The relationship between the two linear
nant functions is either the number of groups minus composites is expressed by the canonical corre-
1 or the number of independent variables, whichev- lation coefficient, RC, which, when squared,
er is smaller. The first discriminant function is the indicates the proportion of variance accounted
linear combination of predictors that maximizes the for in the analysis. When there is more than one
ratio of between-group to within-group variance. dimension of covariation in the two sets of vari-
The second function is the linear combination that ables, more than one canonical correlation can be
maximizes this ratio, after the effect of the first found.
function is removed. Because independent variables Examples of research using canonical correla-
have different weights on the various functions, it is tion are not common, perhaps because of its com-
possible to develop theoretical interpretations based plexity and perhaps because it is problematic as a
on the knowledge of which predictors are important hypothesis-testing technique.
in discriminating among different groups.
Example of canonical correlation:
As with multiple regression analysis, it is possi-
Lucas, Orshan, and Cook (2000) used
ble to use a stepwise approach to enter predictors into
canonical correlation to explore various compo-
a discriminant analysis equation. Discriminant analy-
nents of Pender’s Health Promotion Model. They
sis also produces an index designating the proportion
studied the role of cognitive-perceptual factors
of variance in the dependent variable accounted for
(e.g., self-esteem, perceived health) and modifying
by predictor variables. The index is Wilks’ lambda
factors (e.g., age, income) in various health-
(), which actually indicates the proportion of vari-
promoting behaviors (e.g., stress management)
ance unaccounted for by predictors, or 1 R2.
among community-dwelling older women. Two
Example of discriminant analysis: dimensions of covariation were identified; the
Aminzadeh and Edwards (2000) examined value of RC for the first canonical variate was .74.
predictors of cane use among community-dwelling
older adults. Using a stepwise discriminant analysis,
they determined that age, subjective norms, and atti- Multivariate Analysis of Variance
tudes predicted cane use. The discriminant function
Multivariate analysis of variance (MANOVA) is
accounted for 67% of the variance in cane use
the extension of ANOVA to more than one depen-
( .33), and correctly categorized 91% of the cases.
dent variable. MANOVA is used to test the signif-
icance of differences in group means for two or
more dependent variables, considered simultane-
Canonical Correlation
ously. For instance, if we wanted to examine the
Canonical correlation analyzes the relationship effect of two methods of exercise on both diastolic
between two or more independent variables and and systolic blood pressure, MANOVA would be
two or more dependent variables. Conceptually, appropriate. Researchers often analyze such data
one can think of this technique as an extension of by performing two separate univariate ANOVAs.
multiple regression to more than one dependent Strictly speaking, this practice is incorrect.
variable. Mathematically and interpretatively, the Separate ANOVAs imply that the dependent
532 ■ PART 5 The Analysis of Research Data
variables. Path analysis is not a method for discov- (V1) and illness severity (V2) are exogenous; no
ering causes; rather, it is a method applied to a pre- attempt is made in the model to elucidate what
specified model formulated on the basis of prior causes different nursing actions or different
knowledge and theory. degrees of illness. An endogenous variable, by
Path analytic reports often display results in a contrast, is one whose variation is determined by
path diagram; we use such a diagram (Figure 21-6) other variables in the model. In our example, self-
to illustrate key concepts. This model postulates that care capacity (V3) and length of hospitalization
the dependent variable, patients’ length of hospitaliza- (V4) are endogenous.
tion (V4), is the result of their capacity for self-care Causal linkages are shown on a path diagram
(V3); this, in turn, is affected by nursing actions (V1) by arrows drawn from presumed causes to pre-
and the severity of their illness (V2). This model is a sumed effects. In our illustration, severity of illness
recursive model, which means that the causal flow is is hypothesized to affect length of hospitalization
unidirectional and without feedback loops. In other both directly (path p42) and indirectly through the
words, it is assumed that variable 2 is a cause of vari- mediating variable self-care capacity (paths p32
able 3, and that variable 3 is not a cause of variable 2. and p43). Correlated exogenous variables are indi-
In path analysis, a distinction is made between cated by curved lines, as shown by the curved line
exogenous and endogenous variables. An exoge- between nursing actions and illness severity.
nous variable is a variable whose determinants lie Ideally, the model would totally explain the
outside the model. In Figure 21-6, nursing actions outcome of interest (i.e., length of hospitalization).
V1
Nursing
actions
P 31
r12
V3
P43 V4
Capacity
Length of
for
hospital stay
P 32 self-care
V2
Severity
of
P 42
illness
P4e4
P3e3
e3 e4
In practice, this almost never happens because effects. Polit (1996) and Pedhazur (1997) provide
there are other determinants of the outcome, an overview of path analytic techniques, and
referred to as residual variables. In Figure 21-6, advanced students should refer to books such as the
there are two boxes labeled e, which denote a com- classic text by Blalock (1972).
posite of all determinants of self-care capacity and
hospital stay that are not in the model. If we can Example of path analysis:
identify and measure additional causal forces and Berger and Walker (2001) tested an explana-
incorporate them into the theory, they should be in tory model of fatigue in women receiving adjuvant
the model. breast cancer chemotherapy. The predictor variables
Path analysis solves for path coefficients, in the model included health and functional status,
which are the weights representing the effect of one type of chemotherapy protocol, health-promoting
variable on another. In Figure 21-6, causal paths lifestyle behaviors, nutritional status, interpersonal
indicate that one variable (e.g., V3) is caused by relations, symptom distress, and initial reaction to
another (e.g., V2), yielding a path labeled p32. In the diagnosis of cancer. In the final model, symptom
research reports, path symbols would be replaced distress had the strongest direct path to perceived
by actual path coefficients, which are derived fatigue 48 hours after treatment.
through regression procedures. Path coefficients are
standardized partial regression slopes. For example,
path p32 is equal to 32.1—the beta weight between Linear Structural Relations
variables 2 and 3, holding variable 1 constant. Analysis—LISREL
Because path coefficients are in standard form, they A drawback of path analysis using ordinary least-
indicate the proportion of a standard deviation dif- squares regression is that the method’s validity is
ference in the caused variable that is directly attrib- based on a set of restrictive assumptions, most of
utable to a 1.00 SD difference in the specified which are virtually impossible to meet. First, it is
causal variable. Thus, the path coefficients give us assumed that variables are measured without error,
indication about the relative importance of various but (as discussed in Chapter 18) most measures
determinants. contain some error. Second, it is assumed that
Path analysis estimates path coefficients residuals (error terms) in the different regression
through the use of structural equations. The equations are uncorrelated. This assumption is sel-
structural equations for Figure 21-6 are as follows: dom tenable because error terms often represent
z1 e1 unmeasured individual differences—differences
z2 e2 that are not entirely random. Third, traditional path
z3 p31z1 p32z2 e3; and analysis assumes that the causal flow is unidirec-
z4 p43z3 p42z2 e4 tional or recursive. In reality, causes and effects are
often reciprocal or iterative.
These equations indicate that z1 and z2 (stan- Linear structural relations analysis, usually
dard scores for variables 1 and 2) are determined by abbreviated as LISREL,* is a more general and
outside variables (es); that z3 depends directly on z1 powerful approach that avoids these problems.
and z2 plus outside variables; and that z4 depends Unlike path analysis, LISREL can accommodate
directly on z2, z3, and outside variables. These struc- measurement errors, correlated residuals, and
tural equations could be solved to yield path coeffi-
cients by performing two multiple regressions: by
regressing variable 3 on variables 1 and 2; and by
regressing variable 4 on variables 2 and 3. *LISREL is actually the name of a specific computer program
for analyzing covariance structures and for performing structur-
Path analysis involves a number of procedures al equations modeling, but the software has come to be almost
for testing causal models and for determining synonymous with the approach.
CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 535
nonrecursive models that allow for reciprocal cau- measured variables. The measurement model is
sation. Another attractive feature of LISREL is that essentially a factor analytic model that seeks to
it can be used to analyze causal models involving confirm a hypothesized factor structure. Thus,
latent variables. A latent variable is an unmea- loadings on the factors (the latent variables) pro-
sured variable corresponding to an abstract vide a method for evaluating relationships between
construct. For example, factor analysis yields infor- observed and unobserved variables.
mation about underlying, latent dimensions that are We illustrate with a simplified example involv-
not measured directly. In LISREL, latent variables ing the relationship between two latent variables: a
are captured by two or more measured variables student’s cognitive ability and academic success. In
(manifest variables) that are indicators of the our example, depicted in Figure 21-7, cognitive
underlying construct. ability is captured by two indicators: scores on the
Path analysis uses least-squares estimation verbal and quantitative portions of the GRE.
with its associated assumptions and restrictions. According to the model, people’s test scores are
LISREL uses a different estimation procedure caused by their cognitive ability (and thus the
known as maximum likelihood estimation. arrows indicating a hypothesized causal path), and
Maximum likelihood estimators are ones that esti- are also affected by error (e1 and e2). Moreover, it is
mate the parameters most likely to have generated hypothesized that the error terms are correlated, as
the observed measurements. indicated by the curved line connecting e1 and e2.
LISREL proceeds in two phases. In the first Correlated measurement errors on the GRE might
phase, which corresponds to a confirmatory factor arise as a result of the person’s test anxiety or level
analysis (CFA), a measurement model is tested. of fatigue—factors that would systematically
When there is evidence of an adequate fit of the data depress test scores on both parts of the examination.
to the hypothesized measurement model, the causal The second latent variable is academic success,
model is tested by structural equation modeling. which is captured by undergraduate and graduate
GPAs. The error terms associated with these two
The Measurement Model Phase manifest variables are presumed not to be correlat-
The measurement model stipulates the hypothe- ed but, within the measurement model, the two
sized relationships among latent and manifest vari- latent variables are hypothesized to be correlated.
ables. In path analysis, it is assumed—unrealistical- The hypothesized measurement model then
ly—that underlying constructs on the one hand and would be tested against actual data using LISREL.
the scales used to measure them on the other are The analysis would provide loadings of the
identical. In LISREL, constructs are treated as unob- observed variables on the latent variables, the corre-
served variables measured by two or more fallible lation between the two latent variables, and the cor-
indicators. Thus, a major strength of LISREL is that relation between e1 and e2. The analysis would also
it enables researchers to separate latent variables indicate whether the overall model fit is good, based
from errors, and then to study causal relationships on a goodness-of-fit statistic. If the hypothesized
among latent variables. The process depends on the model is not a good fit to the study data, the mea-
judicious selection of indicators. A faulty selection surement model could be respecified and retested.
of measured variables to represent a latent variable
leads to questions about whether the theory’s con- The Structural Equations Model Phase
structs are really embedded in the causal model. After an adequate measurement model has been
During the measurement model phase, LIS- found, the second phase of LISREL can proceed. As
REL tests three things: (1) causal relationships in path analysis, researchers specify the theoretical
between measured and latent variables, (2) correla- causal model to be tested. As an example, we could
tions between pairs of latent variables, and (3) use LISREL to test a model in which we
correlations among the errors associated with the hypothesized that the latent variable cognitive ability
536 ■ PART 5 The Analysis of Research Data
Cognitive Academic
ability success
e1 e2 e3 e4
caused the latent variable academic success. The LISREL has gained considerable popularity
model being tested would look similar to the one in among nurse researchers, but is a highly complex
Figure 21-7, except that there would be a straight procedure. Readers interested in further informa-
line and an arrow (i.e., a path) from cognitive ability tion on LISREL are urged to consult Jöreskog and
to academic success. There would also be an arrow Sörbom (1997), Kelloway (1998), or Bollen and
leading to academic success from a residual term. Bollen (1989).
In this part of the analysis, LISREL yields
information about the hypothesized causal parame- Example of LISREL:
ters—that is, the path coefficients, which are pre- Mahon and Yarcheski (2001) used LISREL
sented as beta weights. The coefficients indicate the to test two models predicting health outcomes and
expected amount of change in the latent endoge- interpersonal outcomes in relation to depression
nous variable that is caused by a change in the latent among early adolescents. In both models,
causal variable. The LISREL program provides depressed mood contributed fairly strongly to
information on the significance of individual paths. depressive symptoms, which in turn affected health
The residual terms—the amount of unexplained and interpersonal outcomes.
variance for the latent endogenous variables—can
also be calculated from the LISREL analysis. The
O T H E R M U L T I VA R I A T E
overall fit of the causal model to the research data
S TAT I S T I C A L
can be tested by means of several alternative statis-
PROCEDURES
tics. Two such statistics are the goodness-of-fit
index (GFI) and the adjusted goodness-of-fit The statistical procedures described in this and the
index (AGFI). For both indexes, a value of .90 or previous two chapters cover most of the techniques
greater indicates a good fit of the model to the data. for analyzing quantitative data used by nurse
CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 537
displayed. An odds ratio (OR) is the ratio of one prior pregnancies. Figure 21-8 presents part of the
odds to another odds, and thus is an index of relative Statistical Package for the Social Sciences (SPSS)
risk—that is, the risk of an event occurring given one printout for a stepwise multiple regression in which
condition, versus the risk of it occurring given a dif- infant birth weight is the dependent variable and
ferent condition. Logistic regression was rarely used maternal age, smoking status, and number of prior
20 years ago because of the lack of convenient com- pregnancies are predictor variables. We will
puter programs, but that is no longer the case. explain some of the most noteworthy aspects of
this printout.
Example of logistic regression:
The top panel of Figure 21-8 shows that the
Nyamathi, Leake, Longshore, and Gelberg
first variable to enter the regression equation is
(2001) examined the congruity between homeless
mother’s age. The second panel (Model Summary)
women’s self-reports of cocaine use and hair
indicates that R .594–the same value shown in
assays for the presence of cocaine. They used
Figure 20-6 as the bivariate correlation. The value
logistic regression analysis to predict underreport-
of R2 is .353 (.5942), which represents the propor-
ing of cocaine use. Being Latina, younger, and liv-
tion of variance in birth weight accounted for by
ing primarily in shelters were significant predictors
mother’s age.* Maternal age was the first of the
of underreporting. For example, the odds of under-
reporting were twice as high among women whose three predictor variables stepped into the equation
usual living place in the prior month was shelters because it had the highest bivariate correlation with
than among other women (OR 2.01, p .05). birth weight. (The r between birth weight and num-
ber of pregnancies is .32, and the r between birth
weight and smoking status is .24.)
THE COMPUTER AND The next panel (ANOVA) shows the calcula-
M U L T I VA R I A T E tion of the F-ratio in which variability due to
S TAT I S T I C S regression (i.e., to the relationship between birth
Multivariate analyses are invariably done by weight and maternal age) is contrasted with resid-
computer because computations are complex. To ual variability. The value of F (15.252) with 1 and
illustrate computer analyses for two multivariate 28 df is significant at the .001 level, consistent with
techniques, we return to the example described in information from Figure 20-6.
Chapters 19 and 20, involving the implementation The actual regression equation is presented in
of a special prenatal intervention for low-income the next panel (Coefficients). If we wanted to pre-
young women. Data for these examples were pre- dict new values of birth weight based on maternal
sented in Table 19-8. age at birth, the equation would be:
Birth weight (3.119 age) 48.040
Example of Multiple Regression
That is, predicted birth weight values would be
In Chapter 20, we tested the hypothesis that older equal to the regression coefficient (b 3.119)
mothers in the sample had infants with higher birth times values of maternal age (X), plus the value of
weights than younger mothers, using Pearson’s r. the intercept constant (a 48.040).
The calculated value of r was .594, which was In regression, a statistical test can be per-
highly significant, thereby supporting the research formed for each regression coefficient by dividing
hypothesis. b by the standard error for b; this is shown here
Suppose now that we want to test whether we
can significantly improve our ability to predict
infant birth weight by adding two predictor vari- *The adjusted R2 of .330 is the R2 after it has been adjusted to
reflect more closely the goodness of fit of the regression model
ables in a multiple regression analysis: whether the in the population, through a formula that involves sample size
mother smoked while pregnant, and number of and number of independent variables.
Regression
Variables Entered/Removeda
Variables Variables
Model Entered Removed Method
1 Stepwise
(Criteria:
Probability
-of-F-to-en
Mother’s ter <=
age .050,
Probability
-of-F-to-re
move >=
.100)
Model Summary
Adjusted R Std. Error of
Model R R Square Square the Estimate
ANOVAb
Sum of
Model Squares df Mean Square F Sig.
Coefficientsa
Standardi
zed
Unstandardized Coefficients
Coefficients
Excluded Variablesb
Colinearif
Partial y Statistics
Model Beta In t Sig. Correlation Tolerance
FIGURE 21.8 Statistical Package for the Social Sciences (SPSS) printout for a stepwise multiple regression.
539
540 ■ PART 5 The Analysis of Research Data
under the columns t and Sig. The value of t for the the next panel (Tests of between-subjects
age coefficient is 3.905, which is significant at the effects), we see that the F-value for the covariate
.001 level. AGE is 24.358, significant at the .000 level (i.e.,
The final panel (Excluded Variables) shows the beyond the .001 level). After controlling for
two predictors not yet in the equation, number of AGE, the F-value for the independent variable
prior pregnancies and smoking status. The figure GROUP is 8.719, which is significant at the .006
shows that the t values associated with the regres- level. In other words, once AGE is controlled, the
sion coefficients for the two predictors are both research hypothesis about experimental versus
nonsignificant (p .915 and .659, respectively), control differences is supported rather than
once variation due to maternal age is taken into rejected. The total amount of variability
account. Because of this fact, the stepwise regres- explained with the two variables (1777.228),
sion ends at this point, without adding new predic- when contrasted with residual variability
tors. Given the significance criterion of .05 (shown (1703.072), is also significant (F 14.088,
in the top panel of “probability of F to p .000). The multiple R2 for predicting birth
enter .050”), neither of the two predictor vari- weight, based on both AGE and GROUP, is
ables would have significantly added to the predic- .511—substantially more than the R2 between
tion of birth weight, over and above what was maternal age and birth weight alone (.352).
already achieved with maternal age. The next panel shows the overall mean
It is, of course, possible to force the two (104.70) for the sample; the standard error (1.45);
additional predictors into the regression equation and the 95% confidence interval for estimating the
by doing a simultaneous regression. When this is population mean (CI 101.725 to 107.675). The
done (not shown in the table), the value of R bottom panel shows the group means after they are
increases from .594 to .598, a negligible and non- adjusted for maternal age (i.e., the results of an
significant increase. Thus, we conclude that we MCA). The original, unadjusted means for the
cannot improve our ability to predict infant birth experimental and control groups were 107.53 and
weight with the two additional predictors 101.87, respectively (Figure 20-5). After adjusting
available. for maternal age, however, the experimental mean
is 109.08, and the control mean is 100.32, a much
more sizable difference.
Example of Analysis of Covariance
In Chapter 20, we tested the hypothesis that infants
in the experimental group have higher birth
GUIDE TO MULTIVARIATE
weights than infants in the control group, using a t-
STATISTICAL TESTS
test. The computer calculated t to be 1.44, which
was nonsignificant with 28 df. The research With the widespread availability of computers and
hypothesis was therefore rejected. user-friendly programs for performing statistical
Through ANCOVA, we can test the same analysis, multivariate statistics have become
hypothesis controlling for maternal age, which, increasingly common in nursing studies. As with
as we have just seen, is significantly correlated bivariate statistical tests, the selection of a multi-
with birth weight. Figure 21-9 presents the print- variate procedure depends on several factors,
out for ANCOVA for this analysis, with birth including the nature of the research question and
weight as the dependent variable, maternal age the measurement level of the variables. Table 21-8
(AGE) as the covariate, and GROUP (experimen- summarizes some of the major features of several
tal versus control) as the independent variable. of the multivariate statistics discussed in this chap-
The first panel shows that the group variable ter and thus serves as an aid to the selection of an
involves 15 experimentals and 15 controls. In appropriate procedure.
CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 541
Between-Subjects Factors
Value Label N
Treatment 1 Experimenta 15
group l
2 Control 15
1. Grand Mean
Dependent Variable: Infant birth weight in ounces
95% Confidence Interval
Mean Std. Error Lower Bound Upper Bound
104.700a 1.450 101.725 107.675
2. Treatment group
Dependent Variable: Infant birth weight in ounces
95% Confidence Interval
Treatment group Mean Std. Error Lower Bound Upper Bound
Experimental 109.080a 2.074 104.824 113.337
Control 100.320a 2.074 96.063 104.576
• The second phase of factor analysis involves • LISREL proceeds in two sequential phases: the
factor rotation, which enhances the inter- measurement model phase and the structural
pretability of the factors by aligning variables equations modeling phase. The measurement
more distinctly with a particular factor. model stipulates the hypothesized relationship
• Factor loadings shown in a rotated factor between latent variables and manifest variables;
matrix can be examined to identify and name if the measurement model fits the data, the
the underlying dimensionality of the original set hypothesized causal model is then tested.
of variables and to compute factor scores. • Most multivariate procedures described in the
• Discriminant analysis is used to make predic- chapter are based on least-squares estimation,
tions about dependent variables that are cate- but LISREL is based on a different estimation
gorical (i.e., predictions about membership in procedure known as maximum likelihood;
groups) on the basis of two or more predictor maximum likelihood estimates the parameters
variables. most likely to have generated observed data.
• Canonical correlation analyzes the relation- • Logistic regression, another multivariate proce-
ship between two or more independent and two dure based on maximum likelihood estimation,
or more dependent variables; it yields a canoni- is used to predict categorical dependent vari-
cal correlation coefficient, RC. ables. A useful feature of this technique is that it
• Multivariate analysis of variance (MANOVA) yields an odds ratio that is an index of relative
is the extension of ANOVA to situations in risk, that is, the risk of an outcome occurring
which there is more than one dependent variable. given one condition, versus the risk of it occur-
• Causal modeling involves the development and ring given a different condition.
testing of a hypothesized causal explanation of • Survival analysis and other related event histo-
a phenomenon. ry methods are used when the dependent vari-
• Path analysis, a regression-based method for able of interest is a time interval (e.g., time from
testing causal models, involves the preparation onset of a disease to death).
of a path diagram that stipulates hypothesized
causal links among variables. Path analysis tests STUDY ACTIVITIES
recursive models, that is, ones in which causa-
Chapter 21 of the accompanying Study Guide for
tion is presumed to be unidirectional.
Nursing Research: Principles and Methods, 7th
• Path analysis applies ordinary least-squares
ed., offers various exercises and study suggestions
regression to a series of structural equations,
for reinforcing the concepts presented in this chap-
resulting in a series of path coefficients. The
ter. In addition, the following study questions can
path coefficients represent weights associated
be addressed:
with a causal path, in standard deviation units.
• Linear structural relations analysis (LIS- 1. Refer to Figure 21-1. What would the value of
REL), another approach to causal modeling, Y be for the following X values: 8, 1, 3, 6?
does not have as many assumptions and restric- 2. A researcher has examined the relationship
tions as path analysis. LISREL can accommo- between preventive health care attitudes on the
date measurement errors, nonrecursive models one hand and the person’s educational level, age,
that allow for reciprocal causal paths, and corre- and gender on the other. The multiple correla-
lated errors. tion coefficient is .62. Explain the meaning of
• LISREL can analyze causal models involving this statistic. How much variation in attitudinal
latent variables, which are not directly scores is explained by the three predictors? How
measured but which are captured by two or more much is unexplained? What other variables
manifest variables (i.e., measured variables). might improve the power of the prediction?
CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 545
3. Using power analysis, determine the sample Hair, J. F., Anderson, R. E., Tatham, R. L., & Black, W.
size needed to achieve power .80 for (1998). Multivariate data analysis (5th ed.). Upper
.05, when: Saddle River, NJ: Prentice-Hall.
a. estimated R2 .15, k 5 Harrell, F. E. (2001). Regression modeling strategies:
b. estimated R2 .33, k 10 With applications to linear models, logistic regression,
and survival analysis. New York: Springer-Verlag.
4. Which multivariate statistical procedures
Holt, F., Merwin, E., & Stern S. (1996). Alternative sta-
would you recommend using in the following
tistical methods to use with survival data. Nursing
situations: Research, 45, 345–349.
a. A researcher wants to test the effectiveness Hougaard, P. (2000). Analysis of multivariate survival
of a nursing intervention for reducing stress data. New York: Springer-Verlag.
levels among surgical patients, using an Jaccard, J., & Becker, M. A. (2001). Statistics for the
experimental group of patients from one behavioral sciences (4th ed.). Belmont, CA:
hospital and a control group from another Wadsworth.
hospital. Jöreskog, K. G., & Sörbom, D. (1997). LISREL 8: User’s
b. A researcher wants to predict which college reference guide. Mooresville, IN: Scientific Software
students are at risk of a sexually transmitted International.
disease using background information such Kelloway, E. K. (1998). Using LISREL for structural
as gender, academic grades, socioeconomic equation modeling: A researcher’s guide. Thousand
status, religion, and attitudes toward sexual Oaks, CA: Sage.
activity. Kim, J., Kaye, J., & Wright L. K. (2001). Moderating
c. A researcher wants to test the effects of and mediating effects in causal models. Issues in
Mental Health Nursing, 22, 63–75.
three different diets on blood sugar levels
Knapp, T. R. (1994). Regression analysis: What to
and blood pH.
report. Nursing Research, 43, 187–189.
Knapp, T. R., & Campbell-Heider, N. (1989). Numbers
SUGGESTED READINGS of observations and variables in multivariate analy-
ses. Western Journal of Nursing Research, 11,
Methodologic References 634–641.
Aaronson, L. S. (1989). A cautionary note on the use of Lee, E. T. (1992). Statistical methods for survival data
stepwise regression. Nursing Research, 38, 309–311. analysis (2nd ed.). New York: John Wiley & Sons.
Aaronson, L. S., Frey, M., & Boyd, C. J. (1988). Mason-Hawkes, J. & Holm, K. (1989). Causal modeling:
Structural equation models and nursing research: Part A comparison of path analysis and LISREL. Nursing
II. Nursing Research, 37, 315–318. Research, 38, 312–314.
Blalock, H. M., Jr. (1972). Causal inferences in nonex- Musil, C. M., Jones, S. L., & Warner, C. D. (1998).
perimental research. New York: W. W. Norton. Structural equation modeling and its relationship to
Bollen, K., & Bollen, W. (1989). Structural equations multiple regression and factor analysis. Research in
with latent variables. New York: John Wiley. Nursing & Health, 21, 271–281.
Boyd, C. J., Frey, M. A., & Aaronson, L. S. (1988). Pedhazur, E. J. (1997). Multiple regression in behavioral
Structural equation models and nursing research: Part research (3rd ed.). New York: Harcourt College
I. Nursing Research, 37, 249–253. Publishers.
Cohen, J., & Cohen, P. (1984). Applied multiple regres- Polit, D. F. (1996). Data analysis and statistics for nurs-
sion: Correlation analysis for behavioral sciences ing research. Stamford, CT: Appleton & Lange.
(2nd ed.). Hillsdale, NJ: Lawrence Erlbaum. Tabachnick, B. G., & Fidell, L. S. (2000). Using multi-
Goodwin, L. D. (1984). Increasing efficiency and variate statistics (4th ed.). New York: HarperCollins
precision of data analysis: Multivariate vs. univariate College Publishers.
statistical techniques. Nursing Research, 33, Weisberg, S. (1985). Applied linear regression (2nd ed.).
247–249. New York: John Wiley & Sons.
546 ■ PART 5 The Analysis of Research Data
Welkowitz, J., Ewen, R. B., & Cohen, J. (2000). Johnson, J. L., Budz, B., Mackay, M., & Miller, C.
Introductory statistics for the behavioral sciences (1999). Evaluation of a nurse-delivered smoking ces-
(5th ed.). New York: Harcourt College Publishers. sation intervention for hospitalized patients with car-
Wu, Y. B., & Slakter, M. J. (1989). Analysis of covariance diac disease. Heart & Lung, 28, 55–64.
in nursing research. Nursing Research, 38, 306–308. Lucas, J. A., Orshan, S. A., & Cook, F. (2000).
Determinants of health-promoting behavior among
women ages 65 and above living in the community.
Studies Cited in Chapter 21
Scholarly Inquiry for Nursing Practice, 14, 77–100.
Aminzadeh, F., & Edwards, N. (2000). Factors associated Mahon, N. E., & Yarcheski, A. (2001). Outcomes of
with cane use among community dwelling older depression in early adolescents. Western Journal of
adults. Public Health Nursing, 17, 474–483. Nursing Research, 23, 360–375.
Badger, T. A., & Collins-Joyce, P. (2000). Depression, McCarter-Spaulding, D. E. & Kearney, M. H. (2001).
psychosocial resources, and functional ability in Parenting self-efficacy and perception of insufficient
older adults. Clinical Nursing Research, 9, 238–255. breast milk. Journal of Obstetric, Gynecologic, and
Badr (Zahr), L. K. (2001). Quantitative and qualitative Neonatal Nursing, 30, 515–522.
predictors of development for low birth-weight Misener, T. R., & Cox, D. L. (2001). Development of the
infants of Latino background. Applied Nursing Misener Nurse Practitioner Job Satisfaction Scale.
Research, 14, 125–135. Journal of Nursing Measurement, 9, 91–108.
Berger, A. M., & Walker, S. N. (2001). An explanatory Nyamathi, A., Leake, B., Longshore, D., & Gelberg, L.
model of fatigue in women receiving adjuvant breast (2001). Reliability of homeless women’s reports:
cancer chemotherapy. Nursing Research, 50, 42–52. Concordance between hair assay and self report of
Cowan, M. J., Pike, K., & Budzynski, H. K. (2001). cocaine use. Nursing Research, 50, 165–171
Psychosocial nursing therapy following sudden car- Tarkka, M. T., Paunonen, M., & Laippala, P. (2000).
diac arrest: Impact on two-year survival. Nursing First-time mothers and child care when the child is 8
Research, 50, 68–76. months old. Journal of Advanced Nursing, 31, 20–26.
Fuller, B. F. (2000). Fluctuations in established infant pain
behaviors. Clinical Nursing Research, 9, 298–316.