Patrick Dattalo - Analysis of Multiple Dependent Variables

Download as pdf or txt
Download as pdf or txt
You are on page 1of 191
At a glance
Powered by AI
The document discusses various research methods and statistical techniques used in social work research.

The document discusses research methods that are used in social work research.

Methods like structural equation modeling, multiple regression, qualitative research methods and more are discussed.

Analysis of Multiple Dependent Variables

POCKET GUIDES TO
SOCIAL WORK RESEARCH METHODS
Series Editor
Tony Tripodi, DSW
Professor Emeritus, Ohio State University
Determining Sample Size: The Dissertation:
Balancing Power, Precision, and Practicality From Beginning to End
Patrick Dattalo Peter Lyons and Howard J. Doueck
Preparing Research Articles Cross-Cultural Research
Bruce A. Thyer Jorge Delva, Paula Allen-Meares, and
Systematic Reviews and Meta-Analysis Sandra L. Momper
Julia H. Littell, Jacqueline Corcoran, and Secondary Data Analysis
Vijayan Pillai Thomas P. Vartanian
Historical Research Narrative Inquiry
Elizabeth Ann Danto Kathleen Wells
Confirmatory Factor Analysis Structural Equation Modeling
Donna Harrington Natasha K. Bowen and Shenyang Guo
Randomized Controlled Trials: Finding and Evaluating Evidence:
Design and Implementation for Systematic Reviews and
Community-Based Psychosocial Evidence-Based Practice
Interventions Denise E. Bronson and Tamara S. Davis
Phyllis Solomon, Mary M. Cavanaugh, Policy Creation and Evaluation:
and Jeffrey Draine Understanding Welfare Reform in the
Needs Assessment United States
David Royse, Michele Staton-Tindall, Richard Hoefer
Karen Badger, and J. Matthew Webster Grounded Theory
Multiple Regression with Discrete Julianne S. Oktay
Dependent Variables Systematic Synthesis of Qualitative
John G. Orme and Terri Combs-Orme Research
Developing Cross-Cultural Measurement Michael Saini and Aron Shlonsky
Thanh V. Tran Quasi-Experimental Research Designs
Intervention Research: Bruce A. Thyer
Developing Social Programs Conducting Research in Juvenile and
Mark W. Fraser, Jack M. Richman, Maeda J. Criminal Justice Settings
Galinsky, and Steven H. Day Michael G. Vaughn, Carrie Pettus-Davis,
Developing and Validating Rapid and Jeffrey J. Shook
Assessment Instruments Qualitative Methods for Practice
Neil Abell, David W. Springer, Research
and Akihito Kamata Jeffrey Longhofer, Jerry Floersch,
Clinical Data-Mining: and Janet Hoy
Integrating Practice and Research Culturally Competent Research:
Irwin Epstein Using Ethnography as a Meta-Framework
Strategies to Approximate Random Mo Yee Lee and Amy Zaharlick
Sampling and Assignment Using Complexity Theory for Research and
Patrick Dattalo Program Evaluation
Analyzing Single System Design Data Michael Wolf-Branigin
William R. Nugent Analysis of Multiple Dependent
Survival Analysis Variables
Shenyang Guo Patrick Dattalo
PATRICK DATTALO

Analysis of Multiple
Dependent Variables

1
3
Oxford University Press is a department of the University of Oxford.
It furthers the University’s objective of excellence in research, scholarship,
and education by publishing worldwide.

Oxford New York


Auckland Cape Town Dar es Salaam Hong Kong Karachi
Kuala Lumpur Madrid Melbourne Mexico City Nairobi
New Delhi Shanghai Taipei Toronto

With offices in
Argentina Austria Brazil Chile Czech Republic France Greece
Guatemala Hungary Italy Japan Poland Portugal Singapore
South Korea Switzerland Thailand Turkey Ukraine Vietnam

Oxford is a registered trademark of Oxford University Press in the UK and certain other
countries.

Published in the United States of America by


Oxford University Press
198 Madison Avenue, New York, NY 10016

© Oxford University Press 2013


All rights reserved. No part of this publication may be reproduced, stored in a
retrieval system, or transmitted, in any form or by any means, without the prior
permission in writing of Oxford University Press, or as expressly permitted by law,
by license, or under terms agreed with the appropriate reproduction rights organization.
Inquiries concerning reproduction outside the scope of the above should be sent to the
Rights Department, Oxford University Press, at the address above.

You must not circulate this work in any other form


and you must impose this same condition on any acquirer.

Library of Congress Cataloging-in-Publication Data


Dattalo, Patrick.
Analysis of multiple dependent variables / Patrick Dattalo.
p. cm.—(Pocket guides to social work research methods)
Includes bibliographical references and index.
ISBN 978–0–19–977359–6 (pbk. : alk. paper)
1. Social service—Statistical methods. 2. Multivariate analysis.
I. Title.
HV29.D384 2013
519.5’35—dc23
2012029873

1 3 5 7 9 8 6 4 2
Printed in the United States of America
on acid-free paper
For Marie
This page intentionally left blank
Contents

1 Basic Concepts and Assumptions 1

2 Multivariate Analysis of Variance 22

3 Multivariate Analysis of Covariance 63

4 Multivariate Multiple Regression 87

5 Structural Equation Modeling 109

6 Choosing among Procedures for the Analysis


of Multiple Dependent Variables 149

References 157

Index 171

vii
This page intentionally left blank
Analysis of Multiple Dependent Variables
This page intentionally left blank
1

Basic Concepts and


Assumptions

INTRODUCTION
Multivariate procedures allow social workers and other human services
researchers to analyze multidimensional social problems and interven-
tions in ways that minimize oversimplification. The term multivariate
is used here to describe analyses in which several dependent variables
(DVs) are being analyzed simultaneously. The term univariate is used to
describe analyses in which one DV is being analyzed. DVs are also called
criterion, response, y-variables, and endogenous variables. Independent
variables (IVs) are also called predictor, explanatory, x-variables, and
exogenous variables. For consistency, only the terms independent and
dependent variables will be used here. Examples of multivariate analy-
ses include investigations of relationships between personality traits and
aspects of behavior, program characteristics and effectiveness, and client
characteristics and service needs. The purposes of multivariate analyses
include (1) data structure simplification and reduction, (2) description
of relationships, and (3) prediction.
Using multivariate analysis for data structure simplification and reduc-
tion involves identifying and interpreting concepts called latent variables,
or emergent variables. According to Huberty (1994), “what is being sought
is an attribute or trait that may not be directly observable, and is not

1
2 Analysis of Multiple Dependent Variables

tangible; it is a latent variable often considered to comprise observable


variables that are in some way combined (usually in a linear manner)”
(p. 623).
Using multivariate analysis for description involves identifying and
specifying the relative importance of IVs to anticipate the values of DVs;
that is, description involves determining an ordering of IVs by answering
the question: How well does the estimated relationship between a set of
IVs and a set of DVs (i.e., a model) perform without each IV?
Using multivariate analysis for prediction involves identifying and
specifying one or more IVs that can help anticipate the values of DVs.
That is, the purpose of predictive analysis is to model the relationship
between one or more IVs and one or more DVs, and to use this model to
estimate values of the DVs when only values of the IVs are available.
Examples of multivariate statistical procedures to predict and describe
relationships include multivariate multiple regression (MMR), multi-
variate analysis of variance (MANOVA), and multivariate analysis of
covariance (MANCOVA). Structural equation modeling (SEM) may be
used for data simplification and reduction, description, and prediction.

RATIONALE FOR MULTIVARIATE ANALYSIS


When multiple DVs are being modeled, a researcher faces a choice among
whether to (1) analyze each DV separately; (2) aggregate the DVs before
analysis (e.g., compute standard scores for each DV and add the scores
of each DV to obtain a composite score for all DVs); or (3) perform a
multivariate analysis, such as MANOVA. One approach that can help to
clarify which of the aforementioned three choices is best is to examine
the correlations among DVs. Low correlations (Pearson’s r < .20) suggest
the need to analyze each DV separately; moderate correlations (r = .20 to
.50) suggest the need to analyze DVs together (i.e., a multivariate analy-
sis); and high positive correlations (r > .50) suggest the need to aggregate
the DVs before analysis (Stevens, 2009).
For example, MANOVA is conducted when the multiple DVs are
conceptually or statistically related. When the DVs represent different
dimensions of a multidimensional concept (e.g., clients satisfaction),
they are conceptually related. DVs are statistically correlated when a mea-
sure of association, such as Pearson’s, r, results . . . in a weak to moderate
Basic Concepts and Assumptions 3

coefficient. Whether a researcher conducts a multivariate analysis or


multiple univariate analyses the intercorrelations among DVs should be
reported (Huberty & Morris, 1989; Tabachnick & Fidell, 2007).
When there are moderate to low correlations among DVs, and a
researcher wishes to retain separate measures of these DVs, an additional
benefit is multiple operationalism. Operationalism refers to the process
of representing constructs by a specific set of steps or operations. In other
words, operationalism helps a researcher clarify the process of using
empirical data to measure a concept (i.e., an abstract idea or theoretical
construct), and to allow others to evaluate the validity of that measure-
ment process.
Multiple Operationalism. This term refers to the use of two or more
measures to represent a concept. From this perspective, the use of multi-
ple measures of a single concept provides a better chance of fully repre-
senting that concept. Campbell and Fiske (1959) explain that

“For any data taken from a single operation, there is a subinfinity of


interpretations possible; a subinfinity of concepts, or combinations of
concepts, that it could represent. Any single operation, as representative
of concepts, is equivocal.” (p. 101)

Boring (1953) stated, “ . . . as long as a new construct has only the sin-
gle operational definition that it received at birth, it is just a construct;
when it gets two alternative operational definitions, it is beginning to be
validated” (p. 183).
Multiple DVs may be suggested by theory, previously conducted
empirical investigations, or practice experience. Moreover, as Campbell
and Fiske (1959) have suggested, even if the primary interest is in only
one DV, multiple operationalization may provide a more valid assess-
ment of that DV. However, difficulty may arise in the interpretation of
relationships among multiple DVs if there is reason to suspect that these
variables are correlated. Consequently, the examination of correlations
among DVs may provide a greater understanding than can be attained by
considering these variables separately. The four multivariate procedures
discussed in this book facilitate the analysis of correlated DVs.
Control of Type I Error. A second reason to conduct a multivariate
analysis is to control type I error. Type I error is the probability of incor-
rectly identifying a statistically significant effect. Type I error is a false
4 Analysis of Multiple Dependent Variables

positive, and is the process of incorrectly rejecting the null hypothesis


in favor of the alternative or research hypothesis. In an empirical inves-
tigation in which two or more comparisons are made from the data,
there are two kinds of type I error: (1) the comparison-wise error rate
is the probability of a type I error set by the researcher for evaluating
each comparison, and (2) the study-wise error rate (sometimes termed
experiment-wise error rate) is the probability of making at least one
type I error when performing a set of comparisons. The study-wise error
rate (SW) is defined as 1 − (1 − α)k, where α (alpha) is the significance
level used for each test and k is equal to the number of comparisons.
For example, if alpha is defined as p < 0.05, and tests are assumed to be
independent, then, for a model with five predictor variables, SW ≤ 1 −
(1 − 0.05)5, or 1 − 0.77 = 0.33. Therefore, whenever multiple inferential
statistical tests are performed, there is a potential problem of probability
pyramiding. Use of conventional levels of type I error probabilities for
each test in a series of statistical tests may yield an unacceptably high type
I error probability across all of the tests. If a researcher has a sound theo-
retical reason for performing sequential univariate or bivariate hypoth-
esis tests, then he or she might consider an adjustment to the overall type
I error probability. One approach is to use an additive inequality (i.e.,
Bonferroni). For m tests, the alpha level for each test (α1) is given by
the overall alpha level (αm) divided by m; this function of α1 should be
considered as an upper bound for the overall alpha, αm. Approaches to
controlling type I error are discussed further in chapter 2.
Multivariate statistical procedures, however, are not panaceas; there
are costs associated with their benefits. For, example, Bray and Maxwell
(1985) state,

. . . the indiscriminant use of multiple dependent measures is never an


appropriate substitute for a well-conceived study with a select number of
variables . . . In fact, the use of too many dependent variables may hinder
the chances of finding a significant result because of low statistical power
or it may result in spurious findings due to chance. (pp. 9–10)

Following Enders (2003), the perspective taken here is to strongly encour-


age researchers to “think carefully when thinking multivariately” (p. 53).
Accordingly, which procedure is most appropriate depends on the (1) pur-
pose of the analysis; (2) sample size; and (3) tenability of assumptions.
Basic Concepts and Assumptions 5

These three issues will be discussed for MANOVA, MANCOVA, MMR,


and SEM, which will be presented as alternative statistical procedures for
analyzing models with more than one DV (see the chapter summaries
below).
The remaining portion of this chapter is organized as follows: (1)
audience and background; (2) level and type of examples; (3) defini-
tion of basic terms and concepts; (4) data assumptions; and (5) chapter
summaries.

ORGANIZATION OF THE BOOK


Audience and Background
This book is intended primarily for producers and consumers of social
work research, although it is expected that it will also be useful to other
fields that use experiments and quasiexperiments, such as psychology,
education, evaluation, medicine, and other health-related disciplines.
Readers with a background in basic statistics are the primary audience.
Ideally, readers will have taken a graduate-level statistics course with
content that includes an introduction to ANOVA and regression model-
ing. However, each procedure will be presented here in sufficient detail
to enable readers to implement that procedure without supplemental
reading.

Level and Type of Examples


Discussion draws on the social work literature and on the literature from
a variety of related disciplines. Note that all data and related files are
posted on this book’s companion website. Detailed social work–related
examples are as follows:

1. Annotated examples of MANOVA and MANCOVA are provided


using PASW. Computer output is presented as well as an
explanation of how the output was produced. Instructions for
using Stata and SAS to conduct MANOVA and MANCOVA also
are provided;
2. An annotated example of MMR is provided using Stata.
Computer output is presented as well as an explanation of how
6 Analysis of Multiple Dependent Variables

the output was produced. Instructions for using PASW and SAS
to conduct MMR also are provided; and
3. An annotated example of SEM is provided using AMOS.
Computer output is presented as well as an explanation of how
the output was produced.

Assumptions
The term statistical model is used here to describe mathematical state-
ments about how variables are associated. Ideally, models are constructed
to capture a pattern of association suggested by a theory. Models are con-
structed and fitted to statistically test theories that purport to explain
relationships between and among variables. Statistical models include IVs
and DVs. Variables also may be described in terms of order. Higher-order
variables (e.g., interactions or moderators) are constructed from
lower-order variables (i.e., main effects).
Model building concerns strategies for selecting an optimal set of IVs
that explain the variance in one or more DVs (Schuster, 1998). A statisti-
cal model is said to be misspecified if it incorrectly includes unimportant
variables or excludes important variables (Babyak, 2004). Including more
variables than are needed is termed overfitting. Overfitting occurs when
a model contains too many IVs in relation to sample size, and when IVs
are intercorrelated. An overfitted model captures not only true regulari-
ties reflected in the data, but also chance patterns that reduce the model’s
predictive accuracy. Overfitting yields overly optimistic model results;
that is, findings that appear in an overfitted model may not exist in the
population from which a sample has been drawn. In contrast, underfit-
ting occurs when a model has too few IVs. Underfitting results in omis-
sion bias, and, consequently, may have poor predictive ability because of
the lack of detail in a regression model.
The simplest statistical model consists of IVs whose relationships
with a DV are separate, and, consequently, there is additivity, or no inter-
action. A more complicated model is when the effect of one IV (X1) on a
DV (Y) depends on another IV (X2). The relationship between X1 and X2
with Y is called a statistical interaction (or moderation). In a statistical
test of interaction, the researcher explores this interaction in the regres-
sion model by multiplying X1 by X2, which produces X1X2, which is called
Basic Concepts and Assumptions 7

an interaction of moderator term. Conversely, additivity means that the


effect of one IV on the DV does not depend on the value of another IV.
In a subsequent discussion, data being analyzed are assumed to follow
the general linear model (GLM). One useful perspective on the GLM is
as an extension of linear multiple regression, sometimes termed ordinary
least squares regression. Multiple regression models a DV as a function
of a set of IVs. This model seeks to explain a proportion of the variance
in a DV at a statistically significant level, and can establish the relative
predictive importance of the IVs.
The multiple regression equation for sample estimates takes the form

y = b1x1 + b2x2 + ⋅ ⋅ ⋅ + bnxn + c (1.1)

The bi’s are the regression coefficients, representing the amount the DV
(y) changes when the IV (e.g., x1) changes one unit. The c is the constant,
or where the regression line intercepts the y-axis, and represents the value
of the DV(y) when all IVs equal zero.
Ordinary least squares (OLS) regression derives its name from the
method used to estimate the best-fit regression line: a line such that
the sum of the squared deviations of the distances of all the points from
the line is minimized. More specifically, the regression surface (i.e., a line
in simple regression, or a plane or higher-dimensional surface in mul-
tiple regression) expresses the best predicted value of the DV (Y), given
the values on the IVs (X’s). However, phenomena of interest to social
workers and other applied researchers are never perfectly predictable,
and usually there is variation between observed values of a DV and those
values predicted by OLS regression. The deviation of observed versus pre-
dicted values is called the residual value. Because the goal of regression
analysis is to fit a linear function of the X variables as closely as possible
to the observed Y variable, the residual values for the observed points can
be used to devise a criterion for the “best fit.” Specifically, in regression
problems, the linear function is computed for which the sum-of-squared
deviations of the observed points from values predicted by that func-
tion are minimized. Consequently, this general procedure is referred to
as least squares estimation. Power terms (e.g., x2, x3) can be added IVs to
explore curvilinear effects. Cross-product terms (e.g., x1x2) can be added
as IVs to explore interaction effects.
8 Analysis of Multiple Dependent Variables

The GLM may be extended to include multiple DVs (i.e., multi-


variate). This extension also allows multivariate tests of significance to be
calculated when responses on multiple DVs are correlated. Separate uni-
variate tests of significance for correlated DVs are not independent and
may result in an erroneous decision about the statistical significance of
predicted relationships. Multivariate tests of significance of linear combi-
nations of multiple DVs also may provide insight into which dimensions
of the DVs are, and are not, related to the IVs.
Important assumptions of the GLM, including (1) independence of
observations, (2) model is specified correctly, (3) no missing data or data
missing at random, (4) sound measurement, (5) orthogonality, (6) mul-
tivariate normality, (7) absence of outliers, (8) linearity, and (9) homo-
scedasticity (Stevens, 2009; Tabachnick & Fidell, 2007) are summarized
below. Assumptions and remedies for the violation these assumptions are
discussed further within the context of MANOVA, MANCOVA, MMR,
and SEM.
Independence of Observations. This assumption is that the selection
of any study participant is not related to the selection of any other partic-
ipant. Kenny and Judd (1986) discuss three commonly assumed patterns
of nonindependence: (1) nonindependence due to groups, (2) non-
independence due to sequence, and (3) nonindependence due to space.
Nonindependence due to groups can occur for a variety of reasons. For
example, if members of the same group (e.g., students in a single school)
are sampled, these students may be more similar in terms of measured
and unmeasured characteristics. If characteristics are unmeasured, than
the errors associated with their selection are not independent. If children
are grouped in classrooms, group linkage by classrooms is likely because
of a common teacher, classroom, or textbook. Social interaction may
result in group linkage. In research on small groups, observations within
groups are likely to be linked because the members of the groups interact
during the course of the experimental session. Nonindependence due to
sequence occurs when observations are repeatedly taken from a single unit
over time. Such a data structure is commonly referred to as time-series
data. Nonindependence due to space can occur when observations are
made of units arranged in space (e.g., houses on a block, blocks within a
city, or states within a country) may be spatially nonindependent. That is,
observations that are nearer together in space may be more or less similar
than those farther away in space.
Basic Concepts and Assumptions 9

According to Stevens (2009), violation of the independence of obser-


vations assumption is very serious. Only a small amount of dependence
among observations has a dramatic effect on type I error. Specifically,
when errors are positively correlated, nonindependence of within-group
sample values increases type I error rates and at the same time decreases
type II error probabilities. Moreover, the aforementioned situation does
not improve with larger sample size; it becomes worse. Consequently, if
there is reason to suspect dependence among observations, the researcher
should consider using a more stringent level of significance (e.g., .01
rather than .05).
The independence of observations assumption fundamentally is a
design issue (i.e., it is based on the way data are collected). Consequently,
if nonindependence of observations is suspected, the most effective strat-
egy is to use a research design that explicitly incorporates nonindepen-
dent (i.e., correlated) observations, such as multilevel modeling (Hox,
2002). For example, medical research applications often involve hierar-
chical data structures (i.e., nested) such as patients within hospitals or
physicians within hospitals; for example, assessing differences in mor-
tality rates across hospitals relative to a specific condition or procedure.
Data are collected on random samples of patients nested within each
hospital. In this application, it might be appropriate to adjust for covari-
ates at both the patient-level (such as patient age, patient gender, and the
severity of the index diagnosis) and at the hospital-level (such as hospital
size and hospital teaching status). Hierarchical linear models, sometimes
called multilevel linear models, nested models, mixed linear models,
random-effects models, random parameter models, split-plot designs,
or covariance components models handle parameters that vary at more
than one level.
The basic logic behind these models is that a group mean (i.e., the
potential contextual effect) can explain unique variance over and above
an individual variable of the same name (Bickel, 2007). So, for instance,
average group work hours may explain unique variance in individual
well-being above individual reports of work hours. This occurs because
there is no mathematical reason why the group-level relationship between
means must be the same as the individual-level relationship. When
the slope of the group-mean relationship differs from the slope of the
individual-level relationship, then a contextual effect is present. To esti-
mate contextual regression models, regression is used to simultaneously
10 Analysis of Multiple Dependent Variables

test the significance of the individual and group mean variables. If the
group-mean variable is significant in this model it indicates that the
individual-level and group-level slopes are significantly different, and
one has evidence of a contextual effect.
Model Is Specified Correctly. As discussed above, this assumption
concerns the accuracy of the statistical model being tested. Several tests
of specification have been proposed. Ramsey’s Regression Specification
Error Test (RESET) is an F-test of differences of R2 under linear versus
nonlinear assumptions (Ramsey, 1969). More specifically, the RESET
evaluates whether nonlinear combinations of the estimated values help
explain the DV. The basic assumption of the RESET is that if nonlinear
combinations of the explanatory variables have any power in explaining
the endogenous variable, then the model is misspecified. That is, for a
linear model that is properly specified, nonlinear transforms of the fit-
ted values should not be useful in predicting the DV. While Stata, for
example, labels the RESET as a test of omitted variables, it only tests if
any nonlinear transforms of the specified DVs variables have been omit-
ted. RESET does not test for other relevant linear or nonlinear variables
that have been omitted.
The RESET (and other specification tests) should not be viewed as
a substitute for a good literature review, which is critical to identify a
model’s variables. As a rule of thumb, the lower the overall effect (e.g.,
R2 in multiple regression with a single DV), the more likely it is that
important variables have been omitted from the model and that exist-
ing interpretations of the model will change when the model is correctly
specified. Perhaps, specification error may be best identified when the
research relies on model comparison as opposed to the testing of one
model to assess the relative importance of the IVs.
Missing Data. Missing data in empirical social work research can
substantially affect results. Common causes of missing data include par-
ticipant nonresponse and research design. That is, some participants may
decline to provide certain information, some data (e.g., archival) may
not be available for all participants, and information may be purposely
censored, for example, to protect confidentiality. There is a rich body of
literature on missing data analysis. However, there is no consensus in the
methodological literature about what constitutes excessive missingness
(Enders, 2010; Little & Rubin, 2002). Based on whether missing data are
dependent on observed values, patterns of missing data are classified into
Basic Concepts and Assumptions 11

three categories: missing completely at random (MCAR), missing at


random (MAR), and missing not at random (MNAR).
Missing completely at random (MCAR) exists when missing values are
randomly distributed across all observations. Missing at random (MAR)
is a condition that exists when missing values are not randomly distrib-
uted across all observations but are randomly distributed within one or
more subsamples (i.e., missing more among whites than non-whites, but
random within each subsample). When neither data are not MCAR or
MAR, they are systematically missing or missing not at random (MNAR)
(Heitjan, 1997; Kline, 2011). Biased parameter estimates may result in
data that are systematically missing.
Missing data can be handled by either deletion or imputation tech-
niques (Kline, 1998; Little & Rubin, 1987). Deletion techniques involve
excluding participants with missing data from statistical calculations.
Imputation is the process of replacing missing data with values that
are based on the values of other variables in a data set. Unlike dele-
tion, imputation retains sample size, and minimizes loss of statistical
power. Imputation techniques are generally classified as either single
or multiple. Single imputation assigns one estimate for each missing
data point. Basically, multiple imputation is an extension of the single
imputation idea, in which each missing value is replaced by a set of
m > 1 plausible values to generate m apparently complete data sets.
These m data sets are then analyzed by standard statistical software, and
the results are combined using techniques suggested by Rubin (1987) to
give parameter estimates and standard errors that take into account the
uncertainty attributable to the missing data values. There is an increas-
ing availability of software to perform multiple imputation, includ-
ing PASW version 19. Free software to perform multiple imputation
includes NORM, which is available as a standalone or an S-Plus version.
Version 2.03 of NORM can be downloaded from http://sites.stat.psu.
edu/~jls/norm203.exe
Sound Measurement. Data analysis is influenced at least in part by
how and how well concepts have been operationalized. A thorough dis-
cussion of measurement theory and psychometric technique is beyond
the scope of this text. However, one important principle related to how
concepts have been operationalized is level of measurement. Two impor-
tant principles related to how well concepts have been operationalized are
reliability and validity.
12 Analysis of Multiple Dependent Variables

Levels of measurement are expressions that refer to the theory of


scale types developed by the psychologist Stanley Smith Stevens. Stevens
(1946, 1951) argued that all empirical measures may be categorized
as nominal, ordinal, interval, or ratio, from lowest to highest levels of
measurement. A variable may be defined as an operationalized concept
that changes (e.g., gender) and therefore is a logical grouping of alterna-
tive empirical values (e.g., female or male). The level of measurement,
then, refers to the relationship among these possible alternative empiri-
cal values. Stevens argued that an understanding of a variable’s level
of measurement helps to clarify how a variable changes and therefore,
elucidates how a concept is empirically measured. Knowing the level of
measurement of a variable helps to identify which statistical procedure
is most appropriate, since the level of measurement determines the level
of mathematical precision.
Stevens’s theory of scale types posits a hierarchy of levels of measure-
ment. The GLM assumes that variables are operationalized at the highest
practical level of measurement. At lower levels of measurement, assump-
tions tend to be less restrictive and data analyses tend to be less sensitive.
As one moves from lower to higher levels in the hierarchy, the current
level includes all of the characteristics of the one below it, plus something
new. At the nominal level, alternative empirical values are mutually exclu-
sive categories; no ordering of alternative values (e.g., more versus less) is
implied. For example, gender is a nominal level variable. At the ordinal
level alternative empirical values may be ranked. For example, on a sur-
vey you might code educational attainment as 0 = less than high school,
1 = some high school, 2 = high school degree; 3 = some college; 4 = col-
lege degree; 5 = post college. In this measure, higher numbers mean more
education. But, the distance between values is not assumed to be equal.
At the interval level the distance between values is assumed to be equal.
For example, on the Fahrenheit temperature scale, the distance from 30
to 40 is same as distance from 70 to 80. The interval between values is
interpretable, and, consequently, the average of a group of values on an
interval variable may be computed. At the ratio level of measurement
there is an absolute zero. This means that you can construct a meaning-
ful fraction (or ratio) with a ratio variable. Weight is a ratio variable. It
should be noted that although Stevens’s classification is widely adopted,
it is not universally accepted (cf. Velleman & Wilkinson, 1993; Chrisman,
1998).
Basic Concepts and Assumptions 13

Reliability refers to consistency. A reliable measure is one that yields


consistent results (given an unchanged measured phenomenon) over
repeated measurements. Validity refers to accuracy. A valid measure is
one that measures what it purports to measure. Measurement reliability
is necessary, but not sufficient, for measurement validity. Accordingly,
for example, if scale does not accurately measure what it is supposed
to measure, there is no reason to use it, even if it measures consistently
(reliably). It is expected that measures are sufficiently valid, and reliabil-
ity is a logical first step for demonstrating validity (i.e., to be accurate a
measure must be consistent).
A valid measure (e.g., scale), which by definition is a reliable measure,
should yield an outcome (e.g., score) that has minimum measurement
error. According to classical measurement theory, an observed score is
composed of (1) a true value and (2) error. It is assumed that measure-
ment error is minimal. Measurement error is either systematic or ran-
dom. As the term implies, random error differentially affects members
of a sample. For instance, study participants’ moods can inflate or deflate
their performance. Because the sum of all random errors in a distribu-
tion equals zero, random error adds variability to data, but do not affect
average performance. Consequently, random error is sometimes referred
to as noise. In contrast, systematic error tends to affect all members of
a sample. Consequently, systematic measurement error is problematic.
Systematic error can be detected by assessing reliability and validity.
A detailed discussion of strategies to assess measurement reliabil-
ity and validity is beyond the scope of this chapter, but commonly used
approaches to assessing reliability include internal consistency, test–
retest, and parallel forms. Commonly used approaches to assessing valid-
ity include face, content, criterion, construct, and factorial. See Rubin and
Babbie (2010) for a detailed discussion of strategies to assess reliability
and validity.
Commonly used strategies to minimize both random and system-
atic measurement error include: (1) pilot testing of measurement instru-
ments by collecting feedback from respondents about clarity in terms
of question organization and meaning; (2) ensuring consistency in how
measurement instruments are used by training proctors, interview-
ers, and observers; (3) verifying data are inputted accurately for com-
puter analysis; and (4) using multiple measures of the same concept to
triangulate responses.
14 Analysis of Multiple Dependent Variables

Absence of Multicollinearity is when variables are highly correlated,


and singularity is when variables are perfectly correlated. As a rule of
thumb, intercorrelation among IVs above .80 signals a possible problem.
Likewise, high multicollinearity is signaled when high R2 and significant
F-tests of the model occur in combination with nonsignificant t-tests of
coefficients. Large standard errors because of multicollinearity result in
a reduced probability of rejecting the null hypothesis (i.e., power) and
wide confidence intervals. Under multicollinearity, estimates are unbi-
ased, but assessments of the relative strength of the explanatory variables
and their joint effect are unreliable. That is, under multicollinearity, beta
weights and R2 cannot be interpreted reliably even though predicted val-
ues are still the best estimate using the given IVs.
Orthogonality. Perfect no association between IVs is preferred so
that each IV adds to the ability of the set of IVs to predict values of the
DV. Consequently, if the relationship between each IV and each DV is
orthogonal, the effect of an individual IV may be isolated.
Commonly used approaches to identifying multicollinearity include
(1) inspection of bivariate correlations among IVS; (2) calculating toler-
ance, which is defined as 1 – R2, where R2 is the multiple R of a given
IV regressed on all other IVs. If the tolerance value is less than some
threshold, usually .20, multicollinearity is suspected; and (3) calculating
the variance inflation factor (VIF), which is the reciprocal of tolerance.
The rule of thumb is that VIF > 4.0 (some researchers prefer VIF > 5.0)
when multicollinearity is a problem.
Multivariate Normality. Univariate normality is a necessary, but not
sufficient, condition for multivariate normality to exist (Stevens, 2009).
Multivariate normality exists when each variable has a normal distribu-
tion about fixed values on all other variables. Two other properties of a
multivariate normal distribution are as follows: (1) any linear combina-
tions of the variables are normally distributed; and (2) all subsets of the
set of variables have multivariate normal distributions. This latter prop-
erty implies that all pairs of variables must be bivariate normal.
A multivariate normal distribution has zero skew and zero kurtosis.
Skew is the tilt (or lack of it) in a distribution. That is, residuals are
not symmetrically distributed about the mean error. Negative skew is
right-leaning and positive skew is left-leaning. An S-shaped pattern of
deviations indicates that the residuals have excessive kurtosis. Kurtosis is
the peakedness of a distribution. Leptokurtosis is a distribution with “fat
Basic Concepts and Assumptions 15

tails,” and there is a relatively small proportion of values both above and
below the mean residual. Platykurtosis is a distribution with “thin tails,”
and there are a relatively large proportion of values both above and below
the mean residual.
According to Stevens (2009), deviation from multivariate normal-
ity has only a small effect on type I error. Multivariate skewness appears
to have a negligible effect on power. However, Olsen (1974) found that
platykurtosis does have an effect on power, and the severity of the effect
increases as platykurtosis spreads from one to all groups.
Several strategies to assess multivariate normality are as follows:

1. Bivariate normality, for correlated variables, implies that the


scatterplots for each pair of variables will be elliptical; the higher
the correlation, the thinner the ellipse. Therefore, as a partial
check on multivariate normality, scatterplots for each pair
of variables could inspected for the hoped for approximately
elliptical pattern;
2. Another approach is to focus on the distribution of a model’s
error or residuals. For a data set, residual values are the
differences between values predicted by a model and observed
values of data. To assess the normality of residuals, predicted
values of a DV are plotted on the y-axis of a graph and residual
values of this same DV are plotted on the x-axis this graph. Under
perfect normality, the plot will be a 45-degree line. Univariate
normality is necessary, but not sufficient, for multivariate
normality. However, the examination of residuals may allow a
researcher to assess both the normality of residuals, and the more
important homoscedasticity assumption discussed below;
3. Looney (1995) suggests using battery of tests of univariate
normality. That is, as a first step in assessing multivariate normality
is to test each variable separately for univariate normality. Of the
many procedures available for assessing univariate normality,
two of the most commonly used are (1) an examination of
skewness and kurtosis, and (2) the Shapiro–Wilk (1965) W test.
Many authors have recommended that both approaches be
used, as neither is uniformly superior to the other in detecting
nonunivariate alternatives. If either test strongly indicates a
departure from normality, then normality is rejected; and
16 Analysis of Multiple Dependent Variables

4. DeCarlo (1997) recommends that, as a first step in assessing


multivariate normality, each variable be separately tested for
univariate normality. Because several tests are performed, a
Bonferroni correction can be used to control the type I error
rate (discussed in detail in chapter 2). Although univariate
normality is a necessary condition for multivariate normality,
it is not sufficient, which means that a nonnormal multivariate
distribution can have normal marginals. Therefore, if univariate
normality is not rejected, then the next step is to check for
multivariate normality. DeCarlo (1997) provides a macro that
can be used to assess both univariate and multivariate normality.
For univariate data, the macro calculates the kurtosis statistic b2,
and the skewness statistic √b1. It also provides two omnibus tests:
K2, which sums the two chi-squares for skewness and kurtosis,
and a score (Lagrange multiplier) test, which is a function of √b1
and b2.

In addition to the univariate statistics, the macro provides the follow-


ing measures of multivariate normality: (a) Mardia’s (1970) multivariate
kurtosis; (b) Srivistava’s (1984) and Small’s (1980) tests of multivari-
ate kurtosis and skew, both of which are discussed by Looney (1995);
(c) an omnibus test of multivariate normality based on Small’s statistics
(see Looney, 1995); (d) a list of the five cases with the largest squared
Mahalanobis distances; (e) a plot of the squared Mahalanobis distances,
which is useful for checking multivariate normality and for detecting
multivariate outliers; and (f) Bonferroni adjusted critical values for test-
ing for a single multivariate outlier by using the Mahalanobis distance, as
discussed by Penny (1996). This macro is demonstrated in the annotated
example of MANOVA in chapter 2.
Absence of Outliers. Unusual or extreme values are defined as obser-
vations that appear to be inconsistent with other observations in the data
set. Outliers can occur by chance in any distribution. In the case of nor-
mally distributed data, roughly 1 in 22 observations will differ by twice
the standard deviation or more from the mean, and 1 in 370 will devi-
ate by three times the standard deviation. Outliers, however, may result
from data entry error. However, outliers also may be the result of incor-
rect distributional assumptions (e.g., treating count data as normally
distributed data).
Basic Concepts and Assumptions 17

MANOVA is sensitive to the effect of outliers, which may affect the


type I error rate; OLS regression analysis also is sensitive to outliers.
Outliers in OLS regression can overstate the coefficient of determination
(R2), and give erroneous values for the slope and intercept.
Within the context of the assumption of the absence of outliers,
there are three important characteristics of potentially errant or seem-
ingly unusual observations (data points). The first is leverage. In regres-
sion, high leverage cases are those with unusual values on the IVs. In
MANOVA, high leverage cases are those with an unusual combination of
scores on the DVs. The second is discrepancy (or distance2) between pre-
dicted and the observed values on the DV. The third is influence, which
reflects the product of leverage and discrepancy.
Lorenz (1987) suggests using Cook’s D as the focus of a procedure for
pre-screening for outliers. Cook’s distance measure D provides an over-
all measure of the impact of an observation on the estimated regression
coefficient. Just because the residual plot or Cook’s distance measure test
identifies an observation as an outlier does not mean one should auto-
matically eliminate the point. Some possible approaches to working with
outliers are as follows:

1. Transformation of data is one way to reduce the impact of outliers


because the most commonly used expressions, square roots and
logarithms, shrink larger values to a much greater extent than
they shrink smaller values. However, transformations may not fit
into the theory of the model or they may affect its interpretation.
Taking the log of a variable does more than make a distribution
less skewed; it changes the relationship between the original
variable and the other variables in the model. In addition, most
commonly used transformations require non-negative data or
data that are greater than zero, so they are not always the answer;
2. Deletion of outliers. When in doubt, report model results with
and without outliers; and
3. Use of methods that are robust in the presence of outliers, such as
robust regression models.

At a minimum, researchers should consider fitting a model with


and without outliers, and compare the coefficients, mean-squared error
and, R2 from the two models. It also may be useful for the researcher
18 Analysis of Multiple Dependent Variables

to conjecture about possible reasons for outliers. Conjecture may facili-


tate conclusions about the representativeness of a sample with outliers.
Conjecture also may assist in the design of future studies by suggesting
possible strategies to prevent missing data.
Linearity. This is the assumption that there is a straight line relation-
ship between variables. Violations of linearity are of concern because if a
linear model is fitted to data that are nonlinearly related, predictions are
likely to have large residuals. Nonlinearity is usually most evident in a
plot of the observed versus predicted values, or a plot of residuals versus
predicted values. Points should be symmetrically distributed around a
diagonal line in the former plot, or a horizontal line in the latter plot.
When nonlinearity is present, one solution is to apply a nonlinear
transformation to the IVs or DVs. For example, if the data are strictly posi-
tive, a log transformation may be feasible. Another possibility is to consider
adding another IV that is a nonlinear function of one of the IVs (e.g., X2).
Homoscedasticity. The assumption that the variability in scores for
one variable is equal at all values of another variable is termed homosce-
dasticity. Violations of homoscedasticity make it difficult to estimate stan-
dard error, and usually results in confidence intervals that are too wide or
too narrow. That is, heteroscedasticity can result in biased parameter esti-
mates, and standard error estimates that are either too large or too small.
Consequently, heteroscedasticity can increase type I error, or increase type
II error. One approach to detecting violations of homoscedasticity is to
examine plots of residuals versus predicted values for evidence that resid-
uals are a function of predicted values. However, heteroscedasticity also
can be a result of a violation of the linearity or normality assumptions,
and may also be remedied by addressing those violations.
It is important to note that rarely will a strategy satisfy all assump-
tions of the GLM. The flexibility of the general linear model allows it to
be applied to a variety of research designs. When available, information
is presented about the performance of a strategy when the assumptions
of the GLM are not tenable.

CHAPTER SUMMARIES
MANOVA, MANCOVA, MMR, and SEM are discussed in terms
of (1) purpose of the analysis; (2) important assumptions; (3) key
Basic Concepts and Assumptions 19

concepts and analytical steps; (4) sample size and power requirements;
(5) strengths and limitations; (6) an annotated example; (7) reporting
the results of an analysis; and (8) additional examples from the applied
research literature.

Chapter 2—Multivariate Analysis of Variance


MANOVA is discussed. In analysis of variance (ANOVA), the mean dif-
ferences between three or more groups on a single DV are evaluated.
Social workers and other applied researchers may also want to evaluate
the mean differences between groups on two or more DVs. That is, in
MANOVA, there is at least one IV (i.e., group) with two or more levels
(i.e., subgroups), and at least two DVs. Whether groups differ signifi-
cantly on one or more optimally weighted linear combinations of two or
more DVs may be evaluated. Some researchers argue that MANOVA is
preferable to performing a series of ANOVAs (one for each DV) because
(1) multiple ANOVAs can capitalize on chance and (2) ANOVA ignores
intercorrelations among DVs. In contrast, MANOVA controls for inter-
correlations among DVs.

Chapter 3—Multivariate Analysis of Covariance


MANCOVA is similar to MANOVA, but allows a researcher to control
for the effects of supplementary continuous IVs, termed covariates. In an
experimental design, covariates are usually the variables not controlled
by the experimenter, but still affect the DVs.

Chapter 4—Multivariate Multiple Regression


MMR, which models the relationship between several DVs and a set of
IVs, is presented. The DVs may be the multiple dimensions of a single
concept. For example, one might conceptualize consumer satisfaction as
consisting of three dimensions: (1) congruence between provider and
consumer service delivery expectations; (2) provider responsiveness,
including courtesy and timeliness; and (3) effectiveness of service inter-
ventions. MMR estimates the same coefficients and standard errors as
one would obtain using separate OLS regressions (one for each crite-
rion variable). In addition, MMR is a joint estimator and also estimates
20 Analysis of Multiple Dependent Variables

between-equation covariances. This means that it is possible to test coef-


ficients (IVs) across equations (DVs).
That is, although MMR’s multivariate tests (multivariate F-tests) take
into account the correlations among the criterion variables, the regres-
sion coefficients and univariate F-values identified by MMR will be
exactly the same as those obtained from performing one regression anal-
ysis for each DV. For q DVs, q regression formulas may be estimated. The
utility of MMR lies in its ability to test significance across and between
the q equations. At least three kinds of statistical significance questions
may be asked. First, is there a relationship between the DVs taken as a
set and the IVs taken as a set? Second, is a particular IV or subset of the
IVs related to the set of DVs? Third, is the relationship between a set of
IVs and a particular DV statistically significant, when controlling for the
association among the set of DVs?

Chapter 5—Structural Equation Models


SEM, also termed analysis of covariance structures, is discussed. SEM
refers to a hybrid model that integrates path analysis and factor analysis.
SEM is used when data consist of multiple indicators for each variable
(called latent variables or factors) and specified paths connecting the
latent variables. Thinking of SEM as a combination of factor analysis
(to be discussed briefly in chapter 4) and path analysis (also to be dis-
cussed briefly in chapter 4) ensures consideration of SEM’s two primary
components: the measurement model and the structural model. The
measurement model describes the relationships between observed vari-
ables and the construct or constructs those variables are hypothesized
to measure. The structural model describes interrelationships among
constructs.
When the measurement model and the structural model are consid-
ered together, the model is termed the composite or full structural model.
In SEM, statistical power is the ability to detect and reject a poor model.
In contrast to traditional hypothesis testing, the goal in SEM analysis is
to produce a nonsignificant result (i.e., to fail to reject the null of no dif-
ference between the proposed and the perfect model). The null hypoth-
esis is assessed by forming a discrepancy function between the proposed
model (specified by the researcher) and the perfect or saturated model
(one with no constraints that will always fit any data perfectly).
Basic Concepts and Assumptions 21

Various discrepancy functions can be formed depending on the par-


ticular minimization algorithm being used (e.g., maximum likelihood),
but the goal remains the same: to derive a test statistic that has a known
distribution, and then to compare the obtained value of the test statistic
against tabled values in order to make a decision about the null hypoth-
esis. Because in SEM the researcher is attempting to develop a theoretical
model that accounts for all the covariances among the measured items, a
nonsignificant difference between the proposed model and the saturated
model is argued to be suggestive of support for the proposed model.

Chapter 6—Choosing among Procedures for the Analysis of Multiple


Dependent Variables
Conclusions and recommendations are summarized, and MANOVA,
MANCOVA, MMR, and SEM are compared and contrasted.
2

Multivariate Analysis
of Variance: Overview
and Key Concepts

This chapter presents MANOVA as a generalization of the t-test and


ANOVA. The t-test is a strategy to test hypotheses about differences
between two groups on a single mean. When there are more than two
means, it is possible to use a series of t-tests to evaluate hypotheses about
the difference between each pair of group means (e.g., for three groups,
there are three unique pairs of means). However, conducting multiple
t-tests can lead to inflation of the type I error rate (see chapter 1 for a
discussion of type I and type II errors). Consequently,

1. ANOVA is used to test hypotheses about differences between


three or more groups on a single mean; and
2. MANOVA is a strategy to test hypotheses about differences
between two or more groups on two or more means (i.e., a vector
of means).

As a generalization of the t-test and ANOVA, MANOVA may be


understood as the ratio between two measures of multivariate variance.
Univariate variance is defined as the average squared deviation of each
value of a variable from that variable’s mean.

22
Multivariate Analysis of Variance 23

Multivariate variance is defined as the simultaneous average squared


deviation of each value of a variable on multiple means. Univariate vari-
ance and multivariate variance will be discussed further in the ANOVA
and MANOVA sections below.

THE t-TEST
Frequently, hypotheses concern differences between the means of two
groups in a sample. For example, a hypothesis could state that, in a ran-
dom sample, the average age of males does not equal the average age
of females. The t-test may be used to test the significance of this differ-
ence in average age between males and females. The variances of the
two sample means may be assumed to be equal or unequal. There are
two basic designs for comparing the mean of one group on a particu-
lar variable with the mean of another group on that same variable. Two
samples are independent if the data in one sample are unrelated to the
data in the other sample. Two samples are paired (also termed correlated
or dependent samples) if each data point in one sample is matched to a
unique data point in the second sample. An example of a paired sample
is a pre-test/post-test study design in which all participants are measured
on a DV before and after an intervention.
William Gosset, who published under the pseudonym of Student,
noted that using a sample’s standard deviation to estimate the popula-
tion’s standard deviation is unreliable for small samples (Student, 1908).
This unreliability is because the sample standard deviation tends to
underestimate the population standard deviation. As a result, Gosset
described a distribution that permits the testing of hypotheses from nor-
mally distributed populations when the population mean is not known.
This distribution is the t-distribution or Student’s t.
The t-test is used when a population’s variance is not known and the
sample’s size is small (N < 30). In fact, even when N > 30, the t-distribution
is more accurate than the normal distribution for assessing probabilities,
and, therefore, t is the distribution of choice in studies that rely on sam-
ples to compare two means.
The t-distribution is similar to the normal distribution when the esti-
mate of variance is based on many degrees of freedom (df) (i.e., larger
samples), but has relatively more scores in its tails when there are fewer
24 Analysis of Multiple Dependent Variables

Normal

t(4)

–8 –7 –6 –5 –4 –3 –2 –1 0 1 2 3 4 5 6 7 8
Figure 2.1 t versus Normal Distribution.

degrees of freedom (i.e., smaller samples). The t-distribution has n − 1


degrees of freedom, where n is the sample size. Figure 2.1 shows the
t distribution with four df and the standard normal distribution. Note
that the normal distribution has relatively more scores in the center of
the distribution and the t-distribution has relatively more in the tails.
The t-distribution, therefore, is leptokurtic. Because the t-distribution is
leptokurtic (i.e., flatter), the percentage of the distribution, for example,
within 1.96 standard deviations of the mean, is less than the 95% for the
normal distribution.
Important properties of the t-distribution are as follows: (1) differ-
ent for different sample sizes; (2) generally bell-shaped, but with smaller
sample sizes shows increased variability; (3) as the sample size increases,
approaches a normal distribution; (4) symmetrical about the mean;
(5) the mean of the distribution is equal to 0; and (6) variance is equal to
v / (v − 2), where v is the degrees of freedom and v ≥ 2.
The t-statistic is similar to the z-statistic: both statistics are expressed
as the deviation of a sample mean from a population mean in terms of
a standardized measure of distance (i.e., the standard error of the mean
and the standard deviation, respectively). The t-test evaluates this differ-
ence in terms of the probability of obtaining an observed difference as
large as or larger than the sample difference, if the null hypothesis of no
Multivariate Analysis of Variance 25

difference is true. The α-level (e.g., .05) defines “likely.” That is, if αis set
at p = .05, then unlikely is .05 or less. If the αof an observed difference
is equal to or less than .05, then the observed difference is unlikely if the
null hypothesis is true. As a result, the null hypothesis is treated as false
and rejected. In contrast, if the α of the observed difference is greater
than .05, then the null is retained.
The following procedure is used to test whether a difference between
two means in a sample is likely to be true at some level of probability
(e.g., .05) for the population:

1. Formulate a hypothesis (variously referred to as the research,


alternative, or H1) about the difference;
2. State what would be true if H0 is false (i.e., the null hypothesis
or H0);
3. Measure the difference in a random sample drawn from a
population;
4. Assuming either that the variances around the means of the two
groups being compared are equal or are unequal, evaluate the
probability of obtaining the sample results if the H0 is true; and
5. If these results are unlikely, reject the H0.

ANOVA
When there are more than two means, conducting multiple t-tests can
lead to inflation of the type I error rate. This refers to the fact that the more
comparisons that are conducted at, for example, α = .05, the more likely
that a statistically significant comparison will be identified. For example,
if the comparisons are independent, with five means there are 10 possible
pair-wise comparisons. Doing all possible pair-wise comparisons on five
means would increase the overall chance of a type I error to

α* = 1 − (1 − α)10 = 1 − .599 = .401 (2.1)

For five comparisons at α = .05 there is a 40.1% chance of making a type


I error, instead of 5%. The overall chance of a type I error rate in a par-
ticular series of pair-wise comparisons is referred to as the family-wise
error rate.
26 Analysis of Multiple Dependent Variables

Introduced by Fisher (1918), ANOVA is used to test hypotheses about


differences between two or more means without increasing the type I
error rate. ANOVA may be defined as the ratio of two univariate vari-
ances: (1) sum-of-squares-between (or SSB) is a measure of the variabil-
ity of the values of a variable (e.g., age) around the grand mean (i.e., the
mean of several groups) of the variable; and (2) sum-of-squares-within,
also termed SSW, residual sum-of-squares, or sum-of-squares error,
is a measure of the variability within respective groups around that
group’s mean for the same variable (e.g., age). That is, the total varia-
tion (sum-of-squares-total or SST) of the observations is partitioned (or
divided) into a part attributable to differences between group means and
a part attributable to the differences between observations in the same
group, or SST = SSB + SSW, and SSB/SSW, equals F (also termed F-ratio
and F-test), which is used to assess the variability of the group means.
The shape of the distribution of the values of the F distribution
depends on the sample size. More precisely, it depends on two degrees
of freedom parameters: one for the numerator (SSB) and one for the
denominator (MSW). SSB has k − 1 degrees of freedom, and MSW has
N − k degrees of freedom, where N is the total number of observations for
the groups, and k is the number of groups. To summarize:

dfSSB = k – 1 (2.2)

dfSSW = N – k (2.3)

and

F = SSB/SSW (2.4)

If the means of the populations represented by the groups are equal,


then, within the limits of random variation, the SSB and SSW should
be equal; that is, F = 1. Deviations from 1 are assessed by comparing the
observed F to the distribution of values of F if the null hypothesis that
group means are equal is true.
More specifically, the larger the value of F, the greater the evidence
that group means differ. The observed value of F is compared with a
table of values of the F-distribution. It should be noted that, with two
groups, one-way analysis of variance is equivalent to the two-sample
t-test, and F = t2.
Multivariate Analysis of Variance 27

Necessary assumptions for the F-test include the following: (1) the
response variable is normally distributed in each group; and (2) the
groups have equal variances. Note that the aforementioned assumptions
are parametric assumptions and that ANOVA (as illustrated here) also
requires that the observations are independent. F is an extended fam-
ily of distributions, which varies as a function of a pair of df (one for
SSB and one for SSW). F is positively skewed. Figure 2.2 illustrates an F
distribution with degrees of freedom (df) = df1,df2, or df = 5,10. F ratios,
like the variance estimates from which they are derived, cannot have a
value less than zero. Consequently, in one sense, F is a one-tailed prob-
ability test. However, the F-ratio is sensitive to any pattern of differences
among means. Therefore, F should be more appropriately treated as a
two-tailed test.
When the overall F-test is not significant (i.e., assuming the hull
hypothesis is true, the observed value of F is likely or greater than a
pre-established α-level, such as p = .05, analysis usually terminates. When
differences are statistically significant (i.e., assuming the hull hypothesis
is true, the observed value of F is unlikely or less than a pre-established
α-level, such as p = .05), the researcher will often identify specific differ-
ences among groups with post hoc tests. That is, with ANOVA, if the null
hypothesis is rejected, then at least two groups are different in terms of the
mean for the group on the DV. To determine which groups are different,
post hoc tests are performed using some form of correction. Commonly
used post hoc tests include Bonferonni, Tukey b, Tukey, and Scheffé. Post
hoc tests are discussed further within the context of MANOVA.
In general, post hoc tests will be robust in those situations where the
one-way ANOVA’s F-test is robust, and will be subject to the same poten-
tial problems with unequal variances, particularly when the sample sizes
are unequal.

α
Figure 2.2 The F Distribution
28 Analysis of Multiple Dependent Variables

MANOVA
Developed as a theoretical construct by Wilks (1932), MANOVA designs
are substantially more complicated than ANOVA designs, and, therefore,
there can be ambiguity about the relationship of each IV with each DV.
In MANOVA, there is at least one IV (i.e., group) with two or more lev-
els (i.e., subgroups), and at least two DVs. MANOVA designs evaluate
whether groups differ on at least one optimally weighted linear combina-
tion (i.e., composite means or centroids) of at least two DVs.
MANOVA, therefore, is an ANOVA with several DVs. The testing of
multiple DVs is accomplished by creating new DVs that maximize group
differences. The gain in power obtained from decreased within-group
sum of squares might be offset by the loss in degrees of freedom. One
degree of freedom is lost for each DV. Some researchers argue that
MANOVA is preferable to performing a series of ANOVAs (i.e., one for
each DV) because (1) multiple ANOVAs can capitalize on chance (i.e.,
inflate α or increase type I error); and (2) ANOVA ignores intercorrela-
tions among DVs; in contrast, MANOVA controls for intercorrelations
among DVs. See chapter 1 for additional discussion of intercorrelations
among DVs and inflated α−levels
A two-group MANOVA-termed Hotelling’s T 2 is used when one IV
has two groups and there are several DVs (Hotelling, 1931). For example,
there might be two DVs, such as score on an academic achievement test
and attention span in the classroom, and two levels of types of educa-
tional therapy, emphasis on perceptual training versus emphasis on aca-
demic training. It is inappropriate to use separate t tests for each DV to
evaluate differences between groups because this strategy inflates type I
error. Instead, Hotelling’s T 2 is used to see if groups differ on the two DVs
combined. The researcher asks if there are non-chance differences in the
centroids (average on the combined DVs) for the two groups. Hotelling’s
T 2 is a special case of MANOVA, just as the t-test is a special case of
ANOVA, when the IV has only two groups.
One-way MANOVA evaluates differences among optimally weighted
linear combinations for a set of DVs when there are more than two level
levels (i.e., groups) of one IV (factor). Any number of DVs may be used;
the procedure addresses correlations among them, and controls for type
I error. Once statistically significant differences are found, planned and
post hoc comparisons (discussed below) are available to assess which
DVs are influenced by the IV.
Multivariate Analysis of Variance 29

Factorial MANOVA is an extension of MANOVA to designs with


more than one IV and multiple DVs. For example, a researcher may wish
to explore differences in academic achievement (DV) and attention span
(DV) based on gender (a between-subjects IV) and educational therapy
(another between-subjects IV). In this case, the analysis is a two-way
between-subjects factorial MANOVA that tests the main effects of gen-
der and type of educational therapy and their interaction on the cen-
troids of the DVs.
As stated earlier, ANOVA may be defined as the ratio of two uni-
variate variances. Analogously, MANOVA may be defined as the ratio
of two multivariate variances; multivariate variance is a measure of the
simultaneous dispersion of values around multiple means. More spe-
cifically, MANOVA may be described as the ratio of the determinants of
two variance–covariance matrices. An understanding of this definition
requires an understanding of three key concepts: (1) matrix, (2) determi-
nant of a matrix, and (3) variance–covariance matrix. For readers inter-
ested in a more detailed discussion of these concepts, an introduction to
matrix algebra is posted on this book’s website.

1. An m × n matrix is a rectangular array of real numbers with m


rows and n columns. Rows are horizontal and columns are vertical;
2. One measure of multivariate variance is the determinant of a
⎡A B ⎤
matrix. The 2×2 matrix A = ⎢ ⎥ has a determinant, defined as
⎣C D ⎦
A = ad − bc. A determinant is a real number associated with every square
matrix. Mathematically, therefore, the determinant is the product of the
elements on the main diagonal minus the product of the elements off
the main diagonal, and it: (1) is a real number; (2) can be a negative
number; and (3) only exists for square matrices. An important rule about
determinants is that |A| is always 0 if, and only if, at least two vectors of
a matrix are linearly dependent. A vector is defined as a row (horizontal)
or a column (vertical) of numbers. In general, the determinant increases
with increasing independence of the vectors constituting the matrix.
The determinant can be conceptualized as the area (two-dimensional)
or volume (three- or more dimensional) of a geometrical shape, which
is spanned by the vectors of a matrix (e.g., a parallelogram with sides
formed by two row and two column vectors). A parallelogram is a quad-
rilateral that has two pairs of parallel sides. The area of the parallelogram
30 Analysis of Multiple Dependent Variables

is the absolute value of the determinant of the matrix formed by the vec-
tors representing the parallelogram’s sides. The more similar the vectors
(i.e., they point in the same direction), the smaller the area or volume and
the smaller the determinant. In higher dimensions, the analog of volume
is called hypervolume and the same conclusion can be drawn by the same
argument: the hypervolume of the parallel-sided region determined by
k vectors in k dimensions is the absolute value of the determinant whose
entries are their components in the directions of the these vectors.
For example, in Figure 2.3, parallelogram 1 has a greater volume
(e.g., determinant) than parallelogram 2. The parallelogram tilts, and the
length of its sides remains the same.
For nonsquare matrices, it can be shown that there are always some vec-
tors (either rows or columns) that are not independent of the other vectors.
Therefore, the determinant of such a nonsquare matrix is always zero; and

3. An important square matrix is the variance–covariance matrix.


The variance–covariance matrix has variances in its diagonal and cova-
riance in its off-diagonal elements. Thus, the variance–covariance matrix
V can be defined as

⎡ σ x2 cov xxy ⎤
V=⎢ ⎥
⎣cov yx σ y2 ⎦

Because MANOVA may be expressed as the ratio of two multivariate mea-


sures of variance, MANOVA may be expressed as the ratio of the deter-
minants of two variance–covariance matrices. That is, the sum-of-squares
between-groups, the sum-of-squares-within, and the total sum-of-squares
measures used in an ANOVA are replaced by variance–covariance

Parallelogram 1 Parallelogram 2

Figure 2.3 A Determinant as the Area of a Parallelogram.


Multivariate Analysis of Variance 31

matrices: one matrix for between-groups sum-of-squares (the hypothesis


or effect matrix, H), one matrix for within-groups sum-of-squares (the
error or residual matrix, E), and one for total sum-of-squares (the total
matrix, T). The values in the main diagonal of these matrices are the uni-
variate sums-of-squares for each variable. The other elements are the
sums-of-cross products between any two of the variables.
H0 for a one-way MANOVA is that the population effect of the groups
or treatments is zero with respect to all linear combinations of the response
variables. This is equivalent to no difference between population centroids
(multivariate means). This H0 can be tested by using a multivariate F-test
that is based on the ratio of the multivariate variances of two matrices, such
as their determinants. One commonly used multivariate F-test is Wilk’s
lamda (Λ) (Stevens, 2009; Tabachnick & Fidell, 2007), which is the ratio
of the determinants of the sum-of-squares-within-groups matrix (i.e.,| E |)
and the total sum-of-squares matrix (i.e.,| T |) or

|E|
− (2.6)
|T|

Wilk’s Λis an inverse criterion: the smaller the value of Λ, the more evi-
dence for the relationship of the IVs with the DVs. If there is no associa-
tion between the two sets of variables, then Λwould approach 1.
Note that other multivariate F-tests rely on alternative measures of
multivariate variance, such as the trace and an eigenvalue. The trace of
an n-by-n square matrix A is defined to be the sum of the elements on
the main diagonal (the diagonal from the upper left to the lower right) of
A. An eigenvalue provides quantitative information about the variance
in a portion of a matrix. Specifically, if A is a linear transformation repre-
sented by a matrix A such that AX = λX for some scalar λ, thenλ is called
the eigenvalue of A with corresponding eigenvector X.
Pillai–Bartlett trace is the trace of H/T (i.e., the variance between
groups). Hotelling–Lawley trace is the ratio of the determinants H and
E or |H|/|E|. Roy’s largest root is the largest eigenvalue of H/E (i.e., the
eigenvalue of the linear combination that explains most of the variance
and covariance between groups).
The sampling distributions of these statistics are usually converted
to approximate F-ratio statistics (Tabachnick & Fidell, 2007), and they
will generally produce similar results with more than two groups. The
32 Analysis of Multiple Dependent Variables

Pillai–Bartlett trace is the most conservative of these four F-tests, but is a


viable alternative if there are reasons to suspect that the assumptions of
MANOVA are untenable (Stevens, 2009). In terms of power, none of the
aforementioned F-tests is always the choice with the greatest statistical
power. Wilk’s is the most widely used and consequently more likely to be
familiar to readers (Warner, 2008).

Assumptions of MANOVA
MANOVA assumes that data being analyzed follow the GLM (Stevens,
2009; Tabachnick & Fidell, 2007). These assumptions are discussed in
further detail in chapter 1, and demonstrated in the annotated example
below. In particular, the tenability of following assumptions should be
assessed prior to conducting a MANOVA:

1. Multivariate normality of the DVs (robust to moderate violations


with respect to type 1 error);
2. Absence of Outliers (sensitive to outliers, especially with a small
sample size);
3. Homoscedasticity (robust to moderate violation if groups sizes
are approximately equal; largest/smallest < 1.5); and
4. Low to moderate correlation among the DVs.

The MANOVA Procedure


MANOVA may be conceptualized as follows:

1. The omnibus null hypothesis is evaluated with a multivariate


F-test;
2. Overall model fit is assessed; and
3. Pairs of group means whose difference are statistically significant
are identified.

Evaluating the Omnibus Null Hypothesis


To test the omnibus null hypothesis, multivariate F- tests (discussed ear-
lier in this chapter) are used. These multivariate F-tests include Wilk’s
Λ, Hotelling’s trace, Pillai–Bartlett trace, and Roy’s largest root. The
Multivariate Analysis of Variance 33

purpose of this omnibus hypothesis test, then, is to help to prevent inflat-


ing the family-wise α-level. As discussed in chapter 1, the family-wise
α-level increases as the number of separate hypothesis tests increases.
Consequently, if separate tests are performed for each DV, the probability
of obtaining a false significant value would increase in direct propor-
tion to the number of DVs being tested; that is, the power of the test
decreases. If the multivariate F is significant, overall model fit may be
assessed, and pairs of means whose difference are statistically significant
are identified.

Assessing Overall Model Fit


One measure of overall model fit is an effect size. In the context of a
one-way MANOVA, an effect size is a retrospective measure of the amount
of variability in the set of DVs that are explained by the IV(s). One multi-
variate measure of effect size is Wilks’ λ(discussed above), which reflects
the ratio of within-group variance to total variance. That is, as discussed
above, Wilk’s Λ is an inverse criterion: the smaller the value of Λ, the
more evidence for a relationship between the IVs and the DVs. If there is
no association between the IVs and the DVs, then Λwould approach 1.
Another measure of effect size for a MANOVA model is partial eta-
squared (Ș2). As calculated by SPSS partial eta-squared is defined as
1 − Wilk’s Λ. In SPSS, partial eta square measures the contribution of
each factor without controlling for the effects of other factors, and, con-
sequently, it is technically not a true partial measure (Pierce, Block, &
Agunis, 2004). In SPSS, the values of partial eta-squared of all DVs can
sum to greater than 100%.

Exploring Multivariate Group Differences


The H0 for a one-way MANOVA is that there are no multivariate group
differences (i.e., there are no group differences on linear combinations
of the DVs). Accordingly, when conducting a MANOVA, researchers fre-
quently will pose specific questions regarding the nature of multivariate
group differences. One issue that is continuously debated by research-
ers, then, concerns how best to follow-up a statistically significant overall
F-test when conducting a MANOVA (Howell, 2009). As discussed earlier,
as the number of pair-wise statistical comparisons (e.g., t-tests) increases,
34 Analysis of Multiple Dependent Variables

the overall type I error rate increases. This problem also may occur when
a MANOVA is followed-up with multiple comparisons between groups
on linear combinations of the DVs.
There are two basic approaches to assessing multivariate group
differences: (1) perform the omnibus test first, followed by a study of
some comparisons between pairs of means; or (2) proceed directly to
group comparisons and answer specific research questions. That is, some
authors (c.f. Huberty & Smith, 1982; Wilcox, 1987) question the need
to test the overall null hypothesis, and argue that this second approach
may be appropriate when there is an a priori ordering of the DVs, which
implies that a specific set of hypotheses be tested. Howell (2009) explains
that the hypotheses tested by multivariate F-tests and a multiple compar-
ison tests are different. Multivariate F-tests distribute differences among
groups across the number of degrees of freedom for groups. This has
the effect of diluting the overall F-value if, for example, several group
means are equal to each other but different one other mean. In general,
follow-up tests, sometimes termed comparisons or contrasts, are a priori
or planned, or post hoc or unplanned. An a priori comparison is one
that the researcher has decided to test prior to an examination of the
data. These comparisons are theory-driven and part of a strategy of con-
firmatory data analysis. A post hoc comparison is one that a researcher
decided to test after observing all or part of the data. These comparisons
are data-driven and are part of an exploratory data analysis strategy.
The following comparison strategies are discussed here: (1) multiple
ANOVAs (Leary & Altmaier, 1980); (2) two-group multivariate compari-
sons (Stevens, 2009); (3) step-down analysis (SDA) (Roy & Bargmann,
1958); and (4) simultaneous confidence intervals (SCI) (Smithson, 2003).
Multiple ANOVAs and SCIs are univariate perspectives. Two group mul-
tivariate comparisons and SDA are multivariate perspectives.
It should be noted that descriptive discriminant analysis (DDA) also
has been suggested (Huberty, 1994). In DDA, IVs are linearly combined to
create a composite IV that maximally differentiates the groups. Although
mathematically identical to a one-way MANOVA, DDA emphasizes clas-
sification and prediction. The first step in any DDA is to derive discrim-
inant functions that are linear combinations of the original variables.
The first discriminant function is the linear combination of variables
that maximizes the ratio of between-groups to within-groups vari-
ance. Subsequent discriminant functions are uncorrelated with previous
Multivariate Analysis of Variance 35

discriminant functions best separate groups using the remaining varia-


tion. The number of discriminant functions that can be extracted equals
the number of groups minus one.
MANOVA and DDA may be viewed as providing complemen-
tary information. In MANOVA, the focus tends to be on which groups
differed significantly. In DDA the focus tends to be on which discrimi-
nating variables were useful in differentiating groups. That is, DDA tends
to focus on a linear composite of the multiple DVs. It may be argued
that if the focus is primarily on the linear composite, DDA may be suf-
ficient, and the researcher should consider using it instead of MANOVA.
If, however, the researcher is primarily interested in the individual DVs in
the analysis, while controlling for the correlations among the DVs, then
MANOVA, may be sufficient. In accordance with MANOVA’s emphasis
on tests of hypotheses about group differences, subsequent discussion
will focus on follow-up group comparison strategies.
Multiple ANOVAs with a Priori or Post Hoc Comparisons. A sig-
nificant multivariate F-test indicates that there are group differences with
respect to their means on the DVs. One popular univariate approach to
following up a statistically significant multivariate F-test is to compare
planned univariate F-tests and unadjusted group means for each of
the DVs.
Another popular variation on the aforementioned univariate strat-
egy is to compare unplanned univariate F-tests and adjusted group
means for each of the DVs. Post hoc approaches discussed here include
(1) Bonferroni, (2) Tukey honestly significant difference (HSD), and (3)
Scheffé. These three tests are a popular subset of possible post hoc cor-
rections, and provide a continuum of choices for conservative to mod-
erate tests of simple and complex comparisons (Klockars, Hancock,
& McAweeney, 1995; Kromrey & La Rocca, 1995; Seaman, Levin, &
Serlin,1991;Toothacker, 1993).
The Bonferroni procedure. The Bonferroni, a commonly used post
hoc test, is relatively simple to compute. The per-comparison α for each
individual significance test between means is set as the study-wise α
level divided by the number of comparisons to be made (Hsu, 1996).
For example, in a one-way MANOVA with five groups (k = 4) and one
DV, there are k (k −1)/2 = 6 possible pair-wise comparisons of group
means. If the researcher wants to limit the overall study-wise type I error
rate for the set of six comparisons to .05, then the per-comparison αfor
36 Analysis of Multiple Dependent Variables

each individual significance test between means is the study-wiseα level


divided by the number of comparisons to be made, or .05/10 = .005. The
Bonferroni procedure is conservative, and some researchers prefer less
conservative methods of limiting the risk of type I error.
The Tukey’s HSD (Honestly Significant Difference) test. The Tukey
HSD uses a sampling distribution called the Studentized Range or
Student’s q, which is similar to a t-distribution (Tukey, 1953). Like t, the
q-distribution depends on the number of participants within groups, but
the shape of this distribution also depends on k, the number of groups.
As the number of groups increases, the number of pair-wise compari-
sons also increases. To protect against inflated risk of type I error, larger
differences between group means are required for rejection of the null
hypothesis as k increases.
The distribution of the Studentized range statistic is broader and flat-
ter than the t distribution and has thicker tails. Consequently, the critical
values of q are larger than the corresponding critical values of t. The Tukey
HSD test could be applied by computing a q value for each comparison
of means and then checking to see if the obtained q for each compari-
son exceeded the critical value of q from a table of the Studentized range
statistic.
More specifically, q corresponds to the sampling distribution of the
largest difference between two means coming from a set of A means
(when A = 2, the q distribution corresponds to the usual Student’s t).
In practice, one computes a criterion denoted qobserved, which evalu-
ates the difference between the means of two groups. This criterion is
computed as:

Mi Mj
q observed = (2.7)
⎛ 1⎞
MSerror
⎝s⎠

where Mi and Mj are the group means being compared MSerror is mean
square within, which is an estimate of the population variance based on
the average of all variances within the several samples, and S is the number
of observations per group (the groups are assumed to be of equal size).
Once the qobserved is computed, it is then compared with a qcritical value
from a table of critical values (see this book’s companion webpage). The
value of qcritical depends upon the a-level, the degrees of freedom v = N ¡K,
Multivariate Analysis of Variance 37

where N is the total number of participants and K is the number of


groups, and a parameter R, which is the number of means being tested.
For example, in a group of K = 5 means ordered from smallest to largest,
M1 < M2 < M3 < M4 < M5; R = 5 when comparing M5 to M1; R = 3 when
comparing M3 to M1. The Tukey HSD is considered to be a moderately
conservative post hoc test, and is preferred for a large number of simple
(i.e., pairwise) comparisons, or for making complex comparisons (Hsu,
1996). A simple contrast compares pairs of means. A complex comparison
compares means to linear combinations of means or linear combinations
to mean to each other. For example, a researcher wonders if participants
in a control group will experience fewer suicidal ideations compared to
the average number of suicidal ideations of participants who received
one of three alternative intervention (exp1, exp2, and exp3) strategies. The
H0 for this comparison with contrast coefficients of (1, −1/3, −1/3, −1/3)
is as follows:

μ exp + μ expp 2 + μ exp 3


μ wait list = (2.8)
3

The Scheffé Test. The Scheffé test uses F tables versus Studentized
range tables (Scheffé, 1953). If the overall null hypothesis is rejected,
F -values are computed simultaneously for all possible comparison pairs.
These F-values are larger than the value of the overall F test. The formula
simply modifies the F-critical value by taking into account the number of
groups being compared: (a –1) Fcrit. The new critical value represents the
critical value for the maximum possible family-wise error rate. This test
is the most conservative of all post hoc tests discussed here. Compared
to Tukey’s HSD, Scheffé has less power when making pair-wise compari-
sons (Hsu, 1996). That is, although the Scheffé test has the advantage
of maintaining the study-wise significance level, it does so at the cost of
the increased probability of type II errors. Some authors (cf. Brown &
Melamed, 1990) argue that he Scheffé test is used most appropriately if
there is a need for unplanned comparisons. Toothacker (1993) recom-
mends the Scheffé test for complex comparisons, or when the number of
comparisons is large.
One limitation of using multiple ANOVAs to follow-up a significant
multivariate F-test result is that univariate and multivariate tests use dif-
ferent information. If multiple ANOVAs with unadjusted or adjusted
38 Analysis of Multiple Dependent Variables

mean comparisons are performed, a researcher may assume that if the


multivariate null hypothesis is rejected, then at least one of these univar-
iate F-tests will be significant. However, the univariate tests measure the
relationship between one DV variable and IV, and the multivariate test
uses information about the correlations among the DVs. Consequently,
there is no necessary relationship between multivariate significance and
univariate significance. That is, rejection of multivariate test does not
guarantee that there exists at least one significant univariate F-value
(Stevens, 2009). When the correlation among DVs is strong, error on
any one DV is accounted for by at least one other DV. As a result, the
multivariate error term smaller, and the multivariate test has greater
power than the univariate test. This greater power of the multivari-
ate test means that it is more likely to identify a difference among the
group means on the DVs using a multivariate versus a univariate test if
a difference exists.
Two-Group Multivariate Comparisons. As outlined by Stevens
(1972), the two-group method involves using the Hotelling’s T-test
statistic (Hotelling, 1931) to simultaneously compare all possible pairs
of groups on the set of DVs. As an example, if the MANOVA involved
one IV with three levels measured on three DVs, the two-group analysis
would involve the multivariate comparison of (1) groups one and two,
(2) groups one and three, and (3) groups two and three on the three
DVs simultaneously. In general, a significant MANOVA will produce at
least one significant T2. Then, researchers often use univariate t-tests, to
explore these statistically significant differences between groups and the
DVs. Stevens (1972) compared this approach with DDA, and found that
the results, in terms of identifying group differences, were similar, though
the two group approach did not allow for direct identification of group
differences on individual DVs.
Stepdown Analysis. Introduced by Roy and Bargmann (1958),
stepdown analysis (SDA) is a multistep procedure. Prior to beginning
a SDA, the researcher places the DVs in descending order of theoreti-
cal importance. It is important to note that this ordering should be
based on theoretic issues and not on statistics. Logical orderings of
DVs arise when study participants have been tested repeatedly over
time, when measures involve progressively more complex behaviors,
or whenever there exists a systematic progression from one DV to
another.
Multivariate Analysis of Variance 39

In step one of a SDA, an analysis of variance (ANOVA) is conducted


in which group means on the most important response variable are
compared. In step two, the test in step one serves as a covariate and the
means of the groups on the second most important DV are compared
using an ANCOVA (i.e., in step one there is only one DV in the model).
This stepping procedure continues for each DV in turn with DVs at a
higher level of importance serving as covariates for those of lesser impor-
tance. Testing stops if a significant F-value is encountered, and the H0 is
rejected. Conversely, if no F-value is significant, H0 is accepted. To control
the type I error rate, Tabachnick and Fidell, (2007) recommend the use of
a Bonferroni procedure (discussed earlier).
Mudholkar and Subbaiah (1980) cited several potential advantages
of using the SDA procedure as a follow up to a significant MANOVA
result: (1) simplicity; (2) detailed results for specific variables and groups;
(3) useful with small samples; and (4) results for large samples that are
equivalent to the omnibus likelihood ratio test. Step-down analysis also
has limitations. A potential weakness is the need to order the DVs in terms
of theoretical importance. Empirical evidence suggests that the quantita-
tive conclusions reached by researchers can differ substantially depend-
ing upon the ordering of DVs (Kowlowsky & Caspy, 1991; Mudholkar &
Subbaiah, 1988; Stevens, 1972). In fact, Stevens (2009) argues that a SDA
procedure may not be appropriate when a clear a priori ordering of the
DVs is not possible. Stevens (1996) also explains that each step in an
ordering of DVs is not independent of other steps, and the type I error
rate across all of steps in an SDA may not be known. Consequently, the
order of DVs may affect the results of an SDA.
Mudholkar and Subbaiah (1980) suggest that when an ordering of
DVs is not feasible, SDA may still be used as a post hoc procedure if the
observed variables first are subjected to a data-reduction analysis, such
as creating linear combinations of DVs with factor analysis. These linear
combinations then could be used with for a SDA, with linear combina-
tions of DVs accounting for a greater percentage of the total variance
of all DVs in a model placed higher in the sequence. Kowlowsky and
Caspy (1991) observed that by applying the SDA to multiple sequences
of the response variables a researcher may engage in meaningful data
exploration and testing of multiple hypotheses.
Research examining the performance of the SDA procedure, includes
a Monte Carlo simulation by Subbaiah and Mudholkar (1978). Using
40 Analysis of Multiple Dependent Variables

simulated data sets from the multivariate normal distribution Subbaiah


and Mudholkar (1978) found that the SDA procedure was able to main-
tain nominal (or expected) type I error rates (0.01 or 0.05) while attain-
ing power values above 0.8 for most studied conditions. Stevens (1972)
used both SDA and discriminant function analysis to compare four
groups on eight response variables. The two methods yielded similar
results in terms of identifying variables on which the groups differed,
while the application of univariate ANOVAs produced a markedly differ-
ent outcome. According to Finch (2007), when both the assumptions of
normality and equality of covariance matrices are not met, the observed
type I error rate for the first DV tested in a sequence was inflated, while in
all other cases it was near the nominal 0.05 level. This result should serve
to encourage researchers using the SDA to assess the viability of both the
normality and equality of covariance matrix assumptions prior to mak-
ing use of the technique.
The essential point of the step-down procedure is that the DVs should
be arranged in descending order of importance. For example, with two
DVs the compound (or multivariate) hypothesis is then decomposed into
two component (or univariate) hypotheses. The first component hypoth-
esis concerns the marginal univariate distribution of the most important
DV; the second component hypothesis concerns the conditional univari-
ate distribution of the next most important DV given the most important
DV. Each of the aforementioned component hypotheses is evaluated by
using a univariate F-test. The compound hypothesis is accepted if each of
the univariate hypotheses is accepted. The component univariate tests are
independent, if the compound hypothesis is true (Roy, 1958).
Simultaneous Confidence Intervals (CIs). A CI is a range of values
around which a population parameter (e.g., true mean) is likely to lie in
the long run (Neyman, 1952). For example, assuming a normal distribu-
tion, if samples of the same size are drawn repeatedly from a popula-
tion and a 95% CI is calculated around each sample’s mean (i.e., plus or
minus two standard errors from the mean), then 95% of these intervals
should contain the population mean. A second interpretation of a CI is
as a significance test (Schmidt & Hunter, 1997). From this perspective,
if the value of 0 falls within an interval, the estimate is not statistically
significant at the level of the CI.
When multiple CIs are calculated, each with confidence level 1 − α, the
simultaneous or overall confidence level will be 1 − mαwhere m equals
Multivariate Analysis of Variance 41

the number of CIs being calculated. Therefore, one approach to control-


ling the simultaneous confidence level is to set the level of each confidence
interval equal to 1 − α/m. For example, if five CIs are calculated, and the
researcher wants to maintain the overall CI at the 95% level, then each of
the five CIs would be calculated at the 99% CI. This method of forming
simultaneous confidence intervals (SCIs) is called the Bonferroni method.
Bonferroni-type simultaneous confidence intervals based on the Student’s
t distribution for the contrast differences across all DVs are available in
SPSS, SAS, and Stata (http://www.ats.ucla.edu/stat/stata/examples/alsm/
alsmstata17.htm) Alternative approaches to calculating SCIs include
Scheffé and Tukey, which are based on the F-distribution and Studentized
Range or Student’s q-distribution, respectively (discussed further below).
According to Maxwell and Delaney (2003), no one comparison method
is uniformly best. If all pair-wise comparisons are of interest, Tukey’s is
the most powerful strategy. If only a subset of pair-wise comparisons
is required, Bonferroni is preferred. Using the Scheffé method almost
always sacrifices power, unless at least one of the comparisons to be tested
is complex. If multiple simultaneous CIs are provided by a software pack-
age (e.g., Stata and SPSS), the researcher should select the CI with the
narrowest width. Note that although there is no general consensus, a
recommended strategy to follow up a significant omnibus test is used to
guide the analysis in the annotated example describe below.

Model Validation
In the last stage of MANOVA, the model should be validated. If sam-
ple size permits, one approach to validation is sample splitting, which
involves creating two subsamples of the data and performing a MANOVA
on each subsample. Then, the results can be compared. Differences in
results between subsamples suggest that these results may not generalize
to the population.

Sample Size Requirements


Statistical power is the ability of a statistical test to detect an effect if the
effect exists (Cohen, 1988). That is, power is the probability of reject-
ing H0 when a particular alternative hypothesis (Ha) is true. Power
can be defined only in the context of a specific set of parameters, and,
42 Analysis of Multiple Dependent Variables

accordingly, reference to an alternative hypothesis, expressed as an effect


size (discussed subsequently), is necessary. The power of a test, therefore,
is defined as 1 − beta (E), where E (or type II error) is the probability of
falsely accepting H0 when Ha is true. More specifically, power is a func-
tion of α and E, where α (or type I error) is the probability of falsely
rejecting H0 when Ha is false. As α increases, E decreases with a corre-
sponding increase in power. Also, power is a function of effect size (i.e.,
the true alternative hypothesis, such as an expected difference between
groups, how small the effect can be and still be of substantive interest, or
the expected size of the effect). As effect size increases, power increases.
Finally, power is a function of sample size and variance. In general, the
variance of the distribution of an effect size decreases as N increases; and,
as this variance decreases, power increases.
The logic behind power analysis is compelling: a study should be
conducted only if it relies on a sample size that is large enough to provide
an adequate and a prespecified probability of finding an effect if an effect
exists. If a sample is too small, the study could miss important differ-
ences; too large, the study could make unnecessary demands on respon-
dent privacy and time or waste valuable resources. The benefits of power
analysis are at least twofold: sample size selection is made on a rational
basis rather than using rules-of-thumb, and the researcher must specify
the size of the effect that is substantively interesting.
A widely used measure of effect size in MANOVA is Cohen’s (1988) f 2.
For multiple correlation and regression, f 2 can be expressed as a function
of R 2 (the multiple correlation coefficient):

R2
f 2= . (2.10)
1− R 2

However, in the case of MANOVA, the relationship between f 2 and R 2 is


complex. Insight into this relationship between f 2 and R 2 is provided in
Tables 10.2.1, 10.2.2, and 10.2.3 in Cohen (1988). A review of these tables
can assist researchers who are planning studies and who are familiar with
R 2 make decisions with regard to the minimum or expected value of f 2.
The strategy for determining sample size recommended here and
demonstrated in the annotated example below is to use GPower, which
is a free power analysis program available at http://www.psycho.unidues
seldorf.de/aap/projects/gpower/.
Multivariate Analysis of Variance 43

STRENGTHS AND LIMITATIONS OF MANOVA


MANOVA has the following analytical strengths:

1. By measuring several DVs the researcher improves the chance of


discovering what changes as a result of different treatments and
their interactions;
2. Protection against inflated type I error as a result of multiple tests
of (likely) correlated DVs; and
3. Group differences, not evident in separate ANOVAs, may become
apparent. Consequently, MANOVA, which considers DVs in
combination, may be more powerful than separate ANOVAs.

MANOVA has the following analytical limitations:

1. The assumption of equality of variance–covariance matrices


is more difficult to satisfy than its bivariate analog (equality of
variance);
2. There may be ambiguity in interpretation of the effects of IVs on
any single DV;
3. The situations in which MANOVA is more powerful than
ANOVA are quite limited; often MANOVA is considerably less
powerful than ANOVA, particularly in finding significant group
differences for a particular DV. Even moderately correlated DVs
diminish the power of MANOVA; and
4. When there is more than one DV, there may be redundancy in
the measurement of outcomes.

ANNOTATED EXAMPLE
A study is conducted to test a model that compares satisfaction with agency
services and satisfaction with parenting by client race. Satisfaction with
agency services and satisfaction with parenting are operationalized as scale
scores. Race is operationalized as clients’ self-reports of their racial group
(coded 1 = Caucasian, 2 = Asian American, 3 = African American). Both
the satisfaction with agency services and satisfaction with parenting scales
are considered reliable. Both scales have associated Cronbach’s alphas
greater than .80. Cronbach’s alpha is the most common form of internal
44 Analysis of Multiple Dependent Variables

consistency reliability coefficient. By convention, a lenient cut-off of .60


is common in exploratory research; alpha should be at least .70 or higher
to retain an item in an “adequate” scale; and many researchers require a
cut-off of .80 for a “good scale.” Cronbach’s alpha is discussed further in
the section on standard measures and scales, along with other coefficients
such as Cohen’s kappa (cf. Rubin & Babbie, 2010). Cronbach’s alpha can
be interpreted as the percentage of variance the observed scale would
explain in the hypothetical true scale composed of all possible items in
the universe. Alternatively, it can be interpreted as the correlation of the
observed scale with all possible other scales measuring the same thing
and using the same number of items (Nunnally & Bernstein, 1994).
The data analysis strategy used here in this annotated example is
consistent with the recommendations of several prominent authors
(cf. Tabachnick & Fidell, 2007; Stevens, 2009) to use Roy–Bargmann
stepdown analysis and simultaneous confidence intervals to follow-up
comparisons to elucidate a significant MANOVA result. Researchers are
strongly encouraged to plan data analysis in detail and include a small
number of planned follow-up comparisons after a statistically significant
overall F-test. According to Stevens (2009), although “large” is difficult
to define, the following considerations mitigate against the use of a large
number of DVs: (1) small or negligible differences on many DVs may
obscure real differences of a few DVs; (2) the power of the multivariate
tests generally declines and the number of DVs is increased; and (3) reli-
ability of variables may be a problem in behavioral science research. It is
probably wise to combine highly similar response measures, particularly
when basic measurements tend individually to be unreliable.
There are at least three advantages to the proposed strategy. First, the
strategy encourages the researcher to think about effects of interest, and
to relate these to some underlying theory or to beneficial outcomes of the
research. Second, the strategy encourages the researcher to think multi-
variately, and not to ignore response variable intercorrelations. Third, the
researcher gets a clearer perspective on the relative importance on terms
of their contributions to group differences.
Accordingly, the researcher decides to perform a one-way MANOVA
analysis to investigate the relationship between a client’s race (IV), and
her/his satisfaction with the adoption agency (DV1), and satisfaction
with parenting (DV2). Note that throughout this annotated example
Multivariate Analysis of Variance 45

SPSS, Stata, and SAS commands are numbered in sequence and high-
lighted in Courier. The data set used in this annotated example, enti-
tled Annotated_Example1_FULL_N=66.sav, may be downloaded from
this book’s companion website. These data are in SPSS format, but may
be imported into Stata and SAS.

Determining the Minimally Sufficient Sample Size


GPower, which is a free power analysis program available at http://www.
psycho.uni-duesseldorf.de/aap/projects/gpower/, is used to determine
the minimally sufficient sample size. Forα= .05, E= .20, three groups,
two DVs, and f 2 = 0.15 (moderate), a total sample size of approximately
66 is adequate.

Prescreening for Multivariate Normality


The tenability of the multivariate normality assumption was evaluated
by using an SPSS macro developed by DeCarlo (1997). Open the file
Annotated_Example1_FULL_N=66.sav in SPSS and type the following
commands (numbered in sequence and highlighted in Courier into
an SPSS’s syntax window):

1. include ‘C:\Users\pat\Desktop\MANOVA_Ex_1\
normtest.sps’
2. normtest vars = satisfaction_
parenting,satisfaction_adopt_agency
3. execute.

The first line of the aforementioned syntax includes (calls) the macro
entitled normtest.sps, which is located in the directory c:\spsswin\,
and the second line invokes the macro for the variables satisfaction_
parenting,satisfaction_adopt_agency. This macro is available from http://
www.columbia.edu/~ld208/ and this book’s companion website.

1. Select run → all.

The results of DeCarlo’s (1997) are summarized in Figure 2.4. For brev-
ity, the focus here is on Small’s (1980) and Srivistava’s (1984) tests of
46 Analysis of Multiple Dependent Variables

Tests of multivariate skew:

Small’s test (chisq)


Q1 df p-value
1.7716 2.0000 .4124

Srivastava’s test
chi (b1p) df p-value
.4892 2.0000 .7830

Tests of multivariate kurtosis:

A variant of small’s test (chisq)


VQ2 df p-value
5.3119 2.0000 .0702

Srivastava’s test
b2p N(b2p) p-value
2.7672 –.5459 .5851

Mardia’s test
b2p N(b2p) p-value
7.9961 –.0040 .9968

Omnibus test of multivariate normality:

(based on Small’s test, chisq)


VQ3 df p-value
7.0834 4.0000 .1315

Figure 2.4 Tests of Multivariate Skew.

multivariate kurtosis and skew, Mardia’s (1970) multivariate kurtosis,


and an omnibus test of multivariate normality based on Small’s statis-
tics (see Looney, 1995). However, all other tests calculated by the macro
were consistent with these results. That is, all tests of multivariate skew,
multivariate kurtosis, and an omnibus test of multivariate normality
were not statistically significant at p = .05. Each of these tests evaluates
the H0 that the current distribution (these data) equals the multivari-
ate normal distribution. It is hoped that this H0 will fail to be rejected
(accepted). These results suggest that the assumption of multivariate
normality is tenable. See chapter 1 for additional information about
these tests.
Multivariate Analysis of Variance 47

Stata Commands to Prescreen for Multivariate Normality


Mvtest performs multivariate tests of univariate, bivariate, and multi-
variate normality. All multiple-sample tests provided by mvtest assume
independent samples.

1. findit mvtest
2. install
3. mvtest normality satisfaction_parenting
satisfaction_adopt_agency, stats(all)

SAS Commands to Prescreen for Multivariate Normality


The %MULTNORM macro provides tests and plots of multivariate nor-
mality. A test of univariate normality is also given for each of the vari-
ables. The macro is available from http://support.sas.com/kb/24/983.
html. To use the macro, type the following commands in the SAS editor
window:

1. %inc “<location of your file containing the


MULTNORM macro>”;
2. %multnorm(data=cork, var=n e s w,
plot=mult, hires=no)

For this second command statement, DATA = SAS data set to be analyzed.
If the DATA = option is not supplied, the most recently created SAS data
set is used. VAR = the list of variables to used. Individual variable names,
separated by blanks, must be specified. PLOT = MULT requests a high-
or low-resolution chi-square quantile–quantile (Q-Q) plot of the squared
Mahalanobis distances of the observations from the mean vector. PLOT
= UNI requests high-resolution univariate histograms of each variable
with overlaid normal curves and additional univariate tests of normality.
Note that the univariate plots cannot be produced if HIRES = NO. PLOT
= BOTH (the default) requests both of the above. PLOT = NONE sup-
presses all plots and the univariate normality tests. HIRES = YES (the
default) requests that high-resolution graphics be used when creating
plots. You must set the graphics device (GOPTIONS DEVICE =) and
48 Analysis of Multiple Dependent Variables

any other graphics-related options before invoking the macro. HIRES =


NO requests that the multivariate plot be drawn with low-resolution.
The univariate plots are not available with HIRES = NO.
For these data, the following commands were used:

1. %inc “<location of your file containing the


MULTNORM . . . macro>”;
2. %multnorm(data=file name, var =
satisfaction_parenting satisfaction_adopt_
agency, plot = mult, hires=no)

Prescreening for Absence of Outliers


The tenability of the absence of outliers assumption was evaluated by
using SPSS to calculate Cook’s distance, D, which provides an overall
measure of the impact of an observation on the estimated MANOVA
model. To calculate Cook’s D in SPSS, proceed as follows:

1. Click Analyze → General Linear Model → Multivariate


2. Move the DVs to the Dependent Variables box
3. Move the IVs to the Fixed Factor(s) box
4. Click the Model button, and be sure that the radio button Full
Factorial is selected
5. Click the Save button and select Cook’s distance
6. Click OK

As a result of the aforementioned steps, two variables are added


to the data file (one for each DV). These variables are named COO_1
and COO_2. Values for these two variables are the values of D for each
DV. Observations with larger values of D than the rest of the data are
those which have unusual leverage. Some authors suggest as a cut-off
for detecting influential cases values of D greater than 4/(n − k − 1),
where n is the number of cases and k is the number of independents.
Others suggest D > 1 as the criterion to constitute a strong indication
of an outlier problem, with D > 4/n the criterion to indicate a possible
problem. Figure 2.5 summarizes Cook’s D cutoffs based on the three
Multivariate Analysis of Variance 49

Criterion Calculated Value for These Data

D greater than 4/(n – k – 1) 4/(66-2-1) = .06

D>1 1

D > 4/n 4/66 = .06

Figure 2.5 Output of DeCarlo’s Macro.

aforementioned criteria. For these data, no value of D exceeded .06. These


results suggest that the assumption of absence of outliers is tenable.

Stata Commands to Prescreen for Absence of Outliers


1. regress Race satisfaction_parenting
satisfaction_adopt_agency
2. _predict cooksd

Cook’s D values will be added to the data file as COO_1 and COO_2.

SAS Commands to Prescreen for Absence of Outliers


1. proc reg data=“file path”;
2. model file Race=satisfaction_parenting
satisfaction_adopt_agency;
3. output out= resdata cookd=cooksd
4. run;

Cook’s D values will be saved in a new file named “resdata.”

Prescreening for Homoscedasticity


MANOVA assumes that for each group (each cell in the factor design
matrix) the covariance matrix is similar. Box’s M tests the null hypothesis
that the observed covariance matrices of the DVs are equal across groups.
To calculate Box’s M in SPSS, proceed as follows:

1. Click Analyze → General Linear Model → Multivariate


50 Analysis of Multiple Dependent Variables

2. Move the DVs to the Dependent Variables box


3. Move the IVs to the Fixed Factor(s) box
4. Click the Model button, and be sure that the radio button Full
Factorial is selected
5. Click the Options button and select Estimates of effect size and
homogeneity tests
6. Click OK

Researchers do not want M to be significant in order to conclude


there is insufficient evidence that the covariance matrices differ.
Note, however, that the F-test is quite robust even when there are
departures from this assumption. Here M is not significant, and
consequently, the assumption of homoscedasticity is tenable (see
Figure 2.6).

Stata Commands to Prescreen for Homoscedasticity


1. mvtest covariances satisfaction_parenting
satisfaction_adopt_agency, by(Race)

SAS Commands to Prescreen for Homoscedasticity


1. proc discrim;
2. class Race;
3. var satisfaction_parenting
satisfaction_adopt_agency;
4. pool=test wcov;
5. run;

Box’s M 12.046
F 1.906
df1 6
df2 31697.215
Sig. .076

Figure 2.6 Box’s Test of Equality of Covariance Matrices.


Multivariate Analysis of Variance 51

Prescreening for Homogeneity of Regressions


Since a Roy–Bargmann stepdown analysis is planned, data are pre-
screened for homogeneity of regressions. This assumption, also termed
homogeneity of the regression hyperplanes or parallelism, requires that
the regression slope between the covariate and DV is the same (homo-
geneous) for all groups. According to Stevens (2002), a violation of this
assumption means that there is a statistically significant covariate-by-IV
interaction. Conversely, if the interaction is not statistically significant
this assumption is met.
To test this assumption for a MANOVA with one DV, a model that
contains one covariate (i.e., the lower importance DV)-by-IV interaction
term is tested for each pair of groups. It is hoped that this interaction
term does not achieve statistical significance (e.g., p > 0.05). If there is
more than one DV, then, one covariate-by-IV interaction term is sequen-
tially included in the model for each DV, and each sequential model is
tested for each pair of groups. It is hoped that all covariate-by-IV interac-
tion terms are not statistically significant.
To test the homogeneity of regressions for a model that contains a cli-
ent’s race (IV), her/his satisfaction with the adoption agency (DV1), and
satisfaction with parenting (DV2) proceed as follows:

1. MANOVA satisfaction_adopt_agency by race


(1,3) with satisfaction_parenting
2. /analysis = satisfaction_adopt_agency
3. /design = satisfaction_parenting,race,
satisfaction_parenting by race.

1. MANOVA satisfaction_adopt_agency by race


(1,2) with satisfaction_parenting
2. /analysis = satisfaction_adopt_agency
3. /design = satisfaction_parenting,race,
satisfaction_parenting by race.

1. MANOVA satisfaction_adopt_agency by race


(2,3) with satisfaction_parenting
2. /analysis = satisfaction_adopt_agency
3. /design = satisfaction_parenting,race,
satisfaction_parenting by race.
52 Analysis of Multiple Dependent Variables

Tests of significance for satisfaction_adopt_agency using


UNIQUE sums of squares
Source of Variation SS DF MS F Sig of F

WITHIN+RESIDUAL 173.94 60 2.90


SATISFACTION_PARENTI 2.11 1 2.11 .73 .397
NG
RACE 13.38 2 6.69 2.31 .108
SATISFACTION_PARENTI 13.39 2 6.69 2.31 .108
NG BY RACE

(Model) 55.59 5 11.12 3.83 .004


(Total) 229.53 65 3.53

R-Squared = .242
Adjusted R-Squared = .179

Figure 2.7 Group 1 versus Group 3.

In each set of SPSS commands separates the grouping variable


(race) from the covariates with the key word with (i.e., with
satisfaction_parenting).
These three sets of SPSS commands yield the following three out-
puts (Figures 2.7, 2.8, and 2.9). Note that the focus of each analysis is
on the statistical significance of the interaction between race and sat-
isfaction with parenting (highlighted with a rectangle in the output).

Tests of significance for satisfaction_adopt_agency using


UNIQUE sums of squares
Source of Variation SS DF MS F Sig of F

WITHIN+RESIDUAL 62.20 31 2.01


SATISFACTION_PARENTI 5.08 1 5.08 2.53 .122
NG
RACE 6.50 1 6.50 3.24 .082
SATISFACTION_PARENTI 7.42 1 7.42 3.70 .064
NG BY RACE

(Model) 35.97 3 11.99 5.98 .002


(Total) 98.17 34 2.89

R-Squared = .366
Adjusted R-Squared = .305

Figure 2.8 Group 1 versus Group 2.


Multivariate Analysis of Variance 53

Tests of significance for satisfaction_adopt_agency using


UNIQUE sums of squares
Source of Variation SS DF MS F Sig of F

WITHIN+RESIDUAL 156.63 46 3.41


SATISFACTION_PARENTI 4.41 1 4.41 1.30 .261
NG
RACE 13.28 1 13.28 3.90 .054
SATISFACTION_PARENTI 13.11 1 13.11 3.85 .056
NG BY RACE

(Model) 14.09 3 4.70 1.38 .261


(Total) 170.72 49 3.48

R-Squared = .083
Adjusted R-Squared = .023

Figure 2.9 Group 2 versus Group 3.

In each analysis the race and satisfaction with parenting interaction


term is not statistically significant at p = 0.05. Therefore, homogene-
ity of regression was achieved for all components of the stepdown
analysis.

Stata Commands
1. findit xi3
2. install
3. xi3: regress satisfaction_adopt_agency
g.race* satisfaction_parenting

SAS Commands
1. proc glm data = file name;
2. class race;
3. model satisfaction_adopt_agency =
race satisfaction_parenting
race*satisfaction_parenting;
4. run;
5. quit;
54 Analysis of Multiple Dependent Variables

Performing the MANOVA


In SPSS, proceed as follows:

1. Click Analyze → General Linear Model → Multivariate


2. Move the DVs to the Dependent Variables box.
3. Move the IVs to the Fixed Factor(s) box
4. Click the Model button, and be sure that the radio button Full
Factorial is selected
5. Click OK

The “Multivariate Tests” section of SPSS’s output reports simultane-


ous tests of each factor’s effect on the dependent groups. SPSS offers four
alternative multivariate significance tests: Pillai’s Trace, Wilk’s Lamda,
Hotelling’s Trace, and Roy’s Largest Root. The significance of the F tests
show if that effect is significant. For these data, all multivariate tests are
statistically significant at <.05. The partial eta squared values suggest a
low to moderate model fit. The focus here is on Wilk’s Lambda, with
a partial eta squared value of 0.110. Significant differences were found
among the three categories of race on the DVs, with Wilk’s Lambda =
.793, F(4,124) = 3.816, p <.01 (see Figure 2.10).
A Roy–Bargmann stepdown analysis was conducted to further
explore the statistical significance of differences on the DVs by race
(see Figure 2.8). It must be noted that the homogeneity of regressions
assumption is necessary to perform a Roy–Bargmann stepdown analy-
sis. Recall that according to the above analysis, this assumption is tenable

Multivariate TestsC
Effect Partial Eta
Value F Hypothesis df Error df Sig. Squared
Intercept Pillai’s Trace .999 48837.040a 2.000 62.000 .000 .999
Wilks’ Lambda .001 48837.040a 2.000 62.000 .000 .999
Hotelling’s Trace 1575.388 48837.040a 2.000 62.000 .000 .999
Roy’s Largest Root 1575.388 48837.040a 2.000 62.000 .000 .999
Race Pillai’s Trace .211 3.724 4.000 126.000 .007 .106
Wilks’ Lambda .793 3.816a 4.000 124.000 .006 .110
Hotelling’s Trace .256 3.904 4.000 122.000 .005 .113
Roy’s Largest Root .233 7.342b 2.000 63.000 .001 .189
a. Exact statistic
b. The statistic is an upper bound on F that yields a lower bound on the significance level.
c. Design: Intercept+Race

Figure 2.10 Multivariate Tests.


Multivariate Analysis of Variance 55

for the current model. All DVs were judged to be sufficiently reliable
to warrant stepdown analysis, since the reported reliability coefficients
for and were and respectively. Both the satisfaction with agency services
and satisfaction with parenting scales are considered reliable. Both scales
have associated Cronbach’s alphas greater than .80. To conduct a Roy–
Bargmann stepdown analysis with two DVs, one ANCOVA is performed
with the most import DV as the DV, and the next most important DV
as a covariate. (Note that, for example, with three DVs, an ANCOVA is
performed with the most important DV as the DV, and the next most
important DV as a covariate. Then, an ANCOVA is performed with the
next most important DV as the DV, and the next two most important
DVs as the covariates.)
To conduct a Roy–Bargmann stepdown analysis with two DVs, pro-
ceed as follows in SPSS:

1. Click Analyze → General Linear Model → Univariate


2. Move the DV satisfaction_adopt_agency to the Dependent
Variables box.
3. Move the DV satisfaction_parenting to the Covariate(s) box
4. Click the Model button, and be sure that the radio button Full
Factorial is selected
5. Click OK

Results of the stepdown analysis are summarized in Figure 2.11. No


unique contribution to predicting differences among racial categories was

Dependent Variable: satisfaction_adopt_agency


Source Type III Sum
of Squares df Mean Square F Sig.
Corrected Model .088a 1 .088 .024 .876
Intercept 219.117 1 219.117 61.120 .000
satisfaction_parenting .088 1 .088 .024 .876
Error 229.442 64 3.585
Total 169963.000 66
Corrected Total 229.530 65
a. R Squared = .000 (Adjusted R Squared = –.015)
Figure 2.11 Tests of Between-Subject Effects.
56 Analysis of Multiple Dependent Variables

made by satisfaction with agency services (satisfaction_adopt_agency),


F(1, 66) = 0.024, p = .876.
To further explore differences among racial categories on satisfac-
tion with agency services, Table 2.1 displays the mean level of satisfac-
tion with agency services simultaneous confidence intervals by racial
group. African American respondents report the highest level of satisfac-
tion with agency services followed by Asian American, and Caucasian
American respondents. Table 2.1 was produced with the following SPSS
commands:

1. Click Analyze → General Linear Model → Multivariate


2. Move the DVs to the Dependent Variables box
3. Move the IVs to the Fixed Factor(s) box
4. Click the Model button, and be sure that the radio button Full
Factorial is selected
5. Click the Options button, move Race into the Display Means for
window
6. Check Compare main effects
7. For Confidence interval adjustment and select HSD(none)
8. Click Continue
9. Click OK

Stata Commands to Perform a MANOVA


1. manova satisfaction_parenting satisfaction_
adopt_agency = Race

Table 2.1 Estimates


Estimates
Dependent Race Mean Std. Error 95% Confidence Interval
Variable
Lower Bound Upper Bound

satisfaction_ 1 50.500 .446 49.609 51.391


parenting 2 49.737 .409 48.919 50.555
3 50.323 .320 49.682 50.963
satisfaction_ 1 49.312 .431 48.451 50.174
adopt_agency 2 51.000 .396 50.209 51.791
3 51.258 .310 50.639 51.877
Multivariate Analysis of Variance 57

2. anova satisfaction_parenting Race


3. regress
4. anova satisfaction_adopt_agency Race
5. regress

Note that step 3 and 5 calculate simultaneous confidence intervals.

SAS Commands to Perform a MANOVA


In SAS the proc glm procedure uses the following general form:

1. Proc glm;
2. class IV;
3. model DVs = IV;
4. means IV / LSD alpha = p-value;
5. manova h=IV;
6. run;

Note that the aforementioned commands are defined as CLASS


<names of the variables to be used as ANOVA factors in the model>; and
MODEL <dependent variable(s)> = <independent variable(s)>.
For this analysis, the following commands were used:

1. proc glm;
2. class race;
3. model satisfaction_parenting satisfaction_
adopt_agency = race;
4. means race / LSD alpha = .05;
5. manova h=race;
6. run;

Note that Step 4 calculates simultaneous confidence intervals.

SAS Commands to Perform a Roy–Bargmann Stepdown Analysis


1. proc glm;
2. class race;
3. model satisfaction_adopt_agency = race
satisfaction_parenting;
58 Analysis of Multiple Dependent Variables

4. means race / LSD alpha = .05;


5. manova h=race satisfaction_parenting
race*satisfaction_parenting;
6. run;

Stata Commands to Perform a Roy–Bargmann Stepdown Analysis


manova satisfaction_adopt_agency = RACE c.satisfaction_parenting

REPORTING THE RESULTS OF A MANOVA


1. Restate in summary form the reason(s) for the analysis, and
the basic components of the model(s) tested including the DV
and IVs;
2. Describe how the assumptions underlying the model were tested,
and, if not tenable, treated. Include a discussion of missing
data, outliers, equality of variance–covariance matrices, and
correlation among correlation among DVs;
3. Report the results of at least one multivariate F-test;
4. Report a measure of effect size;
5. If the multivariate F-test is significant, describe how further the
statistical significance of differences on the DVs by group were
further explored; and
6. Summarize the results of model testing in terms of the study’s
hypotheses.

Results
This study tested a model that compared a client’s satisfaction with an
adoption agency’s services and their satisfaction with being an adoptive
parent by the client’s race (coded 1 = Caucasian, 2 = Asian American,
3 = African American). Satisfaction with agency services and satisfaction
with parenting were operationalized as scale scores. Race was operation-
alized as clients’ self-reports of their racial group.
A one-way multivariate analysis of variance (MANOVA) was per-
formed to determine the effect of being a member of three racial groups
(i.e., African, Asian, and Caucasian American) on a client’s satisfaction
Multivariate Analysis of Variance 59

with an adoption agency’s services and their satisfaction with being an


adoptive parent.
No problems were noted for missing data, since 0% was missing, out-
liers, or multivariate normality. The tenability of the absence of outli-
ers assumption was evaluated using Cook’s distance, D, which provides
an overall measure of the impact of an observation on the estimated
MANOVA model. Observations with larger values of D than the rest of
the data are those which have unusual leverage. For these data, no value
of D exceeded .06. These results suggest that the assumption of absence of
outliers is tenable. The tenability of the multivariate normality assump-
tion was evaluated by using an SPSS macro developed by DeCarlo (1997).
All tests of multivariate skew, multivariate kurtosis, and an omnibus test
of multivariate normality were not statistically significant at p = .05. These
results suggest that the assumption of multivariate normality is tenable.
Box’s M was used to test the assumption (i.e., H0) of equality of vari-
ance–covariance matrices. Box’s M equaled 12.046, F(6, 31697) = 1.906,
p = .076, which means that equality of variance-covariance matrices can
be assumed.
Significant differences were found among the three categories of race
on the DVs, with Wilk’s Lambda = .793, F(4,124) = 3.816, p < .01. That
is, there are statistically significant differences on the DVs based on race.
A Roy–Bargmann stepdown analysis and simultaneous confidence inter-
vals, analysis were conducted to further explore these differences.
For these data, at a statistically significant level, there are mean dif-
ferences between racial groups on level of satisfaction with the adoption
agency (see Table 2.2). At a statistically significant level, African American
respondents report the highest level of satisfaction with the adoption

Table 2.2 Means, and Standard Deviations for each Dependent Variable
by Race
Caucasian Asian American African
N = 16 N = 19 American
Mean(SD) Mean(SD) N = 31
Mean(SD)
Satisfaction w/ Adoption 49.31(1.078)* 51.00(1.764)* 51.26(1.949)*
Agency
Satisfaction w/ Parenting 50.50(1.826) 40.74(1.485) 50.32(1.922)
*Significant at p < .05
60 Analysis of Multiple Dependent Variables

agency followed by Asian American, and Caucasian American respon-


dents (see Table 2.2). Although these results are statistically significant,
they may not be considered practically significant because group means
on the satisfaction agency scale ranged from 49.31 to 51.26.

ADDITIONAL EXAMPLES OF MANOVA FROM THE APPLIED RESEARCH LITERATURE


Colarossi, L. G. (2001). Adolescent gender differences in social support: structure,
function, and provider type. Social Work Research, 25(4), 233–242.
This study used survey data on 364 adolescents to examine gender differ-
ences in perceptions of support across three different constructs: (1) structural
support related to the number of adults versus friends; (2) the quantity of
support provided by mothers, fathers, peers, and teachers; (3) satisfaction with
support from friends and family. Results of a MANOVA indicated that young
women report a greater number of supportive friends and receive more fre-
quent support from their friends than do young men. However, young men are
just as satisfied with friend support as are young women. No gender differences
were found in the number of adult supporters, but males received more fre-
quent support from fathers. Nonparental adults emerged as important sources
of support for both genders. Implications for practice with adolescents are
discussed.

Jewell, J. D., & Stark, K. D. (2003). Comparing the family environments of adoles-
cents with conduct disorder or depression. Journal of Child and Family Studies,
12(1), 77–89.
This study attempted to differentiate the family environments of youth
with Conduct Disorder (CD) compared to youth with a depressive disorder.
Participants were 34 adolescents from a residential treatment facility. The
K-SADS-P was used to determine the youth’s diagnosis, while their family
environment was assessed by the Self Report Measure of Family Functioning
Child Version. A MANOVA was used to compare the two diagnostic groups on
seven family environment variables. Results indicated that adolescents with
CD described their parents as having a permissive and ambiguous discipline
style, while adolescents with a depressive disorder described their relation-
ship with their parents as enmeshed. A discriminant function analysis, using
the two family environment variables of enmeshment and laissez-faire family
style as predictors, correctly classified 82% of the participants. Implications
for treatment of youth with both types of diagnoses and their families are dis-
cussed. (Journal abstract.).
Multivariate Analysis of Variance 61

Mowbray, C. ; Oyserman, D ; Bybee, D. ; MacFarlane, P. (2002). Parenting of moth-


ers with a serious mental illness: differential effects of diagnosis, clinical his-
tory, and other mental health variables. Social Work Research, 26(4), 225–231.
This study examined the effects of mental illness on parenting in a large
urban-based sample of women with serious mental illness. Seventy percent of
the sample were women from ethnic minority groups, average age mid-30s;
all had care responsibility for at least one minor child. Diagnostic Interview
Schedule modules were administered; the women were interviewed to obtain
information on parenting, clinical history, and current functioning. MANOVA
results suggested that diagnosis had a small but significant negative effect on par-
enting attitudes and behaviors, and there were race-by-diagnosis interactions.
However, current symptoms mediated the effects of diagnosis and chronicity
on parenting stress, and current symptomatology and community functioning
partially mediated the effects of diagnosis on parenting satisfaction.

O’Hare T. (1998). Alcohol expectancies and excessive drinking contexts in young


adults. Social Work Research, 22(1), p44–50.
Despite the growth of alcohol expectancy research, few investigations have
examined the association between beliefs in the reinforcing effects of alcohol
and situations in which young people drink excessively. The current study
of 315 youthful drinkers examined the relationship between three alcohol
expectancies (increased social assertiveness, tension reduction, and enhanced
sexual pleasure), as measured by the Alcohol Expectancy Questionnaire, and
three drinking situation subscales (convivial, personal-intimate, and nega-
tive coping) from the Drinking Context Scale. MANOVA showed significant
main effects for the three expectancy measures. Univariate results showed
that expectancies of social assertiveness and tension reduction varied directly
with all three excessive drinking contexts. The expectancy of enhanced sexual
pleasure, however, varied significantly with personal-intimate drinking only.
Practice implications with youthful drinkers are discussed.

Thevos, A. K., Thomas, S. E., & Randall, C. L. (2001). Social support in alco-
hol dependence and social phobia: Treatment comparisons. Research on Social
Work Practice, 11, 458–472.
This study investigated whether different alcoholism treatment approaches
differentially impact social support scores in individuals with concurrent alco-
hol dependence and social phobia. Individuals (N = 397) were selected retro-
spectively from a larger pool of participants enrolled in a multisite randomized
clinical trial on treatment matching. Three standard treatments were delivered
over 12 weeks: Cognitive-Behavioral Therapy (CBT), Twelve Step Facilitation
Therapy (TSF), and Motivational Enhancement Therapy (MET). MANOVA
62 Analysis of Multiple Dependent Variables

was used to analyze social support measures to test the effects of treatment
group and gender. For men, there was significant improvement on two mea-
sures of social support regardless of treatment group. Women who received
CBT or TSF had better support outcomes than women who received MET.

Tsan, J. Y. (2007). Personality and gender as predictors of online counseling use.


Journal of Technology in Human Services, 25(3), 39–56.
Extraversion, neuroticism, and gender as predictors of online counseling
help seeking behavior were investigated. A total of 176 college student
participants, 30 men and 146 women, were given the NEO-PI R and
ATSPPH-S that assessed, respectively, their personalities and attitudes toward
seeking professional psychological help through different modes: traditional
face-to-face counseling, video-conferencing, e-mail, instant text message,
and microphone. Results were analyzed using MANOVA and MANCOVA.
Subjects were grouped by scores on personality variables: low, medium,
and high. Findings suggest that attitudes toward different modes of seeking
counseling were associated primarily with gender and extraversion but not
with neuroticism.
3

Multivariate Analysis
of Covariance

In ANOVA, the mean differences between three or more groups on a


single DV are evaluated. ANCOVA assesses group differences on a DV
after the effects of one or more covariates are statistically removed. By
utilizing the relationship between the covariate(s) and the DV, ANCOVA
can increase the power of an analysis.
MANCOVA is an extension of ANCOVA to relationships where a
linear combination of DVs is adjusted for differences on one or more
covariates. The adjusted linear combination of DVs is the combina-
tion that would be obtained if all participants had the same scores on
the covariates. That is, MANCOVA is similar to MANOVA, but allows
a researcher to control for the effects of supplementary continuous IVs,
termed covariates. In an experimental design, covariates are usually the
variables not controlled by the experimenter, but still affect the DVs.
Consequently, although not as effective as random assignment, including
covariates may reduce both systematic and within-group error by equal-
izing groups being compared on important characteristics.
For example, a researcher is planning a study to compare levels of
self-esteem and depression among Caucasian, African American, and
Mexican American juvenile offenders in a substance-abuse treatment pro-
gram. Before and after completing the treatment, program participants

63
64 Analysis of Multiple Dependent Variables

will be administered a self-esteem scale and a depression scale. It may be


useful to adjust the DV scores for pretreatment differences in self-esteem
and depression. Here the covariates are pretests of the DVs, a classic use
of covariance analysis. After adjustment for pretreatment score differ-
ences, post test score differences may be more accurately attributed to
treatment.

ASSUMPTIONS OF MANCOVA
For MANCOVA, all of the assumptions for MANOVA apply, with the fol-
lowing additions: (1) covariates are measured without error; (2) a linear
relationship between the DV and the covariates; and (3) homogeneity of
the regression hyperplanes. This assumption, also termed homogeneity
of regressions or parallelism, requires that the regression slopes between
the covariate and DV are the same (homogeneous) for all groups.
According to Stevens (2002), a violation of this assumption means that
there is a statistically significant covariate-by-IV interaction. Conversely,
if the interaction is not statistically significant, this assumption is met.
To test this assumption for a MANCOVA with one covariate, then,
a model that contains a covariate-by-IV interaction term is tested. It is
hoped that this interaction term does not achieve statistical significance
(e.g., p > .05). If there is more than one covariate, then, for each covariate,
a covariate-by-IV interaction term is included in the model. It is hoped
that all covariate-by-IV interaction terms are not statistically significant.
MANCOVA AND MANOVA are similar in terms of (1) evaluating
the omnibus null hypothesis, (2) assessing overall model fit, (3) explor-
ing multivariate group differences, and (4) model validation.

SAMPLE SIZE REQUIREMENTS


As in MANOVA, the measure of effect size in MANCOVA is f 2 (Cohen,
1988). Resources for estimating sample size for MANCOVA are difficult
to identify. One approach is to adapt the aforementioned sample size esti-
mation strategy for MANOVA. That is, use GPower, which is a free power
analysis program available at http://www.psycho.uniduesseldorf.de/
aap/projects/gpower/ and adjust the denominator df. If k is the number
Multivariate Analysis of Covariance 65

of cells (IVs by DVs) in the design and g is the number of covariates, then
groups = k + g. This strategy for determining sample size is demonstrated
in the annotated example below.

STRENGTHS AND LIMITATIONS OF MANCOVA


MANCOVA has the following analytical strengths:

1. May provide protection against inflated type I error as a result of


multiple tests of correlated DVs;
2. Group differences, not evident in separate ANCOVAs, may
become apparent;
3. Addition of covariates may reduce systematic bias; and
4. Addition of covariates may reduce within group or error variance.

Regarding analytical strengths three and four, the goal is to obtain


maximum adjustment of the DV with minimum loss of degrees of free-
dom for the error term. Each covariate “costs” one degree of freedom for
error. Also, when there is substantial correlation between a covariate and
the DV, the gain in terms of power often offsets the loss of power due to
reduced degrees of freedom. However, with multiple covariates, a point
of diminishing returns is quickly reached, particularly if the covariates
are correlated with one another.
In studies with relatively small group sizes (< 20), it is particularly
imperative to consider the use of no more than three covariates, because
for small or medium effect sizes (f 2 ≤ .15) power will be low for small
group size. Huitema (1980) recommends limiting the number of covari-
ates as follows:

C (J − )
< .10. (3.1)
N
This formula can be re-written as follows:

C = (.10N) − (J − 1), (3.2)

where C is the number of covariates, J is the number of groups, and N is


total sample size. Therefore, for a three-group problem with a total of 60
66 Analysis of Multiple Dependent Variables

participants, then C < 6 − 2 = 4; that is, four or more covariates are used,
then estimates of the adjusted means are likely to be unstable.
All covariates should be correlated with the DV, and none should be
substantially correlated with each other. If covariates are substantially
correlated with each other, they may not add significantly to reduction
of error, or they may cause computational difficulties such as multicol-
linearity. In addition, to avoid confounding of the intervention effect with
a change on the covariate, one should only use pretest or other informa-
tion gathered before the intervention begins as covariates. If a covariate
is used that is measured after the intervention beings and that variable
was affected by the intervention, then the change on the covariate may
be correlated with change of the DV. Consequently, when the covariate
adjustment is made, part of the intervention effect is removed.
MANCOVA has the following analytical limitations:

1. The assumption of equality of variance–covariance matrices is more


difficult to satisfy than its bivariate analog of equality of variance;
2. There may be ambiguity in interpretation of the effects of IVs on
any single DV;
3. Moderately correlated DVs may diminish the power of
MANCOVA; and
4. When there is more than one DV, there may be redundancy in
the measurement of outcomes.

ANNOTATED EXAMPLE
A study is conducted to test a model that compares satisfaction with
agency services and satisfaction with parenting by client race, controlling
for client self-efficacy. Satisfaction with agency services, satisfaction with
parenting, and client self-efficacy were operationalized as scale scores. Race
was operationalized as clients’ self-reports of their racial group. The three
scales have associated Cronbach’s alphas greater than .80. See chapter 2
for a discussion of reliability.
The data analysis strategy used here is consistent with the recom-
mendations of several prominent authors (cf. Tabachnick & Fidell, 2007;
Stevens, 2009) to use Roy–Bargmann stepdown analysis and simul-
taneous confidence intervals to follow-up comparisons to elucidate
Multivariate Analysis of Covariance 67

a significant MANOVA result. See chapter 2 for a discussion of this data


analysis strategy.
Accordingly, the researcher decides to perform a one-way MANCOVA
to investigate the relationship between a client’s race (IV), and her/his
satisfaction with the adoption agency (DV1), and satisfaction with parent-
ing (DV2), while controlling for self-efficacy (i.e., the covariate). Note that
throughout this annotated example SPSS, Stata, and SAS commands are
numbered in sequence and highlighted in Courier. The data set used
in this annotated example, entitled Annotated_Example1_FULL_N=54.
sav, may be downloaded from this book’s companion website. These data
are in SPSS format, but may be imported into Stata and SAS.

Determining the Minimally Sufficient Sample Size


GPower, which is a free power analysis program available at http://www.
psycho.uni-duesseldorf.de/aap/projects/gpower/ is used to determine the
minimally sufficient sample size. The following assumptions are made:
α = .05, β = .20, f 2 = 0.15 (moderate), the number of groups is three,
and the number of DVs equals the number of DVs plus the number of
covariates, or three. A total sample size of 54 is adequate.

Prescreening for Multivariate Normality


The tenability of the mulitvariate normality assumption was evaluated
by using an SPSS macro developed by DeCarlo (1997). To use the macro,
open the file Annotated_Example1_FULL_N=54.sav in SPSS, and type
the following commands (numbered in sequence and highlighted in
Courier) into a new SPSS syntax window:

1. include ‘C:\Users\pat\Desktop\MANCOVA_Ex_1\
normtest.sps’
2. normtest vars = satisfaction_
parenting,satisfaction_adopt_agency,
self_efficacy/.
3. execute.

The first line of the aforementioned syntax includes (calls) the macro
entitled normtest.sps, which is located in the directory C:\Users\pat\
68 Analysis of Multiple Dependent Variables

Desktop\MANCOVA_Ex_1\, and the second line invokes the macro for


the variables satisfaction_parenting, satisfaction_adopt_agency, self_effi-
cacy. This macro is available from http://www.columbia.edu/~ld208/
and this book’s companion website.

1. Select run→all.

The results of DeCarlo’s (1997) are summarized in Figure 3.1. For brev-
ity, the focus here is on Small’s (1980) and Srivistava’s (1984) tests of
multivariate kurtosis and skew, Mardia’s (1970) multivariate kurtosis,
and an omnibus test of multivariate normality based on Small’s statis-
tics (see Looney, 1995). However, all other tests calculated by the macro
were consistent with these results. That is, all tests of multivariate skew,

Tests of multivariate skew:

Small’s test (chisq)


O1 df p-value
4.8470 3.0000 .1833

Srivastava’s test
chi (b1p) df p-value
7.5667 3.0000 .0559

Tests of multivariate kurtosis:

A variant of small’s test (chisq)


VO2 df p-value
6.4722 3.0000 .0908

Srivastava’s test
b2p N (b2p) p-value
3.2197 .5707 .5682

Mardia’s test
b2p N (b2p) p-value
15.9375 .6289 .5294

Omnibus test of multivariate normality:

(based on Small’s test, chisq)


VO3 df p-value
11.3192 6.0000 .0790

Figure 3.1 Tests of Multivariate Skew.


Multivariate Analysis of Covariance 69

multivariate kurtosis, and an omnibus test of multivariate normality


were not statistically significant at p = .05. Each of these tests evaluates
the H0 that the current distribution (these data) equals the multivari-
ate normal distribution. It is hoped that this H0 will fail to be rejected
(accepted). These results suggest that the assumption of multivariate
normality is tenable. See chapter 1 for additional information about
these tests.

Stata Commands to Prescreen for Multivariate Normality


Mvtest performs multivariate tests of univariate, bivariate, and multi-
variate normality. All multiple-sample tests provided by mvtest assume
independent samples.

1. findit mvtest
2. install
3. mvtest normality satisfaction_parenting
satisfaction_adopt_agency self_efficacy,
stats(all)

SAS Commands to Prescreen for Multivariate Normality


The %MULTNORM macro provides tests and plots of multivariate nor-
mality. A test of univariate normality is also given for each of the vari-
ables. The macro is available from http://support.sas.com/kb/24/983.
html. To use the macro, type the following commands in the SAS editor
window:

1. %inc “<location of your file containing the


MULTNORM macro>”;
2. %multnorm(data=cork, var=n e s w,
plot=mult, hires=no)

For this second command statement, DATA = SAS data set to be ana-
lyzed. If the DATA = option is not supplied, the most recently created SAS
data set is used. VAR = the list of variables to used. Individual variable
names, separated by blanks, must be specified. PLOT = MULT requests
a high- or low-resolution chi-square quantile–quantile (Q-Q) plot of
70 Analysis of Multiple Dependent Variables

the squared Mahalanobis distances of the observations from the mean


vector. PLOT = UNI requests high-resolution univariate histograms
of each variable with overlaid normal curves and additional univariate
tests of normality. Note that the univariate plots cannot be produced if
HIRES = NO.PLOT = BOTH (the default) requests both of the above.
PLOT = NONE suppresses all plots and the univariate normality tests.
HIRES = YES (the default) requests that high-resolution graphics be
used when creating plots. You must set the graphics device (GOPTIONS
DEVICE =) and any other graphics-related options before invok-
ing the macro. HIRES = NO requests that the multivariate plot be
drawn with low-resolution. The univariate plots are not available with
HIRES = NO.
For these data, the following commands were used:

1. %inc “<location of your file containing the


MULTNORM macro>”;
2. %multnorm(data=file name, var =
satisfaction_parenting satisfaction_
adopt_agency self_efficacy, plot = mult,
hires=no)

Prescreening for Absence of Outliers


The tenability of the absence of outliers assumption was evaluated by
using SPSS to calculate Cook’s distance, D, which provides an overall
measure of the impact of an observation on the estimated MANOVA
model. To calculate Cook’s D in SPSS, proceed as follows:

1. Click Analyze → General Linear Model → Multivariate


2. Move the DVs to the Dependent Variables box
3. Move the IVs to the Fixed Factor(s) box
4. Move the covariate(s) to the Covariate(s) box
5. Click the Model button, and be sure that the radio button Full
Factorial is selected
6. Click the Save button and select Cook’s distance
7. Click OK
Multivariate Analysis of Covariance 71

Criterion Calculated Value for These Data


D greater than 4/(n - k - 1) 4/(54-2-1) = .07

D>1 1

D > 4/n 4/54 = .07

Figure 3.2 Three Citeria for Evaluating Cook’s D.

As a result of the aforementioned steps, two variables are added to


the data file (one for each DV). These variables are named COO_1 and
COO_2. Values for these two variables are the values of D for each DV.
Observations with larger values of D than the rest of the data are those
which have unusual leverage. Some authors suggest as a cut-off for detect-
ing influential cases, values of D greater than 4/(n − k − 1), where n is the
number of cases and k is the number of independents. Others suggest
D > 1 as the criterion to constitute a strong indication of an outlier prob-
lem, with D > 4/n the criterion to indicate a possible problem. Figure 3.2
summarizes Cook’s D cutoffs based on the three aforementioned crite-
ria. For these data, five cases had values greater than .07 (.08–to .11).
If they appear atypical, cases with values of D greater than .07 may be
deleted. See chapter 1 for additional discussion about the management of
outliers. Alternatively, MANCOVA results should be reported with and
without the outliers. Note that for these cases, MANCOVA results with
and without cases that had values on one or both DVS that exceeded
.07 yielded equivalent results. Consequently, only results for all cases are
reported below.

Stata Commands to Prescreen for Absence of Outliers


1. regress Race satisfaction_parenting
satisfaction_adopt_agency self_efficacy
2. _predict cooksd

Cook’s D values will be added to the data file as coo_1 and COO_2.

SAS Commands to Prescreen for Absence of Outliers


1. proc reg data=“file path”;
72 Analysis of Multiple Dependent Variables

2. model file Race=satisfaction_parenting


satisfaction_adopt_agency self_efficacy;
3. output out= resdata cookd=cooksd
4. run;

Cook’s D values will be saved in a new file named “resdata.”

Prescreening for Homoscedasticity


MANOVA assumes that for each group (each cell in the factor-design
matrix) the covariance matrix is similar. Box’s M tests the null hypothesis
that the observed covariance matrices of the DVs are equal across groups.
To calculate Box’s M in SPSS, proceed as follows:

1. Click Analyze → General Linear Model → Multivariate


2. Move the DVs to the Dependent Variables box
3. Move the IVs to the Fixed Factor(s) box
4. Move the covariate(s) to the Covariate(s) box
5. Click the Model button, and be sure that the radio button Full
Factorial is selected
6. Click the Options button and select Homogeneity tests
7. Click OK

Researchers do not want M to be significant in order to conclude there


is insufficient evidence that the covariance matrices differ. Note, however,
that the F-test is quite robust even when there are departures from this
assumption. Here M is significant, and, consequently, these suggest that
these data exhibit low to moderate heterogeneity (p = 0.45). MANCOVA
is robust against moderate violations of the homoscedasticity assump-
tion (see Figure 3.3).

Stata Commands to Prescreen for Homoscedasticity


mvtest covariances satisfaction_parenting
satisfaction_adopt_agency self_efficacy,
by(Race)
Multivariate Analysis of Covariance 73

Box’s M 13.658
F 2.147
df1 6
df2 64824.923
Sig. .045

Figure 3.3 Box’s Test of Equality of Covariance Matrices.

SAS Commands to Prescreen for Homoscedasticity


1. proc discrim;
2. class Race;
3. var satisfaction_parenting satisfaction_
adopt_agency self_efficacy;
4. pool=test wcov;
5. run;

Prescreening for Homogeneity of Regressions


Because a Roy–Bargmann stepdown analysis is planned, data are pre-
screened for homogeneity of regressions. This assumption, also termed
homogeneity of the regression hyperplanes or parallelism, requires that
the regression slope between the covariate and DV is the same (homo-
geneous) for all groups. According to Stevens (2002), a violation of this
assumption means that there is a statistically significant covariate-by-IV
interaction. Conversely, if the interaction is not statistically significant
this assumption is met.
To test this assumption for a MANOVA with one DV, a model that
contains one covariate (i.e., the lower importance DV)-by-IV interaction
term is tested for each pair of groups. It is hoped that this interaction
term does not achieve statistical significance (e.g., p > .05). If there is
more than one DV, then, one covariate-by-IV interaction term is sequen-
tially included in the model for each DV, and each sequential model is
tested for each pair of groups. It is hoped that all covariate-by-IV interac-
tion terms are not statistically significant.
74 Analysis of Multiple Dependent Variables

To test the homogeneity of regressions assumption for a model that


contains a client’s race (IV), her/his satisfaction with the adoption agency
(DV1), and satisfaction with parenting (DV2), assuming satisfaction with
the adoption agency is logically prior to satisfaction with parenting, pro-
ceed as follows:

1. File → New → Syntax


2. Paste the following three blocks of commands into this
window:
MANOVA satisfaction_adopt_agency by race
(1,3) with satisfaction_parenting
self_efficacy
/analysis = satisfaction_adopt_agency
/design = satisfaction_parenting,race,
satisfaction_parenting by race self_
efficacy by race.
MANOVA satisfaction_adopt_agency by race
(1,2) with satisfaction_parenting
self_efficacy
/analysis = satisfaction_adopt_agency
/design = satisfaction_parenting,race,
satisfaction_parenting by race self_
efficacy by race.
MANOVA satisfaction_adopt_agency by race
(2,3) with satisfaction_parenting
self_efficacy
/analysis = satisfaction_adopt_agency
/design = satisfaction_parenting,race,
satisfaction_parenting by race self_
efficacy by race.
3. Run → All

In each set of commands, SPSS separates the grouping variable (race)


from the covariates with the keyword with (i.e., with satisfaction_
parenting self_efficacy).
These three sets of SPSS commands yield the following three outputs
(Figures 3.4, 3.5, and 3.6). Note that the focus of each analysis is on the
statistical significance of the interaction between race and satisfaction
Multivariate Analysis of Covariance 75

Tests of significance for satisfaction_adopt_agency using


UNIQUE sums of squares
Source of Variation SS DF MS F Sig of F

WITHIN+RESIDUAL 116.09 46 2.52


SATISFACTION_PARENTI .00 1 .00 .00 .986
NG
RACE 7.35 2 3.68 1.46 .244
SATISFACTION_PARENTI 18.28 2 9.14 3.62 .055
NG BY RACE
SELF_EFFICACY BY RAC 2.71 2 1.35 .54 .588
E

(Model) 55.91 7 7.99 3.17 .008


(Total) 172.00 53 3.25

R-Squared = .325
Adjusted R-Squared = .222
Figure 3.4 Group 1 versus Group 3.

Tests of significance for satisfaction_adopt_agency using


UNIQUE sums of squares
Source of Variation SS DF MS F Sig of F

WITHIN+RESIDUAL 59.98 31 1.93


SATISFACTION_PARENTI 3.76 1 3.76 1.95 .173
NG
RACE .15 1 .15 .08 .785
SATISFACTION_PARENTI 5.77 1 5.77 2.98 .094
NG BY RACE
SELF_EFFICACY BY RAC 2.54 1 2.54 1.31 .261
E

(Model) 34.99 4 8.75 4.52 .005


(Total) 94.97 35 2.71

R–Squared = .368
Adjusted R–Squared = .287
Figure 3.5 Group 1 versus Group 2.
76 Analysis of Multiple Dependent Variables

Tests of significance for satisfaction_adopt_agency using


UNIQUE sums of squares
Source of Variation SS DF MS F Sig of F

WITHIN+RESIDUAL 98.3 31 3.17


SATISFACTION_PARENTI .06 1 .06 .02 .894
NG
RACE 6.45 1 6.45 2.03 .164
SATISFACTION_PARENTI 19.23 1 19.23 6.06 .052
NG BY RACE
SELF_EFFICACY BY RAC 1.32 1 1.32 .42 .523
E

(Model) 21.19 4 5.30 1.67 .182


(Total) 119.56 35 3.42

R–Squared = .177
Adjusted R–Squared = .071
Figure 3.6 Group 2 versus Group 3.

with parenting and race and self efficacy (highlighted with rectangles in
the output). In each analysis the race and satisfaction with parenting is
not statistically significant at p = .05. Therefore, homogeneity of regres-
sion was achieved for all components of the stepdown analysis.

Stata Commands
1. findit xi3
2. install
3. xi3: regress satisfaction_adopt_
agency g.race* satisfaction_parenting
g.race*self_efficacy

SAS Commands
1. proc glm data = file name;
2. class race;
3. model satisfaction_adopt_agency = race
satisfaction_parenting race*satisfaction_
parenting race*self_efficacy;
4. run;
5. quit;
Multivariate Analysis of Covariance 77

Performing the MANOVA


In SPSS, proceed as follows:

1. Click Analyze → General Linear Model → Multivariate


2. Move the DVs to the Dependent
Variables box
3. Move the IVs to the Fixed Factor(s)
box
4. Move the covariate(s) to the Covariate(s)
box
5. Click the Model button, and be sure that the radio button Full
Factorial is selected
6. Click OK

The “Multivariate Tests” section reports simultaneous tests each


factor’s effect on the dependent groups. SPSS offers four alternative
multivariate significance tests: Pillai’s Trace, Wilk’s Lamda, Hotelling’s
Trace, and Roy’s Largest Root. The significance of the F tests show if that
effect is significant. For these data, all multivariate tests are statistically
significant at <.05. The partial eta squared values suggest a low to
moderate model fit. The focus here is on Wilk’s Lambda, with a partial
eta squared value of 0.128. Significant differences were found among the
three categories of race on the DVs, with Wilk’s Lambda = .760, F(4,98) =
3.600, p < .01 (see Figure 3.7).
A Roy–Bargmann stepdown analysis was conducted to further
explore the statistical significance of differences on the DVs by race
(see Figure 3.8). It must be noted that the homogeneity-of-regressions
assumption is necessary to perform a Roy–Bargmann stepdown analy-
sis. Recall that according to the above analysis, this assumption is ten-
able for the current model. All DVs were judged to be sufficiently reliable
to warrant stepdown analysis, since the reported reliability coefficients
for and were and respectively. Both the satisfaction with agency services
and satisfaction with parenting scales are considered reliable. Both scales
have associated Cronbach’s alphas greater than .80. To conduct a Roy–
Bargmann stepdown analysis with two DVs, one ANCOVA is performed
with the most import DV as the DV, and the next most important DV as
a covariate. (Note that, for example, with three DVs, an ANCOVA is per-
formed with most important DV as the DV, and the next most important
78 Analysis of Multiple Dependent Variables

Multivariate TestsC

Effect Partial Eta


Value F Hypothesis df Error df Sig. Squared
a 2.000 49.000 .734
Intercept Pillai’s Trace .734 67.568 .000
Wilks’ Lambda .266 a 2.000 49.000 .000 .734
67.568
a 2.000 49.000
Hotelling’s Trace 2.758 67.568 .000 .734
2.758 a 49.000 .000 .734
Roy’s Largest Root 67.568 2.000
a
self_efficacy Pillai’s Trace .060 1.570 2.000 49.000 .218 .060
.940 a 49.000 .218
Wilks’ Lambda 1.570 2.000 .060
a
Hotelling’s Trace .064 1.570 2.000 49.000 .218 .060
a 49.000
Roy’s Largest Root .064 1.570 2.000 .218 .060

Race Pillai’s Trace .241 3.424 4.000 100.000 .011 .120


a
Wilks’ Lambda .760 3.600 4.000 98.000 .009 .128
Hotelling’s Trace .314 3.768 4.000 96.000 .007 .136
Roy’s Largest Root .309 7.734b 2.000 50.000 .001 .236

a. Exact statistic
b. The statistic is an upper bound on F that yields a lower bound on the significanc level.
c. Design: Intercept+self_efficacy+Race

Figure 3.7 Multivariate Tests.

DV as a covariate. Then, an ANCOVA is performed with the next most


important DV as the DV, and the next two most important DVs as the
covariates.)
To conduct a Roy–Bargmann stepdown analysis with two DVs, pro-
ceed as follows in SPSS:

1. Click Analyze → General Linear Model → Univariate

Dependent Variable:satisfaction_adopt_agency
Source Type III Sum
of Squares df Mean Square F Sig.
Corrected Model 40.400a 4 10.100 3.761 .010
Intercept 140.908 1 140.908 52.466 .000
self_efficacy 6.634 1 6.634 2.470 .122
satisfaction_parenting .074 1 .074 .027 .869
Race 37.085 2 18.542 6.904 .002
Error 131.600 49 2.686
Total 136978.000 54
Corrected Total 172.000 53
a. R Squared = .235 (Adjusted R Squared = .172)
Figure 3.8 Tests of Between-Subjects Effects.
Multivariate Analysis of Covariance 79

2. Move the DV satisfaction_adopt_agency to the Dependent


Variable box
3. Move the IV to the Fixed Factor(s) box
4. Move satisfaction_parenting and self_efficacy to the Covariate(s)
box
5. Click the Model button, and be sure that the radio button Full
Factorial is selected
6. Click OK

Results of the stepdown analysis are summarized in Figure 2.11. No


unique contribution to predicting differences among racial categories
was made by self_efficacy (self efficacy), F(1,54) = 2.47. p = .122 and satis-
faction with agency services (satisfaction_adopt_agency), F(1, 54) = 0.027,
p = .869.
To further explore differences among racial categories on satisfac-
tion with agency services, Table 3.1 displays the mean level of satisfaction
with agency services with simultaneous confidence intervals by racial
group. African American respondents report the highest level of satisfac-
tion with agency services followed by Asian American, and Caucasian
American respondents. Table 3.1 was produced with the following SPSS
commands:

1. Click Analyze → General Linear Model → Multivariate


2. Move the DVs to the Dependent Variables box
3. Move the IVs to the Fixed Factor(s) box
4. Move the covariate(s) to the Covariate(s) box
5. Click the Model button, and confirm that the radio button Full
Factorial is selected (this is SPSS’s default)
6. Click the Options button, move Race into the Display Means
for window
7. Check Compare main effects
8. For Confidence interval adjustment, select HSD(none)
9. Click Continue
10. Click OK
80 Analysis of Multiple Dependent Variables

Table 3.1 Estimates


Estimates
Dependent Race Mean Std. Error 95% Confidence Interval
Variable
Lower Bound Upper Bound
satisfaction_ 1 50.4963a .433 49.626 51.366
parenting
2 49.8173a .431 48.951 50.683
3 50.0203a .443 49.130 50.910
satisfaction_ 1 49.1302a .387 48.353 49.906
adopt_agency
2 50.7633a .385 49.990 51.536
3 51.1082a .396 50.313 51.902
a.
Covariates appearing in the model are evaluated at the following values: self efficacy = 50.24.

Stata Commands to Perform a MANCOVA


1. manova satisfaction_parenting satisfaction_
adopt_agency = Race c. self_efficacy
2. anova satisfaction_parenting Race
3. regress
4. anova satisfaction_adopt_agency Race
5. regress

Note that step 3 and 5 calculate simultaneous confidence intervals.

Stata Commands to Perform a Roy–Bargmann Stepdown Analysis

manova satisfaction_adopt_agency = RACE


c.satisfaction_parenting c.self_efficacy

SAS Commands to Perform a MANOVA


In SAS the proc glm procedure uses the following general form:

1. Proc glm;
2. class IV;
Multivariate Analysis of Covariance 81

3. model DVs = IV;


4. means IV / LSD alpha = p-value;
5. manova h=IV;
6. run;

Note that the options command CLASS is followed by the names of the
variables to be used as MANOVA IVs (factors) in the model. The options
command MODEL follows the format of names of DV(s) = names of
IV(s).
For this analysis, the following commands were used:

1. proc glm;
2. class race;
3. model satisfaction_parenting satisfaction_
adopt_agency
= race;
4. means race / LSD alpha =.05;
5. manova h=race;
6. run;

Note that Step 4 calculates simultaneous confidence intervals.

SAS Commands to Perform a Roy–Bargmann Stepdown Analysis


1. proc glm;
2. class race;
3. model satisfaction_adopt_agency = race
satisfaction_parenting;
4. means race / LSD alpha =.05;
5. manova h=race satisfaction_parenting
race*satisfaction_parenting;
6. run;

REPORTING THE RESULTS OF A MANCOVA


1. Restate in summary form the reason(s) for the analysis, and the
basic components of the model(s) tested including the DV and
IVs;
82 Analysis of Multiple Dependent Variables

2. Describe how the assumptions underlying the model were


tested, and if not tenable treated. Include a discussion of missing
data, outliers, equality of variance–covariance matrices, and
correlation among correlation among DVs;
3. Report the results of at least one multivariate F-test;
4. Report a measure of effect size;
5. If the multivariate F-test is significant, describe how further the
statistical significance of differences on the DVs by group were
further explored; and
6. Summarize the results of model testing in terms of the study’s
hypotheses.

Results
This study tested a model that compared a client’s satisfaction with an
adoption agency’s services and their satisfaction with being an adoptive
parent by the client’s race (coded 1 = Caucasian, 2 = Asian, 3 = African
American), controlling for client self-efficacy. Satisfaction with agency
services, satisfaction with parenting, and client self-efficacy were opera-
tionalized as scale scores. Race was operationalized as clients’ self-reports
of their racial group.
A one-way multivariate analysis of covariance (MANCOVA) was
performed to determine the effect of being a member of three racial
groups on a client’s satisfaction with an adoption agency’s services and
their satisfaction with being an adoptive parent, controlling for client
self-efficacy.
No problems were noted for missing data, outliers, or multivari-
ate normality. The tenability of the absence of outliers assumption was
evaluated using Cook’s distance, D, which provides an overall measure
of the impact of an observation on the estimated MANCOVA model.
Observations with larger values of D than the rest of the data are those
which have unusual leverage.
For these data, five cases had values greater than .07 (.08–.11). If they
appear atypical, cases that contain values of D greater than 0.07 may be
deleted. See chapter 1 for additional discussion about the management of
outliers. Alternatively, MANCOVA results should be reported with and
without the outliers. Note that for these cases, MANCOVA results with
and without cases that had values on one or both DVS that exceeded
Multivariate Analysis of Covariance 83

.07 yielded equivalent results. Consequently, only results for all cases are
reported below. These results suggest that the assumption of absence of
outliers is tenable.
The tenability of the multivariate normality assumption was
evaluated by using an SPSS macro developed by DeCarlo (1997). All
tests of multivariate skew, multivariate kurtosis, and an omnibus test
of multivariate normality were not statistically significant at p = .05.
These results suggest that the assumption of multivariate normality is
tenable.
Box’s M was used to test the assumption (i.e., H0) of equality of vari-
ance–covariance matrices. Box’s M equaled 13.658, F(6, 64825) = 2.147,
p = .045, which means that equality of variance–covariance matrices can-
not be assumed.
Significant differences were found among the three categories of race
on the DVs, with Wilk’s Lambda = .760, F(4,98) = 3.600, p < .01. A Roy–
Bargmann stepdown analysis and simultaneous confidence intervals,
analysis were conducted to further explore these differences.
For these data, at a statistically significant level, controlling for cli-
ent self-efficacy, there are mean differences between racial groups on
level of satisfaction with the adoption agency (see Table 3.2). That is, at
a statistically significant level, controlling for client self-efficacy. African
American respondents report the highest level of satisfaction with the
adoption agency followed by Asian American and Caucasian American
respondents. Although these results are statistically significant, they

Table 3.2 Means and Standard Deviations for each Dependent Variable by Race
Caucasian 95% CI Asian 95% CI African 95% CI
N=18 N=18 American
Mean(SE) Mean(SE) N=18
Mean(SE)
Satisfaction 50.50 49.26– 49.82 48.95– 50.02 49.13–
w/Parenting (.43) 51.37 (.43) 50.69 (.44) 50.91
Satisfaction 49.13 48.35– 50.76 49.99– 51.11 50.31–
w/Adoption (.39)* 49.91 (.39)* 51.54 (.40)* 51.90
Agency
*Significant at p <.05
84 Analysis of Multiple Dependent Variables

may not be considered practically significant since group means on the


satisfaction with adoption agency scale ranged from 49.13 to 51.10.

ADDITIONAL EXAMPLES OF MANOVA FROM THE APPLIED RESEARCH LITERATURE

Jones, R., Yates, W. R., Williams, S., Zhou, M., & Hardman, L. (1999). Outcome
for adjustment disorder with depressed mood: Comparison with other mood
disorders. Journal of Affective Disorders, 55 (1), 55–61.
Retrospective data were used to evaluate the construct validity of the adjust-
ment disorder diagnostic category. The data primarily consisted of SF-36
Health Status Survey responses by a large group of adult psychiatric outpatients
before treatment and again six months after beginning treatment. Respondents
were divided into five diagnostic groups, and MANOVA, MANCOVA and chi
square were used to clarify relationships among diagnoses, sociodemographic
data and SF-36 scores.

Ashford, J. A. (2006). Child protective service relationships on parental attitudes


in the juvenile dependency process. Research on Social Work Practice, 16 (6),
582–590.
This pilot study used MANCOVA to examine the attitudes of parents in the
child dependency process to determine whether their perceptions of fairness,
trustworthiness, and satisfaction with the juvenile dependency system differed
across types of relationships: relationships with judges or child protective ser-
vice (CPS) workers. The study relied on a convenience sample of (N = 40)
parents who were administered structured interviews with Likert-type items
after being in relationships with the authorities for 6 months. The findings
supported the study’s hypotheses about the nature of the relationships in the
family drug court process and the relative contributions of relational versus
self-interest factors in explaining variations in parental attitudes.

Diana M. DiNitto, Deborah K. Webb and Allen Rubin. (2002). The Effectiveness of
an Integrated Treatment Approach for Clients With Dual Diagnoses, Research
on Social Work Practice, 12 (5), 621–641.
A MANCOVA tested the effectiveness of adding a psychoeducationally ori-
ented group therapy intervention, Good Chemistry Groups, to standard inpa-
tient chemical-dependency services for clients dually diagnosed with mental
and substance dependence disorders. Ninety-seven clients were randomly
assigned to an experimental group (n = 48) and a control group (n = 49).
Outcome variables included drug and alcohol use, participation in self-help
support group meetings, incarceration days, psychiatric symptoms, psychiatric
Multivariate Analysis of Covariance 85

inpatient admissions, compliance with prescribed psychotropic medication


plans, and composite scores on the Addiction Severity Index. No significant
treatment effects were found on any of the outcome variables.

Farooqi, A., Hägglöf, B., Sedin, G., Gothefors, L., & Serenius, F. (2007). Mental
health and social competencies of 10- to 12-year-old children born at 23 to
25 weeks of gestation in the 1990s: A Swedish national prospective follow-up
study. Pediatrics, 120, 118–133.
The study investigated a national cohort of extremely immature children with
respect to behavioral and emotional problems and social competencies, from
the perspectives of parents, teachers, and children themselves. MANCOVA of
parent-reported behavioral problems revealed no interactions, but significant
main effects emerged for group status (extremely immature versus control), fam-
ily function, social risk, and presence of a chronic medical condition, with all effect
sizes being medium and accounting for 8% to 12% of the variance. MANCOVA of
teacher-reported behavioral problems showed significant effects for group status
and gender but not for the covariates mentioned above. According to the teachers’
ratings, extremely immature children were less well adjusted to the school envi-
ronment than were control subjects. However, a majority of extremely immature
children (85%) were functioning in mainstream schools without major adjust-
ment problems. results. Reports from children showed a trend toward increased
depression symptoms compared with control subjects.

Painter, K. (2009). Multisystemic therapy as community-based treatment for


youth with severe emotional disturbance. Research on Social Work Practice,
19(3), 314–324.
Using MANCOVA, this study compares multisystemic therapy (MST) to
family skills training combined with case management in community mental
health for emotionally disturbed youth. Youth who received MST experienced
more improved mental health symptoms, less juvenile justice involvement,
and improvement across the linear combination of school functioning, family
functioning, mental health symptoms, substance abuse, risk of self-harm, and
disruptive or aggressive behavior than did youth who received usual services.
Both groups experienced significant improvement in youth functioning.

Pomeroy, E. C., Kiam, R., & Abel, E. M. (1999). The effectiveness of a psycho-
educational group for HIV-infected/affected incarcerated women. Research on
Social Work Practice, 9 (2), 171–187.
This study evaluated the effectiveness of a psychoeducational group inter-
vention for HIV/AIDS-infected and affected women at a large southeastern
county jail facility. A MANCOVA yielded significant differences between the
experimental and comparison groups. Subsequent analysis of covariance for
86 Analysis of Multiple Dependent Variables

each dependent variable indicated significant differences between groups as


well. Effect sizes ranged from moderate to strong. The psychoeducational
group intervention appeared to be effective in alleviating depression, anxi-
ety, and trauma symptoms among women inmates infected and affected by
HIV/AIDS.

Rubin, A., Bischofshausen, S., Conroy-Moore, K., Dennis, K., Hastie, M., Melnick,
L., Reeves, D., & Smith, T. (2001). The effectiveness of EMDR in a child guid-
ance center. Research on Social Work Practice, 11 (4), 435–457.
This study evaluated the effectiveness of adding Eye Movement
Desensitization and Reprocessing (EMDR) to the routine treatment regi-
men of child therapists. MANCOVA found no significant differences in Child
Behavior Checklist scores between groups. Subanalyses conducted for 33 cli-
ents with elevated pretest scores found moderate effect sizes that approached,
but fell short of, statistical significance. These findings raise doubts about
notions that EMDR produces rapid and dramatic improvements with children
whose emotional and behavioral problems are not narrowly connected to a
specific trauma and who require improvisational deviations from the standard
EMDR protocol. Further research is needed in light of the special difficulties
connected to implementing the EMDR protocol with clients like those in this
study.
4

Multivariate Multiple
Regression

OVERVIEW AND KEY TERMS


Less frequently termed canonical regression, multivariate multiple
regression (MMR) is used to model the linear relationship between more
than one IV and more than one DV. MMR is multiple because there is
more than one IV. MMR is multivariate because there is more than one
DV. MMR was developed by Bartlett (1938) as an extension of Hotelling’s
(1935, 1936) canonical correlation analysis. Although the term canoni-
cal regression did not appear in this early literature, it was used later by
Tintner (1950) and Bartlett (1951) in reference to Bartlett (1938). MMR
is a logical extension of OLS regression. MMR estimates the same regres-
sion coefficients and standard errors that would be obtained by using
separate OLS regression equations for each DV. However, MMR also may
be used to test an omnibus H0 and composite hypotheses for a model.
The omnibus null hypothesis is that all regression coefficients are equal
zero across all DVs. Tests of individual IVs or subsets of IVs across DVs
are termed composite hypotheses (Cramer & Nicewander, 1979). These
composite hypotheses may be also viewed as sequential or hierarchical
multiple regressions. As used here the term hierarchical multiple regres-
sion (not to be confused with hierarchical linear models) is similar to

87
88 Analysis of Multiple Dependent Variables

stepwise regression, but a researcher, not the computer, determines the


order of entry of variables, or sets of variables.
This chapter begins with an introduction to building and refin-
ing linear regression models. The remaining discussion focuses on the
MMR procedure, and is organized as follows: (1) estimating multivariate
regression parameters; (2) testing the omnibus hypothesis; (3) assessing
overall model fit; (4) testing composite hypotheses; (5) model validation;
(6) sample size requirements; (7) strengths and limitations of MMR;
(8) annotated example; (9) reporting the results of a MMR analysis;
(10) results of the annotated example; and (11) additional examples from
the applied research literature. MMR will be demonstrated with Stata.
References to resources for users of SPSS and SAS also are provided.
As described in chapter 1, model building concerns strategies for
selecting an optimal set of IVs that explain the variance in one or more
DVs (Schuster, 1998). There are three basic types of OLS regression mod-
els: fixed, random, and mixed effects (Winer, 1971). A fixed effect is one
that is assumed to be measured without error. It is also assumed that the
values of a fixed effect in one study are the same as the values of the fixed
variable in another study. A random effect is assumed to be a value drawn
from a larger population of values. The values of random effects repre-
sent a random sample of all possible values of that effect. Consequently,
the results obtained with a random effect can be generalized to all other
possible values of that random variable. Because they involve infer-
ence, random effects models are less powerful. Random effects models
are sometimes referred to as Model II or variance component models.
Analyses using both fixed and random effects are called mixed effects
models (Rencher, 1998). Discussion in this chapter focuses on the testing
of fixed effects models.

Model Refinement: Dependent Variables


After identifying theoretically important variables, and before testing
hypotheses about a model, a researcher may wish to maximize parsimony
by empirically reducing the number of DVs in the model.
Stepdown analysis (SDA). Introduced by Roy and Bargmann (1958),
with this procedure, the DVs in a model are examined sequentially. SDA
is recommended by Stevens (2009) and Tabachnick and Fidell (2007) as
a follow-up procedure for multivariate analysis of variance (MANOVA).
Multivariate Multiple Regression 89

However, SDA may also be appropriately used as a follow-up to an MMR,


if, based on a statistically significant multivariate F-test, the omnibus H0
is rejected. See chapter 2 for a detailed discussion of SDA.

Model Refinement: Independent Variables


Having identified theoretically important variables, and before proceed-
ing to test hypotheses, a researcher may wish to maximize parsimony by
empirically reducing the number of IVs in the model.
Stepwise procedures. These strategies involve identifying regression
models in stages. Stepwise procedures were developed before personal
computers when time on mainframe computers was at a premium, and
when researchers were considering the problem of what to do when
there may be more predictors than observations (Cohen & Cohen, 1983).
Stepwise procedures include forward selection, backwards elimination,
and stepwise regression. These approaches add or remove variables
one-at-a-time until some stopping rule is satisfied. Forward selection
starts with an empty model. Backward elimination starts with all of the
predictors in the model. Stepwise regression is similar to forward selec-
tion except that variables are removed from the model if they become
nonsignificant as other predictors are added.
More specifically, in stage one of a forward selection procedure, the
IV best correlated with the DV is included in the equation. In the second
stage, the IV with the highest partial correlation with the DV, controlling
for the first IV, is entered. This process is repeated, while at each stage
controlling for previously entered IVs, until the addition of a remaining
IV does not increase R2 by a statistically significant amount, or until all
IVs are entered.
Backward elimination begins with all IVs and removes IVs one at
a time until the elimination of one makes a significant difference in R2.
Backward elimination has an advantage over forward selection and step-
wise regression because it is possible for a set of variables to have con-
siderable predictive capability even though any subset of them does not.
Forward selection and stepwise regression will fail to identify these vari-
able sets. This is because in forward selection or stepwise regression if IVs
do not predict well individually, they will not be entered into a model.
Backward elimination starts with all IVs in the model, and, consequently,
joint predictive capability may be measured.
90 Analysis of Multiple Dependent Variables

All-Possible-Regressions. This procedure moves beyond stepwise


regression and tests all possible subsets of the set of IVs. If there are k
potential independent variables, then there are 2k − 1 distinct subsets of
variables to be tested. For example, if you have 10 potential independent
variables, the number of subsets to be tested is 210 − 1 = 1023, and if you
have 20 potential independent variables, the number is 220, or more than
one million. A variation of all possible regressions is best-subsets regres-
sion in which a researcher can limit the search based on the number and
type of variables to be modeled. For example, Minitab’s (www.Minitab.
com) best-subsets procedure identifies the best-fitting regression models
that can be constructed with the IVs specified by a researcher. Minitab
examines all possible subsets of the predictors, beginning with all mod-
els containing one predictor, and then all models containing two predic-
tors, and so on. By default, Minitab displays the two best models for each
number of predictors. For example, for a regression model that contains
three IVs, Minitab can report the best and second best one-IV models,
followed by the best and second best two-IV models, followed by the full
model containing all three IVs.
The two most commonly used criteria for ranking all-possible-sub-
sets and best subsets models are adjusted R2 and the Mallows Cp statistic
(Hocking, 1976; Mallows, 1973). The latter statistic is related to adjusted
R2, but includes a heavier “penalty” for increasing the number of IVs.
Defined as

RSS(p )
Mallows C p = − (n p) (4.1)
s2
where n is the sample size, p is the number of IVs plus the intercept,
RSS(p) is the residual sum-of -squares from a model containing p param-
eters, and s2 is the mean residual sum-of-squares from the model con-
taining p parameters. Recall from chapter 2, that residual sum-of-squares
(also termed SSW, sum-of-squares-within, and sum-of-squares error) is
a measure of the variability within respective groups around that group’s
mean for the same variable (e.g., age).
The values of Cp are typically positive and greater than one, where
lower values are better. Models that yield the best (lowest) values of Cp
will tend to be similar to those that yield the best (highest) values of
adjusted R2, but the exact ranking may be slightly different. Compared
Multivariate Multiple Regression 91

to adjusted R2, the Cp criterion tends to favor models with fewer param-
eters, so it is perhaps more robust to model overfitting. Generally, plots
of R2 and Cp versus the number of variables are examined to determine
an optimal model.
SPSS does not compute Mallow’s Cp directly from its menu system.
However, a syntax file that can be used to calculate Cp may be downloaded
from http://www.spsstools.net/Syntax/RegressionRepeatedMeasure/
DoAll-SubsetsRegressions.txt. For Stata users, rsquare is a module that
will calculate Mallow’s Cp. The rsquare module can be downloaded from
within Stata by typing findit rsquare. For SAS users, the SAS command
file, model_selection.sas, which can be downloaded from http://www-
personal.umich.edu/~kwelch/workshops/regression/sas/model_
selection.sas calculates Cp.
When all-possible-subsets and best subsets models are reported, it
is essential to describe how the model was derived. It is impossible to
determine from the numerical results whether a set of IVs was specified
before data collection or was obtained by using a selection procedure for
finding the “best” model. Parameter estimates and ANOVA tables do not
change to reflect which variable selection procedure was used. Results are
the same as what would have been obtained if that set of IVs had been
specified in advance.
An increasing number of applied researchers believe that computer-
controlled model-enhancement procedures, such as stepwise regression
and all-possible-subsets regression, are most appropriate in exploratory
research. For theory testing, a researcher should base selection of vari-
ables and their order of entry into a model on theory, not on a computer
algorithm. Menard (1995), for example, writes,

There appears to be general agreement that the use of computer-con-


trolled stepwise procedures to select variables is inappropriate for theory
testing because it capitalizes on random variations in the data and pro-
duces results that tend to be idiosyncratic and difficult to replicate in any
sample other than the sample in which they were originally obtained.
(p. 54)

At each stage in the stepwise process, a computer program must


fit a multiple regression model to the variables in the model to obtain
an F-to-remove value, and the program must fit a separate regression
92 Analysis of Multiple Dependent Variables

model for each of the variables not in the model in order to obtain their
F-to-enter values. It should be noted that statistical software packages,
such as SPSS, do not fit all models from scratch. Instead, the stepwise
search process, for example, is performed by a sequence of transforma-
tions to the correlation matrix of the variables in the model. That is,
variables are only read in once, and the sequence of adding or removing
variables and recalculating the F-statistics requires an updating opera-
tion on the correlation matrix, which is called “sweeping.”
The nominal significance level (e.g., .05) used at each step in stepwise
regression is subject to inflation, such that by the last step, the true signif-
icance level may be much higher, increasing the chances of type I errors.
That is, when many IVs are considered, and there is nothing theoretically
compelling about any of them before the data are collected, probability
theory suggests that at least one IV will achieve statistical significance
(Draper, Guttman, & Lapczak, 1979). Therefore, as more tests are per-
formed, the probability that one or more achieve statistical significance
because of chance (type I error) increases. This phenomenon (discussed
in chapter 2), sometimes termed probability pyramiding or inflated alpha,
explains why IVs that are theoretically unimportant sometimes achieve
statistical significance. Stepwise regression, therefore, usually results in
measures (F-test, t-tests, R2, standard error of the estimate, prediction
intervals) that are biased toward too much strength in the relationship
between the DV and the IVs. It is incorrect to call them statistically sig-
nificant because the reported values do not accurately reflect of the selec-
tion procedure.
Also troublesome is when there are missing data. Stepwise procedures
must exclude observations that are missing for any of the potential IVs.
Sometimes one or more of the IVs in the final model are no longer sta-
tistically significant when the model is fitted to the data set that includes
missing observations that had been deleted, even when these values are
missing at random.
Perhaps, the fundamental problem with computer-controlled meth-
ods is that they often substitute for thinking about a problem. Statistical
computing packages available today do our arithmetic for us in a way
that was totally unthinkable thirty years ago. When solving regres-
sion equations with many variables could take weeks using a desk cal-
culator, researchers were understandably reluctant to embark on the
Multivariate Multiple Regression 93

computational tasks without considerable thought (Searle, 1987). Today,


however, with the increasing availability of affordable user-friendly soft-
ware and powerful personal computers, minimal statistical knowledge is
needed for generating what can be voluminous and sophisticated arith-
metic. However, this minimal knowledge probably provides an inad-
equate perspective on what the output of statistical software packages,
such as SPSS, means, and how to use it.
Hierarchical multiple regression. What is needed is a strategy that
can distinguish between IVs that are better predictors than those that
appear to be better predictors because of chance alone. One alterna-
tive to computer-driven model testing procedures for independent
variables, such as stepwise regression and all-possible-regressions, is
hierarchical multiple regression (HMR). In HMR a researcher, not
the software, determines the order of entry of the variables (Bryk,
Raudenbush, Seltzer, & Congdon, 1988). F-tests are used to compute
the significance of each added variable (or set of variables) to the expla-
nation reflected in R2. This hierarchical procedure is an alternative to
comparing betas for purposes of assessing the importance of the inde-
pendents. In more complex forms of hierarchical regression, the model
can involve a series of intermediate variables that are dependents with
respect to some other independents, but are themselves independents
with respect to the ultimate dependent variable. Hierarchical multiple
regression, then, can involve a series of regressions for each intermedi-
ate as well as for the ultimate dependent variable. As discussed above,
in MMR, tests of individual IVs or subsets of IVs across DVs are termed
composite hypotheses (Cramer & Nicewander, 1979). These composite
hypotheses may be also viewed as sequential or hierarchical multiple
regressions.

THE MULTIVARIATE MULTIPLE REGRESSION PROCEDURE


Estimating Multivariate Regression Parameters
This process is analogous to estimating parameters in OLS regression.
Specifically, OLS regression is used to predict the variance in a DV based
on linear combinations of IVs. OLS regression also can determine the
relative predictive importance of the IVs in a regression model. Power
94 Analysis of Multiple Dependent Variables

terms can be added as IVs to measure curvilinear relationships with the


DV. Cross-product terms can be added as IVs to measure interaction
effects. The estimated values of the IVs (and the value of the y-intercept)
can be used to construct a prediction equation.
OLS regression seeks to find the line that best predicts the DV from
one or more IVs. This line is sometimes referred to as the line-of-best-fit.
That is, the goal of regression is to minimize the sum of the squares of
the vertical distances of the points from the line-of-best-fit. The line that
best predicts the DV from an IV is defined in terms of a slope for each
IV and a y-intercept. The slope quantifies the gradient of the line, and
equals the change in the DV for each unit change in an IV. If the slope
is positive, the DV increases as an IV increases. If the slope is negative,
the DV decreases as an IV increases. The y-intercept is the value of the
line when all IVs equal zero. It defines the elevation of the line on the
y-axis.
The OLS regression equation for a sample takes the form y = b1x1 +
b2x2 + ⋅ ⋅ ⋅ + bnxn + c. The b’s are the regression coefficients or slopes. The c
is the y-intercept. The standardized versions of the b coefficients are the
beta weights. Within the context of regression analysis, standardization
refers to the practice of redefining regression equations in terms of stan-
dard deviation units. Standardized coefficients, then, are the estimates
resulting from an analysis performed on variables that have been stan-
dardized so that they have variances equal to one.
Advocates of standardized regression coefficients point out that the
coefficients are the same regardless of an independent variable’s underly-
ing scale. Changes of scale are trivial in one sense, for they do not affect
the underlying reality or the degree of fit of a linear model to data. Critics
of standardized regression coefficients, however, argue that standardiza-
tion may have an impact on the ability to evaluate the relative impor-
tance of different explanatory variables or the relative importance of a
given variable in two or more different populations. That is, there is no
reason why a change of one SD in one predictor should be equivalent
to a change of one SD in another predictor. Some variables are easier to
change (e.g., amount of time on the Internet); other variables are more
difficult to change (e.g., number of cigarettes smoked); still others are
impossible to change (e.g., age). See, for example, Gelman (2008) and
(Kim, 1981) for detailed discussions of the strengths and weaknesses of
standardized regression coefficients.
Multivariate Multiple Regression 95

Standardized regression coefficients can be calculated in two ways,


with both yielding equivalent solutions (Bring, 1994). One way is first to
standardize all variables,

xi xi
x i* = , i ,..., k ,
si
(4.2)
yi y
y =
*
,
sy

where x–i and y– are the means of each variable in the sample and Si and
Sy are the standardized variables. A second way to standardize regres-
sion coefficients is by multiplying them by the ratio between the stan-
dard deviation of the respective IV (Si) and the standard deviation of the
DV (Sy),

Bi βli (s i / s y ), (4.3)

where Βi is the standardized regression coefficient. Traditional sources


of detailed discussions multiple OLS regression include (Cohen, Cohen,
West, & Aiken, 2003; Keith, 2006).
In MMR, for each DV, there are as many regression coefficients as
there are IVs, plus one coefficient for the intercept. Therefore, if there
are two DVs and three IVs, there are 2 × (3 + 1) = 8 regression coef-
ficients. The estimated IVs, like their OLS regression counterparts, are
unbiased and have minimum variance. Finn (1974) provides a succinct
mathematical explanation of MMR. Briefly, in the fixed effects regres-
sion model, each DV in a sample of n observations may be expressed
as a linear function of a set of IVs plus a random error, ε. The number
of IVs (x) is denoted by q, and the βs are the regression coefficients as
follows:

y1 = β0 + β1x11 + β2x12 + · · · +βq x1q + ε1


y2 = β0 + β1x21 + β2x22 + · · · +βq x2q + ε2
.
. (4.4)
.
yn = β0 + β1xn1 + β2xn2 + · · · +βq xnq + εn
96 Analysis of Multiple Dependent Variables

Using matrix notation, the aforementioned models for the n observa-


tions can be written more concisely as follows:

⎛ y 1 ⎞ ⎛ 1 x 11 x 12 " x 1q ⎞ ⎛ β0 ⎞ ⎛ ε1 ⎞
⎜ y ⎟ ⎜1 x ⎜ ⎟
x 22 " x 2q ⎟ ⎜ β1 ⎟ ⎜ ε2 ⎟
⎜ 2⎟ = ⎜ 21
⎟⎜ ⎟ + (4.5)
⎜ # ⎟ ⎜# # # # ⎟⎜ # ⎟ ⎜ ⎟
⎜⎝ y ⎟⎠ ⎜ 1 x ⎟⎜ ⎟ ⎜ # ⎟
n ⎝ n1 x n2 " x nq ⎠ ⎝ βq ⎠ ⎜⎝ ε ⎟⎠
n

or as Y = XΒ + ε. In OLS regression, the single variable y1 has a unique


role in the equation. In MMR, all DVs in the vector Y are treated analo-
gously, and Y is the most predictable linear combination of the jointly
dependent variable. “Most predictable” may be defined in a least squares
sense as minimizing the sum of square residuals in proportion to the
variance of the linear combination of the DVS, or in the sense of maxi-
mizing the likelihood of the model among all linear combinations of the
DVs. See chapter 5 for a discussion of maximum likelihood estimation
(MLE).
MMR estimates the same coefficients and standard errors as would
be obtained by using separate OLS regressions. MMR examines each DV
separately in relation to a linear combination of all predictor variables
without imposing any structure across the several resulting regression
equations. The IVs are not transformed into mutually uncorrelated vari-
ates, and coefficient estimation is identical to estimating regression equa-
tions separately for each DV (Cramer & Nicewander, 1979). However,
as mentioned above, MMR also may be used to test an omnibus null
hypothesis and composite hypotheses for a model (discussed below); it is
these two capabilities of MMR that distinguish it from OLS regression.

Testing the Omnibus Null Hypothesis


The omnibus null hypothesis is that all regression coefficients equal zero
across all DVs. The purpose of the omnibus hypothesis test, then, is to
help to prevent inflating the study-wise alpha level. If separate tests are
performed for each DV, the probability of obtaining a false significant
value would increase in direct proportion to the number of DVs being
tested; that is, the power of the test decreases. To evaluate the omnibus null
hypothesis, multivariate F- tests are used, which include Wilk’s lambda,
Multivariate Multiple Regression 97

Hotelling’s trace, Pillai’s trace, and Roy’s largest root. See chapter 2 for a
detailed discussion of these multivariate F-tests.

Assessing Overall Model Fit


Goodness-of-fit in MMR is determined in a way that is identical to OLS
regression. In MMR, the value of R2 can be computed for each equation
separately to study the effectiveness of each relationship in accounting
for observed variation. However, MMR does not yield an overall associa-
tion measure. The R2 values (one for each DV) may be correlated, and,
consequently, may not indicate the unique variance contribution of each
equation to the total set of equations. That is, the R2 values for each DV
may not indicate the unique variance explained for that DV by its set of
IVs as a proportion of total variance explained for all DVs.
Put more technically, the Ri2 statistics are conventional multiple cor-
relation coefficients and reflect assessments of predictive ability (Cramer
& Nicewander, 1979). Although the mean of all Ri2 values is equal to the
total redundancy (Muller, 1981), the Ri2 values for individual regression
equations are not analogs of the proportions of variance of the multiple
DVs explained by individual redundancy predictor variates, Ry|pi2, nor are
they analogs of specific canonical correlations squared, Rci2. Canonical
correlation analysis (CCA) seeks to explain the relationship between two
or more sets of variables. More specifically, in CCA, the objective is to
find linear combinations of the variables within each set such that the two
linear combinations of variables have maximum correlation. The several
Ri2 values, which may be found on the diagonal of the matrix product
RyxRxx−1Rxy, are not hierarchically ordered, nor are the values of the DVs
predicted by the separate regression equations necessarily uncorrelated.

Testing Composite Hypotheses


If, based on a statistically significant multivariate F-test, the omni-
bus null hypothesis is rejected, MMR may be used to test IVs across
different models (e.g., the DVs). Tests of IVs or subsets of IVs across
DVs are termed composite hypotheses. That is, MMR also estimates
the between-equation covariances. Anderson (1999; 2002) describes
a methodology for calculating asymptotic variances and covariances
for all coefficient estimates in standard canonical correlation analysis
98 Analysis of Multiple Dependent Variables

under the assumption that the data are generated by a structural linear
model with normally distributed variables and disturbances. According
to Anderson (1999; 2002), if these conditions hold, the coefficients of
the first canonical pair correspond to those in a MMR, and the asymp-
totic distribution of the sample canonical correlations and coefficients
of the canonical variates may be used for statistical inference about the
coefficients.

Model Validation
In the last stage of MMR, the model should be validated. If sample size
permits, one approach to validation is sample splitting, which involves
creating two subsamples of the data and performing an MMR analysis
on each subsample. Then, the results can be compared. Differences in
results between subsamples suggest that these results may not generalize
to the population.

Sample Size Requirements


There is little discussion about the specific sample-size requirements
of MMR. However, it is generally agreed that MMR is a special case of
canonical correlation analysis (CCA). Stevens (1996) provides a thor-
ough discussion of the sample size for CCA. To estimate the canonical
loadings, only for the most important canonical function, Stevens rec-
ommends a sample size at least 20 times the number of variables in the
analysis. To arrive at reliable estimates for two canonical functions, a
sample size of at least 40 – 60 times the number of variables in the analy-
sis is recommended.
Another perspective on estimating sample size for CCA is provided
by Barcikowski and Stevens (1975). These authors suggest that CCA may
detect stronger canonical correlations (e.g., R > .7), even with relatively
small samples (e.g., n = 50). Weaker canonical correlations (e.g., R = .3)
require larger sample sizes (n > 200) to be detected. Researchers should
consider combining both perspectives to triangular on a minimally suf-
ficient sample size for CCA. That is, they should consider the number
of canonical functions to be interpreted, and the relative strength of
the canonical loadings of the variables represented by the functions of
interest.
Multivariate Multiple Regression 99

STRENGTHS AND LIMITATIONS OF MMR


MMR has the following analytical strengths:

1. Efficiency in computation; and


2. Familiarity in interpretation.
MMR has the following analytical limitations:
1. Causality cannot be established with regression models alone;
2. Results assume perfect model specification in terms of all IVs and
their functional form (e.g., product term);
3. Sensitivity to outliers, heteroscedasticity of residuals,
multicollonearity among IVs, and nonlinearity of relationships
between the IVs and the DV.

ANNOTATED EXAMPLE
A study is conducted to test a model that predicts post-adoption service
utilization and positive adoption. Specifically, the study tests a model that
includes (1) factors influencing the utilization of post-adoption services
(parents’ perceptions of self-efficacy, relationship satisfaction between par-
ents, and attitudes toward adoption) as IVs and (2) service utilization and
positive adoption outcomes (satisfaction with parenting and satisfaction
with adoption agency) as DV. All variables were operationalized as scale
scores. The researcher decides to perform a MMR analysis to investigate
further the relationship between the following (1) IVs: parents’ percep-
tions of self-efficacy, relationship satisfaction between parents, and attitudes
toward adoption, and (2) the following DVs: service utilization, and satis-
faction with parenting. The next section presents Stata commands, which
are numbered in sequence and in Courier. These Stata commands use
the format DVs = IVs.

1. manova service_util_i satisfaction_


parenting = self_efficacy relationship_sat
attitude_adoption,continuous(self_efficacy
relationship_sat attitude_adoption)

Figure 4.1 illustrates the results of the manova command. The overall
F tests the null hypothesis that regression coefficients for all IVs equal zero
for all DVs. The multivariate F is based on the sum of squares between
100 Analysis of Multiple Dependent Variables

Number of obs = 300

W = Wilks’ lambda L = Lawley-Hotelling trace


P = Pillai’s trace R = Roy’s largest root
Source Statistic df F(df1, df2) = F Prob>F
Model W 0.2372 3 6.0 590.0 103.57 0.0000 e
P 0.9788 6.0 592.0 94.57 0.0000 a
L 2.3054 6.0 588.0 112.96 0.0000 a
R 1.7993 3.0 296.0 177.53 0.0000 u
Residual 296

self_effi~y W 0.9155 1 2.0 295.0 13.62 0.0000 e


P 0.0845 2.0 295.0 13.62 0.0000 e
L 0.0924 2.0 295.0 13.62 0.0000 e
R 0.0924 2.0 295.0 13.62 0.0000 e
relations~t W 0.8343 1 2.0 295.0 29.29 0.0000 e
P 0.1657 2.0 295.0 29.29 0.0000 e
L 0.1986 2.0 295.0 29.29 0.0000 e
R 0.1986 2.0 295.0 29.29 0.0000 e
attitude_~n W 0.3593 1 2.0 295.0 263.02 0.0000 e
P 0.6407 2.0 295.0 263.02 0.0000 e
L 1.7832 2.0 295.0 263.02 0.0000 e
R 1.7832 2.0 295.0 263.02 0.0000 e
Residual 296

Total 299
e = exact, a = approximate, u = upper bound on F

Figure 4.1 Stata MANOVA Output.

and within groups, and on the sum of crossproducts; that is, it considers
correlations between the criterion variables (see chapter two for a more
detailed discussion).
The F- and p-values for all four tests under the section labeled
“Model,” Wilk’s lambda, Lawley–Hotelling trace, Pillai’s trace, and Roy’s
largest root, are statistically significant (p < .001). Because the overall
multivariate tests are significant, it is concluded that there are differences
among the DVs as a function of one or more IVs.

2. mvreg service_util_i satisfaction_


parenting = self_efficacy relationship_sat
attitude_adoption

Figure 4.2 illustrates the results of the mvreg command, which first pro-
vides the multiple R2 values for each DV (equation), with associated F- and
p-values. Second, the output provides unstandardized regression coeffi-
cients, standard errors, t-values, p-values, and 95% confidence intervals
Multivariate Multiple Regression 101

for each IV in each model. In Stata mvreg is the command used for MMR
estimates. The output from the mvreg command is similar to the output
from the regress command, except that there are three equations (one
for each DV) instead of one. The coefficients (and all of the output) are
interpreted in the same way as they are for any OLS regression. To be clear,
the “R-sq” in Figure 4.2 corresponds to the R2 from the regress com-
mand, not the adjusted R2. If regressions were performed for each out-
come variable, the same coefficients, standard errors, t- and p-values, and
confidence intervals as shown above would be obtained.
MMR does not provide an overall association measure. The Ri2
statistics are conventional multiple correlation coefficients and reflect
assessments of predictive ability. The Ri2 values predicted by the sepa-
rate regression equations necessarily uncorrelated. That is, the R2 values
for each DV may not indicate the unique variance explained for those
DVs by its set of IVs as a proportion of total variance explained for
all DVs.
As mentioned, if a separate regression was run for each DV, the same
coefficients, standard errors, t- and p-values, and confidence intervals as
shown above would be obtained. The use of the test command is one of
the compelling reasons for conducting a multivariate regression analysis.
One advantage of using the mvreg command is that tests of coefficients
(IVs) across the DVs may be run. Accordingly, the researcher tests the null
hypothesis that the coefficients for the IVs test self_efficacy, relationship_sat,
and attitude_adoption equal 0 in the equations for each of the two DVs.

Equation obs Parms RMSE “R-sq” F P

service_ut~n 300 4 .4099478 0.3366 50.06577 0.0000


satisfacti~g 300 4 2.301917 0.6423 177.1453 0.0000

Coef. Std. Err. t P>| t | [95% Conf. Interval]


service_ut~n
self_effic~y –.0252346 .0049885 –5.06 0.000 –.035052 –.0154172
relationsh~t –.0419408 .0054803 –7.65 0.000 –.052726 –.0311556
attitude)_a~n –.0028436 .0048637 –0.58 0.559 –.0124154 .0067281
_cons 3.996216 .3564754 11.21 0.000 3.294669 4.697764

satisfacti~g
self_effic~y –.0402116 .0280111 –1.44 0.152 –.0953378 .0149146
relationsh~t –.0193035 .0307725 –0.63 0.531 –.0798641 .041257
attitude)_a~n .6267131 .0273102 22.95 0.000 .5729664 .6804598
_cons 21.56306 2.001662 10.77 0.000 17.62377 25.50235

Figure 4.2 Results of the Stata’s mvreg Command.


102 Analysis of Multiple Dependent Variables

3. test self_efficacy relationship_sat


attitude_adoption (Please see Figure 4.3)
(1) [service_utilization]self_efficacy = 0
(2) [satisfaction_parenting]self_efficacy = 0
(3) [service_utilization]relationship_sat = 0
(4) [satisfaction_parenting]relationship_sat = 0
(5) [service_utilization]attitude_adoption = 0
(6) [satisfaction_parenting]attitude_adoption = 0
F(6, 296) = 113.73
Prob > F = 0.0000
Figure 4.3 Results of the Stata’s test Command.

4. As a follow-up to the test command, the researcher calculated


standardized regression coefficients (i.e., Beta) to facilitate a
comparison of the relative contribution of IVs in the two models
(one for each DV).

regress service_util_i self_efficacy


relationship_sat attitude_adoption, beta
regress satisfaction_parenting self_efficacy
relationship_sat attitude_adoption, beta

Figures 4.4 and 4.5 illustrate the results of these two analyses. In
Figure 4.4 for the DV service utilization, the three variables seem to make
equal contributions. In Figure 4.5 for the DV satisfaction with parenting,
the variable attitude toward adoption seems to make the greatest relative
contribution (Beta = 0.8011).

SPSS Commands to Perform a MMR


GLM service_utilization satisfaction_
parenting BY self_efficacy relationship_sat
attitude_adoption

service_ut~i Beta

self_effic~y –.0530787
relationsh~t –.0470319
attitude_a~n .0221673
_cons .
Figure 4.4 DV is service utilization: Standardized Regression Coefficients.
Multivariate Multiple Regression 103

satisfacti~g Beta

self_effic~y –.0557719
relationsh~t –.0244611
attitude_a~n .8011388
_cons .
Figure 4.5 DV is satisfaction with parenting: Standardized Regression Coefficients.

/METHOD=SSTYPE(3)
/INTERCEPT=INCLUDE
/CRITERIA=ALPHA(.05)
/DESIGN= self_efficacy relationship_sat
attitude_adoption

SAS Commands to Perform a MMR


proc reg data = “g:\SAS\hsb2”;
model read socst = write math science;
mtest / details print;
run;
quit;
model DV1 DV2 = IV1 + IV2 + IV3

REPORTING THE RESULTS OF A MMR ANALYSIS


1. Restate in summary form the reason(s) for the analysis, and the
basic components of the model(s) tested including the DV and IVs;
2. Report how the assumptions underlying the model were tested.
Describe missing data, and whether it is plausible to assume
that they are missing at random. Describe and discuss how
(a) outliers were identified and treated; (b) multicollinearity was
identified and treated; (c) linearity was identified and treated;
and (d) heteroscedasticity was identified and treated;
3. Report the adjusted R2 for the whole model. Recall that with a
large sample, IVs in a model may be statistically significant, but
only explain a low proportion of the variability of the DV;
4. Report the regression coefficients, standard errors, and confidence
intervals for the IVs. Although PASW will give a negative t-value
if the corresponding regression coefficient is negative, you should
104 Analysis of Multiple Dependent Variables

drop the negative sign when reporting the results. Degrees of


freedom (df) for both F and t values must be given. Usually dfs
are written as subscripts. For F the numerator degrees of freedom
are given first. Dfs may also be place in parentheses, or reported
explicitly, e.g., F(3,12) = 4.32 or F = 4.32, df = 3, 12. Significance
levels can either be reported exactly (e.g., p = .032), or in terms of
conventional levels (e.g. p < 0.05). There are arguments in favor
of either approach, but, at a minimum, presentation should be
reported in a consistent way. Beware of highly significant F or
t values, whose significance levels will be reported by PASW as,
for example, 0.0000. It is an act of statistical illiteracy to write
p = 0.0000; significance levels can never be zero; distributions
such as t, F, and chi-square are asymptotic, and, consequently, the
probability observing any value for these distributions, assuming
the null hypothesis is true, is never equal to zero. Therefore,
significance levels should be presented, for example, as p < .00005.
F and t values are conventionally reported to two decimal places,
and R2adj values to the nearest percentage point (sometimes to one
additional decimal place). For coefficients, you should be guided
by the sample size: for samples in the range of 100 to 1000, three
significant figures usually are sufficient; and
5. Summarize the results of model testing in terms of the study’s
hypotheses.

RESULTS OF THE ANNOTATED EXAMPLE


A multivariate multiple regression analysis was conducted to test a model
that predicts post-adoption service utilization and positive adoption out-
comes. Specifically, the model includes (1) factors influencing the uti-
lization of post-adoption services (parents’ perceptions of self-efficacy,
relationship satisfaction between parents, and attitudes toward adoption)
as independent variables and (2) service utilization and positive adop-
tion outcomes (satisfaction with parenting and satisfaction with adoption
agency) as dependent variables.
Wilk’s lambda, which was used to test the omnibus hypothesis
that all beta coefficients across all DVs equal zero, was statistically sig-
nificant: F(6, 590) = 103.57, p < .001. Consequently, it was concluded
Multivariate Multiple Regression 105

that that one or more independent variables are statistically significant


predictors of one or more dependent variables. To further explore the
relationships between the independent and each dependent variable,
each dependent variable was regressed on all three independent vari-
ables. For the model with service utilization as the dependent variable,
R2 = 0.3366, F = 50.0677, p < .01. For the model with satisfaction with
parenting as the dependent variable, R2 = 0.6423, F = 177.1453, p < .01.
These results suggest that the three independent variables are better pre-
dictors (i.e., explain more variance) of satisfaction with parenting than of
service utilization.
To facilitate a comparison of the relative contribution of the IVs in
the two models, standardized regression coefficients were calculated. For
the DV service utilization, the three variables seem to make equal con-
tributions (see Figure 4.4). For the DV satisfaction with parenting, the
variable attitude toward adoption seems to make the greatest relative con-
tribution (Beta = 0.8011) (see Figure 4.5).

ADDITIONAL EXAMPLES FROM THE APPLIED RESEARCH LITERATURE


Barrera, M., & Garrison-Jones, C. (1992). Family and peer social support as spe-
cific correlates of adolescent depressive symptoms. Journal of Abnormal Child
Psychology, 20(1), 1–16.
Family and peer support were distinquished to determine if these sources
of support were differentially related to depression symptoms as measured by
the Child Assessment Schedule (K. Hodges et al; see record 1982-29404-001).
Step-down multivariate multiple-regression analyses showed that depres-
sion symptoms were uniquely predicted by social relationship variables after
accounting for the effects of anxiety and conduct disorder symptoms. Results
are consistent with the assertion that social supports significantly contribute to
the experience of depressive symptoms by adolescents.

Cargill, B. R., Emmons, K. M., Kahler, C.W., & Brown, R. A. (2001). Relationship
among alcohol use, depression, smoking behavior, and motivation to quit
smoking with hospitalized smokers. Psychology of Addictive Behaviors, 15(3),
272–275.
Relationships among depression, alcohol use, and motivation to quit smok-
ing were examined in a sample of hospitalized smokers. Multivariate multiple
regression and logistic regression analyses indicated that participants with
depressed mood were more likely to have a history of problematic drinking.
106 Analysis of Multiple Dependent Variables

Participants with depressed mood and a history of problematic drinking were


more likely to be nicotine dependent and anticipated greater difficulty refrain-
ing from smoking while hospitalized. Overall, depression and alcohol use had
stronger associations with smoking-related variables than with smoking cessa-
tion motivation variables.

Flores, L. Y., Navarro, R. L., & DeWitz, S. J. (2008). Mexican American high school
students’ postsecondary educational goals: Applying social cognitive career
theory. Journal of Career Assessment, 16(4), 489–501.
A multivariate multiple regression analysis predicting the educational
goal aspirations and expectations of Mexican American high school students
was examined based on Lent, Brown, and Hackett’s Social Cognitive Career
Theory and prior research findings with Mexican American samples. No gen-
der or generational status differences were found in educational aspirations
or expectations; however, participants reported higher educational aspirations
than educational expectations. In addition, results of a multivariate multiple
regression analysis suggested that Anglo-oriented acculturation was signifi-
cantly positively related to educational goal expectations and educational goal
aspirations. Mexican-oriented acculturation, college self-efficacy, and college
outcome expectations were not significantly related to Mexican American
students’ educational goals aspirations or expectations. Results are discussed
as they relate to improving the educational achievement among Mexican
American youth.

Henderson, M. J., Saules, K. K., & Galen, L. W. (2004). The predictive validity
of the University of Rhode Island Change Assessment Questionnaire in a
heroin-addicted polysubstance abuse sample. Psychology of Addictive Behaviors,
18(2), 106–112.
The purpose of this investigation was to examine the predictive utility
of the stages-of-change scales of the University of Rhode Island Change
Assessment Questionnaire in a heroin-addicted polysubstance-abusing
treatment sample. Participants completed the URICA at the beginning of
a 29-week treatment period that required thrice-weekly urine drug screens.
Multivariate multiple regression analysis indicated that after controlling for
demographic variables, substance abuse severity, and treatment assignment,
the stages of change scales added significant variance to the prediction of
heroin- and cocaine-free urine samples. The Maintenance scale was posi-
tively related to cocaine-free urines and length in treatment. The implica-
tions of these findings for treatment and for measuring readiness among
individuals using multiple substances while taking maintenance medications
are discussed.
Multivariate Multiple Regression 107

Lynch, S. M., & Graham-Bermann, S. A. (2004). Exploring the relationship


between positive work experiences and women’s sense of self in the context of
partner abuse. Psychology of Women Quarterly, 28(2), 159–167.
This study examined the relationships among partner abuse, work quality,
and women’s sense of self. In particular, we explored the potential for women’s
work to serve as an alternative source of feedback for the self in the context
of partner abuse. The sample consisted of working women who reported
experiencing a range of partner abuse. Relationships among partner abuse,
work quality, and three self-constructs were tested using multivariate mul-
tiple regression. Work quality was significantly and positively associated with
self at work and general self-esteem and approached significance for self at
home. There were no significant associations between partner abuse and self
at work. Partner abuse was negatively and significantly associated with self at
home and approached significance for self-esteem. These varied results sup-
port the importance of assessing multiple aspects of the self and the potential
of women’s work to be a resource in the context of partner abuse.

Miville, M. L., & Constantine, M. G. (2006). Sociocultural predictors of psycho-


logical help-seeking attitudes and behavior among Mexican American college
students. Cultural Diversity and Ethnic Minority Psychology, 12(3), 420–432.
Sociocultural variables of acculturation, enculturation, cultural congru-
ity, and perceived social support were used as predictors of psychological
help-seeking attitudes and behaviors among 162 Mexican American college
students. Multivariate multiple regression analyses indicated that higher
cultural congruity, lower perceived social support from family, and higher
perceived social support from significant others were significant predictors
of positive help-seeking attitudes. In addition, higher acculturation into the
dominant society, lower perceived social support from family, and lower
perceived social support from friends were significantly predictive of greater
help-seeking behavior. Implications for research and practice are discussed.

Miville, M. L., Darlington, P., Whitlock, B., & Mulligan, T. (2005). Integrating
identities: The relationships of racial, gender, and ego identities among white
college students. Journal of College Student Development, 46(2), 157–175.
The authors proposed that racial and gender identities were related to ego
identities based on common themes that exist across these different dimen-
sions of identity. A sample of White college students completed the White
Racial Identity Attitude Scale, the Womanist Identity Attitude Scale or Men’s
Identity Attitude Scale, and the Extended Objective Measure of Ego Identity
Status. Multivariate multiple regression analyses revealed that all ego identity
statuses were significantly related to gender and/or racial identity statuses for
108 Analysis of Multiple Dependent Variables

both women and men. Implications for practice, limitations, and directions
for future research are discussed.

Reynolds, A. L., & Constantine, M. G. (2007). Cultural adjustment difficulties


and career development of international college students. Journal of Career
Assessment, 15(3), 338–350.
This study examined the extent to which two dimensions of cultural adjust-
ment difficulties (i.e., acculturative distress and intercultural competence
concerns) predicted two specific career development outcomes (i.e., career
aspirations and career outcome expectations) in a sample of international col-
lege students from Africa, Asia, and Latin America. Although no significant
differences among the participants were found by region of origin and gender,
multivariate multiple regression analyses indicated that higher levels of accul-
turative distress were predictive of lower levels of career outcome expecta-
tions among these international students. Furthermore, findings revealed that
greater intercultural competence concerns were predictive of lower career aspi-
rations and lower career outcome expectations. Implications of the findings
for career counseling with African, Asian, and Latin American international
students are discussed.

Tang, T., & Kim, J. K. (1999). The meaning of money among mental health
workers: The endorsement of money ethic as related to organizational citizen-
ship, behavior, job satisfaction, and commitment. Public Personnel Management,
28(1), 15–26.
Exploratory and confirmatory factor analyses were conducted to examine
the measurement and dimensions of the six-item Money Ethic Scale (MES)
using a sample of mental health workers. Results showed that the items of the
new MES had very low and negligible cross-loadings and the interfactor cor-
relations were small. Therefore, the three factors (Budget, Evil, and Success)
measured fairly independent constructs. In addition, the results of a multi-
variate multiple regression showed that the linear combination of the factors
Budget, Evil, and Success was a significant predictor of the linear combina-
tion of organizational citizenship behavior, job satisfaction, and organizational
commitment.
5

Structural Equation
Modeling

OVERVIEW AND KEY TERMS


Structural equation modeling (SEM), also referred to as causal modeling
and covariance structure analysis, is used to evaluate the consistency of
substantive theories with empirical data (cf. Bollen, 1989; Guo, Perron, &
Gillespie, 2009; Kaplan, 2000; Kline, 2011; Loehlin, 2004; Mulaik, 2009).
SEM is appropriate with experimental, nonexperimental, cross-sectional,
longitudinal, one-level (non-nested), multilevel (nested), linear, and
nonlinear data. SEM is a hybrid model that integrates path analysis and
factor analysis. SEM is related to factor analysis because it may be used to
test hypothesized relationships between unmeasured or latent variables
and observed or empirical indicators of latent variables. SEM is related
to path analysis because it may be used to test hypothesized relationships
between constructs. Thinking of SEM as a combination of factor analysis
and path analysis ensures consideration of SEM’s two primary compo-
nents: the measurement model and the structural model.
The measurement model describes the relationships between
observed variables and the construct or constructs those variables are
hypothesized to measure (Bollen, 1989). Confirmatory factor analysis
is used in testing the measurement model. A latent variable is defined

109
110 Analysis of Multiple Dependent Variables

more accurately to the extent that the measures that define it are strongly
related to one another. If, for example, one measure is only weakly cor-
related with two other measures of the same construct, then that con-
struct will be poorly defined. Ideally, each indicator is a separate measure
of the hypothesized latent variable. In SEM, measurement is recognized
as error-prone. By explicitly modeling measurement error, SEM seeks to
derive unbiased estimates for the relations between latent constructs.
Multi-item scales pose challenges for SEM if all the items are used
as indicators of a latent construct (Cattell, 1956). For instance, a model
could have too many parameters to estimate relative to the available
sample size, resulting in reduced power to detect important parameters.
In addition, it might not fit the data sufficiently well because individual
items may have less than ideal measurement properties, leading to the
rejection of a plausible model.
When describing strategies for incorporating multi-item scales into
SEM, it is useful to distinguish among factors or latent variables, items
or observed variables or items, and groups of items or parcels. Three
basic strategies for incorporating lengthy scales into SEM are as fol-
lows: (1) including all items individually and (2) combining items (e.g.,
summed or averaged) into one or more subsets or parcels. The practice
of parceling items within scales or subscales has received considerable
attention in the structural equation modeling literature (cf. Bandalos,
2002; Little, Cunningham, Shahar, & Widaman, 2002; MacCallum et al.,
1999; Nasser & Takahashi, 2003). Item parceling can reduce the dimen-
sionality and number of parameters estimated, resulting in more stable
parameter estimates and proper solutions of model fit. When items are
severely nonnormal or are coarsely categorized, research suggests item
parceling improves the normality and continuity of the indicators and
estimates of model fit are enhanced as compared to the original items
(Bandalos, 2002).
Item parceling’s potential psychometric benefits notwithstanding,
this strategy has been controversial. One concern is that parceling results
in a loss of information about the relative importance of individual items
(Marsh & O’Neill, 1984), because items are implicitly weighted equally
in parcels (Bollen &Lennox, 1991). Another concern was that parcel-
ing of ordinal scales results in indicators with undefined values, poten-
tially changing the original relations between the indicators and latent
variables, for instance, from nonlinear to linear relations (Coanders,
Structural Equation Modeling 111

Satorra, & Saris, 1997). Parceling binary or trichotomous items could


result in limited range as opposed to the latent trait scale, thereby biasing
variance and covariance parameters in SEM (Wright, 1999). Compared
with using individual items, parceling could underestimate the relations
of the latent variables if the reliability of the scale is low (Shevlin, Miles, &
Bunting, 1997).
Studies suggest that single item parcels work best when constructed on
unidimensional structures and for items having five response categories
(e.g., Bagozzi and Edwards;1998; Kishton & Widaman, 1994; Little et al.,
2002; Schau, Stevens, Dauphinee, & Del Vecchio, 1995) parceling is one
desirable option. Consequently, to facilitate this introduction to SEM, the
scales used in the simulated data annotated example below are assumed
to be one-dimensional with high Cronbach’s alphas. Consequently, the
approach used to incorporate relatively lengthy scales into this example’s
model is to use the sum of the scale as an indicator of a latent construct.
The structural model describes interrelationships among constructs.
Equations in the structural portion of the model specify the hypothesized
relationships among latent variables.
SEM utilizes two basic types of variables: exogenous and endog-
enous. Exogenous variables are analogous to independent variables.
In SEM terminology, other variables regress on exogenous variables.
Endogenous variables are analogous to dependent variables. A variable
that regresses on a variable is always an endogenous variable, even if this
same variable is also used as a variable to be regressed on at other points
in a model. A variable that is directly observed and measured is called a
manifest or indicator variable. A variable that is not directly measured
is a latent variable. The “factors” in a factor analysis are latent variables.
Relationships among exogenous variables (latent or manifest) may be
described as covariances. Covariances are analogous to correlations in
that they are defined as nondirectional relationships among independent
latent variables. Covariances are unstandardized correlations. That is,
covariances are to correlations as unstandardized regression coefficients
are to standardized regression coefficients.
Direct effects are relationships among measured and latent variables
(Please see Figure 5.1). Direct relationships are recursive or nonrecur-
sive. A recursive relationship is directional and consists of a single path
from one variable to another (i.e., X→Y in Figure 5.1). A nonrecursive
relationship is mutual or nondirectional.
112 Analysis of Multiple Dependent Variables

M
a b

c′
X Y
Figure 5.1 A Recursive Relationship.

Mediation in its simplest form represents the addition of a third vari-


able to this X→Y relationship so that it is hypothesized that X causes
the mediator, M, and M causes Y, so X→M→Y (Please see Figure 5.1).
Indirect effects (sometimes referred to as mediated effects) are the rela-
tionships between exogenous variables and endogenous variables that
are mediated by one or variables. Indirect effects are estimated statisti-
cally as the product of direct effects (i.e., ab in Figure 5.1). Mediation may
be full or partial. Indirect effects are estimated statistically as the product
of direct effects. Total effects are the sum of all direct and indirect effects
of one variable on another (i.e., a b + c in Figure 5.1).
A simple statistical model consists of exogenous variables whose rela-
tionships with a dependent variable are separate. That is, the effects of the
exogenous variables are additive, and, consequently, there is additivity,
or no interaction. A more complicated model is when the effect of one
exogenous variable X1 on a endogenous variable Y depends on another
exogenous variable X2 (i.e., moderator). The relationship between X1
and X2 with Y is termed moderation. Essentially, moderator and media-
tor relationships are the foundation of structural equation modeling.
When the measurement model and the structural model are con-
sidered together, the model is termed the composite or full structural
model. SEM is a largely confirmatory, rather than exploratory, technique.
That is, a researcher is more likely to use SEM to determine whether a
certain model is valid, rather than use SEM to discover a suitable model.
Consequently, because in SEM the researcher is attempting to develop
a theoretical model that accounts for all the covariances among the
measured items, a nonsignificant difference (accept or fail to reject H0)
between the proposed model and the saturated or perfect model is argued
to be suggestive of support for the proposed model. Various discrepancy
functions can be formed depending on the particular minimization algo-
rithm being used (e.g., maximum likelihood), but the goal remains the
same: to derive a test statistic that has a known distribution, and then
Structural Equation Modeling 113

compare the obtained value of the test statistic against tabled values in
order to make a decision about the null hypothesis.
A structural equation model implies a structure of the covariance
matrix of the measures (see chapter 2 for a discussion of covariance
matrices). Once the model’s parameters have been estimated, the result-
ing model-implied covariance matrix can then be compared to an empir-
ical or data-based covariance matrix. If the two matrices are consistent
with one another, then the structural equation model can be considered
a plausible explanation for relations between the measures.
From another perspective, assume a set of numbers X related to
another set of numbers Y by the equation Y = 4X, then the variance of
Y must be 16 times that of X, so you can test the hypothesis that Y and
X are related by the equation Y = 4X indirectly by comparing the vari-
ances of the Y and X variables. This idea generalizes, in various ways, to
several variables interrelated by a group of linear equations. The rules
become more complex, the calculations more difficult, but the basic mes-
sage remains the same: to test whether variables are interrelated through
a set of linear relationships by examining the variances and covariances
of the variables.
Model fit or goodness-of-fit may be assessed by examining the results
of the analysis, in particular the solution (i.e., parameter estimates, stan-
dard errors, correlations of parameter estimates, squared multiple cor-
relations, coefficients of determination), the overall fit (i.e., chi-square
based and non–chi-squared comparative fit indices), and the detailed
assessment of fit (i.e., standardized residuals and modification indices).
The remaining discussion is organized as follows: (1) assumptions,
(2) the procedure, (3) sample-size requirements, (4) strengths and limi-
tations, (5) annotated example, (6) reporting results, (7) results, and
(8) additional examples of SEM from the applied research literature.

ASSUMPTIONS OF SEM
For SEM, all of the assumptions for MANOVA apply, together with the
following extension of the assumption that the model is specified cor-
rectly. Because SEM is a confirmatory technique, full model must be
defined a priori. Within the context of SEM, this assumption, termed
identification, specifies the requirements of an appropriate model.
114 Analysis of Multiple Dependent Variables

As discussed above, thinking of SEM as a combination of factor analysis


and path analysis ensures consideration of SEM’s two primary compo-
nents: the measurement model and the structural model. Accordingly,
identification concerns whether the parameter of a model (i.e., a set of
equations) can be estimated. In SEM, both the structural and the mea-
surement models must be identified. These issues are analogous to the
GLM assumptions that a model is correctly specified and measurement
error has been minimized, respectively.

Structural Identification
Model identification is a complex topic and a comprehensive mathemati-
cal discussion is beyond the scope of this book. However, some insight
into identification is essential for researchers to competently perform
structural equation modeling. Essentially, the following discussion
focuses on the t-rule, one of several tests associated with identification.
Other tests associated with identification will be briefly described and
sources of more comprehensive discussion will be recommended. More
extensive discussions of model identification within the context of SEM
are provided by Bollen (1989) and Kline (2011).
A statistical model is structurally identified if the known informa-
tion available implies that there is one best value for each parameter in
the model whose value is not known. Structural models must be identi-
fied for the overall SEM to be identified. That is, a model is identified if
the unknown parameters in the model only are functions of identified
parameters and these functions lead to unique solutions (Bollen, 1989).
Models for which there are an infinite number of possible parameter
estimate values are said to be underidentified. For example, a theoretical
model suggests that X + Y = 10. One possible solution is that X = 5 and
Y = 5, another is that X = 2 and Y = 8, but there are many possible solu-
tions for this problem; that is, there is indeterminacy, or the possibility
that the data fit more than one implied theoretical model equally well. If
a model is underidentified, then it will remain under identified regardless
of sample size. Models that are not identified should be respecified.
Models in which there is only one possible solution for each param-
eter estimate are said to be just-identified. Finally, models that have more
than one possible solution (but one best or optimal solution) for each
parameter estimate are considered overidentified. Usually, overidentified
Structural Equation Modeling 115

F1

V1 V2

e1 e2

Figure 5.2 An SEM Model.

models are used in SEM because these models allow a researcher to test
statistical hypotheses (Loehlin, 2004).
Heuristics are available to help to determine whether a model is struc-
turally identified. One commonly used heuristic is the t-Rule. This rule
states that there must be more known pieces of information (i.e., inputs)
than unknown pieces of information (i.e., parameters to be estimated) to
calculate a unique solution. If this condition is not satisfied, the model is
not identified. If this condition is satisfied, the model may be identified.
Consider Figure 5.2
This model contains one factor, F1, two observed variables, V1 and V2,
and two error variances or residuals, e1 and e2. This model requires that
four parameters be estimated: the factor’s variance, the two error variances,
and one factor loading. To estimate the number of inputs available to esti-
mate the aforementioned four parameters, use the following formula:

[Q(Q + 1)] / 2

where Q represents the number of measured variables in the model. In this


model, there are two observed variables, V1 and V2, and [2(2 + 1)]/2 = 3.
It is not possible to estimate four unknown parameters from three inputs.
There are three available inputs, but there are four unknown parameters
to estimate, and overall, the model has 3 – 4 = −1 degrees of freedom. This
model is underidentified.
Now consider Figure 5.3.
116 Analysis of Multiple Dependent Variables

F1 F2

V1 V2 V3 V4

e1 e2 e3 e4

Figure 5.3 An Underidentified SEM Model

This second model has [4(4 + 1)] / 2 = 10 available degrees of free-


dom because there are four observed variables used in the model. This
model has one degree of freedom. This model is structurally identified.
In fact, it is overidentified because there is one positive degree of freedom
present. That is, 10 minus four error variances, two factor loadings, two
factor variances, and one covariance between the factors.
The t-Rule provides necessary, but not sufficient, conditions for
identification. That is, if the t-Rule test is not passed, the model is not
identified. If the test is passed, the model still may not be identified.
Other heuristics to help to determine whether a model is identified
include the Null B Rule, the Recursive Rule, the Order Condition Test,
and the Rank Condition Test. The Null B Rule applies to models in
which there is no relationship between the endogenous variables, which
is rare. The Null B Rule provides sufficient, but not necessary condi-
tions for identification. The Recursive Rule is sufficient but not neces-
sary for identification. To be recursive, there must not be any feedback
loops among endogenous variables. The Order Condition Test sug-
gests that the number of variables excluded from each equation must
be at least P − 1, where P is the number of equations. If equations
are related, then the model is underidentified. This is a necessary, but
not a sufficient condition to claim identification. The Rank Condition
test is both necessary and sufficient, making it one of the better tests
of identification. Passing this test means the model is appropriate for
further analyses.
Structural Equation Modeling 117

SEM software programs such as AMOS perform identification checks


as part of the model-fitting process. They usually provide warnings about
underidentification conditions. AMOS, for example, displays the follow-
ing message: The specified model is probably unidentified. In order to
achieve identifiability, it will probably be necessary to impose additional
constraints. The (probably) unidentified parameters are marked. The
method that AMOS uses for determining that a model is unidentified,
and for determining how many additional constraints are required to
make the model identified, is fallible. However, it is usually right. In order
to decide whether a parameter is identified, or whether an entire model
is identified, AMOS examines the rank of the matrix of approximate
second derivatives, and of some related matrices. There are objections
to this approach in principle (Bentler & Weeks, 1980; McDonald, 1982).
There are also practical problems in determining the rank of a matrix in
borderline cases. Because of these difficulties, identifiability of a model
should be judged a priori. With complex models, this may be impossible,
so a researcher will have to rely on AMOS’s numerical determination.

Empirical Identification
Measurement models must also be empirically identified for the overall
SEM to be identified (Kenny & Judd, 1986). A model in which at least
one parameter estimate is unstable is empirically underidentified. As
discussed above, in SEM, the measurement model describes the rela-
tionships between observed variables and the construct or constructs
those variables are hypothesized to measure. The measurement model
of SEM allows the researcher to evaluate how well his or her observed
(measured) variables combine to identify underlying hypothesized con-
structs. Confirmatory factor analysis is used in testing the measurement
model, and the hypothesized factors are referred to as latent variables.
The measures chosen by the researcher define the latent variables in the
measurement model. A latent variable is defined more accurately to the
extent that the measures that define it are strongly related to one another.
Accordingly, in building measurement models, multiple-indicator mea-
surement models (Hunter & Gerbing, 1982) are preferred because they
allow the most unambiguous assignment of meaning to the estimated
constructs. The reason for this is that with multiple-indicator measure-
ment models, each estimated construct is defined by at least two measures,
118 Analysis of Multiple Dependent Variables

and each measure is intended as an estimate of only one construct. In


contrast, measurement models that contain correlated measurement
errors or that have indicators that load on more than one estimated
construct do not represent unidimensional construct measurement. As
a result, assignment of meaning to such estimated constructs can be
problematic.
More specifically, for a measurement model to be empirically identi-
fied, latent variables must be scaled. Scaling the latent variable creates one
less unknown. Because latent variables are unobserved, they do not have
a predefined unit of measurement; consequently, the researcher needs
to set the unit of measurement (Brown, 2006). There are two ways to
do this. One option is to make it the same as that of one of the indicator
variables. The second option is to set the variance equal to 1 for the latent
variable. In general, the first option is the more popular. Scaling the latent
variable has been compared to converting currency (Brown, 2006). For
example, in creating a latent variable for cost of living across the United
States, United Kingdom, and France, a researcher has three indicators—
one in U.S. dollars, one in British pounds, and the other in Euros. Dollars,
pounds, and Euros all have different scales of measurement, but the
latent variable can be scaled to any one of these. If scaled to U.S. dollars,
the latent variable will be interpretable in terms of dollars. But, the latent
variable could also be scaled to either pounds or Euros—whichever will
be most interpretable and meaningful for the intended audience.
In addition, for a measurement model to be empirically identified at
least one of the following three conditions must hold for each construct
in the model:

1. The construct has at least three indicators whose errors are


uncorrelated with each other. Note that correlated error terms in
measurement models represent the hypothesis that the unique
variances of the associated indicators overlap; that is, they
measure something in common other than the latent constructs
that are represented in the model. Another way to describe an
error correlation is as an unanalyzed association, which means
that the specific nature of the shared “something” is unknown.
Correlated within-factor measurement errors may imply, for
example, (a) the presence of another common factor and (b)
direct causal relations among indicators (e.g., a response bias set
Structural Equation Modeling 119

where one item partly “causes” the response to the next item in a
survey);
2. The construct has at least two indicators whose errors are
uncorrelated and either (a) both the indicators of the construct
correlate with a third indicator of another construct but neither
of the two indicators’ errors is correlated with the error of that
third indicator, or (b) the two indicators’ loadings are set equal to
each other; and
3. The construct has one indicator and (a) its error variance is
fixed to zero or some other a priori value (e.g., the quantity one
minus the reliability times the indicator’s variance), or (b) there
is a variable that can serve as an instrumental variable in the
structural model and the error in the indicator is not correlated
with that instrumental variable. Note that for a variable to be
a valid instrument, then, it must be (a) correlated with the
dependent variable of a model and (b) only affect the dependent
variable through an independent variable.

The remedy for all forms of underidentification is to try to locate the


source of the identification problem and determine if the source is empir-
ical underidentification or structural underidentification. For structural
underidentification, the only remedy is to respecify the model. Empirical
underidentification may be correctable by collecting more data or respeci-
fying the model. (Hayduk 1996), for example, suggests that we can learn
a good bit about whether a model is identified by working with it. If the
software fails to converge to a solution, or matrices cannot be inverted,
the model is probably underidentified. If the model is estimated, but pro-
duces absurd results, it may be underidentified. A researcher also may esti-
mate a model by specifying different starting values for MLE parameter
estimates. If consistent results are obtained, then the model is probably
identified. Note that this will be discussed in more detail below.

THE SEM PROCEDURE


Model Specification
Often, the most difficult part of SEM, model specification involves deter-
mining every relationship and parameter in the model that is of interest
120 Analysis of Multiple Dependent Variables

to the researcher. That is, prior to any data collection or analysis, the
researcher describes a model to be confirmed. Available information is
used to decide which variables to include in the theoretical model, which
implicitly also involves which variables not to include in the model and
how these variables are related. A model will be misspecified to the extent
that the relationships hypothesized do not capture the observed relation-
ships. The following basic rules are used when drawing a model:

1. Latent variables are represented with circles and measured


variables are represented with squares;
2. Lines with an arrow in one direction show a hypothesized direct
relationship between the two variables. It should originate at the
causal variable and point to the variable that is caused;
3. Absence of a line indicates there is no causal relationship between
the variables;
4. Lines with an arrow in both directions should be curved and this
demonstrates a bi-directional relationship (i.e., a covariance);
5. Covariance arrows should only be allowed for exogenous
variables; and
6. For every endogenous variable, a residual term should be added
in the model. Generally, a residual term is a circle with the letter E
written in it, which stands for error.

Model Estimation
In SEM, the parameters of a proposed model are estimated by minimiz-
ing the discrepancy between the empirical (sample) covariance matrix,
S, and a covariance matrix, Σ, implied by the model (population). When
elements in the matrix S minus the elements in the matrix Σ equal zero
(S − Σ, = 0), then one has a perfect model fit to the data. The model
estimation process uses a fitting function or estimation procedures to
minimize the difference between Σ and S. Several fitting functions are
available. In AMOS, the following estimation procedures are available:
unweighted or ordinary least squares (ULS or OLS), generalized least
squares (GLS), asymptotically distribution free (ADF), scale-free least
squares (SFLS), and maximum likelihood (ML).
The least-squares criterion minimizes the sum of squared residuals
between the observed and predicted values of y. In the regression setting,
Structural Equation Modeling 121

this criterion is only optimal if the assumption of homoscedasticity


is satisfied. When this assumption is violated, weighted least-squares
(WLS) regression can be used instead, which minimizes a weighted sum
of squares, with the weights reflecting the different variances of indi-
vidual elements. In SEM, the assumption of homogeneity is never plau-
sible, because in the place of y we have the very different elements of the
sample covariance matrix, whose variances have no reason to be the same
(these “variances of the variances,” or fourth-order moments, are related
to each variable’s kurtosis). Moreover, while in regression we assume that
the observations are independent of each other, in SEM the elements of
the sample covariance matrix are not in fact independent, and additional
weights related to their covariances also need to be estimated. Thus, in
SEM, LS estimation is rarely the optimal choice.
GLS and ADF, also referred to in the literature as AGLS, arbitrary
distribution generalized least squares, differ in the assumptions the
researcher must make about the data and in the choice of weights. GLS
is appropriate when the variables have no excess kurtosis, so that the
weights are greatly simplified. This estimator is appropriate when the
data are normally distributed, for example. ADF does not require any
assumptions and estimates all the weights from the data before using
them in a fitting function. Because estimating these weights accurately
requires large samples, this method almost never works well unless the
sample size is very large (perhaps a thousand or more) or the model is
very simple.
In scale-free least squares estimation (SFLS), the minimum of the
fitting function is independent of the scale of the variables. For example,
with estimation that is not scale-free, the solutions based on the cova-
riance matrix input and the correlation matrix input can differ. The
GLS and ML estimation methods also are sale free, which means that
if we transform the scale of one or more of our observed variables, the
untransformed and the transformed variables will yield estimates that
are properly related, that is, that differ by the transformation.
Maximum likelihood (ML) estimates model parameters that have
the greatest chance of reproducing the observed data. First detailed by
Fisher in 1920, ML proposes the estimation of a parameter by that value
for which the likelihood function is at a maximum for the given data.
Essentially, the principle is that the value of the parameter under which
the obtained data could have had highest probability of arising must
122 Analysis of Multiple Dependent Variables

be intuitively our best estimator of the population value. For example,


suppose a researcher seeks to identify the best estimate of the popula-
tion mean given a particular sample mean. The sample mean is the ML
estimator for the population mean. That is, the researcher chooses from
among several estimates of the population mean by selecting the esti-
mate that maximizes the likelihood that the discrepancy between that
estimated population mean and the observed sample information must
be attributed to sampling error.
In SEM, the statistical criterion minimized in ML estimation is related
to discrepancies between the observed covariances and those predicted
by the researcher’s model. The mathematics of how ML estimation actu-
ally goes about generating a set of parameter estimates that minimize
its fitting function are complex, and it is beyond the scope of this sec-
tion to describe them in detail. An accessible presentation is available in
Nunnally & Bernstein (1994).
Because MLE has no algebraic formulas similar to the equations used
in OLS regression, its solution requires a computer capable of examining
many parameter sets until the best choice is identified. The procedure
begins with an initial estimate. A series of iterations, or cycles, produces
new estimates and compares them to the previous ones. The iterations
continue until successive estimates differ from the preceding ones by
less than a specified small amount. Computer programs usually issue a
warning message if iterative estimation is unsuccessful. When this occurs,
whatever final set of estimates was derived by the computer may warrant
little confidence. A tactic is to increase the probability of convergence is
to change the program’s default limit on the number of iterations to a
higher value (e.g., from 50 to 100). Allowing the computer more “tries”
may lead to a converged solution.
When a raw data file is analyzed, standard ML estimation assumes
there are no missing values. For large samples, MLE parameter estimates
are unbiased, efficient, and normally distributed, and thus allow signifi-
cance tests. More specifically, ML estimation assumes that the population
distribution for the endogenous variables is multivariate normal. A grow-
ing body of research suggest that although the values of parameter esti-
mates generated by ML are relatively robust against nonnormality, results
of statistical tests may be positively biased, which means that they lead
to the rejection of the null hypothesis too often. (c.f. Satorra & Bentler,
1994; Hoyle & Panter, 1995). Because ML is so widely available and is
Structural Equation Modeling 123

the most widely researched estimator among those otherwise available, it


has been suggested that the use of an estimation method other than ML
requires explicit justification (Hoyle & Panter, 1995). If characteristics of
the data raise question as to the appropriateness of ML, then the results
of alternative estimation procedures might be reported in summary form
if they contradict ML results or in a footnote if they corroborate them.
Alternatively, researchers may consider applying linear transformations,
such as square root and logarithmic, which are discussed in chapter 1.
However complex the mathematics of ML estimation, the interpreta-
tion of ML estimates for path models is relatively straightforward. Path
coefficients are interpreted as regression coefficients in multiple regres-
sion, which means that they control for correlations among multiple pre-
sumed causes. In standard ML estimation, standard errors are calculated
only for the unstandardized solution. This means that results of statisti-
cal tests (i.e., ratios of parameter estimates over their standard errors) are
available only for the unstandardized solution.
Also, the level of statistical significance of an unstandardized esti-
mate does not automatically apply to its standardized counterpart.
Furthermore, standard ML estimation may derive incorrect standard
errors when the variables are standardized. However, there are basically
two ways to obtain more accurate estimated standard errors for the stan-
dardized solution. Some SEM computer programs, such as AMOS, use
bootstrapping to generate standard errors for standardized estimates.
Another method is constrained estimation, which is available as a user-
specified option in some SEM computer programs including SEPATH
and RAMONA.

Model Fit and Interpretation


Finding a statistically significant theoretical model that also has practi-
cal and substantive meaning is the primary goal of using SEM to test
theories. Consequently, once estimated, the model’s fit to the data must
be evaluated. A model’s fit refers to its ability to reproduce the data (i.e.,
usually the variance–covariance matrix). The objective is to determine
whether the associations among measured and latent variables in the
researcher’s estimated model adequately reflect the observed associations
in the data. The perspective commonly taken by social workers and other
applied researchers is that approximating observed data is acceptable and
124 Analysis of Multiple Dependent Variables

can result in important contributions to the literature. It should be noted


that a good-fitting model is not necessarily a valid model. Accordingly,
researchers also should consider the implications of accepting one
explanatory model over another in terms of theory-related criteria. More
specifically, researchers should evaluate fit in terms of (1) statistical signif-
icance and strength of estimated parameters; (2) variance accounted for
in endogenous observed and latent variables; and (3) how well the overall
model fits the observed data, as indicated by a variety of fit indices.
Multiple indices are available to evaluate model fit. See Figure 5.4 for a
summary of commonly reported fit indices, including generally accepted
thresholds. These fit indices reported by most software programs, includ-
ing AMOS. SEM software packages, such as AMOS, produce a large num-
ber of alternative measures of model fit. These additional measures are
discussed in the context of the annotated example presented below. See
Kaplan (2009) or Mulaik (2009) for more information about other fit
statistics in SEM.
Following Kline (2011), Boomsma (2000), MacCallum and Austin
(2000), and McDonald and Ho (2002), the following fit indexes are
reviewed here: (1) model chi-square, (2) goodness of fit Index (GFI),
(3) comparative fit index (CFI), (4) root mean square error of approxima-
tion (RMSEA), and (5) standardized root mean square residual (SRMR).
The chi-square (χ2) test is a conventional null hypothesis significance
test (NHST). According to Kline (2011), as the most basic measure of
model fit, the χ2 test is the product (N – 1) FML, where FML is the value of
the statistical criterion (fit function) minimized in ML estimation and
(N – 1) is one less than the sample size. In large samples and assum-
ing multivariate normality, the product (N – 1) FML follows a central
chi-square distribution with degrees of freedom equal to that of the
researcher’s model, or dfM. This statistic is also known as the likelihood
ratio chi-square or generalized likelihood ratio.
The null hypothesis (H0) is that the postulated model holds in the
population; that is, the sample covariance matrix equals the popula-
tion covariance matrix. AMOS reports the value of chi-square as CMIN.
In contrast to traditional significance testing, the researcher prefers a
nonsignificant chi-square, since this result indicates that the predicted
model is congruent with the observed data. The higher the probability
level (p-value) associated with chi square, the better the fit. A significant
chi-square indicates lack of satisfactory model fit.
Structural Equation Modeling 125

Index Shorthand Acceptable Fit

Chi-square (CMIN) χ2 test Ratio of χ2 to df ≤ 3


Root Mean Square Residual RMR ≤ .05
Goodness of Fit Index GFI ≥ .95
Adjusted Goodness of Fit Index AGFI ≥ .95
Parsimony Goodness of Fit Index PGFI ≥ .95
Normed Fit Index NFI ≥ .95
Relative Fit Index RFI ≥ .95
Incremental Fit Index IFI ≥ .95
Tucker-Lewis Index TLI ≥ .95
Comparative Fit Index CFI ≥ .95
Parsimony Normed Fit Index PNFI Larger values indicate
better fit
Parsimony Comparative Fit Index PCFI Larger values indicate
better fit
Root Mean Square Error of the RMSEA ≤ .05
Approximation
Akaike Information Criterion AIC Smaller values indicate
better fit
Browne-Cudeck Criterion BCC Smaller values indicate
better fit
Bayes Information Criterion BIC Smaller values indicate
better fit
Consistent AIC Criterion CAIC Smaller values indicate
better fit
Expected Cross-Validation Index ECVI Smaller values indicate
better fit
Modified Expected Cross-Validation MECVI Smaller values indicate
Index better fit

Figure 5.4 Commonly Reported for Fit Indices.

The chi-square (χ2) test has at least two limitations. First, the
chi-square test offers only a dichotomous decision strategy implied by a
statistical decision rule and cannot be used to quantify the degree of fit
along a continuum with some prespecified boundary. Second, as with
most statistics, large sample sizes increase power, resulting in signifi-
cance with small effect sizes. Consequently, a nonsignificant χ2 may be
unlikely, although the model may be a close fit to the observed data.
Despite these limitations, researchers almost universally report the χ2
(Martens, 2005).
126 Analysis of Multiple Dependent Variables

The Goodness-of-Fit Index (GFI) estimates the proportion of cova-


riances in the sample data matrix explained by the model. That is, the
GFI estimates how much better the researcher’s model fits compared
with no model at all (Jöreskog, 2004). The range of values for this pair of
approximate fit indexes is generally 0–1.0, where 1.0 indicates the best fit.
A general formula is

C res
GFI = 1 − (5.1)
C tot

where Cres and Ctot estimate, respectively, the residual and total variabil-
ity in the sample covariance matrix. The numerator in the right side of
Equation 5.1 is related to the sum of the squared covariance residuals,
and the denominator is related to the total sum of squares in the data
matrix. GFI should by equal to or greater than .90 to indicate good fit.
GFI is less than or equal to 1. A value of 1 indicates a perfect fit. GFI
tends to be larger as sample size increases. GFI >.95 indicates good fit.
GFI index is roughly analogous to the multiple R-square in multiple
regression because it represents the overall amount of the covariation
among the observed variables that can be accounted for by the hypoth-
esized model. One limitation of the GFI is that its expected values vary
with sample size. Another limitation is that values of the GFI sometimes
fall outside of the range 0 – 1.
The Comparative Fit Index (CFI) (Bentler, 1990) measures the rela-
tive improvement in the fit of the researcher’s model over that of a base-
line model, typically the independence or null model, which specifies no
relationships among variables. CFI ranges from 0 to 1.0. Values close to 1
indicates a very good fit, >.9 or close to .95 indicates good fit, by conven-
tion, CFI should be equal to or greater than .90 to accept the model. The
formula is

χ2M − df M
CFI = 1 − (5.2)
χ2B − df B

where the numerator and the denominator of the expression on the right
side of the equation estimates the chi square noncentrality parameter for,
respectively, the researcher’s model and the baseline model. Note that
noncentrality parameters reflect the extent to which the null hypothesis
Structural Equation Modeling 127

is false. For example, the traditional χ2 test assumes that the null hypoth-
esis is true (χ2 = 0) in the population. This test relies on the “central”
distribution of χ2 values. Because the researcher is hoping not to reject
the null hypothesis, it is argued that it is more appropriate to test the
alternative hypothesis (Ha). This test of Ha would rely on a “noncentral”
chi-square distribution that assumes Ha is true in the population. This
approach to model fit uses a chi-square equal to the df for the model
as having a perfect fit (as opposed to χ2 = 0). Thus, the noncentrality
parameter estimate is calculated by subtracting the df of the model from
the chi-square (χ2 − df).
CFI is relatively insensitive to sample size (Fan, Thompson, and Wang,
1999). However, one limitation is that the null hypothesis that the base-
line model is better than the independence model is almost always true.
This is because the assumption of zero covariances among the observed
variables is improbable in most studies. Although it is possible to specify
a different, more plausible baseline model—such as one that allows the
exogenous variables only to covary—and compute by hand the value of
an incremental fit index with its equation, this is rarely done in practice.
Widaman and Thompson (2003) describe how to specify more plausible
baseline models.
The root-mean-square error of the approximation (RMSEA) is
another index based that is based on the noncentrality parameter (Steiger,
1990). The formula is

χ2M − df M
RMSEA = (5.3)
df M (N − 1)

The numerator is the noncentrality parameter, and the denominator


is the product of the model degrees of freedom and one less than the
sample size. That is, the value of the RMSEA decreases as there are more
degrees of freedom (greater parsimony) or a larger sample size, keeping
all else constant. However, the RMSEA does not necessarily favor models
with more degrees of freedom. This is because the effect of the correction
for parsimony diminishes as the sample size becomes increasingly large
(Kline, 2011).
Hu and Bentler (1995) suggest values below .06 indicate good fit. The
RMSEA values are classified into four categories: close fit (.00 – .05), fair fit
(.05 – .08), mediocre fit (.08 – .10), and poor fit (over .10). RMSEA tends
128 Analysis of Multiple Dependent Variables

to improve as we add variables to the model, especially with larger sample


size. A confidence interval can be computed for this index. According to
Kline (2011), the population parameter estimated by the RMSEA is often
designated as ε. Because the RMSEA’s distribution values are known, a
confidence interval around the point estimate of the RMSEA can be con-
structed to indicate the level of its precision. In computer output, the
lower and upper bounds of the 90% confidence interval for ε are often
printed along with the sample value of the RMSEA, the point estimate
of ε. Using this confidence interval, evaluating the null hypothesis can
be examined more precisely. In using these confidence intervals, a null
hypothesis could be rejected in favor of accepting the alternative if the
entire range of the confidence interval is less than .05. So, like the χ2 sta-
tistic, it is possible to use RMSEA to evaluate the null hypothesis that a
model fits the data exactly. According to Kline (2011), the width of this
confidence interval is generally larger in smaller samples, which indicates
less precision. The bounds of the confidence interval for ε may not be
symmetrical around the sample value of the RMSEA, and, ideally, the
lower bound equals zero. If the upper bound of the confidence interval
for ε exceeds a value that may indicate “poor fit,” then the model warrants
less confidence.
Unlike the chi-squared statistic, RMSEA is less affected by sample
size problems. Relatively little information is currently available on the
performance of RMSEA when data are nonnormal. However, available
information suggests that RMSEA may perform less optimally when there
are large sample sizes and relatively small degrees of freedom (Quintana
& Maxwell, 1999). The interpretation of the RMSEA and the lower and
upper bounds of its confidence interval depends on the assumption that
this statistic follows noncentral chi-square distributions. Evidence sug-
gests that the aforementioned assumption may not be tenable that (cf.
Olsson, Foss, & Breivik, 2004; Yuan, 2005). RMSEA tends to impose a
harsher penalty for complexity on smaller models with relatively few
variables or factors. This is because smaller models may have relatively
few degrees of freedom, but larger models may have more “room” for
higher dfM values. Consequently, the RMSEA may favor larger models
(Breivik & Olsson, 2001).
The standardized root-mean-square residual (SRMR) is based on
transforming both the sample covariance matrix and the predicted cova-
riance matrix into correlation matrices. The SRMR is thus a measure of
Structural Equation Modeling 129

the mean absolute correlation residual, the overall difference between the
observed and predicted correlations. The Hu and Bentler (1999) thresh-
old of SRMR ≤ .08 for acceptable fit was not a very demanding standard.
This is because if the average absolute correlation residual is around .08,
then many individual values could exceed this value, which would indi-
cate poor explanatory power at the level of pairs of observed variables.
It is better to actually inspect the matrix of correlation residuals and
describe their pattern as part of a diagnostic assessment of fit than just to
report the summary statistic SRMR.
In addition to considering overall model fit, it is important to con-
sider the significance of estimated parameters, which are analogous to
regression coefficients. As with regression, a model that fits the data quite
well but has few significant parameters would be meaningless. At a mini-
mum, the researcher should inspect model estimates to determine if pro-
posed parameters were significant and in the expected direction.

Model Modification
Rarely is a proposed model the best-fitting model. Consequently, modifi-
cation, also termed respecification, may be needed. This involves adjusting
the estimated model by freeing (estimating) or setting (not estimating)
parameters. Post hoc modifications of the model are often based on
modification indices. Improvement in fit is measured by a reduction in
chi-square, which makes the chi-square fit index less likely to be found
significant (recall a finding of significance corresponds to rejecting the
model as one that fits the data). For each fixed and constrained param-
eter (coefficient), the modification index reflects the predicted decrease
in chi-square if a single fixed parameter or equality constraint is removed
from the model by eliminating its path, and the model is re-estimated.
One arbitrary rule of thumb is to consider eliminating paths associated
with parameters whose modification index exceeds 10. However, another
approach is to eliminate the parameter with the largest MI, then see the
effect as measured by the chi-square fit index.
Modification is a controversial topic, which has been likened to the
debate about post hoc comparisons in ANOVA (MacCallum & Austin,
2000; McDonald & Ho, 2002). The suggested modifications, however,
may or may not be supported on theoretical grounds. As with ANOVA
and regression, problems with model modification include capitalization
130 Analysis of Multiple Dependent Variables

on chance and results that are specific to a sample because they are data
driven. Although there is disagreement regarding the acceptability of
post hoc model modification, statisticians and applied researchers alike
emphasize the need to clearly state when there was post hoc modification
rather than imply that analyses were a priori.
Researchers are urged not to make too many changes based on modi-
fication indices, even if such modifications seem sensible on theoretical
grounds. Note that SEM takes a confirmatory approach to model testing;
one does not try to find the best model or theory via data using SEM.
Rather than data-driven post hoc modifications (which may be very
inconsistent over repeated samples), it is often more defensible to con-
sider multiple alternative models a priori. That is, multiple models (e.g.,
based on competing theories or different sides of an argument) should
be specified prior to model fitting, and the best-fitting model should be
selected among the alternatives. Because a more complex model, assum-
ing it is identified, will generally produce better fit, and different models
can produce the same fit, theory is imperative in model testing.
In conclusion, it is worth noting that although SEM allows the test-
ing of causal hypotheses, a well-fitting SEM model does not and can-
not prove causal relations without satisfying the necessary conditions for
causal inference (e.g., time precedence, robust relationship in the pres-
ence or absence of other variables). A selected well-fitting model in SEM
is like a retained null hypothesis in conventional hypothesis testing; it
remains plausible among perhaps many other models that are not tested
but may produce the same or better level of fit. SEM users are cautioned
not to make unwarranted causal claims. Replications of findings with
independent samples are recommended, especially if the models are
obtained with post hoc modifications.

STRENGTHS AND LIMITATIONS OF SEM


Strengths and Limitations of SEM
SEM has the following analytical strengths:

1. Enables researchers to answer a set of interrelated research


questions by modeling the relationships among multiple IVs
and DVs simultaneously. That is, SEM permits complicated
Structural Equation Modeling 131

variable relationships to be expressed through hierarchical or


nonhierarchical, recursive or nonrecursive structural equations;
2. Allows researchers to specify latent variable models that provide
separate estimates of relations among latent constructs and
their manifest indicators (the measurement model) and of the
relations among constructs (the structural model);
3. Availability of measures of global fit that can provide a summary
evaluation of even complex models that involve a large number
of linear equations. Most alternative procedures that might be
used in place of SEM (e.g., MANOVA and MANCOVA) to test
such models would provide only separate “minitests” of model
components that are conducted on an equation-by-equation
basis; and
4. Allows researchers to directly test the model of interest, also
termed the theoretical or alternative hypothesis, directly. In
most statistical contexts, the researcher’s theoretical hypothesis
is aligned with the alternative hypothesis rather than the null
hypothesis. In contrast, in SEM the theoretical hypothesis is often
aligned with the null hypothesis, which specifies that the model
fits exactly or at least approximately (e.g., MacCallum et al. 1993).
It should be noted that between-group comparisons of factor
means, for examples, represents an exceptions to this conclusion;

SEM has the following analytical limitations:

1. A well-fitting SEM model does not demonstrate causality. That is,


a well-fitting model in SEM is plausible among other models that
are not tested but may produce the same or better level of fit;
2. Cannot be used as an exploratory modeling strategy. Theory
should be used to guide model specification;
3. Like all statistical strategies, SEMs are approximations of
reality. For example, important variables may be omitted. Such
omissions present a misleading picture of the measurement.

Although the problem of omitted variables is not unique to SEM,


the implications of model misspecification in SEM must be recognized.
According to Tomarken and Waller (2005), users often underestimate
the importance of residual variance and covariance terms for generating
132 Analysis of Multiple Dependent Variables

a model with acceptable fit. More specifically, if residual variance and


covariance terms were to be omitted, at least some models would fit
poorly. Tomarken and Waller (2005) note a paradox: “the residual
parameterizations afforded by SEM software can mask the limitations of
a rather incomplete model” (p. 49); and

4. Measures of global fit do not directly test lower-order


components of a model, such as path coefficients and relevant
quantities that can be derived from such parameters, such as
the proportion of variance in an endogenous variable that
is accounted for by the specified predictors in the model.
Researchers must be aware that a model can fit perfectly yet
be associated with problematic lower-order components (e.g.,
parameter estimates that are biased, small in magnitude or
opposite to theoretical expectations).

ANNOTATED EXAMPLE
A researcher plans to examine the relationship between factors that influ-
ence postadoption service utilization and positive adoption outcomes.
Specifically, the study tests a model that links (1) factors influencing the
utilization of postadoption services (parents’ perceptions of self-efficacy,
relationship satisfaction between parents, and attitudes toward adoption)
with (2) service utilization, and (3) positive adoption outcomes (satisfac-
tion with parenting and satisfaction with adoption agency). See Figure 5.5.
Note that latent variables are not included in the current model, since

e2
1

e1 Satisfaction with
Self Efficacy
Adoption Agency
1

Attitude toward Service e3


Adoption Utilization 1

Satisfaction with
Relationship
Parenting
Satisfaction

Figure 5.5 Annotated Example.


Structural Equation Modeling 133

instruments are unidimensional with a high reliability. Unidimensionality


and reliability of instruments is discussed further in chapter 2 (Annotated
Example section) and in chapter 6.

Determining the Minimally Sufficient Sample Size


As discussed earlier in this chapter, structural equation modeling (SEM)
does not use raw data. Instead, in SEM the variance (or covariance)
matrix is used. The number of observations in SEM is defined as the
number of covariances in the matrix rather than the number of cases
in a data set. The number of observations or covariances in SEM can be
calculated as follows:

v(v + 1)/2,

where v is the number of variables in the model. In the current model,


there are seven variables (i.e., self-efficacy, satisfaction with marital rela-
tionship, attitude toward adoption, knowledge of agency services, service uti-
lization, satisfaction with parenting, and satisfaction with adoption agency).
Therefore, the number of variables in this example is 6(6 + 1)/2, or 21.
The researcher determines sample size by following the approach pro-
posed by MacCallum and Browne (1993), which was described earlier in
this chapter. This approach uses the RMSEA (Brown, 2006; Hu & Bentler,
1999; Steiger, 1990). This index weighs absolute fit, which declines when-
ever a parameter is removed from the model, against model complexity,
such that the benefits of parsimony are considered along with fit (Steiger,
1990). Models fitting with RMSEA <.05 are usually considered as having
a “close” fit, .05 to .08 as having a “fair” fit, .08 to .10 as having a “medio-
cre” fit, and above .10 as having a “poor” fit (MacCallum et al., 1993).
A program called NIESEM performs power analysis according to
the approach proposed by MacCallum and Browne, (1993). NIESEM
is free and available for download from http://rubens.its.unimelb.edu.
au/∼dudgeon/. To estimate sample size with NIESEM, the researcher
proceeds as follows:

1. Selects Power calculations


2. Selects estimate N for given power
3. Inputs power equals 0.80
134 Analysis of Multiple Dependent Variables

4. Inputs the null hypothesized RMSEA value equals 0.00


5. Inputs the alternative hypothesized RMSEA value equals 0.05
6. Inputs the chosen alpha level equals 0.05
7. Inputs the degrees of freedom of the model equals 6(6 + 1)/2 =
21
8. Inputs the number of groups in the model equals 1
9. Clicks OK

The output is as follows (see Figure 5.6) and the estimated sample size
equals 408.
An alternative strategy for estimating sample size for SEM is provided
by the following webpage: http://timo.gnambs.at/en/scripts/power-
forsem. This webpage generates syntax for SPSS and R to estimate sam-
ple size for various measures of model fit, including RMSEA, GFI, AGFI.
This webpage http://www.datavis.ca/sasmac/csmpower.html provides an
SAS macro, csmpower, to calculate retrospective or prospective power
computations for SEM using the method of MacCallum and Browne
(1993). Their approach allows for testing a null hypothesis of “not-good-
fit,” so that a significant result provides support for good fit. Effect size
in this approach is defined in terms of a null hypothesis and alterna-
tive hypothesis value of the root-mean-square error of approximation
(RMSEA) index. These values, together with the degrees of freedom (df)
for the model being fitted, the sample size (n), and error rate (alpha),
allow power to be calculated.

CSM Power Analysis

RMSEA Null Value = 0.000


RMSEA Alternative Value = 0.050
Alpha significance level= 0.050
Degrees of freedom = 21
Number of groups = 1

Desired power = 0.800

Estimated sample size = 408

Figure 5.6 CSM Power.


Structural Equation Modeling 135

To perform the SEM analysis in the researcher proceeds as


follows:

1. Selects the data set that will be used to test the model by
clicking File Data Files File Name to browse to and select the
file;
2. Draws the model diagram by using the Diagram dropdown
menu (see Figure 5.7);
3. Draws observed variables by selecting Draw Observed, then uses
the cursor to draw the five observed variables (rectangles) in the
model;
4. Names an observed variable by right-clicking on a rectangle, and
under text tab, adds the variable name;
5. Draws latent variables (not included in the current mode) by
selecting Draw Observed, and then uses the cursor to draw the
five observed variables (rectangles) in the model;
6. Names a latent variable by right-clicking on a rectangle, and
under text tab, adding the variable name;
7. Draws paths by selecting Draw Path, and using the cursor;
8. Draws error terms by selecting Draw Unique Variable and using
the cursor; and
9. Names an error term by right-clicking a rectangle, and under text
tab adding the error term’s name.
Once the model is illustrated,
10. Clicks View Analysis Properties Estimation tab;
11. Selects Maximum likelihood;
12. Selects Fit saturated and independence models;
13. Clicks Output tab;
14. Selects Standardized estimates;
15. Selects Residual moments;
16. Selects Modification indices;
17. Selects Indirect, direct, and total effects;
18. Selects Covariances of estimates;
19. Selects Correlations of estimates;
20. Selects Tests for normality and outliers;
21. Once the model is illustrated and analysis properties are selected,
clicks Analyze, Calculate Estimates; and
136 Analysis of Multiple Dependent Variables

Figure 5.7 Diagram Dropdown Menu.

22. To view results clicks View Text Output.


(Note that the toolbar icons may also be used to view analysis
properties, calculate estimates, and view output.)
AMOS next prints out a large number of alternative measures
of model fit (see Figure 5.3). Each measure is calculated for three
models. “Default model” is the “current model” as specified by the
researcher. The “independence model” is the model in which variables
are assumed to be uncorrelated with the dependent(s), so if the fit for
“your model” is no better than for the “independence model,” then
your model should be rejected. The “saturated model” is one with no
constraints and will always fit any data perfectly, so normally your
model will have a measure of fit between the saturated and indepen-
dence models.
The output is as follows: NPAR is the number of parameters being esti-
mated in the model and is not a measure of fit. AMOS reports the value of
the χ2 test as CMIN, and its associated p-value as P(CMIN). If P(CMIN)
is less than .05, we reject null hypothesis that the data are a perfect fit to
Structural Equation Modeling 137

the model. By this criterion the present model is not rejected. Recall that
nonsignificant chi-square (e.g., p > .05) indicates that the parameters that
were estimated for the model fit the data (Please see Figure 5.8).
CMIN/DF is the χ2 test divided by the current model’s degrees of
freedom (df). Some researchers allow values as large as 5 as being an
adequate fit, but conservative use calls for rejecting models with rela-
tive chi-square greater than 3. By this criterion the current model is not
rejected (Please see Figure 5.8).
One important problem with chi-square is that with large samples
significance is easy to obtain. Given that large sample are recommended
for the SEM technique, this presents a dilemma. A solution has been to
develop what are called fit indexes which are based on the chi-square but
which control in some way for sample size.
RMR is the Root-Mean-Square Residual, which is the square root of
the mean squared amount by which the sample covariances differ from
the estimated covariances, estimated on the assumption that your model
is correct; the smaller the RMR, the better the fit. An RMR of zero indi-
cates a perfect fit. The closer the RMR to 0 for a model being tested, the
better the model fit, and an RMR value smaller than .05 suggests good fit.
For these data, RMR equals 9.370, which suggests that the current model
is not a good fit (Please see Figure 5.9).
GFI is the Goodness of Fit Index. GFI varies from 0 to 1, but theoreti-
cally can yield meaningless negative values. By convention, GFI should
by equal to or greater than .90 to accept the model. By this criterion the
present model (GFI = .979) is accepted (Please see Figure 5.9).
AGFI is the Adjusted Goodness of Fit Index. AGFI is a variant of GFI.
AGFI also varies from 0 to 1, but theoretically can yield meaningless neg-
ative values. AGFI should also be at least .90. By this criterion the present
model (AGFI = .913) is accepted (Please see Figure 5.9).

Model NPAR CMIN DF P(CMIN) CMIN/DF


Default model
(current model) 16 6.499 5 .261 1.300

Saturated model
21 .000 0
Independence model
6 87.629 15 .000 5.842

Figure 5.8 AMOS CMIN Output.


138 Analysis of Multiple Dependent Variables

PGFI is the Parsimony Goodness of Fit Index. It is a variant of


GFI that penalizes GFI by multiplying it times the ratio formed by the
degrees of freedom in your model and degrees of freedom in the inde-
pendence model. The PGFI ranges from 0 to 1, with higher values indi-
cating a more parsimonious fit. There is no standard for how high the
index should be to indicate parsimonious fit. This index is best used to
compare two competing theoretical models. That is, calculate PGFI for
each model and choose the model with the highest level of parsimoni-
ous fit (Please see Figure 5.9).
NFI is the Normed Fit Index, which varies from 0 to 1, with 1 = per-
fect fit. By convention, NFI values below .90 indicate a need to respecify
the model. By this criterion the present model (NFI = .926) is accepted
(Please see Figure 5.10).
RFI is the Relative Fit Index, which is not guaranteed to vary from 0
to 1. RFI close to 1 indicates a good fit. By this criterion the present model
(RFI = .778) is accepted (Please see Figure 5.10).
IFI is the Incremental Fit Index, which is not guaranteed to vary
from 0 to 1. IFI close to 1 indicates a good fit and values above .90
an acceptable fit. By this criterion the present model (IFI = .982) is
accepted (Please see Figure 5.10).
TLI is the Tucker–Lewis Index, also called the Bentler–Bonett
non-normed fit index (NNFI). TLI is not guaranteed to vary from 0 to
1. TLI close to 1 indicates a good fit. By this criterion the present model
(TLI = .938) is accepted (Please see Figure 5.10).
CFI is the Comparative Fit Index, which varies from 0 to 1. CFI close to
1 indicates a very good fit, and values above .90 an acceptable fit. By this cri-
terion the present model (CFI = .979) is accepted (Please see Figure 5.10).
Note that the DELTA and RHO headings are alternative names for
these measures.

Model RMR GFI AGFI PGFI

Default model
(current model) 9.370 .979 .913 .233

Saturated model .000 1.000

Independence model 53.318 .772 .681 .552

Figure 5.9 RMR, GFI, AGFI, PGFI Output


Structural Equation Modeling 139

PRATIO is the Parsimony Ratio, which is the ratio of the degrees


of freedom in your model to degrees of freedom in the independence
(null) model. PRATIO is not itself a goodness-of-fit test, but is used in
goodness-of-fit measures like PNFI and PCFI that reward parsimonious
models with relatively few parameters to estimate in relation to the num-
ber of variables and relationships in the model (Please see Figure 5.11).
PNFI is the Parsimony Normed Fit Index, equal to the PRATIO times
NFI. There is no standard for how high the index should be to indicate
parsimonious fit. This index is best used to compare two competing the-
oretical models (Please see Figure 5.11).
PCFI is the Parsimony Comparative Fit Index, equal to PRATIO
times CFI. There is no standard for how high the index should be to indi-
cate parsimonious fit. This index is best used to compare two competing
theoretical models (Please see Figure 5.11).
NCP is the noncentrality parameter. It and FO are used in the compu-
tation of RMSEA, the root-mean-square error of approximation, which
incorporates the discrepancy function criterion (comparing observed
and predicted covariance matrices) and the parsimony criterion (see
above). For each, LO 90 and HI 90 indicate 90% confidence limits on the
coefficient (Please see Figure 5.12).
FMIN is the minimum fit function. It can be used as an alternative
to CMIN to compute CFI, NFI, NNFI, IFI, and other fit measures (Please
see Figure 5.13).
RMSEA is the Root-Mean-Square Error of the Approximation.
The RMSEA values usually are classified into four categories: close
fit (.00 – .05), fair fit (.05 – .08), mediocre fit (.08 – .10), and poor
fit (over .10). By convention, there is good model fit if RMSEA less
than or equal to .05. For these data, RMSEA equals .055, which sug-
gests that the model is a fair fit. PCLOSE tests the null hypothesis
that RMSEA is no greater than .05. Since PCLOSE equals .394, we do

NFI RFI IFI TLI CFI


Model
Delta1 rho1 Delta2 rho2

Default model .926 .778 .982 .938 .979

Saturated model 1.000 1.000 1.000

Independence model .000 .000 .000 .000 .000

Figure 5.10 NFI, RFI, IFI, TLI, CFI Output.


140 Analysis of Multiple Dependent Variables

Model PRATIO PNFI PCFI


Default model
(current model) .333 .309 .326

Saturated model .000 .000 .000

Independence model 1.000 .000 .000

Figure 5.11 PRATIO, PHFI, PCFI Output.

not reject the null hypothesis and conclude that RMSEA is no greater
than .05 (Please see Figure 5.14).
AIC is the Akaike Information Criterion. The AIC measure indicates
a better fit when it is smaller. The measure is not standardized and is
not interpreted for a given model. For two models estimated from the
same data set, the model with the smaller AIC is to be preferred. The
AIC makes the researcher pay a “penalty” for every parameter that is
estimated. The absolute value of AIC has relatively little meaning; rather
the focus is on the relative size, the model with the smaller AIC being
preferred (Please see Figure 5.15).
BCC is the Browne–Cudeck Criterion, also called the Cudeck &
Browne single-sample cross-validation index. The BCC should be close
to .9 to conclude a model is a good fit. BCC penalizes for model com-
plexity (lack of parsimony) more than AIC. For two models estimated
from the same data set, the model with the smaller BCC is to be pre-
ferred (Please see Figure 5.15).
BIC is the Bayes Information Criterion, also known as Akaike’s
Bayesian information criterion (ABIC). BIC penalizes for sample size as
well as model complexity. Specifically, BIC penalizes for additional model
parameters more severely than does AIC. For two models estimated from

Model NCP LO 90 HI 90

Default model 1.499 .000 12.357

Saturated model .000 .000 .000

Independence model 72.629 46.848 105.919

Figure 5.12 NCP Output.


Structural Equation Modeling 141

Model FMIN F0 LO 90 HI 90

Default model .066 .015 .000 .125

Saturated model .000 .000 .000 .000

Independence model .885 .734 .473 1.070

Figure 5.13 FMIN Output.

the same data set, the model with the smaller BIC is to be preferred
(Please see Figure 5.15).
CAIC is the Consistent AIC Criterion, which also penalizes for sam-
ple size as well as model complexity (lack of parsimony). The penalty is
greater than AIC or BCC but less than BIC. For two models estimated
from the same data set, the model with the smaller CAIC is to be pre-
ferred (Please see Figure 5.15).
ECVI is the Expected Cross-Validation Index. It is another variant on
AIC. For two models estimated from the same data set, the model with
the smaller ECVI is to be preferred. MECVI is the Modified Expected
Cross-Validation Index. It is a variant on BCC. MECVI is a variant on
BCC, differing in scale factor. Compared to ECVI, a greater penalty is
imposed for model complexity. Lower is better between models. For two
models estimated from the same data set, the model with the smaller
MECVI is to be preferred (Please see Figure 5.16).
Hoelter’s critical N or Hoetler index is the largest sample size at
which the researcher would accept the model at the .05 or .01 levels. This
offers a perspective on the chi-square fit index, which has the problem
that the larger the sample size, the more likely the rejection of the model
and the more likely a Type II error. In this case, actual sample size was
408 and the model was accepted. If the sample size had been only 164, it
would have been accepted at the .05 level (Please see Figure 5.17).

Model RMSEA LO 90 HI 90 PCLOSE

Default model .055 .000 .158 .394

Independence model .221 .178 .267 .000

Figure 5.14 RMSEA Output.


142 Analysis of Multiple Dependent Variables

Model AIC BCC BIC CAIC

Default model 38.499 40.934 80.182 96.182

Saturated model 42.000 45.196 96.709 117.709

Independence model 99.629 100.542 115.260 121.260

Figure 5.15 AIC, BCC, BIC, CAIC Output.

To calculate the Standardized Root-Mean-Square Residual (SRMR)


in AMOS,

1. Select Analyze Calculate Estimates;


2. Select Plugins Standardized RMR, and a blank Standardized
RMR dialog box is displayed; and
3. Re-select Analyze Calculate Estimates, and the Standardized
RMR dialog will display SRMR.

For these data, SRMR equals .0513. Therefore, the model is a fair
to good fit. SRMR ≤ .08 suggests good fit. Additionally, the matrix of
correlation was inspected. No unusual coefficients were observed. No
modification index was greater than zero, and, consequently, no model
modifications were suggested.
In addition to considering overall model fit, it is important to con-
sider the significance of estimated parameters, which are analogous to
regression coefficients. As with regression, a model that fits the data
quite well but has few significant parameters would be meaningless. At

Model ECVI LO 90 HI 90 MECVI

Default model .389 .374 .499 .413

Saturated model .424 .424 .424 .457

Independence model 1.006 .746 1.343 1.016

Figure 5.16 ECVI, MECVI Output.


Structural Equation Modeling 143

HOELTER HOELTER
Model
.05 .01

Default model 469 431

Independence model 329 335

Figure 5.17 Hoelter’s Critical N Output.

a minimum, the researcher should inspect model estimates to deter-


mine if proposed parameters were significant and in the expected
direction.
For the final model (see Figure 5.18), correlation coefficients for
exogenous variables, and unstandardized and standardized regression
coefficients for endogenous variables were calculated. Of note for the
exogenous variables, Self Efficacy and Relationship satisfaction were mod-
erately correlated (r = .47).
Of note for the endogenous variables, the standardized coefficients
between Relationship Satisfaction and Service Utilization was −.52, Self
Efficacy and Service Utilization was −.15, Relationship Satisfaction and
Satisfaction with Parenting was .16, Service Utilization and Satisfaction
with Parenting was .19, and Satisfaction with Parenting and Satisfaction
with Adoption Agency was .28. That is, the best predictor of Satisfaction
with Parenting is Service Utilization, and the best predictor of Service
Utilization is Sat with Parenting.

e2

e1 Satisfaction with
Self Efficacy
–.15 Adoption Agency
.01
.10
.47 Attitude toward .09 Service e3 .28
Adoption Utilization
–.52 .19
.9
Satisfaction with
Relationship –16
Parenting
Satisfaction

Figure 5.18 The Final Model.


144 Analysis of Multiple Dependent Variables

REPORTING THE RESULTS OF A SEM


1. Sample selection strategy and size with justifications (e.g.,
power analysis);
2. Extent and handling of issues related to missing data,
normality, outliers, and multicollinearity;
3. Model identification;
4. Software program used;
5. Estimation method;
6. Model fit;
7. Model coefficients;
8. Residual analysis (i.e., predicted and actual covariance matrix
examination);
9. Model modification, including rationale; and
10. Diagram of final model.

RESULTS
The hypothesized model is described in Figure 5.5. An SEM analysis was
performed on hypothetical data from 408 clients of an adoption agency
(one mean score for each adoptive couple for each concept). The analysis
was performed using AMOS version 18. The assumptions of no missing
data, multivariate normality, linearity, and absence of multivariate outli-
ers, and perfect multicollinearity were evaluated. Each of the aforemen-
tioned assumptions seems tenable.
The hypothesized model was tested with maximum likelihood esti-
mation. The hypothesized model appears to be a good fit to the data.
P(CMIN) equals .261, and the present model is not rejected. CMIN/
DF equals 1.30. By this criterion the current model is not rejected. GFI
equals .979. By this criterion the present model is not rejected. CFI equals
.979. By this criterion the present model is not rejected. RMSEA equal
to .00 − .05 indicates close fit. As PCLOSE equals .394, we do not reject
the null hypothesis and conclude that RMSEA is no greater than .05.
By this criterion the present model is not rejected. SRMR equals .0513.
Therefore, the model is a fair to good fit. SRMR ≤ .08 suggests good fit.
Additionally, the matrix of correlation was inspected. No unusual coef-
ficients were observed. No modification index was greater than zero, and,
consequently, no model modifications were suggested. No modification
Structural Equation Modeling 145

index was greater than zero, and, consequently, no model modifications


were suggested.
In addition to considering overall model fit, it is important to con-
sider the size of the estimated parameters, which are analogous to regres-
sion coefficients. As with regression, a model that fits the data quite well
but has few practically significant parameters would be meaningless. At
a minimum, the researcher should inspect model estimates to determine
if proposed parameters were of practically significant magnitudes and in
the expected directions.
For the final model (see Figure 5.16), correlation coefficients for
exogenous variables, and standardized regression coefficients for endog-
enous variables were calculated. Of note for the exogenous variables,
Self Efficacy and Relationship satisfaction were moderately correlated
(r = .47). Of note for the endogenous variables, the standardized coeffi-
cients between Relationship Satisfaction and Service Utilization was −.52,
Self Efficacy and Service Utilization was −.15, Relationship Satisfaction and
Satisfaction with Parenting was .16, Service Utilization and Satisfaction
with Parenting was .19, and Satisfaction with Parenting and Satisfaction
with Adoption Agency was .28. That is, the best predictor of Satisfaction
with Parenting is Service Utilization, and the best predictor of Service
Utilization is Sat with Parenting.

ADDITIONAL EXAMPLES OF SEM FROM THE APPLIED RESEARCH LITERATURE


Benda, B. B. (2003). Test of a structural equation model of comorbidity among
homeless and domiciled military veterans. Journal of Social Service Research,
29(1), 1–35.
This exploratory study of 600 Vietnam era military veterans was designed
to test a hypothesized theoretical model of comorbidity (substance abuse and
depression) among domiciled and homeless persons. The model tested repre-
sented an integration of stress process and social support models found in the
literatures on substance abuse and on depression. Caregiver attachment and
early abuse also were used to elaborate on the integrated stress-social support
theoretical model. Using structural equation modeling, all relationships in the
hypothesized model were supported by data from domiciled veterans, except
attachments to caregivers were not related to peer associations with substance
users, and these associations were not related to depression. In contrast, all
24 relationships hypothesized in the model tested were supported among
146 Analysis of Multiple Dependent Variables

homeless veterans. Conceptual and practice implications of the findings were


discussed.

Fandrem, H., Strohmeier, D., & Roland, E. (2009). Bullying and victimization
among native and immigrant adolescents in Norway: The role of proactive and
reactive aggressiveness. The Journal of Early Adolescence, 29(6), 898–923.
This study compares levels of bullying others, victimization, and aggressiveness
in native Norwegian and immigrant adolescents living in Norway and shows how
bullying is related to proactive and reactive aggressiveness. The sample consists
of 2,938 native Norwegians and 189 immigrant adolescents in school grades 8, 9,
and 10. Data were collected via self-assessments. SEMs were conducted separately
for girls and boys in both groups. The levels of victimization, reactive and pro-
active aggressiveness were the same for both native Norwegians and immigrant
adolescents but there was a significant difference in the levels of bullying others.
Compared with the native Norwegians, immigrant adolescents were found to be
at higher risk of bullying others. Structural models revealed significantly stronger
relations between affiliation-related proactive aggressiveness and bullying others
in immigrant boys compared with the other groups. This indicates that the wish
for affiliation is an important mechanism of bullying others in immigrant boys.
The authors also suggest further research and the practical importance of the
findings for prevention of targeting immigrant adolescents.

Jung, J. S. (2009). The relationships of flexible work schedules, workplace sup-


port, supervisory support, work-life balance, and the well-being of working
parents. Journal of Social Service Research, 35(2), 93–104.
This study, using a secondary dataset from the 2002 National Study of the
Changing Workforce, examines how working parents cope with work demands
and family responsibilities. The design is a study on the relationships of flexible
work schedule, workplace support, supervisory support, and work-life bal-
ance on the well-being of working parents employing the SEM. In this study,
employee well-being is an endogenous latent construct. Work-schedule flex-
ibility, workplace support, supervisory support, and work-life balance are latent
exogenous constructs. This information will assist social workers in developing
more effective intervention efforts in the workplace, with the ultimate goal of
increasing the quality of life.

McGowan, B. G., Auerbach, C., & Strolin-Goltzman, J. S. (2009) Turnover in


the child welfare workforce: A different perspective. Journal of Social Service
Research, 35(3), 228–235.
Child welfare agencies across the country are experiencing a workforce
crisis involving high staff turnover rates. The purpose of this study was
Structural Equation Modeling 147

to determine which of the organizational, personal, and supervisory vari-


ables identified in prior research on this topic are most associated with
intent to leave among employees in urban and rural child welfare settings.
Four-hundred-and-forty-seven employees in 13 child welfare agencies partici-
pated in a survey addressing organizational, personal, and supervisory factors
related to turnover. ANOVA, logistical regression, and SEM were used in the
data analysis. The organizational and supervisory variables identified as sig-
nificant in the logistic regression, as in earlier research, were not significant
when the data were subjected to structural equation modeling. Instead, find-
ings suggest that career satisfaction and satisfaction with paperwork are the
key determinants of workers’ intention to stay.

Moon, S. S. (2009). Ecological influences on school achievement in a diverse


youth sample: the mediating role of substance use. Journal of Human Behavior
in the Social Environment, 19(5), 572–591.
This study’s purpose was to examine the extent to which closeness to fam-
ily, peers, and school was associated with substance use and school achieve-
ment, based on the integrative model of ecological theory, social attachment
theory, and social learning theory. A secondary data analysis was conducted
on the first wave of the National Longitudinal Study of Adolescent Health.
The final sample yielded 3,147 boys and 3,356 girls. A SEM was employed
to test a hypothesized model. School closeness was found to be a primary
ecological factor that significantly influenced school achievement while sub-
stance use emerged as a critical mediator of this relationship. Family closeness
was negatively associated with school achievement. Also, substance use did
not have a mediating function in the relationship between peer closeness and
school achievement. No gender difference was found, except the relationship
between family closeness and school achievement, in that family closeness
had a significant, direct effect on school achievement among only boys but
not girls.

Owens, T. J. (2009). Depressed mood and drinking occasions across high school:
comparing the reciprocal causal structures of a panel of boys and girls. Journal
of Adolescence, 32 (4), 763–780.
Does adolescent depressed mood portend increased or decreased drink-
ing? Is frequent drinking positively or negatively associated with emotional
well-being? Do the dynamic relations between depression and drinking differ
by gender? Using block-recursive SEMs, we explore the reciprocal short-term
effects (within time, t) and the cross-lagged medium-term effects (t + 1 year)
and long-term effects (t + 2 years) of depressed mood and monthly drinking
occasions. Data come from the high school waves of the Youth Development
148 Analysis of Multiple Dependent Variables

Study, a randomly selected panel of 1015 ninth graders followed to 12th


grade. The author found that for both genders, depressed mood consistently
decreased short-term drinking in each grade measured. However, depression
increased drinking for both genders in the medium-term but only for girls
in the long-term. In the other direction, drinking tended to increase depres-
sion in the short-term only among ninth-grade boys and 12th-grade girls.
Observed trends and differences in the magnitude of the reciprocal effects vary
by gender, with drinking being especially deleterious to emotional well-being
for boys early in high school (10th grade) but for girls on the cusp of the post-
high school world (12th grade).
6

Choosing among
Procedures for the
Analysis of Multiple
Dependent Variables

Chapters 2, 3, and 4 have discussed the following traditional or


first-generation statistical tools: MANOVA, MANCOVA, and MMR,
respectively. Chapter 5 discussed SEM, which is a more recent or sec-
ond generation addition to the statistical toolbox of social workers and
other applied researchers. In one sense, the aforementioned multivariate
procedures are variations on a single theme: the general linear model.
Nonetheless, there are differences among these procedures that make
each more appropriate given the (1) purpose of the analysis; (2) sample
size; (3) tenability of the GLM’s assumptions; and (4) type of DV system
being modeled.
This chapter (1) summarizes similarities and differences between
MANOVA, MANCOVA, MMR, and SEM, and (2) offers suggestions
to guide their differential use. Figure 6.1 compares these four multi-
variate procedures on selected criteria. Remaining discussion compares
and contrasts MANOVA and MANCOVA versus MMR, MANOVA and
MANCOVA versus SEM, and MMR versus SEM.

149
Criterion MANOVA MANCOVA MMR SEM

Research Objective Focuses on mean differences Focuses on mean differences Test models that focuses on Test models that Focus on
while controlling for other relationships between a set of individual paths between
variable that may affect these IVs and a set of DVs exogenous and endogenous
differences variables
Important Assumptions GLM GLM GLM GLM
Variables Modeled Multiple Independent (IV) Multiple Independent (IV) Multiple Independent (IV) Multiple Exogenous
Multiple Dependent (DV) Multiple Dependent (DV) Multiple Dependent (DV) Multiple Endogenous
Level of Measurement IVs = Nominal IVs = Nominal IVs = All Exogenous = All
DVs = Interval/Ratio DVs = Interval/Ratio DVs = Interval/Ratio Endogenous = All
Covariates = Interval/Ratio
Minimum Sample Size Smaller Smaller Smaller Larger
Criteria to Evaluate the Utility of the Eta Squared Eta Squared Multiple R-squared Various, including model chi
Model Tested square, CFI, RMSEA
Modeling Capabilities Path of all IVs to a linear Path of all IVs to a linear Path of all IVs to a of each DV Paths to observed and latent
combination of all DVs combination of all DVs, while endogenous variables
controlling for other variable
that may affect these differences
Type of Variance Common Common Common Common
Modeled Specific
Error
Moderating Relationships Modeled? Yes Yes Yes Yes
Mediating Relationships Modeled? No No Yes Yes
Causation Established? No No No No

Figure 6.1 Comparison of Four Multivariate Procedures on Selected Criteria.


Choosing among Procedures for the Analysis of Multiple Dependent Variables 151

MANOVA AND MANCOVA VERSUS MMR


As stated above, MANOVA shares the same underlying mathemati-
cal model with linear regression: the GLM. Neither MANOVA/
MANCOVA nor MMR make assumptions about the distributions of the
IVs. Consequently, the individual IVs do not need to be normally dis-
tributed in order to perform MANOVA/MANCOVA or a MMR. Both
MANOVA/MANCOVA and MMR assume that a model’s residuals (ε
are normally distributed random variables with the same variance (i.e.,
homoscedastic). Heteroscedasticity can be problematic in all three of the
aforementioned procedures. With MMR, strong heteroscedasticity may
cause the variance around the estimated slope and intercept to be under-
estimated (Miller, 1986), potentially leading to overestimates of statisti-
cal significance. In MANOVA/MANCOVA, heteroscedasticity alters the
assumptions underlying the F-test and may cause the p-values to be over-
or underestimated (Miller, 1986).
Although they have the same underlying mathematical framework,
MANOVA/MANCOVA and MMR are different in several important
ways. MANOVA and MANCOVA seek to answer different questions,
and, consequently, test different hypotheses compared with MMR. MMR
quantitatively describes the relationship between a DV and one or more
IVs. MANOVA and MANCOVA describe whether a set of DVs differs
among three or more levels of an IV. Regression has been applied most
often to data obtained from correlational or nonexperimental research
to describe and predict DV values on the basis of a model constructed
from the relationship between IVs and DVs. In contrast, MANOVA and
MANCOVA have been applied most frequently to experimental data.
Imprecise or inaccurate estimates of the independent variables are a par-
ticular concern for regression, which explicitly assumes that all predic-
tors are measured without error, or at least with much less error than
the response variable Y. Violation of this assumption leads to errors in
variables (EIV) and biased parameter estimates. For example, in simple
linear regression, EIV bias regression slopes toward zero (Sokal & Rohlf,
1995), potentially altering biological conclusions and complicating the
use of regression models in further research. Most important, the gen-
eral linear model assumes that the relationship between Y and X can
be described using a linear equation, so that regression is inappropri-
ate when the relationship cannot be made linear in the parameters (e.g.,
through transformations or polynomial terms).
152 Analysis of Multiple Dependent Variables

Cottingham, Lennon, and Brown (2005) argue that because linear


regression, and, consequently, MMR, requires fewer parameters to model
a relationship, it is generally a more powerful statistical approach than
MANOVA/MANCOVA. As discussed above, power is defined as the prob-
ability of detecting an effect when that effect is present (i.e., the prob-
ability of rejecting the null hypothesis when the null hypothesis is false).
Assuming constant values for sample size, alpha, and beta, the fewer the
parameters (i.e., IVs in the model), the greater the statistical power of an
analysis.
For simplicity, the following example focuses on ANOVA and regres-
sion. This example can be extended to include MANOVA/MANCOVA
and MMR. Suppose that a researcher designs a study to quantify the
effect of an intervention on the level of depression of a study’s partici-
pants by using five levels of that intervention (i.e., 1 session, 5 sessions,
10 sessions, 15 sessions, and 20 sessions). A simple linear regression to
account for the effect of the intervention would have two parameters:
intercept and slope. On the other hand, a one-way ANOVA model for the
same study would require five parameters, each specifying the mean for
an intervention level. This difference in the number of parameters grows
as the number of interventions increases. There is one important excep-
tion. When there are only two levels per factor (i.e., per IV), the power
of ANOVA is always equivalent to the power of regression because both
have the same number of parameters. Accordingly, a one-way ANOVA
with two levels of the IV has the same power as a simple (i.e., bivariate)
linear regression, and a two-way ANOVA with two levels per factor has
the same power as a multiple regression model with main effects and an
interaction term.

MANOVA AND MANCOVA VERSUS SEM


MANOVA often is recommended when a set of DVs constitutes a vari-
able system, also referred to as a factor or latent variable (e.g., Huberty
& Morris, 1989). For example, a researcher may measure a factor such
as social interaction with three indicators: time spent with friends, time
spent with family, and time spent with coworkers. An alternative way to
conduct an analysis when a set of DVs constitutes a variable system is to
compare group means with SEM (Kline, 2011). Cole, Maxwell, Arvey, and
Choosing among Procedures for the Analysis of Multiple Dependent Variables 153

Salas (1993), for example, have argued that the choice between MANOVA
and SEM should be guided by the question under investigation and by
the type of DV system being modeled.
More specifically, both SEM and MANOVA may be used to test mod-
els that link latent variables (or factors) and the empirical indicators of
those latent variables. But, the type of link between latent variables and
empirical indicators differs for the two approaches. In SEM models, the
direction is from the latent variables to the indicators; in contrast, in
MANOVA models, the direction is from the empirical indicators to the
latent variables. Following Bollen (1989) and Bollen and Lennox (1991),
in SEM, empirical indicators of a latent variable are referred to as effect
indicators, and in MANOVA, empirical indicators of a latent variable are
referred to cause indicators. It is crucial in deciding which analysis to
conduct to distinguish between DVs (i.e., exogenous variables) that are
affected by latent variables and DVs (i.e., exogenous variables) that cause
latent variables.
While it may seem reasonable to expect indicators of the same latent
variable to be positively related with each other, indicators of a latent
variable may not be related to each other. More specifically, empirical
indicators should be related to each other if they are the effects of a latent
variable. For example, to measure self-esteem, a person may be asked to
indicate whether he or she agrees or disagrees with the statements: (1) I
believe that I am a good person; and (2) I am happy with who I am. A per-
son with high self-esteem should agree with both statements; in contrast,
a person with low self-esteem would probably disagree with both state-
ments. Because each indicator depends on or is caused by self-esteem,
both of the aforementioned indicators should be positively correlated
with each other. That is, indicators that depend on the same variable
should be associated with one another if they are valid measures (Rubin,
2010). In contrast, when indicators are the cause rather than the effect
of a latent variable, these indicators may correlate positively, correlated
negatively, or be uncorrelated (Rubin, 2010). For example, gender and
race could be used as indicators of the variable “exposure to discrimina-
tion.” Being nonwhite or female increases the likelihood of experiencing
discrimination, so both are good indicators of the variable. But, race and
gender of individuals would not be expected to be strongly associated.
In summary, in SEM, empirical indicators (i.e., measured variables)
are hypothesized to be linear combinations of latent variables plus error;
154 Analysis of Multiple Dependent Variables

that is, arrows are directed from latent to empirical indicators. In SEM,
measured variables are effect indicators within a latent variable sys-
tem in that the indicators are affected by the factors (Bollen &Lennox,
1991; MacCallum & Browne, 1993). In MANOVA and MANCOVA,
researchers may hypothesize that a latent variable (e.g., program effec-
tiveness) is a linear combination of measured variables (e.g., increased
client self-efficacy, congruence between client expectations and program
performance, and client reports of program responsiveness in terms of
respectful and timely worker behaviors). Under these conditions, mea-
sured variables may be best conceptualized as cause indicators (Bollen &
Lennox. 1991; MacCallum & Browne, 1993).

MMR VERSUS SEM


First generation multivariate methods, such as MMR, are appropri-
ate for evaluating constructs and relationships between constructs. As
described earlier, MMR is an extension of OLS regression. MMR esti-
mates the same regression coefficients and standard errors that would
be obtained by using separate OLS regressions equations for each DV.
However, MMR also may be used to test an omnibus null hypothesis
and composite hypotheses for a model. The omnibus null hypothesis is
that all regression coefficients equal zero across all DVs. Tests of indi-
vidual IVs or subsets of IVs across DVs are termed composite hypotheses
(Cramer & Nicewander, 1979).
In MMR, the variables used to specify and test a model are assumed
to be measured without error, error terms (or residuals) are not intercor-
related, and variables in the model are unidirectional (does not incor-
porate feedback loops among variables). These assumptions are rarely,
if ever, met in nonexperimental social research. In MMR, the simultane-
ous evaluation of model construct relationships is not possible; evalua-
tion has to be performed in sequential steps. In contrast, SEM allows the
simultaneous analysis of all the variables in a model. In addition, SEM is
capable of testing relationships between latent and observed variables.
Given the advantages of SEM over OLS regression, when would one
ever want to use MMR? Jaccard and Wan (1996) state that regression
may be preferred to SEM when there are substantial departures from
the SEM assumptions of multivariate normality of the indicators and
Choosing among Procedures for the Analysis of Multiple Dependent Variables 155

small sample sizes, SEM requires relatively large samples. Moreover,


SEM focuses on testing overall models, whereas significance tests are of
single effects. While many of the measures used in SEM can be assessed
for significance, significance testing is less important in SEM than in
other multivariate techniques. In other techniques, significance testing
is usually conducted to establish that we can be confident that a finding
is different from the null hypothesis, or, more broadly, that an effect
can be viewed as “real.” In SEM the purpose is usually to determine if
one model conforms to the data better than an alternative model. It is
acknowledged that establishing this does not confirm “reality” as there
is always the possibility that an unexamined model may conform to the
data even better. More broadly, in SEM the focus is on the strength of
conformity of the model with the data, which is a question of associa-
tion, not significance.
In conclusion, researchers prefer to make causal inferences from ran-
domized experiments. However, often, experiments are impractical or
unethical. Consequently, a significant proportion of our social science
knowledge is derived from nonexperimental (i.e. observational) studies.
In nonexperimental investigations, judgment is required to assess the
probable impact of measured and unmeasured bias such as confounding
variables, and mathematical equations (i.e., statistical models) are fre-
quently used to adjust for confounding and other sources of bias. The
challenges of making causal statements based on nonexperimental data
are well known. However, it is not that causal statements based on nonex-
perimental data are impossible, but that they usually are difficult make.
The perspective taken here is that statistical models (i.e., multivariate
procedures) are an important design feature in observation or nonex-
perimental studies. For example, in a SEM context, even a correctly speci-
fied and tested theoretical model (e.g., one that includes all necessary
variables and paths) can fit poorly and yield highly biased estimates if the
study is poorly designed. Multivariate procedures allow social workers
and other human services researchers to analyze multidimensional social
problems and interventions in ways that minimize oversimplification.
Multivariate procedures allow researchers to use multiple operational-
ism, which refers to the use of two or more measures to represent a con-
cept. The use of multiple measures of a single concept provides a better
chance of fully representing that concept. A second reason to conduct a
multivariate analysis is to control type I error. However, there are costs
156 Analysis of Multiple Dependent Variables

associated with the use of multivariate procedures. The use of too many
DVs may reduced power or result in spurious findings due to chance.
The proper selection of an analytical strategy is a crucial part of
the research study. Which strategy is most appropriate depends on the:
(1) purpose of the analysis; (2) sample size; (3) tenability of assump-
tions; and (4) type of DV system being modeled. These three issues have
been discussed for MANOVA, MANCOVA, MMR, and SEM, which were
presented as alternative statistical procedures for analyzing models with
more than one DV.
References

Anderson, T. W. (1999). Asymptotic theory for canonical correlation analysis.


Journal of Multivariate Analysis, 70, 1–29.
Anderson, T. W. (2002). Canonical correlation analysis and reduced rank regres-
sion in autoregressive models. Annals of Statistics, 30, 1134–1154.
Babyak, M. A. (2004). What you see may not be what you get: A brief, nontech-
nical introduction to overfitting in regression-type models. Psychosomatic
Medicine, 66, 411–421.
Bagozzi, R. P., & Edwards, J. R. (1998). A general approach for representing con-
structs in organizational research. Organizational Research Methods, 1, 45–87.
Bagozzi, R. P., Fornell, C., & Larcker, D. F. (1981). Canonical correlation analy-
sis as a special case of a structural relations model. Multivariate Behavioral
Research, 16, 437–454.
Bagozzi, R. P., Yi, Y, & Singh, S. (1991). On the use of structural equation models
in experimental designs: Two extensions. International Journal of Research in
Marketing, 8, 125–140.
Baggaley, A. R. (1981). Multivariate analysis: An introduction for consumers of
behavioral research. Evaluation Review, 5, 123–131.
Barcikowski, R., & J. P. Stevens . (1975). A Monte Carlo study of the stability of
canonical correlations, canonical weights, and canonical variate-variable cor-
relations. Multivariate Behavioral Research, 10, 353–364.
Bandalos, D. L. (2002). The effects of item parceling on goodness-of-fit and
parameter estimate bias in structural equation modeling. Structural Equation
Modeling, 9, 78–102.
Bartlett, M.S. (1938). Further aspects of the theory of multiple correlation.
Proceedings of the Cambridge Philosophical Society, 34, 33–40.

157
158 References

Bartlett, M. S. (1948). Internal and external factor analysis. British Journal of


Psychology, 1, 73–81.
Bartlett, M. S. (1951). The effect of standardization on a approximation in factor
analysis. Biometrika, 38, 337–344.
Barrera, M., & Garrison-Jones, C. (1992). Family and peer social support as spe-
cific correlates of adolescent depressive symptoms. Journal of Abnormal Child
Psychology, 20(1), 1–16.
Bentler, P. M. (1980). Multivariate analysis with latent variables: Causal model-
ing. Annual Review of Psychology, 31, 419–456.
Bentler, P. M. (1990). Comparative fit indexes in structural models. Psychological
Bulletin, 107(2), 238–246.
Bentler, P. M., & Weeks, D. G. (1980). Linear structural equations with latent vari-
ables. Psychometrika, 45, 289–308.
Berg, A. J., Ingersoll, G. M., & Terry, R. L. (1985). Psychological Reports, 56(1),
115–122.
Berk, R. A. (2003). Regression analysis: A constructive critique. Thousand Oaks,
CA: Sage Publications.
Bickel, Robert. (2007). Multilevel analysis for applied research: It’s just regression!
New York: Guilford Press.
Bollen, K. A. (1989). Structural equations with latent variables. New York: Wiley.
Bollen, K., & Lennox, R. (1991). Conventional wisdom on measurement: A struc-
tural equation perspective. Psychological Bulletin, 110, 305–314.
Boomsma, A. (2000). Reporting analyses of covariance structures. Structural
Equation Modeling, 7, 461–483.
Boring, E. G. (1953). The role of theory in experimental psychology. The American
Journal of Psychology, 69(2), 169–184.
Bray, J. H., & Maxwell, S. E. (1985). Multivariate analysis of variance. Newbury
Park, CA: Sage Publications.
Breivik, E., & Olsson, U. H. (2001). Adding variables to improve fit: The effect
of model size on fit assessment in LISREL. In R. Cudeck, S. Du Toit, & D.
Sörbom (Eds.), Structural equation modeling: Present and future (pp. 169–194).
Lincolnwood, IL: Scientific Software International.
Bryk, A., S. W., Raudenbush, Seltzer, M., & Congdon, R. T. (1988). An introduction
to HLM. Chicago: University of Chicago.
Bring, J. (1994). How to standardize regression coefficients. The American
Statistician, 48(3), 209–213.
Brown, T. A. (2006), Confirmatory factor analysis for applied research. New York:
Guilford Press.
Brown, S. R., & Melamed, L. M. (1990). Experimental design and analysis.
Thousand Oaks, CA: Sage Publications.
References 159

Bruning, J. M., & Kintz, B. L. (1987). Computational handbook of statistics.


Glenview, Illinois: Scott, Foresman and Company.
Campbell, D. T., & Fiske, D. W. (1959). Convergent and discriminant validation
by the multitrait-multimethod matrix. Psychological Bulletin, 56(2), 81–103.
Cargill, B. R., Emmons, K. M., Kahler, C.W., & Brown, R. A. (2001). Relationship
among alcohol use, depression, smoking behavior, and motivation to quit smok-
ing with hospitalized smokers. Psychology of Addictive Behaviors, 15(3), 272–275.
Carroll, J. D., & Green, P. E. (1997). Mathematical tools for applied multivariate
analysis. New York: Academic Press.
Cattell, R. J. (1956). Validation and intensification of the sixteen personality fac-
tor questionnaire. Journal of Clinical Psychology, 12, 205–214.
Coanders, G., Satorra, A., & Saris, W. E. (1997). Alternative approaches to struc-
tural equation modeling of ordinal data: A Monte Carlo study. Structural
Equation Modeling, 4, 261–282.
Cohen, J., Cohen, P., West, S. G., & Aiken, L. S. (2003). Applied multiple regres-
sion/correlation analysis for the behavioral sciences. Mahwah, NJ: Erlbaum.
Cohen, J., & Cohen, P. (1983). Applied multiple regression/correlation analysis
for the behavioral sciences. Hillsdale, NJ: Erlbaum.
Cohen, J. (1988). Statistical power analysis for the behavioral sciences (2nd ed.).
Hillsdale, NJ: Erlbaum.
Cohen, J. (1994). The earth is round (p < .05). American Psychologist, 49(12),
997–102.
Cleroux, R., & Lazraq, A. (2002). Testing the signicance of the successive compo-
nents in redundancy analysis. Psychometrika, 67, 411–419.
Cliff, N. (1983). Some cautions concerning the application of causal modeling
methods. Multivariate Behavioral Research, 18, 81–105.
Cliff, N., & Krus, J. D. (1976). Interpretation of canonical analysis: Rotated vs.
unrotated solutions. Psychometrika, 41, 35–42.
Cocozzelli, C., & Constable, R T. (1985). An empirical analysis of the relation
between theory and practice in clinical social work. Journal of Social Service
Research, 9(1), 47–64.
Cohen, J., & Cohen, P. (1983). Applied multiple regression/correlation analysis for
the behavioral sciences. New Hersey: Lawrence Erlbaum Associates.
Cohen, J. (1988). Statistical power analysis for the behavioral sciences (2nd ed.).
Hillsdale, NJ: Erlbaum.
Cole, D. A., Maxwell, S. E., Arvey, R., & Salas, E. (1993). Multivariate group com-
parisons of variable systems. Psychological Bulletin, 114(1), 174–184.
Cole, D. A., & Maxwell, S. E. (1985). Multitrait-multimethod comparisons across
populations: A confirmatory factor analytic approach. Multivariate Behavioral
Research, 20, 389–417.
160 References

Cook, J. A., Razzano, L., & Cappelleri, J. C. (1996). Canonical correlation anal-
ysis of residential and vocational outcomes following psychiatric rehabilitation.
Evaluation and Program Planning, 19(4), 351–363.
Cooley, W. W., & Lohnes, P. R. (1971). Evaluation research in education. New York:
Irvington.
Cottingham, K. L., Lennon, J. T., & Brown, B. L. (2005). Knowing when to draw
the line: designing more informative ecological experiments. Frontiers in
Ecology and the Environment, 3(3), 145–152.
Cox, D. R., & Small N. J. H. (1978). Testing multivariate normality. Biometrika
65(2), 263–272.
Cramer, E. M., & Nicewander, W. A. (1979). Some symmetric, invariant measures
of multivariate association. Psychometrika, 44, 43–54.
Chrisman, N. R. (1998). Rethinking levels of measurement for cartography.
Cartography and Geographic Information Science, 25(4), 231–242.
DeCarlo, L. T. (1997). On the meaning and use of kurtosis. Psychological Methods,
2(3), 292–307.
Draper, N. R. and Smith, H. (1998). Applied Regression Analysis. New York: John
Wiley.
Draper, N. R., Guttman, I., & Lapczak, L. (1979). Actual rejection levels in a cer-
tain stepwise test. Communications in Statistics, 8, 99–105.
Duffy, M. E., Wood, R. Y., & Morris, S. (2001). The influence of demograph-
ics, functional status and comorbidity on the breast self-examination profi-
ciency of older African-American women. Journal of National Black Nurses
Association, 12(1), 1–9.
Edgeworth, F. Y. (1886). Progressive means. Journal of the Royal Statistical Society,
49, 469–475.
Enders, C. K. (2003). Performing multivariate group comparisons following a
statistically significant MANOVA. Measurement and Evaluation in Counseing
and Development, 36, 40–56.
Enders, C. K. (2010). Applied missing data analysis. New York: The Guilford
Press.
Fan, X. (1997). Canonical correlation analysis and structural equation modeling:
What do they have in common? Structural Equation Modeling, 4, 65–79.
Fan, X., Thompson, B., & Wang, L. (1999). The effects of sample size, estima-
tion methods, and model specification on SEM fit indices. Structural Equation
Modeling, 6, 56–83.
Finch, W. H. (2007). Performance of the Roy-Bargmann stepdown procedure as
a follow up to a significant MANOVA. Multiple Regression Viewpoints, 33(1),
12–22.
Finn, J. D. (1974). A general model for multivariate analysis. New York: Holt.
References 161

Finn, J. (1978). Multivariance: Univariate and multivariance analysis of variance,


covariance and regression. Chicago: National Education Resources.
Fisher, R. A. (1918). The correlation between relatives on the supposition of
Mendelian inheritance. Philosophical Transactions of the Royal Society of
Edinburgh, 52, 399–433.
Fisher, R. A. (1925). Statistical methods for research workers. Edinburgh, England:
Oliver and Boyd.
Fisher, R. A. (1935). The design of experiments. Edinburgh, England: Oliver and
Boyd.
Flores, L. Y., Navarro, R. L., & DeWitz, S. J. (2008). Mexican American high school
students’ postsecondary educational goals: Applying social cognitive career
theory. Journal of Career Assessment, 16(4), 489–501.
Fox, J. (1991). Regression diagnostics. Thousand Oaks, CA: Sage Publications.
Friedman, J. H., & Rafsky, L. C. (1979). Multivariate generalizations of the
Wald-Wolfowitz and Smirnov two sample tests. Annals of Statistics, 7,
697–717.
Galton, F. (1886). Regression toward mediocrity in hereditary stature. Journal of
the Anthropological Institute, 15, 246–263.
Gelman, A. (2008). Scaling regression inputs by dividing by two standard devia-
tions. Statistics in Medicine, 27, 2865–2873.
Gunst, R. F., & Mason, R. (1980). Regression analysis and its applications: A
data-oriented approach. New York: Marcel Dekker, Inc.
Guo, B., Perron, B. E., & Gillespie, D. F. (2009). A systematic review of structural
equation modelling in social work research. British Journal of Social Work, 39,
1556–1574.
Hair, J. F., Anderson, R. E., Tatham, R. L., & Black, W. C. (1998). Multivariate data
analysis. New York: Prentice Hall.
Hall, R. J., Snell, A. F., Foust, M. S. (1999). Item parceling strategies in SEM:
Investigating the subtle effects of unmodeled secondary constructs.
Organizational Research Methods, 2, 757–765.
Hayduk, L. A. (1996). LISREL issues, debates and strategies. Baltimore: Johns
Hopkins University Press.
Heitjan, D. (1997). Annotation: What can be done about missing data? Approaches
to imputation. American Journal of Public Health, 87, 548–550.
Henderson, M. J., Saules, K. K., & Galen, L. W. (2004). The predictive validity
of the University of Rhode Island Change Assessment Questionnaire in a
heroin-addicted polysubstance abuse sample. Psychology of Addictive Behaviors,
18(2), 106–112.
Hocking, R. R. (1976). The analysis and selection of variables in linear regression.
Biometrics, 32(1), 1–49.
162 References

Hotelling, H. (1931). The generalization of Student’s ratio. Annals of Mathematical


Statistics, 2(3), 360–378.
Hotelling, H. (1935). The most predictable criterion. Journal of Educational
Psychology, 26, 139–142.
Hotelling, H. (1936). Relations between two sets of variates. Biometrika, 28,
321–377.
Hsu, J. (1996). Multiple comparisons: Theory and methods. New York, NY:
Chapman & Hall.
Howell, D. C. (2009). Statistical methods for psychology. Belmont, CA: Wadsworth.
Hox, J. J. (2002). Multilevel analysis: Techniques and applications. Mahwah, NJ:
Lawrence Erlbaum.
Hoyle, R. H., & Panter, A. T. (1995). Writing about structural equation models.
In R. H. Hoyle (Ed.), Structural equation modeling (pp. 158–176). Thousand
Oaks, CA: Sage.
Hu, L.-T., & Bentler, P. (1995). Evaluating model fit. In R. H. Hoyle (Ed.),
Structural Equation Modeling. Concepts, Issues, and Applications (pp. 76–99).
London: Sage.
Hu, Li-tze, & Bentler, P. M. (1999). Cutoff criteria for fit indexes in covariance
structure analysis: conventional criteria versus new alternatives. Structural
Equation Modeling, 6(1), 1–55.
Huberty, C.J. (1994). Applied discriminant Analysis. New York: John Wiley and
Sons, Inc.
Hubety, C. J. (1994a). Why multivariate analyses? Educational and Psychological
Measurement, 64, 620–627.
Huberty, C., & Morris, J. D. (1989). Multivariate analysis versus multiple univari-
ate analyses. Psychological Bulletin, 105(2), 302–308.
Huberty, C. J., & Olejnik, S. (2006). Applied MANOVA and discriminant Analysis.
New Jersey: John Wiley & Sons, Inc.
Huberty, S. J., & Smith, J. D. (1982). The study of effects in MANOVA. Multivariate
Behavioral Research, 17, 417–432.
Huitema, B. E. (1980). The analysis of covariance and alternatives. New York:
Wiley.
Hunter. J. E., & Gerbing, D. W. (1982). Unidimensional measurement, second-or-
der factor analysis, and causal models. In B. M. Staw & L. L. Cummings (Eds.),
Research in organizational behavior(pp. 267–299). Greenwich, CT: JAI Press.
Jaccard J., & Guilamo-Ramos V. (2002). Analysis of variance frameworks in clini-
cal child and adolescent psychology: issues and recommendations. Journal of
Clinical Child and Adolescent Psychology, 31(1):130–46.
Jaccard, J., & Wan, C. K. (1996). LISREL analyses of interaction effects in multiple
regression. Newbury Park, CA: Sage.
References 163

Jagannathan, R., & Camasso, M. J. (1996). Risk assessment in child protective


services: A canonical analysis of the case management function. Child Abuse
and Neglect, 20(7), 599–612.
Jöreskog, K. G. (1971). Statistical analysis of sets of congeneric tests. Psychometrika,
36, 109–133.
Jöreskog, K. G. (2004). On chi-squares for the independence model and fit measures
in LISREL. Retrieved from December 27, 2010 from http://www.ssicentral.
com/lisrel/techdocs/ftb.pdf
Joreskog, K. G., & Sorbom, D. (1991). LISREL 7: A guide to the program and appli-
cations. Chicago: SPSS.
Kane, M. N., Hamlin, E. R. & Hawkins, W. E. (2002). Which clinical methods
are associated with better preparing students and practitioners for managed
care, state licensing, and other important areas? Professional Development: The
International Journal of Continuing Social Work Education, 5(2), 15–27.
Kaplan, D. (2000). Structural equation modeling: Foundations and extensions.
Thousand Oaks, CA: Sage.
Keith, T. Z. (2006). Multiple regression and beyond. New York: Pearson.
Kenny, D. A. & Judd, C. M. (1986). Consequences of violating the independence
assumption in analysis of variance. Psychiological Bulletin, 99, 422–432.
Kerlinger, F. N., & Pedhazur, E. J. (1973) Multiple regression in behavioral research.
New York: Holt, Rinehart & Winston.
Kim, J. O. (1981). Standardization in causal analysis. Sociological Methods and
Research, 10, 187–210.
Kishton, J. M., & Widaman, K. F. (1994). Unidimensional versus domain repre-
sentative parceling of questionnaire items: An empirical example. Educational
and Psychological Measurement, 54, 757–765.
Kline, R. B. (2011). Principles and practice of structural equation modeling. New
York: Guilford Press.
Klockars, A. J. and Sax, G. (1986). Multiple comparisons. Thousand Oaks, CA:
Sage Publications.
Klockars, A. J., Hancock, G. R., & McAweeney, M. J. (1995). Power of unweighted
and weighted versions of simultaneous and sequential multiple-comparison
procedures. Psychological Bulletin, 118, 300–307.
Knapp, T. R. (1978). Canonical correlation analysis: A general parametric signifi-
cance testing system. Psychological Bulletin, 85(2), 410–416.
Kowlowsky, M., & Caspy, T. (1991). Stepdown analysis of variance: A refinement.
Journal of Organizational Behavior, 12, 555–559.
Kromrey, J. D., & La Rocca, M. A. (1995). Power and Type I error rates of new
pairwise multiple comparison procedures under heterogeneous variances.
Journal of Experimental Education, 63, 343–362.
164 References

Larsen, J. J., & Juhasz, A. M. (1985). The effects of knowledge of child develop-
ment and social-emotional maturity on adolescent attitudes toward parenting.
Adolescence, 20(80), 823–39.
Leary, M. R., & Altmaier, E. M. (1980). Type I error in counseling research: A plea
for multivariate analyses. Journal of Counseling Psychology, 27, 611–615.
Loehlin, J. C. (2004). Latent variable models: An introduction to factor, path, and
structural equation analysis. Mahwah, NJ: Erlbaum.
Lorenz, Frederick O. (1987). Teaching about influence in simple regression.
Teaching Sociology, 15(2), 173–177.
Little, R. J., & Rubin, D. B. (2002). Statistical analysis with missing data. New York:
John Wiley & Sons.
Little, T. D., Cunningham, W. A., Shahar, G., & Widaman, K. F. (2002). To par-
cel or not to parcel: Exploring the question, weighing the merits. Structural
Equation Modeling, 9, 151–173.
Loehlin, J. C. (2004). Latent variable models: An introduction to factor, path, and
structural equation analysis. Mahwah, NJ: Erlbaum.
Looney, S. W. (1995). How to use tests for univariate normality to assess multi-
variate normality. The American Statistician, 49(1), 64–70.
Lorenz, F. O. (1987). Teaching influence in simple regression. Teaching Sociology,
15(2), 173–177.
Lynch, S. M., & Graham-Bermann, S. A. (2004). Exploring the relationship
between positive work experiences and women’s sense of self in the context of
partner abuse. Psychology of Women Quarterly, 28(2), 159–167.
MacCallum, R. C., & Austin, J. T. (2000). Applications of structural equation mod-
eling in psychological research. Annual Review of Psychology, 51, 201–226.
MacCallum, R. C., & Browne, M. W. (1993). The use of causal indicators in covariance
structure models: Some practical issues. Psychological Bulletin, 114, 533–541.
MacCallum, R. C., Wegener, D. T., Uchino, B. N., & Fabrigar L. R. (1993). The
problem of equivalent models in applications of covariance structure analysis.
Psychological Bulletin, 114, 185–99.
MacCallum, R. C., Widaman, K. F., Zhang, S., & Hong, S. (1999). Sample size in
factor analysis. Psychological Methods, 4, 84–99.
Mallows, C. L. (1973). Some comments on CP. Technometrics, 15(4), 661–675.
Marsh, H. W., & O’Neill, R. (1984). Self Description Questionnaire III: The con-
struct validity of multidimensional self-concept ratings by late adolescents.
Journal of Educational Measurement, 21, 153–174.
Martens, M. P. (2005). Future directions of structural equation modeling in
counseling psychology. The Counseling Psychologist, 33(3), 375–382.
Menard, S. (1995). Applied logistic regression analysis. Thousand Oaks, CA: Sage
Publications.
References 165

Mardia, K. V. (1970). Measures of multivariate skewness and kurtosis with appli-


cations. Biomnetrika, 57, 519–530.
Maxwell, S. E., & Delaney, H. D. (2003). Designing experiments and analyzing
data: A model comparison perspective. New York: Routledge.
McDonald, R. P. (1982). A note on the investigation of local and global identifi-
ability. Psychometrika, 47(1), 101–103.
McDonald, R. P., & Ho, M. H. R. (2002). Principles and practice in reporting
structural equation analyses. Psychological Methods, 7, 64–82.
Miller, R. G. (1986). Beyond ANOVA: basics of applied statistics. New York, NY:
Wiley Press.
Miville, M. L., & Constantine, M. G. (2006). Sociocultural predictors of psy-
chological help-seeking attitudes and behavior among Mexican American
college students. Cultural Diversity and Ethnic Minority Psychology, 12(3),
420–432.
Miville, M. L., Darlington, P., Whitlock, B., & Mulligan, T. (2005). Integrating
identities: The relationships of racial, gender, and ego identities among white
college students. Journal of College Student Development, 46(2), 157–175.
Mudholkar, G. S., & Subbaiah, P. (1980). A review of step-down procedures for
Multivariate Analysis of Variance. In R. P. Gupta, (Ed.), Multivariate Statistical
Analysis (pp.161–178). Amsterdam: North-Holland.
Mudholkar, G. S., & Subbaiah, P. (1988). On a fisherian detour of the step-down
procedure for MANOVA. Communications in Statistics: Theory and Methods,
17, 599–611
Mulaik, S. A. (2009) Linear causal modeling with structural equations. New York:
Chapman & Hall.
Muller, K. E. (1981). Relationships between redundancy analysis, canonical cor-
relation, and multivariate regression. Psychometrika, 46, 139–142.
Nasser, F., & Takahashi, T. (2003). The effect of using item parcels on ad hoc
goodness-of-fit indexes in confirmatory factor analysis: An example using
Sarason’s Reactions to Tests. Applied Measurement in Education, 16, 75–97.
Neyman, J. (1952). Lectures and conferences on mathematical statistics and prob-
ability. Washington, DC: U.S. Department of Agriculture.
Nunnally, J. C., & Bernstein, I. H. (1994). Psychometric theory. New York:
McGraw-Hill.
Olsen, C. L. (1974). Comparative robustness of six tests in multivariate analysis of
variance. Journal of the American Statistical Association, 69, 894–908.
Olsson, U. H., Foss, T., & Breivik, E. (2004). Two equivalent discrepancy functions
for maximum likelihood estimation: Do their test statistics follow a non-central
chi-square distribution under model misspecification? Sociological Methods
and Research, 32, 453–500.
166 References

Pearson, K. (1896). Regression, heredity and panmixia. Philosophical Transactions


of the Royal Society, 187, 253–267.
Pearson, K. (1908). On the generalized probable error in multiple normal cor-
relation. Biometrika, 6, 59–68.
Pedhazur, E. (1982). Multiple regression in behavioral research. New York: Holt,
Rinehart and Winston.
Penny, K. I. (1996). Appropriate critical values when testing for a single multivari-
ate outlier by using the Mahalanobis distance. Applied Statistics, 45, 73–81.
Pierce, C. A., Block, R. A., & Agunis, H. (2004). Cautionary note on report-
ing eta-squared values from multifactor ANOVA designs. Educational and
Psychological Measurement, 64(6), 916–924.
Pierre, M. R., & Mahalik J. R. (2005). Examining African self-consciousness and
black racial identity as predictors of black men’s psychological well-being.
Cultural Diversity & Ethnic Minority Psychology, 11(1), 28–40.
Quintana, S. M., & Maxwell, S. E. (1999). Implications of recent developments
in structural equation modeling for counseling psychology. The Counseling
Psychologist, 27, 485–527.
Ramsey, J. B. (1969) Tests for specification errors in classical linear least squares
regression analysis. Journal of the Royal Statistical Society B, 31(2), 350–371.
Rencher, A. C. (1998), Multivariate statistical inference and applications. New
York: Wiley.
Rencher, A. C. (1992). Interpretation of canonical discriminant functions, canonical
variates, and principal components. The American Statistician, 46(3), 217–225.
Reynolds, A. L., & Constantine, M. G. (2007). Cultural adjustment difficulties
and career development of international college students. Journal of Career
Assessment, 15(3), 338–350.
Rindfleisch, N., & Foulk, R. C. (1992). Factors that influence the occurrence and
the seriousness of adverse incidents in residential facilities. Journal of Social
Service Research, 16(3/4), 65–87.
Roy, J. (1958). Step-down procedure in multivariate analysis. Annals of
Mathematical statistics, 29(4), 1177–1187.
Roy, S. N., & Bargmann, R. . (1958). Tests of multiple independence and the asso-
ciated confidence bounds. The Annals of Mathematical Statistics, 29, 491–503.
Rubin, E. R. (2010). The practice of social research. Belmont, CA: Wadsworth.
Rubin, A. (2007). Statistics for evidence-based practice and evaluation. Belmont,
CA: Thompson Brooks/Cole.
Rubin, D. B. (1987). Multiple imputation for nonresponse in surveys. New York:
John Wiley & Sons.
Rubin, A., & Babbie, E. R. (2010). Research methods for social work. Belmont, CA:
Thompson Brooks/Cole.
References 167

Satorra, A., & Bentler, P. (1994). Corrections to test statistics and standard
errors in covariance structure analysis. In A. von Eye and C.C. Clogg (eds.),
Latent variable analysis: Applications to developmental research (pp. 399–419).
Newbury Park: Sage.
Schau, C., Stevens, J., Dauphinee, T. L., & Del Vecchio, A. (1995). The develop-
ment and validation of the Survey of Attitudes toward Statistics. Educational
and Psychological Measurement, 55, 868–875.
Scheffé, H. (1953). A method for judging all contrasts in the analysis of variance.
Biometrika, 40, 87–104.
Schmidt, F. L., & Hunter, J. E. (1997). Eight common but false objections to
the discontinuation of significance testing in the analysis of research data.
In L. L. Harlow, S. A. Mulaik, & J. H. Steiger (Eds.), What if there were no
significance tests? (pp. 37–64). Mahwah, NJ: Erlbaum.
Schuster, C. (1998). Regression analysis for social sciences. New York: Academic
Press.
Seaman, M. A., Levin, J. R., & Serlin, R. C. (1991). New developments in pair-
wise multiple comparisons: Some powerful and practicable procedures.
Psychological Bulletin, 110, 577–586.
Searle, S. (1987). Linear models for unbalanced data. New York: John Wiley &
Sons, Inc.
Shapiro, S. S., & Wilk, M. B. (1965). An analysis of variance test for normality
(complete samples). Biometrika, 52, 591–611.
Shevlin, M., Miles, J. N. V., & Bunting B. P. (1997). Summated rating scales: A
Monte Carlo investigation of the effects of reliability and collinearity in regres-
sion models. Personality and Individual Differences, 23, 665–676.
Sokal, R. R., & Rohlf, F. J. (1995). Biometry. New York, NY: WH Freeman.
Small, N. J. H. (1980). Marginal skewness and kurtosis in testing multivariate
normality. Applied Statistics, 29, 85–87.
Smith, S. P., & Jain, A. K. (1988). A test to determine the multivariate normality
of a dataset. IEEE Transactions on Pattern Analysis and Machine Intelligence,
10(5), 757–761.
Smithson, M. (2003). Confidence intervals. Thousand Oaks, CA: Sage.
Sorbom, D. (1981). Structural equation models with structured means. In K.
G. Joreskog & H. Wolds, (Eds.), Systems under indirect observation: Causality,
structure and prediction (pp. 23–69). Amsterdam: North-Holland.
Srivistava, M. S. (1984). A measure of skewness and kurtosis and a graphical
method for assessing multivariate normality. Statistics and Probability Letters,
2, 263–276.
Steiger, J. H. (1990). Structural model evaluation and modification: An interval
estimation approach. Multivariate Behavioral Research, 25, 173–180.
168 References

Stevens, J. P. (1972). Four methods of analyzing between variation for the K-group
MANOVA problem. Multivariate Behavioral Research, 7, 499–522.
Stevens, J. P. (1973). Step-down analysis and simultaneous confidence intervals in
MANOVA. Multivariate Behavioral Research, 8(3), 391–402.
Stevens, J. P. (1980). Power of the multivariate analysis of variance tests.
Psychological Bulletin, 88(3), 728–737.
Stevens, J. (1996). Applied multivariate statistics for the social sciences. New York:
Routledge.
Stevens, J. (2002). Applied multivariate statistics for the social sciences. New York:
Routledge.
Stevens, J. (2009). Applied multivariate statistics for the social sciences. New York:
Routledge.
Stevens, S. S. (1946). On the theory of scales of measurement. Science, 103,
677–680.
Stevens, S.S. (1951). Mathematics, measurement and psychophysics. In S. S.
Stevens (Ed.), Handbook of experimental psychology (pp. 1–49). New York:
Wiley.
Student. (1908). The probable error of a mean. Biometrika, 6, 1–25.
Subbaiah, P., & Mudholkar, G. S. (1978). A comparison of two tests for the sig-
nificance of a mean vector. Journal of the American Statistical Association, 73,
414–418.
Tabachnick, B. G., & Fidell, L. S. (2007). Using multivariate statistics. Boston:
Allyn and Bacon.
Takane, Y., & Hwang, H. (2005). On a test of dimensionality in redundancy analy-
sis. Psychometrika, 70(2), 271–281.
Tang, T., & Kim, J. K. (1999). The meaning of money among mental health work-
ers: The endorsement of money ethic as related to organizational citizenship,
behavior, job satisfaction, and commitment. Public Personnel Management,
28(1), 15–26.
Thompson, B. (1984). Canonical correlation analysis: Use and interpretation.
Beverly Hills, CA: Sage.
Thompson, B. (1991). A primer on the logic and use on canonical correlation
analysis. Measurement and Evaluation in Counseling and Development, 24(2),
80–95.
Timm, N. H. (1975). Multivariate analysis with applications in education and psy-
chology. Monterey, California: Brookes-Cole.
Tintner, G. (1950). Some formal relations in multivariate analysis. Journal of the
Royal Statistical Society, Series B (Methodological), 12, 95–101.
Tomarken, A. J., & Waller, N. G. (2005). Structural equation modeling: Strengths,
limitations, and misconceptions. Annual Review of Clinical Psychology, 1,
31–65.
References 169

Toothacker, L. E. (1993). Multiple comparisons procedures. Thousand Oaks, CA:


Sage Publications.
Tukey, J. W. (1953). The problem of multiple comparisons. Unpublished manu-
script, Princeton University: Princeton, N. J.
van den Wollenberg, A. L. (1977). Redundancy analysis: An alternative for canon-
ical analysis. Psychometrika, 42, 207–219.
Velleman, P. F., & Wilkinson, L. (1993). Nominal, ordinal, interval, and ratio
typologies are misleading. The American Statistician, 47(1), 65–72.
Warner, R. M. (2008). Applied statistics. Los Angeles, CA: Sage Publications.
Widaman, K. F., & Thompson, J. S. (2003). On specifying the null model for
incremental fit indexes in structural equation modeling. Psychological Methods,
8, 16–37.
Wilcox, R. R. (1987). New designs in analysis of variance. Annual Review of
Psychology, 38, 29–60.
Wilks, S. S. (1932). Certain generalizations in the analysis of variance. Biometrika,
24, 471–494.
Winer, B. J. (1971). Statistical principles in experimental designs. New York:
McGraw-Hill.
Wright, B. D. (1999). Fundamental measurement for psychology. In S. E.
Embretson & S. I. Hershberger, (Eds.). The new rules of measurement: What
every psychologist and educator should know (pp. 65–104). Mahwah, NJ:
Erlbaum.
Yuan, K.-H. (2005). Fit indices versus test statistics. Multivariate Behavioral
Research, 40, 115–148.
Yule, G. U. (1907). On the theory of correlation for any number of variables treated
by a new system of notation. Proceedings of the Royal Society, 79, 182–193.
This page intentionally left blank
Index

Abel, E. M., 85–86 Applied discriminant Analysis (Huberty),


absence of multicollinearity, 14 1–2
additive inequality, 4 Applied logistic regression analysis
additivity, 6 (Menard), 91
adjusted R2, 90 Applied multivariate statistics for the social
“Adolescent gender differences in social sciences (Stevens, J.), 9, 15, 39, 44, 88
support: structure, function, and a priori comparisons, 34
provider type” (Colarossi), 60 multiple ANOVAs with, 35
“Alcohol expectancies and excessive Arvey, R., 152–53
drinking contexts in young adults” Ashford, J. A., 84
(O’Hare), 61 “Asymptotic theory for canonical
all-possible-regressions, 90–93 correlation analysis” (Anderson),
AMOS outputs, 137–42, 137f, 138f, 139f, 97–98
140f, 141f, 142f, 143f Auerbach, C., 146–47
analysis of covariance (ANCOVA), 63
The analysis of covariance and alternatives Babbie, E. R., 13
(Huitema), 65–66 backward elimination, description of, 89
analysis of covariance structures. See Barcikowski, R., 98
structural equation modeling Bargmann, R., 38, 88
analysis of variance (ANOVA), 25–27 Barrera, M., 105
description of, 19, 22 Benda, B. B., 145–46
with post hoc comparisons, 35–38 Bentler, P., 127, 129
with a priori comparisons, 35 Bernstein, I. H., 122
“An analysis of variance test for normality best-subsets regression, 90
(complete samples)” (Shapiro and beta weights, definition of, 94
Wilk), 15–16 between-equation covariances, 97–98
ANCOVA. See analysis of covariance Bischofshausen, S., 86
Anderson, T. W., 97–98 bivariate normality, 15
ANOVA. See analysis of variance Bollen, K. A., 114, 153

171
172 Index

Bonferroni correction, 16 definition of, 93, 97, 154


Bonferroni method, 41–42 testing, 97–98
Bonferroni procedure, description of, 35–36 compound hypothesis, 40
Boring, E. G., 3 confidence intervals (CIs), 40–41, 56,
Bray, J. H., 4 79–80, 80t
Brown, B. L., 152 Conroy-Moore, K., 86
Brown, R. A., 105–6 “Consequences of violating the
Browne, M. W., 133–35, 134f independence assumption in analysis
“Bullying and victimization among native of variance” (Kenny and Judd), 8
and immigrant adolescents in Constantine, M. G., 107–8
Norway: The role of proactive and contextual effect, 9–10
reactive aggressiveness” (Fandrem, “Conventional wisdom on measurement:
Strohmeier, and Roland), 146 A structural equation perspective”
Bybee, D., 61 (Bollen and Lennox), 153
“Convergent and discriminant validation
Campbell, D. T., 3 by the multitrait-multimethod
canonical correlation analysis (CCA), 97 matrix” (Campbell and Fisk), 3
“Canonical correlation analysis and Cook’s D, 17
reduced rank regression in correction for parsimony, 127
autoregressive models” (Anderson), “The correlation between relatives on
97–98 the supposition of Mendelian
Cargill, B. R., 105–6 inheritance” (Fisher), 26
Caspy, T., 39 Cottingham, K. L., 152
CCA. See canonical correlation analysis covariances, definition of, 111
“Certain generalizations in the analysis of Cronbach’s alpha, 43–44
variance” (Wilks), 28 cross-product terms, added as IVs, 7, 94
CFI. See Comparative Fit Index “Cultural adjustment difficulties and
“Child protective service relationships on career development of international
parental attitudes in the juvenile college students” (Reynolds and
dependency process” (Ashford), 84 Constantine), 107–8
chi-square test, 124–25, 125f “Cutoff criteria for fit indexes in
CIs. See confidence intervals covariance structure analysis:
Cohen, J., 42 conventional criteria versus new
Colarossi, L. G., 60 alternatives” (Hu and Bentler), 129
Cole, D. A., 152–53
Comparative Fit Index (CFI), 126–27 Darlington, P., 107
Comparative robustness of six tests in data
multivariate analysis of variance deletion techniques for missing, 11
(Olsen), 15 operationalized concepts influencing
“Comparing the family environments of analysis of, 11
adolescents with conduct disorder or statistical models missing, 10–11
depression” (Jewell and Stark), 60 stepwise procedures and missing, 92
“A comparison of two tests for the DDA. See descriptive discriminant analysis
significance of a mean vector” DeCarlo, L. T., 16, 45–46, 67–68
(Mudholkar and Subbaiah), 39–40 degrees of freedom (df), 23–24, 115
comparison-wise error rate, definition of, 4 Delaney, H. D., 41
complex comparison, 37 deletion techniques, for missing data, 11
component hypotheses, 40 Dennis, K., 86
composite hypotheses, 87–88 dependent variables (DVs), 1
Index 173

conceptually related, 2 EIV. See errors in variables


GLM extended to include, 8 emergent variables. See latent variables
multivariate procedures for analysis of, Emmons, K. M., 105–6
149, 150f empirical identification, 117–19
OLS regression models and, 88–89 Enders, C. K., 4
statistically related, 2–3 endogenous variables, 111
“Depressed mood and drinking occasions errors in variables (EIV), 151
across high school: comparing the estimation procedures, 120
reciprocal causal structures of a panel exogenous variables, 111
of boys and girls” (Owens), 147–48 experiment-wise error rate. See study-wise
descriptive discriminant analysis (DDA), error rate
34–35 “Exploring the relationship between
Designing experiments and analyzing data: positive work experiences and
A model comparison perspective women’s sense of self in the context
(Maxwell and Delaney), 41 of partner abuse” (Lynch and
determinant, 29–30, 30f Graham-Bermann), 106–7
DeWitz, S. J., 106
df. See degrees of freedom factor analysis, 109
DiNitto, Diana M., 84–85 “Family and peer social support as specific
direct effects, 111 correlates of adolescent depressive
discrepancy, 17 symptoms” (Barrera and Garrison-
functions, 112 Jones), 105
discriminant functions, 34–35 family-wise error rate, definition of, 25
DVs. See dependent variables family-wise D-level, 32–33
Fandrem, H., 146
“Ecological influences on school Farooqi, A., 85
achievement in a diverse youth F-distribution, 26–27, 27f
sample: the mediating role of Fidell, L. S., 88
substance use” (Moon), 147 Finch, W. H., 40
“The Effectiveness of an Integrated Finn, J., 95
Treatment Approach for Clients Fisher, R. A., 26, 121
With Dual Diagnoses” (DiNitto, Fiske, D. W., 3
Webb, and Rubin), 84–85 fitting, 6
“The effectiveness of a psychoeducational fixed effect
group for HIV-infected/affected definition of, 88
incarcerated women” (Pomeroy, description of, 88
Kiam, and Abel), 85–86 Flores, L. Y., 106
“The effectiveness of EMDR in a forward selection, description of, 89
child guidance center” (Rubin, “Four methods of analyzing between
Bischofshausen, Conroy-Moore, variation for the K-group MANOVA
Dennis, Hastie, Melnick, Reeves, and problem” (Stevens, J. P.), 38, 40
Smith), 86 F-ratio, 26, 27
“The effect of standardization on a F-tests, 26–27, 96–97
approximation in factor analysis” full structural model, 112–13
(Bartlett), 87 “Further aspects of the theory of multiple
effect size correlation” (Bartlett), 87
definition of, 33
description of, 33 Galen, L. W., 106
measures of, 33 Garrison-Jones, C., 105
174 Index

Gelman, A., 94 of multicollinearity, 14


“The generalization of Student’s ratio” structural, 114–17, 115f, 116f
(Hotelling), 87 imputation, 11
general linear model (GLM) independence of observations, 8–10
assumptions of, 8 independent samples, 23
description of, 7–8 independent variables (IVs)
DVs included with, 8 cross-product terms added as, 7, 94
GFI. See goodness-of-fit index description of, 1
GLM. See general linear model OLS regression models and, 89–93
goodness-of-fit, 97 power terms as added, 7
goodness-of-fit index (GFI), 126 indicator variable, 111
Gosset, William. See Student indirect effects, 112
Gothefors, L., 85 influence, 17
GPower, 42, 64–65 “Integrating identities: The relationships
sample size determined by, 45, 67 of racial, gender, and ego identities
Graham-Bermann, S. A., 106–7 among white college students”
(Miville, Darlington, Whitlock, and
Hägglöf, B., 85 Mulligan), 107
Hardman, L., 84 interaction effects, 94
Hastie, M., 86 interactions. See higher-order variables
Henderson, M. J., 106 item parcels, 110–11
heteroscedasticity, 18 IVs. See independent variables
heuristics, 115
hierarchical data structures, 9 Jaccard, J., 154–55
hierarchical multiple regressions (HMR), Jewell, J. D., 60
87–88 Jones, R., 84
description of, 93 Judd, C. M., 8
higher-order variables, lower-order Jung, J. S., 146
variables, constructed from, 6
HMR. See hierarchical multiple regressions Kahler, C. W., 105–6
homogeneity of regressions, 51–53, 52f, 53f Kaplan, D., 124
description of, 64, 73 Kenny, D. A., 8
prescreening for, 73–76, 75f, 76f Kiam, R., 85–86
homoscedasticity Kim, J. K., 108
definition of, 18 Kline, R. B., 114, 124, 128
description of, 18 “Knowing when to draw the line:
prescreening for, 49–50, 72–73, 73f designing more informative
Hotelling’s T2, 28 ecological experiments”
Howell, D. C., 34 (Cottingham, Lennon, and Brown,
“How to use tests for univariate normality B. L.), 152
to assess multivariate normality” Kowlowsky, M., 39
(Looney), 15 kurtosis, definition of, 14
Hu, Li-tze, 127, 129 kurtosis statistic b2, 16
Huberty, C. J., 1–2
Huitema, B. E., 65–66 Lagrange multiplier test. See score test
latent variables, 1–2
identification definition of, 111
description of, 113–14 empirical indicators of, 109
empirical, 117–19 scaling, 117–18
Index 175

SEM used to test relationships “The meaning of money among mental


between, 109 health workers: The endorsement
Lennon, J. T., 152 of money ethic as related to
Lennox, R., 153 organizational citizenship, behavior,
leptokurtosis, definition of, 14–15 job satisfaction, and commitment”
levels of measurement, 12 (Tang and Kim), 108
leverage, 17 measurement
Linear causal modeling with structural error, 13
equations (Mulaik), 124 levels of, 12
linearity, 18 models, 109–10, 117–19
line-of-best-fit, 94 nominal level of, 12
LISREL analyses of interaction effects in ordinal level of, 12
multiple regression (Jaccard and random error of, 13
Wan), 154–55 reliability, 13
Looney, S. W., 15 in statistical models, 11–13
Lorenz, F. O., 17 systematic error of, 13
lower-order variables, higher-order validity, 13
variables constructed from, 6 “A measure of skewness and kurtosis and
Lynch, S. M., 106–7 a graphical method for assessing
multivariate normality” (Srivistava),
MacCallum, R. C., 133–35, 134f 45–46, 68–69
MacFarlane, P., 61 “Measures of multivariate skewness and
Mahalanobis distances, 16 kurtosis with applications” (Mardia),
main effects. See lower-order variables 45–46, 68–69
Mallows Cp, 90–91 mediation, 112
MANCOVA. See multivariate analysis of Melnick, L., 86
covariance Menard, S., 91
MANOVA. See multivariate analysis of “Mental health and social competencies
variance of 10- to 12-year-old children
manova command, 99–100, 100f born at 23 to 25 weeks of gestation
MAR. See missing at random in the 1990s: A Swedish national
Mardia, K. V., 45–46, 68–69 prospective follow-up study”
“Marginal skewness and kurtosis in testing (Farooqi, Hägglöf, Sedin, Gothefors,
multivariate normality” (Small), and Serenius), 85
45–46, 68–69 “Mexican American high school students’
“Mathematics, measurement and postsecondary educational goals:
psychophysics” (Stevens, S.), 12 Applying social cognitive career
matrices theory” (Flores, Navarro, and
determinant of, 29–30, 30f DeWitz), 106
error, 30–31 Minitab, 90
hypothesis, 30–31 missing at random (MAR), 10–11
m × n, 29 missing completely at random (MCAR),
total, 30–31 10–11
traces of, 31–32 missing not at random (MNAR), 10–11
variance-covariance, 30–31 misspecification, 6
maximum likelihood (ML), 121–23 Miville, M. L., 107
Maxwell, S. E., 4, 41, 152–53 mixed effects, definition of, 88
MCAR. See missing completely at random ML. See maximum likelihood
McGowan, B. G., 146–47 MMR. See multivariate multiple regression
176 Index

MNAR. See missing not at random SEM vs., 152–54


model building, definition of, 6, 88 strengths and limitations of, 65–66
model fit, 113, 123–29, 125f Multivariate analysis of variance (Bray and
model modification, 129–30 Maxwell), 4
moderation, 112. See also statistical multivariate analysis of variance
interaction (MANOVA)
moderator and mediator relationships, 112 assumptions of, 32
moderators. See higher-order variables cause indicators in, 153, 154
modification indices, 129 definition of, 29
“A Monte Carlo study of the stability of description of, 19, 22, 28–32
canonical correlations, canonical examples of, 43–62, 46f, 49f, 50f, 52f,
weights, and canonical variate- 53f, 54f, 55f, 56t, 59t
variable correlations” (Barcikowski factorial, 29
and Stevens), 98 limitations of, 43
Moon, S. S., 147 MMR vs., 151–52
Mowbray, C., 61 model validation in, 41
Mudholkar , G. S., 39–40 one-way, 28
Mulaik, S. A., 124 outliers influencing, 17
Mulligan, T., 107 performing, 54–57, 54f, 55f, 56t, 77–81,
multicollinearity, 14 78f, 80t
%MULTINORM macro, 69–70 procedure, 32
Multiple comparisons procedures results of, 58–60, 59t
(Toothacker), 37 SEM vs., 152–54
multiple imputation, description of, 11 strengths of, 43
multiple operationalism, 3 two-group, 28
“Multisystemic therapy as a community- two-way between-subjects factorial, 29
based treatment for youth with types of, 28–29
severe emotional disturbance” “Multivariate group comparisons of
(Painter), 85 variable systems” (Cole, Maxwell,
Multivariance: Univariate and Arvey, and Salas), 152–53
multivariance analysis of variance, multivariate group differences, 33–41
covariance and regression (Finn), 95 multivariate multiple regression (MMR)
multivariate analyses between-equation covariances
for description, 2 estimated by, 97–98
description of, 1–2 description of, 19–20, 87–93
examples of, 1 examples of, 99–108, 100f, 101f, 102f,
for prediction, 2 103f
purposes of, 1 goodness-of-fit in, 97
rationale for, 2–5 limitations of, 99
multivariate analysis of covariance MANCOVA vs., 151–52
(MANCOVA) MANOVA vs., 151–52
assumptions of, 64 model validation, 98
description of, 19, 63 multivariate regression parameters for,
examples of, 63–64, 66–86, 68f, 71f, 73f, estimating, 93–96
75f, 76f, 78f, 80t, 83t OLS regression distinguished from, 96
MMR vs., 151–52 overview, 87–93
performing, 80 performing, 102–3
results of, 81–84, 83t procedure, 93–98
sample size requirements, 64–65 results of, 103–5
Index 177

sample size requirements of, 98 multiple, 3


SEM vs., 154–56 Order Condition Test, 116
strengths of, 99 ordinal level of measurement, 12
multivariate normality, 14–16 ordinary least squares (OLS) regression
prescreening for, 45–48, 46f, 67–70, 68f description of, 7
multivariate procedures DVs and models of, 88–89
description of, 1, 155–56 MMR distinguished from, 96
for multiple DVs, 149, 150f models, 88–93
multivariate statistical procedures, refinement, 88–93
examples of, 2 orthogonality, 14
multivariate variance, definition of, 23 “Outcome for adjustment disorder with
mvreg command, 100–101, 101f depressed mood: Comparison with
mvtest, 69 other mood disorders” (Jonesa,
Yatesb, Williamsc, Zhouc, and
Navarro, R. L., 106 Hardmand), 84
nominal level of measurement, 12 outliers
nonindependence of observations, 8–9 absence of, 16–18, 48–49, 70–72, 71f
nonlinear transformation, example of, 18 MANOVA influenced by, 17
nonrecursive relationship, 111 overfitting, 6
NORM, 11 Owens, T. J., 147–48
normality Oyserman, D., 61
bivariate, 15
multivariate, 14–16 Painter, K., 85
Null B Rule, 116 paired samples, example of, 23
Nunnally, J. C., 122 pair-wise comparisons, 25
“Parenting of mothers with a serious
observations mental illness: differential effects
independence of, 8–10 of diagnosis, clinical history, and
nonindependence of, 8–9 other mental health variables”
O’Hare, T., 61 (Mowbray, Oyserman, Bybee, and
Olsen, C. L., 15 MacFarlane), 61
OLS regression. See ordinary least squares partial eta-squared, definition of, 33
regression path analysis, 109
omnibus hypothesis test, study-wise alpha “Performance of the Roy-Bargmann
level and, 96 stepdown procedure as a follow
omnibus null hypothesis, 87, 154 up to a significant MANOVA”
evaluating, 32–33 (Finch), 40
testing, 96–97 “Performing multivariate group
“On specifying the null model for comparisons following a
incremental fit indexes in structural statistically significant MANOVA”
equation modeling” (Widaman and (Enders), 4
Thompson), 127 “Personality and gender as predictors of
“On the meaning and use of kurtosis” online counseling use” (Tsan), 62
(DeCarlo), 16, 45–46, 67–68, 68f platykurtosis, definition of, 15
“On the theory of scales of measurement” Pomeroy, E. C., 85–86
(Stevens, S.), 12 post hoc tests, 27, 34. See also specific post
operationalism hoc tests
data analysis influenced by, 11 multiple ANOVAs with, 35–38
definition of, 3 power terms, as added IVs, 7
178 Index

“The predictive validity of the University “The role of theory in experimental


of Rhode Island Change Assessment psychology” (Boring), 3
Questionnaire in a heroin-addicted root-mean-square of approximation
polysubstance abuse sample” (RMSEA), 127–28
(Henderson, Saules, and Galen), 106 Roy, S. N., 38, 88
Principles and practice of structural Roy-Bargmann stepdown analysis, 52f,
equation modeling (Kline), 114, 124, 54–56, 55f
128 performing, 57–58, 77–79, 78f, 80
probability pyramiding, 4, 92 Rubin, A., 11, 13, 84–85, 86
“The probable error of a mean” (Student), 23
Psychometric theory (Nunnally and Salas, E., 152–53
Bernstein), 122 samples
GPower determining size of, 45, 67
Ramsey’s Regression Specification Error independent, 23
Test (RESET), 10 paired, 23
Randall, C. L., 61–62 size requirements, 41–42, 45, 64–65, 67,
random effect, description of, 88 98, 133–43, 134f, 136f, 137f, 138f,
random error, description of, 13 139f, 140f, 141f, 142f, 143f
Rank Condition test, 116 Saules, K. K., 106
recursive relationship, 111, 112f scale-free least squares estimation
Recursive Rule, 116 (SFLS), 121
Reeves, D., 86 “Scaling regression inputs by dividing by two
regress command, 101, 101f standard deviations” (Gelman), 94
“Relationship among alcohol use, Scheffé test, 37–38
depression, smoking behavior, and score test, 16
motivation to quit smoking with SDA. See stepdown analysis
hospitalized smokers” (Cargill, Sedin, G., 85
Emmons, Kahler, and Brown, R. A.), SEM. See structural equation modeling
105–6 Serenius, F., 85
“The relationships of flexible work SFLS. See scale-free least squares
schedules, workplace support, estimation
supervisory support, work-life Shapiro, S. S., 15
balance, and the well-being of simple contrast, 37
working parents” (Jung), 146 single imputation, description of, 11
reliability of measurement, definition singularity, definition of, 14
of, 13 skew
Research methods for social work (Rubin definition of, 14
and Babbie), 13 multivariate, 45–46, 46f, 68–69, 68f
RESET. See Ramsey’s Regression skewness statistic —b1, 16
Specification Error Test slope, 94
residual sum-of-squares, 26 Small , N. J. H., 45–46, 68–69
residual value, 7 Smith, T., 86
“A review of step-down procedures for “Social support in alcohol dependence
Multivariate Analysis of Variance” and social phobia: Treatment
(Mudholkar and Subbaiah), 39 comparisons” (Thevos, Thomas, and
Reynolds, A. L., 107–8 Randall), 61–62
RMSEA. See root-mean-square of social work, 5–6
approximation “Sociocultural predictors of psychological
Roland, E., 146 help-seeking attitudes and behavior
Index 179

among Mexican American college problems with, 92


students” (Miville and Constantine), Stevens, J., 9, 15, 39, 44
107 Stevens, J. P., 38, 40, 98
“Some formal relations in multivariate Stevens, Stanley Smith, 12
analysis” (Tintner), 87 Strohmeier, D., 146
Srivistava, M. S., 45–46, 68–69 Strolin-Goltzman, J. S., 146–47
SRMR. See standardized root-mean-square structural equation modeling (SEM), 2
residual assumptions of, 113–19, 115f, 116f
SSB. See sum-of-squares-between description of, 20–21
SST. See sum-of-squares-total effect indicators in, 153, 154
SSW. See sum-of-squares-within examples of, 132–48, 132f, 134f, 136f,
standardized regression coefficients, 137f, 138f, 139f, 140f, 141f, 142f,
94–95 143f
standardized root-mean-square residual key terms, 109–13, 112f
(SRMR), 128–29 latent variables, used to test
Stark, K. D., 60 relationships between, 109
statistical interaction limitations of, 131–32
definition of, 6 MANCOVA vs., 152–54
description of, 6–7 MANOVA vs., 152–54
Statistical methods for psychology (Howell), MMR vs., 154–56
34 model estimation in, 120–23
statistical model model modification in, 129–30
absence of outliers in, 16–18 model specification in, 119–20
data missing from, 10–11 overview, 109–13, 112f
description of, 6 procedure, 119–30, 125f
just-identified, 114 results of, 144–45
measurement in, 11–13 sample size in, 133–43, 134f, 136f, 137f,
overidentified, 114–16, 116f 138f, 139f, 140f, 141f, 142f, 143f
residuals, 15 software programs, 117, 123
specified correctly, 10 strengths of, 130–31
structurally identified, 114 Structural equation modeling: Foundations
underidentified, 114, 115, 115f, 119 and extensions (Kaplan), 124
statistical power, 41–42 “Structural equation modeling: Strengths,
Statistical power analysis for the behavioral limitations, and misconceptions”
sciences (Cohen), 42 (Tomarken and Waller), 131–32
Statistics for evidence-based practice and Structural equations with latent variables
evaluation (Rubin), 11 (Bollen), 114, 153
stepdown analysis (SDA), 38–40 structural identification, 114–17, 115f, 116f
description of, 88–89 structural model, 111
“Stepdown analysis of variance: A Student, 23
refinement” (Kowlowsky and Student’s q, 36–37
Caspy), 39 Student’s t. See t-distribution
stepwise procedures study-wise alpha level, omnibus hypothesis
backward elimination, 89 test and, 96
description of, 89 study-wise error rate (SW), 4
forward selection, 89 Subbaiah, P., 39–40
missing data and, 92 sum-of-squares-between (SSB), 26
stepwise regression, 89 sum-of-squares error. See residual sum-of-
stepwise regression, 89 squares
180 Index

sum-of-squares-total (SST), 26 Auerbach, and Strolin-Goltzman),


sum-of-squares-within (SSW), 26 146–47
SW. See study-wise error rate two-group multivariate comparisons, 38
sweeping, 92 type I error
systematic error, description of, 13 control of, 3–5
definition of, 3
Tabachnick, B. G., 88 description of, 3–4
Tang, T., 108 inflation of rate of, 25
t-distribution, 23–25, 24f kinds of, 4
“Teaching influence in simple regression”
(Lorenz), 17 underfitting, 6
test command, 102, 102f univariate procedures, description of, 1
“Test of a structural equation model univariate variance, definition of, 22
of comorbidity among homeless “The use of causal indicators in covariance
and domiciled military veterans” structure models: Some practical
(Benda), 145–46 issues” (MacCallum and Browne),
“Tests of multiple independence and the 133–35, 134f
associated, confidence bounds” (Roy Using multivariate statistics (Tabachnick
and Bargmann), 88 and Fidell), 88
“Tests of multiple independence and the
associated confidence bounds” (Roy validity of measurement, 13
and Bargmann), 38 variance inflation factor (VIF), 14
theory of scale types, 12
Thevos, A. K., 61–62 Waller, N. G., 131–32
Thomas, S. E., 61–62 Wan, C. K., 154–55
Thompson , J. S., 127 Webb, Deborah K., 84–85
tolerance, definition of, 14 Whitlock, B., 107
Tomarken, A. J., 131–32 Widaman , K. F., 127
Toothacker, L. E., 37 Wilk, M. B., 15
total effects, 112 Wilks, S. S., 28
t-Rule, 115 Wilk’s /, 31, 32, 33
Tsan, J. Y., 62 Williams, S., 84
t-test, 23–25
definition of, 22 Yates, W. R., 84
Tukey’s Honestly Significant Difference y-intercept, 94
(HSD) test, 36–37
“Turnover in the child welfare workforce: Zhou, M., 84
A different perspective” (McGowan, z-statistic, 24

You might also like