1.1 Why Should I Do Multivariable Analysis?
We live in a multivariable world. Most events, whether medical, political, social, or personal, have multiple causes. And these causes are related to one another. Multivariable analysisFootnote 1 is a statistical tool for determining the relative contributions of different causes to a single event or outcome.
Clinical researchers, in particular, need multivariable analysis because most diseases have multiple causes, and prognosis is usually determined by a large number of factors. Even for those infectious diseases that are known to be caused by a single pathogen, a number of factors affect whether an exposed individual becomes ill, including the characteristics of the pathogen (e.g., virulence of strain), the route of exposure (e.g., respiratory route), the intensity of exposure (e.g., size of inoculum), and the host response (e.g., immunologic defense).
Multivariable analysis allows us to sort out the multifaceted nature of risk factors and their relative contribution to outcome. For example, observational epidemiology has taught us that there are a number of risk factors associated with premature mortality, notably: smoking, a sedentary lifestyle, obesity, elevated cholesterol, and hypertension. Note that I did not say that these factors cause premature mortality. Statistics alone cannot prove that a relationship between a risk factor and an outcome are causal.Footnote 2 Causality is established on the basis of biological plausibility and rigorous study designs, such as randomized controlled trials, which eliminate sources of potential bias.
Identification of risk factors of premature mortality through observational studies has been particularly important because you cannot randomize people to many of the conditions that cause premature mortality, such as smoking, sedentary lifestyle, or obesity. And yet these conditions tend to occur together; that is, people who smoke tend to exercise less and be more likely to be obese. How does multivariable analysis separate the independent contribution of each of these factors? Let’s consider the case of exercise. Numerous studies have shown that persons who exercise live longer than persons with sedentary lifestyles. But if the only reason that persons who exercise live longer is that they are less likely to smoke and more likely to eat low-fat meals leading to lower cholesterol, then initiating an exercise routine would not change a person’s life expectancy.
The Aerobics Center Longitudinal Study tackled this important question.Footnote 3 They evaluated the relationship between exercise and mortality in 25,341 men and 7,080 women. All participants had a baseline examination between 1970 and 1989. The examination included a physical examination, laboratory tests, and a treadmill evaluation to assess physical fitness. Participants were followed for an average of 8.4 years for the men and 7.5 years for the women.
Table 1.1 compares the characteristics of survivors to persons who had died during the follow-up. You can see that there are a number of significant differences between survivors and decedents among men and women. Specifically, survivors were younger, had lower blood pressure, lower cholesterol, were less likely to smoke, and were more physically fit (based on the length of time they stayed on the treadmill and their level of effort).

Although the results are interesting, Table 1.1 does not answer our basic question: Does being physically fit independently increase longevity? It doesn’t answer the question because whereas the high-fitness group was less likely to die during the study period, those who were physically fit may just have been younger, been less likely to smoke, or had lower blood pressure.
To determine whether exercise is independently associated with mortality, the authors performed proportional hazards analysis, a type of multivariable analysis. The results are shown in Table 1.2. If you compare the number of deaths per thousand person-years in men, you can see that there were more deaths in the low-fitness group (38.1) than in the moderate/high fitness group (25.0). This difference is reflected in the elevated relative risk for lower fitness (38.1/25.0 = 1.52). These results are adjusted for all of the other variables listed in the table. This means that low fitness is associated with higher mortality, independent of the effects of other known risk factors for mortality, such as smoking, elevated blood pressure, cholesterol, and family history. A similar pattern is seen for women.

Stratified analysis assesses the effect of a risk factor on outcome while holding another variable constant.
Was there any way to answer this question without multivariable analysis? One could have performed stratified analysis. Stratified analysis assesses the effect of a risk factor on outcome while holding another variable constant. So, for example, we could compare physically fit to unfit persons separately among smokers and nonsmokers. This would allow us to calculate a relative risk for the impact of fitness on mortality, independent of smoking. This analysis is shown in Table 1.3.

Unlike the multivariable analysis in Table 1.2, the analyses in Table 1.3 are bivariate.Footnote 4 We see that the mortality rate is greater among those at low fitness compared to those at moderate/high fitness, both among smokers (48.0 vs. 29.4) and among nonsmokers (44.0 vs. 20.1). This stratified analysis shows that the effect of fitness is independent of smoking status.
But what about all of the other variables that might affect the relationship between fitness and longevity? You could certainly stratify for each one individually, proving that the effect of fitness on longevity is independent not only of smoking status, but also independent of elevated cholesterol, elevated blood pressure, and so on. However, this would only prove that the relationship is independent of these variables taken singly.
To stratify by two variables (smoking and cholesterol), you would have to assess the relationship between fitness and mortality in four groups (smokers with high cholesterol; smokers with low cholesterol; nonsmokers with high cholesterol; nonsmokers with low cholesterol). To stratify by three variables (smoking status, cholesterol level, and elevated blood pressure [yes/no]), you would have to assess the relationship between fitness and mortality in eight groups (2 × 2 × 2 = 8); add elevated glucose (yes/no) and you would have 16 groups (2 × 2 × 2 × 2 = 16); add age (in six decades) and you would have 96 groups (2 × 2 × 2 × 2 × 6 = 96); and we haven’t even yet taken into account all of the variables in Table 1.1 that are associated with mortality.
With each stratification variable you add, you increase the number of subgroups for which you have to individually assess whether the relationship between fitness and mortality holds. Besides producing mountains of printouts and requiring a book (rather than a journal article) to report your results, you would likely have an insufficient sample size in some of these subgroups, even if you started with a large sample size. For example, in the Aerobics Center Longitudinal Study there were 25,341 men but only 601 deaths. With 96 subgroups, assuming uniform distributions, you would expect only about six deaths per subgroup. But, in reality, you wouldn’t have uniform distributions. Some samples would be very small, and some would have no outcomes at all.
Multivariable analysis overcomes this limitation. It allows you to simultaneously assess the impact of multiple independent variables on outcome. But there is (always) a cost: The model makes certain assumptions about the nature of the data. These assumptions are sometimes hard to verify. We will take up these issues in Chapters 3, 4, and 9.
1.2 What Are Confounders? How Does Multivariable Analysis Help Me to Deal with Them?
The ability of multivariable analysis to simultaneously assess the independent contribution of a number of risk factors to outcome is particularly important when you have “confounding.” Confounding occurs when the apparent association between a risk factor and an outcome is affected by the relationship of a third variable to the risk factor and the outcome; the third variable is called a confounder.
A confounder is associated with the risk factor and causally related to the outcome.
For a variable to be a confounder, the variable must be associated with the risk factor and causally related to the outcome (Figure 1.1).

Figure 1.1 Relationships among risk factor, confounder, and outcome.
A classically taught example of confounding is the relationship between carrying matches and developing lung cancer (Figure 1.2). Persons who carry matches have a greater chance of developing lung cancer; the confounder is smoking. This example is often used to illustrate confounding because it is easy to grasp that carrying matches cannot possibly cause lung cancer.

Figure 1.2 Relationships among carrying matches, smoking, and lung cancer.
Stratified analysis can be used to assess and eliminate confounding. If you stratify by smoking status you will find that carrying matches is not associated with lung cancer. That is, there will be no relationship between carrying matches and lung cancer when you look separately among smokers and nonsmokers (Figure 1.2). The statistical evidence of confounding is the difference between the unstratified and the stratified analysis. In the unstratified analysis the chi-squared test would be significant and the odds ratio for the impact of matches on lung cancer would be significantly greater than one. In the two stratified analyses (smokers and nonsmokers), carrying matches would not be significantly associated with lung cancer; the odds ratio would be one in both strata. This differs from the example of stratified analysis in Table 1.3 where exercise was significantly associated with mortality for both smokers and nonsmokers.
Most clinical examples of confounding are more subtle and harder to diagnose than the case of matches and lung cancer. Let’s look at the relationship between smoking and prognosis in patients with coronary artery disease following angioplasty (the opening of clogged coronary vessels with the use of a wire and a balloon).
Everyone knows (although the cigarette companies long claimed ignorance) that smoking increases the risk of death. Countless studies, including the Aerobics Center Longitudinal Study (Table 1.2), have demonstrated that smoking is associated with increased mortality. How then can we explain the results of Hasdai and colleagues?Footnote 5 They followed 5,437 patients with coronary artery disease who had angioplasty. They divided their sample into nonsmokers, former smokers (quit at least six months before procedure), recent quitters (quit immediately following the procedure), and persistent smokers. The relative risk of death with the 95% confidence intervals are shown in Table 1.4.
| Bivariate | Nonsmokers | Former smokers | Recent quitters | Persistent smokers |
|---|---|---|---|---|
| Relative risk of death | 1.0 (ref.) | 1.08 (0.92–1.26) | 0.56 (0.40–0.77) | 0.74 (0.59–0.94) |
How can the risk of death be lower among persons who persistently smoke than those who never smoked? In the case of recent quitters, you would expect their risk of death to return toward normal only after years of not smoking – and even then you wouldn’t actually expect quitters to have a lower risk of death than nonsmokers.
Before you assume that there is something wrong with this study, several other studies have found a similar relationship between smoking and better prognosis among patients with coronary artery disease after thrombolytic therapy. This effect has been named the “smoker’s paradox.”Footnote 6 What is behind the paradox? Look at Table 1.5. As you can see, compared to nonsmokers and former smokers, quitters and persistent smokers are younger, have had angina for a shorter period of time, are less likely to have diabetes and hypertension, and have less severe coronary artery disease (i.e., more one-vessel disease and less three-vessel disease). Given this, it is not so surprising that the recent quitters and persistent smokers have a lower risk of death than nonsmokers and former smokers: They are younger and have fewer underlying medical problems than the nonsmokers and former smokers.

Compare the bivariate (unadjusted) risk of death to the multivariable risk of death (Table 1.6). Note that in the multivariable analysis the researchers adjusted for those differences, such as age and duration of angina, that existed among the four groups.

With statistical adjustment for the baseline differences between the groups, the former smokers and persistent smokers have a significantly greater risk of death than nonsmokers – a much more sensible result. (The recent quitters also have a greater risk of death than the nonsmokers, but the confidence intervals of the relative risk do not exclude one.) The difference between the bivariate and multivariable analysis indicates that confounding is present. The advantage of multivariable analysis over stratified analysis is that it would have been difficult to stratify for age, duration of angina, diabetes, hypertension, and extent of coronary artery disease.
1.3 What Are Suppressors? How Does Multivariable Analysis Help Me to Deal with Them?
Suppressor variables are a type of confounder. As with confounders, a suppressor is associated with the risk factor and the outcome (Figure 1.3). The difference is that on bivariate analysis there is no effect seen between the risk factor and the outcome. But when you adjust for the suppressor, the relationship between the risk factor and the outcome becomes significant.
Unlike a typical confounder, when you have a suppressor you won’t see any bivariate association between the risk factor and the outcome until you adjust for the suppressor.
Identifying and adjusting for suppressors can lead to important findings. For example, it was unknown whether taking antiretroviral treatment would prevent HIV seroconversion among healthcare workers who sustained a needle stick from a patient who was HIV-infected. For several years, healthcare workers who had an exposure were offered zidovudine treatment, but they were told that there was no efficacy data to support its use. A randomized controlled trial was attempted, but it was disbanded because healthcare workers did not wish to be randomized.
Since a randomized controlled trial was not possible, a case-control study was performed instead.Footnote 7 The cases were healthcare workers who sustained a needle stick and had seroconverted. The controls were healthcare workers who sustained a needle stick but had remained HIV-negative. The question was whether the proportion of persons taking zidovudine would be lower in the group who had seroconverted (the cases) than in the group who had not become infected (the controls). The investigators found that the proportion of cases using zidovudine was lower (9 of 33 cases or 27%) than the proportion of controls using zidovudine (247 of 679 controls or 36%), but the difference was not statistically significant (probability [P] = 0.35). Consistent with this nonsignificant trend, the odds ratio shows that zidovudine was protective (0.7), but the 95% confidence intervals were wide and did not exclude one (0.3–1.4).
However, it was known that healthcare workers who sustained an especially serious exposure (e.g., a deep injury or who stuck themselves with a needle that had visible blood on it) were more likely to choose to take zidovudine than healthcare workers who had more minor exposures. Also, healthcare workers who had serious exposures were more likely to seroconvert.
When the researchers adjusted their analysis for severity of injury using multiple logistic regression, zidovudine use was associated with a significantly lower risk of seroconversion (odds ratio [OR] = 0.2; 95% confidence interval (CI) = 0.1 – 0.6; P < 0.01). Thus, we have an example of a suppressor effect as shown in Figure 1.4. Severity of exposure is associated with zidovudine use and causally related to seroconversion. Zidovudine use is not associated with seroconversion in bivariate analyses but becomes significant when you adjust for severity of injury.

Figure 1.4 Relationships among zidovudine, severity of injury, and seroconversion.
Although this multivariable analysis demonstrated the efficacy of zidovudine on preventing seroconversion by incorporating the suppressor variables, it should be remembered that multivariable analysis cannot adjust for other potential biases in the analysis. For example, the cases and controls for this study were not chosen from the same population, raising the possibility that selection bias may have influenced the results. Nonetheless, on the strength of this study, postexposure prophylaxis with antiretroviral treatment became the standard of care for healthcare workers who sustained needle sticks from HIV-contaminated needles.
1.4 What Are Interactions? How Does Multivariable Analysis Help Me to Deal with Them?
An interaction occurs when the impact of a risk factor on outcome is changed by the value of a third variable. Interaction is sometimes referred to as effect modification, since the effect of the risk factor on outcome is modified by another variable.
An interaction occurs when the impact of a risk factor on outcome is changed by the value of a third variable.
An interaction is illustrated in Figure 1.5. The risk factor’s effect on outcome (solid lines) differs depending on the value of the interaction variable (whether it is 1 or 0). The dotted line indicates the relationship without consideration of the interaction effect.

Figure 1.5 Illustration of an interaction effect.
Figure 1.5Long description
The illustration depicts the relationship of the risk factor's effect on outcomes with or without the consideration of the interaction effect. The effect differs depending on the values of the interaction variables 0 or 1 depicted by solid lines. The effect of the risk factors is direct on the outcome without the consideration of the interaction effect depicted by dotted lines.
In extreme cases, an interaction may completely reverse the relationship between the risk factor and the outcome. This would occur when the risk factor increased the likelihood of outcome at one value of the interaction variable but decreased the likelihood of outcome at a different value of the interaction variable. More commonly, the effect of the risk factor on the outcome is stronger (or weaker) at certain values of the third variable.
As with confounding, stratification can be used to identify an interaction. By stratifying by the interaction variable, you can observe the effect of a risk factor on outcome at the different values of the interaction variable. You can statistically test whether the association between a risk factor and an outcome at different levels of the interaction variable are statistically different from one another using a chi-squared test for homogeneity.
However, as with the use of stratification to eliminate confounding, use of stratification to demonstrate interaction has limitations. It is cumbersome to stratify by more than one or two variables; yet you may have multiple interactions in your data. Whereas stratification will accurately quantify the effect of the risk factor on the outcome at different levels of the interaction variable, this analysis will not be adjusted for the other variables in your model (e.g., confounders) that may affect the relationship between risk factor and outcome. Multivariable analysis allows you to include interaction terms and assess them while adjusting for other variables.
For example, Zucker and colleagues evaluated whether specific signs or symptoms of myocardial infarction were different in men than in women presenting to the emergency department with chest pain or other symptoms of acute cardiac ischemia.Footnote 8
In Table 1.7 you can see the association between the independent variables and confirmed diagnoses of acute myocardial infarction. The coefficients and odds ratios are from a multiple logistic regression model. The authors found three significant interactions involving gender: male gender and ST elevation (on electrocardiogram), male gender and congestive heart failure, and male gender and white race.
What do these interactions mean? Let’s use the interaction involving male gender and ST elevations as an example (I have put these two variables and their interaction term in bold print). Note that men were more likely than women to have an acute myocardial infarction (OR = 1.6), even after adjusting for other variables associated with an infarction. Similarly, ST elevations were more likely to indicate ischemia (OR = 8.1). Given this, you would expect that males with ST elevations would have a markedly higher risk of myocardial infarction (1.6 × 8.1 = 13.0) than women (1.0 × 8.1 = 8.1) (the wonderful property of odds ratios that allows you to multiply them this way is explained in Section 9.7).
The multiplication of the odds ratios of gender and ST elevations would lead you to believe that men with ST elevations would have a significantly higher risk of an acute myocardial infarction than women (13.0 vs. 8.1). In fact, the risk for men and women with ST elevations was similar. This is reflected in the negative coefficient for male gender × ST elevations and the odds ratio of 0.6. If you multiply out the odds ratio for the interaction of male gender with ST elevations, men with ST elevations (1.6 × 8.1 × 0.6 = 7.8) and women with ST elevations (1.0 × 8.1× 1.0 = 8.1) have a similar risk of myocardial infarction.
ST elevations are highly specific for (although not diagnostic of) myocardial infarction. It is not surprising, therefore, that the risks of myocardial infarction are similar in men and women with ST elevations. Had being male made it even worse to have ST elevations the coefficient would have been positive, the odds ratio would have been greater than one, and we would have seen an even greater difference between the risk of an acute myocardial infarction for men and for women in the presence of ST elevations than the difference between 13.0 and 8.1.
Because interaction effects can be difficult to assess and interpret, I will return to this topic in Sections 7.3, 8.4, and 9.7.












