- We are sorry, but NCBI web applications do not support your browser and may not function properly. More information

- Journal List
- NIHPA Author Manuscripts
- PMC3003934

# Variable Selection for Qualitative Interactions

^{1}E-mail: lgunter, jizhu, samurphy @umich.edu

## Abstract

In this article we discuss variable selection for decision making with focus on decisions regarding when to provide treatment and which treatment to provide. Current variable selection techniques were developed for use in a supervised learning setting where the goal is prediction of the response. These techniques often downplay the importance of interaction variables that have small predictive ability but that are critical when the ultimate goal is decision making rather than prediction. We propose two new techniques designed specifically to find variables that aid in decision making. Simulation results are given along with an application of the methods on data from a randomized controlled trial for the treatment of depression.

**Keywords:**decision making, variable selection, depression, machine learning

## 1 Introduction

This article describes two new methods to select variables useful for decision making. Applications that deal with decision making occur in many different fields such as computer science, engineering, economics and medicine. In medicine, deciding when a patient needs treatment and which treatment is best are critical decisions. Clinical trials, particularly those involving heterogenous patients, collect a large amount of potentially useful information that can aid in making these decisions. However, in clinical practice much of this information is expensive, time-consuming, and/or burdensome to collect. Thus variable selection is needed to help inform the clinicians which variables are most important.

Variable selection techniques have been developed to enhance prediction, but their use in decision making has not been well tested. In our research we have found these techniques often miss or downplay the importance of certain interaction variables that are key to making decisions. The variable selection techniques we propose focus on finding these important interactions.

This work is motivated in part by the the Nefazodone CBASP trial data. The Nefazodone CBASP trial (Keller et al., 2000) was a randomized controlled trial conducted to compare the efficacy of three alternate treatments for patients with chronic depression. The study randomized 681 patients with non-psychotic chronic major depressive disorder (MDD) to either Nefazodone, cognitive behavioral-analysis system of psychotherapy (CBASP) or the combination of the two treatments. Analysis of the trial data showed the combination treatment to be superior to the two singleton treatments overall. We wanted know whether this relationship held true for all subsets of patients, and if not, to discover which patient characteristics help to determine the optimal depression treatment for an individual patient.

The remainder of this article is organized as follows: Sections 2 and 3 give background material on optimal decision making and discuss what makes a variable important for decision making. Section 4 provides two ranking techniques designed to find variables useful for decision making followed by an algorithm for using these techniques. Section 5 presents some simulation experiments, and Section 6 illustrates the methods using data from the Nefazodone CBASP study. Concluding remarks are given in Section 7.

## 2 Optimal Decision Making

We consider variable selection in the simplest decision-making setting in which one must decide between two actions. The idea is to use observations about a subject *X* = (*X*_{1}*, X*_{2}*,* …, *X _{p}*), to choose a treatment action

*A*. Following the action a response occurs. The response,

*R*, gives us some indication of the desirability of the chosen action. The goal is to choose actions that maximize the response. A policy,

*π*, is a stochastic or deterministic decision rule mapping the space of observations,

*X*, to the space of the action,

*A*. In other words,

*π*defines the probability for choosing action

*A*=

*a*given the observations

*X*=

*x*. So the goal can be restated as finding a policy

*π*

^{*}that maximizes the response.

A simple example of a decision-making problem is a clinical trial to test two alternative drug treatments. The observation vector, *X*, would consist of baseline variables, such as the patient’s background, medical history and current symptoms. The action would be the treatment assigned to the patient and the response could be the patient’s condition or symptoms after receiving treatment. The goal is to determine which treatment is optimal for any given future patient, using the data obtained in the trial.

Alternate policies can be compared via the expected mean response, called the Value of a policy (Sutton and Barto, 1998). Let the distribution of *X* be a fixed distribution *f*, and let the distribution of *R* given (*X, A*) be a fixed distribution *g*. Then when actions are chosen according to a policy *π*, the trajectory (*X, A, R*) has distribution

If *E _{π}*[ ] denotes the expectation over the above distribution, then the Value of

*π*is

The optimal policy, *π*^{*}, is then defined as

or equivalently

If we knew the multivariate distribution of (X,A,R), the best treatment for future use could be found by calculating *E* [*R*| *X* = *x, A* = *a*] for every possible (*x, a*) combination and then selecting the action leading to the highest conditional expectation of *R* for each *x*. In practice, however, we do not know this distribution. So we must use data to estimate the optimal future treatment. We do this by first estimating *E* [*R*| *X* = *x, A* = *a*] for each (*x, a*) using a predictive model and learner, such as a multiple linear regression. We then use the estimated regression function to ‘estimate’ the best future treatment for each *x*. For example, if we used the data to estimate *E* [*R*| *X* = *x, A* = *a*] by

for *a* {0, 1}, our estimated optimal future treatment actions would be

## 3 Variable Selection

There are multiple reasons why variable selection might be necessary in a decision making application. One reason is that finding the optimal policy becomes more difficult as the number of spurious variables included in the model increases. Thus, careful variable selection could lead to better policies. Also, due to limited resources, only a small number of variables may be possible to collect when enacting a policy in a real world setting. Researchers are often unsure which variables would be most important to collect. Variable selection techniques could help identify these variables. In addition, policies with fewer variables are often easier to understand, so variable selection can improve interpretability.

Currently, variable selection for decision making in many fields is predominantly guided by expert opinion. Expert opinion can be a good starting place when there is sufficient domain knowledge and expertise. Some predictive variable selection techniques, such as Lasso (Tibshirani, 1996), have been suggested (Loth, Davy and Preux, 2006). In clinical trials, a combination of predictive variable selection techniques and statistical testing of a small number of interaction variables suggested by expert opinion are most commonly used (ALLHAT Collaborative Research Group, 2002; Krystal et al., 2001; Reynolds et al., 2006). Little research has been carried out to evaluate these techniques in decision making or suggest how they might be improved.

When selecting variables for decision making, a distinction should be made between variables that are included merely to facilitate estimation as opposed to variables involved in the decision rules. *Predictive* variables are variables used to reduce the variability and increase the accuracy of the estimator. Variables that help prescribe the optimal action for a given patient are *prescriptive* variables (Hollon and Beck, 2004). For optimal estimation results, it is best to select both types of variables. However, only *prescriptive* variables need to be collected when implementing the policy.

For a variable to be *prescriptive*, it must have a qualitative interaction with the action (Peto, 1982). A variable *X _{j}* is said to qualitatively interact with the action,

*A*, if there exists at least two distinct, non-empty sets,

*S*

_{1},

*S*

_{2}

*space*(

*X*) for which

_{j}

for all *x _{j}*

_{1}

*S*

_{1}, and

*x*

_{j}_{2}

*S*

_{2}. These variables are useful for decision making because they help decipher which action is optimal for each individual patient.

To illustrate this idea, see the plots in Figure 1. These plots depict different possible relationships between the conditional mean of *R*, *A* and a particular *X _{j}*, when averaging over all other

*X*≠

_{i}, i*j*. Figure 1(a), shows a variable,

*X*

_{1}, which does not interact with the action. Figure 1(b) shows a variable,

*X*

_{2}, that interacts with the action,

*A*, but does not qualitatively interact with the action. In both plots, the optimal action is

*A*= 1. Knowledge of

*X*

_{1}or

*X*

_{2}is useful for predicting the response for a given action, but should not affect which action should be chosen. Figure 1(c), shows a variable,

*X*

_{3}, which qualitatively interacts with the action. We can see that the optimal action in this plot is

*A*= 0 when

*X*

_{3}≤ .5, and

*A*= 1 when

*X*

_{3}> .5. Knowledge of

*X*

_{3}impacts the best choice of the action and likewise the response, thus it is important for decision making.

The degree to which a prescriptive variable is useful depends on two factors:

*Interaction*: the magnitude of the interaction between the variable and the action. For an action with two possible values,*A*{0, 1}, this is the degree to which the following quantity varies as*x*varies

*Proportion*: the proportion of patients whose optimal choice of action changes given a knowledge of the variable. If*a*^{*}= arg max[_{a}E*R*|*A*=*a*], this is the proportion of patients for which the following holds:$$\underset{a}{argmax}E[R\mid X=x,A=a]\ne {a}^{\ast}.$$(3)

Consider the plots in Figure 2. Figure 2(a) shows the relationship between the conditional mean of *R*, *A*, and a variable *X*_{4}, with an underlying plot giving the distribution of *X*_{4}. Figures 2(b), 2(c) are similar to Figure 2(a), but for variables *X*_{5} and *X*_{6}. Notice that *X*_{4} and *X*_{5} have the same distribution. However, the interaction between *X*_{4} and *A* is much stronger than the interaction between *X*_{5} and *A*. Therefore, the effect of choosing the optimal action is much greater given *X*_{4} than it is given *X*_{5}. Now notice that *X*_{4} and *X*_{6} have the same relationship with the conditional mean of *R* and *A* but are distributed differently. The distribution of *X*_{4} is centered at the point of intersection, so half of the subjects would do better choosing *A* = 0 over *A* = 1. Whereas, the proportion of patients benefiting from choosing *A* = 0 is much smaller with *X*_{6}. Thus *X*_{4} would be more useful in decision making than *X*_{5} or *X*_{6}.

Since both of these factors also affect the predictive ability of a qualitative interaction, it may not be readily apparent why current variable selection techniques designed for prediction are not well equipped to find these prescriptive variables. One reason current variable selection methods aimed at prediction may have problems detecting prescriptive variables could be due to the way prescriptive variables occur in nature. In real world applications, individual variables, rather than interactions between variables, tend to explain most of the variation in the outcome and thus are most important for good prediction. This individual effect a variable has on the response is often referred to as the ‘main effect’ of the variable and most variable selection techniques are good at finding main effects. Furthermore, while non-qualitative treatment-covariate interactions do occur quite frequently in real world applications, it is commonly assumed that qualitative interactions are rare in nature (Cohen, 1988; Yusuf et al., 1991).

There is an abundance of literature discussing qualitative interactions (e.g. Byar and Corle, 1977; Gail and Simon, 1985; Peto, 1982;Lagakos, 2001; Shuster and Van Eys, 1983; Yusuf et al., 1991; Senn, 2001). Much of the statistical literature suggests that the search for qualitative interactions should be severely limited and qualitative interactions that are found should be initially mistrusted (Peto, 1982; Lagakos, 2001; Yusuf et al., 1991; Senn, 2001). This point of view is fueled by a myriad of papers publishing claims of finding a qualitative interaction during exploratory data analysis of a controlled trial followed by subsequent studies in which the interaction did not replicate (Beta-Blocker Pooling Project Research Group, 1988; Yusuf et al., 1991).

Skepticism concerning the validity of qualitative interactions is partially due to the way many clinical trials are conducted. Entry criteria are restrictive for many clinical trials. This results in data with minimal variation in the *X* variables, representing only a small subset of the treatable population. In this case it is often reasonable to assume there are no genuine qualitative interactions within the range of the data. However, this does not imply that genuine qualitative interactions do not exist over the range of *X* for the entire treatable population. For this reason, the methods we present are most useful when applied to data representative of the entire treatable population (or at least a substantial proportion of the population).

Skepticism also exists due to the way analyses of clinical trials are reported. Since journals traditionally publish only significant results, it is tempting to comb through the data in post hoc analysis looking for anything that is significant and interesting. Many times this “data fishing” includes looking for significant qualitative interactions. When significance levels are not corrected for the number of tests performed, researchers can often find at least one significant qualitative interaction spuriously. This problem would not occur as much if researchers were more forthcoming to journals about the number of tests they performed and the significance levels they used (Assmann et al., 2000). However, it is important to note that genuine qualitative interactions with small effect sizes will be undetectable in some data sets, especially those of small sample size. Despite this skepticism, medical researchers continue to look for qualitative interactions. They look for them because it is an underlying goal of clinical research to find the best treatment for each individual patient

Our goal is to develop approaches that assist in finding qualitative interactions, but are less susceptible to finding spurious results. To address the concern that the proposed methods are equivalent to testing large numbers of interactions with uncorrected significance levels, we thoroughly test them on simulated data generated without qualitative interactions. Beyond this, we feel it is also important to emphasize that the goal of these methods is not to find the ‘correct’ underlying model. Rather, the driving force for this variable selection is to facilitate and improve decision making by reducing the number of variables that need to be considered when constructing a policy.

## 4 Qualitative Interaction Ranking

As we discussed in the previous section, variable selection in decision making should focus on variables that qualitatively interact with the action. In this section we will present two variable-ranking techniques that rank the variables in *X* based on their potential for a qualitative interaction with the action variable. We conclude the section with a proposed complete algorithm for variable selection in a decision making application.

The first variable-ranking method is based upon the two usefulness factors for a qualitative variable discussed in the previous section (see quantities (2) and (3)). Assume we have a data set of *n* subjects, with *p* baseline observations taken on each subject, making up the *n*×*p* observation matrix *X*. Also assume that in the data the action, *A* = {0, 1} is randomized. The response is denoted by *R*. Consider the evaluation of the *jth* variable, *X _{j}* (the

*jth*column of

*X*). Then given an estimator of

*E*[

*R*|

*X*=

_{j}*x*=

_{j}, A*a*], say

*Ê*[

*R*|

*X*=

_{j}*x*=

_{j}, A*a*], define the following quantities for

*j*= 1, …,

*p*:

and

where 1{·} is 1 if ‘·’ is true and 0 otherwise and *a*^{*} = arg max* _{a}Ê* [

*R*|

*A*=

*a*] is the overall optimal action.

Here *D _{j}* is a measure of the magnitude of the interaction. The

*P*is a measure of the proportion of subjects affected by a change in the optimal choice of action due to the inclusion of an interaction involving

_{j}*X*. These two quantities can be combined to make a score,

_{j}*U*, for ranking the variables:

_{j}The first term in parentheses provides the relative (as compared to the other variables in *X*) magnitude of *X _{j}*’s interaction with the action; the second term in parentheses provides the relative proportion of affected subjects in

*X*used to select the action. The

_{j}*U*is a product because we want to select

_{j}*X*only if both

_{j}*D*and

_{j}*P*are relatively large. The first variable ranking procedure will rank variables in terms of their

_{j}*U*.

_{j}The second ranking procedure looks directly at the expected increase in the estimated optimal Value due to the knowledge of the variable *X _{j}*. Recall, the Value is the expected response. The ranking procedure estimates the quantity described by Parmigiani (2002) as the value of information. Define the score

*S*as

_{j}Both of these scores, *U* and *S* can be used to rank the variables. They have been defined generically to allow different models for *E*[*R*|*X, A*]. In the numerical section that follows, we use a linear model to estimate the conditional expectation and obtain *Ê*.

Although not explicitly shown in the notation, predictive variables may also be used in the estimation of the conditional expectation. When testing for the interaction between *X _{j}* and

*A*, researchers often prefer to maintain a hierarchical ordering (Wu and Hamada, 2000) and thus the main effect of the variable

*X*and the main effect of the action should be included. This helps to avoid finding spurious interactions that may appear because the main effect is important but is not included in the estimation. It is also wise to include other important main effects of the variables in

_{j}*X*on

*R*to help reduce variability in the estimation.

### 4.1 Variable Selection Algorithm

The following is an overview of an algorithm for variable selection.

**Select Important Predictors:**Select important predictive variables of*R*among (*X, A***X*) using a Lasso with the penalty parameter chosen by Bayesian Information Criterion (BIC).**Rank Interactions Individually:**Rank the variables in*X*using either*U*or*S*. Use the main effect variables selected in step 1 to help decrease the variability in the estimator*Ê*. Select the top*H*variables in rank, where*H*= the number of variables having non-zero*U*or*S*scores.**Create Nested Subsets of Chosen Predictive and Prescriptive Variables:**- Collect the following
*K*variables:- the predictive variables chosen in step 1,
- the main effects of the top
*H*ranked variables in step 2, and - the interactions between
*A*and the top*H*ranked variables in step 2.

- Run a weighted Lasso using a weighting scheme that satisfies the following properties:
- all main effect variables and all interaction variables chosen in step 1 only are given a weight
*w*= 1; - all interaction variables chosen in step 2 are given a weight 0 <
*w*≤ 1 which is a non-increasing function of the U or S score.

- Create
*K*nested subsets based on the order of entry of the*K*variables in the weighted Lasso in the previous step.

**Select Subset Using Adjusted Gain in Value Criterion:**- For each subset
*k*= 1, …,*K*, estimate the maximal Value, e.g.- use the subset to estimate
*Ê*; - estimate the optimal policy, ${\widehat{\pi}}_{k}^{\ast}(x)=arg{max}_{a}\widehat{E}[R\mid X=x,A=a]$;
- estimate the Value of ${\widehat{\pi}}_{k}^{\ast}$ by:$${\widehat{V}}_{k}=\frac{1}{n}\sum _{i=1}^{n}\widehat{E}[R\mid X={x}_{i},A={\widehat{\pi}}_{k}^{\ast}({x}_{i})].$$

- Select the subset,
*k*^{*}, that has the highest Adjusted Gain in Value (AGV) criterion:$$AG{V}_{k}=\frac{{\widehat{V}}_{k}-{\widehat{V}}_{0}}{{\widehat{V}}_{{m}^{\ast}}-{\widehat{V}}_{0}}\left(\frac{{m}^{\ast}}{k}\right),$$where*m*^{*}= arg maxand_{k}_{k}_{0}is the estimated Value of the policy ${\widehat{\pi}}_{0}^{\ast}=arg{max}_{a}\widehat{E}[R\mid A=a]$.

In step 1 we use Lasso to find the variables among (*X, A* * *X*) that are important predictors of *R*. We chose Bayesian Information Criterion to select the penalty parameter (Zou, Hastie and Tibshirani, 2007) because of its conservative nature to ensure only strong predictors enter the model. Predictive variables are important for reducing variability in the estimations. However, predictive variables are only part of the puzzle, so we add to step 1 a few more steps to help our algorithm select both prescriptive and predictive variables. In step 2 we look for qualitative interactions individually using an approach which rates each variable in *X* based on its potential for a qualitative interaction with the action. We look at each of the interaction variables individually to avoid problems with collinearity. In steps 3 and 4 we seek to further refine the set of variables collected in steps 1 and 2. In step 3 we seek a quick way to navigate through the space of all possible combinations of the variables collected in steps 1 and 2. Thus we chose to create nested subsets from the variables based on order of selection in a weighted Lasso. This ordering by the weighted Lasso gives us a joint ranking of all the variables selected in steps 1 and 2. We use the weighting scheme in the weighted Lasso to balance the importance of both predictive and prescriptive variables in the decision making process. Since Lasso favors variables that are predictive we offset this by down-weighting the prescriptive variables. In step 4 we select between the different subsets using the AGV criterion, a criterion that trades o3 between the complexity and the observed Value of each of the models.

The AGV criterion selects the subset of variables with the maximum proportion of increase in Value per variable. It is similar in idea to the adjusted *R*^{2} value. The model with *m*^{*} = arg max* _{k}_{k}* variables is akin to a saturated model, because the addition of more variables does not improve the Value of the model. Thus the denominator is the observed maximum gain in value, among the different variable subsets, divided by

*m*

^{*}, an estimate of the degrees of freedom used to achieve that gain in Value. The numerator then measures the gain in Value of the intermediate model, the model with

*k*variables, divided by

*k*, the estimated degrees of freedom needed to achieve that gain in Value.

An alternate way to look at the AGV criterion is that the quotient (* _{k}* −

_{0})/(

_{m}_{*}−

_{0}) compares the gain in Value for the current subset of variables against the maximum gain in Value over all the subsets of variables. Ideally this term stays fairly stationary whenever a main effect variable is added to the model and increases when a qualitative interaction is added to the model. Thus this quotient is expected to be approximately monotone increasing with

*k*. The quotient,

*m*

^{*}/

*k*, acts as a penalty on the inclusion of variables that do not substantially increase the Value. We include main effect variables in the counts

*m*

^{*}and

*k*because each main effect variable that is included decreases the degrees of freedom. Also, the inclusion of main effects in the counts quickly deflates the quotient as

*k*increases, leading to a less severe penalty on larger models. This is helpful since there is often many more useful predictive variables than prescriptive variables.

In the next section we test this algorithm on simulated data. We reference the algorithm as Method U or Method S depending on the scoring function *U* or *S* that was used in step 2. For the weighting scheme in step 3(b) we tried multiple different schemes (inverse, exponential, etc.). In practice the weighting scheme that worked best is listed below:

- all predictive variables are given a weight
*w*= 1; - all prescriptive variables are given a weight $w=1-\frac{U}{max(U)+\epsilon}$ or $w=1-\frac{S}{max(S)+\epsilon}$, respectively.

The *ε* term in the weight is needed to ensure *w* ≠ 0. So *ε* can be thought of as a stabilizing factor, but it can also be thought of as the balancing factor between prescriptive and predictive variables (i.e. large *ε* favors predictive variables, small *ε* favors prescriptive variables). In experimentation we found *ε* = *H/n* to be a good value.

## 5 Simulations

To test the performance of the new techniques, we ran them on realistically designed simulation data and compared the results to using Lasso (Tibshirani, 1996). Lasso was used to select from the set of main effects of *X*, and the interactions between *A* and each variable in *X*. The main effect of *A* was not subject to selection, that is the coefficient of *A* was unconstrained by the *L*_{1} penalty function. We tested two different methods for choosing the penalty parameter. The first method we used was the Bayesian Information Criterion (BIC) as defined in Zou and Zou et al. (2007). We reference this method as BIC Lasso. Zou et al. (2007) recommend this method when using Lasso primarily for variable selection. The second method we used for choosing the penalty parameter was 5-fold cross-validation on the prediction error of the Lasso model (Tibshirani, 1996). This is a standard method for choosing the penalty parameter and we reference this method as CV Lasso. Note that our method uses Lasso as well, but only to select predictive variables in step 1 and to order variables in step 3(a).

To generate realistic simulation data, we randomly selected rows, with replacement from *X*, the observation matrix from the Nefazodone CBASP trial data. We generated new actions, *A*, and new responses, *R*, that covered a wide variety of models. We report results for the following generative models:

- Main effects of
*X*only, no treatment effect and no interactions with treatment; - Main effects of
*X*, moderate treatment effect and no interactions with treatment; - Main effects of
*X*, moderate treatment effect, multiple medium to small non-qualitative interactions with treatment, no qualitative interaction with treatment; - Main effects of
*X*, small treatment effect, small qualitative interaction with a binary variable, no non-qualitative interactions; - Main effects of
*X*, small treatment effect, small qualitative interaction with a continuous variable, no non-qualitative interactions; - Main effects of
*X*, small treatment effect, multiple moderate to small non-qualitative interactions with treatment, small to moderate qualitative interaction with a binary variable and treatment; - Main effects of
*X*, small treatment effect, multiple small non-qualitative interactions with treatment, small qualitative interaction with a continuous variable and treatment.

For each generative model, we used main effect coefficients for the variables *X*, estimated in an analysis of the real data set. In generative models 3–7 we randomly selected variables from the Nefazodone CBASP data for each treatment covariate interaction and used these same variables for each repetition. The treatment, qualitative interaction and non-qualitative interaction coe3-cients were set using a variant of Cohen’s D effect size measure Cohen (1988) shown below:

We altered this formula by replacing the marginal variance, *V ar*(*R*), with the conditional variance of the response *V ar*(*R*|*X, A*). However, we maintained the definitions of ‘small’ and ‘moderate’ effect sizes suggested by Cohen (1988) as *D* = 0.2 and *D* = 0.5 respectively. Thus the effects are slightly smaller than the traditional definition.

For each generative model, we ran CV Lasso, BIC Lasso, Method U and Method S to see which interaction variables were selected by each method. We repeated this 1000 times and recorded the percentage of time each variable was selected for each method and the sign of the coefficient of each interaction selected.

For each repetition, we also calculated the following statistic for each method

where *V _{π}*

_{*}is the Value of the true optimal policy,

*π*

^{*},

*V*is the Value of an ‘agnostic’ policy

_{π}*π*which gives equal probability to each action and

*V*

_{}_{*}is the Value of the estimated optimal policy given the selected variables. We estimated the policy

^{*}by first fitting a linear model of the selected variables on the response using the training set and then optimizing the fitted model with respect to the action.

The statistic T gives the percentage of gain in Value when using the estimated optimal policy as opposed to an ‘agnostic’ policy relative to the percentage of gain in Value when using the true optimal policy as opposed to an agnostic policy. We compared the new methods with both of the Lasso competitors by looking at the difference in their T statistics. The results are listed in Tables 1 and and2.2. Differences denoted with a * are significantly different from zero using a two sided t-test with *α* = .05. Note that since generative model 1 has no treatment effect and no interactions with treatment, all policies will have the same Value resulting in an undefined T statistic. The tables also list the average number of spurious interactions selected by each method and the selection percentage of the qualitative interaction (if one existed) over the 1000 repetitions.

^{*}are significantly different from

**...**

^{*}are significantly different

**...**

Looking over Table 1 we see that BIC Lasso tends to include a slightly smaller number of spurious interactions, as expected, due to its conservative nature (Zou et al., 2007). It’s conservative nature is also a bonus in terms of the average Value in the rare situation when no interactions exist in the generative model (generative model 2). However, the use of this method results in a dramatic loss in the average Value when a qualitative interaction does exist because the qualitative interaction is often left out.

Table 2 shows that while CV Lasso is good at selecting the qualitative interaction, it tends to include several more spurious interactions than the new methods. This leads to a significant loss in the average Value due to policies with bad decisions based on the spurious variables selected when using this method.

Overall, we found that the two new methods perform well. While the competing Lasso methods each have their appeal in terms of selection, when considering the average Value returns and the purpose of the variable selection, the new methods appear more advantageous.

## 6 Application: Nefazodone CBASP Trial

To apply this method to a real data set we suggest augmenting this algorithm in two ways. First we use bootstrap sampling (Efron and Tibshirani, 1993) of the original data to give a measure of reliability on the results. That is, take 1000 bootstrap samples of the data, run the algorithm and record the interaction variables that are selected along with the sign of the interaction coefficient for each bootstrap sample. This will give a percentage of time each interaction variable is selected by the method. Define the adjusted selection percentage to be the absolute value of the number of times an interaction is selected with a positive coefficient minus the number of times an interaction is selected with a negative coefficient. This adjustment eliminates variables that, across the bootstrap samples, do not consistently interact in one direction with the action.

Second we construct a threshold to determine which interaction variables to include in the final model. The threshold estimates the selection percentages we would expect to see if the data contained no interactions. To compute the threshold, we first remove the interaction effects within the data by randomly reassigning the observed values for the interaction variables to different subjects. In other words, permute the *X* values of the *X* * *A* interactions in the (*X, A, X* * *A*) model matrix. After obtaining 100 permuted data sets, on each permuted data set we run the same analysis of taking 1000 bootstrap samples, running the algorithm and recording the selection percentage of each interaction variable over the 1000 bootstrap samples. We then calculate the maximum adjusted selection percentage over the *p* interaction variables for each permuted data set. The threshold is then set to be the (1 − *α*)*th* percentile over the 100 maximum selection percentages. We found in simulations that the threshold effectively controlled the family-wise error rate to be approximately *α* giving us in any given experiment (1 − *α*)% confidence that a variable with a selection percentage above this threshold interacts with the action.

This augmentation by bootstrap resampling and thresholding helps to stabilize the results and it is possible to apply it to other variable selection algorithms, not just the new methods suggested in this paper. In simulations we found the bootstrap resampling and thresholding also effectively controlled the family-wise error for BIC Lasso. The bootstrap resampling can also be done with CV Lasso, however, this threshold was far too conservative to control the family-wise error rate for the CV Lasso.

To demonstrate these new methods along with this augmentation we applied them to a real data set dealing with depression. As introduced previously, the Nefazodone CBASP trial (Keller et al., 2000) was conducted to compare the efficacy of three alternate treatments for patients with chronic depression. We applied the methods to pinpoint if any of the patient characteristics might help to determine the optimal depression treatment for each patient.

The study randomized 681 patients with non-psychotic chronic major depressive disorder (MDD) to either Nefazodone, cognitive behavioral-analysis system of psychotherapy (CBASP) or the combination of the two treatments. For detailed study design and primary analysis see Keller et al. (2000). We considered *p* = 61 baseline covariates for our observation matrix *X*; these variables are listed in Table 3. The outcome, *R*, was the 24-item Hamilton Rating Scale for Depression score (Hamilton, 1967), observed post treatment. For simplicity, we only allowed the action to vary between two treatments at a time. Since the primary analysis of the data showed the combination treatment to be superior to either individual treatment alone, we ran the variable selection techniques twice: the first time with the action varying between the combination treatment and Nefazodone alone, and the second time with the action varying between the combination treatment and CBASP alone.

The results of our first analysis comparing the combination treatment to Nefazodone alone are shown in Table 3 and Figure 3. The adjusted selection percentages for each variable are listed in Table 3 along with 80% and 90% thresholds at the bottom (i.e. alpha equal to 0.2 and 0.1). Figure 3 shows plots of these adjusted selection percentages and thresholds. The *x*-axis in each plot corresponds to the variable numbers listed in Table 3. The horizontal dashed lines are 80% thresholds and the horizontal solid lines are 90% thresholds.

Only the adjusted selection percentages from the bootstrap resampling of CV Lasso are plotted in the first plot. Absent a working threshold, it is not very clear which variables should be selected for further analysis. The next three plots are for BIC Lasso, method U and method S. All three of these methods had one variable with an adjusted selection percentage exceeding the 80% threshold. For BIC Lasso, this variable was variable 34, *Obsessive Compulsive Disorder*, whereas, for both of the new methods the variable was variable 38, past history of *Alcohol Dependence*. Further analysis of the two variables confirmed that the interaction with *Obsessive Compulsive Disorder* and the action was non-qualitative in the data, whereas the interaction between past *Alcohol Dependence* and the action had good potential for being qualitative. More study should be done to determine the usefulness of past history of *Alcohol Dependence* for selecting treatments. Also, in this study around 20% of subjects in each group left the study early. Here *R* is the last observed Hamilton Rating, which is a worst case scenario under the assumption that depressed subjects who drop out of the study do not improve. Although we included all available good predictors of *R* in the model of *E*[*R*|*X, A*], it may be the case that unobserved determinants of dropout provide an alternate explanation for the apparent qualitative interaction with past *Alcohol Dependence*.

The results of our second analysis comparing the combination treatment to CBASP alone are shown in Figure 4. The figure shows plots of the adjusted selection percentages for each method along with 80% thresholds. The *x*-axis in each plot corresponds to the variable numbers listed in Table 3. The horizontal dashed lines are 80% thresholds. As shown in the plot, no variables were selected by either of the new methods or BIC Lasso. For brevity we forgo listing the individual selection percentages. This analysis suggests there are no true qualitative interactions between the baseline covariates and the two treatment options. Many researchers believe this is the most likely scenario in medical decision-making applications. We conclude that the combination treatment is better than CBASP alone for all patient subsets tested.

## 7 Discussion

In this article, we discussed when a variable is important in decision making and why variable selection techniques designed for prediction may not perform well in a decision-making setting. We presented two new techniques explicitly designed to select variables for decision making. These techniques focus on interaction variables that are good candidates for playing a role in the actual decision rules.

It should be noted that Lasso treats the indicator variables used to model a categorical variable as separate variables. It is well known that this can lead to over-selection of categorical variables with many categories. Consequently, the proposed method is subject to this problem. Therefore we recommend using Group Lasso Yuan and Lin (2006); Zhao, Roca and Yu (2006), OSCAR Bondell and Reich (2008) or elastic net type penalty Zou and Hastie (2005) in step 3 of the algorithm when applying it to a data set with many multi-category variables.

The entire algorithm, including bootstrap sampling and thresholding, takes approximately 30 hours to run in Matlab on a 3 Ghz Intel Xeon X5355 processor for a data set of *p* = 60 baseline covariates and *n* = 400 subjects. The algorithm would require far less computation time if a more theoretically justified threshold could be determined, rather than using a permutation based threshold. This is an area for future work.

More research is needed to determine the oracle consistency properties of this algorithm and its performance on problems where *p* > *n*. Adjusting these methods to deal with dropout is also an open issue. Our long term goal is to extend these methods to settings with multiple decision time points.

## Acknowledgments

We wish to thank Martin Keller and the investigators who conducted the trial ‘A Comparison of Nefazodone, the Cognitive Behavioral-analysis System of Psychotherapy, and Their Combination for Treatment of Chronic Depression’, for use of their data, and gratefully acknowledge Bristol-Myers Squibb for helping fund the study. We also acknowledge financial support from NIDA grants R21 DA019800, K02 DA15674, P50 DA10075, NIMH grant R01 MH080015 and NSF grants DMS 0505432 and DMS 0705532, and technical support from the Betty Jo Hay Chair in Mental Health, A. John Rush, MD at the University of Texas Southwestern Medical Center, Dallas.

## Footnotes

**Publisher's Disclaimer: **This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.

## References

- Assmann SF, Pocock SJ, Enos LE, Kasten LE. Subgroup analysis and other (mis)uses of baseline data in clinical trials. The Lancet. 2000;355:1064–1069. [PubMed]
- Beta-Blocker Pooling Project Research Group. The beta-blocker pooling project (bbpp): Subgroup findings from randomized trials in post infarction patients. The Eropean Heart Journal. 1988;9:8–16. [PubMed]
- Bondell HD, Reich BJ. Simultaneous regression shrinkage, variable selection, and supervised clustering of predictors with OSCAR. Biometrics. 2008;64:115–123. [PMC free article] [PubMed]
- Byar DP, Corle DK. Selecting optimal treatment in clinical trials using covariate information. Journal of Chronic Diseases. 1977;30:445–459. [PubMed]
- Cohen J. Statistical Power Analysis for the Behavioral Sciences. 2. Lawrence Earlbaum Associates; Hillsdale, NJ: 1988.
- Efron B, Tibshirani R. An introduction to the bootstrap. Chapman and Hall; New York: 1993.
- Gail M, Simon R. Testing for qualitative interactions between treatment effects and patient subsets. Biometrics. 1985;41:361–372. [PubMed]
- Hamilton M. Development of a rating scale for primary depressive illness. British Journal of Social and Clinical Psychology. 1967;6:278–296. [PubMed]
- Hollon SD, Beck AT. Cognitive and cognitive behavioral therapies. In: Lambert MJ, editor. Garfield and Bergin’s Handbook of Psychotherapy and Behavior Change: An Empirical Analysis. 5. New York: Wiley; 2004. pp. 447–492.
- Keller MB, McCullough JP, Klein DN, Arnow B, Dunner DL, Gelenberg AJ, Marekowitz JC, Nemero3 CB, Russell JM, Thase ME, Trivedi MH, Zajecka J. A comparison of nefazodone, the cognitive behavioral-analysis system of psychotherapy, and their combination for treatment of chronic depression. New England Journal of Medicine. 2000;342:331–336. [PubMed]
- Krystal JH, Cramer JA, Krol WF, Kirket GF, Rosenheck R. Naltrexone in the treatment of alcohol dependence. New England Journal of Medicine. 2001;345:1734–1739. [PubMed]
- Lagakos S. The challenge of subgroup analyses-reporting without distorting. New England Journal of Medicine. 2006;354:1667–1669. [PubMed]
- Loth M, Davy M, Preux P. IEEE International Symposium on Approximate Dynamic Programming and Reinforcement. Hawaii, USA: Springer; 2006. Sparse temporal difference learning using lasso.
- Parmigiani G. Modeling in Medical Decision Making: a Baysian Approach. Wiley; West sussex, England: 2002.
- Peto R. Statistical aspects of cancer trials. In: Halnan KE, editor. Treatment of Cancer. London, UK: Chapman; 1982. pp. 867–871.
- Reynolds CF, Dew MA, Pollock BG, Mulsant BH, Frank E, Miller MD, Houck PR, Mazumdar S, Butters MA, Stack JA, Schlernitzauer MA, Whyte EM, Gildengers A, Karp J, Lenze E, Szanto K, Bensasi S, Kupfer DJ. Maintenance treatment of major depression in old age. New England Journal of Medicine. 2006;345:1130–1138. [PubMed]
- Senn SJ. Individual therapy: new dawn or false dawn. Drug Information Journal. 2001;35:1479–1494.
- Shuster J, Van Eys J. Interaction between prognostic factors and treatment. Controlled Clinical Trials. 1983;4:209–214. [PubMed]
- Sutton RS, Barto AG. Reinforcement Lerning: An Introduction. MIT Press; Cambridge, MA: 1998.
- The ALLHAT O3cers and Coordinators for the ALLHAT Collaborative Research Group. Major outcomes in moderately hypercholesterolemic, hypertensive patients randomized to pravastatin vs usual care: The antihypertensive and lipid-lowering treatment to prevent heart attack trial(allhat-llt) Journal of the American Medical Association. 2002;288:2998–3007. [PubMed]
- Tibshirani R. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society, Series B. 1996;58:267–288.
- Wu CF, Hamada M. Experiments: Planning, Analysis, and Parameter Design Optimization. Wiley; New York: 2000.
- Yuan M, Lin Y. Model selection and estimation in regression with grouped variables. Journal of the Royal Statistical Society, Series B. 2006;68:49–67.
- Yusuf S, Wittes J, Probstfield J, Tyrole HA. Analysis and interpretation of treatment effects in subgroups of patients in randomized clinical trials. Journal of the American Medial Association. 1991;266:93–98. [PubMed]
- Zhao P, Rocha G, Yu B. Grouped and hierarchical model selection through composite absolute penalties. Department of Statistics University of California; Berkeley: 2006. Technical Report 703.
- Zou H, Hastie T. Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society, Series B. 2005;67:301–320.
- Zou H, Hastie T, Tibshirani R. On the “degrees of freedom” of the lasso. The Annals of Statistics. 2007;35:2173–2192.

## Formats:

- Article |
- PubReader |
- ePub (beta) |
- PDF (814K)

- Variable selection for qualitative interactions in personalized medicine while controlling the family-wise error rate.[J Biopharm Stat. 2011]
*Gunter L, Zhu J, Murphy S.**J Biopharm Stat. 2011 Nov; 21(6):1063-78.* - Variable selection for optimal treatment decision.[Stat Methods Med Res. 2013]
*Lu W, Zhang HH, Zeng D.**Stat Methods Med Res. 2013 Oct; 22(5):493-504. Epub 2011 Nov 23.* - Decision aids for people facing health treatment or screening decisions.[Cochrane Database Syst Rev. 2009]
*O'Connor AM, Bennett CL, Stacey D, Barry M, Col NF, Eden KB, Entwistle VA, Fiset V, Holmes-Rovner M, Khangura S, et al.**Cochrane Database Syst Rev. 2009 Jul 8; (3):CD001431. Epub 2009 Jul 8.* - Parameter selection for and implementation of a web-based decision-support tool to predict extubation outcome in premature infants.[BMC Med Inform Decis Mak. 2006]
*Mueller M, Wagner CL, Annibale DJ, Knapp RG, Hulsey TC, Almeida JS.**BMC Med Inform Decis Mak. 2006 Mar 1; 6:11. Epub 2006 Mar 1.* - Decision aids for people facing health treatment or screening decisions.[Cochrane Database Syst Rev. 2001]
*O'Connor AM, Stacey D, Rovner D, Holmes-Rovner M, Tetroe J, Llewellyn-Thomas H, Entwistle V, Rostom A, Fiset V, Barry M, et al.**Cochrane Database Syst Rev. 2001; (3):CD001431.*

- Experimental Design and Primary Data Analysis Methods for Comparing Adaptive Interventions[Psychological methods. 2012]
*Nahum-Shani I, Qian M, Almirall D, Pelham WE, Gnagy B, Fabiano G, Waxmonsky J, Yu J, Murphy S.**Psychological methods. 2012 Dec; 17(4)10.1037/a0029372* - Q-Learning: A Data Analysis Method for Constructing Adaptive Interventions[Psychological methods. 2012]
*Nahum-Shani I, Qian M, Almirall D, Pelham WE, Gnagy B, Fabiano G, Waxmonsky J, Yu J, Murphy S.**Psychological methods. 2012 Dec; 17(4)478-494* - Estimating Optimal Treatment Regimes from a Classification Perspective[Stat. 2012]
*Zhang B, Tsiatis AA, Davidian M, Zhang M, Laber E.**Stat. 2012 Jan 1; 1(1)103-114* - On Bayesian Methods of Exploring Qualitative Interactions for Targeted Treatment[Statistics in medicine. 2012]
*Chen W, Ghosh D, Raghunathan TE, Norkin M, Sargent DJ, Bepler G.**Statistics in medicine. 2012 Dec 10; 31(28)3693-3707* - Variable Selection for Optimal Treatment Decision[Statistical methods in medical research. 20...]
*Lu W, Zhang HH, Zeng D.**Statistical methods in medical research. 2013 Oct; 22(5)493-504*

- PubMedPubMedPubMed citations for these articles

- Variable Selection for Qualitative InteractionsVariable Selection for Qualitative InteractionsNIHPA Author Manuscripts. Jan 30, 2011; 1(8)42PMC

Your browsing activity is empty.

Activity recording is turned off.

See more...