+ All Categories
Home > Documents > Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992)...

Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992)...

Date post: 06-Oct-2020
Category:
Upload: others
View: 0 times
Download: 0 times
Share this document with a friend
49
Longitudinal data and regression trees Random effects (RE-EM) trees Unbiased regression trees MODel-basEd RaNdom effects (MODERN) trees Goodness-of-fit and regression trees Future work Regression Trees for Longitudinal and Clustered Data: Methods, Applications, and Extensions Jeffrey S. Simonoff (New York University) Joint work with Denis Larocque, Rebecca J. Sela, and Wei Fu Modern Modeling Methods (M 3 ) Conference, May 25, 2016 Regression Trees for Longitudinal and Clustered Data
Transcript
Page 1: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Regression Trees for Longitudinal and ClusteredData: Methods, Applications, and Extensions

Jeffrey S. Simonoff (New York University)

Joint work with Denis Larocque, Rebecca J. Sela, and Wei Fu

Modern Modeling Methods (M3) Conference, May 25, 2016

Regression Trees for Longitudinal and Clustered Data

Page 2: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Outline of talkLongitudinal data and regression trees

Longitudinal data modelingRegression trees

Random effects (RE-EM) treesEstimationApplication to real dataPerformance of RE-EM trees

Unbiased regression treesUnbiased variable selection for regression treesPerformance of unbiased RE-EM treeApplication to real data

MODel-basEd RaNdom effects (MODERN) treesLinear functions at nodesPerformance of MODERN treeApplication to real data

Goodness-of-fit and regression treesTesting for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

Future workRegression Trees for Longitudinal and Clustered Data

Page 3: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Longitudinal data modelingRegression trees

Longitudinal data

Panel or longitudinal data, in which we observe many individualsover multiple periods, offers a particularly rich opportunity forunderstanding and prediction, as we observe the different pathsthat a variable might take across individuals. Such data, often on alarge scale, are seen in many applications:

I test scores of students over time

I blood levels of patients over time

I transactions by individual customers over time

I tracking of purchases of individual products over time

Regression Trees for Longitudinal and Clustered Data

Page 4: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Longitudinal data modelingRegression trees

Longitudinal data

The analysis of longitudinal data is especially rewarding with largeamounts of data, as this allows the fitting of complex or highlystructured functional forms to the data. Conversely, “big data”often come in clustered and longitudinal form, a structure that istypically ignored (or at least underutilized) in modern machinelearning methods.

We observe a panel of individuals i = 1, ..., I at times t = 1, ...,Ti .A single observation period for an individual (i , t) is termed anobservation; for each observation, we observe a vector ofcovariates, xit = (xit1, ..., xitK )′, and a numerical response, yit .

Regression Trees for Longitudinal and Clustered Data

Page 5: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Longitudinal data modelingRegression trees

Longitudinal data models

Because we observe each individual multiple times, we may findthat the individuals differ in systematic ways; e.g., y may tend tobe higher for all observation periods for individual i than for otherindividuals with the same covariate values because ofcharacteristics of that individual that do not depend on thecovariates. This pattern can be represented by an “effect” specificto each individual (for example, an individual-specific intercept)that shifts all predicted values for individual i up by a fixed amount.

yit = Zitbi + f (xit1, ..., xitK ) + εit

Regression Trees for Longitudinal and Clustered Data

Page 6: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Longitudinal data modelingRegression trees

Fixed and random effects

I If f is linear in the parameters and the bi are taken as fixed orpotentially correlated with the predictors, then this is a linearfixed effects model (analysis of covariance).

I If f is linear in the parameters and the bi are assumed to berandom and uncorrelated with the predictors, then the modelis a linear mixed effects model (with random effects bi ).

Conceptually, random effects are appropriate when the observedset of individuals can be viewed as a sample from a largepopulation of individuals, while fixed effects are appropriate whenthe observed set of individuals represents the only ones aboutwhich there is interest.

Regression Trees for Longitudinal and Clustered Data

Page 7: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Longitudinal data modelingRegression trees

Modeling for large data sets

The linear mixed effects model assumes a simple parametric formfor f , which might be too restrictive an assumption; when there isa large number of individuals, a more complex functional formcould be supported. Furthermore, K may be very large, requiringmodel selection, and linear models cannot include variables withmissing values as easily as many data mining methods can.

We focus on regression trees. A regression tree is a binary tree,where each non-terminal node is split into two nodes based on thevalues of a single predictor. This method allows for interactionsbetween variables and can represent a variety of functions of thepredictors.

Regression Trees for Longitudinal and Clustered Data

Page 8: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Longitudinal data modelingRegression trees

Regression tree for National Longitudinal Survey of Youth(NLSY) logged wages data

experp < 0.001

1

≤ 3.792 > 3.792

experp < 0.001

2

≤ 2.373 > 2.373

hgcp < 0.001

3

≤ 9 > 9

ueratep < 0.001

4

≤ 8.4> 8.4

n = 882y = 1.754

5n = 720

y = 1.649

6

ueratep = 0.001

7

≤ 14.1> 14.1

n = 677y = 1.836

8n = 70

y = 1.666

9

hgcp < 0.001

10

≤ 9 > 9

ueratep = 0.006

11

≤ 6.463> 6.463

n = 248y = 1.906

12n = 445

y = 1.781

13

n = 401y = 1.917

14

experp < 0.001

15

≤ 7.466 > 7.466

ueratep < 0.001

16

≤ 5.495 > 5.495

racep = 0.001

17

{Hispanic, White}Black

hgcp = 0.021

18

≤ 9 > 9

n = 372y = 2.058

19n = 182

y = 2.183

20

n = 244y = 1.971

21

hgcp < 0.001

22

≤ 9 > 9

racep = 0.05

23

{Hispanic, White}Black

n = 689y = 1.915

24n = 150

y = 1.813

25

experp = 0.036

26

≤ 5.013> 5.013

n = 197y = 1.931

27n = 260

y = 2.045

28

hgcp < 0.001

29

≤ 9 > 9

ueratep = 0.028

30

≤ 3.195> 3.195

n = 21y = 2.458

31n = 450

y = 2.051

32

racep < 0.001

33

{Hispanic, White} Black

ueratep = 0.006

34

≤ 3.295 > 3.295

n = 18y = 2.745

35exper

p = 0.022

36

≤ 8.331> 8.331

experp = 0.021

37

≤ 8.005> 8.005

n = 63y = 2.282

38n = 33

y = 1.978

39

n = 200y = 2.346

40

n = 80y = 2.063

41

Regression Trees for Longitudinal and Clustered Data

Page 9: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Longitudinal data modelingRegression trees

Previous research

Historically most approaches to extending tree models tolongitudinal or clustered data were based on concepts frommultivariate response data (the repeated responses for a particularindividual are treated as a multivariate response from thatindividual, and the splitting criterion is modified accordingly):

I Gillo and Shelly (1974)

I Segal (1992)

I De’Ath (2002) (mvpart)

I Larsen and Speckman (2004)

I Loh and Zheng (2013) (GUIDE)

Regression Trees for Longitudinal and Clustered Data

Page 10: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Longitudinal data modelingRegression trees

Previous research

This approach has several challenges:

I It requires the same number of time points for all individuals.

I It uses a single set of predictors for all of the observation periods, which meansthat either time-varying (observation-level) predictors cannot be used, orpredictor values from later time periods can potentially be used to predictresponses from earlier ones even though that is probably contextually unrealistic.

I It cannot be used for the prediction of future periods for the same individuals ina direct way.

I Missing data is a challenge.

Hajjem et al. (2011) and Sela and Simonoff (2012) independently proposed an

approach that accounts for the longitudinal structure of the data while avoiding these

difficulties.

Regression Trees for Longitudinal and Clustered Data

Page 11: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

EstimationApplication to real dataPerformance of RE-EM trees

“EM”-type algorithm

yit = Zitbi + f (xit1, ..., xitK ) + εit

If the random effects, bi , were known, the model implies that wecould fit a regression tree to yit − Zitbi to estimate f , using forexample CART. If the fixed effects, f , were known and can berepresented as a linear function, then we could estimate therandom effects using a traditional mixed effects linear model withfixed effects corresponding to the fitted values, f (xi ). Thisalternation between the estimation of different parameters isreminiscent of (although is not) the EM algorithm, as used by Lairdand Ware (1982); for this reason, we call the resulting estimator aRandom Effects/EM Tree, or RE-EM Tree. Hajjem et al. refer tothis as the MERT (mixed effects regression tree) method.

Regression Trees for Longitudinal and Clustered Data

Page 12: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

EstimationApplication to real dataPerformance of RE-EM trees

Estimation of a RE-EM Tree

I The fitting of the regression tree uses built-in methods formissing data, such as probabilistic or surrogate split.

I The fitting of the random effects portion of the model can bebased on either independence within individuals, or a specifiedautocorrelation structure.

I Multilevel hierarchies (e.g., classrooms within schools withinschool districts within counties) are easily handled.

I Nodes are defined at the observation level, not the individuallevel; that is, different observations of the same individual endup in different (terminal) nodes. This is why observation-level(time-varying) covariates are easily accommodated.

Regression Trees for Longitudinal and Clustered Data

Page 13: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

EstimationApplication to real dataPerformance of RE-EM trees

Transaction data set

We apply this method to a dataset on third-party sellers onAmazon Web Services to predict the prices at which software titlesare sold based on the characteristics of the competing sellers(Ghose, 2005). The goal is to use the tree structure of the RE-EMtree to describe the factors that appear to influence prices. Wealso use the dataset to compare the predictive performance of theRE-EM tree to that of alternative methods through two types ofleave-one-out cross validation.

The data consist of 9484 transactions for 250 distinct softwaretitles; thus, there are I = 250 individuals in the panel with avarying number of observations Ti per individual.

Regression Trees for Longitudinal and Clustered Data

Page 14: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

EstimationApplication to real dataPerformance of RE-EM trees

Transaction data set

I Target variable: the price premium that a seller can command(the difference between the price at which the good is soldand the average price of all of the competing goods in themarketplace).

I Predictor variablesI The seller’s own reputation (total number of comments, the

number of positive and negative comments received frombuyers, the length of time that the seller has been in themarketplace)

I The characteristics of its competitors (the number ofcompetitors, the quality of competing products, and theaverage reputation of the competitors, and the average pricesof the competing products).

Regression Trees for Longitudinal and Clustered Data

Page 15: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

EstimationApplication to real dataPerformance of RE-EM trees

Tree ignoring longitudinal structure

Regression Trees for Longitudinal and Clustered Data

Page 16: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

EstimationApplication to real dataPerformance of RE-EM trees

RE-EM Tree

Regression Trees for Longitudinal and Clustered Data

Page 17: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

EstimationApplication to real dataPerformance of RE-EM trees

Cross-validated RMSE accuracy

Method Excluding ExcludingObservations Titles

Linear Model 95.88 96.92LM with RE 73.62 461.48LM with RE - AR(1) 74.75 387.18rpart 69.66 89.38RE-EM Tree 64.54 88.53RE-EM Tree - AR(1) 63.88 87.90FE-EM Tree 65.67 91.10

Regression Trees for Longitudinal and Clustered Data

Page 18: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

EstimationApplication to real dataPerformance of RE-EM trees

Properties of RE-EM trees

I When the true data generation process is a treeI RE-EM tree is best

I When the true data generation process is a linear modelI linear mixed effects model is best for small samplesI RE-EM tree is as good as the linear model with random effects

when T or I are large for most types of predictions

I When the true data generation is a complex polynomial modelwith interactions the relative performance of the tree andlinear model methods are similar to when it is a linear model.

Regression Trees for Longitudinal and Clustered Data

Page 19: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

EstimationApplication to real dataPerformance of RE-EM trees

Properties of RE-EM trees

I RE-EM tree provides more accurate estimates of randomeffects in almost all situations.

I When the true data generation process is a tree RE-EM treeprovides best estimates of true fixed effects.

I When the true data generation process is a linear orpolynomial model

I linear mixed effects model provides best estimates of true fixedeffects for small samples

I RE-EM tree is as good as linear model when T or I are large

I Autocorrelation hurts all models, but hurts linear models more.

Regression Trees for Longitudinal and Clustered Data

Page 20: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Unbiased variable selection for regression treesPerformance of unbiased RE-EM treeApplication to real data

Variable selection bias

Tree methods like CART suffer from a variable selection (splitting)bias, in that the algorithm is more likely to split on variables with alarger number of possible split points. This bias is introducedbecause the tree is constructed based on maximization of asplitting criterion over all possible splits simultaneously; that is, thechoice of which variable to split on and where the split should beare made in a single step. As a result of this, in general, standardmeasures of impurity will prefer a variable that has been randomlypartitioned into a larger number of values as a candidate forsplitting, even though the additional partition is random.

Regression Trees for Longitudinal and Clustered Data

Page 21: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Unbiased variable selection for regression treesPerformance of unbiased RE-EM treeApplication to real data

Avoiding variable selection bias

Several authors have proposed approaches that avoid this bias. Inthe multivariate response / longitudinal framework GUIDE (Loh andZheng, 2013) and MELT (Eo and Cho, 2014) use χ2 goodness-of-fittests based on residuals to assess whether a variable should besplit, with the best split set then found for that variable. In theRE-EM tree formulation, any variable selection bias comes fromthe use of CART as the underlying tree method using yit − Zit bi asthe responses, but there is no requirement that CART be used forthis; if a tree method that has unbiased variable selection is usedinstead, the resultant RE-EM tree should inherit that lack of bias.

Regression Trees for Longitudinal and Clustered Data

Page 22: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Unbiased variable selection for regression treesPerformance of unbiased RE-EM treeApplication to real data

Conditional inference trees

We replace CART with the conditional inference tree proposed byHothorn et al. (2006). This method is based on a hypothesis testingapproach, in which the process of choosing variables on which to split isstopped when the hypothesis that all of the conditional distributions of ygiven Xj equal the unconditional distribution cannot be rejected. Thetesting is based on a permutation version of each conditional distribution,addressing the bias problem (since the p-value for the test of associationof y and Xj is not related to the number of potential splitting points ofXj). The split point itself can be determined by any criterion, and unlikeCART, no pruning procedure is necessary (avoiding the randomness ofthe 10-fold cross-validation pruning procedure).

The algorithm that implements this method is available in the R

packages party and partykit.

Regression Trees for Longitudinal and Clustered Data

Page 23: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Unbiased variable selection for regression treesPerformance of unbiased RE-EM treeApplication to real data

Properties of unbiased RE-EM trees

I CART-based RE-EM trees inherit the tendency to split onvariables with more possible split points, but the unbiasedRE-EM tree completely corrects for that phenomenon.

I The unbiased tree has lower error in estimating fixed effects.

I The unbiased tree has much better performance at recoveringstructure when the true first split variable is binary and thereis a correlated continuous predictor present in the data.

Regression Trees for Longitudinal and Clustered Data

Page 24: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Unbiased variable selection for regression treesPerformance of unbiased RE-EM treeApplication to real data

NLSY logged wages data

We examine data from the National Longitudinal Survey of Youth(NLSY), focusing on wage data that were also analyzed by Singerand Willett (2003). The data consist of 888 high school dropouts,ages 14-17, with the goal being to model hourly wages (inconstant 1990 dollars). Predictors include race (White, Black, orHispanic), hgc (highest grade of schooling completed), which aretime-invariant, and exper (duration of work experience in years),ged (whether the respondent had earned a high school equivalencydegree at the time) and uerate (unemployment rate at the time).

The CART-based RE-EM tree splits only on experience at roughly3 and 7.5 years, which seems an unlikely “true” result.

Regression Trees for Longitudinal and Clustered Data

Page 25: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Unbiased variable selection for regression treesPerformance of unbiased RE-EM treeApplication to real data

NLSY logged wages data; unbiased RE-EM tree

experp < 0.001

1

≤ 2.938 > 2.938

hgcp < 0.001

2

≤ 9 > 9

experp < 0.001

3

≤ 1.417> 1.417

n = 975y = 1.691

4uerate

p = 0.004

5

≤ 5.684> 5.684

n = 187y = 1.86

6n = 731y = 1.75

7

experp = 0.002

8

≤ 1.103> 1.103

n = 340y = 1.774

9n = 568y = 1.87

10

experp < 0.001

11

≤ 7.465 > 7.465

hgcp < 0.001

12

≤ 9 > 9

ueratep < 0.001

13

≤ 5.495> 5.495

experp < 0.001

14

≤ 4.562> 4.562

n = 218y = 1.91

15n = 404

y = 2.027

16

n = 1133y = 1.893

17

experp < 0.001

18

≤ 5.276> 5.276

n = 582y = 1.977

19race

p = 0.002

20

{Hispanic, White}Black

n = 292y = 2.134

21n = 106

y = 1.994

22

hgcp < 0.001

23

≤ 9 > 9

n = 472y = 2.025

24race

p < 0.001

25

{Hispanic, White}Black

n = 314y = 2.278

26n = 80

y = 2.023

27

Regression Trees for Longitudinal and Clustered Data

Page 26: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Linear functions at nodesPerformance of MODERN treeApplication to real data

Drawback of standard regression trees

A potential weakness of these methods is that they restrict theexpected response at each terminal node to be a constant. Eo andCho (2014) proposed MELT, a regression tree method forlongitudinal data that provides an estimated slope at each terminalnode for a linear function of time. This method cannot handletime-varying covariates in a direct fashion, and since it does notprovide an estimate of the intercept cannot be used for prediction.

Regression Trees for Longitudinal and Clustered Data

Page 27: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Linear functions at nodesPerformance of MODERN treeApplication to real data

Functionals at the terminal nodes

A modification of the basic MERT/RE-EM idea that allows forlinear functions of predictors at terminal nodes rather than simplymean responses was proposed by Larocque and Simonoff (2015)and Fokkema et al. (2015). This is done through the use ofmodel-based partitioning, as is discussed in Zeileis et al. (2008).Burgin and Ritschard (2015) proposed a similar idea for ordinalmultinomial responses.

Regression Trees for Longitudinal and Clustered Data

Page 28: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Linear functions at nodesPerformance of MODERN treeApplication to real data

Implementing a linear function of time at the nodes

We will focus on the most natural situation for longitudinal data,in which the functional form is a linear function of time, producingdifferent growth curves for different subsets of observations. Theunderlying model is

yit = Zitbi + f (xit1, ..., xitk , timeit) + εit ,

The algorithm proceeds by alternating between estimating aregression tree (splitting on the x variables) with a linear functionof time (β0 + β1 × time) at each node, assuming that ourestimates of the random effects are correct, and estimating therandom effects, assuming that the model-based regression tree iscorrect. This results in a MODel-basEd RaNdom effects tree, or aMODERN tree.

Regression Trees for Longitudinal and Clustered Data

Page 29: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Linear functions at nodesPerformance of MODERN treeApplication to real data

Properties of trees

I When it can be applied MELT’s slope estimates are far morevariable than MODERN’s, and it cannot be used forprediction, so it falls short in routine applications.

I MODERN is competitive with REEM/MERT when slopes areall zero, but much better when slopes are nonzero.

I Performance of MODERN is relatively insensitive to thevalues of β0 and (nonzero) β1.

I Longitudinal trees (including MODERN) are more effectivewhen there are time-varying splitting variables, as they resultin lower variability of intercept estimates (this is related to theestimation error of b). Slope estimates are unaffected.

Regression Trees for Longitudinal and Clustered Data

Page 30: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Linear functions at nodesPerformance of MODERN treeApplication to real data

MODERN tree for NLSY logged wages data (experience astime)

hgcp < 0.001

1

≤ 9 > 9

ueratep < 0.001

2

≤ 5.395 > 5.395

ueratep = 0.019

3

≤ 3.195 > 3.195

Node 4 (n = 74)

●●●

●●

●●

●●

●●

●●●

●●●

●●

●●

●●●

●●

●●

●●

●●

−1.2 13.9

1

4

Node 5 (n = 1022)

●●●

●●

●●●

●●

●●

●●

●●●●

●●

●●

●●

●●●

●●

●●

●●●

●●●●●

●●

●●

●●●●●

●●

●●

●●

●●

●●

●●●

●●

●●●●

●●●●

●●

●●●

●●●●

●●●

●●

●●●

●●●

●●

●●

●●

●●●

●●

●●

●●●

●●

●●●

●●

●●

●●●

●●

●●●

●●

●●

●●

●●●

●●

●●●

●●

●●

●●●

●●

●●

●●

●●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●●

●●●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●●●

●●

●●

●●●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●●●

●●

●●

●●

●●●●●

●●

●●

●●

●●●●

●●●

●●●

●●

●●●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●●●●

●●

●●

●●

●●●●●●

●●

●●

●●

●●

−1.2 13.9

1

4

ueratep < 0.001

6

≤ 8.967 > 8.967

Node 7 (n = 1889)

●●

●●

●●●

●●

●●

●●●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●●●●●●●

●●

●●

●●●

●●●●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●●●

●●●

●●

●●●

●●

●●●●

●●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●●●

●●

●●

●●●●

●●

●●●

●●

●●

●●

●●

●●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●●

●●

●●

●●

●●●●

●●

●●●●

●●

●●

●●

●●●

●●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●●

●●

●●

●●

●●

●●

●●●●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●●●●

●●

●●●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●●●

●●

●●●●

●●●

●●

●●

●●

●●

●●

●●●

●●●

●●●●

●●

●●●

●●

●●

●●

●●

●●●●

●●●

●●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●●

●●

●●

●●

●●●

●●●

●●

●●

●●

●●●●●

●●●

●●●

●●

●●●●

●●

●●●●●●●

●●

●●

●●

●●●

●●

●●

●●●

●●●

●●

●●

●●

●●●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●●●

●●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●●●

●●

●●

●●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●●

●●

●●

●●●

●●

●●

●●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●●●

−1.2 13.9

1

4

Node 8 (n = 1135)

●●

●●

●●

●●

●●

●●●●

●●

●●●

●●

●●●

●●

●●

●●●

●●●●

●●●

●●

●●

●●●●●

●●●

●●

●●

●●

●●

●●

●●

●●●

●●●

●●●

●●

●●

●●

●●●

●●●●

●●●

●●●

●●

●●

●●

●●

●●

●●●

●●●

●●

●●

●●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●●

●●

●●●

●●●●

●●

●●●

●●●

●●

●●

●●

●●●●

●●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●●●

●●●

●●

●●

●●●

●●

●●

●●

●●

●●●

●●●

●●

●●●

●●

●●

●●●

●●●

●●

●●

●●

●●●●

●●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●●

●●

●●●

●●

●●

●●●●

●●

●●

●●

●●

●●

●●●●

●●

●●●

●●

●●

●●●

●●

●●

●●

●●

●●●

●●●

●●

●●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●●●

●●●

●●

●●

●●●●●

●●

●●

●●

●●

●●

●●●●●

●●

●●●

●●●

●●

●●●

●●

●●

●●

●●●●●

●●●

●●●

●●●

●●

●●

●●

●●●

●●●

−1.2 13.9

1

4

racep < 0.001

9

Black {Hispanic, White}

Node 10 (n = 650)

●●

●●

●●●●

●●

●●

●●

●●

●●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●●

●●

●●

●●●

●●●

●●●

●●

●●

●●

●●

●●

●●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●●

●●

−1.2 13.9

1

4

ueratep < 0.001

11

≤ 5.495 > 5.495

Node 12 (n = 427)

●●

●●

●●

●●

●●●

●●

●●

●●●

●●

●●

●●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●●

●●

●●

●●●●

●●●

●●

●●

●●

●●

●●●

●●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●●●

●●

●●

●●

●●

●●

●●

●●●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●

−1.2 13.9

1

4

Node 13 (n = 1205)

●●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●●●

●●

●●

●●

●●

●●

●●

●●

●●●●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●●

●●

●●●

●●

●●

●●

●●●

●●●

●●●

●●

●●●●

●●

●●●

●●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●●

●●

●●

●●

●●●●

●●

●●●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●●

●●●●●

●●

●●

●●

●●

●●

●●●●●●

●●

●●

●●

●●

●●●●

●●●

●●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●●

●●●

●●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●●●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●●●

●●

●●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●●

●●●

●●

●●

●●

●●●

●●●●

●●

●●●

●●

●●

●●●

●●

●●

●●

●●

●●

●●

●●

●●●

●●

●●

●●

●●

●●●●

●●

●●●

●●

●●

●●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

−1.2 13.9

1

4

Regression Trees for Longitudinal and Clustered Data

Page 31: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Testing for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

Linear mixed effects models and goodness-of-fit

Recall the general mixed effects model

yit = Zitbi + f (xit1, ..., xitK ) + εit .

The most common choice of f is of course the linear model

yit = Zitbi + Xitβ + εit ,

assuming errors ε that are normally distributed with constantvariance. This model has the advantage of simplicity ofinterpretation, but as is always the case, if the assumptions of themodel do not hold inferences drawn can be misleading. Suchmodel violations include nonlinearity and heteroscedasticity. Ifspecific violations are assumed, tests such as likelihood ratio testscan be constructed, but omnibus goodness-of-fit tests would beuseful to help identify unspecified model violations.

Regression Trees for Longitudinal and Clustered Data

Page 32: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Testing for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

Regression trees and goodness-of-fit

The idea discussed here is a simple one that has (perhaps) beenunderutilized through the years: since the errors are supposed tobe unstructured if the model assumptions hold, examining theresiduals using a method that looks for unspecified structure canbe used to identify model violations. A natural method for this is aregression tree.

Miller (1996) proposed using a CART regression tree for thispurpose in the context of identifying unmodeled nonlinearity inlinear least squares regression, terming it a diagnostic tree.

Regression Trees for Longitudinal and Clustered Data

Page 33: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Testing for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

Regression trees and goodness-of-fit

Su, Tsai, and Wang (2009) altered this idea slightly bysimultaneously including both linear and tree-based terms in onemodel, terming it an augmented tree, assessing whether thetree-based terms are deemed necessary in the joint model. Theyalso note that building a diagnostic tree using squared residuals asa response can be used to test for heteroscedasticity.

The diagnostic trees are not meant to replace examination ofresiduals or more focused (and powerful) tests of specific modelviolations; rather, they are an omnibus tool to add to the dataanalyst’s toolkit to try to help identify unspecified mixed effectsmodel violations.

Regression Trees for Longitudinal and Clustered Data

Page 34: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Testing for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

Proposed method

We propose adapting the diagnostic tree idea tolongitudinal/clustered data using RE-EM trees as follows:

I Fit the linear mixed effects model.

I Fit a RE-EM tree to the residuals from this model to explorenonlinearity.

I Fit a RE-EM tree to the absolute residuals from the model toexplore heteroscedasticity (squared residuals are morenon-Gaussian and lead to poorer performance).

A final tree that splits from the root node rejects the null model.The structure of the tree can help suggest the form of the violationof assumptions.

Regression Trees for Longitudinal and Clustered Data

Page 35: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Testing for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

Properties of the tests

I The test based on the standard pruning rule for RE-EM hasroughly .05 (or lower) Type I error. The unbiased tree basedon the conditional inference tree by definition has the correctsize.

I Simulations indicate good power to identify both differentslopes and polynomial terms when that nonlinear structure isunknown to the analyst.

I The test for heteroscedasticity has good power to identify itwhen it is related to a predictor or to the response eventhough that structure is not theorized.

Regression Trees for Longitudinal and Clustered Data

Page 36: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Testing for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

Spruce tree growth

Diggle, Liang, and Zeger (1994) and Venables and Ripley (2002)discuss a longitudinal growth study. The response is the log-size of79 Sitka spruce trees, two-thirds of which were grown inozone-enriched chambers, measured at five time points.

Regression Trees for Longitudinal and Clustered Data

Page 37: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Testing for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

The data

o

oo

o o

o

o

o

oo

o

oo

oo

o

o

o

o

o

o

o

o

o

o

o

o

o

o o

o

o

o

oo

o

o

o

oo

o

o

o

o

o

o

o

o

oo

o

o

o

o

o

o

o

o

ooo

o

o

oo

o

o

o

o

o

o

oo

o

o

o

o

o

oo

o

o

o

oo

o

o

oo

o

o

o

o

oo

o

o

o

oo

o

o

o

oo

o

o

o

oo

o

o

o

o

oo

o

o

o

o

o

o

o

oo

160 180 200 220 240 260

23

45

6

Time

Siz

e

o

o

o

o

o

o o

o

o o

o

o

oo o

o

o

oo o

o

o

o

o

o

o

o

o

o

o

oo

o

oo

o

o

o

oo

o

o

o

oo

o

o

o oo

o

o

oo o

o

o

o

oo

o

o

o

oo

o

o

oo

o

o

o oo

o

o

o

o

o o

oo

oo

o

o

o

oo

o

o

o

oo o

o

o

o

oo

o

oo

oo

o

o

o

oo

o

o

o

oo

o

oo o

o

o

o

o

o

o

o

o

o

o

o

o

o

o

oo

o

o

oo

o

o

o

o

o o

o

o

oo

o

o

o

o

oo

oo

o

o o

o

o

oo o

o

oo

oo

o

o

o

oo

o

o

oo o

o

o

o

oo

o

o

o

oo

o

oo

oo

o

o

o

oo

o

o

o

o o

o

o

o

o o

o

o

o

oo

o

o

o

oo

o

o

o

o oo

o

o

oo

o

o

o

o o

o

oo

oo

oo

o

o o

o

o

o

oo

o

o

o

o o

o

o

o

o o

o

oo

o o

o

o

o

o o

Regression Trees for Longitudinal and Clustered Data

Page 38: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Testing for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

The data

160 180 200 220 240 260

23

45

6

Time

Siz

e

Regression Trees for Longitudinal and Clustered Data

Page 39: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Testing for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

Linear mixed effects model

160 180 200 220 240 260

23

45

6

Time

Siz

e

Regression Trees for Longitudinal and Clustered Data

Page 40: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Testing for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

Linear mixed effects model

o

oo

o o

o

o

o

oo

o

oo

oo

o

o

o

o

o

o

o

o

o

o

o

o

o

o o

o

o

o

oo

o

o

o

oo

o

o

o

o

o

o

o

o

oo

o

o

o

o

o

o

o

o

ooo

o

o

oo

o

o

o

o

o

o

oo

o

o

o

o

o

oo

o

o

o

oo

o

o

oo

o

o

o

o

oo

o

o

o

oo

o

o

o

oo

o

o

o

oo

o

o

o

o

oo

o

o

o

o

o

o

o

oo

160 180 200 220 240 260

23

45

6

Time

Siz

e

o

o

o

o

o

o o

o

o o

o

o

oo o

o

o

oo o

o

o

o

o

o

o

o

o

o

o

oo

o

oo

o

o

o

oo

o

o

o

oo

o

o

o oo

o

o

oo o

o

o

o

oo

o

o

o

oo

o

o

oo

o

o

o oo

o

o

o

o

o o

oo

oo

o

o

o

oo

o

o

o

oo o

o

o

o

oo

o

oo

oo

o

o

o

oo

o

o

o

oo

o

oo o

o

o

o

o

o

o

o

o

o

o

o

o

o

o

oo

o

o

oo

o

o

o

o

o o

o

o

oo

o

o

o

o

oo

oo

o

o o

o

o

oo o

o

oo

oo

o

o

o

oo

o

o

oo o

o

o

o

oo

o

o

o

oo

o

oo

oo

o

o

o

oo

o

o

o

o o

o

o

o

o o

o

o

o

oo

o

o

o

oo

o

o

o

o oo

o

o

oo

o

o

o

o o

o

oo

oo

oo

o

o o

o

o

o

oo

o

o

o

o o

o

o

o

o o

o

oo

o o

o

o

o

o o

Regression Trees for Longitudinal and Clustered Data

Page 41: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Testing for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

Diagnostic tree for lack of fit

The tree-based nonlinearity test indicates lack of fit of the linearmixed effects model, related to time.

A natural alternative model is one allowing for different slopes forthe treatment and control groups, but that does not correct thelack of fit.

Regression Trees for Longitudinal and Clustered Data

Page 42: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Testing for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

Alternatives to the linear mixed model

As the diagnostic trees suggest, the problem is in the linearformulation of the effect of time. If time is treated as a categoricalpredictor, the apparent lack of fit disappears, as the diagnostic treehas no splits.

In fact, the pattern of responses in the diagnostic tree suggeststhat a quadratic term in time could account for the nonlinearity,and the diagnostic tree based on the quadratic model indicates nolack of fit.

Regression Trees for Longitudinal and Clustered Data

Page 43: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Testing for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

The data

160 180 200 220 240 260

23

45

6

Time

Siz

e

Regression Trees for Longitudinal and Clustered Data

Page 44: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Testing for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

Quadratic mixed effects model

160 180 200 220 240 260

23

45

6

Time

Siz

e

Regression Trees for Longitudinal and Clustered Data

Page 45: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Testing for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

Quadratic mixed effects model

o

oo

o o

o

o

o

oo

o

oo

oo

o

o

o

o

o

o

o

o

o

o

o

o

o

o o

o

o

o

oo

o

o

o

oo

o

o

o

o

o

o

o

o

oo

o

o

o

o

o

o

o

o

ooo

o

o

oo

o

o

o

o

o

o

oo

o

o

o

o

o

oo

o

o

o

oo

o

o

oo

o

o

o

o

oo

o

o

o

oo

o

o

o

oo

o

o

o

oo

o

o

o

o

oo

o

o

o

o

o

o

o

oo

160 180 200 220 240 260

23

45

6

Time

Siz

e

o

o

o

o

o

o o

o

o o

o

o

oo o

o

o

oo o

o

o

o

o

o

o

o

o

o

o

oo

o

oo

o

o

o

oo

o

o

o

oo

o

o

o oo

o

o

oo o

o

o

o

oo

o

o

o

oo

o

o

oo

o

o

o oo

o

o

o

o

o o

oo

oo

o

o

o

oo

o

o

o

oo o

o

o

o

oo

o

oo

oo

o

o

o

oo

o

o

o

oo

o

oo o

o

o

o

o

o

o

o

o

o

o

o

o

o

o

oo

o

o

oo

o

o

o

o

o o

o

o

oo

o

o

o

o

oo

oo

o

o o

o

o

oo o

o

oo

oo

o

o

o

oo

o

o

oo o

o

o

o

oo

o

o

o

oo

o

oo

oo

o

o

o

oo

o

o

o

o o

o

o

o

o o

o

o

o

oo

o

o

o

oo

o

o

o

o oo

o

o

oo

o

o

o

o o

o

oo

oo

oo

o

o o

o

o

o

oo

o

o

o

o o

o

o

o

o o

o

oo

o o

o

o

o

o o

Regression Trees for Longitudinal and Clustered Data

Page 46: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Testing for model violationsPerformance of tree-based lack-of-fit testsApplication to real data

Treating time as categorical

An additional interaction of the treatment and (categorical) timeeffects is statistically significant, but has higher AIC and BICvalues than the additive model, reinforcing that from a practicalpoint of view the fit of the simpler model is adequate.

Heteroscedasticity diagnostic trees for all models do not split.

Regression Trees for Longitudinal and Clustered Data

Page 47: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

Conclusion and future work

Random forests for these methods are discussed in Hajjem et al. (2014) andLarocque and Simonoff (2015).

Longitudinal data often come with a time-to-event (survival) aspect; forexample, repeat visits of patients to their doctor, with particular interest in howchanges in a patient’s health relate to survival time.

I Longitudinal data are inherently time-varying, but existing work ontime-varying survival trees is very limited. Can such trees be formulatedin an effective way? Yes — see forthcoming Arxiv paper Fu and Simonoff(2016).

I Can tree-based methods be built that jointly model longitudinal andtime-to-event data?

I Can time-to-event trees be based directly on survival time, rather thanindirectly on hazard functions?

Regression Trees for Longitudinal and Clustered Data

Page 48: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

References

I Fokkema, M., Smits, N., Zeileis, A., Hothorn, T., and Kelderman, H.(2015), “Detecting Treatment-Subgroup Interactions in Clustered DataWith Generalized Linear Mixed-Effects Model Trees,” University ofInnsbruck Working Papers in Economics and Statistics No. 2015-10.

I Fu, W. and Simonoff, J.S. (2015), “Unbiased Regression Trees forLongitudinal Data,” Computational Statistics and Data Analysis, 88,53-74.

I Hajjem, A., Bellavance, F., and Larocque, D. (2011), “Mixed EffectsRegression Trees for Clustered Data,” Statistics and Probability Letters,81, 451-459.

I Hajjem, A., Bellavance, F. and Larocque, D. (2014), “Mixed EffectsRandom Forest for Clustered Data,” Journal of Statistical Computationand Simulation, 84, 1313-1328.

Regression Trees for Longitudinal and Clustered Data

Page 49: Modern Modeling Methods Conference 2013 - Regression ......I Gillo and Shelly (1974) I Segal (1992) I De’Ath (2002) (mvpart) I Larsen and Speckman (2004) I Loh and Zheng (2013) (GUIDE)

Longitudinal data and regression treesRandom effects (RE-EM) trees

Unbiased regression treesMODel-basEd RaNdom effects (MODERN) trees

Goodness-of-fit and regression treesFuture work

References

I Larocque, D. and Simonoff, J.S. (2015), “Model-Based Mixed EffectsTrees for Longitudinal and Clustered Data,” Proceedings of the 30thInternational Workshop on Statistical Modelling, Volume 1, Linz, Austria,279-284.

I Sela, R.J. and Simonoff, J.S. (2012), “RE-EM Trees: A Data MiningApproach for Longitudinal and Clustered Data,” Machine Learning, 86,169-207.

I Simonoff, J.S. (2013), “Regression Tree-Based Diagnostics for LinearMultilevel Models,” Statistical Modelling, 13, 459-480.

The R package REEMtree used to construct RE-EM trees based on rpart is

available from CRAN. A function that adapts this to unbiased trees based on

ctree is available at people.stern.nyu.edu/jsimonof/unbiasedREEM/. The

glmertree package on R-Forge fits model-based regression trees.

Regression Trees for Longitudinal and Clustered Data


Recommended