ESALQ Course on Models for Longitudinal and Incomplete Data
|
|
|
- Pierce Beasley
- 10 years ago
- Views:
Transcription
1 ESALQ Course on Models for Longitudinal and Incomplete Data Geert Verbeke Biostatistical Centre K.U.Leuven, Belgium Geert Molenberghs Center for Statistics Universiteit Hasselt, Belgium ESALQ Course, Diepenbeek, November 2006
2 Contents 0 Related References... 1 I Continuous Longitudinal Data 6 1 Introduction A Model for Longitudinal Data Estimation and Inference in the Marginal Model Inference for the Random Effects ESALQ Course on Models for Longitudinal and Incomplete Data i
3 II Marginal Models for Non-Gaussian Longitudinal Data 64 5 The Toenail Data The Analgesic Trial Generalized Linear Models Parametric Modeling Families Generalized Estimating Equations A Family of GEE Methods III Generalized Linear Mixed Models for Non-Gaussian Longitudinal Data Generalized Linear Mixed Models (GLMM) Fitting GLMM s in SAS IV Marginal Versus Random-effects Models Marginal Versus Random-effects Models ESALQ Course on Models for Longitudinal and Incomplete Data ii
4 V Non-linear Models Non-Linear Mixed Models VI Incomplete Data Setting The Scene Proper Analysis of Incomplete Data Weighted Generalized Estimating Equations Multiple Imputation VII Topics in Methods and Sensitivity Analysis for Incomplete Data An MNAR Selection Model and Local Influence Interval of Ignorance Concluding Remarks ESALQ Course on Models for Longitudinal and Incomplete Data iii
5 VIII Case Study: Age-related Macular Degeneration The Dataset Weighted Generalized Estimating Equations Multiple Imputation ESALQ Course on Models for Longitudinal and Incomplete Data iv
6 Chapter 0 Related References Aerts, M., Geys, H., Molenberghs, G., and Ryan, L.M. (2002). Topics in Modelling of Clustered Data. London: Chapman & Hall. Brown, H. and Prescott, R. (1999). Applied Mixed Models in Medicine. New York: John Wiley & Sons. Crowder, M.J. and Hand, D.J. (1990). Analysis of Repeated Measures. London: Chapman & Hall. Davidian, M. and Giltinan, D.M. (1995). Nonlinear Models For Repeated Measurement Data. London: Chapman & Hall. ESALQ Course on Models for Longitudinal and Incomplete Data 1
7 Davis, C.S. (2002). Statistical Methods for the Analysis of Repeated Measurements. New York: Springer. Demidenko, E. (2004). Mixed Models: Theory and Applications. New York: John Wiley & Sons. Diggle, P.J., Heagerty, P.J., Liang, K.Y. and Zeger, S.L. (2002). Analysis of Longitudinal Data. (2nd edition). Oxford: Oxford University Press. Fahrmeir, L. and Tutz, G. (2002). Multivariate Statistical Modelling Based on Generalized Linear Models (2nd ed). New York: Springer. Fitzmaurice, G.M., Laird, N.M., and Ware, J.H. (2004). Applied Longitudinal Analysis. New York: John Wiley & Sons. Goldstein, H. (1979). The Design and Analysis of Longitudinal Studies. London: Academic Press. ESALQ Course on Models for Longitudinal and Incomplete Data 2
8 Goldstein, H. (1995). Multilevel Statistical Models. London: Edward Arnold. Hand, D.J. and Crowder, M.J. (1995). Practical Longitudinal Data Analysis. London: Chapman & Hall. Hedeker, D. and Gibbons, R.D. (2006). Longitudinal Data Analysis. New York: John Wiley & Sons. Jones, B. and Kenward, M.G. (1989). Design and Analysis of Crossover Trials. London: Chapman & Hall. Kshirsagar, A.M. and Smith, W.B. (1995). Growth Curves. New York: Marcel Dekker. Leyland, A.H. and Goldstein, H. (2001) Multilevel Modelling of Health Statistics. Chichester: John Wiley & Sons. Lindsey, J.K. (1993). Models for Repeated Measurements. Oxford: Oxford ESALQ Course on Models for Longitudinal and Incomplete Data 3
9 University Press. Littell, R.C., Milliken, G.A., Stroup, W.W., Wolfinger, R.D., and Schabenberger, O. (2005). SAS for Mixed Models (2nd ed.). Cary: SAS Press. Longford, N.T. (1993). Random Coefficient Models. Oxford: Oxford University Press. McCullagh, P. and Nelder, J.A. (1989). Generalized Linear Models (second edition). London: Chapman & Hall. Molenberghs, G. and Verbeke, G. (2005). Models for Discrete Longitudinal Data. New York: Springer-Verlag. Pinheiro, J.C. and Bates D.M. (2000). Mixed effects models in S and S-Plus. New York: Springer. Searle, S.R., Casella, G., and McCulloch, C.E. (1992). Variance Components. ESALQ Course on Models for Longitudinal and Incomplete Data 4
10 New-York: Wiley. Senn, S.J. (1993). Cross-over Trials in Clinical Research. Chichester: Wiley. Verbeke, G. and Molenberghs, G. (1997). Linear Mixed Models In Practice: A SAS Oriented Approach, Lecture Notes in Statistics 126. New-York: Springer. Verbeke, G. and Molenberghs, G. (2000). Linear Mixed Models for Longitudinal Data. Springer Series in Statistics. New-York: Springer. Vonesh, E.F. and Chinchilli, V.M. (1997). Linear and Non-linear Models for the Analysis of Repeated Measurements. Basel: Marcel Dekker. Weiss, R.E. (2005). Modeling Longitudinal Data. New York: Springer. Wu, H. and Zhang, J.-T. (2006). Nonparametric Regression Methods for Longitudinal Data Analysis. New York: John Wiley & Sons. ESALQ Course on Models for Longitudinal and Incomplete Data 5
11 Part I Continuous Longitudinal Data ESALQ Course on Models for Longitudinal and Incomplete Data 6
12 Chapter 1 Introduction Repeated Measures / Longitudinal data Examples ESALQ Course on Models for Longitudinal and Incomplete Data 7
13 1.1 Repeated Measures / Longitudinal Data Repeated measures are obtained when a response is measured repeatedly on a set of units Units: Subjects, patients, participants,... Animals, plants,... Clusters: families, towns, branches of a company, Special case: Longitudinal data ESALQ Course on Models for Longitudinal and Incomplete Data 8
14 1.2 Rat Data Research question (Dentistry, K.U.Leuven): How does craniofacial growth depend on testosteron production? Randomized experiment in which 50 male Wistar rats are randomized to: Control (15 rats) Low dose of Decapeptyl (18 rats) High dose of Decapeptyl (17 rats) ESALQ Course on Models for Longitudinal and Incomplete Data 9
15 Treatment starts at the age of 45 days; measurements taken every 10 days, from day 50 on. The responses are distances (pixels) between well defined points on x-ray pictures of the skull of each rat: ESALQ Course on Models for Longitudinal and Incomplete Data 10
16 Measurements with respect to the roof, base and height of the skull. Here, we consider only one response, reflecting the height of the skull. Individual profiles: ESALQ Course on Models for Longitudinal and Incomplete Data 11
17 Complication: Dropout due to anaesthesia (56%): #Observations Age (days) Control Low High Total Remarks: Much variability between rats, much less variability within rats Fixed number of measurements scheduled per subject, but not all measurements available due to dropout, for known reason. Measurements taken at fixed time points ESALQ Course on Models for Longitudinal and Incomplete Data 12
18 1.3 Prostate Data References: Carter et al (1992, Cancer Research). Carter et al (1992, Journal of the American Medical Association). Morrell et al (1995, Journal of the American Statistical Association). Pearson et al (1994, Statistics in Medicine). Prostate disease is one of the most common and most costly medical problems in the United States Important to look for markers which can detect the disease at an early stage Prostate-Specific Antigen is an enzyme produced by both normal and cancerous prostate cells ESALQ Course on Models for Longitudinal and Incomplete Data 13
19 PSA level is related to the volume of prostate tissue. Problem: Patients with Benign Prostatic Hyperplasia also have an increased PSA level Overlap in PSA distribution for cancer and BPH cases seriously complicates the detection of prostate cancer. Research question (hypothesis based on clinical practice): Can longitudinal PSA profiles be used to detect prostate cancer in an early stage? ESALQ Course on Models for Longitudinal and Incomplete Data 14
20 A retrospective case-control study based on frozen serum samples: 16 control patients 20 BPH cases 14 local cancer cases 4 metastatic cancer cases Complication: No perfect match for age at diagnosis and years of follow-up possible Hence, analyses will have to correct for these age differences between the diagnostic groups. ESALQ Course on Models for Longitudinal and Incomplete Data 15
21 Individual profiles: ESALQ Course on Models for Longitudinal and Incomplete Data 16
22 Remarks: Much variability between subjects Little variability within subjects Highly unbalanced data ESALQ Course on Models for Longitudinal and Incomplete Data 17
23 Chapter 2 A Model for Longitudinal Data Introduction The 2-stage model formulation Example: Rat data The general linear mixed-effects model Hierarchical versus marginal model Example: Rat data Example: Bivariate observations ESALQ Course on Models for Longitudinal and Incomplete Data 18
24 2.1 Introduction In practice: often unbalanced data: unequal number of measurements per subject measurements not taken at fixed time points Therefore, multivariate regression techniques are often not applicable Often, subject-specific longitudinal profiles can be well approximated by linear regression functions This leads to a 2-stage model formulation: Stage 1: Linear regression model for each subject separately Stage 2: Explain variability in the subject-specific regression coefficients using known covariates ESALQ Course on Models for Longitudinal and Incomplete Data 19
25 2.2 A 2-stage Model Formulation Stage 1 Response Y ij for ith subject, measured at time t ij, i =1,...,N, j =1,...,n i Response vector Y i for ith subject: Y i =(Y i1,y i2,...,y ini ) Stage 1 model: Y i = Z i β i + ε i ESALQ Course on Models for Longitudinal and Incomplete Data 20
26 Z i is a (n i q) matrix of known covariates β i is a q-dimensional vector of subject-specific regression coefficients ε i N(0, Σ i ),oftenσ i = σ 2 I ni Note that the above model describes the observed variability within subjects ESALQ Course on Models for Longitudinal and Incomplete Data 21
27 2.2.2 Stage 2 Between-subject variability can now be studied from relating the β i to known covariates Stage 2 model: β i = K i β + b i K i is a (q p) matrix of known covariates β is a p-dimensional vector of unknown regression parameters b i N(0,D) ESALQ Course on Models for Longitudinal and Incomplete Data 22
28 2.3 Example: The Rat Data Individual profiles: ESALQ Course on Models for Longitudinal and Incomplete Data 23
29 Transformation of the time scale to linearize the profiles: Age ij t ij =ln[1+(age ij 45)/10)] Note that t =0corresponds to the start of the treatment (moment of randomization) Stage 1 model: Y ij = β 1i + β 2i t ij + ε ij, j =1,...,n i Matrix notation: Y i = Z i β i + ε i with Z i = 1 t i1 1 t i2.. 1 t ini ESALQ Course on Models for Longitudinal and Incomplete Data 24
30 In the second stage, the subject-specific intercepts and time effects are related to the treatment of the rats Stage 2 model: β 1i = β 0 + b 1i, β 2i = β 1 L i + β 2 H i + β 3 C i + b 2i, L i, H i,andc i are indicator variables: L i = 1 if low dose 0 otherwise H i = 1 if high dose 0 otherwise C i = 1 if control 0 otherwise ESALQ Course on Models for Longitudinal and Incomplete Data 25
31 Parameter interpretation: β 0 : average response at the start of the treatment (independent of treatment) β 1, β 2,andβ 3 : average time effect for each treatment group ESALQ Course on Models for Longitudinal and Incomplete Data 26
32 2.4 The General Linear Mixed-effects Model A 2-stage approach can be performed explicitly in the analysis However, this is just another example of the use of summary statistics: Y i is summarized by β i summary statistics β i analysed in second stage The associated drawbacks can be avoided by combining the two stages into one model: Y i = Z i β i + ε i β i = K i β + b i = Y i = Z i K i } {{ } X i β + Z i b i + ε i = X i β + Z i b i + ε i ESALQ Course on Models for Longitudinal and Incomplete Data 27
33 General linear mixed-effects model: Y i = X i β + Z i b i + ε i b i N(0,D), ε i N(0, Σ i ), b 1,...,b N, ε 1,...,ε N independent Terminology: Fixed effects: β Random effects: b i Variance components: elements in D and Σ i ESALQ Course on Models for Longitudinal and Incomplete Data 28
34 2.5 Hierarchical versus Marginal Model The general linear mixed model is given by: Y i = X i β + Z i b i + ε i b i N(0,D), ε i N(0, Σ i ), b 1,...,b N, ε 1,...,ε N independent It can be rewritten as: Y i b i N(X i β + Z i b i, Σ i ), b i N(0,D) ESALQ Course on Models for Longitudinal and Incomplete Data 29
35 It is therefore also called a hierarchical model: AmodelforY i given b i Amodelforb i Marginally, we have that Y i is distributed as: Y i N(X i β,z i DZ i +Σ i ) Hence, very specific assumptions are made about the dependence of mean and covariance onn the covariates X i and Z i : Implied mean : X i β Implied covariance : V i = Z i DZ i +Σ i Note that the hierarchical model implies the marginal one, NOT vice versa ESALQ Course on Models for Longitudinal and Incomplete Data 30
36 2.6 Example: The Rat Data Stage 1 model: Y ij = β 1i + β 2i t ij + ε ij, j =1,...,n i Stage 2 model: β 1i = β 0 + b 1i, β 2i = β 1 L i + β 2 H i + β 3 C i + b 2i, Combined: Y ij =(β 0 + b 1i )+(β 1 L i + β 2 H i + β 3 C i + b 2i )t ij + ε ij = β 0 + b 1i +(β 1 + b 2i )t ij + ε ij, if low dose β 0 + b 1i +(β 2 + b 2i )t ij + ε ij, if high dose β 0 + b 1i +(β 3 + b 2i )t ij + ε ij, if control. ESALQ Course on Models for Longitudinal and Incomplete Data 31
37 Implied marginal mean structure: Linear average evolution in each group Equal average intercepts Different average slopes Implied marginal covariance structure (Σ i = σ 2 I ni ): Cov(Y i (t 1 ), Y i (t 2 )) = 1 t 1 D 1 t 2 + σ 2 δ {t1,t 2 } = d 22 t 1 t 2 + d 12 (t 1 + t 2 )+d 11 + σ 2 δ {t1,t 2 }. Note that the model implicitly assumes that the variance function is quadratic over time, with positive curvature d 22. ESALQ Course on Models for Longitudinal and Incomplete Data 32
38 A model which assumes that all variability in subject-specific slopes can be ascribed to treatment differences can be obtained by omitting the random slopes b 2i from the above model: Y ij =(β 0 + b 1i )+(β 1 L i + β 2 H i + β 3 C i )t ij + ε ij = β 0 + b 1i + β 1 t ij + ε ij, if low dose β 0 + b 1i + β 2 t ij + ε ij, if high dose β 0 + b 1i + β 3 t ij + ε ij, if control. This is the so-called random-intercepts model The same marginal mean structure is obtained as under the model with random slopes ESALQ Course on Models for Longitudinal and Incomplete Data 33
39 Implied marginal covariance structure (Σ i = σ 2 I ni ): Cov(Y i (t 1 ), Y i (t 2 )) = 1 D 1 + σ 2 δ {t1,t 2 } = d 11 + σ 2 δ {t1,t 2 }. Hence, the implied covariance matrix is compound symmetry: constant variance d 11 + σ 2 constant correlation ρ I = d 11 /(d 11 + σ 2 ) between any two repeated measurements within the same rat ESALQ Course on Models for Longitudinal and Incomplete Data 34
40 2.7 Example: Bivariate Observations Balanced data, two measurements per subject (n i =2), two models: Model 1: Random intercepts + heterogeneous errors V = 1 1 (d) (1 1) + σ σ 2 2 = d + σ 2 1 d d d+ σ 2 2 Model 2: Uncorrelated intercepts and slopes + measurement error V = d d σ σ 2 = d 1 + σ 2 d 1 d 1 d 1 + d 2 + σ 2 ESALQ Course on Models for Longitudinal and Incomplete Data 35
41 Different hierarchical models can produce the same marginal model Hence, a good fit of the marginal model cannot be interpreted as evidence for any of the hierarchical models. A satisfactory treatment of the hierarchical model is only possible within a Bayesian context. ESALQ Course on Models for Longitudinal and Incomplete Data 36
42 Chapter 3 Estimation and Inference in the Marginal Model ML and REML estimation Fitting linear mixed models in SAS Negative variance components Inference ESALQ Course on Models for Longitudinal and Incomplete Data 37
43 3.1 ML and REML Estimation Recall that the general linear mixed model equals Y i = X i β + Z i b i + ε i b i N(0,D) ε i N(0, Σ i ) independent The implied marginal model equals Y i N(X i β,z i DZ i +Σ i ) Note that inferences based on the marginal model do not explicitly assume the presence of random effects representing the natural heterogeneity between subjects ESALQ Course on Models for Longitudinal and Incomplete Data 38
44 Notation: β: vector of fixed effects (as before) α: vector of all variance components in D and Σ i θ =(β, α ) : vector of all parameters in marginal model Marginal likelihood function: L ML (θ) = N i=1 (2π) n i/2 V i (α) 1 2 exp 1 2 (Y i X i β) Vi 1 (α)(y i X i β) If α were known, MLE of β equals where W i equals V 1 i. β(α) = N i=1 X iw i X i 1 N i=1 X iw i y i, ESALQ Course on Models for Longitudinal and Incomplete Data 39
45 In most cases, α is not known, and needs to be replaced by an estimate α Two frequently used estimation methods for α: Maximum likelihood Restricted maximum likelihood ESALQ Course on Models for Longitudinal and Incomplete Data 40
46 3.2 Fitting Linear Mixed Models in SAS A model for the prostate data: ln(psa ij +1) = β 1 Age i + β 2 C i + β 3 B i + β 4 L i + β 5 M i +(β 6 Age i + β 7 C i + β 8 B i + β 9 L i + β 10 M i ) t ij +(β 11 Age i + β 12 C i + β 13 B i + β 14 L i + β 15 M i ) t 2 ij + b 1i + b 2i t ij + b 3i t 2 ij + ε ij. SAS program (Σ i = σ 2 I ni ): proc mixed data=prostate method=reml; class id group timeclss; model lnpsa = group age time group*time age*time time2 group*time2 age*time2 / solution; random intercept time time2 / type=un subject=id; run; ESALQ Course on Models for Longitudinal and Incomplete Data 41
47 ML and REML estimates for fixed effects: Effect Parameter MLE (s.e.) REMLE (s.e.) Age effect β (0.013) (0.014) Intercepts: Control β (0.919) (0.976) BPH β (1.026) (1.090) L/R cancer β (0.997) (1.059) Met. cancer β (1.022) (1.086) Age time effect β (0.020) (0.021) Time effects: Control β (1.359) (1.473) BPH β (1.511) (1.638) L/R cancer β (1.469) (1.593) Met. cancer β (1.499) (1.626) Age time 2 effect β (0.008) (0.009) Time 2 effects: Control β (0.549) (0.610) BPH β (0.604) (0.672) L/R cancer β (0.590) (0.656) Met. cancer β (0.598) (0.666) ESALQ Course on Models for Longitudinal and Incomplete Data 42
48 ML and REML estimates for variance components: Effect Parameter MLE (s.e.) REMLE (s.e.) Covariance of b i : var(b 1i ) d (0.083) (0.098) var(b 2i ) d (0.187) (0.230) var(b 3i ) d (0.032) (0.041) cov(b 1i,b 2i ) d 12 = d (0.113) (0.136) cov(b 2i,b 3i ) d 23 = d (0.076) (0.095) cov(b 3i,b 1i ) d 13 = d (0.043) (0.053) Residual variance: var(ε ij ) σ (0.002) (0.002) Log-likelihood ESALQ Course on Models for Longitudinal and Incomplete Data 43
49 Fitted average profiles at median age at diagnosis: ESALQ Course on Models for Longitudinal and Incomplete Data 44
50 3.3 Negative Variance Components Reconsider the model for the rat data: Y ij =(β 0 + b 1i )+(β 1 L i + β 2 H i + β 3 C i + b 2i )t ij + ε ij REML estimates obtained from SAS procedure MIXED: Effect Parameter REMLE (s.e.) Intercept β (0.325) Time effects: Low dose β (0.228) High dose β (0.231) Control β (0.285) Covariance of b i : var(b 1i ) d (1.123) var(b 2i ) d ( ) cov(b 1i,b 2i ) d 12 = d (0.381) Residual variance: var(ε ij ) σ (0.145) REML log-likelihood ESALQ Course on Models for Longitudinal and Incomplete Data 45
51 This suggests that the REML likelihood could be further increased by allowing negative estimates for d 22 In SAS, this can be done by adding the option nobound to the PROC MIXED statement. Results: Parameter restrictions for α d ii 0,σ 2 0 d ii IR, σ 2 IR Effect Parameter REMLE (s.e.) REMLE (s.e.) Intercept β (0.325) (0.313) Time effects: Low dose β (0.228) (0.198) High dose β (0.231) (0.198) Control β (0.285) (0.254) Covariance of b i : var(b 1i ) d (1.123) (1.019) var(b 2i ) d ( ) (0.169) cov(b 1i,b 2i ) d 12 = d (0.381) (0.357) Residual variance: var(ε ij ) σ (0.145) (0.165) REML log-likelihood ESALQ Course on Models for Longitudinal and Incomplete Data 46
52 Note that the REML log-likelihood value has been further increased and a negative estimate for d 22 is obtained. Brown & Prescott (1999, p. 237) : ESALQ Course on Models for Longitudinal and Incomplete Data 47
53 Meaning of negative variance component? Fitted variance function: Var(Y i (t)) = 1 t D 1 t + σ 2 = d 22 t 2 +2 d 12 t + d 11 + σ 2 = 0.287t t The suggested negative curvature in the variance function is supported by the sample variance function: ESALQ Course on Models for Longitudinal and Incomplete Data 48
54 This again shows that the hierarchical and marginal models are not equivalent: The marginal model allows negative variance components, as long as the marginal covariances V i = Z i DZ i + σ 2 I ni are positive definite The hierarchical interpretation of the model does not allow negative variance components ESALQ Course on Models for Longitudinal and Incomplete Data 49
55 3.4 Inference for the Marginal Model Estimate for β: β(α) = N i=1 X iw i X i with α replacedbyitsmlorremlestimate 1 N i=1 X iw i y i Conditional on α, β(α) is multivariate normal with mean β and covariance Var( β) = = N i=1 X iw i X i N i=1 X iw i X i 1 N 1 i=1 X iw i var(y i )W i X i N i=1 X iw i X i 1 In practice one again replaces α by its ML or REML estimate ESALQ Course on Models for Longitudinal and Incomplete Data 50
56 Inference for β: Wald tests t-andf -tests LR tests (not with REML) Inference for α: Wald tests LR tests (even with REML) Caution 1: Marginal vs. hierarchical model Caution 2: Boundary problems ESALQ Course on Models for Longitudinal and Incomplete Data 51
57 Chapter 4 Inference for the Random Effects Empirical Bayes inference Example: Prostate data Shrinkage Example: Prostate data ESALQ Course on Models for Longitudinal and Incomplete Data 52
58 4.1 Empirical Bayes Inference Random effects b i reflect how the evolution for the ith subject deviates from the expected evolution X i β. Estimation of the b i helpful for detecting outlying profiles This is only meaningful under the hierarchical model interpretation: Y i b i N(X i β + Z i b i, Σ i ) b i N(0,D) Since the b i are random, it is most natural to use Bayesian methods Terminology: prior distribution N(0,D) for b i ESALQ Course on Models for Longitudinal and Incomplete Data 53
59 Posterior density: f(b i y i ) f(b i Y i = y i ) = f(y i b i ) f(b i ) f(y i b i ) f(b i ) f(y i b i ) f(b i ) db i... exp 1 2 (b i DZ iw i (y i X i β)) Λ 1 i (b i DZ iw i (y i X i β)) for some positive definite matrix Λ i. Posterior distribution: b i y i N (DZ iw i (y i X i β), Λ i ) ESALQ Course on Models for Longitudinal and Incomplete Data 54
60 Posterior mean as estimate for b i : b i (θ) =E [b i Y i = y i ] = b i f(b i y i ) db i = DZ iw i (α)(y i X i β) Parameters in θ are replaced by their ML or REML estimates, obtained from fitting the marginal model. b i = b i ( θ) is called the Empirical Bayes estimate of b i. Approximate t- andf -tests to account for the variability introduced by replacing θ by θ, similar to tests for fixed effects. ESALQ Course on Models for Longitudinal and Incomplete Data 55
61 4.2 Example: Prostate Data We reconsider our linear mixed model: ln(psa ij +1) = β 1 Age i + β 2 C i + β 3 B i + β 4 L i + β 5 M i +(β 6 Age i + β 7 C i + β 8 B i + β 9 L i + β 10 M i ) t ij +(β 11 Age i + β 12 C i + β 13 B i + β 14 L i + β 15 M i ) t 2 ij + b 1i + b 2i t ij + b 3i t 2 ij + ε ij. In SAS the estimates can be obtained from adding the option solution to the random statement: random intercept time time2 / type=un subject=id solution; ods listing exclude solutionr; ods output solutionr=out; ESALQ Course on Models for Longitudinal and Incomplete Data 56
62 The ODS statements are used to write the EB estimates into a SAS output data set, and to prevent SAS from printing them in the output window. In practice, histograms and scatterplots of certain components of b i are used to detect model deviations or subjects with exceptional evolutions over time ESALQ Course on Models for Longitudinal and Incomplete Data 57
63 ESALQ Course on Models for Longitudinal and Incomplete Data 58
64 Strong negative correlations in agreement with correlation matrix corresponding to fitted D: Dcorr = Histograms and scatterplots show outliers Subjects #22, #28, #39, and #45, have highest four slopes for time 2 and smallest four slopes for time, i.e., with the strongest (quadratic) growth. Subjects #22, #28 and #39 have been further examined and have been shown to be metastatic cancer cases which were misclassified as local cancer cases. Subject #45 is the metastatic cancer case with the strongest growth ESALQ Course on Models for Longitudinal and Incomplete Data 59
65 4.3 Shrinkage Estimators b i Consider the prediction of the evolution of the ith subject: Y i X i β + Z i b i = X i β + Z i DZ iv i 1 (y i X i β) = ( I ni Z i DZ iv 1 i = Σ i Vi 1 X i β + ( I ni Σ i Vi 1 ) Xi β + Z i DZ iv i 1 y i ) yi, Hence, Ŷi is a weighted mean of the population-averaged profile X i β and the observed data y i, with weights Σ i V i 1 and I ni Σ i V i 1 respectively. ESALQ Course on Models for Longitudinal and Incomplete Data 60
66 Note that X i β gets much weight if the residual variability is large in comparison to the total variability. This phenomenon is usually called shrinkage : The observed data are shrunk towards the prior average profile X i β. This is also reflected in the fact that for any linear combination λ b i of random effects, var(λ b i ) var(λ b i ). ESALQ Course on Models for Longitudinal and Incomplete Data 61
67 4.4 Example: Prostate Data Comparison of predicted, average, and observed profiles for the subjects #15 and #28, obtained under the reduced model: ESALQ Course on Models for Longitudinal and Incomplete Data 62
68 Illustration of the shrinkage effect : Var( b i ) = , D = ESALQ Course on Models for Longitudinal and Incomplete Data 63
69 Part II Marginal Models for Non-Gaussian Longitudinal Data ESALQ Course on Models for Longitudinal and Incomplete Data 64
70 Chapter 5 The Toenail Data Toenail Dermatophyte Onychomycosis: Common toenail infection, difficult to treat, affecting more than 2% of population. Classical treatments with antifungal compounds need to be administered until the whole nail has grown out healthy. New compounds have been developed which reduce treatment to 3 months Randomized, double-blind, parallel group, multicenter study for the comparison of two such new compounds (A and B) for oral treatment. ESALQ Course on Models for Longitudinal and Incomplete Data 65
71 Research question: Severity relative to treatment of TDO? patients randomized, 36 centers 48 weeks of total follow up (12 months) 12 weeks of treatment (3 months) measurements at months 0, 1, 2, 3, 6, 9, 12. ESALQ Course on Models for Longitudinal and Incomplete Data 66
72 Frequencies at each visit (both treatments): ESALQ Course on Models for Longitudinal and Incomplete Data 67
73 Chapter 6 The Analgesic Trial single-arm trial with 530 patients recruited (491 selected for analysis) analgesic treatment for pain caused by chronic nonmalignant disease treatment was to be administered for 12 months we will focus on Global Satisfaction Assessment (GSA) GSA scale goes from 1=very good to 5=very bad GSA was rated by each subject 4 times during the trial, at months 3, 6, 9, and 12. ESALQ Course on Models for Longitudinal and Incomplete Data 68
74 Research questions: Evolution over time Relation with baseline covariates: age, sex, duration of the pain, type of pain, disease progression, Pain Control Assessment (PCA),... Investigation of dropout Frequencies: GSA Month 3 Month 6 Month 9 Month % % % % % % % % % % % % % % % % % % % 3 1.4% Tot ESALQ Course on Models for Longitudinal and Incomplete Data 69
75 Missingness: Measurement occasion Month 3 Month 6 Month 9 Month 12 Number % Completers O O O O Dropouts O O O M O O M M O M M M Non-monotone missingness O O M O O M O O O M O M O M M O M O O O M O O M M O M O M O M M ESALQ Course on Models for Longitudinal and Incomplete Data 70
76 Chapter 7 Generalized Linear Models The model Maximum likelihood estimation Examples McCullagh and Nelder (1989) ESALQ Course on Models for Longitudinal and Incomplete Data 71
77 7.1 The Generalized Linear Model Suppose a sample Y 1,...,Y N of independent observations is available All Y i have densities f(y i θ i,φ) which belong to the exponential family: f(y θ i,φ)=exp { φ 1 [yθ i ψ(θ i )] + c(y, φ) } θ i the natural parameter Linear predictor: θ i = x i β φ is the scale parameter (overdispersion parameter) ESALQ Course on Models for Longitudinal and Incomplete Data 72
78 ψ( ) is a function, generating mean and variance: E(Y )=ψ (θ) Var(Y )=φψ (θ) Note that, in general, the mean μ and the variance are related: Var(Y ) = φψ [ ψ 1 (μ) ] = φv(μ) The function v(μ) is called the variance function. The function ψ 1 which expresses θ as function of μ is called the link function. ψ is the inverse link function ESALQ Course on Models for Longitudinal and Incomplete Data 73
79 7.2 Examples The Normal Model Model: Y N(μ, σ 2 ) Density function: f(y θ, φ) = 1 2πσ 2 exp 1 σ 2(y μ)2 =exp 1 σ 2 yμ μ2 2 + ln(2πσ 2 ) 2 y2 2σ 2 ESALQ Course on Models for Longitudinal and Incomplete Data 74
80 Exponential family: θ= μ φ= σ 2 ψ(θ) =θ 2 /2 c(y, φ) = ln(2πφ) 2 y2 2φ Mean and variance function: μ= θ v(μ) =1 Note that, under this normal model, the mean and variance are not related: φv(μ) = σ 2 The link function is here the identity function: θ = μ ESALQ Course on Models for Longitudinal and Incomplete Data 75
81 7.2.2 The Bernoulli Model Model: Y Bernoulli(π) Density function: f(y θ, φ) =π y (1 π) 1 y =exp{y ln π +(1 y) ln(1 π)} =exp y ln π + ln(1 π) 1 π ESALQ Course on Models for Longitudinal and Incomplete Data 76
82 Exponential family: θ=ln ( ) π φ=1 1 π ψ(θ) = ln(1 π) = ln(1 + exp(θ)) c(y, φ) =0 Mean and variance function: μ= v(μ) = exp θ 1+exp θ = π exp θ (1+exp θ) 2 = π(1 π) Note that, under this model, the mean and variance are related: φv(μ) = μ(1 μ) The link function here is the logit link: θ =ln ( μ 1 μ ) ESALQ Course on Models for Longitudinal and Incomplete Data 77
83 7.3 Generalized Linear Models (GLM) Suppose a sample Y 1,...,Y N of independent observations is available All Y i have densities f(y i θ i,φ) which belong to the exponential family In GLM s, it is believed that the differences between the θ i can be explained through a linear function of known covariates: θ i = x i β x i is a vector of p known covariates β is the corresponding vector of unknown regression parameters, to be estimated from the data. ESALQ Course on Models for Longitudinal and Incomplete Data 78
84 7.4 Maximum Likelihood Estimation Log-likelihood: l(β,φ) = 1 φ i [y iθ i ψ(θ i )] + i c(y i,φ) The score equations: S(β) = i μ i β v 1 i (y i μ i ) = 0 Note that the estimation of β depends on the density only through the means μ i and the variance functions v i = v(μ i ). ESALQ Course on Models for Longitudinal and Incomplete Data 79
85 The score equations need to be solved numerically: iterative (re-)weighted least squares Newton-Raphson Fisher scoring Inference for β is based on classical maximum likelihood theory: asymptotic Wald tests likelihood ratio tests score tests ESALQ Course on Models for Longitudinal and Incomplete Data 80
86 7.5 Illustration: The Analgesic Trial Early dropout (did the subject drop out after the first or the second visit)? Binary response PROC GENMOD can fit GLMs in general PROC LOGISTIC can fit models for binary (and ordered) responses SAS code for logit link: proc genmod data=earlydrp; model earlydrp = pca0 weight psychiat physfct / dist=b; run; proc logistic data=earlydrp descending; model earlydrp = pca0 weight psychiat physfct; run; ESALQ Course on Models for Longitudinal and Incomplete Data 81
87 SAS code for probit link: proc genmod data=earlydrp; model earlydrp = pca0 weight psychiat physfct / dist=b link=probit; run; proc logistic data=earlydrp descending; model earlydrp = pca0 weight psychiat physfct / link=probit; run; Selected output: Analysis Of Parameter Estimates Standard Wald 95% Chi- Parameter DF Estimate Error Confidence Limits Square Pr > ChiSq Intercept PCA WEIGHT PSYCHIAT PHYSFCT Scale NOTE: The scale parameter was held fixed. ESALQ Course on Models for Longitudinal and Incomplete Data 82
88 Chapter 8 Parametric Modeling Families Continuous outcomes Longitudinal generalized linear models Notation ESALQ Course on Models for Longitudinal and Incomplete Data 83
89 8.1 Continuous Outcomes Marginal Models: E(Y ij x ij )=x ijβ Random-Effects Models: E(Y ij b i, x ij )=x ijβ + z ijb i Transition Models: E(Y ij Y i,j 1,...,Y i1, x ij )=x ijβ + αy i,j 1 ESALQ Course on Models for Longitudinal and Incomplete Data 84
90 8.2 Longitudinal Generalized Linear Models Normal case: easy transfer between models Also non-normal data can be measured repeatedly (over time) Lack of key distribution such as the normal [= ] A lot of modeling options Introduction of non-linearity No easy transfer between model families crosssectional longitudinal normal outcome linear model LMM non-normal outcome GLM? ESALQ Course on Models for Longitudinal and Incomplete Data 85
91 8.3 Marginal Models Fully specified models & likelihood estimation E.g., multivariate probit model, Bahadur model, odds ratio model Specification and/or fitting can be cumbersome Non-likelihood alternatives Historically: Empirical generalized least squares (EGLS; Koch et al 1977; SAS CATMOD) Generalized estimating equations Pseudo-likelihood ESALQ Course on Models for Longitudinal and Incomplete Data 86
92 Chapter 9 Generalized Estimating Equations General idea Asymptotic properties Working correlation Special case and application SAS code and output ESALQ Course on Models for Longitudinal and Incomplete Data 87
93 9.1 General Idea Univariate GLM, score function of the form (scalar Y i ): S(β) = N i=1 μ i β v 1 i (y i μ i )=0 with v i = Var(Y i ) In longitudinal setting: Y =(Y 1,...,Y N ): where S(β) = i j μ ij β v 1 ij (y ij μ ij ) = N i=1 D i [V i (α)] 1 (y i μ i ) = 0 D i is an n i p matrix with (i, j)th elements μ ij β y i and μ i are n i -vectors with elements y ij and μ ij Is V i n i n i diagonal or more complex? ESALQ Course on Models for Longitudinal and Incomplete Data 88
94 V i = Var(Y i ) is more complex since it involves a set of nuisance parameters α, determining the covariance structure of Y i : in which V i (β, α) =φa 1/2 i (β)r i (α)a 1/2 i (β) A 1/2 i (β) = vi1 (μ i1 (β)) v ini (μ ini (β)) and R i (α) is the correlation matrix of Y i, parameterized by α. Same form as for full likelihood procedure, but we restrict specification to the first moment only Liang and Zeger (1986) ESALQ Course on Models for Longitudinal and Incomplete Data 89
95 9.2 Large Sample Properties As N where N(ˆβ β) N(0,I 1 0 ) I 0 = N i=1 D i[v i (α)] 1 D i (Unrealistic) Conditions: α is known the parametric form for V i (α) is known This is the naive purely model based variance estimator Solution: working correlation matrix ESALQ Course on Models for Longitudinal and Incomplete Data 90
96 9.3 Unknown Covariance Structure Keep the score equations BUT S(β) = N i=1 [D i] [V i (α)] 1 (y i μ i )=0 suppose V i (.) is not the true variance of Y i but only a plausible guess, a so-called working correlation matrix specify correlations and not covariances, because the variances follow from the mean structure the score equations are solved as before ESALQ Course on Models for Longitudinal and Incomplete Data 91
97 The asymptotic normality results change to N(ˆβ β) N(0,I 1 0 I 1 I 1 0 ) I 0 = N i=1 D i[v i (α)] 1 D i I 1 = N i=1 D i[v i (α)] 1 Var(Y i )[V i (α)] 1 D i. This is the robust empirically corrected sandwich variance estimator I 0 is the bread I 1 is the filling (ham or cheese) Correct guess = likelihood variance ESALQ Course on Models for Longitudinal and Incomplete Data 92
98 The estimators ˆβ are consistent even if the working correlation matrix is incorrect An estimate is found by replacing the unknown variance matrix Var(Y i ) by (Y i ˆμ i )(Y i ˆμ i ). Even if this estimator is bad for Var(Y i ) it leads to a good estimate of I 1, provided that: replication in the data is sufficiently large same model for μ i is fitted to groups of subjects observation times do not vary too much between subjects A bad choice of working correlation matrix can affect the efficiency of ˆβ Care needed with incomplete data (see Part VI) ESALQ Course on Models for Longitudinal and Incomplete Data 93
99 9.4 The Working Correlation Matrix V i = V i (β, α,φ)=φa 1/2 i (β)r i (α)a 1/2 i (β) Variance function: A i is (n i n i ) diagonal with elements v(μ ij ),theknownglm variance function. Working correlation: R i (α) possibly depends on a different set of parameters α. Overdispersion parameter: φ, assumed 1 or estimated from the data. The unknown quantities are expressed in terms of the Pearson residuals Note that e ij depends on β. e ij = y ij μ ij v(μij ). ESALQ Course on Models for Longitudinal and Incomplete Data 94
100 9.5 Estimation of Working Correlation Liang and Zeger (1986) proposed moment-based estimates for the working correlation. Corr(Y ij,y ik ) Estimate Independence 0 Exchangeable α ˆα = 1 N AR(1) α ˆα = 1 N N i=1 1 n i (n i 1) j k e ij e ik N i=1 1 n i 1 j ni 1 e ij e i,j+1 Dispersion parameter: ˆφ = 1 N N i=1 1 n i n i j=1 e2 ij. Unstructured α jk ˆα jk = 1 N N i=1 e ij e ik ESALQ Course on Models for Longitudinal and Incomplete Data 95
101 9.6 Fitting GEE The standard procedure, implemented in the SAS procedure GENMOD. 1. Compute initial estimates for β, using a univariate GLM (i.e., assuming independence). 2. Compute Pearson residuals e ij. Compute estimates for α and φ. Compute R i (α) and V i (β, α) =φa 1/2 i (β)r i (α)a 1/2 i (β). 3. Update estimate for β: β (t+1) = β (t) N 4. Iterate until convergence. i=1 D ivi 1 D i 1 N i=1 D iv 1 Estimates of precision by means of I 1 0 and/or I 1 0 I 1 I 1 0. i (y i μ i ). ESALQ Course on Models for Longitudinal and Incomplete Data 96
102 9.7 Special Case: Linear Mixed Models Estimate for β: β(α) = i=1 X iw i X i with α replacedbyitsmlorremlestimate Conditional on α, β has mean N 1 N i=1 X iw i Y i E [ β(α) ] = N i=1 X iw i X i 1 N i=1 X iw i X i β = β provided that E(Y i )=X i β Hence, in order for β to be unbiased, it is sufficient that the mean of the response is correctly specified. ESALQ Course on Models for Longitudinal and Incomplete Data 97
103 Conditional on α, β has covariance Var( β) = N i=1 X iw i X i 1 N i=1 X iw i Var(Y i )W i X i N i=1 X iw 1 i X i = N i=1 X iw 1 i X i Note that this model-based version assumes that the covariance matrix Var(Y i ) is correctly modelled as V i = Z i DZ i +Σ i. An empirically corrected version is: Var( β) = N i=1 X iw i X i 1 } {{ } } {{ } } {{ } BREAD N i=1 X iw i Var(Y i )W i X i MEAT N i=1 X iw i X i BREAD 1 ESALQ Course on Models for Longitudinal and Incomplete Data 98
104 Chapter 10 A Family of GEE Methods Classical approach Prentice s two sets of GEE Linearization-based version GEE2 Alternating logistic regressions ESALQ Course on Models for Longitudinal and Incomplete Data 99
105 10.1 Prentice s GEE where N i=1 D iv 1 i (Y i μ i ) = 0, N i=1 E iw 1 i (Z i δ i ) = 0 Z ijk = (Y ij μ ij )(Y ik μ ik ) μij (1 μ ij )μ ik (1 μ ik ), δ ijk = E(Z ijk ) The joint asymptotic distribution of N(ˆβ β) and N( ˆα α) normal with variance-covariance matrix consistently estimated by N A 0 B C Λ 11 Λ 12 Λ 21 Λ 22 AB 0 C ESALQ Course on Models for Longitudinal and Incomplete Data 100
106 where A = B = C = N i=1 N i=1 N i=1 D iv 1 i E i W 1 i E i W 1 i 1 D i, E 1 N i i=1 1 E i, E i W 1 i Z i β N i=1 D i V 1 i 1 D i, Λ 11 = N i=1 Λ 12 = N i=1 Λ 21 = Λ 12, Λ 22 = N i=1 D iv i 1 Cov(Y i )Vi 1 D i, D iv i 1 Cov(Y i, Z i )Wi 1 E i, E iw i 1 Cov(Z i )Wi 1 E i, and Statistic Estimator Var(Y i ) (Y i μ i )(Y i μ i ) Cov(Y i, Z i ) (Y i μ i )(Z i δ i ) Var(Z i ) (Z i δ i )(Z i δ i ) ESALQ Course on Models for Longitudinal and Incomplete Data 101
107 10.2 GEE Based on Linearization Model formulation Previous version of GEE are formulated directly in terms of binary outcomes This approach is based on a linearization: with y i = μ i + ε i η i = g(μ i ), η i = X i β, Var(y i ) = Var(ε i )=Σ i. η i is a vector of linear predictors, g(.) is the (vector) link function. ESALQ Course on Models for Longitudinal and Incomplete Data 102
108 Estimation (Nelder and Wedderburn 1972) Solve iteratively: where N i=1 X iw i X i β = N i=1 W iy i, W i = F iσ 1 i F i, y i = ˆη i +(y i ˆμ i )F 1 i, F i = μ i η i, Σ i = Var(ε), μ i = E(y i ). Remarks: y i is called working variable or pseudo data. Basis for SAS macro and procedure GLIMMIX For linear models, D i = I ni and standard linear regression follows. ESALQ Course on Models for Longitudinal and Incomplete Data 103
109 The Variance Structure Σ i = φa 1/2 i (β)r i (α)a 1/2 i (β) φ is a scale (overdispersion) parameter, A i = v(μ i ), expressing the mean-variance relation (this is a function of β), R i (α) describes the correlation structure: If independence is assumed then R i (α) =I ni. Other structures, such as compound symmetry, AR(1),...can be assumed as well. ESALQ Course on Models for Longitudinal and Incomplete Data 104
110 10.3 GEE2 Model: Marginal mean structure Pairwise association: Odds ratios Correlations Working assumptions: Third and fourth moments Estimation: Second-order estimating equations Likelihood (assuming 3rd and 4th moments are correctly specified) ESALQ Course on Models for Longitudinal and Incomplete Data 105
111 10.4 Alternating Logistic Regression Diggle, Heagerty, Liang, and Zeger (2002) and Molenberghs and Verbeke (2005) When marginal odds ratios are used to model association, α can be estimated using ALR, which is almost as efficient as GEE2 almost as easy (computationally) than GEE1 μ ijk as before and α ijk =ln(ψ ijk ) the marginal log odds ratio: logit Pr(Y ij =1 x ij )=x ij β logit Pr(Y ij =1 Y ik = y ik )=α ijk y ik +ln μ ij μ ijk 1 μ ij μ ik + μ ijk ESALQ Course on Models for Longitudinal and Incomplete Data 106
112 α ijk can be modelled in terms of predictors the second term is treated as an offset the estimating equations for β and α are solved in turn, and the alternating between both sets is repeated until convergence. this is needed because the offset clearly depends on β. ESALQ Course on Models for Longitudinal and Incomplete Data 107
113 10.5 Application to the Toenail Data The model Consider the model: Y ij Bernoulli(μ ij ), log μ ij 1 μ ij = β 0 + β 1 T i + β 2 t ij + β 3 T i t ij Y ij : severe infection (yes/no) at occasion j for patient i t ij : measurement time for occasion j T i : treatment group ESALQ Course on Models for Longitudinal and Incomplete Data 108
114 Standard GEE SAS Code: proc genmod data=test descending; class idnum timeclss; model onyresp = treatn time treatn*time / dist=binomial; repeated subject=idnum / withinsubject=timeclss type=exch covb corrw modelse; run; SAS statements: The REPEATED statements defines the GEE character of the model. type= : working correlation specification (UN, AR(1), EXCH, IND,...) modelse : model-based s.e. s on top of default empirically corrected s.e. s corrw : printout of working correlation matrix withinsubject= : specification of the ordering within subjects ESALQ Course on Models for Longitudinal and Incomplete Data 109
115 Selected output: Regression parameters: Analysis Of Initial Parameter Estimates Standard Wald 95% Chi- Parameter DF Estimate Error Confidence Limits Square Intercept treatn time treatn*time Scale Estimates from fitting the model, ignoring the correlation structure, i.e., from fitting a classical GLM to the data, using proc GENMOD. The reported log-likelihood also corresponds to this model, and therefore should not be interpreted. The reported estimates are used as starting values in the iterative estimation procedure for fitting the GEE s. ESALQ Course on Models for Longitudinal and Incomplete Data 110
116 Analysis Of GEE Parameter Estimates Empirical Standard Error Estimates Standard 95% Confidence Parameter Estimate Error Limits Z Pr > Z Intercept treatn time <.0001 treatn*time Analysis Of GEE Parameter Estimates Model-Based Standard Error Estimates Standard 95% Confidence Parameter Estimate Error Limits Z Pr > Z Intercept <.0001 treatn time <.0001 treatn*time The working correlation: Exchangeable Working Correlation Correlation ESALQ Course on Models for Longitudinal and Incomplete Data 111
117 Alternating Logistic Regression type=exch logor=exch Note that α now is a genuine parameter ESALQ Course on Models for Longitudinal and Incomplete Data 112
118 Selected output: Analysis Of GEE Parameter Estimates Empirical Standard Error Estimates Standard 95% Confidence Parameter Estimate Error Limits Z Pr > Z Intercept treatn time <.0001 treatn*time Alpha <.0001 Analysis Of GEE Parameter Estimates Model-Based Standard Error Estimates Standard 95% Confidence Parameter Estimate Error Limits Z Pr > Z Intercept treatn time <.0001 treatn*time ESALQ Course on Models for Longitudinal and Incomplete Data 113
119 Linearization Based Method GLIMMIX macro: %glimmix(data=test, procopt=%str(method=ml empirical), stmts=%str( class idnum timeclss; model onyresp = treatn time treatn*time / solution; repeated timeclss / subject=idnum type=cs rcorr; ), error=binomial, link=logit); GLIMMIX procedure: proc glimmix data=test method=rspl empirical; class idnum; model onyresp (event= 1 ) = treatn time treatn*time / dist=binary solution; random _residual_ / subject=idnum type=cs; run; ESALQ Course on Models for Longitudinal and Incomplete Data 114
120 Both produce the same results The GLIMMIX macro is a MIXED core, with GLM-type surrounding statements The GLIMMIX procedure does not call MIXED, it has its own engine PROC GLIMMIX combines elements of MIXED and of GENMOD RANDOM residual is the PROC GLIMMIX way to specify residual correlation ESALQ Course on Models for Longitudinal and Incomplete Data 115
121 Results of Models Fitted to Toenail Data Effect Par. IND EXCH UN GEE1 Int. β (0.109;0.171) (0.134;0.173) (0.166;0.173) T i β (0.157;0.251) 0.012(0.187;0.261) 0.072(0.235;0.246) t ij β (0.025;0.030) (0.021;0.031) (0.028;0.029) T i t ij β (0.039;0.055) (0.036;0.057) (0.047;0.052) ALR Int. β (0.157;0.169) T i β (0.222;0.243) t ij β (0.023;0.030) T i t ij β (0.039;0.052) Ass. α 3.222( ;0.291) Linearization based method Int. β (0.112;0.171) (0.142;0.174) (0.171;0.172) T i β (0.160;0.251) 0.011(0.196;0.262) 0.036(0.242;0.242) t ij β (0.025;0.030) (0.022;0.031) (0.038;0.034) T i t ij β (0.040:0.055) (0.038;0.057) (0.058;0.058) estimate (model-based s.e.; empirical s.e.) ESALQ Course on Models for Longitudinal and Incomplete Data 116
122 Discussion GEE1: All empirical standard errors are correct, but the efficiency is higher for the more complex working correlation structure, as seen in p-values for T i t ij effect: Structure p-value IND EXCH UN Thus, opting for reasonably adequate correlation assumptions still pays off, in spite of the fact that all are consistent and asymptotically normal Similar conclusions for linearization-based method ESALQ Course on Models for Longitudinal and Incomplete Data 117
123 Model-based s.e. and empirically corrected s.e. in reasonable agreement for UN Typically, the model-based standard errors are much too small as they are based on the assumption that all observations in the data set are independent, hereby overestimating the amount of available information, hence also overestimating the precision of the estimates. ALR: similar inferences but now also α part of the inferences ESALQ Course on Models for Longitudinal and Incomplete Data 118
124 Part III Generalized Linear Mixed Models for Non-Gaussian Longitudinal Data ESALQ Course on Models for Longitudinal and Incomplete Data 119
125 Chapter 11 Generalized Linear Mixed Models (GLMM) Introduction: LMM Revisited Generalized Linear Mixed Models (GLMM) Fitting Algorithms Example ESALQ Course on Models for Longitudinal and Incomplete Data 120
126 11.1 Introduction: LMM Revisited We re-consider the linear mixed model: Y i b i N(X i β + Z i b i, Σ i ), b i N(0,D) The implied marginal model equals Y i N(X i β,z i DZ i +Σ i ) Hence, even under conditional independence, i.e., all Σ i equal to σ 2 I ni,amarginal association structure is implied through the random effects. The same ideas can now be applied in the context of GLM s to model association between discrete repeated measures. ESALQ Course on Models for Longitudinal and Incomplete Data 121
127 11.2 Generalized Linear Mixed Models (GLMM) Given a vector b i of random effects for cluster i, it is assumed that all responses Y ij are independent, with density f(y ij θ ij,φ)=exp { φ 1 [y ij θ ij ψ(θ ij )] + c(y ij,φ) } θ ij is now modelled as θ ij = x ij β + z ij b i As before, it is assumed that b i N(0,D) Let f ij (y ij b i, β,φ) denote the conditional density of Y ij given b i, the conditional density of Y i equals f i (y i b i, β,φ)= n i j=1 f ij(y ij b i, β,φ) ESALQ Course on Models for Longitudinal and Incomplete Data 122
128 The marginal distribution of Y i is given by f i (y i β,d,φ)= f i (y i b i, β,φ) f(b i D) db i = n i j=1 f ij(y ij b i, β,φ) f(b i D) db i where f(b i D) is the density of the N(0,D) distribution. The likelihood function for β, D, andφ now equals L(β,D,φ)= N i=1 f i(y i β,d,φ) = N i=1 n i j=1 f ij(y ij b i, β,φ) f(b i D) db i ESALQ Course on Models for Longitudinal and Incomplete Data 123
129 Under the normal linear model, the integral can be worked out analytically. In general, approximations are required: Approximation of integrand Approximation of data Approximation of integral Predictions of random effects can be based on the posterior distribution f(b i Y i = y i ) Empirical Bayes (EB) estimate : Posterior mode, with unknown parameters replaced by their MLE ESALQ Course on Models for Longitudinal and Incomplete Data 124
130 11.3 Laplace Approximation of Integrand Integrals in L(β,D,φ) can be written in the form I = e Q(b) db Second-order Taylor expansion of Q(b) around the mode yields Q(b) Q( b)+ 1 2 (b b) Q ( b)(b b), Quadratic term leads to re-scaled normal density. Hence, I (2π) q/2 Q ( b) 1/2 e Q( b). Exact approximation in case of normal kernels Good approximation in case of many repeated measures per subject ESALQ Course on Models for Longitudinal and Incomplete Data 125
131 11.4 Approximation of Data General Idea Re-write GLMM as: Y ij = μ ij + ε ij = h(x ijβ + z ijb i )+ε ij with variance for errors equal to Var(Y ij b i )=φv(μ ij ) Linear Taylor expansion for μ ij : Penalized quasi-likelihood (PQL): Around current β and b i Marginal quasi-likelihood (MQL): Around current β and b i = 0 ESALQ Course on Models for Longitudinal and Incomplete Data 126
132 Penalized quasi-likelihood (PQL) Linear Taylor expansion around current β and b i : Y ij h(x ij β + z ij b i ) + h (x ij β + z ij b i )x ij(β β) + h (x ij β + z ij b i )z ij(b i b i )+ε ij μ ij + v( μ ij )x ij(β β)+v( μ ij )z ij(b i b i )+ε ij In vector notation: Y i μ i + V i X i (β β)+ V i Z i (b i b i )+ε i Re-ordering terms yields: Y 1 i V i (Y i μ i )+X i β + Z i b i X i β + Z i b i + ε i, Model fitting by iterating between updating the pseudo responses Y i and fitting the above linear mixed model to them. ESALQ Course on Models for Longitudinal and Incomplete Data 127
133 Marginal quasi-likelihood (MQL) Linear Taylor expansion around current β and b i = 0: Y ij h(x ij β) + h (x ij β)x ij(β β) + h (x ij β)z ijb i + ε ij μ ij + v( μ ij )x ij(β β)+v( μ ij )z ijb i + ε ij In vector notation: Y i μ i + V i X i (β β)+ V i Z i b i + ε i Re-ordering terms yields: Y 1 i V i (Y i μ i )+X i β X i β + Z i b i + ε i Model fitting by iterating between updating the pseudo responses Y i and fitting the above linear mixed model to them. ESALQ Course on Models for Longitudinal and Incomplete Data 128
134 PQL versus MQL MQL only performs reasonably well if random-effects variance is (very) small Both perform bad for binary outcomes with few repeated measurements per cluster With increasing number of measurements per subject: MQL remains biased PQL consistent Improvements possible with higher-order Taylor expansions ESALQ Course on Models for Longitudinal and Incomplete Data 129
135 11.5 Approximation of Integral The likelihood contribution of every subject is of the form f(z)φ(z)dz where φ(z) is the density of the (multivariate) normal distribution Gaussian quadrature methods replace the integral by a weighted sum: f(z)φ(z)dz Q q=1 w qf(z q ) Q is the order of the approximation. The higher Q the more accurate the approximation will be ESALQ Course on Models for Longitudinal and Incomplete Data 130
136 The nodes (or quadrature points) z q are solutions to the Qth order Hermite polynomial The w q are well-chosen weights The nodes z q and weights w q are reported in tables. Alternatively, an algorithm is available for calculating all z q and w q for any value Q. With Gaussian quadrature, the nodes and weights are fixed, independent of f(z)φ(z). With adaptive Gaussian quadrature, the nodes and weights are adapted to the support of f(z)φ(z). ESALQ Course on Models for Longitudinal and Incomplete Data 131
137 Graphically (Q =10): ESALQ Course on Models for Longitudinal and Incomplete Data 132
138 Typically, adaptive Gaussian quadrature needs (much) less quadrature points than classical Gaussian quadrature. On the other hand, adaptive Gaussian quadrature is much more time consuming. Adaptive Gaussian quadrature of order one is equivalent to Laplace transformation. ESALQ Course on Models for Longitudinal and Incomplete Data 133
139 11.6 Example: Toenail Data Y ij is binary severity indicator for subject i at visit j. Model: Y ij b i Bernoulli(π ij ), log π ij 1 π ij = β 0 + b i + β 1 T i + β 2 t ij + β 3 T i t ij Notation: T i : treatment indicator for subject i t ij : time point at which jth measurement is taken for ith subject Adaptive as well as non-adaptive Gaussian quadrature, for various Q. ESALQ Course on Models for Longitudinal and Incomplete Data 134
140 Results: Gaussian quadrature Q =3 Q =5 Q =10 Q =20 Q =50 β (0.31) (0.39) (0.32) (0.69) (0.43) β (0.38) 0.19 (0.36) 0.47 (0.36) (0.80) (0.57) β (0.03) (0.04) (0.05) (0.05) (0.05) β (0.05) (0.07) (0.07) (0.07) (0.07) σ 2.26 (0.12) 3.09 (0.21) 4.53 (0.39) 3.86 (0.33) 4.04 (0.39) 2l Adaptive Gaussian quadrature Q =3 Q =5 Q =10 Q =20 Q =50 β (0.59) (0.40) (0.45) (0.43) (0.44) β (0.64) (0.54) (0.59) (0.59) (0.59) β (0.05) (0.04) (0.05) (0.05) (0.05) β (0.07) (0.07) (0.07) (0.07) (0.07) σ 4.51 (0.62) 3.70 (0.34) 4.07 (0.43) 4.01 (0.38) 4.02 (0.38) 2l ESALQ Course on Models for Longitudinal and Incomplete Data 135
141 Conclusions: (Log-)likelihoods are not comparable Different Q can lead to considerable differences in estimates and standard errors For example, using non-adaptive quadrature, with Q =3, we found no difference in time effect between both treatment groups (t = 0.09/0.05,p=0.0833). Using adaptive quadrature, with Q =50, we find a significant interaction between the time effect and the treatment (t = 0.16/0.07,p=0.0255). Assuming that Q =50is sufficient, the final results are well approximated with smaller Q under adaptive quadrature, but not under non-adaptive quadrature. ESALQ Course on Models for Longitudinal and Incomplete Data 136
142 Comparison of fitting algorithms: Adaptive Gaussian Quadrature, Q =50 MQL and PQL Summary of results: Parameter QUAD PQL MQL Intercept group A 1.63 (0.44) 0.72 (0.24) 0.56 (0.17) Intercept group B 1.75 (0.45) 0.72 (0.24) 0.53 (0.17) Slope group A 0.40 (0.05) 0.29 (0.03) 0.17 (0.02) Slope group B 0.57 (0.06) 0.40 (0.04) 0.26 (0.03) Var. random intercepts (τ 2 ) (3.02) 4.71 (0.60) 2.49 (0.29) Severe differences between QUAD (gold standard?) and MQL/PQL. MQL/PQL may yield (very) biased results, especially for binary data. ESALQ Course on Models for Longitudinal and Incomplete Data 137
143 Chapter 12 Fitting GLMM s in SAS Proc GLIMMIX for PQL and MQL Proc NLMIXED for Gaussian quadrature ESALQ Course on Models for Longitudinal and Incomplete Data 138
144 12.1 Procedure GLIMMIX for PQL and MQL Re-consider logistic model with random intercepts for toenail data SAS code (PQL): proc glimmix data=test method=rspl ; class idnum; model onyresp (event= 1 ) = treatn time treatn*time / dist=binary solution; random intercept / subject=idnum; run; MQL obtained with option method=rmpl Inclusion of random slopes: random intercept time / subject=idnum type=un; ESALQ Course on Models for Longitudinal and Incomplete Data 139
145 12.2 Procedure NLMIXED for Gaussian Quadrature Re-consider logistic model with random intercepts for toenail data SAS program (non-adaptive, Q =3): proc nlmixed data=test noad qpoints=3; parms beta0=-1.6 beta1=0 beta2=-0.4 beta3=-0.5 sigma=3.9; teta = beta0 + b + beta1*treatn + beta2*time + beta3*timetr; expteta = exp(teta); p = expteta/(1+expteta); model onyresp ~ binary(p); random b ~ normal(0,sigma**2) subject=idnum; run; Adaptive Gaussian quadrature obtained by omitting option noad ESALQ Course on Models for Longitudinal and Incomplete Data 140
146 Automatic search for optimal value of Q in case of no option qpoints= Good starting values needed! The inclusion of random slopes can be specified as follows: proc nlmixed data=test noad qpoints=3; parms beta0=-1.6 beta1=0 beta2=-0.4 beta3=-0.5 d11=3.9 d12=0 d22=0.1; teta = beta0 + b1 + beta1*treatn + beta2*time + b2*time + beta3*timetr; expteta = exp(teta); p = expteta/(1+expteta); model onyresp ~ binary(p); random b1 b2 ~ normal([0, 0], [d11, d12, d22]) subject=idnum; run; ESALQ Course on Models for Longitudinal and Incomplete Data 141
147 Part IV Marginal Versus Random-effects Models ESALQ Course on Models for Longitudinal and Incomplete Data 142
148 Chapter 13 Marginal Versus Random-effects Models Interpretation of GLMM parameters Marginalization of GLMM Example: Toenail data ESALQ Course on Models for Longitudinal and Incomplete Data 143
149 13.1 Interpretation of GLMM Parameters: Toenail Data We compare our GLMM results for the toenail data with those from fitting GEE s (unstructured working correlation): GLMM GEE Parameter Estimate (s.e.) Estimate (s.e.) Intercept group A (0.4356) (0.1656) Intercept group B (0.4478) (0.1671) Slope group A (0.0460) (0.0277) Slope group B (0.0601) (0.0380) ESALQ Course on Models for Longitudinal and Incomplete Data 144
150 The strong differences can be explained as follows: Consider the following GLMM: Y ij b i Bernoulli(π ij ), log = β 0 + b i + β 1 t ij 1 π ij The conditional means E(Y ij b i ), as functions of t ij,aregivenby π ij E(Y ij b i ) = exp(β 0 + b i + β 1 t ij ) 1+exp(β 0 + b i + β 1 t ij ) ESALQ Course on Models for Longitudinal and Incomplete Data 145
151 The marginal average evolution is now obtained from averaging over the random effects: exp(β 0 + b i + β 1 t ij ) E(Y ij ) = E[E(Y ij b i )] = E 1 + exp(β 0 + b i + β 1 t ij ) exp(β 0 + β 1 t ij ) 1 + exp(β 0 + β 1 t ij ) ESALQ Course on Models for Longitudinal and Incomplete Data 146
152 Hence, the parameter vector β in the GEE model needs to be interpreted completely different from the parameter vector β in the GLMM: GEE: marginal interpretation GLMM: conditional interpretation, conditionally upon level of random effects In general, the model for the marginal average is not of the same parametric form as the conditional average in the GLMM. For logistic mixed models, with normally distributed random random intercepts, it can be shown that the marginal model can be well approximated by again a logistic model, but with parameters approximately satisfying β RE β M = c 2 σ > 1, σ 2 = variance random intercepts c =16 3/(15π) ESALQ Course on Models for Longitudinal and Incomplete Data 147
153 For the toenail application, σ was estimated as , such that the ratio equals c2 σ = The ratio s between the GLMM and GEE estimates are: GLMM GEE Parameter Estimate (s.e.) Estimate (s.e.) Ratio Intercept group A (0.4356) (0.1656) Intercept group B (0.4478) (0.1671) Slope group A (0.0460) (0.0277) Slope group B (0.0601) (0.0380) Note that this problem does not occur in linear mixed models: Conditional mean: E(Y i b i )=X i β + Z i b i Specifically: E(Y i b i = 0) =X i β Marginal mean: E(Y i )=X i β ESALQ Course on Models for Longitudinal and Incomplete Data 148
154 The problem arises from the fact that, in general, E[g(Y )] g[e(y )] So, whenever the random effects enter the conditional mean in a non-linear way, the regression parameters in the marginal model need to be interpreted differently from the regression parameters in the mixed model. In practice, the marginal mean can be derived from the GLMM output by integrating out the random effects. This can be done numerically via Gaussian quadrature, or based on sampling methods. ESALQ Course on Models for Longitudinal and Incomplete Data 149
155 13.2 Marginalization of GLMM: Toenail Data As an example, we plot the average evolutions based on the GLMM output obtained in the toenail example: P (Y ij =1) = E E exp( b i t ij ), 1+exp( b i t ij ) exp( b i t ij ), 1+exp( b i t ij ) ESALQ Course on Models for Longitudinal and Incomplete Data 150
156 Average evolutions obtained from the GEE analyses: P (Y ij =1) = exp( t ij ) 1+exp( t ij ) exp( t ij ) 1+exp( t ij ) ESALQ Course on Models for Longitudinal and Incomplete Data 151
157 In a GLMM context, rather than plotting the marginal averages, one can also plot the profile for an average subject, i.e., a subject with random effect b i = 0: P (Y ij =1 b i =0) = exp( t ij ) 1+exp( t ij ) exp( t ij ) 1+exp( t ij ) ESALQ Course on Models for Longitudinal and Incomplete Data 152
158 13.3 Example: Toenail Data Revisited Overview of all analyses on toenail data: Parameter QUAD PQL MQL GEE Intercept group A 1.63 (0.44) 0.72 (0.24) 0.56 (0.17) 0.72 (0.17) Intercept group B 1.75 (0.45) 0.72 (0.24) 0.53 (0.17) 0.65 (0.17) Slope group A 0.40 (0.05) 0.29 (0.03) 0.17 (0.02) 0.14 (0.03) Slope group B 0.57 (0.06) 0.40 (0.04) 0.26 (0.03) 0.25 (0.04) Var. random intercepts (τ 2 ) (3.02) 4.71 (0.60) 2.49 (0.29) Conclusion: GEE < MQL < PQL < QUAD ESALQ Course on Models for Longitudinal and Incomplete Data 153
159 Part V Non-linear Models ESALQ Course on Models for Longitudinal and Incomplete Data 154
160 Chapter 14 Non-Linear Mixed Models From linear to non-linear models Orange Tree Example Song Bird Example ESALQ Course on Models for Longitudinal and Incomplete Data 155
161 14.1 From Linear to Non-linear Models We have studied: linear models generalized linear models for Gaussian data non-gaussian data for cross-sectional (univariate) data for correlated data (clustered data, multivariate data, longitudinal data) In all cases, a certain amount of linearity is preserved: E.g., in generalized linear models, linearity operates at the level of the linear predictor, describing a transformation of the mean (logit, log, probit,...) This implies that all predictor functions are linear in the parameters: β 0 + β 1 x 1i +...β p x pi ESALQ Course on Models for Longitudinal and Incomplete Data 156
162 This may still be considered a limiting factor, and non-linearity in the parametric functions may be desirable. Examples: Certain growth curves, especially when they include both a growth spurt as well as asymptote behavior towards the end of growth Dose-response modeling Pharmacokinetic and pharmacodynamic models ESALQ Course on Models for Longitudinal and Incomplete Data 157
163 14.2 LMM and GLMM In linear mixed models, the mean is modeled as a linear function of regression parameters and random effects: E(Y ij b i )=x ij β + z ij b i In generalized linear mixed models, apart from a link function, the mean is again modeled as a linear function of regression parameters and random effects: E(Y ij b i )=h(x ij β + z ij b i ) In some applications, models are needed, in which the mean is no longer modeled as a function of a linear predictor x ij β + z ij b i. These are called non-linear mixed models. ESALQ Course on Models for Longitudinal and Incomplete Data 158
164 14.3 Non-linear Mixed Models In non-linear mixed models, it is assumed that the conditional distribution of Y ij, given b i is belongs to the exponential family (Normal, Binomial, Poisson,...). The mean is modeled as: E(Y ij b i )=h(x ij, β, z ij, b i ) As before, the random effects are assumed to be normally distributed, with mean 0 and covariance D. Let f ij (y ij b i, β,φ) be the conditional density of Y ij given b i,andletf(b i D) be the density of the N(0,D) distribution. ESALQ Course on Models for Longitudinal and Incomplete Data 159
165 Under conditional independence, the marginal likelihood equals L(β,D,φ)= N i=1 f i(y i β,d,φ) = N i=1 n i j=1 f ij(y ij b i, β,φ) f(b i D) db i The above likelihood is of the same form as the likelihood obtained earlier for a generalized linear mixed model. As before, numerical quadrature is used to approximate the integral Non-linear mixed models can also be fitted within the SAS procedure NLMIXED. ESALQ Course on Models for Longitudinal and Incomplete Data 160
166 14.4 Example: Orange Trees We consider an experiment in which the trunk circumference (in mm) is measured for 5 orange trees, on 7 different occasions. Data: Response Day Tree 1 Tree 2 Tree 3 Tree 4 Tree ESALQ Course on Models for Longitudinal and Incomplete Data 161
167 The following non-linear mixed model has been proposed: Y ij = β 1 + b i 1 + exp[ (t ij β 2 )/β 3 ] + ε ij b i N(0,σ 2 b) ε ij N(0,σ 2 ) ESALQ Course on Models for Longitudinal and Incomplete Data 162
168 In SAS PROC NLMIXED, the model can be fitted as follows: proc nlmixed data=tree; parms beta1=190 beta2=700 beta3=350 sigmab=10 sigma=10; num = b; ex = exp(-(day-beta2)/beta3); den = 1 + ex; model y ~ normal(num/den,sigma**2); random b ~ normal(beta1,sigmab**2) subject=tree; run; An equivalent program is given by proc nlmixed data=tree; parms beta1=190 beta2=700 beta3=350 sigmab=10 sigma=10; num = b + beta1; ex = exp(-(day-beta2)/beta3); den = 1 + ex; model y ~ normal(num/den,sigma**2); random b ~ normal(0,sigmab**2) subject=tree; run; ESALQ Course on Models for Longitudinal and Incomplete Data 163
169 Selected output: Specifications Data Set Dependent Variable Distribution for Dependent Variable Random Effects Distribution for Random Effects Subject Variable Optimization Technique Integration Method WORK.TREE y Normal b Normal tree Dual Quasi-Newton Adaptive Gaussian Quadrature Parameter Estimates Standard Parameter Estimate Error DF t Value Pr > t beta beta <.0001 beta sigmab sigma ESALQ Course on Models for Longitudinal and Incomplete Data 164
170 Note that the number of quadrature points was selected adaptively to be equal to only one. Refitting the model with 50 quadrature points yielded identical results. Empirical Bayes estimates, and subject-specific predictions can be obtained as follows: proc nlmixed data=tree; parms beta1=190 beta2=700 beta3=350 sigmab=10 sigma=10; num = b + beta1; ex = exp(-(day-beta2)/beta3); den = 1 + ex; ratio=num/den; model y ~ normal(ratio,sigma**2); random b ~ normal(0,sigmab**2) subject=tree out=eb; predict ratio out=ratio; run; ESALQ Course on Models for Longitudinal and Incomplete Data 165
171 We can now compare the observed data to the subject-specific predictions ŷ ij = β 1 + b i 1+exp[ (t ij β 2 )/ β 3 ] ESALQ Course on Models for Longitudinal and Incomplete Data 166
172 14.5 Example: Song Birds At the University of Antwerp, a novel in-vivo MRI approach to discern the functional characteristics of specific neuronal populations in a strongly connected brain circuitry has been established. Of particular interest: the song control system (SCS) in songbird brain. The high vocal center (HVC), one of the major nuclei in this circuit, contains interneurons and two distinct types of neurons projecting respectively to the nucleus robustus archistriatalis, RA or to area X. ESALQ Course on Models for Longitudinal and Incomplete Data 167
173 Schematically, ESALQ Course on Models for Longitudinal and Incomplete Data 168
174 The MRI Data T1-weighted multi slice SE images were acquired (Van Meir et al 2003). After acquisition of a set of control images, MnCl 2 was injected in the cannulated HVC. Two sets of 5 coronal slices (one through HVC and RA, one through area X) were acquired every 15 min for up to 6 7 hours after injection. This results in 30 data sets of 10 slices of each bird (5 controls and 5 testosterone treated). The change of relative SI of each time point is calculated from the mean signal intensity determined within the defined regions of interest (area X and RA) and in an adjacent control area. ESALQ Course on Models for Longitudinal and Incomplete Data 169
175 Initial Study The effect of testosterone (T) on the dynamics of Mn 2+ accumulation in RA and area X of female starling has been established. This has been done with dynamic ME-MRI in individual birds injected with Manganese in their HVC. This was done in a 2-stage approach: The individual SI data, determined as the mean intensity of a region of interest, were submitted to a sigmoid curve fitting providing curve parameters which revealed upon a two-way repeated ANOVA analysis that neurons projecting to RA and those to area X were affected differently by testosterone treatment. However, this approach could be less reliable if the fit-parameters were mutually dependent. ESALQ Course on Models for Longitudinal and Incomplete Data 170
176 ESALQ Course on Models for Longitudinal and Incomplete Data 171
177 Thus: an integrated non-linear modeling approach is necessary: the MRI signal intensities (SI) need to be fitted to a non-linear mixed model assuming that the SI follow a pre-specified distribution depending on a covariate indicating the time after MnCl 2 injection and parameterized through fixed and random (bird-specific) effects. An initial model needs to be proposed and then subsequently simplified. ESALQ Course on Models for Longitudinal and Incomplete Data 172
178 A Model for RA in the Second Period Let RA ij be the measurement at time j for bird i. The following initial non-linear model is assumed: RA ij = (φ 0 + φ 1 G i + f i )T η 0+η 1 G i +n i ij (τ 0 + τ 1 G i + t i ) η 0+η 1 G i +n i + T η 0+η 1 G i +n i ij +γ 0 + γ 1 G i + ε ij The following conventions are used: G i is an indicator for group membership (0 for control, 1 for active) T ij is a covariate indicating the measurement time The fixed effects parameters: ESALQ Course on Models for Longitudinal and Incomplete Data 173
179 the intercept parameters φ 0, η 0, τ 0,andγ 0 the group effect parameters φ 1, η 1, τ 1,andγ 1 If the latter are simultaneously zero, there is no difference between both groups. If at least one of them is (significantly) different from zero, then the model indicates a difference between both groups. There are three bird-specific or random effects, f i, n i,andt i, following a trivariate zero-mean normal and general covariance matrix D The residual error terms ε ij are assumed to be mutually independent and independent from the random effects, and to follow a N(0,σ 2 ) distribution. Thus, the general form of the model has 8 fixed-effects parameters, and 7 variance components (3 variances in D, 3 covariances in D, andσ 2 ). ESALQ Course on Models for Longitudinal and Incomplete Data 174
180 SAS program: data hulp2; set m.vincent03; if time <= 0 then delete; if periode = 1 then delete; run; proc nlmixed data=hulp2 qpoints=3; parms phim=0.64 phimdiff=0 eta=1.88 etadiff=0 tau=3.68 taudiff=0 gamma=-0.01 gdiff=0 d11=0.01 sigma2=0.01 d22=0.01 d33=0.01; teller = (phim + phimdiff * groep + vm ) * (time ** (eta + etadiff * groep + n )); noemer = ((tau + t + taudiff * groep ) ** (eta + etadiff * groep +n) ) + (time ** (eta + etadiff * groep + n)); gemid = teller/noemer + gamma + gdiff * groep; model si_ra ~ normal(gemid,sigma2); random vm t n ~ normal([0, 0, 0],[d11,0,d22, 0, 0, d33]) subject=vogel out=m.eb; run; Gaussian quadrature is used. ESALQ Course on Models for Longitudinal and Incomplete Data 175
181 The covariances in D are set equal to zero, due to computational difficulty. Fitting the model produces the following output. First, the bookkeeping information is provided: Specifications Data Set Dependent Variable Distribution for Dependent Variable Random Effects Distribution for Random Effects Subject Variable Optimization Technique Integration Method WORK.HULP2 SI_RA Normal vm t n Normal vogel Dual Quasi-Newton Adaptive Gaussian Quadrature Dimensions Observations Used 262 Observations Not Used 58 Total Observations 320 Subjects 10 Max Obs Per Subject 28 Parameters 12 Quadrature Points 3 ESALQ Course on Models for Longitudinal and Incomplete Data 176
182 Next, starting values and the corresponding likelihood are provided: Parameters phim phimdiff eta etadiff tau taudiff gamma gdiff d sigma2 d22 d33 NegLogLike The iteration history tells convergence is not straightforward: Iteration History Iter Calls NegLogLike Diff MaxGrad Slope ESALQ Course on Models for Longitudinal and Incomplete Data 177
183 E E-6 NOTE: GCONV convergence criterion satisfied. ESALQ Course on Models for Longitudinal and Incomplete Data 178
184 The fit statistics can be used for model comparison, as always: Fit Statistics -2 Log Likelihood AIC (smaller is better) AICC (smaller is better) BIC (smaller is better) Finally, parameter estimates are provided: Parameter Estimates Standard Parameter Estimate Error DF t Value Pr > t Alpha Lower Upper Gradient phim phimdiff eta < E-6 etadiff tau < taudiff gamma gdiff d sigma < d d ESALQ Course on Models for Longitudinal and Incomplete Data 179
185 Next, a backward selection is conducted, using likelihood ratio tests. First, the variance d 33 of n i was removed. The corresponding test statistic has a χ 2 0:1 null distribution (p =0.4387). Next, fixed-effect parameters γ 0, γ 1, τ 1,andφ 1 are removed. The final model is: RA ij = (φ 0 + f i )T η 0+η 1 G i ij (τ 0 + t i ) η 0+η 1 G i + T η 0+η 1 G i ij + ε ij ESALQ Course on Models for Longitudinal and Incomplete Data 180
186 Overview of parameter estimates and standard errors for initial and final model: Estimate (s.e.) Effect Parameter Initial Final φ (0.0655) (0.0478) φ (0.0923) η (0.1394) (0.0802) η (0.1865) (0.1060) τ (0.2346) (0.1761) τ (0.3250) γ (0.0044) γ (0.0059) Var(f i ) d (0.0092) (0.0101) Var(t i ) d (0.1169) (0.1338) Var(n i ) d (0.0349) Var(ε ij ) σ ( ) ( ) ESALQ Course on Models for Longitudinal and Incomplete Data 181
187 AREA.X at the Second Period The same initial model will be fitted to AREA ij. In this case, a fully general model can be fitted, including also the covariances between the random effects. Code to do so: proc nlmixed data=hulp2 qpoints=3; parms phim= phimdiff= eta= etadiff= tau= taudiff= gamma= gdiff= sigma2= d11= d22= d33= d12=0 d13=0 d23=0; teller = (phim + phimdiff * groep + vm) * (time ** (eta + etadiff * groep + n )); noemer = ((tau + taudiff * groep +t ) ** (eta + etadiff * groep + n) ) + (time ** (eta + etadiff * groep +n)); gemid = teller/noemer + gamma + gdiff * groep; model si_area_x ~ normal(gemid,sigma2); random vm t n ~ normal([0, 0, 0],[d11, d12, d22, d13, d23, d33]) subject=vogel out=m.eb2; run; ESALQ Course on Models for Longitudinal and Incomplete Data 182
188 Model simplification: First, the random n i effect is removed (implying the removal of d 11, d 12,and d 13 ), using a likelihood ratio test statistic with value 4.08 and null distribution χ 2 2:3. The corresponding p = Removal of the random t i effect is not possible since the likelihood ratio equals on 1:2 degrees of freedom (p <0.0001). Removal of the covariance between the random t i and f i effects is not possible (G 2 =4.35 on 1 d.f., p =0.0371). Next, the following fixed-effect parameters were removed: γ 0, γ 1, η 1 and τ 1. The fixed-effect φ 1 was found to be highly significant and therefore could not be removed from the model (G 2 = on 1 d.f., p =0.0011). This indicates a significant difference between the two groups. The resulting final model is: AREA ij = (φ 0 + φ 1 G i + f i )T η 0 ij (τ 0 + t i ) η 0 + T η 0 ij + ε ij. ESALQ Course on Models for Longitudinal and Incomplete Data 183
189 Estimate (s.e.) Effect Parameter Initial Final φ (0.0333) (0.0261) φ (0.0458) (0.0312) η (0.5390) (0.1498) η (0.5631) τ (0.5662) (0.3262) τ (0.6226) γ (0.0032) γ (0.0048) Var(f i ) d (0.0020) (0.0022) Var(t i ) d (0.2365) (0.2881) Var(n i ) d (0.1858) Cov(f i,t i ) d (0.0205) (0.0229) Cov(f i,n i ) d (0.0159) Cov(t i,n i ) d (0.1615) Var(ε ij ) σ ( ) ( ) ESALQ Course on Models for Longitudinal and Incomplete Data 184
190 Time Individual curves, showing data points as well as empirical Bayes predictions for the bird-specific curves, show that the sigmoidal curves describe the data quite well. SI.area.X Time SI.area.X Time SI.area.X Time SI.area.X Time SI.area.X Time Time SI.area.X Time Time SI.area.X Time SI.area.X ESALQ Course on Models for Longitudinal and Incomplete Data 185 SI.area.X SI.area.X
191 SI.area.X Group 0 Group 1 Marg. Group 0 Marg. Goup 1 We can also explore all individual as well as marginal average fitted curves per group. The marginal effect was obtained using the sampling-based method. It is clear that Area.X is higher for most treated birds (group 1) compared to the untreated birds (group 0) Time ESALQ Course on Models for Longitudinal and Incomplete Data 186
192 Part VI Incomplete Data ESALQ Course on Models for Longitudinal and Incomplete Data 187
193 Chapter 15 Setting The Scene Orthodontic growth data Notation Taxonomies ESALQ Course on Models for Longitudinal and Incomplete Data 188
194 15.1 Growth Data Taken from Potthoff and Roy, Biometrika (1964) Research question: Is dental growth related to gender? The distance from the center of the pituitary to the maxillary fissure was recorded at ages 8, 10, 12, and 14, for 11 girls and 16 boys ESALQ Course on Models for Longitudinal and Incomplete Data 189
195 Individual profiles: Much variability between girls / boys Considerable variability within girls / boys Fixed number of measurements per subject Measurements taken at fixed time points ESALQ Course on Models for Longitudinal and Incomplete Data 190
196 15.2 Incomplete Longitudinal Data ESALQ Course on Models for Longitudinal and Incomplete Data 191
197 15.3 Scientific Question In terms of entire longitudinal profile In terms of last planned measurement In terms of last observed measurement ESALQ Course on Models for Longitudinal and Incomplete Data 192
198 15.4 Notation Subject i at occasion (time) j =1,...,n i Measurement Y ij Missingness indicator R ij = 1 if Y ij is observed, 0 otherwise. Group Y ij into a vector Y i =(Y i1,...,y ini ) =(Y o i, Y m i ) Y o i contains Y ij for which R ij =1, Y m i contains Y ij for which R ij =0. Group R ij into a vector R i =(R i1,...,r ini ) D i :timeofdropout:d i =1+ n i j=1 R ij ESALQ Course on Models for Longitudinal and Incomplete Data 193
199 15.5 Framework f(y i,d i θ, ψ) Selection Models: f(y i θ) f(d i Y o i, Y m i, ψ) MCAR MAR MNAR f(d i ψ) f(d i Y o i, ψ) f(d i Y o i, Y m i, ψ) Pattern-Mixture Models: f(y i D i, θ)f(d i ψ) Shared-Parameter Models: f(y i b i, θ)f(d i b i, ψ) ESALQ Course on Models for Longitudinal and Incomplete Data 194
200 f(y i,d i θ, ψ) Selection Models: f(y i θ) f(d i Y o i, Y m i, ψ) MCAR MAR MNAR CC? direct likelihood! joint model!? LOCF? expectation-maximization (EM). sensitivity analysis?! imputation? multiple imputation (MI).. (weighted) GEE! ESALQ Course on Models for Longitudinal and Incomplete Data 195
201 Chapter 16 Proper Analysis of Incomplete Data Simple methods Bias for LOCF and CC Direct likelihood inference Weighted generalized estimating equations ESALQ Course on Models for Longitudinal and Incomplete Data 196
202 16.1 Incomplete Longitudinal Data ESALQ Course on Models for Longitudinal and Incomplete Data 197
203 Data and Modeling Strategies ESALQ Course on Models for Longitudinal and Incomplete Data 198
204 Modeling Strategies ESALQ Course on Models for Longitudinal and Incomplete Data 199
205 16.2 Simple Methods MCAR Complete case analysis: delete incomplete subjects Standard statistical software Loss of information Impact on precision and power Missingness MCAR bias Last observation carried forward: impute missing values Standard statistical software Increase of information Constant profile after dropout: unrealistic Usually bias ESALQ Course on Models for Longitudinal and Incomplete Data 200
206 Quantifying the Bias Dropouts t ij =0 Probability p 0 Treatment indicator T i =0, 1 E(Y ij )=β 0 + β 1 T i + β 2 t ij + β 3 T i t ij Completers t ij =0, 1 Probability 1 p 0 = p 1 Treatment indicator T i =0, 1 E(Y ij )=γ 0 +γ 1 T i +γ 2 t ij +γ 3 T i t ij CC LOCF MCAR 0 (p 1 p 0 )β 2 (1 p 1 )β 3 σ[(1 p 1 )(β 0 + β 1 γ 0 γ 1 ) p 1 (γ 0 + γ 1 + γ 2 + γ 3 )+(1 p 1 )(β 0 + β 1 ) MAR (1 p 0 )(β 0 γ 0 )] p 0 (γ 0 + γ 2 ) (1 p 0 )β 0 γ 1 γ 3 σ[(1 p 1 )(β 0 + β 1 γ 0 γ 1 ) (1 p 0 )(β 0 γ 0 )] ESALQ Course on Models for Longitudinal and Incomplete Data 201
207 16.3 Direct Likelihood Maximization MAR : f(y o i θ) f(d i Y o i, ψ) ignorability Mechanism is MAR θ and ψ distinct Interest in θ Use observed information matrix = Likelihood inference is valid Outcome type Modeling strategy Software Gaussian Linear mixed model SAS proc MIXED Non-Gaussian Generalized linear mixed model SAS proc GLIMMIX, NLMIXED ESALQ Course on Models for Longitudinal and Incomplete Data 202
208 Original, Complete Orthodontic Growth Data Mean Covar # par 1 unstructured unstructured 18 2 slopes unstructured 14 3 = slopes unstructured 13 7 slopes CS 6 ESALQ Course on Models for Longitudinal and Incomplete Data 203
209 Trimmed Growth Data: Simple Methods Method Model Mean Covar # par Complete case 7a = slopes CS 5 LOCF 2a quadratic unstructured 16 Unconditional mean 7a = slopes CS 5 Conditional mean 1 unstructured unstructured 18 distorting ESALQ Course on Models for Longitudinal and Incomplete Data 204
210 Trimmed Growth Data: Direct Likelihood Mean Covar # par 7 slopes CS 6 ESALQ Course on Models for Longitudinal and Incomplete Data 205
211 Growth Data: Comparison of Analyses Data Complete cases LOCF imputed data Model Unstructured group by time mean Unstructured covariance matrix All available data Analysis methods Direct likelihood ML REML MANOVA ANOVA per time point ESALQ Course on Models for Longitudinal and Incomplete Data 206
212 Principle Method Boys at Age 8 Boys at Age 10 Original Direct likelihood, ML (0.56) (0.49) Direct likelihood, REML MANOVA (0.58) (0.51) ANOVA per time point (0.61) (0.53) Direct Lik. Direct likelihood, ML (0.56) (0.68) Direct likelihood, REML (0.58) (0.71) MANOVA (0.48) (0.66) ANOVA per time point (0.61) (0.74) CC Direct likelihood, ML (0.45) (0.62) Direct likelihood, REML MANOVA (0.48) (0.66) ANOVA per time point (0.51) (0.74) LOCF Direct likelihood, ML (0.56) (0.65) Direct likelihood, REML MANOVA (0.58) (0.68) ANOVA per time point (0.61) (0.72) ESALQ Course on Models for Longitudinal and Incomplete Data 207
213 Behind the Scenes R completers N R incompleters Y i1 Y i2 N μ 1 μ 2, σ 11 σ 12 σ 22 Conditional density Y i2 y i1 N(β 0 + β 1 y i1,σ 22.1 ) μ 1 freq. & lik. μ1 = 1 N μ 2 frequentist μ2 = 1 R μ 2 likelihood μ2 = 1 N N i=1 y i1 R i=1 y i2 R i=1 y i2 + N i=r+1 [ y2 + β 1 (y i1 y 1 ) ] ESALQ Course on Models for Longitudinal and Incomplete Data 208
214 Growth Data: Further Comparison of Analyses Principle Method Boys at Age 8 Boys at Age 10 Original Direct likelihood, ML (0.56) (0.49) Direct Lik. Direct likelihood, ML (0.56) (0.68) CC Direct likelihood, ML (0.45) (0.62) LOCF Direct likelihood, ML (0.56) (0.65) Data Mean Covariance Boys at Age 8 Boys at Age 10 Complete Unstructured Unstructured Unstructured CS Unstructured Independence Incomplete Unstructured Unstructured Unstructured CS Unstructured Independence ESALQ Course on Models for Longitudinal and Incomplete Data 209
215 ESALQ Course on Models for Longitudinal and Incomplete Data 210
216 Growth Data: SAS Code for Model 1 IDNR AGE SEX MEASURE SAS code: proc mixed data = growth method = ml; class sex idnr age; model measure = sex age*sex / s; repeated age / type = un subject = idnr; run; Subjects in terms of IDNR blocks ageensures proper ordering of observations within subjects! ESALQ Course on Models for Longitudinal and Incomplete Data 211
217 Growth Data: SAS Code for Model 2 IDNR AGE SEX MEASURE SAS code: data help; set growth; agecat = age; run; proc mixed data = growth method = ml; class sex idnr agecat; model measure = sex age*sex / s; repeated agecat / type = un subject = idnr; run; Time ordering variable needs to be categorical ESALQ Course on Models for Longitudinal and Incomplete Data 212
218 Chapter 17 Weighted Generalized Estimating Equations General Principle Analysis of the analgesic trial ESALQ Course on Models for Longitudinal and Incomplete Data 213
219 17.1 General Principle MAR and non-ignorable! Standard GEE inference correct only under MCAR Under MAR: weighted GEE Robins, Rotnitzky & Zhao (JASA, 1995) Fitzmaurice, Molenberghs & Lipsitz (JRSSB, 1995) Decompose dropout time D i =(R i1,...,r in )=(1,...,1, 0,...,0) ESALQ Course on Models for Longitudinal and Incomplete Data 214
220 Weigh a contribution by inverse dropout probability ν idi P [D i = d i ]= d i 1 (1 P [R ik =0 R i2 =...= R i,k 1 =1]) k=2 P [R idi =0 R i2 =...= R i,di 1 =1] I{d i T } Adjust estimating equations N i=1 1 ν idi μ i β V 1 i (y i μ i ) = 0 ESALQ Course on Models for Longitudinal and Incomplete Data 215
221 17.2 Computing the Weights Predicted values from (PROC GENMOD) output The weights are now defined at the individual measurement level: At the first occasion, the weight is w =1 At other than the last ocassion, the weight is the already accumulated weight, multiplied by 1 the predicted probability At the last occasion within a sequence where dropout occurs the weight is multiplied by the predicted probability At the end of the process, the weight is inverted ESALQ Course on Models for Longitudinal and Incomplete Data 216
222 17.3 Analysis of the Analgesic Trial A logistic regression for the dropout indicator: logit[p (D i = j D i j, )] = ψ 0 + ψ 11 I(GSA i,j 1 =1)+ψ 12 I(GSA i,j 1 =2) +ψ 13 I(GSA i,j 1 =3)+ψ 14 I(GSA i,j 1 =4) +ψ 2 PCA0 i + ψ 3 PF i + ψ 4 GD i with GSA i,j 1 the 5-point outcome at the previous time I( ) is an indicator function PCA0 i is pain control assessment at baseline PF i is physical functioning at baseline GD i is genetic disorder at baseline are used) ESALQ Course on Models for Longitudinal and Incomplete Data 217
223 Effect Par. Estimate (s.e.) Intercept ψ (0.49) Previous GSA= 1 ψ (0.41) Previous GSA= 2 ψ (0.38) Previous GSA= 3 ψ (0.37) Previous GSA= 4 ψ (0.38) Basel. PCA ψ (0.10) Phys. func. ψ (0.004) Genetic disfunc. ψ (0.24) There is some evidence for MAR: P (D i = j D i j) depends on previous GSA. Furthermore: baseline PCA, physical functioning and genetic/congenital disorder. ESALQ Course on Models for Longitudinal and Incomplete Data 218
224 GEE and WGEE: logit[p (Y ij =1 t j, PCA0 i )] = β 1 + β 2 t j + β 3 t 2 j + β 4 PCA0) i Effect Par. GEE WGEE Intercept β (0.47) 2.17 (0.69) Time β (0.33) (0.44) Time 2 β (0.07) 0.12 (0.09) Basel. PCA β (0.10) (0.13) A hint of potentially important differences between both ESALQ Course on Models for Longitudinal and Incomplete Data 219
225 Working correlation matrices: R UN, GEE = R UN, WGEE = ESALQ Course on Models for Longitudinal and Incomplete Data 220
226 17.4 Analgesic Trial: Steps for WGEE in SAS 1. Preparatory data manipulation: %dropout(...) 2. Logistic regression for weight model: proc genmod data=gsac; class prevgsa; model dropout = prevgsa pca0 physfct gendis / pred dist=b; ods output obstats=pred; run; 3. Conversion of predicted values to weights:... %dropwgt(...) ESALQ Course on Models for Longitudinal and Incomplete Data 221
227 4. Weighted GEE analysis: proc genmod data=repbin.gsaw; scwgt wi; class patid timecls; model gsabin = time time pca0 / dist=b; repeated subject=patid / type=un corrw within=timecls; run; ESALQ Course on Models for Longitudinal and Incomplete Data 222
228 Chapter 18 Multiple Imputation General idea Use of MI in practice Analysis of the growth data ESALQ Course on Models for Longitudinal and Incomplete Data 223
229 18.1 General Principles Valid under MAR An alternative to direct likelihood and WGEE Three steps: 1. The missing values are filled in M times = M complete data sets 2. The M complete data sets are analyzed by using standard procedures 3. The results from the M analyses are combined into a single inference Rubin (1987), Rubin and Schenker (1986), Little and Rubin (1987) ESALQ Course on Models for Longitudinal and Incomplete Data 224
230 18.2 Use of MI in Practice Many analyses of the same incomplete set of data A combination of missing outcomes and missing covariates As an alternative to WGEE: MI can be combined with classical GEE MI in SAS: Imputation Task: PROC MI Analysis Task: PROC MYFAVORITE Inference Task: PROC MIANALYZE ESALQ Course on Models for Longitudinal and Incomplete Data 225
231 MI Analysis of the Orthodontic Growth Data The same Model 1 as before Focus on boys at ages 8 and 10 Results Boys at Age 8 Boys at Age 10 Original Data (0.58) (0.51) Multiple Imputation (0.66) (0.81) Between-imputation variability for age 10 measurement Confidence interval for Boys at age 10: [21.08,24.29] ESALQ Course on Models for Longitudinal and Incomplete Data 226
232 Part VII Topics in Methods and Sensitivity Analysis for Incomplete Data ESALQ Course on Models for Longitudinal and Incomplete Data 227
233 Chapter 19 An MNAR Selection Model and Local Influence The Diggle and Kenward selection model Mastitis in dairy cattle An informal sensitivity analysis Local influence to conduct sensitivity analysis ESALQ Course on Models for Longitudinal and Incomplete Data 228
234 19.1 A Full Selection Model MNAR : f(y i θ)f(d i Y i, ψ)dy m i f(y i θ) Linear mixed model f(d i Y i, ψ) Logistic regressions for dropout Y i = X i β + Z i b i + ε i logit [P (D i = j D i j, Y i,j 1,Y ij )] = ψ 0 + ψ 1 Y i,j 1 + ψ 2 Y ij Diggle and Kenward (JRSSC 1994) ESALQ Course on Models for Longitudinal and Incomplete Data 229
235 19.2 Mastitis in Dairy Cattle Infectious disease of the udder Leads to a reduction in milk yield High yielding cows more susceptible? But this cannot be measured directly because of the effect of the disease: evidence is missing since infected cause have no reported milk yield ESALQ Course on Models for Longitudinal and Incomplete Data 230
236 Model for milk yield: Y i1 Y i2 N μ μ +Δ, σ 2 1 ρσ 1 σ 2 ρσ 1 σ 2 σ 2 1 Model for mastitis: logit [P (R i =1 Y i1,y i2 )] = ψ 0 + ψ 1 Y i1 + ψ 2 Y i2 = Y i1 2.54Y i2 = Y i1 2.54(Y i2 Y i1 ) LR test for H 0 : ψ 2 =0:G 2 =5.11 ESALQ Course on Models for Longitudinal and Incomplete Data 231
237 19.3 Criticism Sensitivity Analysis...,estimatingthe unestimable canbeaccomplished only by making modelling assumptions,... The consequences of model misspecification will(...) bemoresevereinthenon-randomcase. (Laird 1994) Change distributional assumptions (Kenward 1998) Local and global influence methods Pattern-mixture models Several plausible models or ranges of inferences Semi-parametric framework (Scharfstein et al 1999) ESALQ Course on Models for Longitudinal and Incomplete Data 232
238 19.4 Kenward s Sensitivity Analysis Deletion of #4 and #5 G 2 for ψ 2 : Cows #4 and #5 have unusually large increments Kenward conjectures: #4 and #5 ill during the first year Kenward (SiM 1998) ESALQ Course on Models for Longitudinal and Incomplete Data 233
239 19.5 Local Influence Verbeke, Thijs, Lesaffre, Kenward (Bcs 2001) Perturbed MAR dropout model: logit [P (D i =1 Y i1,y i2 )] = ψ 0 + ψ 1 Y i1 + ω i Y i2 Likelihood displacement: LD(ω) = 2 [ L ω=0 ( θ, ψ ) L ω=0 ( θω, ψ ) ] ω 0 ESALQ Course on Models for Longitudinal and Incomplete Data 234
240 19.5 Local Influence Verbeke, Thijs, Lesaffre, Kenward (Bcs 2001) Perturbed MAR dropout model: logit [P (D i =1 Y i1,y i2 )] = ψ 0 + ψ 1 Y i1 + ω i Y i2 or ψ 0 + ψ 1 Y i1 + ω i (Y i2 Y i1 ) Likelihood displacement: LD(ω) = 2 [ L ω=0 ( θ, ψ ) L ω=0 ( θω, ψ ) ] ω 0 ESALQ Course on Models for Longitudinal and Incomplete Data 235
241 19.6 Application to Mastitis Data Removing #4, #5 and #66 G 2 = h max : different signs for (#4,#5) and #66 ESALQ Course on Models for Longitudinal and Incomplete Data 236
242 Chapter 20 Interval of Ignorance The Slovenian Public Opinion Survey MAR and MNAR analyses Informal sensitivity analysis Interval of ignorance & interval of uncertainty ESALQ Course on Models for Longitudinal and Incomplete Data 237
243 20.1 The Slovenian Plebiscite Rubin, Stern, and Vehovar (1995) Slovenian Public Opinion (SPO) Survey Four weeks prior to decisive plebiscite Three questions: 1. Are you in favor of Slovenian independence? 2. Are you in favor of Slovenia s secession from Yugoslavia? 3. Will you attend the plebiscite? Political decision: ABSENCE NO Primary Estimand: θ: Proportion in favor of independence ESALQ Course on Models for Longitudinal and Incomplete Data 238
244 Slovenian Public Opinion Survey Data: Independence Secession Attendance Yes No Yes Yes No No Yes No Yes No ESALQ Course on Models for Longitudinal and Incomplete Data 239
245 20.2 Slovenian Public Opinion: 1st Analysis Pessimistic: All who can say NO will say NO θ = =0.694 Optimistic: All who can say YES will say YES θ = = =0.904 Resulting Interval: θ [0.694; 0.904] ESALQ Course on Models for Longitudinal and Incomplete Data 240
246 Resulting Interval: θ [0.694; 0.904] Complete cases: All who answered on 3 questions θ = =0.928? Available cases: All who answered on both questions θ = =0.929? ESALQ Course on Models for Longitudinal and Incomplete Data 241
247 20.3 Slovenian Public Opinion: 2nd Analysis Missing at Random: Non-response is allowed to depend on observed, but not on unobserved outcomes: Based on two questions: Based on three questions: θ =0.892 θ =0.883 Missing Not at Random (NI): Non-response is allowed to depend on unobserved measurements: θ =0.782 ESALQ Course on Models for Longitudinal and Incomplete Data 242
248 20.4 Slovenian Public Opinion Survey Estimator θ Pessimistic bound Optimistic bound Complete cases 0.928? Available cases 0.929? MAR (2 questions) MAR (3 questions) MNAR ESALQ Course on Models for Longitudinal and Incomplete Data 243
249 20.5 Slovenian Plebiscite: The Truth? θ =0.885 Estimator θ Pessimistic bound Optimistic bound Complete cases 0.928? Available cases 0.929? MAR (2 questions) MAR (3 questions) MNAR ESALQ Course on Models for Longitudinal and Incomplete Data 244
250 20.6 Did the MNAR model behave badly? Consider a family of MNAR models Baker, Rosenberger, and DerSimonian (1992) Counts Y r1 r 2 jk j, k =1, 2 indicates YES/NO r 1,r 2 =0, 1 indicates MISSING/OBSERVED ESALQ Course on Models for Longitudinal and Incomplete Data 245
251 Model Formulation E(Y 11jk )=m jk, E(Y 10jk )=m jk β jk, E(Y 01jk )=m jk α jk, E(Y 00jk )=m jk α jk β jk γ jk, Interpretation: α jk : models non-response on independence question β jk : models non-response on attendance question γ jk : interaction between both non-response indicators (cannot depend on j or k) ESALQ Course on Models for Longitudinal and Incomplete Data 246
252 Identifiable Models Model Structure d.f. loglik θ C.I. BRD1 (α, β) [0.878;0.906] BRD2 (α, β j ) [0.869;0.900] BRD3 (α k,β) [0.866;0.897] BRD4 (α, β k ) [0.674;0.856] BRD5 (α j,β) [0.806;0.882] BRD6 (α j,β j ) [0.788;0.849] BRD7 (α k,β k ) [0.697;0.832] BRD8 (α j,β k ) [0.657;0.826] BRD9 (α k,β j ) [0.851;0.884] ESALQ Course on Models for Longitudinal and Incomplete Data 247
253 An Interval of MNAR Estimates θ =0.885 Estimator [Pessimistic; optimistic] [0.694;0.904] Complete cases Available cases MAR (2 questions) MAR (3 questions) MNAR MNAR interval [0.741;0.892] θ ESALQ Course on Models for Longitudinal and Incomplete Data 248
254 20.7 A More Formal Look Statistical Imprecision Statistical Uncertainty Statistical Ignorance ESALQ Course on Models for Longitudinal and Incomplete Data 249
255 Statistical Imprecision: Due to finite sampling Fundamental concept of mathematical statistics Consistency, efficiency, precision, testing,... Disappears as sample size increases Statistical Ignorance: Due to incomplete observations Received less attention Can invalidate conclusions Does not disappear with increasing sample size Kenward, Goetghebeur, and Molenberghs (StatMod 2001) ESALQ Course on Models for Longitudinal and Incomplete Data 250
256 20.8 Slovenian Public Opinion: 3rd Analysis Model Structure d.f. loglik θ C.I. BRD1 (α, β) [0.878;0.906] BRD2 (α, β j ) [0.869;0.900] BRD3 (α k,β) [0.866;0.897] BRD4 (α, β k ) [0.674;0.856] BRD5 (α j,β) [0.806;0.882] BRD6 (α j,β j ) [0.788;0.849] BRD7 (α k,β k ) [0.697;0.832] BRD8 (α j,β k ) [0.657;0.826] BRD9 (α k,β j ) [0.851;0.884] Model 10 (α k,β jk ) [0.762;0.893] [0.744;0.907] Model 11 (α jk,β j ) [0.766;0.883] [0.715;0.920] Model 12 (α jk,β jk ) [0.694;0.904] ESALQ Course on Models for Longitudinal and Incomplete Data 251
257 20.9 Every MNAR Model Has Got a MAR Bodyguard Fit an MNAR model to a set of incomplete data Change the conditional distribution of the unobserved outcomes, given the observed ones, to comply with MAR The resulting new model will have exactly the same fit as the original MNAR model The missing data mechanism has changed This implies that definitively testing for MAR versus MNAR is not possible ESALQ Course on Models for Longitudinal and Incomplete Data 252
258 20.10 Slovenian Public Opinion: 4rd Analysis Model Structure d.f. loglik θ C.I. θ MAR BRD1 (α, β) [0.878;0.906] BRD2 (α, β j ) [0.869;0.900] BRD3 (α k,β) [0.866;0.897] BRD4 (α, β k ) [0.674;0.856] BRD5 (α j,β) [0.806;0.882] BRD6 (α j,β j ) [0.788;0.849] BRD7 (α k,β k ) [0.697;0.832] BRD8 (α j,β k ) [0.657;0.826] BRD9 (α k,β j ) [0.851;0.884] Model 10 (α k,β jk ) [0.762;0.893] [0.744;0.907] Model 11 (α jk,β j ) [0.766;0.883] [0.715;0.920] Model 12 (α jk,β jk ) [0.694;0.904] ESALQ Course on Models for Longitudinal and Incomplete Data 253
259 θ =0.885 Estimator θ [Pessimistic; optimistic] [0.694;0.904] MAR (3 questions) MNAR MNAR interval [0.753;0.891] Model 10 [0.762;0.893] Model 11 [0.766;0.883] Model 12 [0.694;0.904] ESALQ Course on Models for Longitudinal and Incomplete Data 254
260 Chapter 21 Concluding Remarks MCAR/simple CC biased LOCF inefficient MAR direct likelihood easy to conduct not simpler than MAR methods weighted GEE Gaussian & non-gaussian MNAR variety of methods strong, untestable assumptions most useful in sensitivity analysis ESALQ Course on Models for Longitudinal and Incomplete Data 255
261 Part VIII Case Study: Age-related Macular Degeneration ESALQ Course on Models for Longitudinal and Incomplete Data 256
262 Chapter 22 The Dataset Visual Acuity Age-related Macular Degeneration Missingness ESALQ Course on Models for Longitudinal and Incomplete Data 257
263 22.1 Visual Acuity I T I S A P L E A S U R E T O B E W I T H Y O U I N P I R A C I C A B A C A P I T A L O F B R A Z IL ESALQ Course on Models for Longitudinal and Incomplete Data 258
264 22.2 Age-related Macular Degeneration Trial Pharmacological Therapy for Macular Degeneration Study Group (1997) An occular pressure disease which makes patients progressively lose vision 240 patients enrolled in a multi-center trial (190 completers) Treatment: Interferon-α (6 million units) versus placebo Visits: baseline and follow-up at 4, 12, 24, and 52 weeks Continuous outcome: visual acuity: # letters correctly read on a vision chart Binary outcome: visual acuity versus baseline 0 or 0 ESALQ Course on Models for Longitudinal and Incomplete Data 259
265 Missingness: Measurement occasion 4 wks 12 wks 24 wks 52 wks Number % Completers O O O O Dropouts O O O M O O M M O M M M M M M M Non-monotone missingness O O M O O M M O M O O O M O M M ESALQ Course on Models for Longitudinal and Incomplete Data 260
266 Chapter 23 Weighted Generalized Estimating Equations Model for the weights Incorporating the weights within GEE ESALQ Course on Models for Longitudinal and Incomplete Data 261
267 23.1 Analysis of the ARMD Trial Model for the weights: logit[p (D i = j D i j)] = ψ 0 + ψ 1 y i,j 1 + ψ 2 T i + ψ 31 L 1i + ψ 32 L 2i + ψ 34 L 3i with +ψ 41 I(t j =2)+ψ 42 I(t j =3) y i,j 1 the binary outcome at the previous time t i,j 1 = t j 1 (since time is common to all subjects) T i =1for interferon-α and T i =0for placebo L ki =1if the patient s eye lesion is of level k =1,...,4 (since one dummy variable is redundant, only three are used) I( ) is an indicator function ESALQ Course on Models for Longitudinal and Incomplete Data 262
268 Results for the weights model: Effect Parameter Estimate (s.e.) Intercept ψ (0.49) Previous outcome ψ (0.38) Treatment ψ (0.37) Lesion level 1 ψ (0.49) Lesion level 2 ψ (0.52) Lesion level 3 ψ (0.72) Time 2 ψ (0.49) Time 3 ψ (0.44) ESALQ Course on Models for Longitudinal and Incomplete Data 263
269 GEE: logit[p (Y ij =1 T i,t j )] = β j1 + β j2 T i with T i =0for placebo and T i =1for interferon-α t j (j =1,...,4) refers to the four follow-up measurements Classical GEE and linearization-based GEE Comparison between CC, LOCF, and GEE analyses SAS code:... proc genmod data=armdhlp descending; class trt prev lesion time; model dropout = prev trt lesion time ods output obstats=pred; ods listing exclude obstats; run; / pred dist=b; ESALQ Course on Models for Longitudinal and Incomplete Data 264
270 ... proc genmod data=armdwgee; title data as is - WGEE ; weight wi; class time treat subject; model bindif = time treat*time / noint dist=binomial; repeated subject=subject / withinsubject=time type=exch modelse; run; proc glimmix data=armdwgee empirical; title data as is - WGEE - linearized version - empirical ; weight wi; nloptions maxiter=50 technique=newrap; class time treat subject; model bindif = time treat*time / noint solution dist=binary ; random _residual_ / subject=subject type=cs; run; ESALQ Course on Models for Longitudinal and Incomplete Data 265
271 Results: Effect Par. CC LOCF Observed data Unweighted WGEE Standard GEE Int.4 β (0.24;0.24) -0.87(0.20;0.21) -0.87(0.21;0.21) -0.98(0.10;0.44) Int.12 β (0.24;0.24) -0.97(0.21;0.21) -1.01(0.21;0.21) -1.78(0.15;0.38) Int.24 β (0.25;0.25) -1.05(0.21;0.21) -1.07(0.22;0.22) -1.11(0.15;0.33) Int.52 β (0.29;0.29) -1.51(0.24;0.24) -1.71(0.29;0.29) -1.72(0.25;0.39) Tr.4 β (0.32;0.32) 0.22(0.28;0.28) 0.22(0.28;0.28) 0.80(0.15;0.67) Tr.12 β (0.31;0.31) 0.55(0.28;0.28) 0.61(0.29;0.29) 1.87(0.19;0.61) Tr.24 β (0.33;0.33) 0.42(0.29;0.29) 0.44(0.30;0.30) 0.73(0.20;0.52) Tr.52 β (0.38;0.38) 0.34(0.32;0.32) 0.44(0.37;0.37) 0.74(0.31;0.52) Corr. ρ ESALQ Course on Models for Longitudinal and Incomplete Data 266
272 Effect Par. CC LOCF Observed data Unweighted WGEE Linearization-based GEE Int.4 β (0.24;0.24) -0.87(0.21;0.21) -0.87(0.21;0.21) -0.98(0.18;0.44) Int.12 β (0.24;0.24) -0.97(0.21;0.21) -1.01(0.22;0.21) -1.78(0.26;0.42) Int.24 β (0.25;0.25) -1.05(0.21;0.21) -1.07(0.23;0.22) -1.19(0.25;0.38) Int.52 β (0.29;0.29) -1.51(0.24;0.24) -1.71(0.29;0.29) -1.81(0.39;0.48) Tr.4 β (0.32;0.32) 0.22(0.28;0.28) 0.22(0.29;0.29) 0.80(0.26;0.67) Tr.12 β (0.31;0.31) 0.55(0.28;0.28) 0.61(0.28;0.29) 1.85(0.32;0.64) Tr.24 β (0.33;0.33) 0.42(0.29;0.29) 0.44(0.30;0.30) 0.98(0.33;0.60) Tr.52 β (0.38;0.38) 0.34(0.32;0.32) 0.44(0.37;0.37) 0.97(0.49;0.65) σ τ Corr. ρ ESALQ Course on Models for Longitudinal and Incomplete Data 267
273 Chapter 24 Multiple Imputation Settings and Models Results for GEE Results for GLMM ESALQ Course on Models for Longitudinal and Incomplete Data 268
274 24.1 MI Analysis of the ARMD Trial M =10imputations GEE: logit[p (Y ij =1 T i,t j )] = β j1 + β j2 T i GLMM: logit[p (Y ij =1 T i,t j,b i )] = β j1 + b i + β j2 T i, b i N(0,τ 2 ) T i =0for placebo and T i =1for interferon-α t j (j =1,...,4) refers to the four follow-up measurements Imputation based on the continuous outcome ESALQ Course on Models for Longitudinal and Incomplete Data 269
275 Results: Effect Par. GEE GLMM Int.4 β (0.20) -1.46(0.36) Int.12 β (0.22) -1.75(0.38) Int.24 β (0.23) -1.83(0.38) Int.52 β (0.27) -2.69(0.45) Trt.4 β (0.28) 0.32(0.48) Trt.12 β (0.29) 0.99(0.49) Trt.24 β (0.30) 0.67(0.51) Trt.52 β (0.35) 0.52(0.56) R.I. s.d. τ 2.20(0.26) R.I. var. τ (1.13) ESALQ Course on Models for Longitudinal and Incomplete Data 270
276 24.2 SAS Code for MI 1. Preparatory data analysis so that there is one line per subject 2. The imputation task: proc mi data=armd13 seed= out=armd13a simple nimpute=10 round=0.1; var lesion diff4 diff12 diff24 diff52; by treat; run; Note that the imputation task is conducted on the continuous outcome diff, indicating the difference in number of letters versus baseline 3. Then, data manipulation takes place to define the binary indicators and to create a longitudinal version of the dataset ESALQ Course on Models for Longitudinal and Incomplete Data 271
277 4. The analysis task (GEE): proc genmod data=armd13c; class time subject; by _imputation_; model bindif = time1 time2 time3 time4 trttime1 trttime2 trttime3 trttime4 / noint dist=binomial covb; repeated subject=subject / withinsubject=time type=exch modelse; ods output ParameterEstimates=gmparms parminfo=gmpinfo CovB=gmcovb; run; ESALQ Course on Models for Longitudinal and Incomplete Data 272
278 5. The analysis task (GLMM): proc nlmixed data=armd13c qpoints=20 maxiter=100 technique=newrap cov ecov; by _imputation_; eta = beta11*time1+beta12*time2+beta13*time3+beta14*time4+b +beta21*trttime1+beta22*trttime2+beta23*trttime3+beta24*trttime4; p = exp(eta)/(1+exp(eta)); model bindif ~ binary(p); random b ~ normal(0,tau*tau) subject=subject; estimate tau2 tau*tau; ods output ParameterEstimates=nlparms CovMatParmEst=nlcovb AdditionalEstimates=nlparmsa CovMatAddEst=nlcovba; run; ESALQ Course on Models for Longitudinal and Incomplete Data 273
279 6. The inference task (GEE): proc mianalyze parms=gmparms covb=gmcovb parminfo=gmpinfo wcov bcov tcov; modeleffects time1 time2 time3 time4 trttime1 trttime2 trttime3 trttime4; run; 7. The inference task (GLMM): proc mianalyze parms=nlparms covb=nlcovb wcov bcov tcov; modeleffects beta11 beta12 beta13 beta14 beta21 beta22 beta23 beta24; run; ESALQ Course on Models for Longitudinal and Incomplete Data 274
SAS Software to Fit the Generalized Linear Model
SAS Software to Fit the Generalized Linear Model Gordon Johnston, SAS Institute Inc., Cary, NC Abstract In recent years, the class of generalized linear models has gained popularity as a statistical modeling
Introduction to mixed model and missing data issues in longitudinal studies
Introduction to mixed model and missing data issues in longitudinal studies Hélène Jacqmin-Gadda INSERM, U897, Bordeaux, France Inserm workshop, St Raphael Outline of the talk I Introduction Mixed models
STATISTICA Formula Guide: Logistic Regression. Table of Contents
: Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary
Chapter 29 The GENMOD Procedure. Chapter Table of Contents
Chapter 29 The GENMOD Procedure Chapter Table of Contents OVERVIEW...1365 WhatisaGeneralizedLinearModel?...1366 ExamplesofGeneralizedLinearModels...1367 TheGENMODProcedure...1368 GETTING STARTED...1370
How to use SAS for Logistic Regression with Correlated Data
How to use SAS for Logistic Regression with Correlated Data Oliver Kuss Institute of Medical Epidemiology, Biostatistics, and Informatics Medical Faculty, University of Halle-Wittenberg, Halle/Saale, Germany
Analysis of Correlated Data. Patrick J. Heagerty PhD Department of Biostatistics University of Washington
Analysis of Correlated Data Patrick J Heagerty PhD Department of Biostatistics University of Washington Heagerty, 6 Course Outline Examples of longitudinal data Correlation and weighting Exploratory data
Overview. Longitudinal Data Variation and Correlation Different Approaches. Linear Mixed Models Generalized Linear Mixed Models
Overview 1 Introduction Longitudinal Data Variation and Correlation Different Approaches 2 Mixed Models Linear Mixed Models Generalized Linear Mixed Models 3 Marginal Models Linear Models Generalized Linear
SPPH 501 Analysis of Longitudinal & Correlated Data September, 2012
SPPH 501 Analysis of Longitudinal & Correlated Data September, 2012 TIME & PLACE: Term 1, Tuesday, 1:30-4:30 P.M. LOCATION: INSTRUCTOR: OFFICE: SPPH, Room B104 Dr. Ying MacNab SPPH, Room 134B TELEPHONE:
Linear Mixed-Effects Modeling in SPSS: An Introduction to the MIXED Procedure
Technical report Linear Mixed-Effects Modeling in SPSS: An Introduction to the MIXED Procedure Table of contents Introduction................................................................ 1 Data preparation
SPSS TRAINING SESSION 3 ADVANCED TOPICS (PASW STATISTICS 17.0) Sun Li Centre for Academic Computing [email protected]
SPSS TRAINING SESSION 3 ADVANCED TOPICS (PASW STATISTICS 17.0) Sun Li Centre for Academic Computing [email protected] IN SPSS SESSION 2, WE HAVE LEARNT: Elementary Data Analysis Group Comparison & One-way
Longitudinal Data Analysis
Longitudinal Data Analysis Acknowledge: Professor Garrett Fitzmaurice INSTRUCTOR: Rino Bellocco Department of Statistics & Quantitative Methods University of Milano-Bicocca Department of Medical Epidemiology
Review of the Methods for Handling Missing Data in. Longitudinal Data Analysis
Int. Journal of Math. Analysis, Vol. 5, 2011, no. 1, 1-13 Review of the Methods for Handling Missing Data in Longitudinal Data Analysis Michikazu Nakai and Weiming Ke Department of Mathematics and Statistics
GLM I An Introduction to Generalized Linear Models
GLM I An Introduction to Generalized Linear Models CAS Ratemaking and Product Management Seminar March 2009 Presented by: Tanya D. Havlicek, Actuarial Assistant 0 ANTITRUST Notice The Casualty Actuarial
Models for discrete longitudinal data Analysing discrete correlated data using R. Christophe Lalanne
Models for discrete longitudinal data Analysing discrete correlated data using R Christophe Lalanne Novembre 2007 Preface This textbook is intended to the R user already accustomed to the Generalized Linear
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
Lecture 6: Poisson regression
Lecture 6: Poisson regression Claudia Czado TU München c (Claudia Czado, TU Munich) ZFS/IMS Göttingen 2004 0 Overview Introduction EDA for Poisson regression Estimation and testing in Poisson regression
Handling attrition and non-response in longitudinal data
Longitudinal and Life Course Studies 2009 Volume 1 Issue 1 Pp 63-72 Handling attrition and non-response in longitudinal data Harvey Goldstein University of Bristol Correspondence. Professor H. Goldstein
Overview of Methods for Analyzing Cluster-Correlated Data. Garrett M. Fitzmaurice
Overview of Methods for Analyzing Cluster-Correlated Data Garrett M. Fitzmaurice Laboratory for Psychiatric Biostatistics, McLean Hospital Department of Biostatistics, Harvard School of Public Health Outline
11. Analysis of Case-control Studies Logistic Regression
Research methods II 113 11. Analysis of Case-control Studies Logistic Regression This chapter builds upon and further develops the concepts and strategies described in Ch.6 of Mother and Child Health:
Lecture 18: Logistic Regression Continued
Lecture 18: Logistic Regression Continued Dipankar Bandyopadhyay, Ph.D. BMTRY 711: Analysis of Categorical Data Spring 2011 Division of Biostatistics and Epidemiology Medical University of South Carolina
Lecture 14: GLM Estimation and Logistic Regression
Lecture 14: GLM Estimation and Logistic Regression Dipankar Bandyopadhyay, Ph.D. BMTRY 711: Analysis of Categorical Data Spring 2011 Division of Biostatistics and Epidemiology Medical University of South
Random Effects Models for Longitudinal Data
Random Effects Models for Longitudinal Data Geert Verbeke 1,2 Geert Molenberghs 2,1 Dimitris Rizopoulos 3 1 I-BioStat, Katholieke Universiteit Leuven, Kapucijnenvoer 35, B-3000 Leuven, Belgium 2 I-BioStat,
Multivariate Normal Distribution
Multivariate Normal Distribution Lecture 4 July 21, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #4-7/21/2011 Slide 1 of 41 Last Time Matrices and vectors Eigenvalues
Individual Growth Analysis Using PROC MIXED Maribeth Johnson, Medical College of Georgia, Augusta, GA
Paper P-702 Individual Growth Analysis Using PROC MIXED Maribeth Johnson, Medical College of Georgia, Augusta, GA ABSTRACT Individual growth models are designed for exploring longitudinal data on individuals
Department of Epidemiology and Public Health Miller School of Medicine University of Miami
Department of Epidemiology and Public Health Miller School of Medicine University of Miami BST 630 (3 Credit Hours) Longitudinal and Multilevel Data Wednesday-Friday 9:00 10:15PM Course Location: CRB 995
MATH5885 LONGITUDINAL DATA ANALYSIS
MATH5885 LONGITUDINAL DATA ANALYSIS Semester 1, 2013 CRICOS Provider No: 00098G 2013, School of Mathematics and Statistics, UNSW MATH5885 Course Outline Information about the course Course Authority: William
Lecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
VI. Introduction to Logistic Regression
VI. Introduction to Logistic Regression We turn our attention now to the topic of modeling a categorical outcome as a function of (possibly) several factors. The framework of generalized linear models
I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN
Beckman HLM Reading Group: Questions, Answers and Examples Carolyn J. Anderson Department of Educational Psychology I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN Linear Algebra Slide 1 of
The Probit Link Function in Generalized Linear Models for Data Mining Applications
Journal of Modern Applied Statistical Methods Copyright 2013 JMASM, Inc. May 2013, Vol. 12, No. 1, 164-169 1538 9472/13/$95.00 The Probit Link Function in Generalized Linear Models for Data Mining Applications
Prediction for Multilevel Models
Prediction for Multilevel Models Sophia Rabe-Hesketh Graduate School of Education & Graduate Group in Biostatistics University of California, Berkeley Institute of Education, University of London Joint
Technical report. in SPSS AN INTRODUCTION TO THE MIXED PROCEDURE
Linear mixedeffects modeling in SPSS AN INTRODUCTION TO THE MIXED PROCEDURE Table of contents Introduction................................................................3 Data preparation for MIXED...................................................3
Random Effects Models for Longitudinal Data
Chapter 2 Random Effects Models for Longitudinal Data Geert Verbeke, Geert Molenberghs, and Dimitris Rizopoulos Abstract Mixed models have become very popular for the analysis of longitudinal data, partly
i=1 In practice, the natural logarithm of the likelihood function, called the log-likelihood function and denoted by
Statistics 580 Maximum Likelihood Estimation Introduction Let y (y 1, y 2,..., y n be a vector of iid, random variables from one of a family of distributions on R n and indexed by a p-dimensional parameter
Lecture 8: Gamma regression
Lecture 8: Gamma regression Claudia Czado TU München c (Claudia Czado, TU Munich) ZFS/IMS Göttingen 2004 0 Overview Models with constant coefficient of variation Gamma regression: estimation and testing
Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
Linear Classification. Volker Tresp Summer 2015
Linear Classification Volker Tresp Summer 2015 1 Classification Classification is the central task of pattern recognition Sensors supply information about an object: to which class do the object belong
Least Squares Estimation
Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David
APPLIED MISSING DATA ANALYSIS
APPLIED MISSING DATA ANALYSIS Craig K. Enders Series Editor's Note by Todd D. little THE GUILFORD PRESS New York London Contents 1 An Introduction to Missing Data 1 1.1 Introduction 1 1.2 Chapter Overview
A course on Longitudinal data analysis : what did we learn? COMPASS 11/03/2011
A course on Longitudinal data analysis : what did we learn? COMPASS 11/03/2011 1 Abstract We report back on methods used for the analysis of longitudinal data covered by the course We draw lessons for
Model Fitting in PROC GENMOD Jean G. Orelien, Analytical Sciences, Inc.
Paper 264-26 Model Fitting in PROC GENMOD Jean G. Orelien, Analytical Sciences, Inc. Abstract: There are several procedures in the SAS System for statistical modeling. Most statisticians who use the SAS
Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus
Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus Tihomir Asparouhov and Bengt Muthén Mplus Web Notes: No. 15 Version 8, August 5, 2014 1 Abstract This paper discusses alternatives
An extension of the factoring likelihood approach for non-monotone missing data
An extension of the factoring likelihood approach for non-monotone missing data Jae Kwang Kim Dong Wan Shin January 14, 2010 ABSTRACT We address the problem of parameter estimation in multivariate distributions
Assignments Analysis of Longitudinal data: a multilevel approach
Assignments Analysis of Longitudinal data: a multilevel approach Frans E.S. Tan Department of Methodology and Statistics University of Maastricht The Netherlands Maastricht, Jan 2007 Correspondence: Frans
SAS Syntax and Output for Data Manipulation:
Psyc 944 Example 5 page 1 Practice with Fixed and Random Effects of Time in Modeling Within-Person Change The models for this example come from Hoffman (in preparation) chapter 5. We will be examining
Introduction to General and Generalized Linear Models
Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby
Generalized Linear Models
Generalized Linear Models We have previously worked with regression models where the response variable is quantitative and normally distributed. Now we turn our attention to two types of models where the
Factorial experimental designs and generalized linear models
Statistics & Operations Research Transactions SORT 29 (2) July-December 2005, 249-268 ISSN: 1696-2281 www.idescat.net/sort Statistics & Operations Research c Institut d Estadística de Transactions Catalunya
10 Dichotomous or binary responses
10 Dichotomous or binary responses 10.1 Introduction Dichotomous or binary responses are widespread. Examples include being dead or alive, agreeing or disagreeing with a statement, and succeeding or failing
Elements of statistics (MATH0487-1)
Elements of statistics (MATH0487-1) Prof. Dr. Dr. K. Van Steen University of Liège, Belgium December 10, 2012 Introduction to Statistics Basic Probability Revisited Sampling Exploratory Data Analysis -
Using Repeated Measures Techniques To Analyze Cluster-correlated Survey Responses
Using Repeated Measures Techniques To Analyze Cluster-correlated Survey Responses G. Gordon Brown, Celia R. Eicheldinger, and James R. Chromy RTI International, Research Triangle Park, NC 27709 Abstract
Poisson Models for Count Data
Chapter 4 Poisson Models for Count Data In this chapter we study log-linear models for count data under the assumption of a Poisson error structure. These models have many applications, not only to the
Multivariate Logistic Regression
1 Multivariate Logistic Regression As in univariate logistic regression, let π(x) represent the probability of an event that depends on p covariates or independent variables. Then, using an inv.logit formulation
Missing Data: Part 1 What to Do? Carol B. Thompson Johns Hopkins Biostatistics Center SON Brown Bag 3/20/13
Missing Data: Part 1 What to Do? Carol B. Thompson Johns Hopkins Biostatistics Center SON Brown Bag 3/20/13 Overview Missingness and impact on statistical analysis Missing data assumptions/mechanisms Conventional
Power and sample size in multilevel modeling
Snijders, Tom A.B. Power and Sample Size in Multilevel Linear Models. In: B.S. Everitt and D.C. Howell (eds.), Encyclopedia of Statistics in Behavioral Science. Volume 3, 1570 1573. Chicester (etc.): Wiley,
Reject Inference in Credit Scoring. Jie-Men Mok
Reject Inference in Credit Scoring Jie-Men Mok BMI paper January 2009 ii Preface In the Master programme of Business Mathematics and Informatics (BMI), it is required to perform research on a business
Logistic Regression (1/24/13)
STA63/CBB540: Statistical methods in computational biology Logistic Regression (/24/3) Lecturer: Barbara Engelhardt Scribe: Dinesh Manandhar Introduction Logistic regression is model for regression used
PATTERN MIXTURE MODELS FOR MISSING DATA. Mike Kenward. London School of Hygiene and Tropical Medicine. Talk at the University of Turku,
PATTERN MIXTURE MODELS FOR MISSING DATA Mike Kenward London School of Hygiene and Tropical Medicine Talk at the University of Turku, April 10th 2012 1 / 90 CONTENTS 1 Examples 2 Modelling Incomplete Data
Missing data and net survival analysis Bernard Rachet
Workshop on Flexible Models for Longitudinal and Survival Data with Applications in Biostatistics Warwick, 27-29 July 2015 Missing data and net survival analysis Bernard Rachet General context Population-based,
Logistic Regression (a type of Generalized Linear Model)
Logistic Regression (a type of Generalized Linear Model) 1/36 Today Review of GLMs Logistic Regression 2/36 How do we find patterns in data? We begin with a model of how the world works We use our knowledge
Using PROC MIXED in Hierarchical Linear Models: Examples from two- and three-level school-effect analysis, and meta-analysis research
Using PROC MIXED in Hierarchical Linear Models: Examples from two- and three-level school-effect analysis, and meta-analysis research Sawako Suzuki, DePaul University, Chicago Ching-Fan Sheu, DePaul University,
Principle Component Analysis and Partial Least Squares: Two Dimension Reduction Techniques for Regression
Principle Component Analysis and Partial Least Squares: Two Dimension Reduction Techniques for Regression Saikat Maitra and Jun Yan Abstract: Dimension reduction is one of the major tasks for multivariate
Multivariate Statistical Inference and Applications
Multivariate Statistical Inference and Applications ALVIN C. RENCHER Department of Statistics Brigham Young University A Wiley-Interscience Publication JOHN WILEY & SONS, INC. New York Chichester Weinheim
Institute of Actuaries of India Subject CT3 Probability and Mathematical Statistics
Institute of Actuaries of India Subject CT3 Probability and Mathematical Statistics For 2015 Examinations Aim The aim of the Probability and Mathematical Statistics subject is to provide a grounding in
Life Table Analysis using Weighted Survey Data
Life Table Analysis using Weighted Survey Data James G. Booth and Thomas A. Hirschl June 2005 Abstract Formulas for constructing valid pointwise confidence bands for survival distributions, estimated using
Introduction to Longitudinal Data Analysis
Introduction to Longitudinal Data Analysis Longitudinal Data Analysis Workshop Section 1 University of Georgia: Institute for Interdisciplinary Research in Education and Human Development Section 1: Introduction
Sun Li Centre for Academic Computing [email protected]
Sun Li Centre for Academic Computing [email protected] Elementary Data Analysis Group Comparison & One-way ANOVA Non-parametric Tests Correlations General Linear Regression Logistic Models Binary Logistic
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written
Models for Longitudinal and Clustered Data
Models for Longitudinal and Clustered Data Germán Rodríguez December 9, 2008, revised December 6, 2012 1 Introduction The most important assumption we have made in this course is that the observations
An Introduction to Modeling Longitudinal Data
An Introduction to Modeling Longitudinal Data Session I: Basic Concepts and Looking at Data Robert Weiss Department of Biostatistics UCLA School of Public Health [email protected] August 2010 Robert Weiss
Applied Statistics. J. Blanchet and J. Wadsworth. Institute of Mathematics, Analysis, and Applications EPF Lausanne
Applied Statistics J. Blanchet and J. Wadsworth Institute of Mathematics, Analysis, and Applications EPF Lausanne An MSc Course for Applied Mathematicians, Fall 2012 Outline 1 Model Comparison 2 Model
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
A Composite Likelihood Approach to Analysis of Survey Data with Sampling Weights Incorporated under Two-Level Models
A Composite Likelihood Approach to Analysis of Survey Data with Sampling Weights Incorporated under Two-Level Models Grace Y. Yi 13, JNK Rao 2 and Haocheng Li 1 1. University of Waterloo, Waterloo, Canada
MISSING DATA TECHNIQUES WITH SAS. IDRE Statistical Consulting Group
MISSING DATA TECHNIQUES WITH SAS IDRE Statistical Consulting Group ROAD MAP FOR TODAY To discuss: 1. Commonly used techniques for handling missing data, focusing on multiple imputation 2. Issues that could
Logit Models for Binary Data
Chapter 3 Logit Models for Binary Data We now turn our attention to regression models for dichotomous data, including logistic regression and probit analysis. These models are appropriate when the response
BIO 226: APPLIED LONGITUDINAL ANALYSIS COURSE SYLLABUS. Spring 2015
BIO 226: APPLIED LONGITUDINAL ANALYSIS COURSE SYLLABUS Spring 2015 Instructor: Teaching Assistants: Dr. Brent Coull HSPH Building II, Room 413 Phone: (617) 432-2376 E-mail: [email protected] Office
Adequacy of Biomath. Models. Empirical Modeling Tools. Bayesian Modeling. Model Uncertainty / Selection
Directions in Statistical Methodology for Multivariable Predictive Modeling Frank E Harrell Jr University of Virginia Seattle WA 19May98 Overview of Modeling Process Model selection Regression shape Diagnostics
An Application of the G-formula to Asbestos and Lung Cancer. Stephen R. Cole. Epidemiology, UNC Chapel Hill. Slides: www.unc.
An Application of the G-formula to Asbestos and Lung Cancer Stephen R. Cole Epidemiology, UNC Chapel Hill Slides: www.unc.edu/~colesr/ 1 Acknowledgements Collaboration with David B. Richardson, Haitao
Introducing the Multilevel Model for Change
Department of Psychology and Human Development Vanderbilt University GCM, 2010 1 Multilevel Modeling - A Brief Introduction 2 3 4 5 Introduction In this lecture, we introduce the multilevel model for change.
business statistics using Excel OXFORD UNIVERSITY PRESS Glyn Davis & Branko Pecar
business statistics using Excel Glyn Davis & Branko Pecar OXFORD UNIVERSITY PRESS Detailed contents Introduction to Microsoft Excel 2003 Overview Learning Objectives 1.1 Introduction to Microsoft Excel
Statistical Rules of Thumb
Statistical Rules of Thumb Second Edition Gerald van Belle University of Washington Department of Biostatistics and Department of Environmental and Occupational Health Sciences Seattle, WA WILEY AJOHN
Outline. Topic 4 - Analysis of Variance Approach to Regression. Partitioning Sums of Squares. Total Sum of Squares. Partitioning sums of squares
Topic 4 - Analysis of Variance Approach to Regression Outline Partitioning sums of squares Degrees of freedom Expected mean squares General linear test - Fall 2013 R 2 and the coefficient of correlation
LOGISTIC REGRESSION. Nitin R Patel. where the dependent variable, y, is binary (for convenience we often code these values as
LOGISTIC REGRESSION Nitin R Patel Logistic regression extends the ideas of multiple linear regression to the situation where the dependent variable, y, is binary (for convenience we often code these values
Ordinal Regression. Chapter
Ordinal Regression Chapter 4 Many variables of interest are ordinal. That is, you can rank the values, but the real distance between categories is unknown. Diseases are graded on scales from least severe
Chapter 7: Simple linear regression Learning Objectives
Chapter 7: Simple linear regression Learning Objectives Reading: Section 7.1 of OpenIntro Statistics Video: Correlation vs. causation, YouTube (2:19) Video: Intro to Linear Regression, YouTube (5:18) -
BayesX - Software for Bayesian Inference in Structured Additive Regression
BayesX - Software for Bayesian Inference in Structured Additive Regression Thomas Kneib Faculty of Mathematics and Economics, University of Ulm Department of Statistics, Ludwig-Maximilians-University Munich
CHAPTER 9 EXAMPLES: MULTILEVEL MODELING WITH COMPLEX SURVEY DATA
Examples: Multilevel Modeling With Complex Survey Data CHAPTER 9 EXAMPLES: MULTILEVEL MODELING WITH COMPLEX SURVEY DATA Complex survey data refers to data obtained by stratification, cluster sampling and/or
Service courses for graduate students in degree programs other than the MS or PhD programs in Biostatistics.
Course Catalog In order to be assured that all prerequisites are met, students must acquire a permission number from the education coordinator prior to enrolling in any Biostatistics course. Courses are
Chapter 1. Longitudinal Data Analysis. 1.1 Introduction
Chapter 1 Longitudinal Data Analysis 1.1 Introduction One of the most common medical research designs is a pre-post study in which a single baseline health status measurement is obtained, an intervention
2. Making example missing-value datasets: MCAR, MAR, and MNAR
Lecture 20 1. Types of missing values 2. Making example missing-value datasets: MCAR, MAR, and MNAR 3. Common methods for missing data 4. Compare results on example MCAR, MAR, MNAR data 1 Missing Data
EDMS 769L: Statistical Analysis of Longitudinal Data 1809 PAC, Th 4:15-7:00pm 2009 Spring Semester
Instructor Dr. Jeffrey Harring 1230E Benjamin Building Phone: (301) 405-3630 Email: [email protected] Office Hours Tuesday 2:00-3:00pm, or by appointment Course Objectives, Description and Prerequisites
15 Ordinal longitudinal data analysis
15 Ordinal longitudinal data analysis Jeroen K. Vermunt and Jacques A. Hagenaars Tilburg University Introduction Growth data and longitudinal data in general are often of an ordinal nature. For example,
What s New in Econometrics? Lecture 8 Cluster and Stratified Sampling
What s New in Econometrics? Lecture 8 Cluster and Stratified Sampling Jeff Wooldridge NBER Summer Institute, 2007 1. The Linear Model with Cluster Effects 2. Estimation with a Small Number of Groups and
Overview Classes. 12-3 Logistic regression (5) 19-3 Building and applying logistic regression (6) 26-3 Generalizations of logistic regression (7)
Overview Classes 12-3 Logistic regression (5) 19-3 Building and applying logistic regression (6) 26-3 Generalizations of logistic regression (7) 2-4 Loglinear models (8) 5-4 15-17 hrs; 5B02 Building and
