Bayesian Approaches to Handling Missing Data
|
|
|
- Rafe Tyler
- 9 years ago
- Views:
Transcription
1 Bayesian Approaches to Handling Missing Data Nicky Best and Alexina Mason BIAS Short Course, Jan 30, 2012
2 Lecture 1. Introduction to Missing Data Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 1 / 42
3 Lecture Outline Introduction Motivating example Types of missing data Ad hoc methods Statistically principled methods Bayesian concepts and MCMC Advantages of Bayesian models for missing data Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 2 / 42
4 Introduction Missing data are common! Usually inadequately handled in both observational and experimental research For example, Wood et al. (2004) reviewed 71 recently published BMJ, JAMA, Lancet and NEJM papers 89% had partly missing outcome data In 37 trials with repeated outcome measures, 46% performed complete case analysis Only 21% reported sensitivity analysis Sterne et al. (2009) reviewed articles using Multiple Imputation in BMJ, JAMA, Lancet and NEJM from 2002 to articles found, with use doubling over 6 year period However, the reporting was almost always inadequate Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 3 / 42
5 How does missing data arise? known sampling process Base population Study sample target sample Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 4 / 42
6 How does missing data arise? known sampling process Base population Study sample target sample missing non-responders responders observed sample Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 4 / 42
7 How does missing data arise? known sampling process Base population Study sample target sample missing non-responders responders observed sample full partial missing Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 4 / 42
8 How does missing data arise? Base population known sampling process WHY? Study sample target sample missing non-responders responders WHY? observed sample full partial missing Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 4 / 42
9 Motivating example: antidepressant clinical trial 6 centre clinical trial, comparing 3 treatments of depression 367 subjects randomised to one of 3 treatments Subjects rated on Hamilton depression score (HAMD) on 5 weekly visits week 0 before treatment weeks 1-4 during treatment HAMD score takes values 0-50 the higher the score, the more severe the depression Subjects drop-out from week 2 onwards (246 complete cases) Data were previously analysed by Diggle and Kenward (1994) Study objective: are there any differences in the effects of the 3 treatments on the change in HAMD score over time? Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 5 / 42
10 HAMD example: complete cases 50 Individual Profiles Mean Response Profiles HAMD score treatment 1 treatment 2 treatment 3 HAMD score treatment 1 treatment 2 treatment week week Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 6 / 42
11 HAMD example: analysis model Use the variables y, Hamilton depression (HAMD) score t, treatment w, week and for simplicity ignore any centre effects assume linear relationships A suitable analysis model would be a hierarchical model with random intercepts and slopes (we will look at the details of this later) If we had no missing data, we could just fit this model However... Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 7 / 42
12 HAMD example: exploring the missingness Subjects drop-out of the study from week 2 onwards HAMD score is missing for some individuals for 1, 2 or 3 weeks Before analysing the data, we should consider how many subjects dropped out of the study each week? is the level and pattern of drop-out consistent across treatments? Percentage of missingness by treatment and week treat. 1 treat. 2 treat. 3 all treatments week week week Fewer subjects drop-out of treatment 3 Drop-out occurs earlier for treatment 2 Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 8 / 42
13 HAMD example: implications of drop-out It is also very important to think about possible reasons for the missingness Why do you think some subjects dropped out of the HAMD study? Do you think that subjects who dropped out have similar characteristics to individuals who remained in the study? What do you think are the potential problems with only analysing complete cases? Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 9 / 42
14 HAMD example: missing scores treatment 1 treatment 2 treatment 3 HAMD score complete cases dropout at wk 4 dropout at wk 3 dropout at wk 2 HAMD score complete cases dropout at wk 4 dropout at wk 3 dropout at wk 2 HAMD score complete cases dropout at wk 4 dropout at wk 3 dropout at wk week week week Do individuals have similar profiles whether they dropped out or remained in the study? Do all the treatments show the same patterns? Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 10 / 42
15 HAMD example: model of missingness The full data for the HAMD example includes: y, HAMD score (observed and missing values) t, treatment m, a missing data indicator for y, s.t. { 0: yiw observed m iw = 1: y iw missing So we can model y (random effects model) and the probability that y is missing using: m iw Bernoulli(p iw ) We now consider different possibilities for p iw, which depend on the assumptions we make about the missing data mechanism Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 11 / 42
16 Types of missing data Following Rubin, missing data are generally classified into 3 types Consider the mechanism that led to the missing HAMD scores (y) recall: p iw is the probability y iw is missing for individual i in week w Missing Completely At Random (MCAR) missingness does not depend on observed or unobserved data e.g. logit(piw ) = θ 0 Missing At Random (MAR) missingness depends only on observed data e.g. logit(piw ) = θ 0 + θ 1 t i or logit(p iw ) = θ 0 + θ 2 y i0 note: drop-out from week 2; weeks 0 and 1 completely observed Missing Not At Random (MNAR) neither MCAR or MAR hold e.g. logit(piw ) = θ 0 + θ 3 y iw Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 12 / 42
17 Graphical models to represent different missing data mechanisms Graphical models can be a helpful way to visualise different missing data mechanisms and understand their implications for analysis More generally, graphical models are a useful tool for building complex Bayesian models Ingredients of a Bayesian graphical model: Nodes: all random quantities (data and parameters) Edges: associations between the nodes Used to represent a set of conditional independence statements about system of interest See Spiegelhalter, Thomas and Best (1996); Spiegelhalter (1998); Richardson and Best (2003) Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 13 / 42
18 Graphical models: a simple example Ζ X Y Z = genotypes of parents; X, Y = genotypes of 2 children If we know the genotypes of the parents, then the children s genotypes are conditionally independent, i.e. X Y Z or Pr(X, Y Z ) = Pr(X Z )Pr(Y Z ) Factorization thm: Jt distribution, Pr(V ) = Pr(v parents [v]) Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 14 / 42
19 Graphical models: a simple example Ζ X Y If parents genotypes (Z ) are unknown, then the children s genotypes (X, Y ) are marginally dependent Knowing X tells you something about the likely values of Y Same principles apply if X, Y, Z represent random variables (data and parameters) in a statistical model Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 15 / 42
20 Bayesian graphical models: notation A typical regression model of interest β y i Normal(µ i, σ 2 ), i = 1,..., N µ i = x T β β fully specified prior σ 2 µ i y i x i individual i Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 16 / 42
21 Bayesian graphical models: notation yellow circles = random variables (data and parameters) blue squares = fixed constants (e.g. covariates, denominators) black arrows = stochastic dependence red arrows = logical dependence large rectangles = repeated structures (loops) σ 2 β µ i y i individual i x i Directed Acyclic Graph (DAG) contains only directed links (arrows) and no cycles Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 17 / 42
22 Bayesian graphical models: notation β yellow circles = random variables (data and parameters) blue squares = fixed constants (e.g. covariates, denominators) black arrows = stochastic dependence red arrows = logical dependence large rectangles = repeated structures (loops) σ 2 µ i y i individual i x i x is completely observed but y has missing values We usually make no distinction in the graph between random variables representing data or parameters However, for clarity, we will denote a random variable representing a data node with missing values by an orange circle Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 18 / 42
23 Using DAGs to represent missing data mechanisms A typical regression model of interest β µ i x i σ 2 y i Model of Interest individual i Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 19 / 42
24 Using DAGs to represent missing data mechanisms Now suppose x is completely observed, but y has missing values β µ i x i σ 2 y i Model of Interest individual i Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 20 / 42
25 DAG: Missing Completely At Random (MCAR) A regression model of interest + model for probability of y missing β θ µ i x i σ 2 y i p i Model of Interest individual i m i Model of Missingness Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 21 / 42
26 DAG: Missing At Random (MAR) A regression model of interest + model for probability of y missing β θ µ i x i σ 2 y i p i Model of Interest individual i m i Model of Missingness Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 22 / 42
27 DAG: Missing Not At Random (MNAR) A regression model of interest + model for probability of y missing β θ µ i x i σ 2 y i p i Model of Interest individual i m i Model of Missingness Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 23 / 42
28 Joint model: general notation We now use general notation, but in the HAMD example z = (y, t) Let z = (z ij ) denote a rectangular data set i = 1,..., n individuals and j = 1,..., k variables Partition z into observed and missing values, z = (z obs, z mis ) Let m = (m ij ) be a binary indicator variable such that { 0: zij observed m ij = 1: z ij missing Let β and θ denote vectors of unknown parameters Then the joint model (likelihood) of the full data is f (z, m β, θ) = f (z obs, z mis, m β, θ) Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 24 / 42
29 Joint model: integrating out the missingness The joint model, f (z obs, z mis, m β, θ), cannot be evaluated in the usual way because it depends on missing data However, the marginal distribution of the observed data can be obtained by integrating out the missing data, f (z obs, m β, θ) = f (z obs, z mis, m β, θ)dz mis We now look at factorising the joint model Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 25 / 42
30 Joint model: factorisations Two factorisations of the joint model are commonly used 1 selection models f (z obs, z mis, m β, θ) = f (m z obs, z mis, β, θ)f (z obs, z mis β, θ) 2 pattern mixture models f (z obs, z mis, m β, θ) = f (z obs, z mis m, β, θ)f (m β, θ) An advantage of the selection model factorisation is that it includes the model of interest term directly On the other hand, the pattern mixture model corresponds more directly to what is actually observed (i.e. the distribution of the data within subgroups having different missing data patterns) We will focus on the selection model factorisation in this workshop Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 26 / 42
31 Joint model: selection model factorisation The selection model f (z obs, z mis, m β, θ) = f (m z obs, z mis, β, θ)f (z obs, z mis β, θ) simplifies with appropriate conditional independence assumptions f (z obs, z mis, m β, θ) = f (m z obs, z mis, θ)f (z obs, z mis β) f (z obs, z mis β) is the usual likelihood you would specify if all the data had been observed f (m z obs, z mis, θ) represents the missing data mechanism and describes the way in which the probability of an observation being missing depends on other variables (measured or not) and on its own value For some types of missing data, the form of the conditional distribution of m can be simplified Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 27 / 42
32 Simplifying the factorisation for MAR and MCAR Recall we wish to integrate out the missingness f (z obs, m β, θ) = f (z obs, z mis, m β, θ)dz mis = f (m z obs, z mis, θ)f (z obs, z mis β)dz mis MAR missingness depends only on observed data, i.e. f (m z obs, z mis, θ) = f (m z obs, θ) So, f (z obs, m β, θ) = f (m z obs, θ) = f (m z obs, θ)f (z obs β) f (z obs, z mis β)dz mis MCAR missingness is a special case of MAR that does not even depend on the observed data f (m z obs, z mis, θ) = f (m θ) So, f (z obs, m β, θ) = f (m θ)f (z obs β) Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 28 / 42
33 Ignorable/Nonignorable missingness The missing data mechanism is termed ignorable if 1 the missing data are MCAR or MAR and 2 the parameters β and θ are distinct In the Bayesian setup, an additional condition is 3 the priors on β and θ are independent Ignorable means we can ignore the model of missingness, but does not necessarily mean we can ignore the missing data! However if the missing data mechanism is nonignorable, then we cannot ignore the model of missingness Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 29 / 42
34 Assumptions In contrast with the sampling process, which is often known, the missingness mechanism is usually unknown The data alone cannot usually definitively tell us the sampling process But with fully observed data, we can usually check the plausibility of any assumptions about the sampling process e.g. using residuals and other diagnostics Likewise, the missingness pattern, and its relationship to the observations, cannot definitively identify the missingness mechanism Unfortunately, the assumptions we make about the missingness mechanism cannot be definitively checked from the data at hand Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 30 / 42
35 Sensitivity analysis The issues surrounding the analysis of data sets with missing values therefore centre on assumptions We have to decide which assumptions are reasonable and sensible in any given setting - contextual/subject matter information will be central to this ensure that the assumptions are transparent explore the sensitivity of inferences/conclusions to the assumptions Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 31 / 42
36 Ad-hoc methods: complete case analysis (CC) Missing data are ignored and only complete cases analysed Many computer packages do this by default Advantage of CC simple Disadvantages of CC often introduces bias inefficient as data from incomplete cases discarded Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 32 / 42
37 Ad-hoc methods: single imputation Aim to fill in missing values to create single completed (imputed) dataset, which is then analysed using standard methods Types of single imputation include mean imputation last observation carried forward (LOCF) Computationally simple but makes strong assumptions about missingness mechanism (usually MCAR) Ignores uncertainty about imputed missing values Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 33 / 42
38 Last observation carried forward (LOCF) Widely used in clinical trial settings Assumption: all unseen measurements = last seen measurement Is not even valid under the strong assumption of MCAR In the HAMD example HAMD score treatment week if the treatment is working but individuals drop-out early, we will underestimate the treatment effect LOCF changes the shape of the profile, as drop-outs likely to have improved at least a little this effect will vary by treatment, and hence inference about treatment difference will be misleading Ad hoc methods are frequently used, but not recommended Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 34 / 42
39 Statistically principled methods Generate statistical (stochastic) information about the missing values and/or the missingness mechanism, e.g. Multiple imputation (MI) generate K > 1 imputed values for missing observations from appropriate probability distribution Fully model based (e.g. Bayesian) write down statistical model for full data (including missingness mechanism) and base analysis on this model Make weaker assumptions, but more computationally complex to implement In contrast to ad-hoc methods, principled methods are: based on a well-defined statistical model for the complete data, and explicit assumptions about the missing value mechanism the subsequent analysis, inferences and conclusions are valid under these assumptions doesn t mean the assumptions are necessarily true but it does allow the dependence of the conclusions on these assumptions to be investigated Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 35 / 42
40 Bayesian inference Makes fundamental distinction between Observable quantities, D, (i.e. the data) Unknown quantities, Ω, (i.e. statistical parameters) parameters are treated as random variables in the Bayesian framework, we make probability statements about model parameters in the frequentist framework, parameters are fixed non-random quantities and the probability statements concern the data Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 36 / 42
41 Bayesian inference (continued) As with any analysis, we start by positing a model, p(d Ω) This is the sampling distribution of the data, equivalent to the likelihood in classical analyses From a Bayesian point of view Ω is unknown so should have a probability distribution reflecting our uncertainty about it before seeing the data need to specify a prior distribution p(ω) D is known so we should condition on it use Bayes theorem to obtain conditional probability distribution for unobserved quantities of interest given the data: This is the posterior distribution p(ω D) p(ω) p(d Ω) Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 37 / 42
42 Bayesian computational methods Bayesian inference centres around the posterior distribution p(ω D) p(d Ω) p(ω) where Ω is typically a large vector Ω = {Ω 1, Ω 2,..., Ω k } p(d Ω) and p(ω) will often be available in closed form, but p(ω D) is usually not analytically tractable, and we want to obtain marginal posteriors p(ω i D) =... p(ω D) dω ( i) where Ω ( i) denotes the vector of Ω s excluding Ω i calculate properties of p(ω i D), such as E(Ω i D) = Ω i p(ω i D)dΩ i Pr(Ω i > T D) = numerical integration becomes vital T p(ω i D)dΩ i Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 38 / 42
43 Markov Chain Monte Carlo (MCMC) integration Suppose we can draw samples from the joint posterior distribution for Ω, i.e. (Ω (1) 1 Then θ (1) 1,..., Ω(1) k ), (Ω(2) 1,..., Ω(2) k ),..., (Ω(N) 1,..., Ω (N) k ) p(ω D),..., Ω(N) 1 are a sample from the marginal posterior p(ω 1 D) E(g(Ω 1 )) = g(ω 1 )p(ω 1 D)dΩ 1 1 N N i=1 g(ω(i) 1 ) this is Monte Carlo integration Independent sampling from p(ω D) may be difficult BUT dependent sampling from a Markov chain with p(ω D) as its stationary (equilibrium) distribution is easier Theorems exist which prove convergence in limit as N even if the sample is dependent this is Markov Chain Monte Carlo integration Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 39 / 42
44 Bayesian methods for handling missing data Bayesian approach treats missing data as additional unknown quantities for which a posterior distribution can be estimated no fundamental distinction between missing data and unknown parameters In our previous notation D = {z obs, m} Ω = {z mis, β, θ} Inferential machinery available for Bayesian parameter estimation extends automatically to models with missing data Just need to specify appropriate joint model for the observed and missing data and model parameters, and estimate in usual way using MCMC Fully model-based approach to missing data Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 40 / 42
45 Advantages of Bayesian methods Fully Bayesian models are theoretically sound enable coherent model estimation allow uncertainty to be fully propagated Bayesian models can easily be adapted to include partially observed cases incorporate realistic assumptions about the reasons for the missingness Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 41 / 42
46 Coming up It is helpful to distinguish between missing response and missing covariate data (regression context) i.e. we let z = (y, x) where y is the response of interest and x is a set of covariates ignorable and non-ignorable missingness mechanisms, since when mechanism is ignorable, specifying the joint model reduces to specifying f (z obs, z mis β), and ignoring f (m z obs, θ) In Lecture 2, we will look at missing response data Missing covariate data is the subject of Lecture 3 We will then discuss a general strategy for carrying out Bayesian regression analysis with missing data, including a range of recommended sensitivity analyses (Lecture 4) In Lecture 5, we will compare fully Bayesian models with Multiple Imputation and end with a general discussion Bayesian Missing Data Course (Lecture 1) Introduction to Missing Data 42 / 42
47 Lecture 2. Bayesian methods for missing response data Bayesian Missing Data Course (Lecture 2) Missing response data 1 / 26
48 Lecture Outline Missing responses and missing covariates pose different problems In this lecture we will focus on analysing data with missing responses Using the HAMD example introduced in Lecture 1, we will see that dealing with missing responses in a principled way within a Bayesian framework is trivial when the missing data mechanism is ignorable but is very dependent on assumptions when the missing data mechanism is informative This dependence on assumptions makes sensitivity analysis crucial Bayesian Missing Data Course (Lecture 2) Missing response data 2 / 26
49 HAMD example: recall exploratory analysis 50 Individual Profiles Mean Response Profiles HAMD score treatment 1 treatment 2 treatment 3 HAMD score treatment 1 treatment 2 treatment week week Bayesian Missing Data Course (Lecture 2) Missing response data 3 / 26
50 HAMD example: complete case analysis (CC) We start by carrying out a complete case analysis to provide a comparison with preferred methods as a building block for more complex models CC requires the specification of an analysis model For the HAMD data, CC means discarding partial data from 121 out of 367 subjects specifying an analysis model which takes account of the repeated structure (observations are nested within individuals) We will specify a hierarchical model with random intercepts and random slopes, and for simplicity ignore any centre effects assume linear relationships Bayesian Missing Data Course (Lecture 2) Missing response data 4 / 26
51 HAMD example: analysis model Specify a random effects (hierarchical) model: y iw Normal(µ iw, σ 2 ) µ iw = α i + β (treat(i),i) w α i Normal(µ α, σ 2 α) random intercepts β (1,i) Normal(µ β1, σ 2 β 1 ) random slopes β (2,i), β (3,i) follow common prior distributions similar to β (1,i) µ α, µ β1, µ β2, µ β3 Normal(0, 10000) vague prior on hyperparameters σ α, σ β1, σ β2, σ β3 Uniform(0, 100) vague prior on hyperparameters 1 σ 2 Gamma(0.001, 0.001) vague prior on precision y iw = HAMD score for individual i in week w (weeks 0,...,4) treat(i) = treatment indicator of individual i (takes values 1, 2 or 3) w = week of the visit, takes value 0 for visit before treatment and values 1-4 for follow-up visits Bayesian Missing Data Course (Lecture 2) Missing response data 5 / 26
52 HAMD example: interpretation of results Study objective: are there any differences in the effects of the 3 treatments on the change in HAMD score over time? So we are particularly interested in the differences in the slope parameters (contrasts), i.e. 1 µ β1 µ β2 2 µ β1 µ β3 3 µ β2 µ β3 Results for complete case analysis Bayesian Missing Data Course (Lecture 2) Missing response data 6 / 26
53 Recap: selection model Recall, in general, the joint model for a set of data (z) and missing data indicator (m) can be factorised as f(z obs, z mis, m β, θ) = f(m z obs, z mis, θ)f(z obs, z mis β) model of missingness analysis model When the missing data mechanism is ignorable, we can ignore the model of missingness and just fit the analysis model So we need only specify f (z obs, z mis β) In the regression context, we let z = (y, x) where y is the response of interest x is a set of covariates Bayesian Missing Data Course (Lecture 2) Missing response data 7 / 26
54 Missing response data - assuming missing data mechanism is ignorable (I) In this case, z mis = y mis and z obs = (y obs, x) Usually treat fully observed covariates as fixed constants rather than random variables with a distribution Model f (z obs, z mis β) reduces to specification of f (y obs, y mis x, β) f (y obs, y mis x, β) is just the usual likelihood we would specify for fully observed response y Estimating the missing responses y mis is equivalent to posterior prediction from the model fitted to the observed data Imputing missing response data under an ignorable mechanism will not affect estimates of model parameters Bayesian Missing Data Course (Lecture 2) Missing response data 8 / 26
55 Missing response data - assuming missing data mechanism is ignorable (II) In WinBUGS denote missing response values by NA in the data file specify response distribution (likelihood) as you would for complete data missing data are treated as additional unknown parameters WinBUGS will automatically simulate values for the missing observations according to the specified likelihood distribution, conditional on the current values of all relevant unknown parameters Bayesian Missing Data Course (Lecture 2) Missing response data 9 / 26
56 HAMD example: ignorable missing data mechanism Assume the missing data mechanism is ignorable for the HAMD example the probability of the score being missing is not related to the current score (or change in score since the previous week) Use the same model (and WinBUGS code) as for the complete case analysis But the data now includes incomplete records (with the missing values denoted by NA) Bayesian Missing Data Course (Lecture 2) Missing response data 10 / 26
57 HAMD example: impact on treatment comparisons Table: posterior mean (95% credible interval) for the contrasts (treatment comparisons) from random effects models fitted to the HAMD data treatments complete cases all cases 1 v (-0.03,1.00) 0.74 (0.25,1.23) 1 v (-1.06,-0.04) (-1.01,-0.01) 2 v (-1.56,-0.55) (-1.73,-0.77) individuals with missing scores ignored individuals with missing scores included under the assumption that the missingness mechanism is ignorable Including all the partially observed cases in the analysis provides stronger evidence that: treatment 2 is more effective than treatment 1 treatment 2 is more effective than treatment 3 Bayesian Missing Data Course (Lecture 2) Missing response data 11 / 26
58 HAMD example: non-ignorable missing data mechanism Before defining a model for the missing data mechanism think about the process that led to the missingness gather information from the literature seek insight from those involved in the data collection process For the HAMD model patients for whom the treatment is successful and get better may decide not to continue in the study conversely, if they are not showing any improvement or feeling worse, they may seek alternative treatment and drop-out of the study in either case, the probability of obtaining a HAMD score is dependent on the change in the patient s depression over the previous week, i.e. since the last HAMD score Bayesian Missing Data Course (Lecture 2) Missing response data 12 / 26
59 Model for the missing data mechanism Then, translate these findings into a statistical model, e.g. m i Bernoulli(p i ) r logit(p i ) = θ 0 + θ k w ki + δy i k=1 w 1i,..., w ri is a vector of variables which are predictive of the response missingness The inclusion of the response, y changes the missingness assumption from MAR to MNAR provides the link with the analysis model Bayesian Missing Data Course (Lecture 2) Missing response data 13 / 26
60 Recall: DAG for Missing At Random (MAR) A regression model of interest + model for probability of y missing β θ µ i x i σ 2 y i p i Model of Interest individual i m i Model of Missingness For MAR, the only common element (x) in the two models is fully observed Bayesian Missing Data Course (Lecture 2) Missing response data 14 / 26
61 Recall: DAG for Missing Not At Random (MNAR) A regression model of interest + model for probability of y missing β θ µ i x i σ 2 y i p i Model of Interest individual i m i Model of Missingness For MNAR, one of the common elements (y) has missing values which are estimated Bayesian Missing Data Course (Lecture 2) Missing response data 15 / 26
62 Important choices The shape of the relationship between the response and the probability of missingness is a linear relationship adequate? would a more complex shape, e.g. piecewise linear functional form, be better? for some datasets, it may be intuitively plausible that the response is more likely to be missing if it takes high or low values The most appropriate way of including the response with longitudinal data, could use change in response between current and previous values In the absence of any prior knowledge, recommended strategy is to assume a linear relationship between the probability of missingness and the response or change in response (Mason(2009), Chapter 4) Bayesian Missing Data Course (Lecture 2) Missing response data 16 / 26
63 Estimating parameters associated with the response The parameters associated with the response, δ, are identified by the parametric assumptions in the analysis model (AM) the model of missingness (MoM) Missing responses are imputed in a way that is consistent with the distributional assumptions in the AM given their covariates in the AM Thus δ are (at least weakly) identified by the observed data in combination with the other model assumptions Estimation difficulties increase for more complex models of missingness Advisable to use informative priors if possible Bayesian Missing Data Course (Lecture 2) Missing response data 17 / 26
64 How the AM distributional assumptions are used Illustrative example (Daniels & Hogan (2008), Section 8.3.2) Consider a cross-sectional setting with a single response no covariates Suppose we specify a linear MoM, logit(p i ) = θ 0 + δy i Frequency histogram of observed responses y If we assume the AM follows a normal distribution, y i N(µ i, σ 2 ) must fill in the right tail δ > 0 If we assume the AM follows a skew-normal distribution δ = 0 Bayesian Missing Data Course (Lecture 2) Missing response data 18 / 26
65 Uncertainty in the AM distributional assumptions Inference about δ is heavily dependent on the AM distributional assumptions about the residuals in combination with the choice and functional form of the covariates Unfortunately the AM distribution is unverifiable from the observed data when the response is MNAR Different AM distributions lead to different results Hence sensitivity analysis required to explore impact of different plausible AM distributions Bayesian Missing Data Course (Lecture 2) Missing response data 19 / 26
66 HAMD example: non-ignorable missing data mechanism If we think the probability of the score being missing might be related to the current score, then we must jointly fit 1 an analysis model - already defined 2 a model for the missing data mechanism Assume that the probability of score being missing is related to its current value, then model the missing response indicator as m iw Bernoulli(p iw ) logit(p iw ) = θ 0 + δ(y iw ȳ) θ 0, δ mildly informative priors where ȳ is the mean score of observed ys Bayesian Missing Data Course (Lecture 2) Missing response data 20 / 26
67 HAMD example: MAR v MNAR Table: posterior mean (95% credible interval) for the contrasts (treatment comparisons) from random effects models fitted to the HAMD data treatments complete cases 1 all cases (mar) 2 all cases (mnar) 3 1 v (-0.03,1.00) 0.74 (0.25,1.23) 0.75 (0.26,1.24) 1 v (-1.06,-0.04) (-1.01,-0.01) (-0.98,0.05) 2 v (-1.56,-0.55) (-1.73,-0.77) (-1.70,-0.75) 1 individuals with missing scores ignored 2 individuals with missing scores included under the assumption that the missingness mechanism is ignorable 3 individuals with missing scores included under the assumption that the missingness mechanism is non-ignorable Allowing for informative missingness with dependence on the current HAMD score: has a slight impact on the treatment comparisons yields a 95% interval comparing treatments 1 & 3 that includes 0 Bayesian Missing Data Course (Lecture 2) Missing response data 21 / 26
68 HAMD example: sensitivity analysis - MoM Since the true missingness mechanism is unknown and cannot be checked, sensitivity analysis is essential We have already assessed the results of assuming the missingness mechanism is ignorable using an informative missingness mechanism of the form logit(p i ) = θ 0 + δ(y iw ȳ) However, we should also look at alternative informative missingness mechanisms, e.g. allow drop-out probability to be dependent on the change in score logit(p i ) = θ 0 + θ 1 (y i(w 1) ȳ) + δ 2 (y iw y i(w 1) ) allow different θ and δ for each treatment use different prior distributions Bayesian Missing Data Course (Lecture 2) Missing response data 22 / 26
69 HAMD example: sensitivity analysis results Table: posterior mean (95% credible interval) for the contrasts (treatment comparisons) from random effects models fitted to all the HAMD data treatments mar mnar1 mnar2 1 v (0.25,1.23) 0.75 (0.26,1.24) 0.72 (0.23,1.22) 1 v (-1.01,-0.01) (-0.98,0.05) (-1.09,-0.11) 2 v (-1.73,-0.77) (-1.70,-0.75) (-1.80,-0.84) probability of missingness dependent on current score probability of missingness dependent on change in score This is a sensitivity analysis, we do NOT choose the best" model Model comparison with missing data is very tricky we cannot use the DIC automatically generated by WinBUGS on its own (Mason et al., 2012a) The range of results should be presented Bayesian Missing Data Course (Lecture 2) Missing response data 23 / 26
70 HAMD example: sensitivity analysis - AM Sensitivity to the assumptions about the AM should also be explored For HAMD data, possibilities include allow for non-linearity by including a quadratic term µ iw = α i + β treat(i) w + γ treat(i) w 2 include centre effects by allowing a different intercept for each centre instead of using random effects, account for the repeated structure using an autoregressive model to explicitly model the autocorrelation between weekly visits for each individual A comprehensive sensitivity analysis will pair different combinations of AMs and MoMs Bayesian Missing Data Course (Lecture 2) Missing response data 24 / 26
71 Summary and extensions Missing response data is trivial to handle in the Bayesian framework under the assumption of an ignorable missing data mechanism equivalent to posterior prediction from the model fitted to the observed data Using the HAMD example, we have shown a general framework for modelling informative missing response mechanisms by jointly modelling the analysis model of interest a model for the missing data indicator, which is a function of the missing response values (and possibly other observed covariates) Bayesian Missing Data Course (Lecture 2) Missing response data 25 / 26
72 Summary and extensions The most appropriate way of modelling the missing data indicator will be problem specific Some elaboration may be required to accommodate different types of drop-out (e.g. death and recovery) multinomial regression model for categorical missing data indicator Some datasets may have informative drop-in (e.g. medics start to monitor patients if they become more unwell) For longitudinal studies with drop-out, an alternative is to replace the missingness indicator with a variable representing the time to drop-out and model this using (discrete or continuous time) survival techniques Similar selection model approaches are available in likelihood settings, but their implementation may require non-standard maximisation and numerical integration algorithms Bayesian Missing Data Course (Lecture 2) Missing response data 26 / 26
73 Lecture 3. Bayesian methods for missing covariate data Bayesian Missing Data Course (Lecture 3) Missing covariate data 1 / 23
74 Lecture Outline We have now looked at Bayesian methods for missing responses Missing covariates pose some additional problems In this lecture, we look at methods for a single covariate (assuming missing data mechanism is ignorable) multiple covariates (assuming missing data mechanism is ignorable) allowing the missing data mechanism to be informative Bayesian Missing Data Course (Lecture 3) Missing covariate data 2 / 23
75 Missing responses v missing covariates For missing responses, recall that missing values are automatically simulated from the analysis model, f (y obs, y mis x, β) assuming ignorable missingness, no additional sub-models are required For covariates, we will see that we must build an imputation model to predict their missing values regardless of the missingness mechanism, at least one additional sub-model is always required Bayesian Missing Data Course (Lecture 3) Missing covariate data 3 / 23
76 Treatment of covariates with missing values Recall that we wish to evaluate the joint model f (z obs, z mis, m β, θ) In this case, z mis = x mis and z obs = (y, x obs ) To include records with missing covariates (associated response and other covariates observed) we now have to treat covariates as random variables rather than fixed constants must build an imputation model to predict their missing values Typically, the joint model (ignoring the missingness mechanism), f (z obs, z mis β) = f (y, x obs, x mis β) is factorised as f (y, x obs, x mis β) = f (y x obs, x mis, β y )f (x obs, x mis β x ) where β is partitioned into conditionally independent subsets (β y, β x ) Bayesian Missing Data Course (Lecture 3) Missing covariate data 4 / 23
77 Analysis model and covariate imputation model (CIM) The first term in the joint model factorisation, f (y x obs, x mis, β y ), is the usual likelihood for the response given fully observed covariates (analysis model) The second term, f (x obs, x mis β x ) can be thought of as a prior model for the covariates (which are treated as random variables, not fixed constants), e.g. joint prior distribution, say MVN regression model for each variable with missing values It is not necessary to explicitly include response, y, as a predictor in the prior imputation model for the covariates, as its association with x is already accounted for by the first term in the joint model factorisation (unlike multiple imputation) Bayesian Missing Data Course (Lecture 3) Missing covariate data 5 / 23
78 A typical DAG A regression model of interest with fully observed covariates β µ i x i σ 2 y i Model of Interest individual i Note: x and y are completely observed Bayesian Missing Data Course (Lecture 3) Missing covariate data 6 / 23
79 DAG with Missing Covariates A regression model of interest + prior model for covariates β y β x µ i x i σ 2 y i Covariate Imputation Model individual i Model of Interest Note: y is completely observed, but x has missing values (assumed to have ignorable missingness mechanism) Bayesian Missing Data Course (Lecture 3) Missing covariate data 7 / 23
80 Implementation In WinBUGS Denote missing covariate values by NA in the data file Specify usual regression analysis model, which will depend on partially observed covariates In addition, specify prior distribution or regression model for the covariate(s) with missing values WinBUGS will automatically simulate values from the posterior distribution of the missing covariates (which will depend on the prior model for the covariates and the likelihood contribution from the corresponding response variable) Uncertainty about the covariate imputations is automatically accounted for in the analysis model Bayesian Missing Data Course (Lecture 3) Missing covariate data 8 / 23
81 Single covariate with missing values (x) There are 2 obvious ways of building the covariate imputation model 1 Specify a distribution, e.g. if x is continuous specify xi N(ν, ς 2 ) assume vague priors for ν and ς 2 2 Build a regression model relating x i to other observed covariates, e.g. if x is binary specify x i Bernoulli(q i ) s logit(q i ) = φ 0 + φ r z si k=1 φ 0, φ 1,..., φ s prior distribution z1i,..., z si is a vector of fully observed covariates Compare pattern of imputed values with the observed values to check the adequacy of the covariate imputation model (CIM) Bayesian Missing Data Course (Lecture 3) Missing covariate data 9 / 23
82 LBW example: low birth weight data Study objective: is there an association between trihalomethane (THM) concentrations and the risk of full term low birth weight? THM is a by-product of chlorine water disinfection potentially harmful for reproductive outcomes The variables we will use are: Y : binary indicator of low birth weight (outcome) X: binary indicator of THM concentrations (exposure of interest) C: mother s age, baby gender, deprivation index (vector of measured confounders) U: smoking (a partially measured confounder) So z mis = U mis and z obs = (Y, X, C, U obs ) We have data for 8969 individuals, but only 931 have an observed value for smoking 90% of individuals will be discarded if we use CC Bayesian Missing Data Course (Lecture 3) Missing covariate data 10 / 23
83 LBW example: missingness assumptions Assume that smoking is MAR probability of smoking being missing does not depend on whether the individual smokes this assumption is reasonable as the missingness is due to the sample design of the underlying datasets Also assume that the other assumptions for ignorable missingness hold (Lecture 1), so we do not need to specify a model for the missingness mechanism However, since smoking is a covariate, we must specify an imputation model if we wish to include individuals with missing values of smoking in our dataset Bayesian Missing Data Course (Lecture 3) Missing covariate data 11 / 23
84 LBW example: specification of joint model Analysis model: logistic regression for outcome, low birth weight Y i Bernoulli(p i ) logit(p i ) = β 0 + β X X i + β T C C i + β U U i β 0, β X,... Normal(0, ) Imputation model (1): prior distribution for missing smoking U i Bernoulli(q i ) q i Beta(1, 1) Key assumption: distribution of missing values of U is exchangeable with observed distribution Imputed U s will depend on posterior distribution of q estimated from observed U s and on feedback from the analysis model about the relationship between U and Y (adjusted for X and C) In a Bayesian joint model, feedback means we don t need to explicitly include the response, Y, in the imputation model cf multiple imputation, where we do Bayesian Missing Data Course (Lecture 3) Missing covariate data 12 / 23
85 LBW example: Imputation model (2) Often, the simple exchangeability assumption for the prior on the missing U is not reasonable May want to stratify prior by other fully observed covariates, or build a regression model Particularly important to account for any variables related to the missing data mechanism, to strengthen plausibility of MAR assumption Imputation model (2): logistic regression for missing covariate, smoking U i Bernoulli(q i ) logit(q i ) = φ 0 + φ X X i + φ T C C i φ 0, φ X,... Normal(0, ) Note that study design (missing data mechanism) depends on the deprivation index (C), so this has been included in the imputation model Bayesian Missing Data Course (Lecture 3) Missing covariate data 13 / 23
86 LBW example: graphical representation β φ c i p i x i q i y i individual i Model of Interest u i Covariate Imputation Model Bayesian Missing Data Course (Lecture 3) Missing covariate data 14 / 23
87 LBW example: results Odds ratio (95% interval) CC (N=931) All (N=8969) X Trihalomethanes > 60µg/L 2.36 (0.96,4.92) 1.17 (1.01,1.37) C Mother s age (0.32,1.93) 1.05 (0.74,1.41) (0.00,0.51) 0.80 (0.55,1.14) (0.39,3.80) 1.14 (0.73,1.69) C Male baby 0.84 (0.34,1.75) 0.76 (0.58,0.95) C Deprivation index 1.74 (1.05,2.90) 1.34 (1.17,1.53) U Smoking 1.86 (0.73,3.89) 1.92 (0.80,3.82) Reference group CC analysis is very uncertain Extra records shrink intervals for X coefficient substantially Bayesian Missing Data Course (Lecture 3) Missing covariate data 15 / 23
88 LBW example: results Odds ratio (95% interval) CC (N=931) All (N=8969) X Trihalomethanes > 60µg/L 2.36 (0.96,4.92) 1.17 (1.01,1.37) C Mother s age (0.32,1.93) 1.05 (0.74,1.41) (0.00,0.51) 0.80 (0.55,1.14) (0.39,3.80) 1.14 (0.73,1.69) C Male baby 0.84 (0.34,1.75) 0.76 (0.58,0.95) C Deprivation index 1.74 (1.05,2.90) 1.34 (1.17,1.53) U Smoking 1.86 (0.73,3.89) 1.92 (0.80,3.82) Reference group Little impact on U coefficient, reflecting uncertainty in imputations Bayesian Missing Data Course (Lecture 3) Missing covariate data 16 / 23
89 Aside: bias in CC In regression with missing covariates, the conditions under which CC is biased do not fit neatly into the MCAR/MAR/MNAR categorisation (White & Carlin, 2010) For example, consider the case of a single covariate, X, with missing values and fully observed Y CC is unbiased if missingness in X is MCAR MNAR dependent on X CC is biased if missingness in X is MAR dependent on Y MNAR dependent on X and Y Bayesian Missing Data Course (Lecture 3) Missing covariate data 17 / 23
90 Multiple covariates with missing values The covariate imputation model gets more complex if > 1 missing covariates typically need to account for correlation between missing covariates could assume multivariate normality if covariates all continuous for mixed binary, categorical and continuous covariates, could fit latent variable (multivariate probit) model (Chib and Greenberg 1998) We now extend our LBW example to 2 covariates Bayesian Missing Data Course (Lecture 3) Missing covariate data 18 / 23
91 LBW example: two covariates with missing values Assume that our set of partially measured confounders (U) contains 2 variables smoking ethnicity Imputation model: Multivariate Probit for P(U X, C) U i MVN(µ i, Σ) µ i = γ 0 + γ X X i + γc T C i U ij = I(Uij > 0), j = 1, 2 ), µ i = ( µi1 ) ( 1 κ, Σ = κ 1 ( U U i = i1 Ui2 µ i2 κ Uniform( 1, 1); γ 0, γ X, γ C Normal(0, 10000) See Molitor et al. (2009) for further details ) Bayesian Missing Data Course (Lecture 3) Missing covariate data 19 / 23
92 LBW example: two missing covariates - results CC (THM) FBM (THM) 1.87 (0.75,4.00) 1.15 (0.87,1.49) CC (smoking) FBM (smoking) 3.03 (1.13,6.73) 2.89 (1.01,6.83) CC (ethnicity) FBM (ethnicity) 3.50 (0.93,9.34) 3.24 (1.66,5.95) odds ratio Plot shows posterior distribution with median marked. Posterior mean and 95% interval are shown on the right. Bayesian Missing Data Course (Lecture 3) Missing covariate data 20 / 23
93 Informative missingness mechanism for covariates If we assume that smoking is MNAR, then we must add a third part to the model a model of missingness with a missingness indicator variable (n i ) for smoking as the response e.g. n i Bernoulli(r i ) logit(r i ) = θ 0 + θ T W W i + δu i W is a vector of observed variables predictive of the covariate missingness and U is the (possibly missing) smoking variable This Model of Covariate Missingness (MoCM) is similar to the Model of Response Missingness (MoRM) discussed in Lecture 2. Bayesian Missing Data Course (Lecture 3) Missing covariate data 21 / 23
94 Summary and extensions Excluding missing covariates can be biased and potentially very inefficient Handling missing covariates in a Bayesian framework requires specification of an additional sub-model to impute the covariates, even under MCAR or MAR assumptions Recommendations for building the covariate imputation model are broadly the same as for multiple imputation include variables that are related to the imputed variable include variables that are related to the missingness mechansim account for correlation structure between multiple variables with missing values account for hierarchical structure in the covariates if appropriate A key difference from multiple imputation is that it is not necessary to include the response in the covariate imputation model For covariates that are missing not at random (NMAR), a model of missingness must also be specified Bayesian Missing Data Course (Lecture 3) Missing covariate data 22 / 23
95 Summary and extensions Missing responses and covariates Bayesian models with both missing responses and covariates have the potential to become quite complicated, particularly if we cannot assume ignorable missingness However, the Bayesian framework is well suited to the process of building complex models, linking smaller sub-models as a coherent joint model A typical model may consist of 3 (or more) parts: 1 analysis model 2 covariate imputation model 3 model(s) of missingness (for response and/or covariates) In the next lecture, we will look at a strategy for building complex Bayesian models for analysing incomplete data Bayesian Missing Data Course (Lecture 3) Missing covariate data 23 / 23
96 Lecture 4. A general strategy for modelling missing data using Bayesian methods Bayesian Missing Data Course (Lecture 4) Strategy 1 / 41
97 Lecture Outline Modelling data with missing values can be complicated Guidance on the practicalities of approaching this task might help We provide this by proposing a general strategy In this lecture, we present a strategy for a statistically principled investigation of data with missing covariates and/or responses use an illustrative example to demonstrate each step discuss adaptions and extensions Although this strategy was designed for Bayesian modelling, the framework could be adopted for other inference paradigms. Bayesian Missing Data Course (Lecture 4) Strategy 2 / 41
98 Illustrative example Research Question: is a single mother s rate of pay affected by gaining a partner? Data: taken from the Millennium Cohort Study (MCS), which follows 18,000+ children born in the UK at the start of the Millennium includes information about the children s families We use data from sweeps 1 and 2 on the cohort member s mother, with inclusion criteria: single in sweep 1 in work not self-employed Bayesian Missing Data Course (Lecture 4) Strategy 3 / 41
99 Strategy Overview The strategy consists of two parts It allows constructing a base model assessing conclusions from this base model against a selection of well chosen sensitivity analyses the uncertainty from the missing data to be taken into account additional sources of information to be utilised It can be implemented using currently available software, e.g. WinBUGS Bayesian Missing Data Course (Lecture 4) Strategy 4 / 41
100 Advantages of a Bayesian framework Bayesian models are formulated in a modular way ideal for iteratively building and modifying a base model Uncertainty about the imputed missing values is automatically and coherently propagated through the model reflected in the estimates of interest Provides scope for including extra data or other information through additional submodels informative priors Bayesian Missing Data Course (Lecture 4) Strategy 5 / 41
101 Schematic Diagram 1: select AM using complete cases note plausible alternatives 2: add CIM 3: add MoRM AM = Analysis Model CIM = Covariate Imputation Model MoRM = Model of Response Missingness 4: seek additional data BASE MODEL 5: elicit expert knowledge 6: ASSUMPTION SENSITIVITY 7: PARAMETER SENSITIVITY report robustness YES 8: Are conclusions robust? NO determine region of high plausibility assess fit recognise uncertainty Bayesian Missing Data Course (Lecture 4) Strategy 6 / 41
102 Schematic Diagram: constructing a base model 1: select AM using complete cases note plausible alternatives 2: add CIM 3: add MoRM AM = Analysis Model CIM = Covariate Imputation Model MoRM = Model of Response Missingness 6: ASSUMPTION SENSITIVITY 4: seek additional data BASE MODEL 7: PARAMETER SENSITIVITY 5: elicit expert knowledge Strategy consists of two parts: Constructing a base model Assessing conclusions from this base model against a selection of well chosen sensitivity analyses report robustness YES 8: Are conclusions robust? NO determine region of high plausibility assess fit recognise uncertainty Bayesian Missing Data Course (Lecture 4) Strategy 7 / 41
103 Schematic Diagram: sensitivity analysis 1: select AM using complete cases note plausible alternatives 2: add CIM 3: add MoRM AM = Analysis Model CIM = Covariate Imputation Model MoRM = Model of Response Missingness 6: ASSUMPTION SENSITIVITY 4: seek additional data BASE MODEL 7: PARAMETER SENSITIVITY 5: elicit expert knowledge Strategy consists of two parts: Constructing a base model Assessing conclusions from this base model against a selection of well chosen sensitivity analyses report robustness YES 8: Are conclusions robust? NO determine region of high plausibility assess fit recognise uncertainty Bayesian Missing Data Course (Lecture 4) Strategy 8 / 41
104 Before the strategy: step 0 The strategy consists of a series of model building steps Before starting, the missingness should be explored to determine which steps are required? are any other modifications needed? In particular which variables have missing values? what is the extent and pattern of missingness? think about plausible explanations for the missingness? Bayesian Missing Data Course (Lecture 4) Strategy 9 / 41
105 Illustrative example: step 0 For simplification, we restrict our MCS dataset to individuals fully observed in sweep 1 Dataset has 505 individuals, with 37% missing response (pay) and 33% missing covariates sweep 2 data covariates observed missing pay observed missing Survey methodology literature has shown that income non-response is usually non-ignorable Bayesian Missing Data Course (Lecture 4) Strategy 10 / 41
106 Part 1 (steps 1-5): constructing a base model This part involves building a joint model as follows: 1 choose an analysis model 2 add a covariate imputation model 3 add a model of response missingness Optionally, the amount of available information can be increased by incorporating data from other sources and/or expert knowledge The strategy allows informative missingness in the response but assumes that the covariates are MAR However, it can be adapted to reflect alternative assumptions Bayesian Missing Data Course (Lecture 4) Strategy 11 / 41
107 Step 1 - analysis model note plausible alternatives 6: ASSUMPTION SENSITIVITY 1: select AM using complete cases 2: add CIM 3: add MoRM 4: seek additional data BASE MODEL 7: PARAMETER SENSITIVITY 5: elicit expert knowledge Constructing a base model Form an initial Analysis Model (AM) based on complete cases previous knowledge Includes choosing transform for the response model structure set of explanatory variables Allow for report robustness YES 8: Are conclusions robust? NO determine region of high plausibility assess fit hierarchical structure other data complexities recognise uncertainty Error distribution is key assumption Bayesian Missing Data Course (Lecture 4) Strategy 12 / 41
108 Illustrative example: step 1 - AM description Response: log of hourly net pay 4 explanatory variables: Name Description Details age age at interview continuous a reg region of country 1 = London; 2 = other sing b single/partner 1 = single; 2 = partner stratum country by ward type 9 levels c a centred and standardised b always 1 in sweep 1 by definition c 3 strata for England ( advantaged, disadvantaged and ethnic minority ); 2 strata for Wales, Scotland and Northern Ireland ( advantaged and disadvantaged ) Error distribution: t with 4 degrees of freedom (t 4 ) provides robustness to outliers Bayesian Missing Data Course (Lecture 4) Strategy 13 / 41
109 Illustrative example: step 1 - AM equations Alternative: cube root transform log of hourly pay (hpay) Alternative: normal errors robustness to outliers individual random effects y it t 4 (µ it, σ 2 ) µ it = α i + γ s(i) + p stratum specific intercepts k=1 β kx kit & vague priors e.g. β k N(0, ) age (mother s age) reg (London/other) sing (single/partner) Bayesian Missing Data Course (Lecture 4) Strategy 14 / 41
110 Illustrative example: step 1 - AM results Lower pay is associated with gaining a partner between sweeps living outside London Higher pay is associated with increasing age sing age reg 0.08 ( 0.15, 0.01) 0.15 (0.11,0.18) 0.18 ( 0.31, 0.05) parameter estimates Plot shows posterior distribution with median marked. Posterior mean and 95% interval are shown on the right. Bayesian Missing Data Course (Lecture 4) Strategy 15 / 41
111 Step 2 - covariate imputation model note plausible alternatives 1: select AM using complete cases 2: add CIM 3: add MoRM 6: ASSUMPTION SENSITIVITY 4: seek additional data BASE MODEL 7: PARAMETER SENSITIVITY 5: elicit expert knowledge Constructing a base model Add a Covariate Imputation Model (CIM) to produce realistic imputations of any missing covariates See Lecture 3 for details report robustness YES 8: Are conclusions robust? NO determine region of high plausibility assess fit recognise uncertainty Bayesian Missing Data Course (Lecture 4) Strategy 16 / 41
112 Illustrative example: step 2 - CIM description Assume covariates are missing at random (MAR) stratum does not change between sweeps Imputation of missing sweep 2 values required for age, reg and sing For simplicity, missing values of age and reg set before analysis using simple rules reg: assign sweep 1 value age: sweep 1 age + mean of observed differences in age between sweeps For sing we set up a statistical model Bayesian Missing Data Course (Lecture 4) Strategy 17 / 41
113 Illustrative example: step 2 - CIM equations sing i2 Bernoulli(q i ) q i = φ s(i) + (φ age age i2 ) + (φ reg reg i2 ) stratum specific intercepts & vague priors e.g. φ k N(0, ) As a reasonableness check, we compare the distribution of the imputed and observed values of sing 2 observed values: 34% gained a partner imputed values: 36% (27%,45%) gained a partner CIM could be expanded to include age and reg (see Lecture 3) Bayesian Missing Data Course (Lecture 4) Strategy 18 / 41
114 Step 3 - model of response missingness 1: select AM using complete cases note plausible alternatives 6: ASSUMPTION SENSITIVITY 2: add CIM 3: add MoRM 4: seek additional data BASE MODEL 7: PARAMETER SENSITIVITY 5: elicit expert knowledge Constructing a base model Add a Model of Response Missingness (MoRM) to allow informative missingness in the response See Lecture 2 for details report robustness YES 8: Are conclusions robust? NO determine region of high plausibility assess fit recognise uncertainty Bayesian Missing Data Course (Lecture 4) Strategy 19 / 41
115 Illustrative example: step 3 - MoRM description Allow informative missingness in the response by modelling a missing value indicator (m i ) for sweep 2 pay (hpay i2 ) s.t. { 0: hpayi2 observed m i = 1: hpay i2 missing Use a logit model for response missingness, i.e. m i Bernoulli(p i ); logit(p i ) =? Previous work in this area informs choice of predictors of missing income For simplicity, linear relationships are assumed Untransformed hourly pay used in this sub-model Bayesian Missing Data Course (Lecture 4) Strategy 20 / 41
116 Illustrative example: step 3 - MoRM equations missing value indicator for sweep 2 hourly pay m i Bernoulli(p i ) logit(p i ) = θ 0 + r probability of sweep 2 hourly pay being missing level of pay at sweep 1 k=1 θ kw ki + κ hpay i1 + δ (hpay i2 hpay i1 ) sc (social class) eth (ethnic group) ctry (country) & vague priors change in pay between sweeps The inclusion of the term δ (hpay i2 hpay i1 ) allows the response missingness to be MNAR If δ = 0, then we have MAR missingness Bayesian Missing Data Course (Lecture 4) Strategy 21 / 41
117 The joint model (steps 1-3): schematic diagram analysis model parameters model of response missingness parameters response with missingness probability of missingness covariate imputation model parameters covariates with missingness fully observed covariates missingness indicator Bayesian Missing Data Course (Lecture 4) Strategy 22 / 41
118 The joint model (steps 1-3): schematic diagram analysis model parameters analysis model model of response missingness parameters response with missingness probability of missingness covariate imputation model parameters covariates with missingness fully observed covariates missingness indicator Bayesian Missing Data Course (Lecture 4) Strategy 22 / 41
119 The joint model (steps 1-3): schematic diagram analysis model parameters analysis model model of response missingness parameters response with missingness probability of missingness covariate imputation model parameters covariates with missingness fully observed covariates missingness indicator covariate imputation model Bayesian Missing Data Course (Lecture 4) Strategy 22 / 41
120 The joint model (steps 1-3): schematic diagram analysis model parameters model of response missingness analysis model model of response missingness parameters this part required for non-ignorable missingness in the response response with missingness probability of missingness covariate imputation model parameters covariates with missingness fully observed covariates missingness indicator covariate imputation model Bayesian Missing Data Course (Lecture 4) Strategy 22 / 41
121 Illustrative example: results from base model A greater proportion of individuals are imputed to gain a partner observed values: 34% AM+CIM imputed values: 36% (27%,45%) BASE imputed values: 44% (33%,57%) Individuals whose pay decreases substantially between sweeps are more likely to be missing δ: (-0.76,-0.13) The evidence that gaining a partner is associated with lower pay has strengthened CC BASE 0.08 ( 0.15, 0.01) 0.17 ( 0.28, 0.07) parameter estimates for sing Bayesian Missing Data Course (Lecture 4) Strategy 23 / 41
122 The joint model (steps 1-3): schematic diagram analysis model parameters model of response missingness analysis model model of response missingness parameters information from additional sources may help with the estimation of these parameters response with missingness probability of missingness covariate imputation model parameters covariates with missingness fully observed covariates missingness indicator covariate imputation model Bayesian Missing Data Course (Lecture 4) Strategy 24 / 41
123 Step 4 - seek additional data 1: select AM using complete cases note plausible alternatives 6: ASSUMPTION SENSITIVITY report robustness YES 2: add CIM 3: add MoRM 4: seek additional data BASE MODEL 8: Are conclusions robust? 7: PARAMETER SENSITIVITY NO determine region of high plausibility 5: elicit expert knowledge Constructing a base model assess fit Additional data can help with parameter estimation Possible sources include earlier/later sweeps of longitudinal study not under investigation another study on individuals with similar characteristics recognise uncertainty Bayesian Missing Data Course (Lecture 4) Strategy 25 / 41
124 Example: step 4 - incorporating additional data Seek another study with individuals with similar characteristics which includes variables of interest Expand CIM to simultaneously model data from original study and additional study by fitting 2 sets of equations with common coefficients 1 set for imputing the missing covariates in the original study 1 set for modelling the data from the additional study The extra data allows the parameters in the CIM to be estimated with greater accuracy Bayesian Missing Data Course (Lecture 4) Strategy 26 / 41
125 Step 5 - elicit expert knowledge note plausible alternatives 6: ASSUMPTION SENSITIVITY report robustness YES 1: select AM using complete cases 2: add CIM 3: add MoRM 4: seek additional data BASE MODEL 8: Are conclusions robust? 5: elicit expert knowledge 7: PARAMETER SENSITIVITY NO determine region of high plausibility recognise uncertainty Constructing a base model assess fit Expert knowledge can be elicited and incorporated using informative priors Focus on parameters not well identified by the data particularly those associated with the degree of departure from MAR Eliciting priors on parameters directly is difficult A better strategy is elicit information about the probability of response convert to informative priors Bayesian Missing Data Course (Lecture 4) Strategy 27 / 41
126 Example: step 5 - elicit expert knowledge The key variables in the MoRM are: the level of income the change in income Consult survey methodology experts about how these variables are likely to influence the probability of non-response In particular, review the assumption of linear relationships Individuals may be less inclined to disclose their income if the level is either low or high it has changed substantially in either direction Bayesian Missing Data Course (Lecture 4) Strategy 28 / 41
127 Part 2 (steps 6-8): sensitivity analysis Sensitivity analysis is essential because assumptions are untestable from the data There are many possible options, and the appropriate choice is problem dependent We propose two types of sensitivity analysis: 1 an assumption sensitivity 2 a parameter sensitivity Bayesian Missing Data Course (Lecture 4) Strategy 29 / 41
128 Step 6 - assumption sensitivity note plausible alternatives 1: select AM using complete cases 2: add CIM 3: add MoRM 4: seek additional data 6. ASSUMPTION SENSITIVITY report robustness YES BASE MODEL 8: Are conclusions robust? 7: PARAMETER SENSITIVITY NO determine region of high plausibility recognise uncertainty 5: elicit expert knowledge assess fit Assumption sensitivity forms alternative models by changing the assumptions in the different base sub-models Key assumptions include: AM error distribution transformation of the AM response functional form of the MoRM Could also vary explanatory variables Stage 1: change single aspect to assess effect Stage 2: combine several changes Bayesian Missing Data Course (Lecture 4) Strategy 30 / 41
129 Illustrative example: step 6 - assumption sensitivity BASE CASE key features: AM: t4 error distribution AM: covariates {age, reg, sing} AM: log transform of the response MoRM: linear functional form for level and change Examples of assumption sensitivity analysis options (differences from BASE CASE): AM: normal error distribution AM: additional covariate age 2 AM: cube root transform of response MoRM: piecewise linear functional form for level and change in hourly pay Bayesian Missing Data Course (Lecture 4) Strategy 31 / 41
130 Example: results from assumption sensitivity CC 0.08 ( 0.15, 0.01) BASE 0.17 ( 0.28, 0.07) AS ( 0.37, 0.15) AS ( 0.28, 0.07) AS ( 0.22, 0.02) parameter estimates for sing AS1: normal error distribution; AS2: additional covariate; AS3: piecewise linear functional form for MoRM The amount of evidence that gaining a partner is associated with lower pay is very sensitive to different assumptions Bayesian Missing Data Course (Lecture 4) Strategy 32 / 41
131 Step 7 - parameter sensitivity 1: select AM using complete cases note plausible alternatives 6: ASSUMPTION SENSITIVITY report robustness YES 2: add CIM 3: add MoRM 4: seek additional data BASE MODEL 8: Are conclusions robust? NO determine region of high plausibility 5: elicit expert knowledge 7: PARAMETER SENSITIVITY assess fit Parameter sensitivity involves running the base model with the MoRM parameters controlling the extent of the departure from MAR fixed to values in a plausible range Expert knowledge can help with setting the parameter sensitivity range recognise uncertainty Bayesian Missing Data Course (Lecture 4) Strategy 33 / 41
132 Illustrative example: step 7 - parameter sensitivity Recall MoRM equation for Base Case r logit(p i ) = θ 0 + θ k w ki + κ hpay i1 + δ (hpay i2 hpay i1 ) k=1 The value of δ controls the degree of departure from MAR missingness δ is difficult to estimate for a model with vague prior Run a series of models with δ fixed using point prior δ corresponds to the log odds ratio of a missing response per 1 increase in hourly pay values outside the -1 to 1 range are intuitively implausible as the probability of response changes from 1 to 0 abruptly 9 variants: values { 1, 0.75, 0.5, 0.25, 0, 0.25, 0.5, 0.75, 1} δ = 0 variant is equivalent to assuming the response is MAR Bayesian Missing Data Course (Lecture 4) Strategy 34 / 41
133 Example: results from parameter sensitivity BASE δ = 1 δ = 0.75 δ = 0.5 δ = 0.25 δ = 0 δ = 0.25 δ = 0.5 δ = 0.75 δ = parameter estimates for sing Even the direction of the effect is uncertain Bayesian Missing Data Course (Lecture 4) Strategy 35 / 41
134 Illustrative example: parameter sensitivity continued Suppose a piecewise linear functional form (AS3) was considered more plausible, e.g. {... + δ1 change logit(p i ) = i + : change i < δ 2 change i + : change i 0 Run a parameter sensitivity with both δ 1 and δ 2 fixed Graph shows proportional change in pay associated with gaining a partner Results very sensitive to chosen values of δ 1 and δ 2 red circle: MAR (δ 1 = δ 2 = 0) blue triangle: Base Case green diamond: AS3 δ e β sing δ Bayesian Missing Data Course (Lecture 4) Strategy 36 /
135 Step 8 - determine robustness of conclusions 1: select AM using complete cases note plausible alternatives 6: ASSUMPTION SENSITIVITY report robustness YES 2: add CIM 3: add MoRM 4: seek additional data BASE MODEL 8: Are conclusions robust? 7: PARAMETER SENSITIVITY NO determine region of high plausibility recognise uncertainty 5: elicit expert knowledge assess fit Examine results of sensitivity analyses to establish how much the quantities of interest vary If the conclusions are robust, report this Otherwise seek more information (Steps 4 & 5) determine a region of high plausibility recognise uncertainty Bayesian Missing Data Course (Lecture 4) Strategy 37 / 41
136 Assessing model fit A model s fit to observed data can be assessed However, it s fit to unobserved data cannot be assessed sensitivity analysis is essential DIC is routinely used by Bayesian statisticians to compare models, but using DIC in the presence of missing data is not straightforward the DIC automatically generated by WinBUGS is misleading (Mason et al., 2012a) Data not used in model estimation may be helpful in assessing model fit Bayesian Missing Data Course (Lecture 4) Strategy 38 / 41
137 Illustrative example: assessing model fit Data collected from 7 individuals who were originally non-contacts or refusals in sweep 2, after they were re-issued by the fieldwork agency Set these data to missing before fitting models, and use for model assessment Calculate the mean square error (MSE) of the fit of hourly pay for the 7 re-issues as a summary measure of model performance red circle: MAR (δ 1 = δ 2 = 0) blue triangle: Base Case green diamond: AS3 δ MSE of 7 re issues δ Bayesian Missing Data Course (Lecture 4) Strategy 39 / 41
138 Illustrative example: reporting uncertainty The MSE plot helps identify a region of high plausibility Key points to report are: There is evidence that gaining a partner is associated with a decrease in hourly pay However, the magnitude of this decrease is uncertain Our analysis suggests that the proportional decrease lies in the region 0.76 (0.69,0.82) to 0.93 (0.87,1.00) δ 2 MSE of 7 re issues δ Some models run as part of the sensitivity analysis suggest that change in partnership status is associated with an increase in pay, but these models do not fall in the region of high plausibility Bayesian Missing Data Course (Lecture 4) Strategy 40 /
139 Adaptions and extensions There are situations where it may be necessary to adapt this strategy Step 2 can be elaborated to allow MNAR covariates Steps 3 and 7 can be omitted if informative missingness in the response is implausible Could distinguish between different types of non-response set up a missingness indicator with separate categories for each type of non-response model using multinomial regression Bayesian models have the advantage of being fully coherent, but with large datasets or large numbers of covariates with missingness may be computationally challenging to fit In the next lecture, we consider alternative two stage approaches Bayesian Missing Data Course (Lecture 4) Strategy 41 / 41
140 Lecture 5. Comparison with alternative methods Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 1 / 37
141 Lecture Outline In Lectures 1-4 we discussed how to construct a Fully Bayesian Model (FBM) in the presence of missing data However, there are circumstances where implementation is difficult in practice In this lecture, we compare FBM with alternative methods review standard multiple imputation compare performance in practice between the ends of the multiple imputation spectrum revisit the MCAR/MAR/MNAR paradigm discuss future directions of missing data research Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 2 / 37
142 Standard Multiple Imputation (MI) FBM is one of a number of statistically principled methods for dealing with missing data (Lecture 1) Of the alternatives, standard Multiple Imputation is closest in spirit and has a Bayesian justification Multiple imputation was developed by Rubin (1996) Most widely used principled method for handling missing data Usually assumes missingness mechanism is MAR (can be used for MNAR but more tricky) Most useful for handling missing covariates Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 3 / 37
143 The 3 steps of Multiple Imputation 1 Imputation 2 Analysis 3 Pooling Impute (=fill in) the missing entries of the incomplete data sets, not once, but K (typically 5-10) times Imputed values are drawn from a distribution (that can be different for each missing variable) This step results in K complete data sets Analyse each of the K completed data sets using standard methods Combine the K analysis results into a single final result Simple rules exist for combining the K analyses to produce estimates and confidence intervals that incorporate uncertainty about the missing data Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 4 / 37
144 Step 1: Specifying the imputation model This is the hardest step of the MI procedure Need to specify model for full data f (z obs, z mis β) and then generate values for z mis z obs, β If only one variable (say z J z) has missing values, this usually amounts to specifying a regression model with z J as the outcome and z \J as the predictors Under MAR, can learn about relationship between zj and z \J from the cases where z J is observed (i.e. zj obs ), and then predict the missing cases, zj mis, from the fitted regression model More tricky when many variables have missing values - there are two main approaches based on a joint multivariate normal distribution iterate a series of conditional univariate models Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 5 / 37
145 Step 1: Multivariate regression v chained equations Multivariate regression techniques Theoretically sound Generate multiple imputations assuming an underlying multivariate normal distribution Parallels with specifying the CIM in FBM, but must include the response Multivariate imputation by chained equations (MICE) Specifies separate conditional distributions for the missing data in each incomplete variable Does not necessarily correspond to a genuine joint distribution Issue of incompatible conditionals But does provide a flexible way of implementing complex imputation models Both approaches risk a mismatch between the imputation model and the analysis model (often referred to as the problem of congeniality ) Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 6 / 37
146 Step 1: Selecting predictors for imputation model It is generally recommended to include variables that: appear in the main analysis model, including the response are associated with the missingness mechanism explain a considerable amount of variance in the variables to be imputed Better to err on the side of caution and include predictors rather than leave them out Imputation model typically contains richer set of variables/predictors than are of interest in main model should reflect any complexity in the analysis model (e.g. non-linear terms, interactions, hierarchical structure) Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 7 / 37
147 Comparison of FBM and MICE FBM MICE Imputation Model Analysis Model Imputation Model Analysis Model 1 stage procedure fit Imputation and Analysis Models simultaneously imputation model uses joint distribution of all missing variables uses full posterior distribution of missing values 2 stage procedure 1 fit Imputation Model 2 fit Analysis Model imputation model based on a set of univariate conditional distributions uses small number of draws of missing values from their predictive distribution Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 8 / 37
148 Multiple Imputation (MI) spectrum FBM and MICE can be considered the extremes of a MI spectrum feedback from analysis model to imputation model cut averages over small number of draws series of univariate conditional distributions Fully Bayesian Model (FBM) Bayesian MI (feedforward only model) Standard MI with joint multivariate distribution Standard MI with chained equations (MICE) Increasing approximation to FBM Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 9 / 37
149 Pros and cons of FBM and MICE FBM - pros theoretically sound coherent model estimation uncertainty fully propagated can add missingness model to explore informative missingness FBM - cons implementation can be challenging MICE - pros range of readily available packages speed MICE - cons conditional distributions may not correspond to a joint distribution difficult to explore informative missingness But how do FBM and MICE actually perform in practice? We investigate using simulations with missing covariates Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 10 / 37
150 General setup of simulations Generate 1000 simulated data sets with 2 correlated explanatory variables, x and u response, y, dependent on x and u 50% or 90% missingness imposed on u dependent on y Each simulated dataset analysed by a series of models Performance of models assessed for coefficient for u, βu, (true value=-2) coefficient for x, βx, (true value=1) We report average estimate (average of the posterior means) bias (average estimate - true value) coverage rate (proportion of times true value is contained in the 95% interval) interval width (average width of 95% interval) Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 11 / 37
151 Simulation setup: model descriptions We run 5 types of models GOLD: correct analysis model run on complete datasets EXU: excludes u from analysis model CC: complete case analysis FBM: Fully Bayesian Model (analysis and imputation models) MICE: uses 20 imputations GOLD provides performance targets GOLD, EXU, CC, FBM all fitted using WinBUGS software MICE fitted using functions from mice package in R software FBM and MICE have several variants dependent on scenario All models have the correct analysis model Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 12 / 37
152 Non-hierarchical linear simulation Data generated all variables continuous no hierarchical structure 1000 individuals 90% missingness Note: with single covariate with missing values, the approximation of using chained equations disappears Full data generated as follows ( ) (( x 0 MVN u 0 y N(1 + x 2u, 4 2 ) ) ( 1 0.5, MAR missingness imposed on u i with probability p i logit(p i ) = 7 + y i )) Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 13 / 37
153 Non-hierarchical linear simulation - results average coverage interval bias estimate rate width β x GOLD β x EXU β x CC β x FBM β x MICE β u GOLD β u CC β u FBM β u MICE EXU: extreme bias and 0 coverage (β x only) CC: serious bias and very low coverage FBM & MICE: correct most of bias and achieve nominal coverage Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 14 / 37
154 Non-hierarchical linear simulation - conclusion Findings extend to multiple covariates with missingness Even with extreme levels of missingness, FBM and MICE have similar performance with non-complex data We now look at two types of complexity hierarchical structure informative missingness Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 15 / 37
155 Hierarchical linear simulation - description Data generated with hierarchical structure (individuals within clusters) 10 clusters, each with 100 individuals 50% missingness FBM models: 3 variants with different imputation models for u no hierarchical structure (no HS) random intercepts (HS: ri) random intercepts + random slopes on x (HS: ri+rs) MICE: no hierarchical structure in imputation model in theory could run variants with hierarchical structure but implementation difficulties Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 16 / 37
156 Hierarchical linear simulation - equations Generate full data set as follows: x c 0 u c MVN 0 α c 1 ( ) (( xi xc MVN u i u c, ), y i N(α c + x i 2u i, 1) ( )) c indicates cluster level data; i indicates individual level data Impose missingness such that u i is missing with probability p i logit(p i ) = y i Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 17 / 37
157 Hierarchical linear simulation - β u results average coverage interval bias estimate rate width GOLD CC FBM (no HS) FBM (HS: ri) FBM (HS: ri+rs) MICE (no HS) If hierarchical structure ignored in imputation model FBM - slight bias and poor coverage MICE - much worse (no feedback from structure in analysis model) Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 18 / 37
158 Hierarchical linear simulation - β u results average coverage interval bias estimate rate width GOLD CC FBM (no HS) FBM (HS: ri) FBM (HS: ri+rs) MICE (no HS) If hierarchical structure incorporated in imputation model bias corrected nominal coverage rate achieved Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 19 / 37
159 Hierarchical linear simulation - β x results average coverage interval bias estimate rate width GOLD EXU CC FBM (no HS) FBM (HS: ri) FBM (HS: ri+rs) MICE (no HS) Pattern of bias and coverage results similar to β u Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 20 / 37
160 V-shaped informative missingness - description Data generated with no hierarchical structure 100 individuals missingness imposed on u depends on y and u 50% missingness FBM models: 4 variants MAR: no model of covariate missingness MNAR: assumes linear shape (linear) MNAR: allows v-shape (v-shape) MNAR: allows v-shape + priors inform signs of slopes (v-shape+) MICE: MAR, no model of covariate missingness most implementations do not readily extend to MNAR ad hoc sensitivity analysis to MNAR possible by inflating or deflating imputations (van Buuren and Groothuis-Oudshoorn, 2011) Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 21 / 37
161 V-shaped informative missingness - equations Generate full data set as follows: ( ) (( x 0 MVN u 0 y N(1 + x 2u, 4 2 ) ) ( 1 0.5, )) Impose missingness such that u is missing with probability p logit(p) = u + 0.5y Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 22 / 37
162 V-shaped informative missingness - β u results average coverage interval bias estimate rate width GOLD CC MAR: FBM MNAR: FBM (linear) MNAR: FBM (vshape) MNAR: FBM (vshape+) MAR: MICE MAR results in bias and slightly reduced coverage improvements if allow MNAR, even if wrong form further improvements from correct form and even better with informative priors Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 23 / 37
163 V-shaped informative missingness - β x results average coverage interval bias estimate rate width GOLD EXU CC MAR: FBM MNAR: FBM (linear) MNAR: FBM (vshape) MNAR: FBM (vshape+) MAR: MICE MAR results in modest bias (FBM and MICE) wrong MNAR (linear) slightly worse than MAR little gain in correct MNAR over MAR Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 24 / 37
164 Summary of simulation results hierarchical structure βu and β x - clear benefits from incorporating structure in FBM imputation model (unable to assess MICE) informative missingness βu - benefits from correct MNAR β x - no clear benefits from allowing MNAR FBM CIM AM MoCM AM = Analysis Model CIM = Covariate Imputation Model MoCM = Model of Covariate Missingness With FBM for informative missingness, 3 linked models fitted simultaneously Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 25 / 37
165 LBW example: low birth weight The LBW example was introduced in Lecture 3 Objective: estimate the association between trihalomethane concentrations (THM) and the risk of full term low birth weight Key variables are: Y : binary indicator of low birth weight (outcome) X: binary indicator of THM concentrations (exposure of interest) U: smoking and ethnicity (confounders, over 90% of values missing) We compare estimates of odds ratios using models from different points on the MI spectrum FBM: WinBUGS Bayesian MI: WinBUGS using cut function Standard MI: REALCOM-IMPUTE/MLwiN MICE: mice package in R Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 26 / 37
166 LBW example: comparison of imputation strategies CC (MCS only) FBM Bayesian MI Standard MI MICE odds ratio for smoking during pregnancy (u) EXU (HES only) CC (MCS only) FBM Bayesian MI Standard MI MICE odds ratio for Trihalomethanes > 60µg/L (x) Standard MI and MICE: point estimate and 95% confidence interval Other: posterior distribution (tick marks indicate 2.5% quantile, mean & 97.5% quantile) Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 27 / 37
167 Software for MI Many programs now include routines for carrying out MI using a set of univariate conditional distributions (MICE) R package mice (Stef van Buuren and Karin Groothuis-Oudshoorn) R package mi (Masanao Yajima, Yu-Sung Su, M. Grazia Pittau Stata SAS SPSS and Andrew Gelman) Implementations based on a joint multivariate distribution are less widely available REALCOM-IMPUTE: REALCOM macros in conjunction with MLwiN Bayesian models can be implemented using BUGS WinBUGS, OpenBUGS, JAGS Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 28 / 37
168 Practical advice simple complex small dataset and few covariates with missingness FBM MICE FBM large dataset and/or many covariates with missingness MICE Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 29 / 37
169 Hybrid approaches Neither FBM or MICE are currently well suited to analysis of 1 a large dataset and/or many covariates with missingness and 2 complexity e.g. hierarchical structure or informative missingness In these circumstances a hybrid approach may provide a pragmatic alternative We need to choose a convenient point on the MI spectrum Starting from FBM, consider the following approximations impute (multiple times) some covariates prior to fitting a Bayesian model split CIM into several smaller sub-models, based on level of correlation Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 30 / 37
170 Direction for missing data research feedback from analysis model to imputation model cut averages over small number of draws series of univariate conditional distributions Fully Bayesian Model (FBM) Bayesian MI (feedforward only model) Standard MI with joint multivariate distribution Standard MI with chained equations (MICE) Increasing approximation to FBM Where should our starting point be? FBM: improve computational efficiency, more case studies MICE: robust in simple setup, but more work needed for realistic situations Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 31 / 37
171 Concluding Remarks Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 32 / 37
172 MCAR/MAR/MNAR paradigm It is tempting to think that the missingness affecting a variable always slots neatly into one of Rubin s 3 categories From a practical viewpoint the reality is more complex MNAR can take different forms and depart from MAR by different amounts it depends on what else has been observed! If enough additional information is collected, MNAR missingness can be converted (close) to MAR missingness often used as justification for methods which assume MAR However, the MAR or MNAR classification will still be an assumption sensitivity analysis is required REMEMBER: It is not possible to distinguish between MAR and MNAR from the observed data alone Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 33 / 37
173 MCAR/MAR/MNAR paradigm and bias in CC (1) In Lecture 3 we saw that the conditions under which CC is biased with missing covariates cut across the MCAR/MAR/MNAR categorisation For the case of a single covariate, X, with missing values and fully observed Y CC is unbiased if missingness in X is MCAR MNAR dependent on X CC is biased if missingness in X is MAR dependent on Y MNAR dependent on X and Y Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 34 / 37
174 MCAR/MAR/MNAR paradigm and bias in CC (2) For missing responses, distinguish between cross-sectional and longitudinal data Cross-sectional data only MNAR causes bias Longitudinal data if MAR missingness depends on data at previous time-points excluding partially observed records for CC will convert to MNAR results in bias see difference in analysis of complete cases and all cases for HAMD example (Lecture 2) Missing data requires a lot of thought! Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 35 / 37
175 What does the Bayesian approach offer for missing data problems? Bayesian methods are probably the most powerful and most general methods for dealing with missing data Naturally accommodate missing data without requiring new techniques for inference Bayesian framework is well suited to building complex models by linking smaller sub-models into a coherent joint model for the full data Bayesian approach lends itself naturally to sensitivity analysis through different choices of prior distributions encoding assumptions about the missing data process Offers possibility of including informative prior information about missing data process But models can become computationally challenging... Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 36 / 37
176 Thank you for your attention! Bayesian Missing Data Course (Lecture 5) Comparison with alternative methods 37 / 37
177 Bayesian Missing Data Course References Chib, S. and Greenberg, E. (1998). Biometrika, 85, (2), Analysis of multivariate probit models. Daniels, M. J. and Hogan, J. W. (2008). Missing Data In Longitudinal Studies: Strategies for Bayesian Modeling and Sensitivity Analysis. Chapman & Hall. Diggle, P. and Kenward, M. G. (1994). Informative Drop-out in Longitudinal Data Analysis (with discussion). Journal of the Royal Statistical Society, Series C (Applied Statistics), 43, (1), Little, R. J. A. and Rubin, D. B. (2002). Statistical Analysis with Missing Data, (2nd edn). John Wiley and Sons. Mason, A., Richardson, S., and Best, N. (2012a). Two-pronged strategy for using DIC to compare selection models with non-ignorable missing responses. Bayesian Analysis. to appear. Mason, A., Richardson, S., Plewis, I., and Best, N. (2012b). Strategy for modelling non-random missing data mechanisms in observational studies using Bayesian methods. Journal of Official Statistics. to appear. Mason, A. J. (2009). Bayesian methods for modelling non-random missing data mechanisms in longitudinal studies. PhD thesis, Imperial College London. available at Molenberghs, G. and Kenward, M. G. (2007). Missing Data in Clinical Studies, (1st edn). John Wiley and Sons. Molitor, N.-T., Best, N., Jackson, C., and Richardson, S. (2009). Using Bayesian graphical models to model biases in observational studies and to combine multiple data sources: Application to low birth-weight and water disinfection by-products. Journal of the Royal Statistical Society, Series A, 172, (3), Richardson, S. and Best, N. (2003). Bayesian hierarchical models in ecological studies of health-environment effects. Environmetrics, 14, Rubin, D. B. (1996). Multiple Imputation After 18+ Years. Journal of the American Statistical Society, 91, (434), Spiegelhalter, D. J. (1998). Bayesian Graphical Modelling: A Case-Study in Monitoing Health Outcomes. Journal of the Royal Statistical Society, Series C (Applied Statistics), 47, (1), Spiegelhalter, D. J., Thomas, A., and Best, N. G. (1996). Computation on Bayesian Graphical Models. In Bayesian Statistics 5: Proceedings of the 1
178 Fifth Valencia International Meeting, (ed. J. M. Bernardo, J. O. Berger, A. P. Dawid, and A. F. M. Smith), pp Claredon Press. Sterne, J. A. C., Carlin, J. B., Spratt, M., Kenward, M. G., Wood, A. M., and Carpenter, J. R. (2009). Multiple imputation for missing data in epidemiological and clinical research: potential and pitfalls. British Medical Journal, 338, b2393. White, I. R. and Carlin, J. B. (2010). Bias and efficiency of multiple imputation compared with complete-case analysis for missing covariate values. Statistics in Medicine, 29, Wood, A. M., White, I. R., and Thompson, S. G. (2004). Are missing outcome data adequately handled? A review of published randomized controlled trials in major medical journals. Clinical Trials, 1, (4),
Problem of Missing Data
VASA Mission of VA Statisticians Association (VASA) Promote & disseminate statistical methodological research relevant to VA studies; Facilitate communication & collaboration among VA-affiliated statisticians;
A Basic Introduction to Missing Data
John Fox Sociology 740 Winter 2014 Outline Why Missing Data Arise Why Missing Data Arise Global or unit non-response. In a survey, certain respondents may be unreachable or may refuse to participate. Item
Handling missing data in Stata a whirlwind tour
Handling missing data in Stata a whirlwind tour 2012 Italian Stata Users Group Meeting Jonathan Bartlett www.missingdata.org.uk 20th September 2012 1/55 Outline The problem of missing data and a principled
Statistical modelling with missing data using multiple imputation. Session 4: Sensitivity Analysis after Multiple Imputation
Statistical modelling with missing data using multiple imputation Session 4: Sensitivity Analysis after Multiple Imputation James Carpenter London School of Hygiene & Tropical Medicine Email: [email protected]
Handling attrition and non-response in longitudinal data
Longitudinal and Life Course Studies 2009 Volume 1 Issue 1 Pp 63-72 Handling attrition and non-response in longitudinal data Harvey Goldstein University of Bristol Correspondence. Professor H. Goldstein
Imputation of missing data under missing not at random assumption & sensitivity analysis
Imputation of missing data under missing not at random assumption & sensitivity analysis S. Jolani Department of Methodology and Statistics, Utrecht University, the Netherlands Advanced Multiple Imputation,
Review of the Methods for Handling Missing Data in. Longitudinal Data Analysis
Int. Journal of Math. Analysis, Vol. 5, 2011, no. 1, 1-13 Review of the Methods for Handling Missing Data in Longitudinal Data Analysis Michikazu Nakai and Weiming Ke Department of Mathematics and Statistics
MISSING DATA TECHNIQUES WITH SAS. IDRE Statistical Consulting Group
MISSING DATA TECHNIQUES WITH SAS IDRE Statistical Consulting Group ROAD MAP FOR TODAY To discuss: 1. Commonly used techniques for handling missing data, focusing on multiple imputation 2. Issues that could
Missing Data. A Typology Of Missing Data. Missing At Random Or Not Missing At Random
[Leeuw, Edith D. de, and Joop Hox. (2008). Missing Data. Encyclopedia of Survey Research Methods. Retrieved from http://sage-ereference.com/survey/article_n298.html] Missing Data An important indicator
Challenges in Longitudinal Data Analysis: Baseline Adjustment, Missing Data, and Drop-out
Challenges in Longitudinal Data Analysis: Baseline Adjustment, Missing Data, and Drop-out Sandra Taylor, Ph.D. IDDRC BBRD Core 23 April 2014 Objectives Baseline Adjustment Introduce approaches Guidance
Introduction to mixed model and missing data issues in longitudinal studies
Introduction to mixed model and missing data issues in longitudinal studies Hélène Jacqmin-Gadda INSERM, U897, Bordeaux, France Inserm workshop, St Raphael Outline of the talk I Introduction Mixed models
PATTERN MIXTURE MODELS FOR MISSING DATA. Mike Kenward. London School of Hygiene and Tropical Medicine. Talk at the University of Turku,
PATTERN MIXTURE MODELS FOR MISSING DATA Mike Kenward London School of Hygiene and Tropical Medicine Talk at the University of Turku, April 10th 2012 1 / 90 CONTENTS 1 Examples 2 Modelling Incomplete Data
Bayesian Statistics in One Hour. Patrick Lam
Bayesian Statistics in One Hour Patrick Lam Outline Introduction Bayesian Models Applications Missing Data Hierarchical Models Outline Introduction Bayesian Models Applications Missing Data Hierarchical
Missing Data: Part 1 What to Do? Carol B. Thompson Johns Hopkins Biostatistics Center SON Brown Bag 3/20/13
Missing Data: Part 1 What to Do? Carol B. Thompson Johns Hopkins Biostatistics Center SON Brown Bag 3/20/13 Overview Missingness and impact on statistical analysis Missing data assumptions/mechanisms Conventional
Missing Data in Longitudinal Studies: To Impute or not to Impute? Robert Platt, PhD McGill University
Missing Data in Longitudinal Studies: To Impute or not to Impute? Robert Platt, PhD McGill University 1 Outline Missing data definitions Longitudinal data specific issues Methods Simple methods Multiple
Markov Chain Monte Carlo Simulation Made Simple
Markov Chain Monte Carlo Simulation Made Simple Alastair Smith Department of Politics New York University April2,2003 1 Markov Chain Monte Carlo (MCMC) simualtion is a powerful technique to perform numerical
Web-based Supplementary Materials for Bayesian Effect Estimation. Accounting for Adjustment Uncertainty by Chi Wang, Giovanni
1 Web-based Supplementary Materials for Bayesian Effect Estimation Accounting for Adjustment Uncertainty by Chi Wang, Giovanni Parmigiani, and Francesca Dominici In Web Appendix A, we provide detailed
STA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! [email protected]! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct
Model-based Synthesis. Tony O Hagan
Model-based Synthesis Tony O Hagan Stochastic models Synthesising evidence through a statistical model 2 Evidence Synthesis (Session 3), Helsinki, 28/10/11 Graphical modelling The kinds of models that
2. Making example missing-value datasets: MCAR, MAR, and MNAR
Lecture 20 1. Types of missing values 2. Making example missing-value datasets: MCAR, MAR, and MNAR 3. Common methods for missing data 4. Compare results on example MCAR, MAR, MNAR data 1 Missing Data
Basics of Statistical Machine Learning
CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu [email protected] Modern machine learning is rooted in statistics. You will find many familiar
Spatial Statistics Chapter 3 Basics of areal data and areal data modeling
Spatial Statistics Chapter 3 Basics of areal data and areal data modeling Recall areal data also known as lattice data are data Y (s), s D where D is a discrete index set. This usually corresponds to data
Statistics Graduate Courses
Statistics Graduate Courses STAT 7002--Topics in Statistics-Biological/Physical/Mathematics (cr.arr.).organized study of selected topics. Subjects and earnable credit may vary from semester to semester.
Electronic Theses and Dissertations UC Riverside
Electronic Theses and Dissertations UC Riverside Peer Reviewed Title: Bayesian and Non-parametric Approaches to Missing Data Analysis Author: Yu, Yao Acceptance Date: 01 Series: UC Riverside Electronic
Analyzing Structural Equation Models With Missing Data
Analyzing Structural Equation Models With Missing Data Craig Enders* Arizona State University [email protected] based on Enders, C. K. (006). Analyzing structural equation models with missing data. In G.
Dealing with Missing Data
Res. Lett. Inf. Math. Sci. (2002) 3, 153-160 Available online at http://www.massey.ac.nz/~wwiims/research/letters/ Dealing with Missing Data Judi Scheffer I.I.M.S. Quad A, Massey University, P.O. Box 102904
How to choose an analysis to handle missing data in longitudinal observational studies
How to choose an analysis to handle missing data in longitudinal observational studies ICH, 25 th February 2015 Ian White MRC Biostatistics Unit, Cambridge, UK Plan Why are missing data a problem? Methods:
Imputing Missing Data using SAS
ABSTRACT Paper 3295-2015 Imputing Missing Data using SAS Christopher Yim, California Polytechnic State University, San Luis Obispo Missing data is an unfortunate reality of statistics. However, there are
A Latent Variable Approach to Validate Credit Rating Systems using R
A Latent Variable Approach to Validate Credit Rating Systems using R Chicago, April 24, 2009 Bettina Grün a, Paul Hofmarcher a, Kurt Hornik a, Christoph Leitner a, Stefan Pichler a a WU Wien Grün/Hofmarcher/Hornik/Leitner/Pichler
Dealing with Missing Data
Dealing with Missing Data Roch Giorgi email: [email protected] UMR 912 SESSTIM, Aix Marseille Université / INSERM / IRD, Marseille, France BioSTIC, APHM, Hôpital Timone, Marseille, France January
A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution
A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 4: September
Multiple Imputation for Missing Data: A Cautionary Tale
Multiple Imputation for Missing Data: A Cautionary Tale Paul D. Allison University of Pennsylvania Address correspondence to Paul D. Allison, Sociology Department, University of Pennsylvania, 3718 Locust
APPLIED MISSING DATA ANALYSIS
APPLIED MISSING DATA ANALYSIS Craig K. Enders Series Editor's Note by Todd D. little THE GUILFORD PRESS New York London Contents 1 An Introduction to Missing Data 1 1.1 Introduction 1 1.2 Chapter Overview
1 Prior Probability and Posterior Probability
Math 541: Statistical Theory II Bayesian Approach to Parameter Estimation Lecturer: Songfeng Zheng 1 Prior Probability and Posterior Probability Consider now a problem of statistical inference in which
Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus
Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus Tihomir Asparouhov and Bengt Muthén Mplus Web Notes: No. 15 Version 8, August 5, 2014 1 Abstract This paper discusses alternatives
PS 271B: Quantitative Methods II. Lecture Notes
PS 271B: Quantitative Methods II Lecture Notes Langche Zeng [email protected] The Empirical Research Process; Fundamental Methodological Issues 2 Theory; Data; Models/model selection; Estimation; Inference.
BayesX - Software for Bayesian Inference in Structured Additive Regression
BayesX - Software for Bayesian Inference in Structured Additive Regression Thomas Kneib Faculty of Mathematics and Economics, University of Ulm Department of Statistics, Ludwig-Maximilians-University Munich
Econometrics Simple Linear Regression
Econometrics Simple Linear Regression Burcu Eke UC3M Linear equations with one variable Recall what a linear equation is: y = b 0 + b 1 x is a linear equation with one variable, or equivalently, a straight
Lecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
Dr James Roger. GlaxoSmithKline & London School of Hygiene and Tropical Medicine.
American Statistical Association Biopharm Section Monthly Webinar Series: Sensitivity analyses that address missing data issues in Longitudinal studies for regulatory submission. Dr James Roger. GlaxoSmithKline
CHAPTER 3 EXAMPLES: REGRESSION AND PATH ANALYSIS
Examples: Regression And Path Analysis CHAPTER 3 EXAMPLES: REGRESSION AND PATH ANALYSIS Regression analysis with univariate or multivariate dependent variables is a standard procedure for modeling relationships
Towards running complex models on big data
Towards running complex models on big data Working with all the genomes in the world without changing the model (too much) Daniel Lawson Heilbronn Institute, University of Bristol 2013 1 / 17 Motivation
Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
SPSS TRAINING SESSION 3 ADVANCED TOPICS (PASW STATISTICS 17.0) Sun Li Centre for Academic Computing [email protected]
SPSS TRAINING SESSION 3 ADVANCED TOPICS (PASW STATISTICS 17.0) Sun Li Centre for Academic Computing [email protected] IN SPSS SESSION 2, WE HAVE LEARNT: Elementary Data Analysis Group Comparison & One-way
Reject Inference in Credit Scoring. Jie-Men Mok
Reject Inference in Credit Scoring Jie-Men Mok BMI paper January 2009 ii Preface In the Master programme of Business Mathematics and Informatics (BMI), it is required to perform research on a business
Missing Data Sensitivity Analysis of a Continuous Endpoint An Example from a Recent Submission
Missing Data Sensitivity Analysis of a Continuous Endpoint An Example from a Recent Submission Arno Fritsch Clinical Statistics Europe, Bayer November 21, 2014 ASA NJ Chapter / Bayer Workshop, Whippany
CHAPTER 2 Estimating Probabilities
CHAPTER 2 Estimating Probabilities Machine Learning Copyright c 2016. Tom M. Mitchell. All rights reserved. *DRAFT OF January 24, 2016* *PLEASE DO NOT DISTRIBUTE WITHOUT AUTHOR S PERMISSION* This is a
A REVIEW OF CURRENT SOFTWARE FOR HANDLING MISSING DATA
123 Kwantitatieve Methoden (1999), 62, 123-138. A REVIEW OF CURRENT SOFTWARE FOR HANDLING MISSING DATA Joop J. Hox 1 ABSTRACT. When we deal with a large data set with missing data, we have to undertake
Overview. Longitudinal Data Variation and Correlation Different Approaches. Linear Mixed Models Generalized Linear Mixed Models
Overview 1 Introduction Longitudinal Data Variation and Correlation Different Approaches 2 Mixed Models Linear Mixed Models Generalized Linear Mixed Models 3 Marginal Models Linear Models Generalized Linear
Fairfield Public Schools
Mathematics Fairfield Public Schools AP Statistics AP Statistics BOE Approved 04/08/2014 1 AP STATISTICS Critical Areas of Focus AP Statistics is a rigorous course that offers advanced students an opportunity
Handling missing data in large data sets. Agostino Di Ciaccio Dept. of Statistics University of Rome La Sapienza
Handling missing data in large data sets Agostino Di Ciaccio Dept. of Statistics University of Rome La Sapienza The problem Often in official statistics we have large data sets with many variables and
I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN
Beckman HLM Reading Group: Questions, Answers and Examples Carolyn J. Anderson Department of Educational Psychology I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN Linear Algebra Slide 1 of
Imputation and Analysis. Peter Fayers
Missing Data in Palliative Care Research Imputation and Analysis Peter Fayers Department of Public Health University of Aberdeen NTNU Det medisinske fakultet Missing data Missing data is a major problem
A Mixed Model Approach for Intent-to-Treat Analysis in Longitudinal Clinical Trials with Missing Values
Methods Report A Mixed Model Approach for Intent-to-Treat Analysis in Longitudinal Clinical Trials with Missing Values Hrishikesh Chakraborty and Hong Gu March 9 RTI Press About the Author Hrishikesh Chakraborty,
The Basics of Graphical Models
The Basics of Graphical Models David M. Blei Columbia University October 3, 2015 Introduction These notes follow Chapter 2 of An Introduction to Probabilistic Graphical Models by Michael Jordan. Many figures
Sensitivity Analysis in Multiple Imputation for Missing Data
Paper SAS270-2014 Sensitivity Analysis in Multiple Imputation for Missing Data Yang Yuan, SAS Institute Inc. ABSTRACT Multiple imputation, a popular strategy for dealing with missing values, usually assumes
Bayesian Hierarchical Models: Practical Exercises
1 Bayesian Hierarchical Models: Practical Exercises You will be using WinBUGS 1.4.3 for these practicals. All the data and other files you will need for the practicals are provided in a zip file, which
Credit Risk Models: An Overview
Credit Risk Models: An Overview Paul Embrechts, Rüdiger Frey, Alexander McNeil ETH Zürich c 2003 (Embrechts, Frey, McNeil) A. Multivariate Models for Portfolio Credit Risk 1. Modelling Dependent Defaults:
Analyzing Clinical Trial Data via the Bayesian Multiple Logistic Random Effects Model
Analyzing Clinical Trial Data via the Bayesian Multiple Logistic Random Effects Model Bartolucci, A.A 1, Singh, K.P 2 and Bae, S.J 2 1 Dept. of Biostatistics, University of Alabama at Birmingham, Birmingham,
A Bayesian hierarchical surrogate outcome model for multiple sclerosis
A Bayesian hierarchical surrogate outcome model for multiple sclerosis 3 rd Annual ASA New Jersey Chapter / Bayer Statistics Workshop David Ohlssen (Novartis), Luca Pozzi and Heinz Schmidli (Novartis)
A Bayesian Antidote Against Strategy Sprawl
A Bayesian Antidote Against Strategy Sprawl Benjamin Scheibehenne ([email protected]) University of Basel, Missionsstrasse 62a 4055 Basel, Switzerland & Jörg Rieskamp ([email protected])
Introduction to Fixed Effects Methods
Introduction to Fixed Effects Methods 1 1.1 The Promise of Fixed Effects for Nonexperimental Research... 1 1.2 The Paired-Comparisons t-test as a Fixed Effects Method... 2 1.3 Costs and Benefits of Fixed
Linear Classification. Volker Tresp Summer 2015
Linear Classification Volker Tresp Summer 2015 1 Classification Classification is the central task of pattern recognition Sensors supply information about an object: to which class do the object belong
Analyzing Intervention Effects: Multilevel & Other Approaches. Simplest Intervention Design. Better Design: Have Pretest
Analyzing Intervention Effects: Multilevel & Other Approaches Joop Hox Methodology & Statistics, Utrecht Simplest Intervention Design R X Y E Random assignment Experimental + Control group Analysis: t
ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015
ECON 459 Game Theory Lecture Notes Auctions Luca Anderlini Spring 2015 These notes have been used before. If you can still spot any errors or have any suggestions for improvement, please let me know. 1
Missing Data. Paul D. Allison INTRODUCTION
4 Missing Data Paul D. Allison INTRODUCTION Missing data are ubiquitous in psychological research. By missing data, I mean data that are missing for some (but not all) variables and for some (but not all)
Missing data and net survival analysis Bernard Rachet
Workshop on Flexible Models for Longitudinal and Survival Data with Applications in Biostatistics Warwick, 27-29 July 2015 Missing data and net survival analysis Bernard Rachet General context Population-based,
Multiple Choice Models II
Multiple Choice Models II Laura Magazzini University of Verona [email protected] http://dse.univr.it/magazzini Laura Magazzini (@univr.it) Multiple Choice Models II 1 / 28 Categorical data Categorical
A General Approach to Variance Estimation under Imputation for Missing Survey Data
A General Approach to Variance Estimation under Imputation for Missing Survey Data J.N.K. Rao Carleton University Ottawa, Canada 1 2 1 Joint work with J.K. Kim at Iowa State University. 2 Workshop on Survey
An Internal Model for Operational Risk Computation
An Internal Model for Operational Risk Computation Seminarios de Matemática Financiera Instituto MEFF-RiskLab, Madrid http://www.risklab-madrid.uam.es/ Nicolas Baud, Antoine Frachot & Thierry Roncalli
Bayesian Statistics: Indian Buffet Process
Bayesian Statistics: Indian Buffet Process Ilker Yildirim Department of Brain and Cognitive Sciences University of Rochester Rochester, NY 14627 August 2012 Reference: Most of the material in this note
Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression
Logistic Regression Department of Statistics The Pennsylvania State University Email: [email protected] Logistic Regression Preserve linear classification boundaries. By the Bayes rule: Ĝ(x) = arg max
Principle of Data Reduction
Chapter 6 Principle of Data Reduction 6.1 Introduction An experimenter uses the information in a sample X 1,..., X n to make inferences about an unknown parameter θ. If the sample size n is large, then
2. Linear regression with multiple regressors
2. Linear regression with multiple regressors Aim of this section: Introduction of the multiple regression model OLS estimation in multiple regression Measures-of-fit in multiple regression Assumptions
South Carolina College- and Career-Ready (SCCCR) Probability and Statistics
South Carolina College- and Career-Ready (SCCCR) Probability and Statistics South Carolina College- and Career-Ready Mathematical Process Standards The South Carolina College- and Career-Ready (SCCCR)
Service courses for graduate students in degree programs other than the MS or PhD programs in Biostatistics.
Course Catalog In order to be assured that all prerequisites are met, students must acquire a permission number from the education coordinator prior to enrolling in any Biostatistics course. Courses are
Multinomial and Ordinal Logistic Regression
Multinomial and Ordinal Logistic Regression ME104: Linear Regression Analysis Kenneth Benoit August 22, 2012 Regression with categorical dependent variables When the dependent variable is categorical,
Missing Data & How to Deal: An overview of missing data. Melissa Humphries Population Research Center
Missing Data & How to Deal: An overview of missing data Melissa Humphries Population Research Center Goals Discuss ways to evaluate and understand missing data Discuss common missing data methods Know
Using SAS PROC MCMC to Estimate and Evaluate Item Response Theory Models
Using SAS PROC MCMC to Estimate and Evaluate Item Response Theory Models Clement A Stone Abstract Interest in estimating item response theory (IRT) models using Bayesian methods has grown tremendously
More details on the inputs, functionality, and output can be found below.
Overview: The SMEEACT (Software for More Efficient, Ethical, and Affordable Clinical Trials) web interface (http://research.mdacc.tmc.edu/smeeactweb) implements a single analysis of a two-armed trial comparing
Centre for Central Banking Studies
Centre for Central Banking Studies Technical Handbook No. 4 Applied Bayesian econometrics for central bankers Andrew Blake and Haroon Mumtaz CCBS Technical Handbook No. 4 Applied Bayesian econometrics
Adequacy of Biomath. Models. Empirical Modeling Tools. Bayesian Modeling. Model Uncertainty / Selection
Directions in Statistical Methodology for Multivariable Predictive Modeling Frank E Harrell Jr University of Virginia Seattle WA 19May98 Overview of Modeling Process Model selection Regression shape Diagnostics
Validation of Software for Bayesian Models using Posterior Quantiles. Samantha R. Cook Andrew Gelman Donald B. Rubin DRAFT
Validation of Software for Bayesian Models using Posterior Quantiles Samantha R. Cook Andrew Gelman Donald B. Rubin DRAFT Abstract We present a simulation-based method designed to establish that software
Missing data in randomized controlled trials (RCTs) can
EVALUATION TECHNICAL ASSISTANCE BRIEF for OAH & ACYF Teenage Pregnancy Prevention Grantees May 2013 Brief 3 Coping with Missing Data in Randomized Controlled Trials Missing data in randomized controlled
Institute of Actuaries of India Subject CT3 Probability and Mathematical Statistics
Institute of Actuaries of India Subject CT3 Probability and Mathematical Statistics For 2015 Examinations Aim The aim of the Probability and Mathematical Statistics subject is to provide a grounding in
Standard Deviation Estimator
CSS.com Chapter 905 Standard Deviation Estimator Introduction Even though it is not of primary interest, an estimate of the standard deviation (SD) is needed when calculating the power or sample size of
Analysis of Longitudinal Data with Missing Values.
Analysis of Longitudinal Data with Missing Values. Methods and Applications in Medical Statistics. Ingrid Garli Dragset Master of Science in Physics and Mathematics Submission date: June 2009 Supervisor:
Comparison of frequentist and Bayesian inference. Class 20, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom
Comparison of frequentist and Bayesian inference. Class 20, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom 1 Learning Goals 1. Be able to explain the difference between the p-value and a posterior
Classification Problems
Classification Read Chapter 4 in the text by Bishop, except omit Sections 4.1.6, 4.1.7, 4.2.4, 4.3.3, 4.3.5, 4.3.6, 4.4, and 4.5. Also, review sections 1.5.1, 1.5.2, 1.5.3, and 1.5.4. Classification Problems
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written
Statistics 2014 Scoring Guidelines
AP Statistics 2014 Scoring Guidelines College Board, Advanced Placement Program, AP, AP Central, and the acorn logo are registered trademarks of the College Board. AP Central is the official online home
Note on the EM Algorithm in Linear Regression Model
International Mathematical Forum 4 2009 no. 38 1883-1889 Note on the M Algorithm in Linear Regression Model Ji-Xia Wang and Yu Miao College of Mathematics and Information Science Henan Normal University
5 Directed acyclic graphs
5 Directed acyclic graphs (5.1) Introduction In many statistical studies we have prior knowledge about a temporal or causal ordering of the variables. In this chapter we will use directed graphs to incorporate
Markov random fields and Gibbs measures
Chapter Markov random fields and Gibbs measures 1. Conditional independence Suppose X i is a random element of (X i, B i ), for i = 1, 2, 3, with all X i defined on the same probability space (.F, P).
