Estimating health state utility values from discrete choice experiments a

Size: px
Start display at page:

Download "Estimating health state utility values from discrete choice experiments a"

Transcription

1 Estimating health state utility values from discrete choice experiments a QALY space model approach Yuanyuan Gu, Richard Norman, Rosalie Viney Centre for Health Economics Research and Evaluation, University of Technology, Sydney, Australia Corresponding author: Yuanyuan Gu, Centre for Health Economics Research and Evaluation, University of Technology, Sydney, PO BOX 23, Broadway, NSW, 27, Australia. yuanyuan.gu@gmail.com Phone: Fax: Keywords: Average valuation, Bayesian, DCE, EQ-5D, Johnson s SB, QALY space

2 Abstract Using discrete choice experiments (DCEs) to estimate health state utility values has become an important alternative to the standard methods such as the Time Trade-Off (TTO). Studies using DCEs have typically used the conditional logit to estimate the underlying utility function. We show that this approach will lead to the valuation of each health state from an average person in the population. By contrast, the standard approach that has been developed for the TTO method is based on estimating the average valuation for a health state within the population. These are fundamentally conceptually different approaches and have different interpretations in policy evaluation. In this paper we point out that it is also possible to estimate the average valuation for a health state when using DCEs. The estimation approach is based on the mixed logit (MIXL). In particular, we propose two types of models, one using preference space and the other using QALY space, a concept adapted from the willingness-to-pay literature. These methods are applied to a data set collected using the EQ-5D. The results demonstrate that the preferred QALY space model provides lower estimates of the utility values than the conditional logit, with the divergence increasing with worsening health states. 2

3 Introduction For the evaluation of new health technologies, it is conventional to model their effect using the quality-adjusted life year (QALY). QALYs combine quality of life and life expectancy into a summary measure that reflects preferences for these two dimensions of health gain (Pliskin, et al., 98). The use of cost-utility analysis, with outcomes measured in terms of QALYs is now recommended by most health technology agencies internationally. A number of standard generic quality of life instruments have been developed for the purpose of measuring and valuing quality of life to facilitate estimation of QALYs directly from patient reported outcomes (Brazier, 27). These instruments, known as multi-attribute utility instruments describe the health state space in terms of several dimensions of quality of life, and include a preference based scoring algorithm that can be interpreted on a cardinal scale. Typically, standard preference based valuation techniques such as the Standard Gamble (SG) and Time Trade-off (TTO) have been used to derive the scoring algorithms to assign the scores (known as utility values or QALY weights) to the universe of health states described by the instrument. In the past decade, several authors have considered the use of discrete choice experiments (DCEs) to estimate health state utility values, as an alternative to TTO and SG based techniques (Bansback, et al., 22; Coast, et al., 28; Flynn, 2; Hakim and Pathak, 999; Lancsar, et al., 2; Ratcliffe, et al., 29; Ryan, et al., 26; Viney, et al., 23). In the approach developed by Bansback, et al. (22), and used by others, the health state utility values are estimated based on the conditional logit model. Broadly, in this approach, the conditional logit is used to estimate coefficients of the attributes that describe a health profile. Utility decrements associated with any move away from full health can be estimated for each dimension and level by computing the ratios between the estimated coefficients of the non- 3

4 time attributes and that of the time attribute. Utility values assigned to specific health states are then calculated by summing the relevant utility decrements and subtracting them from one. This approach has important conceptual differences from the approach that has been developed for the TTO and the SG. The standard approach that has been used in the QALY literature and in economic evaluation is based on finding the average valuation of a health state for the relevant population. Effectively this involves estimating the health state utility values for each individual in the population and then averaging these individual utility values over the whole population. In contrast, the approach using conditional logit is to find the valuation of a health status from an average person in the population. These are conceptually different approaches and therefore have potentially different interpretations in policy evaluation. In this paper we demonstrate that it is possible to estimate the average valuation for a health state when using DCEs. The estimation approach is based on the mixed logit (MIXL) which allows us to derive the population distributions of utility decrements and then the means of these distributions. In particular, we propose two types of models, one using preference space and the other using QALY space, a concept adapted from the willingnessto-pay (WTP) literature. The QALY space model has several advantages over the preference space model and the most significant one is that it allows us to directly estimate and compare different distribution assumptions for the utility decrements. A specific contribution is made to the estimation of a QALY space model with utility decrements assumed to follow a multivariate Johnson s SB distribution. In the choice modelling literature this type of model has been very difficult to estimate due to an identification problem (Rigby and Burton, 26; Train and Sonnier, 25). In this paper we show that using informative priors on the bounds The time attribute (also called the survival duration attribute) describes the life expectancy in a health state. 4

5 may improve the identification and estimating the bounds with other parameters simultaneously is possible. In this study, we develop methods to estimate utility values for EQ-5D health states although these methods could be applied to other instruments that are based on a linear additive model, such as the SF-6D. These methods are applied to a data set which has been previously used to estimate health state utility values. The utility values estimated from the selected MIXL model and the conditional logit are compared. 2 Valuing EQ-5D health states using DCEs The EQ-5D, developed by the EuroQol Group, is the most widely used multi-attribute utility instrument (Richardson, et al., 2; Szende, et al., 27). It has five dimensions, intended to represent the major areas in which health changes can manifest: mobility, self-care, usual activities, pain/discomfort and anxiety/depression. For the most commonly used version of the EQ-5D, each dimension contains three levels, loosely classified as No Problems, Some Problems, and Extreme Problems. Details are shown in Table. There are 3^5 = 243 potential states in the descriptive system. [Insert Table around here] The traditional approach to value these 243 states has been to administer a TTO preference based task for a sample of health states in a population based sample, and then use regression based modelling to impute the values of the remaining health states (Dolan, 997; Szende, et 5

6 al., 27; Viney, et al., 2). There is an extensive literature on this broad approach including a series of examinations on their limitations which might have led to the current trend of investigating alternative methods (Bosch, et al., 998; Craig, et al., 29; Norman, et al., 2). For example there have been explorations of alternative specifications of the TTO, including Lead-Time and Lag-Time TTOs (Devlin, et al., 2). A review of the development of using DCEs to value health states can be found in Bansback, et al. (22). 2. The DCE data Viney, et al. (23) have developed a DCE based algorithm for the Australian population, and the data from that study are used in the current analysis. This section briefly describes the experiment. The DCE was developed and administered to a sample of the Australian general population. Respondents were asked to choose between health profiles described in terms of EQ-5D profiles and survival attributes. Each choice set included three options: two health profile options and an immediate death option. Each health profile option in a choice set was defined by five attributes covering the dimensions of the EQ-5D and a survival duration attribute. Five survival durations (, 2, 4, 8 and 6 years) were included in the experiment. The third option of immediate death was included to allow for a complete ranking of health profiles over the worse than death to full health utility space. The task for the respondent was to identify which of the three options was considered the best, and which the worst, thus providing a complete ranking within each choice set. An example of a choice set is provided in Figure. [Insert Figure around here] 6

7 Details of the experimental design can be found in Viney, et al. (23). Although each choice set included an immediate death option, only the choice between the two non-death profiles was considered. 2 Therefore the analysis was based on a constructed choice set with only the rankings of these two profiles. A total of,2 individuals consented to participate in the survey and were eligible to participate. Of these, 3 completed they survey, giving a response rate of 92.%. Viney, et al. (23) showed that overall the characteristics of these who completed the task are broadly comparable to the characteristics of the general Australian population. Each respondent faced 5 choice sets, which translates into 5,465 observations. 3 Using conditional logit As Viney, et al. (23) and Bansback, et al. (22) both noted, an additive utility function with life expectancy and the levels of the EQ-5D would be inconsistent with the theoretical framework that underpins QALYs, because the QALY model requires that all health states have the same utility at death, i.e., as survival approaches zero, the systematic component of the utility function should similarly tend to zero. This satisfies the zero condition implicit in the QALY model (Bleichrodt and Johannesson, 997; Bleichrodt, et al., 997). Therefore, the utility of option in choice set for survey respondent is assumed to be U isj TIME X TIME, () isj isj isj isj 2 Flynn, et al. (28) argues that including the immediate death option in the choice modelling violates the random utility theory as some respondents may always choose survival over death no matter what health profiles are provided to them. 7

8 where represents a set of dummy variables relating to the levels of the EQ-5D health state, represents survival, and the error term are i.i.d. Gumbel distributions. It is conventional to use the best level of each dimension as the reference category. In this case excludes the dummies representing the best levels with other elements remaining: MO2, MO3, SC2, SC3, UA2, UA3, PD2, PD3, AD2, and AD3. For example, a health state denoted as 222 should translate into a vector (,,,,,,,,, ). In the current literature, the α and β terms have been assumed to be constant across individuals and based on this assumption equation () leads to the conditional logit model. 3 It is our baseline model and denoted as M. The estimation of α and β does not directly lead to the valuation of health states. An approach is needed to anchor the latent utility scale to the health state utility scale. There are several ways to derive this algorithm (Bansback, et al., 22; Ratcliffe, et al., 29; Viney, et al., 23). The main idea is that the utility value of a health state is its marginal utility of TIME on the latent scale, i.e., U TIME X. In the case of full health, its marginal utility of TIME on the latent scale is U TIME, which needs to be normalised to be under the QALY model. Hence the normalising constant is α and the utility score for a health state is 3 Viney, et al. (23) assume the error term to be normal. In this case equation () indicates a probit model. 8

9 X. The utility decrements are therefore β α. 3. Average valuation versus an average person s valuation As noted by Bansback, et al. (22), the objective was to derive the population mean utility scores for all possible health states, which requires estimation of population mean utility decrements. The conditional logit parameter estimates and represent population mean preferences for attributes that describe a health profile. In effect, the estimate represents an average person in the population whose preference parameters are exactly and. In this case, is actually the estimate of this average person s utility decrements. This is conceptually different from the population mean utility decrement which would be estimated by deriving, for each person in the target population for the person s and, and using this to calculate that individual's utility decrements. The population mean utility decrements are then computed as the average of all the individual decrements. Mathematically, this procedure can be described as which may or may not be close to, i.e., the ratio of means (. It is worth noting that when the TTO approach is used, this issue does not arise. When using TTO, a sample of health states are selected and respondents utility scores for these health states are elicited. These scores are then used as the dependent variable in a model which is 9

10 regressed on. In this case the regression coefficients, representing population mean utility decrements, are directly estimated using least squares (Dolan, 997; Viney, et al., 2). 4 Using MIXL: preference space versus QALY space One possible way to estimate the population mean utility decrements is to use a framework based on random parameters. Equation () can be rewritten as U isj TIME i isj X TIME i isj isj isj (2) where and are both random. The induced model is called the MIXL. Under this framework, we first estimate the distributions of (i.e., the distributions of utility decrements) and then derive the means of these distributions. To find the distribution of the ratio of two random variables is a longstanding problem. It has been particularly investigated in the WTP literature where represents the coefficient of price and represent the coefficients of non-price attributes in a DCE. Hensher and Greene (23) and Daly, et al. (22) discussed the major challenges in this area of research. The first challenge is that may not have finite moments unless is assumed to have some specific distributions such as log-normal. In our case, to assume to be a log-normal random variable is reasonable because represents a person s preference for the duration of life at perfect health condition and should be always positive. The second challenge concerns the extreme values that arise from the reciprocal of a random variable. As long as can take very small values, / will produce quite large numbers.

11 This problem is increasingly acute when s distribution has thick tails (e.g., student t and log-normal). We therefore estimated two MIXL models: M2.: log and follow multivariate normal distribution with mean μ and variance ; M2.2: log and log follow multivariate normal distribution with mean μ and variance. 4 The second model (M2.2) has the advantage of assuring the decrements distributions are strictly negative and the disadvantage of inducing a lot of extreme values. In contrast, the first model (M2.) may suffer less from the extreme values but it cannot guarantee each individual s utility decrements are strictly negative. Another challenge that has not been addressed in the literature is that the distribution of is induced from our assumptions on the distributions of and so it is not possible to directly compare and test the distributions of. In the WTP literature, alternative methods have been developed to meet these challenges (Daly, et al., 22). Among them the most promising effort has been the invention of the WTP space model (Train and Weeks, 25). The name WTP space was proposed as a contrast to the preference space on which the framework described above is based. The WTP space model is essentially a reparameterisation of equation (2) so that the distribution of can be directly assumed and estimated. We adapted this idea to our context and named the approach QALY space model. We now re-parameterise equation (2) as 4 For estimating M2.2 we need to change the signs of the data corresponding to to their opposite. This applies to other models when log-normal distribution is assumed for negative coefficients.

12 U isj i TIMEisj i X isjtimeisj isj (3) where /. Under this new framework we may estimate and compare models that assume different distributions on the utility decrements i. For the EQ-5D DCE data, we estimated three models: M3.: log( and follow multivariate normal distribution with mean μ and variance ; M3.2: log( and log( ) follow multivariate normal distribution with mean μ and variance ; M3.3: log( and log( follow multivariate normal distribution with mean μ and variance, where represents the size of (i.e., the number of utility decrements) and represents a positive unknown scalar parameter. The model M3. and M3.2 assume normal and log-normal distributions for the utility decrements respectively. Both have merits and flaws; the normal distribution has thin tails but cannot ensure everyone has negative decrements while the log-normal distribution is the opposite it can ensure everyone has negative decrements but has a thick right tail that may lead to very large mean estimates. Model M3.3 assumes Johnson s SB distribution for the utility decrements, i.e., exp / exp ) (4) where is normally distributed. It is a special case of Johnson s SB distribution with the lower bound set as and the upper bound to be estimated. 5 This distribution has both merits of normal and log-normal: thin tail and only taking positive numbers. Literature also 5 As the log-normal case, we changed the signs of the data corresponding to to their opposite. Therefore, a decrement s distribution should have a lower bound - and an upper bound. 2

13 shows that a wide variety of distributions such as normal, log-normal, Weibull, and modified beta can be satisfactorily fitted by the Johnson s SB distribution (Yu and Standish, 99). Moreover, it has been shown that Johnson s SB distribution can accommodate data with two modes spiked at the lower and upper bounds (Rigby and Burton, 26). Based on these evidences we expected M3.3 to be the best modelling strategy, especially given that we have limited prior knowledge on the shape of the distributions of utility decrements. 5 Estimation and model comparison The most popular methods for estimating MIXL are simulated maximum likelihood (SML) and Bayesian. Each has relative merits (Regier, et al., 29; Train, 23). The SML method is widely used, as most econometric and statistic software have developed standard routines to estimate MIXL based on this method. 6 However, the Bayesian approach has several clear advantages that suit our case. First, we assume all the random coefficients are correlated which leads to the estimation of a large covariance matrix. The SML method can be very time consuming in this case. And even with large number of simulation draws, convergence is not always guaranteed. In contrast, the Bayesian approach estimates correlated MIXL and uncorrelated MIXL at almost the same speed (Train, 23). Second, the SML method cannot estimate M3.3 without fixing the bounds while the Bayesian approach may estimate the bounds and other parameters simultaneously by using informative priors (we will show this in a moment). Therefore, in this study we chose to use the Bayesian method to estimate all the models including the conditional logit which is a special case of MIXL with its set as 6 For example, in STATA the mixlogit routine (Hole, 27) can be used to estimate the MIXL models in preference space while the gmnl rountine (Gu, et al., 23) can be modified to estimate MIXL models in QALY space or WTP space (Fiebig, et al., 2; Greene and Hensher, 2; Hole and Kolstad, 22). 3

14 an empty matrix. The sampling scheme for estimating the MIXL models in preference space was given in Train (23). The Matlab code written by Kenneth Train was used. 7 It is also straightforward to estimate the MIXL models in QALY space including M3. and M3.2; only a slight modification of the likelihood function is needed. The challenge comes from M3.3. As Train and Sonnier (25) pointed out, in equation (4) the bound parameter is closely related to the variance of and thus the model is under-identified. In the choice modelling literature, this under-identification is usually solved by fixing the s at a series of constants and then selecting the model with the best log-likelihood estimate. This approach is called grid search. The grid search method works well on the univariate case but for the multivariate situation it can be extremely laborious (Rigby and Burton, 26). In our case, we have a dimension multivariate Johnson s SB distribution and to identify the optimal point on the -D space is computationally infeasible. It is therefore necessary to seek an alternative solution. Our approach was based on using informative prior distributions on the bounds so that Bayesian identifiability of the model can be obtained. 8 The priors were log-normal distributions, constructed based on the estimates from M3.2. More specifically, the chosen priors cover the largest 99 th percentile of the log-normal distributions estimated from M3.2, a reasonable assumption of the upper bound of the bound parameter. The bound parameters were sampled as a vector using the random walk Metropolis-Hasting algorithm. 9 In order to confidently use the post burn-in iterates for inference, it is necessary to check that the sampling scheme has converged. We judged convergence visually by running the sampling scheme from three different initial positions and plotted various functionals of the 7 Available from 8 The mechanism is explained in detail in Scheines, et al. (999). 9 In the sampling, we first draw log( ) and then take its exponential. 4

15 iterates on the same graph. Successful convergence was indicated by the overlap of the functionals from the three chains. Following Train (23), we adopted a frequentist interpretation of the Bayesian estimates, i.e., the posterior means and standard deviations were used as the point estimates and standard errors. The decrements distributions were simulated using, random draws. The log-likelihood was calculated at the point estimates using, random draws. We also used AIC as the criterion for model comparisons. We did not use BIC as it penalises sample size heavily and thus, for very large sample sizes such as in this case, it is less informative in distinguishing between models that involve additional parameters. 6 Results 6. Estimation of Conditional logit (M) The parameter estimates of the conditional logit model were given in Table 2. Utility decrements based on were reported in the last column of Table 2. The interpretation of these numbers is that they represent an average person s utility decrements. [Insert Table 2 around here] We also used AICc which penalizes sample size, but due to the large sample size, AICc is almost identical to AIC. 5

16 6.2 Estimation of MIXL using preference space (M2. and M2.2) The parameter estimates of the two MIXL models using preference space were given in Table 3 (M2.) and Table 4 (M2.2). Base on log-likelihood and AIC, both models were substantially better than M. M2.2 also completely dominated M2. in terms of model fit indicating that the log-normal distribution assumption on accommodated the data much better than the normal distribution assumption. [Insert Table 3 around here] [Insert Table 4 around here] Based on these parameter estimates, the distributions of were simulated. The means of these distributions were reported in the tables as the population mean estimates of utility decrements. By comparing these two sets of estimates with the estimates from M, we found that for the size of level 2 decrements (e.g., MO2), overall M2.2 M M2.. For the size of level 3 decrements (e.g., MO3), overall M2.2 M2. M. The differences for the level 3 decrements were particularly significant. To understand these differences, we plotted these simulated distributions of in Figure 2 (for M2.) and Figure 3 (for level 3 decrements from M2.2). [Insert Figure 2 around here] 6

17 [Insert Figure 3 around here] From Figure 2 we can see that all the distributions from M2. have a significant proportion of the distribution greater than zero. This was particularly the case for the level 2 decrements. Given the EQ-5D is designed to be monotonic (level 2 is necessarily worse than level in each dimension), this is a concern. This also explains why the mean decrements for level 2 decrements from M2. were clearly smaller than the estimates from other two models. Another finding is that extreme values existed on both tails. If these extreme values spread out evenly on both sides the mean estimates would not be affected but unfortunately it is not the case. As shown in Figure 3, the problem of outliers is more severe in M2.2. All the distributions have very thick right tails indicating the population mean estimates are in fact determined by a group of extreme individuals. These extreme people may or may not exist in the real world, and it is questionable whether, in the policy making context the resulting valuations of health states should be driven by their valuations. To correct for this concern, a reasonable approach is to drop the % or 2% most extreme values from the simulated data (Daly, et al., 22; Hensher and Greene, 23). In Figure 3, we plotted the decrements distributions again after discarding the 2% most extreme values. They appeared to have much thinner tails. We also re-calculated the means which were reported in the last column of Table 4. The level 3 decrements mean estimates are now very close to those from M2. but still significantly larger than those from M. 7

18 6.3 Estimation of MIXL using QALY space (M3., M3.2, and M3.3) The parameter estimates of the first two MIXL models using QALY space were given in Table 5 (M3.) and Table 6 (M3.2). Base on log-likelihood and AIC, M3.2 was superior to M3. in terms of model fit indicating that the log-normal distribution assumption on the utility decrements was superior to the normal distribution assumption. Indeed, under M3., some decrements estimated distributions had substantial proportions greater than zero, which potentially led to the underestimation of these mean decrements. In the case of UA2, the sign clearly violates the monotonic condition. [Insert Table 5 around here] [Insert Table 6 around here] Another interesting comparison is M2.2 versus M3.2. The two models had very similar model fit with the latter slightly better. They also produced very similar utility decrements distributions indicating that whilst the distribution of from M2.2 is not in closed form it is in fact very close to log-normal distribution. The parameter estimates of the final model M3.3 were given in Table 7. When estimating the model we used informative prior distributions on all the bounds: ~, σ ) where σ was chosen as.6.,.36) covers a range from.25 to 4 (the st and 99 th percentiles). The 99 th percentiles of the log-normal distributions estimated from M3.2 (the smallest.8 and the largest 3.64) all locate well in this range. 8

19 Base on log-likelihood and AIC, M3.3 dominated M3.2, confirming that Johnson s SB is indeed a better distribution than log-normal for describing the utility decrement s distribution. We plotted the estimated distributions from both models in Figure 4 which clearly demonstrates Johnson s SB s advantage over log-normal: its shape is very close to lognormal but has a very thin tail. Unsurprisingly, the mean decrement estimates from this model were close to those from M2.2 and M3.2 where extreme values were discarded. [Insert Table 7 around here] [Insert Figure 4 around here] 7 Discussions and conclusions This study explored different estimation methods to provide estimates of the health state utility values that take better account of the individual heterogeneity in EQ-5D data that have been obtained using DCEs. This is important not only because previous methods do not exploit any of the individual heterogeneity in the raw data, but also because the methods for estimating health state utility values from DCE data need to model explicitly the variance as well as the means of the model parameters to provide population mean estimates of the health state utility values. In this paper we have argued that the previous methods that did not model variance such as conditional logit essentially derive an average person s valuation which is conceptually different from the average valuation from the population, the standard approach used in 9

20 TTO studies. The paper has developed methods to derive an average valuation from the population using DCE data. This average valuation is then more comparable with the TTO approach. Our methods were based on the MIXL framework and two types of models were proposed in this paper. The first is preference space modelling, which derives the distribution of utility decrements by taking the ratio of random variables. A significant problem associated with this approach is that the distributions are induced from our assumptions on these random parameters and so it is difficult to directly compare these induced distributions. For example, in our empirical analysis, we showed that M2.2 did have better model fit than M2.. However, it did not translate into a fact that the mean decrement estimates from the former model were more reasonable than those from the latter. In fact, the estimates from M2.2 were severely affected by extreme values as the induced distributions had very thick right tails. Dropping these extreme values would make the mean estimates more robust but the choice of the appropriate point of truncation is arbitrary. The second approach is based on an adaptation of methods developed in the WTP literature to deal with the drawbacks of preference space models. We have adapted the WTP space model to develop the second type of model in our analysis that is the QALY space model. It is essentially a re-parameterization of the preference space model so that the decrements distributions can be estimated and compared directly. In the empirical analysis we tried three different distribution assumptions for the utility decrements: normal, log-normal, and Johnson s SB. The last of these provided the best model fit. Our analysis showcased the advantages of Johnson s SB distribution over the normal and lognormal distributions the most commonly used ones in choice modelling practice. Johnson s SB distribution has not been widely used since it was first introduced to the choice modelling 2

21 literature by Train and Sonnier (25). The major reason may be the difficulty of its estimation which often needs an extensive search of the bounds. In this paper, we showed that it is also possible to estimate the bounds by using informative priors on them. In the empirical analysis, we identified plausible priors from a model using log-normal assumptions whose estimation showed that the bounds are likely to be smaller than Based on this, the prior distribution was constructed as, σ ) where σ was set as.6. We also did sensitivity analysis by changing σ and found that other values between.5 and would lead to similar results but the convergence of the model became harder as σ increases. By comparing the mean decrement estimates from M3.3 with the estimates from the conditional logit model, we found that the latter appeared to have smaller sizes. The largest differences happened to the level 3 decrements, in particular, MO3 and AD3. It is worth mentioning that when we estimated the conditional logit we did not impose any constraints while for M3.3 we imposed a monotonic constraint on each dimension of the EQ-5D. To explore the impact of doing so, we re-estimated the conditional logit with its β constraint to be negative (i.e. to impose monotonicity), and doing so did not change the parameter estimates at all. In Figure 5 we plotted the predicted values for all 243 health states described by the EQ-5D using estimates from M and M3.3. The ranking of the 243 health states from left to right is based on the predictions from the conditional logit approach. From the graph we can see that the conditional logit provides higher estimates of the utility values for almost all health states, with the divergence increasing with worsening health states. [Insert Figure 5 around here] 2

22 The DCEs offer a valuable alternative approach to the estimation of utility values, and is an area with an increasing international profile. In particular, it can be argued that the task is less onerous for respondents. However, the methods for analysing the data, and then for translating the result into an algorithm for use in economic evaluation remain contentious. We believe that the QALY space model approach outlined in this work represents a sensible way of using these data for this purpose, and should be explored using other generic quality of life instruments. Reference Bansback N, Brazier J, Tsuchiya A, Anis A. 22. Using a discrete choice experiment to estimate health state utility values. Journal of Health Economics 3: Bleichrodt H, Johannesson M The validity of qalys: An experimental test of constant proportional tradeoff and utility independence. Medical Decision Making 7: Bleichrodt N, Wakker P, Johannesson M Characterizing qalys by risk neutrality. Journal of Risk and Uncertainty 5: 7-4. Bosch JL, Hammitt JK, Weinstein MC, Hunink MG Estimating general-population utilities using one binary-gamble question per respondent. Medical Decision Making 8: Brazier J. 27. Measuring and valuing health benefits for economic evaluation. Oxford University Press: Oxford ; New York. Coast J, Flynn TN, Natarajan L, Sproston K, Lewis J, Louviere JJ, et al. 28. Valuing the icecap capability index for older people. Social Science & Medicine 67: Craig BM, Busschbach JJ, Salomon JA. 29. Keep it simple: Ranking health states yields values similar to cardinal measurement approaches. J Clin Epidemiol 62: Daly A, Hess S, Train K. 22. Assuring finite moments for willingness to pay in random coefficient models. Transportation 39:

23 Devlin NJ, Tsuchiya A, Buckingham K, Tilling C. 2. A uniform time trade off method for states better and worse than dead: Feasibility study of the 'lead time' approach. Health Economics 2: Dolan P Modeling valuations for euroqol health states. Medical Care 35: Fiebig DG, Keane MP, Louviere J, Wasi N. 2. The generalized multinomial logit model: Accounting for scale and coefficient heterogeneity. Marketing Science 29: Flynn TN. 2. Using conjoint analysis and choice experiments to estimate qaly values: Issues to consider. Pharmacoeconomics 28: Flynn TN, Louviere JJ, Marley AA, Coast J, Peters TJ. 28. Rescaling quality of life values from discrete choice experiments for use as qalys: A cautionary tale. Population Health Metrics 6. Greene WH, Hensher DA. 2. Does scale heterogeneity across individuals matter? An empirical assessment of alternative logit models. Transportation 37: Gu Y, Hole AR, Knox S. 23. Estimating the generalized multinomial logit model in stata. The Stata Journal in press. Hakim Z, Pathak DS Modelling the euroqol data: A comparison of discrete choice conjoint and conditional preference modelling. Health Economics 8: 3-6. Hensher DA, Greene WH. 23. The mixed logit model: The state of practice. Transportation 3: Hole AR. 27. Fitting mixed logit models by using maximum simulated likelihood. The Stata Journal 7: Hole AR, Kolstad JR. 22. Mixed logit estimation of willingness to pay distributions: A comparison of models in preference and wtp space using data from a health-related choice experiment. Empirical Economics 42: Lancsar E, Wildman J, Donaldson C, Ryan M, Baker R. 2. Deriving distributional weights for qalys through discrete choice experiments. Journal of Health Economics 3: Norman R, King MT, Clarke D, Viney R, Cronin P, Street D. 2. Does mode of administration matter? Comparison of online and face-to-face administration of a time tradeoff task. Qual Life Res 9: Pliskin JS, Shepard DS, Weinstein MC. 98. Utility-functions for life years and healthstatus. Operations Research 28: Ratcliffe J, Brazier J, Tsuchiya A, Symonds T, Brown M. 29. Using dce and ranking data to estimate cardinal values for health states for deriving a preference-based single index from the sexual quality of life questionnaire. Health Economics 8:

24 Regier DA, Ryan M, Phimister E, Marra CA. 29. Bayesian and classical estimation of mixed logit: An application to genetic testing. Journal of Health Economics 28: Richardson J, McKie J, Bariola E. Review and critique of health related multi attribute utility instruments. Centre for Health Economcs, Monash University, 2. Rigby D, Burton M. 26. Modeling disinterest and dislike: A bounded bayesian mixed logit model of the uk market for gm food. Environmental and Resource Economics 33: Ryan M, Netten A, Skatun D, Smith P. 26. Using discrete choice experiments to estimate a preference-based measure of outcome--an application to social care for older people. Journal of Health Economics 25: Scheines R, Hoijtink H, Boomsma A Bayesian estimation and testing of structural equation models. PSYCHOMETRIKA 64: Szende A, Oppe M, Devlin N, editors. Eq-5d value sets: Inventory, comparative review and user guide. Dordrecht, The Netherlands: Springer, 27. Train K. 23. Discrete choice methods with simulation. Cambridge University Press: New York. Train K, Sonnier G. Mixed logit with bounded distributions of correlated partworths. In: Scarpa R, Alberini A, editors. Applications of simulation methods in environmental and resource economics.. Dordrecht, The Netherlands: Springer Publisher, 25:7-34 Train K, Weeks M. Discrete choice models in preference space and willingness-to-pay space. In: Scarpa R, Alberini A, editors. Applications of simulation methods in environmental and resource economics.. Dordrecht, The Netherlands: Springer Publisher, 25:-6. Viney R, Norman R, Brazier J, Cronin P, King M, Ratcliffe J, et al. 23. An australian discrete choice experiment to value eq-5d health states. Health Economics in press. Viney R, Norman R, King MT, Cronin P, Street DJ, Knox S, et al. 2. Time trade-off derived eq-5d weights for australia. Value Health 4: Yu AB, Standish N. 99. A study of particle size distribution. Powder Technology 62:

25 Table. The EQ-5D instrument Dimension Level Description Mobility (MO) I have no problem in walking about 2 I have some problems in walking about 3 I am confined to bed Self-Care (SC) I have no problems with self-care 2 I have some problems washing and dressing myself 3 I am unable to wash and dress myself Usual Activities (UA) I have no problems with performing my usual activities 2 I have some problems with performing my usual activities 3 I am unable to perform my usual activities Pain / Discomfort (PD) Anxiety / Depression (AD) I have no pain or discomfort 2 I have moderate pain or discomfort 3 I have extreme pain or discomfort I am not anxious or depressed 2 I am moderately anxious or depressed 3 I am extremely anxious or depressed 25

26 Table 2. Conditional logit (M) Parameters Utility decrements Attributes Estimate (S.E.) Levels Time.27 (.7) MO2*Time -.3 (.4) MO2 -.2 MO3*Time -.4 (.4) MO SC2*Time -.3 (.5) SC2 -.2 SC3*Time -.8 (.5) SC UA2*Time -.3 (.5) UA2 -. UA3*Time -.5 (.5) UA3 -.9 PD2*Time -.3 (.4) PD2 -. PD3*Time -.3 (.4) PD3 -.5 AD2*Time -.4 (.4) AD2 -.4 AD3*Time -. (.4) AD Log-likelihood -892 No. of parameters AIC

27 Table 3. MIXL using preference space: ~Log-normal and ~Normal (M2.) Parameters Utility decrements Attributes Mean (S.E.) S.D. (S.E.) Levels Mean S.D. Time.53 (.4).79 (.4) MO2*Time -.9 (.3).48 (.3) MO MO3*Time -. (.5).79 (.4) MO SC2*Time -.2 (.4).59 (.4) SC SC3*Time -.6 (.4).7 (.4) SC UA2*Time -.5 (.3).53 (.3) UA UA3*Time -.4 (.4).59 (.4) UA PD2*Time -.6 (.3).47 (.3) PD PD3*Time -.3 (.5).78 (.4) PD AD2*Time -.25 (.3).5 (.3) AD AD3*Time -.86 (.4).79 (.4) AD Log-likelihood -786 No. of parameters 77 AIC

28 Table 4. MIXL using preference space: ~Log-normal and ~Log-normal (M2.2) Parameters Utility decrements Attributes Mean (S.E.) S.D. (S.E.) Levels Original (S.D.) Truncated (S.D.) Time -.2 (.8).63 (.) MO2*Time (.24).72 (.23) MO2 -.3 (.23) -. (.2) MO3*Time -.64 (.8).53 (.9) MO (.79) -.7 (.56) SC2*Time (.25).6 (.2) SC2 -.5 (.3) -.2 (.5) SC3*Time -.43 (.).56 (.2) SC (.6) -.36 (.36) UA2*Time (.35).92 (.3) UA2 -. (.22) -.8 (.) UA3*Time -.87 (.4).7 (.5) UA (.3) -.22 (.2) PD2*Time -3.3 (.28).88 (.22) PD2 -. (.23) -.9 (.) PD3*Time -.74 (.8).64 (.) PD (.82) -.65 (.56) AD2*Time (.2).99 (.9) AD2 -.4 (.22) -.2 (.3) AD3*Time -.4 (.).82 (.) AD (.64) -.49 (.43) Log-likelihood No. of parameters 77 AIC

29 Table 5. MIXL using QALY space: ~Log-normal and ~Normal (M3.) Parameters Utility decrements Attributes Mean (S.E.) S.D. (S.E.) Levels Mean S.D. Time.32 (.2).84 (.4) MO2*Time -.7 (.2).37 (.2) MO MO3*Time -.77 (.3).6 (.3) MO SC2*Time -.3 (.3).42 (.2) SC SC3*Time -.33 (.3).54 (.3) SC UA2*Time.3 (.3).39 (.2) UA UA3*Time -.6 (.3).43 (.2) UA PD2*Time -.4 (.2).37 (.2) PD PD3*Time -.7 (.3).59 (.3) PD AD2*Time -.8 (.2).37 (.2) AD AD3*Time -.53 (.3).56 (.3) AD Log-likelihood -776 No. of parameters 77 AIC

30 Table 6. MIXL using QALY space: ~Log-normal and ~Log-normal (M3.2) Parameters Utility decrements Attributes Mean (S.E.) S.D. (S.E.) Levels Mean S.D. Time.2 (.8).68 (.) MO2*Time (.23).5 (.4) MO MO3*Time -.63 (.4).83 (.5) MO SC2*Time (.2).2 (.4) SC SC3*Time -.4 (.8).3 (.7) SC UA2*Time -3.5 (.33).27 (.7) UA UA3*Time -.83 (.).9 (.8) UA PD2*Time (.25).26 (.4) PD PD3*Time -.74 (.5).9 (.5) PD AD2*Time (.7). (.) AD AD3*Time -.2 (.6).9 (.6) AD Log-likelihood No. of parameters 77 AIC

31 Table 7. MIXL using QALY space: ~Log-normal and ~Johnson s SB (M3.3) Parameters Utility decrements Attributes Mean (S.E.) S.D. (S.E.) Bound (S.E.) Levels Mean S.D. Time.3 (.8).66 (.9) MO2*Time (.63) 2.28 (.7) -.84 (.39) MO MO3*Time -.2 (.32).54 (.2) -.36 (.22) MO SC2*Time -3.9 (.72) 2.95 (.97) -.88 (.27) SC SC3*Time -.2 (.33) 2.48 (.62) -.97 (.8) SC UA2*Time (.92) 2.95 (.82) -.9 (.2) UA UA3*Time -.45 (.43) 2.2 (.7) -.79 (.2) UA PD2*Time (.85) 2.79 (.6) -.57 (.23) PD PD3*Time -.59 (.35).42 (.2) -.53 (.32) PD AD2*Time (.62) 2.94 (.89) -.62 (.) AD AD3*Time -.42 (.35).32 (.7) -.9 (.58) AD Log-likelihood No. of parameters 87 AIC 57 3

32 Figure. An example choice set 32

33 Figure 2. Kernel densities of utility decrements estimated from the preference space model using normal distribution assumption (M2.) 2 MO2 MO SC UA PD AD SC UA PD AD The left panel displays the kernel densities of level 2 decrements and the right panel displays the kernel densities of level 3 decrements. All densities were estimated using, random draws. 33

34 Figure 3. Kernel densities of utility decrements estimated from the preference space model using log-normal distribution assumption (M2.2) 2 No truncation 2 2% truncation MO MO SC SC UA PD AD UA PD AD3 The left panel displays the kernel densities of level 3 decrements estimated using, random draws and the right panel displays the kernel densities of level 3 decrements estimated using these random draws with the smallest 2% discarded. 34

35 Figure 4. Utility decrements distributions estimated from two QALY space models 2 MO SC UA PD AD MO SC UA PD AD The solid lines represent the distributions estimated from the QALY space model using log-normal distribution assumption (M3.2) and the dotted lines represent the distributions estimated from the QALY space model using Johnson s SB distribution assumption (M3.3). The estimated log-normal distributions were all projected to the negative real line. 35

36 Figure 5. Predicted EQ-5D health state utility values.5 Utility Values Health State The solid line represents the predictions from the conditional logit (M) and the dotted line represents the predictions from the preferred QALY space model (M3.3). The ranking of the 243 health states from left to right is based on the predictions from the conditional logit. 36

Well-being and the value of health

Well-being and the value of health Well-being and the value of health Happiness and Public Policy Conference Bangkok, Thailand 8-9 July 2007 Bernard van den Berg Department of Health Economics & Health Technology Assessment, Institute of

More information

CALCULATIONS & STATISTICS

CALCULATIONS & STATISTICS CALCULATIONS & STATISTICS CALCULATION OF SCORES Conversion of 1-5 scale to 0-100 scores When you look at your report, you will notice that the scores are reported on a 0-100 scale, even though respondents

More information

eq5d: A command to calculate index values for the EQ-5D quality-of-life instrument

eq5d: A command to calculate index values for the EQ-5D quality-of-life instrument The Stata Journal (2011) 11, Number 1, pp. 120 125 eq5d: A command to calculate index values for the EQ-5D quality-of-life instrument Juan Manuel Ramos-Goñi Canary Islands Health Care Service Canary Islands,

More information

Time to tweak the TTO: results from a comparison of alternative specifications of the TTO

Time to tweak the TTO: results from a comparison of alternative specifications of the TTO Eur J Health Econ (2013) 14 (Suppl 1):S43 S51 DOI 10.1007/s10198-013-0507-y ORIGINAL PAPER Time to tweak the TTO: results from a comparison of alternative specifications of the TTO Matthijs M. Versteegh

More information

Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus

Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus Tihomir Asparouhov and Bengt Muthén Mplus Web Notes: No. 15 Version 8, August 5, 2014 1 Abstract This paper discusses alternatives

More information

A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution

A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 4: September

More information

II. DISTRIBUTIONS distribution normal distribution. standard scores

II. DISTRIBUTIONS distribution normal distribution. standard scores Appendix D Basic Measurement And Statistics The following information was developed by Steven Rothke, PhD, Department of Psychology, Rehabilitation Institute of Chicago (RIC) and expanded by Mary F. Schmidt,

More information

Centre for Central Banking Studies

Centre for Central Banking Studies Centre for Central Banking Studies Technical Handbook No. 4 Applied Bayesian econometrics for central bankers Andrew Blake and Haroon Mumtaz CCBS Technical Handbook No. 4 Applied Bayesian econometrics

More information

CHAPTER 3 EXAMPLES: REGRESSION AND PATH ANALYSIS

CHAPTER 3 EXAMPLES: REGRESSION AND PATH ANALYSIS Examples: Regression And Path Analysis CHAPTER 3 EXAMPLES: REGRESSION AND PATH ANALYSIS Regression analysis with univariate or multivariate dependent variables is a standard procedure for modeling relationships

More information

Least Squares Estimation

Least Squares Estimation Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David

More information

SUMAN DUVVURU STAT 567 PROJECT REPORT

SUMAN DUVVURU STAT 567 PROJECT REPORT SUMAN DUVVURU STAT 567 PROJECT REPORT SURVIVAL ANALYSIS OF HEROIN ADDICTS Background and introduction: Current illicit drug use among teens is continuing to increase in many countries around the world.

More information

APPLIED MISSING DATA ANALYSIS

APPLIED MISSING DATA ANALYSIS APPLIED MISSING DATA ANALYSIS Craig K. Enders Series Editor's Note by Todd D. little THE GUILFORD PRESS New York London Contents 1 An Introduction to Missing Data 1 1.1 Introduction 1 1.2 Chapter Overview

More information

1. INTRODUCTION...3 EUROQOL GROUP...3 EQ-5D...4 WHAT IS A HEALTH STATE?...7 VERSIONS OF EQ-5D...8

1. INTRODUCTION...3 EUROQOL GROUP...3 EQ-5D...4 WHAT IS A HEALTH STATE?...7 VERSIONS OF EQ-5D...8 Table of contents 1. INTRODUCTION...3 EUROQOL GROUP...3 EQ-5D...4 WHAT IS A HEALTH STATE?...7 VERSIONS OF EQ-5D...8 2. SCORING THE EQ-5D DESCRIPTIVE SYSTEM...9 3. SCORING THE EQ VAS...1 4. CONVERTING EQ-5D

More information

Permutation Tests for Comparing Two Populations

Permutation Tests for Comparing Two Populations Permutation Tests for Comparing Two Populations Ferry Butar Butar, Ph.D. Jae-Wan Park Abstract Permutation tests for comparing two populations could be widely used in practice because of flexibility of

More information

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written

More information

Multiple Imputation for Missing Data: A Cautionary Tale

Multiple Imputation for Missing Data: A Cautionary Tale Multiple Imputation for Missing Data: A Cautionary Tale Paul D. Allison University of Pennsylvania Address correspondence to Paul D. Allison, Sociology Department, University of Pennsylvania, 3718 Locust

More information

PS 271B: Quantitative Methods II. Lecture Notes

PS 271B: Quantitative Methods II. Lecture Notes PS 271B: Quantitative Methods II Lecture Notes Langche Zeng zeng@ucsd.edu The Empirical Research Process; Fundamental Methodological Issues 2 Theory; Data; Models/model selection; Estimation; Inference.

More information

D-optimal plans in observational studies

D-optimal plans in observational studies D-optimal plans in observational studies Constanze Pumplün Stefan Rüping Katharina Morik Claus Weihs October 11, 2005 Abstract This paper investigates the use of Design of Experiments in observational

More information

What is a QALY? What is...? series. Second edition. Health economics. Supported by sanofi-aventis

What is a QALY? What is...? series. Second edition. Health economics. Supported by sanofi-aventis ...? series Supported by sanofi-aventis Second edition Health economics What is a QALY? Ceri Phillips BSc(Econ) MSc(Econ) PhD Professor of Health Economics, Swansea University A quality-adjusted life-year

More information

What s New in Econometrics? Lecture 8 Cluster and Stratified Sampling

What s New in Econometrics? Lecture 8 Cluster and Stratified Sampling What s New in Econometrics? Lecture 8 Cluster and Stratified Sampling Jeff Wooldridge NBER Summer Institute, 2007 1. The Linear Model with Cluster Effects 2. Estimation with a Small Number of Groups and

More information

The Effects of Start Prices on the Performance of the Certainty Equivalent Pricing Policy

The Effects of Start Prices on the Performance of the Certainty Equivalent Pricing Policy BMI Paper The Effects of Start Prices on the Performance of the Certainty Equivalent Pricing Policy Faculty of Sciences VU University Amsterdam De Boelelaan 1081 1081 HV Amsterdam Netherlands Author: R.D.R.

More information

Basics of Statistical Machine Learning

Basics of Statistical Machine Learning CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu Modern machine learning is rooted in statistics. You will find many familiar

More information

Simple linear regression

Simple linear regression Simple linear regression Introduction Simple linear regression is a statistical method for obtaining a formula to predict values of one variable from another where there is a causal relationship between

More information

A Basic Introduction to Missing Data

A Basic Introduction to Missing Data John Fox Sociology 740 Winter 2014 Outline Why Missing Data Arise Why Missing Data Arise Global or unit non-response. In a survey, certain respondents may be unreachable or may refuse to participate. Item

More information

Keep It Simple: Easy Ways To Estimate Choice Models For Single Consumers

Keep It Simple: Easy Ways To Estimate Choice Models For Single Consumers Keep It Simple: Easy Ways To Estimate Choice Models For Single Consumers Christine Ebling, University of Technology Sydney, christine.ebling@uts.edu.au Bart Frischknecht, University of Technology Sydney,

More information

Imputing Missing Data using SAS

Imputing Missing Data using SAS ABSTRACT Paper 3295-2015 Imputing Missing Data using SAS Christopher Yim, California Polytechnic State University, San Luis Obispo Missing data is an unfortunate reality of statistics. However, there are

More information

Local outlier detection in data forensics: data mining approach to flag unusual schools

Local outlier detection in data forensics: data mining approach to flag unusual schools Local outlier detection in data forensics: data mining approach to flag unusual schools Mayuko Simon Data Recognition Corporation Paper presented at the 2012 Conference on Statistical Detection of Potential

More information

OBJECTIVE ASSESSMENT OF FORECASTING ASSIGNMENTS USING SOME FUNCTION OF PREDICTION ERRORS

OBJECTIVE ASSESSMENT OF FORECASTING ASSIGNMENTS USING SOME FUNCTION OF PREDICTION ERRORS OBJECTIVE ASSESSMENT OF FORECASTING ASSIGNMENTS USING SOME FUNCTION OF PREDICTION ERRORS CLARKE, Stephen R. Swinburne University of Technology Australia One way of examining forecasting methods via assignments

More information

Sample Size and Power in Clinical Trials

Sample Size and Power in Clinical Trials Sample Size and Power in Clinical Trials Version 1.0 May 011 1. Power of a Test. Factors affecting Power 3. Required Sample Size RELATED ISSUES 1. Effect Size. Test Statistics 3. Variation 4. Significance

More information

BayesX - Software for Bayesian Inference in Structured Additive Regression

BayesX - Software for Bayesian Inference in Structured Additive Regression BayesX - Software for Bayesian Inference in Structured Additive Regression Thomas Kneib Faculty of Mathematics and Economics, University of Ulm Department of Statistics, Ludwig-Maximilians-University Munich

More information

Mode and Patient-mix Adjustment of the CAHPS Hospital Survey (HCAHPS)

Mode and Patient-mix Adjustment of the CAHPS Hospital Survey (HCAHPS) Mode and Patient-mix Adjustment of the CAHPS Hospital Survey (HCAHPS) April 30, 2008 Abstract A randomized Mode Experiment of 27,229 discharges from 45 hospitals was used to develop adjustments for the

More information

Genetic Discoveries and the Role of Health Economics

Genetic Discoveries and the Role of Health Economics Genetic Discoveries and the Role of Health Economics Collaboration for Outcome Research and Evaluation (CORE) Faculty of Pharmaceutical Sciences University of British Columbia February 02, 2010 Content

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes

More information

5. Multiple regression

5. Multiple regression 5. Multiple regression QBUS6840 Predictive Analytics https://www.otexts.org/fpp/5 QBUS6840 Predictive Analytics 5. Multiple regression 2/39 Outline Introduction to multiple linear regression Some useful

More information

INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition)

INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition) INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition) Abstract Indirect inference is a simulation-based method for estimating the parameters of economic models. Its

More information

EQ-5D-3L User Guide. Basic information on how to use the EQ-5D-3L instrument. Version 5.1 April 2015. Prepared by Mandy van Reenen / Mark Oppe

EQ-5D-3L User Guide. Basic information on how to use the EQ-5D-3L instrument. Version 5.1 April 2015. Prepared by Mandy van Reenen / Mark Oppe EQ-5D-3L User Guide Basic information on how to use the EQ-5D-3L instrument Version 5.1 April 215 Prepared by Mandy van Reenen / Mark Oppe Table of contents 1. INTRODUCTION... 3 1.1. THE EUROQOL GROUP...

More information

Markov Chain Monte Carlo Simulation Made Simple

Markov Chain Monte Carlo Simulation Made Simple Markov Chain Monte Carlo Simulation Made Simple Alastair Smith Department of Politics New York University April2,2003 1 Markov Chain Monte Carlo (MCMC) simualtion is a powerful technique to perform numerical

More information

econstor zbw www.econstor.eu

econstor zbw www.econstor.eu econstor www.econstor.eu Der Open-Access-Publikationsserver der ZBW Leibniz-Informationszentrum Wirtschaft The Open Access Publication Server of the ZBW Leibniz Information Centre for Economics Hess, Stephane

More information

Problem of Missing Data

Problem of Missing Data VASA Mission of VA Statisticians Association (VASA) Promote & disseminate statistical methodological research relevant to VA studies; Facilitate communication & collaboration among VA-affiliated statisticians;

More information

Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.

Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not. Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C

More information

Interpretation of Somers D under four simple models

Interpretation of Somers D under four simple models Interpretation of Somers D under four simple models Roger B. Newson 03 September, 04 Introduction Somers D is an ordinal measure of association introduced by Somers (96)[9]. It can be defined in terms

More information

Standard errors of marginal effects in the heteroskedastic probit model

Standard errors of marginal effects in the heteroskedastic probit model Standard errors of marginal effects in the heteroskedastic probit model Thomas Cornelißen Discussion Paper No. 320 August 2005 ISSN: 0949 9962 Abstract In non-linear regression models, such as the heteroskedastic

More information

Fitting Subject-specific Curves to Grouped Longitudinal Data

Fitting Subject-specific Curves to Grouped Longitudinal Data Fitting Subject-specific Curves to Grouped Longitudinal Data Djeundje, Viani Heriot-Watt University, Department of Actuarial Mathematics & Statistics Edinburgh, EH14 4AS, UK E-mail: vad5@hw.ac.uk Currie,

More information

4. Continuous Random Variables, the Pareto and Normal Distributions

4. Continuous Random Variables, the Pareto and Normal Distributions 4. Continuous Random Variables, the Pareto and Normal Distributions A continuous random variable X can take any value in a given range (e.g. height, weight, age). The distribution of a continuous random

More information

STATISTICA Formula Guide: Logistic Regression. Table of Contents

STATISTICA Formula Guide: Logistic Regression. Table of Contents : Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary

More information

SENSITIVITY ANALYSIS AND INFERENCE. Lecture 12

SENSITIVITY ANALYSIS AND INFERENCE. Lecture 12 This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this

More information

1/27/2013. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2

1/27/2013. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2 PSY 512: Advanced Statistics for Psychological and Behavioral Research 2 Introduce moderated multiple regression Continuous predictor continuous predictor Continuous predictor categorical predictor Understand

More information

Constructing a TpB Questionnaire: Conceptual and Methodological Considerations

Constructing a TpB Questionnaire: Conceptual and Methodological Considerations Constructing a TpB Questionnaire: Conceptual and Methodological Considerations September, 2002 (Revised January, 2006) Icek Ajzen Brief Description of the Theory of Planned Behavior According to the theory

More information

Handling attrition and non-response in longitudinal data

Handling attrition and non-response in longitudinal data Longitudinal and Life Course Studies 2009 Volume 1 Issue 1 Pp 63-72 Handling attrition and non-response in longitudinal data Harvey Goldstein University of Bristol Correspondence. Professor H. Goldstein

More information

Java Modules for Time Series Analysis

Java Modules for Time Series Analysis Java Modules for Time Series Analysis Agenda Clustering Non-normal distributions Multifactor modeling Implied ratings Time series prediction 1. Clustering + Cluster 1 Synthetic Clustering + Time series

More information

PROPERTIES OF THE SAMPLE CORRELATION OF THE BIVARIATE LOGNORMAL DISTRIBUTION

PROPERTIES OF THE SAMPLE CORRELATION OF THE BIVARIATE LOGNORMAL DISTRIBUTION PROPERTIES OF THE SAMPLE CORRELATION OF THE BIVARIATE LOGNORMAL DISTRIBUTION Chin-Diew Lai, Department of Statistics, Massey University, New Zealand John C W Rayner, School of Mathematics and Applied Statistics,

More information

A Procedure for Classifying New Respondents into Existing Segments Using Maximum Difference Scaling

A Procedure for Classifying New Respondents into Existing Segments Using Maximum Difference Scaling A Procedure for Classifying New Respondents into Existing Segments Using Maximum Difference Scaling Background Bryan Orme and Rich Johnson, Sawtooth Software March, 2009 Market segmentation is pervasive

More information

MISSING DATA TECHNIQUES WITH SAS. IDRE Statistical Consulting Group

MISSING DATA TECHNIQUES WITH SAS. IDRE Statistical Consulting Group MISSING DATA TECHNIQUES WITH SAS IDRE Statistical Consulting Group ROAD MAP FOR TODAY To discuss: 1. Commonly used techniques for handling missing data, focusing on multiple imputation 2. Issues that could

More information

Nonparametric adaptive age replacement with a one-cycle criterion

Nonparametric adaptive age replacement with a one-cycle criterion Nonparametric adaptive age replacement with a one-cycle criterion P. Coolen-Schrijner, F.P.A. Coolen Department of Mathematical Sciences University of Durham, Durham, DH1 3LE, UK e-mail: Pauline.Schrijner@durham.ac.uk

More information

Introduction to Fixed Effects Methods

Introduction to Fixed Effects Methods Introduction to Fixed Effects Methods 1 1.1 The Promise of Fixed Effects for Nonexperimental Research... 1 1.2 The Paired-Comparisons t-test as a Fixed Effects Method... 2 1.3 Costs and Benefits of Fixed

More information

Introduction to General and Generalized Linear Models

Introduction to General and Generalized Linear Models Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby

More information

Ordinal Regression. Chapter

Ordinal Regression. Chapter Ordinal Regression Chapter 4 Many variables of interest are ordinal. That is, you can rank the values, but the real distance between categories is unknown. Diseases are graded on scales from least severe

More information

Multivariate Normal Distribution

Multivariate Normal Distribution Multivariate Normal Distribution Lecture 4 July 21, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #4-7/21/2011 Slide 1 of 41 Last Time Matrices and vectors Eigenvalues

More information

The Method of Least Squares

The Method of Least Squares Hervé Abdi 1 1 Introduction The least square methods (LSM) is probably the most popular technique in statistics. This is due to several factors. First, most common estimators can be casted within this

More information

Analyzing Structural Equation Models With Missing Data

Analyzing Structural Equation Models With Missing Data Analyzing Structural Equation Models With Missing Data Craig Enders* Arizona State University cenders@asu.edu based on Enders, C. K. (006). Analyzing structural equation models with missing data. In G.

More information

Tutorial on Markov Chain Monte Carlo

Tutorial on Markov Chain Monte Carlo Tutorial on Markov Chain Monte Carlo Kenneth M. Hanson Los Alamos National Laboratory Presented at the 29 th International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Technology,

More information

Master s Theory Exam Spring 2006

Master s Theory Exam Spring 2006 Spring 2006 This exam contains 7 questions. You should attempt them all. Each question is divided into parts to help lead you through the material. You should attempt to complete as much of each problem

More information

Session 7 Bivariate Data and Analysis

Session 7 Bivariate Data and Analysis Session 7 Bivariate Data and Analysis Key Terms for This Session Previously Introduced mean standard deviation New in This Session association bivariate analysis contingency table co-variation least squares

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Math 541: Statistical Theory II Lecturer: Songfeng Zheng Maximum Likelihood Estimation 1 Maximum Likelihood Estimation Maximum likelihood is a relatively simple method of constructing an estimator for

More information

Two-Sample T-Tests Assuming Equal Variance (Enter Means)

Two-Sample T-Tests Assuming Equal Variance (Enter Means) Chapter 4 Two-Sample T-Tests Assuming Equal Variance (Enter Means) Introduction This procedure provides sample size and power calculations for one- or two-sided two-sample t-tests when the variances of

More information

Lean Six Sigma Analyze Phase Introduction. TECH 50800 QUALITY and PRODUCTIVITY in INDUSTRY and TECHNOLOGY

Lean Six Sigma Analyze Phase Introduction. TECH 50800 QUALITY and PRODUCTIVITY in INDUSTRY and TECHNOLOGY TECH 50800 QUALITY and PRODUCTIVITY in INDUSTRY and TECHNOLOGY Before we begin: Turn on the sound on your computer. There is audio to accompany this presentation. Audio will accompany most of the online

More information

Penalized regression: Introduction

Penalized regression: Introduction Penalized regression: Introduction Patrick Breheny August 30 Patrick Breheny BST 764: Applied Statistical Modeling 1/19 Maximum likelihood Much of 20th-century statistics dealt with maximum likelihood

More information

Marketing Mix Modelling and Big Data P. M Cain

Marketing Mix Modelling and Big Data P. M Cain 1) Introduction Marketing Mix Modelling and Big Data P. M Cain Big data is generally defined in terms of the volume and variety of structured and unstructured information. Whereas structured data is stored

More information

Facebook Friend Suggestion Eytan Daniyalzade and Tim Lipus

Facebook Friend Suggestion Eytan Daniyalzade and Tim Lipus Facebook Friend Suggestion Eytan Daniyalzade and Tim Lipus 1. Introduction Facebook is a social networking website with an open platform that enables developers to extract and utilize user information

More information

The Probit Link Function in Generalized Linear Models for Data Mining Applications

The Probit Link Function in Generalized Linear Models for Data Mining Applications Journal of Modern Applied Statistical Methods Copyright 2013 JMASM, Inc. May 2013, Vol. 12, No. 1, 164-169 1538 9472/13/$95.00 The Probit Link Function in Generalized Linear Models for Data Mining Applications

More information

A THEORETICAL COMPARISON OF DATA MASKING TECHNIQUES FOR NUMERICAL MICRODATA

A THEORETICAL COMPARISON OF DATA MASKING TECHNIQUES FOR NUMERICAL MICRODATA A THEORETICAL COMPARISON OF DATA MASKING TECHNIQUES FOR NUMERICAL MICRODATA Krish Muralidhar University of Kentucky Rathindra Sarathy Oklahoma State University Agency Internal User Unmasked Result Subjects

More information

Extreme Value Modeling for Detection and Attribution of Climate Extremes

Extreme Value Modeling for Detection and Attribution of Climate Extremes Extreme Value Modeling for Detection and Attribution of Climate Extremes Jun Yan, Yujing Jiang Joint work with Zhuo Wang, Xuebin Zhang Department of Statistics, University of Connecticut February 2, 2016

More information

An introduction to Value-at-Risk Learning Curve September 2003

An introduction to Value-at-Risk Learning Curve September 2003 An introduction to Value-at-Risk Learning Curve September 2003 Value-at-Risk The introduction of Value-at-Risk (VaR) as an accepted methodology for quantifying market risk is part of the evolution of risk

More information

Report on the Scaling of the 2014 NSW Higher School Certificate. NSW Vice-Chancellors Committee Technical Committee on Scaling

Report on the Scaling of the 2014 NSW Higher School Certificate. NSW Vice-Chancellors Committee Technical Committee on Scaling Report on the Scaling of the 2014 NSW Higher School Certificate NSW Vice-Chancellors Committee Technical Committee on Scaling Contents Preface Acknowledgements Definitions iii iv v 1 The Higher School

More information

LOGISTIC REGRESSION ANALYSIS

LOGISTIC REGRESSION ANALYSIS LOGISTIC REGRESSION ANALYSIS C. Mitchell Dayton Department of Measurement, Statistics & Evaluation Room 1230D Benjamin Building University of Maryland September 1992 1. Introduction and Model Logistic

More information

Predict the Popularity of YouTube Videos Using Early View Data

Predict the Popularity of YouTube Videos Using Early View Data 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050

More information

Chapter 1 Introduction. 1.1 Introduction

Chapter 1 Introduction. 1.1 Introduction Chapter 1 Introduction 1.1 Introduction 1 1.2 What Is a Monte Carlo Study? 2 1.2.1 Simulating the Rolling of Two Dice 2 1.3 Why Is Monte Carlo Simulation Often Necessary? 4 1.4 What Are Some Typical Situations

More information

2. Linear regression with multiple regressors

2. Linear regression with multiple regressors 2. Linear regression with multiple regressors Aim of this section: Introduction of the multiple regression model OLS estimation in multiple regression Measures-of-fit in multiple regression Assumptions

More information

Association Between Variables

Association Between Variables Contents 11 Association Between Variables 767 11.1 Introduction............................ 767 11.1.1 Measure of Association................. 768 11.1.2 Chapter Summary.................... 769 11.2 Chi

More information

Module 3: Correlation and Covariance

Module 3: Correlation and Covariance Using Statistical Data to Make Decisions Module 3: Correlation and Covariance Tom Ilvento Dr. Mugdim Pašiƒ University of Delaware Sarajevo Graduate School of Business O ften our interest in data analysis

More information

Chapter 6: The Information Function 129. CHAPTER 7 Test Calibration

Chapter 6: The Information Function 129. CHAPTER 7 Test Calibration Chapter 6: The Information Function 129 CHAPTER 7 Test Calibration 130 Chapter 7: Test Calibration CHAPTER 7 Test Calibration For didactic purposes, all of the preceding chapters have assumed that the

More information

Joint models for classification and comparison of mortality in different countries.

Joint models for classification and comparison of mortality in different countries. Joint models for classification and comparison of mortality in different countries. Viani D. Biatat 1 and Iain D. Currie 1 1 Department of Actuarial Mathematics and Statistics, and the Maxwell Institute

More information

LOGIT AND PROBIT ANALYSIS

LOGIT AND PROBIT ANALYSIS LOGIT AND PROBIT ANALYSIS A.K. Vasisht I.A.S.R.I., Library Avenue, New Delhi 110 012 amitvasisht@iasri.res.in In dummy regression variable models, it is assumed implicitly that the dependent variable Y

More information

Solution: The optimal position for an investor with a coefficient of risk aversion A = 5 in the risky asset is y*:

Solution: The optimal position for an investor with a coefficient of risk aversion A = 5 in the risky asset is y*: Problem 1. Consider a risky asset. Suppose the expected rate of return on the risky asset is 15%, the standard deviation of the asset return is 22%, and the risk-free rate is 6%. What is your optimal position

More information

PARTIAL LEAST SQUARES IS TO LISREL AS PRINCIPAL COMPONENTS ANALYSIS IS TO COMMON FACTOR ANALYSIS. Wynne W. Chin University of Calgary, CANADA

PARTIAL LEAST SQUARES IS TO LISREL AS PRINCIPAL COMPONENTS ANALYSIS IS TO COMMON FACTOR ANALYSIS. Wynne W. Chin University of Calgary, CANADA PARTIAL LEAST SQUARES IS TO LISREL AS PRINCIPAL COMPONENTS ANALYSIS IS TO COMMON FACTOR ANALYSIS. Wynne W. Chin University of Calgary, CANADA ABSTRACT The decision of whether to use PLS instead of a covariance

More information

IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem

IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem IEOR 6711: Stochastic Models I Fall 2012, Professor Whitt, Tuesday, September 11 Normal Approximations and the Central Limit Theorem Time on my hands: Coin tosses. Problem Formulation: Suppose that I have

More information

Analysis of Bayesian Dynamic Linear Models

Analysis of Bayesian Dynamic Linear Models Analysis of Bayesian Dynamic Linear Models Emily M. Casleton December 17, 2010 1 Introduction The main purpose of this project is to explore the Bayesian analysis of Dynamic Linear Models (DLMs). The main

More information

Linear Programming for Optimization. Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc.

Linear Programming for Optimization. Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc. 1. Introduction Linear Programming for Optimization Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc. 1.1 Definition Linear programming is the name of a branch of applied mathematics that

More information

MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS

MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS MSR = Mean Regression Sum of Squares MSE = Mean Squared Error RSS = Regression Sum of Squares SSE = Sum of Squared Errors/Residuals α = Level of Significance

More information

What are health utilities?

What are health utilities? What is...? series Supported by sanofi-aventis Second edition Health economics What are Keith Tolley MPhil Health Economist and Director, Tolley Health Economics Ltd Utilities are cardinal values that

More information

Least-Squares Intersection of Lines

Least-Squares Intersection of Lines Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a

More information

Comparing Features of Convenient Estimators for Binary Choice Models With Endogenous Regressors

Comparing Features of Convenient Estimators for Binary Choice Models With Endogenous Regressors Comparing Features of Convenient Estimators for Binary Choice Models With Endogenous Regressors Arthur Lewbel, Yingying Dong, and Thomas Tao Yang Boston College, University of California Irvine, and Boston

More information

Two-Sample T-Tests Allowing Unequal Variance (Enter Difference)

Two-Sample T-Tests Allowing Unequal Variance (Enter Difference) Chapter 45 Two-Sample T-Tests Allowing Unequal Variance (Enter Difference) Introduction This procedure provides sample size and power calculations for one- or two-sided two-sample t-tests when no assumption

More information

1 Teaching notes on GMM 1.

1 Teaching notes on GMM 1. Bent E. Sørensen January 23, 2007 1 Teaching notes on GMM 1. Generalized Method of Moment (GMM) estimation is one of two developments in econometrics in the 80ies that revolutionized empirical work in

More information

Introduction to Matrix Algebra

Introduction to Matrix Algebra Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary

More information

7 Time series analysis

7 Time series analysis 7 Time series analysis In Chapters 16, 17, 33 36 in Zuur, Ieno and Smith (2007), various time series techniques are discussed. Applying these methods in Brodgar is straightforward, and most choices are

More information

Frictional Matching: Evidence from Law School Admission

Frictional Matching: Evidence from Law School Admission Frictional Matching: Evidence from Law School Admission Pascal Courty Mario Pagliero No. 113 June 2009 www.carloalberto.org/working_papers 2009 by Pascal Courty and Mario Pagliero. Any opinions expressed

More information

Statistics in Retail Finance. Chapter 6: Behavioural models

Statistics in Retail Finance. Chapter 6: Behavioural models Statistics in Retail Finance 1 Overview > So far we have focussed mainly on application scorecards. In this chapter we shall look at behavioural models. We shall cover the following topics:- Behavioural

More information

ASSIGNMENT 4 PREDICTIVE MODELING AND GAINS CHARTS

ASSIGNMENT 4 PREDICTIVE MODELING AND GAINS CHARTS DATABASE MARKETING Fall 2015, max 24 credits Dead line 15.10. ASSIGNMENT 4 PREDICTIVE MODELING AND GAINS CHARTS PART A Gains chart with excel Prepare a gains chart from the data in \\work\courses\e\27\e20100\ass4b.xls.

More information

Parameter estimation for nonlinear models: Numerical approaches to solving the inverse problem. Lecture 12 04/08/2008. Sven Zenker

Parameter estimation for nonlinear models: Numerical approaches to solving the inverse problem. Lecture 12 04/08/2008. Sven Zenker Parameter estimation for nonlinear models: Numerical approaches to solving the inverse problem Lecture 12 04/08/2008 Sven Zenker Assignment no. 8 Correct setup of likelihood function One fixed set of observation

More information