Alpaydin E. Introduction to Machine Learning. MIT Press, 2004.
|
|
|
- Morgan Walton
- 10 years ago
- Views:
Transcription
1 Bibliography: Evaluating Predictive Models Last update: 29 July 2007 General Alpaydin E. Introduction to Machine Learning. MIT Press, An excellent introduction to the field of machine learning. As with most books on machine learning, the emphasis is on classification. Chapter 14 describes the assessment and comparison of classification algorithms. Altman DG, Royston P. What do we mean by validating a prognostic model? Statistics in Medicine 2000; 19: This paper examines (i) what is meant by validation of prognostic models, (ii) reviews why it is necessary, and (iii) describes how validations should be carried out; The emphasis is on conceptual rather than technical issues. The idea of validating a prognostic model is generally taken to mean that it works satisfactorily for patients other than those from whose data the model was derived. The authors suggest that it is desirable to distinguish statistical from clinical validity. Statistical validity means that the model is the best that can be found with the available factors, while clinical validity means that the model predicts accurately enough for its purpose of course, this depends crucially on one's view of the aims of the model. The paper spents considerable attention to the problem of overfitting. It is known that analyses that are not prespecified but are data-dependent are liable to overoptimistic conclusions. The data-dependent aspect of most prognostic models stems from the variable selection and discretization procedures. Duda RO, Hart PE, Stork DG, Pattern Classification. Wiley, This classic textbook from 1973 was revised and updated in It covers a broad range of pattern classification techniques. Chapter 2 discusses Bayesian decision theory. Hand DJ, Construction and Assessment of Classification Rules. Wiley, Chapter 6 of this book (Aspects of evaluation) presents a framework for understanding model evaluation concepts. A distinction is made between measuring accuracy, precision, separability, and resemblance. A large number of examples is provided, e.g. the misclassification (i.e. error) rate and the Brier score are accuracy measures. The highly popular misclassification rate is further investigated in Chapter 7, where several crossvalidation schemes (rotation, leave-one-out, bootstrap) for estimating the actual error are discussed. Chapter 7 deals extensively with aspects of classification accuracy. Hastie T, Tibshirani R, Friedman J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction. New York: Springer-Verlag, Chapter 7 of this excellent textbook on statistical learning methods discusses model assessment and selection methods based on loss (accuracy) functions. Mitchell TM. Machine Learning. McGraw-Hill, Another very good textbook on machine learning. Chapter 5 considers classification accuracy, including the statistical estimation of performance.
2 Pepe MS, The Statistical Evaluation of Medical Tests for Classification and Prediction. Oxford University Press, Wasson JH, Sox HC, Neff RK, Goldman L. Clinical prediction rules. New England Journal of Medicine 1985; 313: In this article, the following methodological standards for creating and validating clinical prediction rules are proposed: 1. The event to be predicted (outcome) should be clearly defined, preferrably by biological rather than sociological or behavioral criteria; 2. Predictive findings should be defined precisely and have a similar meaning to anyone who may use them; 3. The list of predictors should not include any criteria that are used in defining the outcome ("blind assessment"; most relevant for diagnostic rules); 4. Characterics of the patient population used to develop the rule should be clearly described; 5. The study site and type of practice where the data was gathered should be described; 6. An unbiased estimate of the rule's performance should be reported; 7. Effects of using the rule should be prospectively measured; and 8. The statistical technique that was used to derive the rule should be described. Thirty-three publications of clinical predictions rules during the years in four leading medical journals were reviewed by these standards. Most of the criteria were met by more than 80% of the studies. However, performance statistics were seldomly reported (11 publications), and effects of clinical use was almost never prospectively measured (2 publications). Ch. 1 Introduction: Predictive Models and Evaluation Abu-Hanna A, Lucas PJF. Prognostic models in medicine, Methods of Information in Medicine 2001; 40: 1 5. Wyatt JC, Altman DG. Commentary: Prognostic models: clinically useful of quickly forgotten? British Medical Journal 1995; 311: Few prognostic models are routinely used to inform difficult clinical decisions. Wyatt and Altman believe that the main reasons why doctors reject published prognostic models are lack of clinical credibility and lack of evidence that a prognostic model can support decisions about patient care (that is, evidence of accuracy, generality, and effectiveness). Ch. 3 Evaluating Probabilities (i) ROC analysis Bamber D. The area above the ordinal dominance graph and the area below the receiver operating characteristic graph. Journal of Mathematical Psychology 1975; 12: In this paper it is shown that the area under the ROC curve (AUC) equals the probability that a randomly chosen positive case was given a higher test value (or higher prediction by a model) than a randomly chosen negative case. Furthermore, it is shown that the estimated AUC is equivalent to the Mann-Whitney U statistics normalized by the number of pairs of negative and positive cases.
3 DeLong ER, DeLong DM, Clarke-Pearson DL. Comparing the areas under two or more correlated receiver operating characteristic curves: a nonparametric approach, Biometrics 1988; 44: A nonparametric method for comparing the areas under the ROC curves of two distinct models on the same dataset. The method is based on the theory of generalized U statistics. Hand DJ, Till RJ. A simple generalization of the area under the ROC curve to multiple class classification problems, Machine Learning 2001; 45(2): Hanley JA, McNeil BJ. The meaning and use of the area under the receiver operating characteristic (ROC) curve. Radiology 1982; 143: Hanley JA, McNeil BJ. A method of comparing the areas under receiver operating characteristic curves derived from the same cases. Radiology 1983; 148: Parametric method for comparing the areas under the ROC curves of two distinct models on the same dataset, assuming a Normal (Gaussian) distribution of the AUC. Superseded by the nonparametric method of DeLong et al. (1988). Lasko TA, Bhagwat JG, Zou KH, Ohno-Machado L. The use of receiver operating characteristic curves in biomedical informatics. Journal of Biomedical Informatics 2005; 38(5): An overview of different methods to estimate and compare areas on the ROC curve, and software packages available for ROC analysis. The paper does not present new methods but summarizes the existing literature. The approach is practical, and aims at readers who need to choose between applying different methods. Metz CE, Basic principles of ROC analysis, Seminars in Nuclear Medicine 1978; 8(4): Charles Metz was one of the people who popularized the use of ROC analysis in medical research. Provost F, Fawcett T, Kohavi R. The case against accuracy estimation for comparing induction algorithms. Proceedings of the 15th International Conference on Machine Learning (ICML 98), pp , Machine Learning research has traditionally concentrated on designing algorithms for building classifier functions, and the predominant evaluation methodology in this field is classification accuracy (error rate) estimation. In this influential paper, the authors argue that estimating classification accuracy is insufficient for comparing competing classifiers and algorithms, because classification accuracy assumes equal misclassification costs and a known marginal class distribution. However, both misclassification costs and marginal class distribution can be unknown at the time of building the model and may even vary from time to time, place to place, and situation to situation where the model is applied. For these reasons, the authors argue that classifiers and induction algorithms should be evaluated and compared using ROC analyses. Using ten datasets from the UCI repository and several standard machine learning algorithms, it is shown that high classification accuracy does not imply domination in ROC space. Therefore, comparing accuracies on benchmark datasets says little, if anything, about classifier performance on real-world tasks. (Note: the authors do not mention the fact that a model may be very imprecise (badly calibrated) even though it dominates other models in ROC space. This is an imperfection of ROC analysis.)
4 Somers RH. A new asymmetric measure of association for ordinal variables, American Sociological Review 1962; 27: Somers' rank correlation D xy is a nonparametric measure of association between ordinal variables, and is related to the concordance index C (= nonparametric AUC) as follows: D xy = 2(C 0.5). Ch. 3 Evaluating Probabilities (ii) Accuracy of probabilities Ash A, Shwartz M. R 2 : a useful measure of model performance when predicting a dichotomous outcome. Statistics in Medicine 1999; 18: Brier GW. Verification of weather forecasts expressed in terms of probability. Monthly Weather Review 1950; 78: 1 3. The principal reference for the Brier inaccuracy score. Cowell RG, Dawid AP, Lauritzen SL, Spiegelhalter DJ. Probabilistic Networks and Expert Systems. Berlin: Springer-Verlag, Chapter 10 of this textbook on probabilistic network models considers the problem of checking models against data. Although some of the methods are specific for Bayesian networks, others are general (Bayesian) statistical tools for evaluating predictive models. Specific attention is spent to the logarithmic score (deviance). Mittlböck M, Schemper M. Explained variation for logistic regression. Statistics in Medicine 1996; 15: Mittlböck and Schemper review 12 statistical measures that have been proposed to quantify the explained variation of binary predictive model (in contrast to what the title suggests, none of the measures is restricted to use in conjunction with logistic regression). Six measures are based on the correlation of estimated probabilities and observed outcome (e.g. Pearson correlation and Somers D), four are based on reduction in dispersion of the outcome (e.g. sum-of-squares R 2, Gini index, classification error), and two are based on model likelihood (likelihood-ratio and Nagelkerke R 2 ). Nagelkerke NJD. A note on a general definition of the coefficient of determination. Biometrika 1991; 78(3): Nagelkerke proposes to use the ratio of log likelihoods of a binary predictive model (e.g. logistic regression) and the 'null' (intercept only) model as a measure of explained variation by the model. This performance statistic is presented by some statistical packages (e.g. SAS) as 'Nagelkerke R 2 '. Although the statistic has some attractive properties (e.g. consistency with classical R 2, consistency with maximum likelihood estimation, independent of sample size), there are serious problems with its interpretation (e.g. see Mitlböck and Schemper, 1996). Redelmeier DA, Bloch DA, Hickam DH. Assessing predictive accuracy: How to compare Brier scores. Journal of Clinical Epidemiology 1991; 44(11): This paper presents a statistical method to compare the Brier scores from two different sets of predictive assessments (predicted probabilities) on a single test set. The method is an extension of Spiegelhalter's test whether a given Brier score is incompatable with the
5 observed outcomes. A problem with the comparison method is that the test statistic depends on the true, unknown probabilities. To solve this problem, the authors suggest to use the mean of both predictions. The paper contains a small example based on probability judgements of five medical students whom independently reviewed the symptoms and elektrocardiograms of 25 patients with recurrent chest pain. Ch. 6 Assessing the fit of a Model le Cessie S, van Houwelingen JC. A goodness-of-fit test for binary regression models, based on smoothing methods. Biometrics 1991; 47: Hosmer DW, Hosmer T, le Cessie S, Lemeshow S. A comparison of goodness-of-fit tests for the logistic regression model. Statistics in Medicine 1997; 16: Experimental comparison of several goodness-of-fit tests for the logistic regression model. Hosmer DW, Lemeshow S. Goodness-of-fit tests for the multiple logistic regression model. Communications in Statistics 1980; A10: This article presents the original Hosmer-Lemeshow goodness-of-fit test for logistic regression model. The test is based on a statistic C that sums squared Pearson residuals in g (usually 10) risk groups, where the grouping is either based on fixed values of the estimated probabilities or on percentiles of the estimated probabilities (the latter approach is often preferrable). It was experimentally shown that the distribution of the statistic C is well approximated by the χ 2 distribution with g λ 1 degrees of freedom, where λ =1 if C is computed from the training data set, and λ=0 otherwise. If C is large, this indicates that, at least in part of the feature space, the estimated probabilities strongly deviate from the true probabilities. The Hosmer-Lemeshow goodness-of-fit test as described here was implemented in many statistical packages and is routinely applied in epidemiological research, even though it was later shown that the statistic may be unstable and the test is therefore unreliable. Hosmer DW, Lemeshow S. Applied Logistic Regression. New York: John Wiley & Sons, 2nd edition, Chapter 5 of this textbook deals extensively with the evaluation of logistic regression models. The evaluation methods described are partly dedicated for logistic regession models (e.g. goodness-of-fit tests) and partly generic (e.g. ROC analysis). Miller ME, Langefeld CD, Tierney WM, Hui SL, McDonald CJ. Validation of probabilistic predictions. Medical Decision Making 1993; 13(1): Moons E, Aerts M, Wets G. A tree-based lack-of-fit for multiple logistic regression. Statistics in Medicine 2004; 23: Ch. 7 Model Validation Bleeker SE, Moll HA, Steyerberg EW, Donders AR, Derksen-Lubsen G, Grobbee DE, Moons KG. External validation is necessary in prediction research: a clinical example. Journal of Clinical Epidemiology 2003; 56(9):
6 Case study in pediatric diagnostic management (predicting bacterial infections) shows that for relatively small data sets, internal validation of prediction models by bootstrap techniques may not be sufficient and indicative for the model's performance in future patients. External validation is essential before implementing prediction models in clinical practice. Davison AC, Hinkley DV. Bootstrap Methods and their Application. Cambridge: Cambridge University Press, Gant V, Rodway S, Wyatt JC. Artificial neural networks: practical considerations for clinical applications. In Clinical Applications of Artificial Neural Networks (Dybowski R, Gant V, eds.), Cambridge: Cambridge University Press, 2001, pp Hadorn DC, Draper D, Rogers WH, Keeler EB, Brook RH. Cross-validation performance of mortality prediction models. Statistics in Medicine 1992; 11(4): Early study on the performance of different modelling techniques (linear regression, logistic regression, Cox regression, CART) in predicting death after acute myocardial infarction. Similar, but more rigorous, studies were conducted by Steyerberg et al. Justice AC, Covinsky KE, Berlin JA. Assessing the generalizability of prognostic information. Annals of Internal Medicine 1999; 130: Miller ME, Hui SL, Tierney WM. Validation techniques for logistic regression models. Statistics in Medicine 1991; 10(8): This paper presents a comprehensive approach to the validation of logistic prediction models. It reviews measures of overall goodness-of-fit, and indices of calibration and refinement. Using a model-based approach developed by Cox, logistic regression diagnostic techniques are adapted for use in model validation. This allows identification of problematic predictor variables in the prediction model as well as influential observations in the validation data that adversely affect the fit of the model. In appropriate situations, recommendations are made for correction of models that provide poor fit. Peek N, Arts DG, Bosman RJ, Van der Voort PH, De Keizer NF. External validation of prognostic models for critically ill patients required substantial sample sizes. Journal of Clinical Epidemiology 2007; 60(5): This study considers the behavior of predictive performance measures that are commonly used in external validation of prognostic models. A resampling scheme was used to investigate the effects of sample size; the domain of application was intensive care. The AUC and Brier score showed large variation with small samples. It was found that substantial sample sizes are required for performance assessment and model comparison in external validation. Standard errors of AUC values were accurate but the power to detect differences in performance was low. Calibration statistics and the associated significance tests are extremely sensitive to sample size, and should not be used in these settings. Instead, D. Cox customization method to repair lack-of-fit problems is recommended. Direct comparison of performance, without statistical analysis, was unreliable with either measure.
7 Schwarzer G, Vach W, Schumacher M. On the misuses of artificial neural networks for prognostic and diagnostic classification in oncology. Statistics in Medicine 2000; 19: Schwarzer et al. present a critical review of applications of artificial neural networks (ANNs) in biomedicine. The flexibility of ANNs is often cited as an advantage, but the authors argue that it must be seen as a major concern. Several common pitfalls are discussed (e.g. fitting implausible functions, incorrect modelling of survival data, and biased estimation of network accuracy), and a review of the literature of ANN applications in oncology is presented. Many of the 43 applications that are discussed show (severe) methodological weaknesses. Steyerberg EW, Harrell FE, Borsboom GJJM, Eijkemans MJC, Vergouwe Y, Habbema JDF, Internal validation of predictive models: efficiency of some procedures for logistic regression analysis, Journal of Clinical Epidemiology 2001; 54(8): The performance of a predictive model is overestimated when simply determined on the sample of subjects that was used to construct the model. Several internal validation methods are available that aim to provide a more accurate estimate of model performance in new subjects. This study evaluated several variants of split-sample, cross-validation and bootstrapping methods with a logistic regression model that included eight predictors for 30-day mortality after an acute myocardial infarction. Random samples of varying size were drawn from a large data set. Split-sample analyses gave overly pessimistic estimates of performance, with large variability. Cross-validation on 10% of the sample had low bias and low variability, but was not suitable for all performance measures. Internal validity could best be estimated with bootstrapping, which provided stable estimates with low bias. Steyerberg EW, Bleeker SE, Moll HA, Grobbee DE, Moons KG. Internal and external validation of predictive models: a simulation study of bias and precision in small samples. Journal of Clinical Epidemiology 2003; 56(5): Simulation study to investigate the accuracy of bootstrap estimates of optimism (internal validation) and the precision of performance estimates in independent validation samples (external validation). Random samples were drawn from a data set on infectious diseases in children, for the development (n=376) and validation (n=179) of logistic regression models. Model development, including the selection of predictors, and validation were repeated in a bootstrapping procedure. The average apparent ROC area was 0.74, which was expected (based on bootstrapping) to decrease by 0.07 to 0.67, whereas the observed decrease in the validation samples was 0.09 to Omitting the selection of predictors from the bootstrap procedure led to a severe underestimation of the optimism (decrease 0.006). The standard error of the observed ROC area in the independent validation samples was large (0.05). So, for external validation, substantial sample sizes should be used for sufficient power to detect clinically important changes in performance as compared with the internally validated estimate. Steyerberg EW, Borsboom GJ, van Houwelingen HC, Eijkemans MJ, Habbema JD. Validation and updating of predictive logistic regression models: a study on sample size and shrinkage. Statistics in Medicine 2004; 23(16): A logistic regression model may be used to provide predictions of outcome for individual patients at another centre than where the model was developed. When empirical data are available from this centre, the validity of predictions can be assessed by comparing observed outcomes and predicted probabilities. Subsequently, the model may be updated to improve predictions for future patients. In this study, a previously published model for
8 predicting 30-day mortality after acute myocardial infarction was validated and updated with external validation samples that varied in size. Heuristic shrinkage approaches were applied in the model revision methods, such that regression coefficients were shrunken towards their re-calibrated values. Parsimonious updating methods were found preferable to more extensive model revisions, which should only be attempted with relatively large validation samples in combination with shrinkage. Terrin N, Schmid CH, Griffith JL, D'Agostino RB, Selker HP. External validity of predictive models: a comparison of logistic regression, classification trees, and neural networks. Journal of Clinical Epidemiology 2003; 56(8): Simulation study that compared the external validity of standard logistic regression (LR1), logistic regression with piecewise-linear and quadratic terms (LR2), classification trees, and neural networks (NNETs). Predictive models were developed on data simulated from a specified population and on data from perturbed forms of the population not representative of the original distribution. All models were tested on new data generated from the population. The performance of LR2 was superior to that of the other model types when the models were developed on data sampled from the population and when they were developed on nonrepresentative data. However, when the models developed using nonrepresentative data were compared with models developed from data sampled from the population, LR2 had the greatest loss in performance. These results highlight the necessity of external validation to test the transportability of predictive models. Vergouwe Y, Steyerberg, EW, Eijkemans MJC, Habbema JDF. Validity of prognostic models: When is a model clinically useful? Seminars in Urologic Oncology 2002; 20(2): Vergouwe et al. distinguish three aspects of validity of prognostic models: (1) agreement between predicted probabilities and observed probabilities (calibration), (2) ability of the model to distinguish subjects with different outcomes (discrimination), and (3) ability of the model to improve the decision-making process (clinical usefulness). Several techniques for visualizing and quantifiying calibration and discrimination are discussed. Clinical usefulness is inspected by considering classification accuracy, sensitivity, and specificity of the model (after choosing a classification threshold), and by estimating the expected decrease in disutility when the model is applied in practice. This is done by comparing the model s classifications and conventional policy by weighing false-positive and falsenegative classified patients according to relative severity. Vergouwe Y, Steyerberg EW, Eijkemans MJ, Habbema JD. Substantial effective sample sizes were required for external validation studies of predictive logistic regression models. Journal of Clinical Epidemiology 2005 ;58(5): Simulation study in the field of oncology (predicting the probability that residual masses of patients treated for metastatic testicular cancer contained only benign tissue) suggests that a minimum of 100 events and 100 nonevents are required for external validation samples. Zhu B-P, Lemeshow S, Hosmer DW, Klar J, Avrunin J, Teres D. Factors affecting the performance of the models in the Mortality Probability Model II system and strategies of customization: A simulation study. Critical Care Medicine 1996; 24:57 63.
Validity of prediction models: when is a model clinically useful?
Chapter 2 Validity of prediction models: when is a model clinically useful? Abstract Prediction models combine patient characteristics to predict medical outcomes. Unfortunately, such models do not always
STATISTICA Formula Guide: Logistic Regression. Table of Contents
: Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary
Adequacy of Biomath. Models. Empirical Modeling Tools. Bayesian Modeling. Model Uncertainty / Selection
Directions in Statistical Methodology for Multivariable Predictive Modeling Frank E Harrell Jr University of Virginia Seattle WA 19May98 Overview of Modeling Process Model selection Regression shape Diagnostics
Internal validation of predictive models: Efficiency of some procedures for logistic regression analysis
Journal of Clinical Epidemiology 54 (2001) 774 781 Internal validation of predictive models: Efficiency of some procedures for logistic regression analysis Ewout W. Steyerberg a, *, Frank E. Harrell Jr
270107 - MD - Data Mining
Coordinating unit: Teaching unit: Academic year: Degree: ECTS credits: 015 70 - FIB - Barcelona School of Informatics 715 - EIO - Department of Statistics and Operations Research 73 - CS - Department of
Data Mining - Evaluation of Classifiers
Data Mining - Evaluation of Classifiers Lecturer: JERZY STEFANOWSKI Institute of Computing Sciences Poznan University of Technology Poznan, Poland Lecture 4 SE Master Course 2008/2009 revised for 2010
Advanced Quantitative Methods for Health Care Professionals PUBH 742 Spring 2015
1 Advanced Quantitative Methods for Health Care Professionals PUBH 742 Spring 2015 Instructor: Joanne M. Garrett, PhD e-mail: [email protected] Class Notes: Copies of the class lecture slides
D-optimal plans in observational studies
D-optimal plans in observational studies Constanze Pumplün Stefan Rüping Katharina Morik Claus Weihs October 11, 2005 Abstract This paper investigates the use of Design of Experiments in observational
11. Analysis of Case-control Studies Logistic Regression
Research methods II 113 11. Analysis of Case-control Studies Logistic Regression This chapter builds upon and further develops the concepts and strategies described in Ch.6 of Mother and Child Health:
On the effect of data set size on bias and variance in classification learning
On the effect of data set size on bias and variance in classification learning Abstract Damien Brain Geoffrey I Webb School of Computing and Mathematics Deakin University Geelong Vic 3217 With the advent
Development and validation of a prediction model with missing predictor data: a practical approach
Journal of Clinical Epidemiology 63 (2010) 205e214 Development and validation of a prediction model with missing predictor data: a practical approach Yvonne Vergouwe a, *, Patrick Royston b, Karel G.M.
Data Mining Practical Machine Learning Tools and Techniques
Ensemble learning Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter 8 of Data Mining by I. H. Witten, E. Frank and M. A. Hall Combining multiple models Bagging The basic idea
Cross-Validation. Synonyms Rotation estimation
Comp. by: BVijayalakshmiGalleys0000875816 Date:6/11/08 Time:19:52:53 Stage:First Proof C PAYAM REFAEILZADEH, LEI TANG, HUAN LIU Arizona State University Synonyms Rotation estimation Definition is a statistical
CS 2750 Machine Learning. Lecture 1. Machine Learning. http://www.cs.pitt.edu/~milos/courses/cs2750/ CS 2750 Machine Learning.
Lecture Machine Learning Milos Hauskrecht [email protected] 539 Sennott Square, x5 http://www.cs.pitt.edu/~milos/courses/cs75/ Administration Instructor: Milos Hauskrecht [email protected] 539 Sennott
Analyzing PETs on Imbalanced Datasets When Training and Testing Class Distributions Differ
Analyzing PETs on Imbalanced Datasets When Training and Testing Class Distributions Differ David Cieslak and Nitesh Chawla University of Notre Dame, Notre Dame IN 46556, USA {dcieslak,nchawla}@cse.nd.edu
Cross Validation techniques in R: A brief overview of some methods, packages, and functions for assessing prediction models.
Cross Validation techniques in R: A brief overview of some methods, packages, and functions for assessing prediction models. Dr. Jon Starkweather, Research and Statistical Support consultant This month
Data quality in Accounting Information Systems
Data quality in Accounting Information Systems Comparing Several Data Mining Techniques Erjon Zoto Department of Statistics and Applied Informatics Faculty of Economy, University of Tirana Tirana, Albania
Regression Modeling Strategies
Frank E. Harrell, Jr. Regression Modeling Strategies With Applications to Linear Models, Logistic Regression, and Survival Analysis With 141 Figures Springer Contents Preface Typographical Conventions
Interpretation of Somers D under four simple models
Interpretation of Somers D under four simple models Roger B. Newson 03 September, 04 Introduction Somers D is an ordinal measure of association introduced by Somers (96)[9]. It can be defined in terms
Statistics Graduate Courses
Statistics Graduate Courses STAT 7002--Topics in Statistics-Biological/Physical/Mathematics (cr.arr.).organized study of selected topics. Subjects and earnable credit may vary from semester to semester.
THE HYBRID CART-LOGIT MODEL IN CLASSIFICATION AND DATA MINING. Dan Steinberg and N. Scott Cardell
THE HYBID CAT-LOGIT MODEL IN CLASSIFICATION AND DATA MINING Introduction Dan Steinberg and N. Scott Cardell Most data-mining projects involve classification problems assigning objects to classes whether
X X X a) perfect linear correlation b) no correlation c) positive correlation (r = 1) (r = 0) (0 < r < 1)
CORRELATION AND REGRESSION / 47 CHAPTER EIGHT CORRELATION AND REGRESSION Correlation and regression are statistical methods that are commonly used in the medical literature to compare two or more variables.
Feature vs. Classifier Fusion for Predictive Data Mining a Case Study in Pesticide Classification
Feature vs. Classifier Fusion for Predictive Data Mining a Case Study in Pesticide Classification Henrik Boström School of Humanities and Informatics University of Skövde P.O. Box 408, SE-541 28 Skövde
Critical Appraisal and Data Extraction for Systematic Reviews of Prediction Modelling Studies: The CHARMS Checklist
Guidelines and Guidance Critical Appraisal and Data Extraction for Systematic Reviews of Prediction Modelling Studies: The CHARMS Checklist Karel G. M. Moons 1" *, Joris A. H. de Groot 1", Walter Bouwmeester
SAS Software to Fit the Generalized Linear Model
SAS Software to Fit the Generalized Linear Model Gordon Johnston, SAS Institute Inc., Cary, NC Abstract In recent years, the class of generalized linear models has gained popularity as a statistical modeling
MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS
MULTIPLE REGRESSION AND ISSUES IN REGRESSION ANALYSIS MSR = Mean Regression Sum of Squares MSE = Mean Squared Error RSS = Regression Sum of Squares SSE = Sum of Squared Errors/Residuals α = Level of Significance
Predictive Data modeling for health care: Comparative performance study of different prediction models
Predictive Data modeling for health care: Comparative performance study of different prediction models Shivanand Hiremath [email protected] National Institute of Industrial Engineering (NITIE) Vihar
Penalized Logistic Regression and Classification of Microarray Data
Penalized Logistic Regression and Classification of Microarray Data Milan, May 2003 Anestis Antoniadis Laboratoire IMAG-LMC University Joseph Fourier Grenoble, France Penalized Logistic Regression andclassification
Classification of Bad Accounts in Credit Card Industry
Classification of Bad Accounts in Credit Card Industry Chengwei Yuan December 12, 2014 Introduction Risk management is critical for a credit card company to survive in such competing industry. In addition
Chapter 6. The stacking ensemble approach
82 This chapter proposes the stacking ensemble approach for combining different data mining classifiers to get better performance. Other combination techniques like voting, bagging etc are also described
Gerry Hobbs, Department of Statistics, West Virginia University
Decision Trees as a Predictive Modeling Method Gerry Hobbs, Department of Statistics, West Virginia University Abstract Predictive modeling has become an important area of interest in tasks such as credit
LOGISTIC REGRESSION ANALYSIS
LOGISTIC REGRESSION ANALYSIS C. Mitchell Dayton Department of Measurement, Statistics & Evaluation Room 1230D Benjamin Building University of Maryland September 1992 1. Introduction and Model Logistic
Comparing the Results of Support Vector Machines with Traditional Data Mining Algorithms
Comparing the Results of Support Vector Machines with Traditional Data Mining Algorithms Scott Pion and Lutz Hamel Abstract This paper presents the results of a series of analyses performed on direct mail
Curriculum Map Statistics and Probability Honors (348) Saugus High School Saugus Public Schools 2009-2010
Curriculum Map Statistics and Probability Honors (348) Saugus High School Saugus Public Schools 2009-2010 Week 1 Week 2 14.0 Students organize and describe distributions of data by using a number of different
An Analysis of Missing Data Treatment Methods and Their Application to Health Care Dataset
P P P Health An Analysis of Missing Data Treatment Methods and Their Application to Health Care Dataset Peng Liu 1, Elia El-Darzi 2, Lei Lei 1, Christos Vasilakis 2, Panagiotis Chountas 2, and Wei Huang
Ensemble Methods. Knowledge Discovery and Data Mining 2 (VU) (707.004) Roman Kern. KTI, TU Graz 2015-03-05
Ensemble Methods Knowledge Discovery and Data Mining 2 (VU) (707004) Roman Kern KTI, TU Graz 2015-03-05 Roman Kern (KTI, TU Graz) Ensemble Methods 2015-03-05 1 / 38 Outline 1 Introduction 2 Classification
Guide to Biostatistics
MedPage Tools Guide to Biostatistics Study Designs Here is a compilation of important epidemiologic and common biostatistical terms used in medical research. You can use it as a reference guide when reading
Evaluation of Predictive Models
Evaluation of Predictive Models Assessing calibration and discrimination Examples Decision Systems Group, Brigham and Women s Hospital Harvard Medical School Harvard-MIT Division of Health Sciences and
College Readiness LINKING STUDY
College Readiness LINKING STUDY A Study of the Alignment of the RIT Scales of NWEA s MAP Assessments with the College Readiness Benchmarks of EXPLORE, PLAN, and ACT December 2011 (updated January 17, 2012)
Decision Trees from large Databases: SLIQ
Decision Trees from large Databases: SLIQ C4.5 often iterates over the training set How often? If the training set does not fit into main memory, swapping makes C4.5 unpractical! SLIQ: Sort the values
Statistics in Retail Finance. Chapter 6: Behavioural models
Statistics in Retail Finance 1 Overview > So far we have focussed mainly on application scorecards. In this chapter we shall look at behavioural models. We shall cover the following topics:- Behavioural
Better credit models benefit us all
Better credit models benefit us all Agenda Credit Scoring - Overview Random Forest - Overview Random Forest outperform logistic regression for credit scoring out of the box Interaction term hypothesis
COURSE PLAN BDA: Biomedical Data Analysis Master in Bioinformatics for Health Sciences. 2015-2016 Academic Year Qualification.
COURSE PLAN BDA: Biomedical Data Analysis Master in Bioinformatics for Health Sciences 2015-2016 Academic Year Qualification. Master's Degree 1. Description of the subject Subject name: Biomedical Data
203.4770: Introduction to Machine Learning Dr. Rita Osadchy
203.4770: Introduction to Machine Learning Dr. Rita Osadchy 1 Outline 1. About the Course 2. What is Machine Learning? 3. Types of problems and Situations 4. ML Example 2 About the course Course Homepage:
REPORT DOCUMENTATION PAGE
REPORT DOCUMENTATION PAGE Form Approved OMB NO. 0704-0188 Public Reporting burden for this collection of information is estimated to average 1 hour per response, including the time for reviewing instructions,
Another Look at Sensitivity of Bayesian Networks to Imprecise Probabilities
Another Look at Sensitivity of Bayesian Networks to Imprecise Probabilities Oscar Kipersztok Mathematics and Computing Technology Phantom Works, The Boeing Company P.O.Box 3707, MC: 7L-44 Seattle, WA 98124
Comparison of Data Mining Techniques used for Financial Data Analysis
Comparison of Data Mining Techniques used for Financial Data Analysis Abhijit A. Sawant 1, P. M. Chawan 2 1 Student, 2 Associate Professor, Department of Computer Technology, VJTI, Mumbai, INDIA Abstract
Examining a Fitted Logistic Model
STAT 536 Lecture 16 1 Examining a Fitted Logistic Model Deviance Test for Lack of Fit The data below describes the male birth fraction male births/total births over the years 1931 to 1990. A simple logistic
Risk pricing for Australian Motor Insurance
Risk pricing for Australian Motor Insurance Dr Richard Brookes November 2012 Contents 1. Background Scope How many models? 2. Approach Data Variable filtering GLM Interactions Credibility overlay 3. Model
Model Validation Techniques
Model Validation Techniques Kevin Mahoney, FCAS kmahoney@ travelers.com CAS RPM Seminar March 17, 2010 Uses of Statistical Models in P/C Insurance Examples of Applications Determine expected loss cost
Logistic Regression. http://faculty.chass.ncsu.edu/garson/pa765/logistic.htm#sigtests
Logistic Regression http://faculty.chass.ncsu.edu/garson/pa765/logistic.htm#sigtests Overview Binary (or binomial) logistic regression is a form of regression which is used when the dependent is a dichotomy
Predicting Bankruptcy of Manufacturing Firms
Predicting Bankruptcy of Manufacturing Firms Martin Grünberg and Oliver Lukason Abstract This paper aims to create prediction models using logistic regression and neural networks based on the data of Estonian
Statistical Rules of Thumb
Statistical Rules of Thumb Second Edition Gerald van Belle University of Washington Department of Biostatistics and Department of Environmental and Occupational Health Sciences Seattle, WA WILEY AJOHN
Students' Opinion about Universities: The Faculty of Economics and Political Science (Case Study)
Cairo University Faculty of Economics and Political Science Statistics Department English Section Students' Opinion about Universities: The Faculty of Economics and Political Science (Case Study) Prepared
Local classification and local likelihoods
Local classification and local likelihoods November 18 k-nearest neighbors The idea of local regression can be extended to classification as well The simplest way of doing so is called nearest neighbor
Solving Regression Problems Using Competitive Ensemble Models
Solving Regression Problems Using Competitive Ensemble Models Yakov Frayman, Bernard F. Rolfe, and Geoffrey I. Webb School of Information Technology Deakin University Geelong, VIC, Australia {yfraym,brolfe,webb}@deakin.edu.au
Predictive Modeling Techniques in Insurance
Predictive Modeling Techniques in Insurance Tuesday May 5, 2015 JF. Breton Application Engineer 2014 The MathWorks, Inc. 1 Opening Presenter: JF. Breton: 13 years of experience in predictive analytics
Direct Marketing Response Models using Genetic Algorithms
From: KDD-98 Proceedings. Copyright 1998, AAAI (www.aaai.org). All rights reserved. Direct Marketing Response Models using Genetic Algorithms Siddhartha Bhattacharyya Information and Decision Sciences
Why High-Order Polynomials Should Not be Used in Regression Discontinuity Designs
Why High-Order Polynomials Should Not be Used in Regression Discontinuity Designs Andrew Gelman Guido Imbens 2 Aug 2014 Abstract It is common in regression discontinuity analysis to control for high order
Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
Service courses for graduate students in degree programs other than the MS or PhD programs in Biostatistics.
Course Catalog In order to be assured that all prerequisites are met, students must acquire a permission number from the education coordinator prior to enrolling in any Biostatistics course. Courses are
SPSS TRAINING SESSION 3 ADVANCED TOPICS (PASW STATISTICS 17.0) Sun Li Centre for Academic Computing [email protected]
SPSS TRAINING SESSION 3 ADVANCED TOPICS (PASW STATISTICS 17.0) Sun Li Centre for Academic Computing [email protected] IN SPSS SESSION 2, WE HAVE LEARNT: Elementary Data Analysis Group Comparison & One-way
Chapter 11 Boosting. Xiaogang Su Department of Statistics University of Central Florida - 1 -
Chapter 11 Boosting Xiaogang Su Department of Statistics University of Central Florida - 1 - Perturb and Combine (P&C) Methods have been devised to take advantage of the instability of trees to create
Identifying At-Risk Students Using Machine Learning Techniques: A Case Study with IS 100
Identifying At-Risk Students Using Machine Learning Techniques: A Case Study with IS 100 Erkan Er Abstract In this paper, a model for predicting students performance levels is proposed which employs three
S03-2008 The Difference Between Predictive Modeling and Regression Patricia B. Cerrito, University of Louisville, Louisville, KY
S03-2008 The Difference Between Predictive Modeling and Regression Patricia B. Cerrito, University of Louisville, Louisville, KY ABSTRACT Predictive modeling includes regression, both logistic and linear,
BOOSTING - A METHOD FOR IMPROVING THE ACCURACY OF PREDICTIVE MODEL
The Fifth International Conference on e-learning (elearning-2014), 22-23 September 2014, Belgrade, Serbia BOOSTING - A METHOD FOR IMPROVING THE ACCURACY OF PREDICTIVE MODEL SNJEŽANA MILINKOVIĆ University
Predicting the Risk of Heart Attacks using Neural Network and Decision Tree
Predicting the Risk of Heart Attacks using Neural Network and Decision Tree S.Florence 1, N.G.Bhuvaneswari Amma 2, G.Annapoorani 3, K.Malathi 4 PG Scholar, Indian Institute of Information Technology, Srirangam,
Data Mining Techniques for Prognosis in Pancreatic Cancer
Data Mining Techniques for Prognosis in Pancreatic Cancer by Stuart Floyd A Thesis Submitted to the Faculty of the WORCESTER POLYTECHNIC INSTITUE In partial fulfillment of the requirements for the Degree
Curriculum for to the PhD Program in Pharmacy Administration
Curriculum for to the PhD Program in Pharmacy Administration Course Hours Course Title BSE 5113 3 Principles of Epidemiology BSE 5163 3 Biostatistics Methods I* BSE 5173 3 Biostatistics Methods II* BSE
Marketing Mix Modelling and Big Data P. M Cain
1) Introduction Marketing Mix Modelling and Big Data P. M Cain Big data is generally defined in terms of the volume and variety of structured and unstructured information. Whereas structured data is stored
Statistical issues in the analysis of microarray data
Statistical issues in the analysis of microarray data Daniel Gerhard Institute of Biostatistics Leibniz University of Hannover ESNATS Summerschool, Zermatt D. Gerhard (LUH) Analysis of microarray data
Validation of measurement procedures
Validation of measurement procedures R. Haeckel and I.Püntmann Zentralkrankenhaus Bremen The new ISO standard 15189 which has already been accepted by most nations will soon become the basis for accreditation
Fairfield Public Schools
Mathematics Fairfield Public Schools AP Statistics AP Statistics BOE Approved 04/08/2014 1 AP STATISTICS Critical Areas of Focus AP Statistics is a rigorous course that offers advanced students an opportunity
Chapter 12 Discovering New Knowledge Data Mining
Chapter 12 Discovering New Knowledge Data Mining Becerra-Fernandez, et al. -- Knowledge Management 1/e -- 2004 Prentice Hall Additional material 2007 Dekai Wu Chapter Objectives Introduce the student to
Predictive Modeling in Workers Compensation 2008 CAS Ratemaking Seminar
Predictive Modeling in Workers Compensation 2008 CAS Ratemaking Seminar Prepared by Louise Francis, FCAS, MAAA Francis Analytics and Actuarial Data Mining, Inc. www.data-mines.com [email protected]
International Journal of Computer Trends and Technology (IJCTT) volume 4 Issue 8 August 2013
A Short-Term Traffic Prediction On A Distributed Network Using Multiple Regression Equation Ms.Sharmi.S 1 Research Scholar, MS University,Thirunelvelli Dr.M.Punithavalli Director, SREC,Coimbatore. Abstract:
Least Squares Estimation
Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David
Data Mining Techniques Chapter 6: Decision Trees
Data Mining Techniques Chapter 6: Decision Trees What is a classification decision tree?.......................................... 2 Visualizing decision trees...................................................
HYBRID PROBABILITY BASED ENSEMBLES FOR BANKRUPTCY PREDICTION
HYBRID PROBABILITY BASED ENSEMBLES FOR BANKRUPTCY PREDICTION Chihli Hung 1, Jing Hong Chen 2, Stefan Wermter 3, 1,2 Department of Management Information Systems, Chung Yuan Christian University, Taiwan
Building risk prediction models - with a focus on Genome-Wide Association Studies. Charles Kooperberg
Building risk prediction models - with a focus on Genome-Wide Association Studies Risk prediction models Based on data: (D i, X i1,..., X ip ) i = 1,..., n we like to fit a model P(D = 1 X 1,..., X p )
Data Mining. Nonlinear Classification
Data Mining Unit # 6 Sajjad Haider Fall 2014 1 Nonlinear Classification Classes may not be separable by a linear boundary Suppose we randomly generate a data set as follows: X has range between 0 to 15
Data Mining and Machine Learning in Bioinformatics
Data Mining and Machine Learning in Bioinformatics PRINCIPAL METHODS AND SUCCESSFUL APPLICATIONS Ruben Armañanzas http://mason.gmu.edu/~rarmanan Adapted from Iñaki Inza slides http://www.sc.ehu.es/isg
Simple Linear Regression Inference
Simple Linear Regression Inference 1 Inference requirements The Normality assumption of the stochastic term e is needed for inference even if it is not a OLS requirement. Therefore we have: Interpretation
Insurance Analytics - analýza dat a prediktivní modelování v pojišťovnictví. Pavel Kříž. Seminář z aktuárských věd MFF 4.
Insurance Analytics - analýza dat a prediktivní modelování v pojišťovnictví Pavel Kříž Seminář z aktuárských věd MFF 4. dubna 2014 Summary 1. Application areas of Insurance Analytics 2. Insurance Analytics
Validation, updating and impact of clinical prediction rules: A review
Journal of Clinical Epidemiology 61 (2008) 1085e1094 REVIEW ARTICLE Validation, updating and impact of clinical prediction rules: A review D.B. Toll, K.J.M. Janssen, Y. Vergouwe, K.G.M. Moons* Julius Center
Getting insights about life cycle cost drivers: an approach based on big data inspired statistical modelling
Introduction A Big Data applied to LCC Conclusion, Getting insights about life cycle cost drivers: an approach based on big data inspired statistical modelling Instituto Superior Técnico, Universidade
Regularized Logistic Regression for Mind Reading with Parallel Validation
Regularized Logistic Regression for Mind Reading with Parallel Validation Heikki Huttunen, Jukka-Pekka Kauppi, Jussi Tohka Tampere University of Technology Department of Signal Processing Tampere, Finland
Optimization of technical trading strategies and the profitability in security markets
Economics Letters 59 (1998) 249 254 Optimization of technical trading strategies and the profitability in security markets Ramazan Gençay 1, * University of Windsor, Department of Economics, 401 Sunset,
Strategies for Identifying Students at Risk for USMLE Step 1 Failure
Vol. 42, No. 2 105 Medical Student Education Strategies for Identifying Students at Risk for USMLE Step 1 Failure Jira Coumarbatch, MD; Leah Robinson, EdS; Ronald Thomas, PhD; Patrick D. Bridge, PhD Background
USING LOGIT MODEL TO PREDICT CREDIT SCORE
USING LOGIT MODEL TO PREDICT CREDIT SCORE Taiwo Amoo, Associate Professor of Business Statistics and Operation Management, Brooklyn College, City University of New York, (718) 951-5219, [email protected]
CHAPTER THREE COMMON DESCRIPTIVE STATISTICS COMMON DESCRIPTIVE STATISTICS / 13
COMMON DESCRIPTIVE STATISTICS / 13 CHAPTER THREE COMMON DESCRIPTIVE STATISTICS The analysis of data begins with descriptive statistics such as the mean, median, mode, range, standard deviation, variance,
STA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! [email protected]! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct
