Health Care and Life Sciences
|
|
|
- Michael Lucas
- 10 years ago
- Views:
Transcription
1 Sensitivity, Specificity, Accuracy, Associated Confidence Interval and ROC Analysis with Practical SAS Implementations Wen Zhu 1, Nancy Zeng 2, Ning Wang 2 1 K&L consulting services, Inc, Fort Washington, PA 2 Octagon Research Solutions, Wayne, PA 1. INTRUDUCTION Diagnosis tests include different kinds of information, such as medical tests (e.g. blood tests, X-rays, MRA), medical signs (clubbing of the fingers, a sign of lung disease), or symptoms (e.g. pain in a particular pattern). Doctor s decisions of medical treatment rely on diagnosis tests, which makes the accuracy of a diagnosis is essential in medical care. Fortunately, the attributes of the diagnosis tests can be measured. For a given disease condition, the best possible test can be chosen based on these attributes. Sensitivity, specificity and accuracy are widely used statistics to describe a diagnostic test. In particular, they are used to quantify how good and reliable a test is. Sensitivity evaluates how good the test is at detecting a positive disease. Specificity estimates how likely patients without disease can be correctly ruled out. ROC curve is a graphic presentation of the relationship between both sensitivity and specificity and it helps to decide the optimal model through determining the best threshold for the diagnostic test. Accuracy measures how correct a diagnostic test identifies and excludes a given condition. Accuracy of a diagnostic test can be determined from sensitivity and specificity with the presence of prevalence. Given the importance of these statistics in disease diagnosis and the terms are easily confused, it is important to get familiar with how they work, it helps us better understand when to use, how to implement them, and how to interpret the results. The importance and popularity of these statistics urges for a thorough review along with practical SAS examples. This paper will focus on the concepts of sensitivity, specificity and accuracy in the context of disease diagnosis: starting with a review of the definitions, how to calculate sensitivity, specificity and accuracy, associated 95% confidence interval and ROC analysis; followed by a practical example of disease diagnosis and related SAS macro code; then moving on to the common issues on interpreting the results of sensitivity, specificity and accuracy; ended by a final remark of the entire paper. 2. SENSITIVITY, SPECIFICITY AND ACCURACY, 95% CINFIDENCE INTERVAL AND ROC CURVE 2.1 SENSITIVITY, SPECIFICITY AND ACCURACY Outcome of the diagnostic test Condition (e.g. Disease) As determined by the Standard of Truth Positive Negative Row Total Positive TP FP Negative FN TN TP+FP (Total number of subjects with positive test) FN + TN (Total number of subjects with negative test) Column total TP+FN (Total number of subjects with given condition) FP+TN (Total number of subjects without given condition) N = TP+TN+FP+FN (Total number of subjects in study) Table 1. Terms used to define sensitivity, specificity and accuracy The ideas described below are summarized in table 1. 1
2 There are several terms that are commonly used along with the description of sensitivity, specificity and accuracy. They are true positive (TP), true negative (TN), false negative (FN), and false positive (FP). If a disease is proven present in a patient, the given diagnostic test also indicates the presence of disease, the result of the diagnostic test is considered true positive. Similarly, if a disease is proven absent in a patient, the diagnostic test suggests the disease is absent as well, the test result is true negative (TN). Both true positive and true negative suggest a consistent result between the diagnostic test and the proven condition (also called standard of truth). However, no medical test is perfect. If the diagnostic test indicates the presence of disease in a patient who actually has no such disease, the test result is false positive (FP). Similarly, if the result of the diagnosis test suggests that the disease is absent for a patient with disease for sure, the test result is false negative (FN). Both false positive and false negative indicate that the test results are opposite to the actual condition. Sensitivity, specificity and accuracy are described in terms of TP, TN, FN and FP. Sensitivity = TP/(TP + FN) = (Number of true positive assessment)/(number of all positive assessment) Specificity = TN/(TN + FP) = (Number of true negative assessment)/(number of all negative assessment) Accuracy = (TN + TP)/(TN+TP+FN+FP) = (Number of correct assessments)/number of all assessments) As suggested by above equations, sensitivity is the proportion of true positives that are correctly identified by a diagnostic test. It shows how good the test is at detecting a disease. Specificity is the proportion of the true negatives correctly identified by a diagnostic test. It suggests how good the test is at identifying normal (negative) condition. Accuracy is the proportion of true results, either true positive or true negative, in a population. It measures the degree of veracity of a diagnostic test on a condition. The numerical values of sensitivity represents the probability of a diagnos tic test identifies patients who do in fact have the disease. The higher the numerical value of sensitivity, the less likely diagnos tic tes t returns false-positive results. For example, if sensitivity = 99%, it means: when we conduct a diagnostic test on a patient with certain disease, there is 99% of chance, this patient will be identified as positive. A test with high sensitivity tents to capture all possible positive conditions without missing anyone. Thus a test with high sensitivity is often used to screen for disease. The numerical value of specificity represents the probability of a test diagnoses a particular disease without giving false-positive results. For example, if the specificity of a test is 99%. It means: when we conduct a diagnostic test on a patient without certain disease, there is 99% chance; this patient will be identified as negative. A test can be very specific without being sensitive, or it can be very sensitive without being specific. Both factors are equally important. A good test is a one has both high sensitivity and specificity. A good example of a test with high sensitive and specificity is pregnancy test. A positive result of pregnancy test almost for sure suggests the subject who took the test is pregnant. A negative result almost certainly rules out the possibility of being pregnant. In addition to the equation show above, accuracy can be determined from sensitivity and specificity, where prevalence is known. Prevalence is the probability of disease in the population at a given time: Accuracy = (sensitivity) (prevalence) + (specificity) (1 - prevalence). The numerical value of accuracy represents the proportion of true positive results (both true positive and true negative) in the selected population. An accuracy of 99% of times the test result is accurate, regardless positive or negative. This stays correct for most of the cases. However, it worth mentioning, the equation of accuracy implies that even if both sensitivity and specificity are high, say 99%, it does not suggest that the accuracy of the test is equally high as well. In addition to sensitivity and specificity, the accuracy is also determined by how common the disease in the selected population. A diagnosis for rare conditions in the population of interest may result in high sensitivity and specificity, but low accuracy. Accuracy needs to be interpreted cautiously. 2.2 ASYMPTOTIC AND EXACT 95% CONFIDENCE INTERVAL The sensitivity, specificity and accuracy are proportions, thus the according confidence intervals can be calculated by using standard methods for proportions 1. Two types of 95% confidence intervals are generally constructed around proportions: asymptotic and exact 95% confidence interval. The exact confidence interval is constructed by using binomial distribution to reach an exact estimate. Asymptotic confidence interval is calculated by assuming a normal approximation of the sampling distribution. The choice of these two types of confidence interval depends on whether the sample proportion is a good approximation of normal distribution. If the number of event is very small or if the sample size is very small, the normal assumption cannot be met. Thus, exact confident interval is desired. SAS example for both types of 95% confidence intervals will be provided in section 3. 2
3 2.3 RECEIVER OPERATING CHARACTERISTICS (ROC) ANALYSIS For a given diagnostic test, the true positive rate (TPR) against false positive rate (FPR) can be measured, where TPR= TP/(TP+FN) And FPR = FP/(FP+TN) As we can see from the above equations, TPR is equivalent to sensitivity and FPR is equivalent to (1 specificity). All possible combinations of TPR and FPR compose a ROC space. One TPR and one FPR together determine a single point in the ROC space, and the position of a point in the ROC space shows the tradeoff between sensitivity and specificity, i.e. the increase in sensitivity is accompanied by a decrease in specificity. Thus the location of the point in the ROC space depicts whether the diagnostic classification is good or not. In an ideal situation, a point determined by both TPR and FPF yields a coordinates (0, 1), or we can say that this point falls on the upper left corner of the ROC space. This idea point indicates the diagnostic test has a sensitivity of 100% and specificity of 100%. It is also called perfect classification. Diagnostic test with 50% sensitivity and 50% specificity can be visualized on the diagonal determined by coordinate (0, 0) and coordinates (1, 0). Theoretically, a random guess would give a point along this diagonal. A point predicted by a diagnostic test fall into the area above the diagonal represents a good diagnostic classification, otherwise a bad prediction. A graphic presentation of what described above is shown in figure 1. 1 Ideal coordinate (0, 1) 0.8 Cut-point 0.6 Random classification Sensitivity (TPR) specificity (FPR) Figure 1: ROC Space: shadow area represents better diagnostic classification 1 A single cut-point of a diagnostic test defines one single point in the ROC space; however, different possible cutpoints of a diagnostic test determine a curve in ROC space, which is also called ROC curve. Like a single point in the 3
4 ROC space, ROC curve is often plotted by using true positive rate (TPR) against false positive rate (FPR) for different cut-points of a diagnostic test, starting from coordinate (0, 0) and ending at coordinate (1, 1). FPR (1 specificity) is represented by x-axis and TPR (sensitivity) is represented by y-axis. Thus, ROC curve is a plot of a test s sensitivity vs. (1-specificity) as well. The interpretation of ROC curve is similar to a single point in the ROC space, the closer the point on the ROC curve to the ideal coordinate, the more accurate the test is. The closer the points on the ROC curve to the diagonal, the less accurate the test is. In addition, (1) the faster the curve approach the ideal point, the more useful the test results are; (2) the slope of the tangent line to a cut-point tells us the ratio of the probability of identifying true positive over true negative, i.e. likelihood ratio (LR) for the test value: LR = sensitivity/(1-specificity), if the ratio is equal to 1, the selected cut-point doesn t add additional information to identify true positive result. If the ratio is greater than 1, the selected cut-point help identify true positive result. If the ratio is less than 1, it decreases disease likelihood (3) the area under ROC curve (AUC) provides a way to measure the accuracy of a diagnostic test. The larger area, the more accurate the diagnostic test is. AUC of ROC curve can be measured by the following equation, Where t = (1 specificity) and ROC (t) is sensitivity. Commonly used classification using AUC for a diagnostic test is summarized in table 2: Table 2: accuracy classification by AUC for a diagnostic test AUC Range Classification 0.9 < AUC < 1.0 Excellent 0.8 < AUC < 0.9 Good 0.7 < AUC < 0.8 Worthless 0.6 < AUC < 0.7 Not good In short, ROC curve is a good tool to select possible optimal cut-point for a given diagnostic test. 3. AN PRACTICAL EXAMPLE WITH SAS CODE A good example explains better. Suppose, if there is an existing test that can always identify the true positive and true negative in determining the presence or absence of haemodynamically relevant stenosis of the renal arteries. However it is very time-consuming and expensive. A more efficient and affordable test is discovered, the preliminary results shows high sensitivity and specificity. A trial is then carried out to confirm the efficacy of this diagnostic test. Assuming the existing test has 100% accuracy, it will be used as the standard of truth, in other words, we believe that the test results always reflect the true stenosis of the renal arteries. A population of patients is enrolled, and both existing test and trial test are performed on each patient in order to obtain comparable results. And the test results are record in the following dataset data1. And the SAS examples used to calculate the sensitivity, specificity, accuracy and associated asymptotic and exact confidence interval are provided below: A dummy dataset is created below: data1, in which test1 is standard of truth, test2 is for trial test result. value 0 for test1 and test2 means disease absent, while 1 means disease present. data data1; input id $ test1 test2; datalines; sub sub sub sub sub sub sub sub sub sub sub sub sub sub
5 sub sub sub sub sub sub sub sub sub sub sub sub sub sub sub sub sub sub sub sub sub sub sub sub sub sub ; Compare the test result with the standard of truth and assigne TP, TN, FP, FN to each test result following the guidance of table 1. data data2; set data1; if test1=1 then do; if test2=1 then result_c12="tp"; else if test2=0 then result_c12="fn"; else if test1=0 then do; if test2=1 then result_c12="fp"; else if test2=0 then result_c12="tn"; proc sort data=data2; by test1 test2 ; Once the results are assigned to different categories, the sensitivity, specificity and accuracy can be easily calculated by using the formula provided in previous section of this table. The following SAS code can be used for the calculations. This data step generates count for sensitivity and specificity. data main1 (drop=id result_c12); set data2; by test1; retain tp tn fp fn; if (first.test1) then do; tp=0; tn=0; fp=0; fn=0; if (result_c12 in ("TP")) then tp=tp+1; if (result_c12 in ("TN")) then tn=tn+1; if (result_c12 in ("FN")) then fn=fn+1; 5
6 if (result_c12 in ("FP")) then fp=fp+1; else ; if (last.test1) then This data step generates accuracy count. data main2; set main1; tntp=tn+tp; fnfp=fn+fp; Total count of sensitivity, specificity and accuracy could be summed up by proc SQL. proc sql; create table main3 as select sum(tp) as tp, sum(tn) as tn, sum(fp)as fp, sum(fn) as fn, sum(tntp) as tntp, sum(fnfp) as fnfp from main2 ; quit; 3.1 SENSITIVITY, SPECIFICITY AND ACCURACY proc sql; create table main4 as select tp/(tp+fn) as sensitivity, tn/(tn+fp) as specificity, (tn+tp)/(tn+tp+fn+fp) as accuracy from main3 ; quit; This MAIN4 dataset shows the sensitivity, specificity and accuracy of the diagnosis method. Sensitivity Specificity Accuracy ASYMPTOTIC AND EXACT 95% CONFIDENCE INTERVAL Create a dataset for calculating 95% confidence interval; the dataset follows the format showing in table 3.2. proc transpose data=main3 out=t_main; var tp tn fn fp tntp fnfp; data table32 (drop=_name_ col1); length group $20; set t_main; count=col1; if _name_="tp" then do; group="sensitivity"; response=0; else if _name_="fn" then do; group="sensitivity"; response=1; else if _name_="tn" then do; group="specificity"; response=0; 6
7 else if _name_="fp" then do; group="specificity"; response=1; else if _name_="tntp" then do; group="accuracy"; response=0; else if _name_="fnfp" then do; group="accuracy"; response=1; Table 3.2 format of the dataset for calculating 95% confidence interval Group Response Count Sensitivity 0 TP Sensitivity 1 FN Specificity 0 TN Specificity 1 FP Accuracy 0 TN + TP Accuracy 1 FN+FP proc freq data= table32; weight count; by group; tables response/alpha=0.05 binomial(p=0.5); exact binomial; The 95% confidence interval outputted from the SAS procedure is listed below: Rate Asymptotic 95% CI Exact 95% CI Sensitivity 0.7 (0.4992, ) (0.4572, ) Specificity 0.85 (0.6935, ) (0.6211, ) Accuracy (0.6456, ) (0.6155, ) 3.3 RECEIVER OPERATING CHARACTERISTICS (ROC) ANALYSIS ROC analysis also allows analyzing sensitivity and specificity simultaneously at different cut-points; this approach better estimates the accuracy of a given trial test by using multiple pairs of sensitivity and specificity. In the previous example, a trial test is to measure the degree of occlusion (stenosis vs. non-stenosis ). With ROC analysis, the trial test evaluates a continuous degree of stenosis, also called the cut-off points, these points are chosen from 20% upwards to 80% percent stenosis. Trial test at each cut-off point returns one sensitivity and one specificity. The ROC analysis plots all sensitivity vs. (1-specificity) at selected cut-offs points by placing each pair of sensitivity and (1- specificity) in ROC space. The area under the curve (AUC) is a parameter indicating the intrinsic accuracy of the diagnostic test in determining the haemodynamically relevant stenosis of renal arteries. The sensitivity and specificity for each cut-point is calculated in the same way described in previous section and recorded in a dataset named DATA_ROC (variable SENS is for sensitivity and spec1 is for 1-Specificity). The SAS example used to generate the ROC curve and AUC are provided below. The dummy data that the example based on is as follow: data data_roc; input order cut sens spec spec1; datalines;
8 ; Cut-points Sensitivity (1-Specificity) 20% % % % % % proc gplot data=data_roc; symbol1 v=square i=j; plot sens*spec1/ vaxis=0 to 1 by 0.1 haxis=0 to 1 by 0.1; label sens="sensitivity" spec1="1-specificy"; Sensitivity Specificity Figure 2. Sample ROC plot: x-axis = (1-specificity), y-axis = sensitivity. The area under the ROC curve represents accuracy of a trial test. ROC curve AUC is determined by multiple cut-points of the trial test, it gives better estimate of accuracy. ROC curve AUC can be calculated by the following data step: data auc; set data_roc end=eof; drop lagx lagy; lagx=lag(spec1); lagy=lag(sens); if order=1 then do; lagx=0; lagy=0; tpzd=(spec1-lagx)*(sens+lagy)/2; sumtpz+tpzd; 8
9 if eof then do; roc_auc=sumtpz+(1-spec1)*(sens+1)/2;; The AUC obtained from the above code and data is According to table 2, the trial test has fairly good accuracy. REFERENCE 1. Gardner MJ, Altman DG. Calculating confidence intervals for proportions and their differences. In: Gardner MJ, Altman DG, eds. Statistics with confidence. London: BMJ Publishing Group, 1989:28-33 ACKNOWLEDGEMENT AND COPYRIGHT INFORMATION SAS and all other SAS Institute Inc. product or service names are registered trademarks or trademarks of SAS Institute Inc. in the USA and other countries. indicates USA registration. Other brand and product names are trademarks of their respective companies. CONTACT INFORMATION Questions and Feedbacks are welcome, please send them to: Wen Zhu K&L Consulting Services, Inc Virginia Drive, Suite 103, Fort Washington, PA Nancy Zeng Octagon Research Solution, Inc. 585 East Swedeford Rd. Suite 200 Wayne, PA ext 5803 Ning Wang Octagon Research Solution, Inc. 585 East Swedeford Rd. Suite 200 Wayne, PA ext
Evaluation & Validation: Credibility: Evaluating what has been learned
Evaluation & Validation: Credibility: Evaluating what has been learned How predictive is a learned model? How can we evaluate a model Test the model Statistical tests Considerations in evaluating a Model
Performance Measures for Machine Learning
Performance Measures for Machine Learning 1 Performance Measures Accuracy Weighted (Cost-Sensitive) Accuracy Lift Precision/Recall F Break Even Point ROC ROC Area 2 Accuracy Target: 0/1, -1/+1, True/False,
Measures of diagnostic accuracy: basic definitions
Measures of diagnostic accuracy: basic definitions Ana-Maria Šimundić Department of Molecular Diagnostics University Department of Chemistry, Sestre milosrdnice University Hospital, Zagreb, Croatia E-mail
Knowledge Discovery and Data Mining
Knowledge Discovery and Data Mining Lecture 15 - ROC, AUC & Lift Tom Kelsey School of Computer Science University of St Andrews http://tom.home.cs.st-andrews.ac.uk [email protected] Tom Kelsey ID5059-17-AUC
Evaluation of Diagnostic Tests
Biostatistics for Health Care Researchers: A Short Course Evaluation of Diagnostic Tests Presented ed by: Siu L. Hui, Ph.D. Department of Medicine, Division of Biostatistics Indiana University School of
PROC LOGISTIC: Traps for the unwary Peter L. Flom, Independent statistical consultant, New York, NY
PROC LOGISTIC: Traps for the unwary Peter L. Flom, Independent statistical consultant, New York, NY ABSTRACT Keywords: Logistic. INTRODUCTION This paper covers some gotchas in SAS R PROC LOGISTIC. A gotcha
SAS Software to Fit the Generalized Linear Model
SAS Software to Fit the Generalized Linear Model Gordon Johnston, SAS Institute Inc., Cary, NC Abstract In recent years, the class of generalized linear models has gained popularity as a statistical modeling
11. Analysis of Case-control Studies Logistic Regression
Research methods II 113 11. Analysis of Case-control Studies Logistic Regression This chapter builds upon and further develops the concepts and strategies described in Ch.6 of Mother and Child Health:
STATISTICA Formula Guide: Logistic Regression. Table of Contents
: Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary
Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
X X X a) perfect linear correlation b) no correlation c) positive correlation (r = 1) (r = 0) (0 < r < 1)
CORRELATION AND REGRESSION / 47 CHAPTER EIGHT CORRELATION AND REGRESSION Correlation and regression are statistical methods that are commonly used in the medical literature to compare two or more variables.
Business Case Development for Credit and Debit Card Fraud Re- Scoring Models
Business Case Development for Credit and Debit Card Fraud Re- Scoring Models Kurt Gutzmann Managing Director & Chief ScienAst GCX Advanced Analy.cs LLC www.gcxanalyacs.com October 20, 2011 www.gcxanalyacs.com
Binary Diagnostic Tests Two Independent Samples
Chapter 537 Binary Diagnostic Tests Two Independent Samples Introduction An important task in diagnostic medicine is to measure the accuracy of two diagnostic tests. This can be done by comparing summary
Performance Metrics for Graph Mining Tasks
Performance Metrics for Graph Mining Tasks 1 Outline Introduction to Performance Metrics Supervised Learning Performance Metrics Unsupervised Learning Performance Metrics Optimizing Metrics Statistical
CHAPTER 4 QUALITY ASSURANCE AND TEST VALIDATION
CHAPTER 4 QUALITY ASSURANCE AND TEST VALIDATION CINDY WEILAND AND SANDRA L. KATANICK Continued innovations in noninvasive testing equipment provide skilled sonographers and physicians with the technology
Paper DV-06-2015. KEYWORDS: SAS, R, Statistics, Data visualization, Monte Carlo simulation, Pseudo- random numbers
Paper DV-06-2015 Intuitive Demonstration of Statistics through Data Visualization of Pseudo- Randomly Generated Numbers in R and SAS Jack Sawilowsky, Ph.D., Union Pacific Railroad, Omaha, NE ABSTRACT Statistics
Paper 208-28. KEYWORDS PROC TRANSPOSE, PROC CORR, PROC MEANS, PROC GPLOT, Macro Language, Mean, Standard Deviation, Vertical Reference.
Paper 208-28 Analysis of Method Comparison Studies Using SAS Mohamed Shoukri, King Faisal Specialist Hospital & Research Center, Riyadh, KSA and Department of Epidemiology and Biostatistics, University
STAT 35A HW2 Solutions
STAT 35A HW2 Solutions http://www.stat.ucla.edu/~dinov/courses_students.dir/09/spring/stat35.dir 1. A computer consulting firm presently has bids out on three projects. Let A i = { awarded project i },
Evaluation of Predictive Models
Evaluation of Predictive Models Assessing calibration and discrimination Examples Decision Systems Group, Brigham and Women s Hospital Harvard Medical School Harvard-MIT Division of Health Sciences and
BEST METHODS FOR SOLVING QUADRATIC INEQUALITIES.
BEST METHODS FOR SOLVING QUADRATIC INEQUALITIES. I. GENERALITIES There are 3 common methods to solve quadratic inequalities. Therefore, students sometimes are confused to select the fastest and the best
Performance Measures in Data Mining
Performance Measures in Data Mining Common Performance Measures used in Data Mining and Machine Learning Approaches L. Richter J.M. Cejuela Department of Computer Science Technische Universität München
Data Mining Practical Machine Learning Tools and Techniques
Credibility: Evaluating what s been learned Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter 5 of Data Mining by I. H. Witten, E. Frank and M. A. Hall Issues: training, testing,
Getting Correct Results from PROC REG
Getting Correct Results from PROC REG Nathaniel Derby, Statis Pro Data Analytics, Seattle, WA ABSTRACT PROC REG, SAS s implementation of linear regression, is often used to fit a line without checking
2. Simple Linear Regression
Research methods - II 3 2. Simple Linear Regression Simple linear regression is a technique in parametric statistics that is commonly used for analyzing mean response of a variable Y which changes according
Survey, Statistics and Psychometrics Core Research Facility University of Nebraska-Lincoln. Log-Rank Test for More Than Two Groups
Survey, Statistics and Psychometrics Core Research Facility University of Nebraska-Lincoln Log-Rank Test for More Than Two Groups Prepared by Harlan Sayles (SRAM) Revised by Julia Soulakova (Statistics)
ABSTRACT INTRODUCTION
Paper SP03-2009 Illustrative Logistic Regression Examples using PROC LOGISTIC: New Features in SAS/STAT 9.2 Robert G. Downer, Grand Valley State University, Allendale, MI Patrick J. Richardson, Van Andel
Performance Metrics. number of mistakes total number of observations. err = p.1/1
p.1/1 Performance Metrics The simplest performance metric is the model error defined as the number of mistakes the model makes on a data set divided by the number of observations in the data set, err =
Biomarker Discovery and Data Visualization Tool for Ovarian Cancer Screening
, pp.169-178 http://dx.doi.org/10.14257/ijbsbt.2014.6.2.17 Biomarker Discovery and Data Visualization Tool for Ovarian Cancer Screening Ki-Seok Cheong 2,3, Hye-Jeong Song 1,3, Chan-Young Park 1,3, Jong-Dae
IBM SPSS Direct Marketing 23
IBM SPSS Direct Marketing 23 Note Before using this information and the product it supports, read the information in Notices on page 25. Product Information This edition applies to version 23, release
1. Classification problems
Neural and Evolutionary Computing. Lab 1: Classification problems Machine Learning test data repository Weka data mining platform Introduction Scilab 1. Classification problems The main aim of a classification
S03-2008 The Difference Between Predictive Modeling and Regression Patricia B. Cerrito, University of Louisville, Louisville, KY
S03-2008 The Difference Between Predictive Modeling and Regression Patricia B. Cerrito, University of Louisville, Louisville, KY ABSTRACT Predictive modeling includes regression, both logistic and linear,
IBM SPSS Direct Marketing 22
IBM SPSS Direct Marketing 22 Note Before using this information and the product it supports, read the information in Notices on page 25. Product Information This edition applies to version 22, release
Leak Detection Theory: Optimizing Performance with MLOG
Itron White Paper Water Loss Management Leak Detection Theory: Optimizing Performance with MLOG Rich Christensen Vice President, Research & Development 2009, Itron Inc. All rights reserved. Introduction
SENSITIVITY ANALYSIS AND INFERENCE. Lecture 12
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this
Chapter 6. The stacking ensemble approach
82 This chapter proposes the stacking ensemble approach for combining different data mining classifiers to get better performance. Other combination techniques like voting, bagging etc are also described
Innovative Techniques and Tools to Detect Data Quality Problems
Paper DM05 Innovative Techniques and Tools to Detect Data Quality Problems Hong Qi and Allan Glaser Merck & Co., Inc., Upper Gwynnedd, PA ABSTRACT High quality data are essential for accurate and meaningful
STT315 Chapter 4 Random Variables & Probability Distributions KM. Chapter 4.5, 6, 8 Probability Distributions for Continuous Random Variables
Chapter 4.5, 6, 8 Probability Distributions for Continuous Random Variables Discrete vs. continuous random variables Examples of continuous distributions o Uniform o Exponential o Normal Recall: A random
Bayes Theorem & Diagnostic Tests Screening Tests
Bayes heorem & Screening ests Bayes heorem & Diagnostic ests Screening ests Some Questions If you test positive for HIV, what is the probability that you have HIV? If you have a positive mammogram, what
What are the place values to the left of the decimal point and their associated powers of ten?
The verbal answers to all of the following questions should be memorized before completion of algebra. Answers that are not memorized will hinder your ability to succeed in geometry and algebra. (Everything
Simple Predictive Analytics Curtis Seare
Using Excel to Solve Business Problems: Simple Predictive Analytics Curtis Seare Copyright: Vault Analytics July 2010 Contents Section I: Background Information Why use Predictive Analytics? How to use
PREDICTING SUCCESS IN THE COMPUTER SCIENCE DEGREE USING ROC ANALYSIS
PREDICTING SUCCESS IN THE COMPUTER SCIENCE DEGREE USING ROC ANALYSIS Arturo Fornés [email protected], José A. Conejero [email protected] 1, Antonio Molina [email protected], Antonio Pérez [email protected],
Chapter 4. Probability and Probability Distributions
Chapter 4. robability and robability Distributions Importance of Knowing robability To know whether a sample is not identical to the population from which it was selected, it is necessary to assess the
Hindrik Vondeling Department of Health Economics University of Southern Denmark Odense, Denmark
Health technology assessment: an introduction with emphasis on assessment of diagnostic tests Hindrik Vondeling Department of Health Economics University of Southern Denmark Odense, Denmark Outline λ Health
Predictive Implications of Stress Testing (Chapt. 14) 1979, Weiner and coworkers. Factors to improve the accuracy of stress testing
Predictive Implications of Stress Testing (Chapt. 14) Sensitivity Specificity Predictive Value Patient Risk 1979, Weiner and coworkers Stress testing has very little diagnostic value. A positive stress
Chapter 6: Constructing and Interpreting Graphic Displays of Behavioral Data
Chapter 6: Constructing and Interpreting Graphic Displays of Behavioral Data Chapter Focus Questions What are the benefits of graphic display and visual analysis of behavioral data? What are the fundamental
Methodologies for Evaluation of Standalone CAD System Performance
Methodologies for Evaluation of Standalone CAD System Performance DB DSFM DCMS OSEL DESE DP DIAM Berkman Sahiner, PhD USFDA/CDRH/OSEL/DIAM AAPM CAD Subcommittee in Diagnostic Imaging CAD: CADe and CADx
SAS Enterprise Guide in Pharmaceutical Applications: Automated Analysis and Reporting Alex Dmitrienko, Ph.D., Eli Lilly and Company, Indianapolis, IN
Paper PH200 SAS Enterprise Guide in Pharmaceutical Applications: Automated Analysis and Reporting Alex Dmitrienko, Ph.D., Eli Lilly and Company, Indianapolis, IN ABSTRACT SAS Enterprise Guide is a member
Evaluation of Diagnostic and Screening Tests: Validity and Reliability. Sukon Kanchanaraksa, PhD Johns Hopkins University
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this
POINT OF SALE FRAUD PREVENTION
2002 IEEE Systems and Information Design Symposium University of Virginia POINT OF SALE FRAUD PREVENTION Student team: Jeff Grossman, Dawn Herndon, Andrew Kaplan, Mark Michalski, Carlton Pate Faculty Advisors:
How to Get More Value from Your Survey Data
Technical report How to Get More Value from Your Survey Data Discover four advanced analysis techniques that make survey research more effective Table of contents Introduction..............................................................2
Data quality in Accounting Information Systems
Data quality in Accounting Information Systems Comparing Several Data Mining Techniques Erjon Zoto Department of Statistics and Applied Informatics Faculty of Economy, University of Tirana Tirana, Albania
t Tests in Excel The Excel Statistical Master By Mark Harmon Copyright 2011 Mark Harmon
t-tests in Excel By Mark Harmon Copyright 2011 Mark Harmon No part of this publication may be reproduced or distributed without the express permission of the author. [email protected] www.excelmasterseries.com
MEASURES OF LOCATION AND SPREAD
Paper TU04 An Overview of Non-parametric Tests in SAS : When, Why, and How Paul A. Pappas and Venita DePuy Durham, North Carolina, USA ABSTRACT Most commonly used statistical procedures are based on the
Big Data: a new era for Statistics
Big Data: a new era for Statistics Richard J. Samworth Abstract Richard Samworth (1996) is a Professor of Statistics in the University s Statistical Laboratory, and has been a Fellow of St John s since
CALCULATIONS & STATISTICS
CALCULATIONS & STATISTICS CALCULATION OF SCORES Conversion of 1-5 scale to 0-100 scores When you look at your report, you will notice that the scores are reported on a 0-100 scale, even though respondents
6.4 Normal Distribution
Contents 6.4 Normal Distribution....................... 381 6.4.1 Characteristics of the Normal Distribution....... 381 6.4.2 The Standardized Normal Distribution......... 385 6.4.3 Meaning of Areas under
Effective Use of SQL in SAS Programming
INTRODUCTION Effective Use of SQL in SAS Programming Yi Zhao Merck & Co. Inc., Upper Gwynedd, Pennsylvania Structured Query Language (SQL) is a data manipulation tool of which many SAS programmers are
ln(p/(1-p)) = α +β*age35plus, where p is the probability or odds of drinking
Dummy Coding for Dummies Kathryn Martin, Maternal, Child and Adolescent Health Program, California Department of Public Health ABSTRACT There are a number of ways to incorporate categorical variables into
Constructing a Table of Survey Data with Percent and Confidence Intervals in every Direction
Constructing a Table of Survey Data with Percent and Confidence Intervals in every Direction David Izrael, Abt Associates Sarah W. Ball, Abt Associates Sara M.A. Donahue, Abt Associates ABSTRACT We examined
Review of Fundamental Mathematics
Review of Fundamental Mathematics As explained in the Preface and in Chapter 1 of your textbook, managerial economics applies microeconomic theory to business decision making. The decision-making tools
An Application of the Cox Proportional Hazards Model to the Construction of Objective Vintages for Credit in Financial Institutions, Using PROC PHREG
Paper 3140-2015 An Application of the Cox Proportional Hazards Model to the Construction of Objective Vintages for Credit in Financial Institutions, Using PROC PHREG Iván Darío Atehortua Rojas, Banco Colpatria
ROC Curve, Lift Chart and Calibration Plot
Metodološki zvezki, Vol. 3, No. 1, 26, 89-18 ROC Curve, Lift Chart and Calibration Plot Miha Vuk 1, Tomaž Curk 2 Abstract This paper presents ROC curve, lift chart and calibration plot, three well known
Chi Squared and Fisher's Exact Tests. Observed vs Expected Distributions
BMS 617 Statistical Techniques for the Biomedical Sciences Lecture 11: Chi-Squared and Fisher's Exact Tests Chi Squared and Fisher's Exact Tests This lecture presents two similarly structured tests, Chi-squared
Generating Randomization Schedules Using SAS Programming Chunqin Deng and Julia Graz, PPD, Inc., Research Triangle Park, North Carolina
Paper 267-27 Generating Randomization Schedules Using SAS Programming Chunqin Deng and Julia Graz, PPD, Inc., Research Triangle Park, North Carolina ABSTRACT Randomization as a method of experimental control
Paper 109-25 Merges and Joins Timothy J Harrington, Trilogy Consulting Corporation
Paper 109-25 Merges and Joins Timothy J Harrington, Trilogy Consulting Corporation Abstract This paper discusses methods of joining SAS data sets. The different methods and the reasons for choosing a particular
PUBLIC HEALTH OPTOMETRY ECONOMICS. Kevin D. Frick, PhD
Chapter Overview PUBLIC HEALTH OPTOMETRY ECONOMICS Kevin D. Frick, PhD This chapter on public health optometry economics describes the positive and normative uses of economic science. The terms positive
Data Mining Classification: Basic Concepts, Decision Trees, and Model Evaluation. Lecture Notes for Chapter 4. Introduction to Data Mining
Data Mining Classification: Basic Concepts, Decision Trees, and Model Evaluation Lecture Notes for Chapter 4 Introduction to Data Mining by Tan, Steinbach, Kumar Tan,Steinbach, Kumar Introduction to Data
Predicting Defaults of Loans using Lending Club s Loan Data
Predicting Defaults of Loans using Lending Club s Loan Data Oleh Dubno Fall 2014 General Assembly Data Science Link to my Developer Notebook (ipynb) - http://nbviewer.ipython.org/gist/odubno/0b767a47f75adb382246
An analysis of suitable parameters for efficiently applying K-means clustering to large TCPdump data set using Hadoop framework
An analysis of suitable parameters for efficiently applying K-means clustering to large TCPdump data set using Hadoop framework Jakrarin Therdphapiyanak Dept. of Computer Engineering Chulalongkorn University
Scoring (manual, automated, automated with manual review)
A. Source and Extractor Author, Year Reference test PMID RefID Index test 1 Key Question(s) Index test 2 Extractor B. Study description Sampling population A Recruitment Multicenter? Enrollment method
WHO STEPS Surveillance Support Materials. STEPS Epi Info Training Guide
STEPS Epi Info Training Guide Department of Chronic Diseases and Health Promotion World Health Organization 20 Avenue Appia, 1211 Geneva 27, Switzerland For further information: www.who.int/chp/steps WHO
Part 3. Comparing Groups. Chapter 7 Comparing Paired Groups 189. Chapter 8 Comparing Two Independent Groups 217
Part 3 Comparing Groups Chapter 7 Comparing Paired Groups 189 Chapter 8 Comparing Two Independent Groups 217 Chapter 9 Comparing More Than Two Groups 257 188 Elementary Statistics Using SAS Chapter 7 Comparing
Mining Life Insurance Data for Customer Attrition Analysis
Mining Life Insurance Data for Customer Attrition Analysis T. L. Oshini Goonetilleke Informatics Institute of Technology/Department of Computing, Colombo, Sri Lanka Email: [email protected] H. A. Caldera
II. DISTRIBUTIONS distribution normal distribution. standard scores
Appendix D Basic Measurement And Statistics The following information was developed by Steven Rothke, PhD, Department of Psychology, Rehabilitation Institute of Chicago (RIC) and expanded by Mary F. Schmidt,
Using simulation to calculate the NPV of a project
Using simulation to calculate the NPV of a project Marius Holtan Onward Inc. 5/31/2002 Monte Carlo simulation is fast becoming the technology of choice for evaluating and analyzing assets, be it pure financial
Simple Regression Theory II 2010 Samuel L. Baker
SIMPLE REGRESSION THEORY II 1 Simple Regression Theory II 2010 Samuel L. Baker Assessing how good the regression equation is likely to be Assignment 1A gets into drawing inferences about how close the
An introduction to OBJECTIVE ASSESSMENT OF IMAGE QUALITY. Harrison H. Barrett University of Arizona Tucson, AZ
An introduction to OBJECTIVE ASSESSMENT OF IMAGE QUALITY Harrison H. Barrett University of Arizona Tucson, AZ Outline! Approaches to image quality! Why not fidelity?! Basic premises of the task-based approach!
Business Analytics using Data Mining Project Report. Optimizing Operation Room Utilization by Predicting Surgery Duration
Business Analytics using Data Mining Project Report Optimizing Operation Room Utilization by Predicting Surgery Duration Project Team 4 102034606 WU, CHOU-CHUN 103078508 CHEN, LI-CHAN 102077503 LI, DAI-SIN
Introduction to. Hypothesis Testing CHAPTER LEARNING OBJECTIVES. 1 Identify the four steps of hypothesis testing.
Introduction to Hypothesis Testing CHAPTER 8 LEARNING OBJECTIVES After reading this chapter, you should be able to: 1 Identify the four steps of hypothesis testing. 2 Define null hypothesis, alternative
Despite its emphasis on credit-scoring/rating model validation,
RETAIL RISK MANAGEMENT Empirical Validation of Retail Always a good idea, development of a systematic, enterprise-wide method to continuously validate credit-scoring/rating models nonetheless received
How To Check For Differences In The One Way Anova
MINITAB ASSISTANT WHITE PAPER This paper explains the research conducted by Minitab statisticians to develop the methods and data checks used in the Assistant in Minitab 17 Statistical Software. One-Way
ROC Graphs: Notes and Practical Considerations for Data Mining Researchers
ROC Graphs: Notes and Practical Considerations for Data Mining Researchers Tom Fawcett Intelligent Enterprise Technologies Laboratory HP Laboratories Palo Alto HPL-23-4 January 7 th, 23* E-mail: [email protected]
Two Correlated Proportions (McNemar Test)
Chapter 50 Two Correlated Proportions (Mcemar Test) Introduction This procedure computes confidence intervals and hypothesis tests for the comparison of the marginal frequencies of two factors (each with
An Application of Weibull Analysis to Determine Failure Rates in Automotive Components
An Application of Weibull Analysis to Determine Failure Rates in Automotive Components Jingshu Wu, PhD, PE, Stephen McHenry, Jeffrey Quandt National Highway Traffic Safety Administration (NHTSA) U.S. Department
AP Physics 1 and 2 Lab Investigations
AP Physics 1 and 2 Lab Investigations Student Guide to Data Analysis New York, NY. College Board, Advanced Placement, Advanced Placement Program, AP, AP Central, and the acorn logo are registered trademarks
Common Tools for Displaying and Communicating Data for Process Improvement
Common Tools for Displaying and Communicating Data for Process Improvement Packet includes: Tool Use Page # Box and Whisker Plot Check Sheet Control Chart Histogram Pareto Diagram Run Chart Scatter Plot
Simulating Chi-Square Test Using Excel
Simulating Chi-Square Test Using Excel Leslie Chandrakantha John Jay College of Criminal Justice of CUNY Mathematics and Computer Science Department 524 West 59 th Street, New York, NY 10019 [email protected]
VI. Introduction to Logistic Regression
VI. Introduction to Logistic Regression We turn our attention now to the topic of modeling a categorical outcome as a function of (possibly) several factors. The framework of generalized linear models
Life Tables. Marie Diener-West, PhD Sukon Kanchanaraksa, PhD
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this
Make Better Decisions with Optimization
ABSTRACT Paper SAS1785-2015 Make Better Decisions with Optimization David R. Duling, SAS Institute Inc. Automated decision making systems are now found everywhere, from your bank to your government to
Accurately and Efficiently Measuring Individual Account Credit Risk On Existing Portfolios
Accurately and Efficiently Measuring Individual Account Credit Risk On Existing Portfolios By: Michael Banasiak & By: Daniel Tantum, Ph.D. What Are Statistical Based Behavior Scoring Models And How Are
An Approach to Detect Spam Emails by Using Majority Voting
An Approach to Detect Spam Emails by Using Majority Voting Roohi Hussain Department of Computer Engineering, National University of Science and Technology, H-12 Islamabad, Pakistan Usman Qamar Faculty,
Testing differences in proportions
Testing differences in proportions Murray J Fisher RN, ITU Cert., DipAppSc, BHSc, MHPEd, PhD Senior Lecturer and Director Preregistration Programs Sydney Nursing School (MO2) University of Sydney NSW 2006
Probit Analysis By: Kim Vincent
Probit Analysis By: Kim Vincent Quick Overview Probit analysis is a type of regression used to analyze binomial response variables. It transforms the sigmoid dose-response curve to a straight line that
In this presentation, you will be introduced to data mining and the relationship with meaningful use.
In this presentation, you will be introduced to data mining and the relationship with meaningful use. Data mining refers to the art and science of intelligent data analysis. It is the application of machine
Cool Tools for PROC LOGISTIC
Cool Tools for PROC LOGISTIC Paul D. Allison Statistical Horizons LLC and the University of Pennsylvania March 2013 www.statisticalhorizons.com 1 New Features in LOGISTIC ODDSRATIO statement EFFECTPLOT
