Introduction to nonparametric regression: Least squares vs. Nearest neighbors
|
|
- Ethelbert Benson
- 1 years ago
- Views:
Transcription
1 Introduction to nonparametric regression: Least squares vs. Nearest neighbors Patrick Breheny October 30 Patrick Breheny STA 621: Nonparametric Statistics 1/16
2 Introduction For the remainder of the course, we will focus on nonparametric regression and classification The regression problem involves modeling how the expected value of a response y changes in response to changes in an explanatory variable x: E(y x) = f(x) Linear regression, as its name implies, assumes a linear relationship; namely, that f(x) = β 0 + β 1 x Patrick Breheny STA 621: Nonparametric Statistics 2/16
3 Parametric vs. nonparametric approaches This reduction of a complicated function to a simple form with a small number of unknown parameters is very similar to the parametric approach to estimation and inference involving the unknown distribution function The nonparametric approach, in contrast, is to make as few assumptions about the regression function f as possible Instead, we will try to use the data as much as possible to learn about the potential shape of f allowing f to be very flexible, yet smooth Patrick Breheny STA 621: Nonparametric Statistics 3/16
4 Simple local models One way to achieve this flexibility is by fitting a different, simple model separately at every point x 0 in much the same way that we used kernels to estimate density As with kernel density estimates, this is done using only those observations close to x 0 to fit the simple model As we will see, it is possible to extend many of the same kernel ideas we have already discussed to smoothly blend these local models to construct a smooth estimate ˆf of the relationship between x and y Patrick Breheny STA 621: Nonparametric Statistics 4/16
5 Introduction Before we do so, however, let us get a general feel for the contrast between parametric and nonparametric classification by contrasting two simple, but very different, methods: the ordinary least squares regression model and the k-nearest neighbor prediction rule The linear model makes huge assumptions about the structure of the problem, but is quite stable Nearest neighbors is virtually assumption-free, but its results can be quite unstable Each method can be quite powerful in different settings and for different reasons Patrick Breheny STA 621: Nonparametric Statistics 5/16
6 Simulation settings To examine which method is better in which setting, we will simulate data from a simple model in which y can take on one of two values: 1 or 1 The corresponding x values are derived from one of two settings: Setting 1: x values are drawn from a bivariate normal distribution with different means for y = 1 and y = 1 Setting 2: A mixture in which 10 sets of means for each class (1, 1) are drawn; x values are then drawn by randomly selecting a mean from the appropriate class and then generating a random bivariate normal observation with that mean A fair competition between the two methods is then how well they do at predicting whether a future observation is 1 or 1 given its x values Patrick Breheny STA 621: Nonparametric Statistics 6/16
7 Linear model results Patrick Breheny STA 621: Nonparametric Statistics 7/16
8 Linear model remarks The linear model seems to classify points reasonably in setting 1 In setting 2, on the other hand, there are some regions which seem questionable For example, in the lower left hand corner of the plot, does it really make sense to predict blue given that all of the nearby points are red? Patrick Breheny STA 621: Nonparametric Statistics 8/16
9 Nearest neighbors Consider then a completely different approach in which we don t assume a model, a distribution, a likelihood, or anything about the problem: we just look at nearby points and base our prediction on the average of those points This approach is called the nearest-neighbor method, and is defined formally as ŷ(x) = 1 k x i N k (x) where N k (x) is the neighborhood of x defined by its k closest points in the sample y i, Patrick Breheny STA 621: Nonparametric Statistics 9/16
10 Nearest neighbor results Patrick Breheny STA 621: Nonparametric Statistics 10/16
11 Nearest neighbor remarks Nearest neighbor seems not to perform terribly well in setting 1, as its classification boundaries are unnecessarily complex and unstable On the other hand, the method seemed perhaps better than the linear model in setting 2, where a complex and curved boundary seems to fit the data better Furthermore, the choice of k plays a big role in the fit, and the optimal k might not be the same in settings 1 and 2 Patrick Breheny STA 621: Nonparametric Statistics 11/16
12 Inference Introduction Of course, it is potentially misleading to judge whether a method is better simply because it fits the sample better What matters, of course, is how well its predictions generalize to new samples Thus, consider generating 100 data sets of size 200, fitting each model, and then measuring how well each method does at predicting 10,000 new, independent observations Patrick Breheny STA 621: Nonparametric Statistics 12/16
13 Simulation results Black line = least squares; blue line = nearest neighbors Setting 1 Setting Misclassification Rate k Patrick Breheny STA 621: Nonparametric Statistics 13/16
14 Remarks Introduction In setting 1, linear regression was always better than nearest neighbors In setting 2, nearest neighbors was usually better than linear regression However, it wasn t always better than linear regression when k was too big or too small, the nearest neighbors method performed poorly In setting 1, the bigger k was, the better; in setting 2, there was a Goldilocks value of k (about 25) that proved optimal in balancing the bias-variance tradeoff Patrick Breheny STA 621: Nonparametric Statistics 14/16
15 Conclusions Thus, Fitting an ordinary linear model is rarely the best we can do On the other hand, nearest-neighbors is rarely stable enough to be ideal, even in modest dimensions, unless our sample size is very large (recall the curse of dimensionality) Patrick Breheny STA 621: Nonparametric Statistics 15/16
16 Conclusions (cont d) These two methods stand on opposite sides of the methodology spectrum with regard to assumptions and structure The methods we will discuss for the remainder of the course involve bridging the gap between these two methods making linear regression more flexible, adding structure and stability to nearest neighbor ideas, or combining concepts from both As with kernel density estimation, the main theme that emerges is the need to apply methods that bring the right mix of flexibility and stability that is appropriate for the data Patrick Breheny STA 621: Nonparametric Statistics 16/16
Local classification and local likelihoods
Local classification and local likelihoods November 18 k-nearest neighbors The idea of local regression can be extended to classification as well The simplest way of doing so is called nearest neighbor
Data Mining Chapter 6: Models and Patterns Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University
Data Mining Chapter 6: Models and Patterns Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University Models vs. Patterns Models A model is a high level, global description of a
Penalized regression: Introduction
Penalized regression: Introduction Patrick Breheny August 30 Patrick Breheny BST 764: Applied Statistical Modeling 1/19 Maximum likelihood Much of 20th-century statistics dealt with maximum likelihood
Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
Data Mining and Data Warehousing. Henryk Maciejewski. Data Mining Predictive modelling: regression
Data Mining and Data Warehousing Henryk Maciejewski Data Mining Predictive modelling: regression Algorithms for Predictive Modelling Contents Regression Classification Auxiliary topics: Estimation of prediction
Introduction to machine learning and pattern recognition Lecture 1 Coryn Bailer-Jones
Introduction to machine learning and pattern recognition Lecture 1 Coryn Bailer-Jones http://www.mpia.de/homes/calj/mlpr_mpia2008.html 1 1 What is machine learning? Data description and interpretation
Local regression II. Patrick Breheny. November 8. Implementations in R Variance estimation Confidence bands Hypothesis testing
Local regression II Patrick Breheny November 8 Patrick Breheny STA 621: Nonparametric Statistics 1/23 The loess function In R, local linear regression is implemented through the loess function, which uses
Data Mining Classification: Alternative Techniques. Instance-Based Classifiers. Lecture Notes for Chapter 5. Introduction to Data Mining
Data Mining Classification: Alternative Techniques Instance-Based Classifiers Lecture Notes for Chapter 5 Introduction to Data Mining by Tan, Steinbach, Kumar Set of Stored Cases Atr1... AtrN Class A B
Principles of Data Mining by Hand&Mannila&Smyth
Principles of Data Mining by Hand&Mannila&Smyth Slides for Textbook Ari Visa,, Institute of Signal Processing Tampere University of Technology October 4, 2010 Data Mining: Concepts and Techniques 1 Differences
Introduction to Predictive Models Book Chapters 1, 2 and 5. Carlos M. Carvalho The University of Texas McCombs School of Business
Introduction to Predictive Models Book Chapters 1, 2 and 5. Carlos M. Carvalho The University of Texas McCombs School of Business 1 1. Introduction 2. Measuring Accuracy 3. Out-of Sample Predictions 4.
Sentiment analysis using emoticons
Sentiment analysis using emoticons Royden Kayhan Lewis Moharreri Steven Royden Ware Lewis Kayhan Steven Moharreri Ware Department of Computer Science, Ohio State University Problem definition Our aim was
Two-sample inference: Continuous data
Two-sample inference: Continuous data Patrick Breheny April 5 Patrick Breheny STA 580: Biostatistics I 1/32 Introduction Our next two lectures will deal with two-sample inference for continuous data As
Chapter 4: Non-Parametric Classification
Chapter 4: Non-Parametric Classification Introduction Density Estimation Parzen Windows Kn-Nearest Neighbor Density Estimation K-Nearest Neighbor (KNN) Decision Rule Gaussian Mixture Model A weighted combination
Introduction to Statistical Machine Learning
CHAPTER Introduction to Statistical Machine Learning We start with a gentle introduction to statistical machine learning. Readers familiar with machine learning may wish to skip directly to Section 2,
Introduction to Machine Learning
Introduction to Machine Learning Prof. Alexander Ihler Prof. Max Welling icamp Tutorial July 22 What is machine learning? The ability of a machine to improve its performance based on previous results:
HT2015: SC4 Statistical Data Mining and Machine Learning
HT2015: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Bayesian Nonparametrics Parametric vs Nonparametric
Regression III: Advanced Methods
Lecture 16: Generalized Additive Models Regression III: Advanced Methods Bill Jacoby Michigan State University http://polisci.msu.edu/jacoby/icpsr/regress3 Goals of the Lecture Introduce Additive Models
CLASSIFICATION AND CLUSTERING. Anveshi Charuvaka
CLASSIFICATION AND CLUSTERING Anveshi Charuvaka Learning from Data Classification Regression Clustering Anomaly Detection Contrast Set Mining Classification: Definition Given a collection of records (training
Supervised and unsupervised learning - 1
Chapter 3 Supervised and unsupervised learning - 1 3.1 Introduction The science of learning plays a key role in the field of statistics, data mining, artificial intelligence, intersecting with areas in
Weighted least squares
Weighted least squares Patrick Breheny February 7 Patrick Breheny BST 760: Advanced Regression 1/17 Introduction Known weights As a precursor to fitting generalized linear models, let s first deal with
Automated Learning and Data Visualization
1 Automated Learning and Data Visualization William S. Cleveland Department of Statistics Department of Computer Science Purdue University Methods of Statistics, Machine Learning, and Data Mining 2 Mathematical
Classifiers & Classification
Classifiers & Classification Forsyth & Ponce Computer Vision A Modern Approach chapter 22 Pattern Classification Duda, Hart and Stork School of Computer Science & Statistics Trinity College Dublin Dublin
STA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct
Supervised Learning (Big Data Analytics)
Supervised Learning (Big Data Analytics) Vibhav Gogate Department of Computer Science The University of Texas at Dallas Practical advice Goal of Big Data Analytics Uncover patterns in Data. Can be used
Decompose Error Rate into components, some of which can be measured on unlabeled data
Bias-Variance Theory Decompose Error Rate into components, some of which can be measured on unlabeled data Bias-Variance Decomposition for Regression Bias-Variance Decomposition for Classification Bias-Variance
Introduction to Machine Learning. Speaker: Harry Chao Advisor: J.J. Ding Date: 1/27/2011
Introduction to Machine Learning Speaker: Harry Chao Advisor: J.J. Ding Date: 1/27/2011 1 Outline 1. What is machine learning? 2. The basic of machine learning 3. Principles and effects of machine learning
Parametric versus Semi/nonparametric Regression Models
Parametric versus Semi/nonparametric Regression Models Hamdy F. F. Mahmoud Virginia Polytechnic Institute and State University Department of Statistics LISA short course series- July 23, 2014 Hamdy Mahmoud
SENSITIVITY ANALYSIS AND INFERENCE. Lecture 12
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this
Regression III: Advanced Methods
Lecture 4: Transformations Regression III: Advanced Methods William G. Jacoby Michigan State University Goals of the lecture The Ladder of Roots and Powers Changing the shape of distributions Transforming
Cross Validation. Dr. Thomas Jensen Expedia.com
Cross Validation Dr. Thomas Jensen Expedia.com About Me PhD from ETH Used to be a statistician at Link, now Senior Business Analyst at Expedia Manage a database with 720,000 Hotels that are not on contract
Multivariate Normal Distribution
Multivariate Normal Distribution Lecture 4 July 21, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #4-7/21/2011 Slide 1 of 41 Last Time Matrices and vectors Eigenvalues
Statistical tests for SPSS
Statistical tests for SPSS Paolo Coletti A.Y. 2010/11 Free University of Bolzano Bozen Premise This book is a very quick, rough and fast description of statistical tests and their usage. It is explicitly
LCs for Binary Classification
Linear Classifiers A linear classifier is a classifier such that classification is performed by a dot product beteen the to vectors representing the document and the category, respectively. Therefore it
MACHINE LEARNING IN HIGH ENERGY PHYSICS
MACHINE LEARNING IN HIGH ENERGY PHYSICS LECTURE #1 Alex Rogozhnikov, 2015 INTRO NOTES 4 days two lectures, two practice seminars every day this is introductory track to machine learning kaggle competition!
Machine Learning Logistic Regression
Machine Learning Logistic Regression Jeff Howbert Introduction to Machine Learning Winter 2012 1 Logistic regression Name is somewhat misleading. Really a technique for classification, not regression.
Statistical Models in Data Mining
Statistical Models in Data Mining Sargur N. Srihari University at Buffalo The State University of New York Department of Computer Science and Engineering Department of Biostatistics 1 Srihari Flood of
An Introduction to Statistical Machine Learning - Overview -
An Introduction to Statistical Machine Learning - Overview - Samy Bengio bengio@idiap.ch Dalle Molle Institute for Perceptual Artificial Intelligence (IDIAP) CP 592, rue du Simplon 4 1920 Martigny, Switzerland
10-810 /02-710 Computational Genomics. Clustering expression data
10-810 /02-710 Computational Genomics Clustering expression data What is Clustering? Organizing data into clusters such that there is high intra-cluster similarity low inter-cluster similarity Informally,
A.II. Kernel Estimation of Densities
A.II. Kernel Estimation of Densities Olivier Scaillet University of Geneva and Swiss Finance Institute Outline 1 Introduction 2 Issues with Empirical Averages 3 Kernel Estimator 4 Optimal Bandwidth 5 Bivariate
CS 2750 Machine Learning. Lecture 1. Machine Learning. http://www.cs.pitt.edu/~milos/courses/cs2750/ CS 2750 Machine Learning.
Lecture Machine Learning Milos Hauskrecht milos@cs.pitt.edu 539 Sennott Square, x5 http://www.cs.pitt.edu/~milos/courses/cs75/ Administration Instructor: Milos Hauskrecht milos@cs.pitt.edu 539 Sennott
Example 1: Calculate and compare RiskMetrics TM and Historical Standard Deviation Compare the weights of the volatility parameter using,, and.
3.6 Compare and contrast different parametric and non-parametric approaches for estimating conditional volatility. 3.7 Calculate conditional volatility using parametric and non-parametric approaches. Parametric
Summary of Probability
Summary of Probability Mathematical Physics I Rules of Probability The probability of an event is called P(A), which is a positive number less than or equal to 1. The total probability for all possible
Predictive Modeling Techniques in Insurance
Predictive Modeling Techniques in Insurance Tuesday May 5, 2015 JF. Breton Application Engineer 2014 The MathWorks, Inc. 1 Opening Presenter: JF. Breton: 13 years of experience in predictive analytics
Lecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
Age to Age Factor Selection under Changing Development Chris G. Gross, ACAS, MAAA
Age to Age Factor Selection under Changing Development Chris G. Gross, ACAS, MAAA Introduction A common question faced by many actuaries when selecting loss development factors is whether to base the selected
Classification and Regression Trees
Classification and Regression Trees Bob Stine Dept of Statistics, School University of Pennsylvania Trees Familiar metaphor Biology Decision tree Medical diagnosis Org chart Properties Recursive, partitioning
Data Mining: An Overview. David Madigan http://www.stat.columbia.edu/~madigan
Data Mining: An Overview David Madigan http://www.stat.columbia.edu/~madigan Overview Brief Introduction to Data Mining Data Mining Algorithms Specific Eamples Algorithms: Disease Clusters Algorithms:
OUTLIER ANALYSIS. Data Mining 1
OUTLIER ANALYSIS Data Mining 1 What Are Outliers? Outlier: A data object that deviates significantly from the normal objects as if it were generated by a different mechanism Ex.: Unusual credit card purchase,
Basics of Statistical Machine Learning
CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu Modern machine learning is rooted in statistics. You will find many familiar
Big Data: a new era for Statistics
Big Data: a new era for Statistics Richard J. Samworth Abstract Richard Samworth (1996) is a Professor of Statistics in the University s Statistical Laboratory, and has been a Fellow of St John s since
Definition The covariance of X and Y, denoted by cov(x, Y ) is defined by. cov(x, Y ) = E(X µ 1 )(Y µ 2 ).
Correlation Regression Bivariate Normal Suppose that X and Y are r.v. s with joint density f(x y) and suppose that the means of X and Y are respectively µ 1 µ 2 and the variances are 1 2. Definition The
Imputing Values to Missing Data
Imputing Values to Missing Data In federated data, between 30%-70% of the data points will have at least one missing attribute - data wastage if we ignore all records with a missing value Remaining data
Regression analysis. MTAT Data Mining Anna Leontjeva
Regression analysis MTAT.03.183 Data Mining 2016 Anna Leontjeva Previous lecture Supervised vs. Unsupervised Learning? Previous lecture Supervised vs. Unsupervised Learning? Iris setosa Iris versicolor
Sample Size and Power in Clinical Trials
Sample Size and Power in Clinical Trials Version 1.0 May 011 1. Power of a Test. Factors affecting Power 3. Required Sample Size RELATED ISSUES 1. Effect Size. Test Statistics 3. Variation 4. Significance
2DI36 Statistics. 2DI36 Part II (Chapter 7 of MR)
2DI36 Statistics 2DI36 Part II (Chapter 7 of MR) What Have we Done so Far? Last time we introduced the concept of a dataset and seen how we can represent it in various ways But, how did this dataset came
17. SIMPLE LINEAR REGRESSION II
17. SIMPLE LINEAR REGRESSION II The Model In linear regression analysis, we assume that the relationship between X and Y is linear. This does not mean, however, that Y can be perfectly predicted from X.
Sampling Distributions and the Central Limit Theorem
135 Part 2 / Basic Tools of Research: Sampling, Measurement, Distributions, and Descriptive Statistics Chapter 10 Sampling Distributions and the Central Limit Theorem In the previous chapter we explained
CS 2750 Machine Learning. Lecture 1. Machine Learning. CS 2750 Machine Learning.
Lecture 1 Machine Learning Milos Hauskrecht milos@cs.pitt.edu 539 Sennott Square, x-5 http://www.cs.pitt.edu/~milos/courses/cs75/ Administration Instructor: Milos Hauskrecht milos@cs.pitt.edu 539 Sennott
Institute of Actuaries of India Subject CT3 Probability and Mathematical Statistics
Institute of Actuaries of India Subject CT3 Probability and Mathematical Statistics For 2015 Examinations Aim The aim of the Probability and Mathematical Statistics subject is to provide a grounding in
Stat 20: Intro to Probability and Statistics
Stat 20: Intro to Probability and Statistics Lecture 19: Confidence Intervals for Percentages Tessa L. Childers-Day UC Berkeley 28 July 2014 By the end of this lecture... You will be able to: Estimate
Marketing Mix Modelling and Big Data P. M Cain
1) Introduction Marketing Mix Modelling and Big Data P. M Cain Big data is generally defined in terms of the volume and variety of structured and unstructured information. Whereas structured data is stored
Quantifying Seasonal Variation in Cloud Cover with Predictive Models
Quantifying Seasonal Variation in Cloud Cover with Predictive Models Ashok N. Srivastava, Ph.D. ashok@email.arc.nasa.gov Deputy Area Lead, Discovery and Systems Health Group Leader, Intelligent Data Understanding
Average Redistributional Effects. IFAI/IZA Conference on Labor Market Policy Evaluation
Average Redistributional Effects IFAI/IZA Conference on Labor Market Policy Evaluation Geert Ridder, Department of Economics, University of Southern California. October 10, 2006 1 Motivation Most papers
Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression
Logistic Regression Department of Statistics The Pennsylvania State University Email: jiali@stat.psu.edu Logistic Regression Preserve linear classification boundaries. By the Bayes rule: Ĝ(x) = arg max
The Big 50 Revision Guidelines for S1
The Big 50 Revision Guidelines for S1 If you can understand all of these you ll do very well 1. Know what is meant by a statistical model and the Modelling cycle of continuous refinement 2. Understand
Cross-validation for detecting and preventing overfitting
Cross-validation for detecting and preventing overfitting Note to other teachers and users of these slides. Andrew would be delighted if ou found this source material useful in giving our own lectures.
Linear and quadratic Taylor polynomials for functions of several variables.
ams/econ 11b supplementary notes ucsc Linear quadratic Taylor polynomials for functions of several variables. c 010, Yonatan Katznelson Finding the extreme (minimum or maximum) values of a function, is
Statistics, Data Mining and Machine Learning in Astronomy: A Practical Python Guide for the Analysis of Survey Data. and Alex Gray
Statistics, Data Mining and Machine Learning in Astronomy: A Practical Python Guide for the Analysis of Survey Data Željko Ivezić, Andrew J. Connolly, Jacob T. VanderPlas University of Washington and Alex
Advanced Research Methods. Instrumental variables (IV) Regression discontinuity design (RDD)
Instrumental variables (IV) Regression discontinuity design (RDD) Lecture 2 INSTRUMENTAL VARIABLES (IV) EXAMPLE Consider once more your favourite training program for unemployed workers. Assume this time
Inferential Statistics
Inferential Statistics Sampling and the normal distribution Z-scores Confidence levels and intervals Hypothesis testing Commonly used statistical methods Inferential Statistics Descriptive statistics are
CS 688 Pattern Recognition Lecture 4. Linear Models for Classification
CS 688 Pattern Recognition Lecture 4 Linear Models for Classification Probabilistic generative models Probabilistic discriminative models 1 Generative Approach ( x ) p C k p( C k ) Ck p ( ) ( x Ck ) p(
Principal Components Regression. Principal Components Regression. Principal Components Regression. Principal Components Regression
Principal Components Regression Principal Components Regression y i = β 0 + β j x ij + ɛ i Usual least squares may be inappropriate if x is high dimensional, especially in comparison to sample size x is
Introduction to knn Classification and CNN Data Reduction
Introduction to knn Classification and CNN Data Reduction Oliver Sutton February, 2012 1 / 29 1 The Classification Problem Examples The Problem 2 The k Nearest Neighbours Algorithm The Nearest Neighbours
Java Modules for Time Series Analysis
Java Modules for Time Series Analysis Agenda Clustering Non-normal distributions Multifactor modeling Implied ratings Time series prediction 1. Clustering + Cluster 1 Synthetic Clustering + Time series
A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution
A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 4: September
Facebook Friend Suggestion Eytan Daniyalzade and Tim Lipus
Facebook Friend Suggestion Eytan Daniyalzade and Tim Lipus 1. Introduction Facebook is a social networking website with an open platform that enables developers to extract and utilize user information
Inference for Regression
Simple Linear Regression Inference for Regression The simple linear regression model Estimating regression parameters; Confidence intervals and significance tests for regression parameters Inference about
Automated Statistical Modeling for Data Mining David Stephenson 1
Automated Statistical Modeling for Data Mining David Stephenson 1 Abstract. We seek to bridge the gap between basic statistical data mining tools and advanced statistical analysis software that requires
Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data
CMPE 59H Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data Term Project Report Fatma Güney, Kübra Kalkan 1/15/2013 Keywords: Non-linear
the number of organisms in the squares of a haemocytometer? the number of goals scored by a football team in a match?
Poisson Random Variables (Rees: 6.8 6.14) Examples: What is the distribution of: the number of organisms in the squares of a haemocytometer? the number of hits on a web site in one hour? the number of
KATE GLEASON COLLEGE OF ENGINEERING. John D. Hromi Center for Quality and Applied Statistics
ROCHESTER INSTITUTE OF TECHNOLOGY COURSE OUTLINE FORM KATE GLEASON COLLEGE OF ENGINEERING John D. Hromi Center for Quality and Applied Statistics NEW (or REVISED) COURSE (KGCOE- CQAS- 747- Principles of
Classification Techniques for Remote Sensing
Classification Techniques for Remote Sensing Selim Aksoy Department of Computer Engineering Bilkent University Bilkent, 06800, Ankara saksoy@cs.bilkent.edu.tr http://www.cs.bilkent.edu.tr/ saksoy/courses/cs551
Continuous Distributions, Mainly the Normal Distribution
Continuous Distributions, Mainly the Normal Distribution 1 Continuous Random Variables STA 281 Fall 2011 Discrete distributions place probability on specific numbers. A Bin(n,p) distribution, for example,
TRANSCRIPT: In this lecture, we will talk about both theoretical and applied concepts related to hypothesis testing.
This is Dr. Chumney. The focus of this lecture is hypothesis testing both what it is, how hypothesis tests are used, and how to conduct hypothesis tests. 1 In this lecture, we will talk about both theoretical
Physics Lab Report Guidelines
Physics Lab Report Guidelines Summary The following is an outline of the requirements for a physics lab report. A. Experimental Description 1. Provide a statement of the physical theory or principle observed
PS 271B: Quantitative Methods II. Lecture Notes
PS 271B: Quantitative Methods II Lecture Notes Langche Zeng zeng@ucsd.edu The Empirical Research Process; Fundamental Methodological Issues 2 Theory; Data; Models/model selection; Estimation; Inference.
A General Framework for Mining Concept-Drifting Data Streams with Skewed Distributions
A General Framework for Mining Concept-Drifting Data Streams with Skewed Distributions Jing Gao Wei Fan Jiawei Han Philip S. Yu University of Illinois at Urbana-Champaign IBM T. J. Watson Research Center
COLLEGE OF SCIENCE. John D. Hromi Center for Quality and Applied Statistics
ROCHESTER INSTITUTE OF TECHNOLOGY COURSE OUTLINE FORM COLLEGE OF SCIENCE John D. Hromi Center for Quality and Applied Statistics NEW (or REVISED) COURSE: COS-STAT-747 Principles of Statistical Data Mining
MATH 10: Elementary Statistics and Probability Chapter 9: Hypothesis Testing with One Sample
MATH 10: Elementary Statistics and Probability Chapter 9: Hypothesis Testing with One Sample Tony Pourmohamad Department of Mathematics De Anza College Spring 2015 Objectives By the end of this set of
C19 Machine Learning
C9 Machine Learning 8 Lectures Hilary Term 25 2 Tutorial Sheets A. Zisserman Overview: Supervised classification perceptron, support vector machine, loss functions, kernels, random forests, neural networks
Two Topics in Parametric Integration Applied to Stochastic Simulation in Industrial Engineering
Two Topics in Parametric Integration Applied to Stochastic Simulation in Industrial Engineering Department of Industrial Engineering and Management Sciences Northwestern University September 15th, 2014
Michael R. Pinsky, M.D., C.M., Dr.h.c., FCCP, MCCM Professor of Critical Care Medicine, Bioengineering, Anesthesiology, Cardiovascular Diseases, and
Michael R. Pinsky, M.D., C.M., Dr.h.c., FCCP, MCCM Professor of Critical Care Medicine, Bioengineering, Anesthesiology, Cardiovascular Diseases, and Clinical & Translational Sciences, Vice Chair for Academic
Appendix 1: Time series analysis of peak-rate years and synchrony testing.
Appendix 1: Time series analysis of peak-rate years and synchrony testing. Overview The raw data are accessible at Figshare ( Time series of global resources, DOI 10.6084/m9.figshare.929619), sources are
BIOE 370 1. Lotka-Volterra Model L-V model with density-dependent prey population growth
BIOE 370 1 Populus Simulations of Predator-Prey Population Dynamics. Lotka-Volterra Model L-V model with density-dependent prey population growth Theta-Logistic Model Effects on dynamics of different functional
Pr(X = x) = f(x) = λe λx
Old Business - variance/std. dev. of binomial distribution - mid-term (day, policies) - class strategies (problems, etc.) - exponential distributions New Business - Central Limit Theorem, standard error
Session 3: Marketing Research
Session 3: Marketing Research Marketing Management Nanda Kumar, Ph.D. Agenda Marketing Research Qualitative Analysis Quantitative Analysis Quantitative Analysis Regression Analysis Analyzing survey data
Interaction between quantitative predictors
Interaction between quantitative predictors In a first-order model like the ones we have discussed, the association between E(y) and a predictor x j does not depend on the value of the other predictors
Introduction. example of a AA curve appears at the end of this presentation.
1 Introduction The High Quality Market (HQM) Corporate Bond Yield Curve for the Pension Protection Act (PPA) uses a methodology developed at Treasury to construct yield curves from extended regressions
SUMAN DUVVURU STAT 567 PROJECT REPORT
SUMAN DUVVURU STAT 567 PROJECT REPORT SURVIVAL ANALYSIS OF HEROIN ADDICTS Background and introduction: Current illicit drug use among teens is continuing to increase in many countries around the world.
EC 112 SAMPLING, ESTIMATION AND HYPOTHESIS TESTING. One and One Half Hours (1 1 2 Hours) TWO questions should be answered
ECONOMICS EC 112 SAMPLING, ESTIMATION AND HYPOTHESIS TESTING One and One Half Hours (1 1 2 Hours) TWO questions should be answered Statistical tables are provided. The approved calculator (that is, Casio