CMPSCI 240: Reasoning about Uncertainty

Size: px
Start display at page:

Download "CMPSCI 240: Reasoning about Uncertainty"

Transcription

1 CMPSCI 240: Reasoning about Uncertainty Lecture 18: Spam Filtering and Naive Bayes Classification Andrew McGregor University of Massachusetts Last Compiled: April 9, 2015

2 Review Total Probability If A 1,..., A n partition Ω then for any event B: P(B) P(B A 1 )P(A 1 ) P(B A n )P(A n ) n P(B A i )P(A i ) Bayes Theorem If A 1,..., A n partition Ω then for any event B: P(A i B) P(A i B) P(B) i1 P(B A i )P(A i ) n j1 P(B A j)p(a j )

3 Spam Filtering When you receive an you have two hypotheses : H 1 is spam H 2 is not spam and note that these are partitioning events. You have some observed data e.g., D contains the word cash How do you use the observed data to pick one of the hypotheses? We want to pick the maximum a posteriori hypothesis (MAP), i.e., the hypothesis H i than maximizes P(H i D).

4 Using Bayes Theorem for Spam Filtering By Bayes theorem: P(H i D) P(H i D) P(D) P(D H i )P(H i ) We need to know the prior probabilities of H 1 and H 2 : P(H 1 )? P(H 2 )? i.e., before we knew anything above the , how likely was the to be spam or not. We also need to know that likelihood of the observed data given each hypothesis: P(D H 1 )? P(D H 2 )? i.e., for each hypothesis, what s the probability that we would see the observed data?

5 Computing Priors and Likelihoods To compute priors and likelihoods we use training or historical data. Suppose that we have analyzed 1000 pervious s and found: 700 of the s were spam 300 of the s were not spam From this, it would be natural to believe that 70% of future s were spam and that 30% of future s were not spam. Hence, we set the priors to be P (H 1 ) 0.7 and P (H 2 ) 0.3. Suppose we have also found that 350 of the spam s include the word cash 100 of the non-spam s include the word cash From this, it s natural to set the likelihoods to be P (D H 1 ) 1/2 and P (D H 2 ) 1/3.

6 Putting it all together If P (H 1 ) 0.7, P (H 2 ) 0.3, P (D H 1 ) 1/2, P (D H 2 ) 1/3 then: P(H 1 D) P(H 1 D) P(D) and similarly P(D H 1 )P(H 1 ) 1/ P(H 2 D) 0.3 1/3 0.1 Therefore the MAP hypothesis is H 1.

7 Combining Observed Data Suppose we have two pieces of observed data: D 1 contains the word cash D 2 contains the word pharmacy The MAP hypothesis is the one maximizing P(H j D 1 D 2 ) where P(H j D 1 D 2 ) P(D 1 D 2 H j )P(H j ) P(D 1 D 2 H 1 )P(H 1 ) + P(D 2 D 2 H 2 )P(H 2 )

8 Naive Bayes Classification In Naive Bayes Classification (NBC) we assume that the observed data is independent conditioned on either of the hypotheses, i.e., P(D 1 D 2 H 1 ) P(D 1 H 1 )P(D 2 H 1 ) P(D 1 D 2 H 2 ) P(D 1 H 2 )P(D 2 H 2 ) And therefore NBC picks the hypothesis H j that maximizes P(D 1 D 2 H j )P(H j ) P(D 1 H j )P(D 2 H j )P(H j ) Hence, we just need to compute the priors P(H 1 ), P(H 2 ) and the likelihoods P(D 1 H 1 ), P(D 2 H 1 ), P(D 1 H 2 ), P(D 2 H 2 ) from the training data.

9 Picking an Hypothesis Suppose we have multiple hypotheses H 1, H 2,... H k and we have priors for these hypotheses P(H 1 ), P(H 2 ),..., P(H k ). If we observe some event D, the MAP hypothesis is the hypothesis H i that maximizes P(H i D) P(D H i)p(h i ) j P(D H j)p(h j ) To find this we need likelihoods P(D H j ) for each hypothesis H j. If we observe multiple events D 1, D 2,..., D t, the MAP hypothesis is the hypothesis H i that maximizes P(H i D 1 D 2... D t ) P(D 1 D 2... D t H i )P(H i ) j P(D 1 D 2... D t H j )P(H j ) In Naive Bayes Classification we assume that for each j: P(D 1 D 2... D t H j ) P(D 1 H j )P(D 2 H j )... P(D t H j )

10 Combing Evidence: Bird Watching You see a blue parrot and you hypothesize that it s a Norwegian Blue Your birdwatching book says: Only 10% of blue parrots are Norwegian Blues Norwegian Blues spend 60% of their time lying down Other blue parrots only spend 20% of their time lying down 80% of Norwegian Blues have lovely plumage 20% of other blue parrots have lovely plumage. The blue parrot you observe is lying down (D 1 ) and has lovely plumage (D 2 ): what is the likelihood it s a Norwegian Blue (N)? P (N D 1 D 2 ) P (D 1 D 2 N) P (N) P (D 1 D 2 N) P (N) + P (D 1 D 2 N c ) P (N c ) Not enough information for P (D 1 D 2 N) & P (D 1 D 2 N c ) Could use NBC: Assume P (D 1 D 2 N) P (D 1 N) P (D 2 N) and P (D 1 D 2 N c ) P (D 1 N c ) P (D 2 N c )

Math 141. Lecture 2: More Probability! Albyn Jones 1. jones@reed.edu www.people.reed.edu/ jones/courses/141. 1 Library 304. Albyn Jones Math 141

Math 141. Lecture 2: More Probability! Albyn Jones 1. jones@reed.edu www.people.reed.edu/ jones/courses/141. 1 Library 304. Albyn Jones Math 141 Math 141 Lecture 2: More Probability! Albyn Jones 1 1 Library 304 jones@reed.edu www.people.reed.edu/ jones/courses/141 Outline Law of total probability Bayes Theorem the Multiplication Rule, again Recall

More information

1 Introductory Comments. 2 Bayesian Probability

1 Introductory Comments. 2 Bayesian Probability Introductory Comments First, I would like to point out that I got this material from two sources: The first was a page from Paul Graham s website at www.paulgraham.com/ffb.html, and the second was a paper

More information

Question 2 Naïve Bayes (16 points)

Question 2 Naïve Bayes (16 points) Question 2 Naïve Bayes (16 points) About 2/3 of your email is spam so you downloaded an open source spam filter based on word occurrences that uses the Naive Bayes classifier. Assume you collected the

More information

Introduction to Bayesian Classification (A Practical Discussion) Todd Holloway Lecture for B551 Nov. 27, 2007

Introduction to Bayesian Classification (A Practical Discussion) Todd Holloway Lecture for B551 Nov. 27, 2007 Introduction to Bayesian Classification (A Practical Discussion) Todd Holloway Lecture for B551 Nov. 27, 2007 Naïve Bayes Components ML vs. MAP Benefits Feature Preparation Filtering Decay Extended Examples

More information

Bayesian Updating with Discrete Priors Class 11, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom

Bayesian Updating with Discrete Priors Class 11, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom 1 Learning Goals Bayesian Updating with Discrete Priors Class 11, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom 1. Be able to apply Bayes theorem to compute probabilities. 2. Be able to identify

More information

Bayes Theorem. Susanna Kujanpää OUAS 2.12.2015

Bayes Theorem. Susanna Kujanpää OUAS 2.12.2015 ayes Theorem Susanna Kujanpää OUS 2.12.2015 ayes Theorem Thomas ayes 1701 1761 This is a theorem with two distinct interpretations. 1 ayesian interpretation: it shows how a subjective degree of belief

More information

Bayes and Naïve Bayes. cs534-machine Learning

Bayes and Naïve Bayes. cs534-machine Learning Bayes and aïve Bayes cs534-machine Learning Bayes Classifier Generative model learns Prediction is made by and where This is often referred to as the Bayes Classifier, because of the use of the Bayes rule

More information

People have thought about, and defined, probability in different ways. important to note the consequences of the definition:

People have thought about, and defined, probability in different ways. important to note the consequences of the definition: PROBABILITY AND LIKELIHOOD, A BRIEF INTRODUCTION IN SUPPORT OF A COURSE ON MOLECULAR EVOLUTION (BIOL 3046) Probability The subject of PROBABILITY is a branch of mathematics dedicated to building models

More information

1 Maximum likelihood estimation

1 Maximum likelihood estimation COS 424: Interacting with Data Lecturer: David Blei Lecture #4 Scribes: Wei Ho, Michael Ye February 14, 2008 1 Maximum likelihood estimation 1.1 MLE of a Bernoulli random variable (coin flips) Given N

More information

The central question: Does the bankruptcy process make sure that firms in bankruptcy are inefficient, and that inefficient firms end up in bankruptcy?

The central question: Does the bankruptcy process make sure that firms in bankruptcy are inefficient, and that inefficient firms end up in bankruptcy? Bankruptcy White (1989) Bankruptcy the closing down of firms that are inefficient, use outdated technologies, or produce products that are in excess supply. The central question: Does the bankruptcy process

More information

When 95% Accurate Isn t Written by: Todd CadwalladerOlsker California State University, Fullerton tcadwall@fullerton.edu

When 95% Accurate Isn t Written by: Todd CadwalladerOlsker California State University, Fullerton tcadwall@fullerton.edu When 95% Accurate Isn t Written by: Todd CadwalladerOlsker California State University, Fullerton tcadwall@fullerton.edu Overview of Lesson In this activity, students will investigate Bayes theorem using

More information

Naive-Bayes Classification Algorithm

Naive-Bayes Classification Algorithm Naive-Bayes Classification Algorithm 1. Introduction to Bayesian Classification The Bayesian Classification represents a supervised learning method as well as a statistical method for classification. Assumes

More information

Lecture 9: Introduction to Pattern Analysis

Lecture 9: Introduction to Pattern Analysis Lecture 9: Introduction to Pattern Analysis g Features, patterns and classifiers g Components of a PR system g An example g Probability definitions g Bayes Theorem g Gaussian densities Features, patterns

More information

An Efficient Spam Filtering Techniques for Email Account

An Efficient Spam Filtering Techniques for Email Account American Journal of Engineering Research (AJER) e-issn : 2320-0847 p-issn : 2320-0936 Volume-02, Issue-10, pp-63-73 www.ajer.org Research Paper Open Access An Efficient Spam Filtering Techniques for Email

More information

Naïve Bayes and Hadoop. Shannon Quinn

Naïve Bayes and Hadoop. Shannon Quinn Naïve Bayes and Hadoop Shannon Quinn http://xkcd.com/ngram-charts/ Coupled Temporal Scoping of Relational Facts. P.P. Talukdar, D.T. Wijaya and T.M. Mitchell. In Proceedings of the ACM International Conference

More information

CS 348: Introduction to Artificial Intelligence Lab 2: Spam Filtering

CS 348: Introduction to Artificial Intelligence Lab 2: Spam Filtering THE PROBLEM Spam is e-mail that is both unsolicited by the recipient and sent in substantively identical form to many recipients. In 2004, MSNBC reported that spam accounted for 66% of all electronic mail.

More information

On Attacking Statistical Spam Filters

On Attacking Statistical Spam Filters On Attacking Statistical Spam Filters Gregory L. Wittel and S. Felix Wu Department of Computer Science University of California, Davis One Shields Avenue, Davis, CA 95616 USA Paper review by Deepak Chinavle

More information

Lecture 9: Bayesian hypothesis testing

Lecture 9: Bayesian hypothesis testing Lecture 9: Bayesian hypothesis testing 5 November 27 In this lecture we ll learn about Bayesian hypothesis testing. 1 Introduction to Bayesian hypothesis testing Before we go into the details of Bayesian

More information

The Mean Value Theorem

The Mean Value Theorem The Mean Value Theorem THEOREM (The Extreme Value Theorem): If f is continuous on a closed interval [a, b], then f attains an absolute maximum value f(c) and an absolute minimum value f(d) at some numbers

More information

Bayesian Analysis for the Social Sciences

Bayesian Analysis for the Social Sciences Bayesian Analysis for the Social Sciences Simon Jackman Stanford University http://jackman.stanford.edu/bass November 9, 2012 Simon Jackman (Stanford) Bayesian Analysis for the Social Sciences November

More information

Statistics and Data Analysis B01.1305

Statistics and Data Analysis B01.1305 Statistics and Data Analysis B01.1305 Professor William Greene Phone: 212.998.0876 Office: KMC 7-78 Home page: www.stern.nyu.edu/~wgreene Email: wgreene@stern.nyu.edu Course web page: www.stern.nyu.edu/~wgreene/statistics/outline.htm

More information

" Y. Notation and Equations for Regression Lecture 11/4. Notation:

 Y. Notation and Equations for Regression Lecture 11/4. Notation: Notation: Notation and Equations for Regression Lecture 11/4 m: The number of predictor variables in a regression Xi: One of multiple predictor variables. The subscript i represents any number from 1 through

More information

Welcome to the HVAC Sales Academy Learning Management System (LMS) Walkthrough:

Welcome to the HVAC Sales Academy Learning Management System (LMS) Walkthrough: Welcome to the HVAC Sales Academy Learning Management System (LMS) Walkthrough: In this walkthrough we will show you the basics of navigating around inside our new LMS system. This should help make it

More information

Probabilities. Probability of a event. From Random Variables to Events. From Random Variables to Events. Probability Theory I

Probabilities. Probability of a event. From Random Variables to Events. From Random Variables to Events. Probability Theory I Victor Adamchi Danny Sleator Great Theoretical Ideas In Computer Science Probability Theory I CS 5-25 Spring 200 Lecture Feb. 6, 200 Carnegie Mellon University We will consider chance experiments with

More information

Lecture 1 Introduction Properties of Probability Methods of Enumeration Asrat Temesgen Stockholm University

Lecture 1 Introduction Properties of Probability Methods of Enumeration Asrat Temesgen Stockholm University Lecture 1 Introduction Properties of Probability Methods of Enumeration Asrat Temesgen Stockholm University 1 Chapter 1 Probability 1.1 Basic Concepts In the study of statistics, we consider experiments

More information

Prediction of Heart Disease Using Naïve Bayes Algorithm

Prediction of Heart Disease Using Naïve Bayes Algorithm Prediction of Heart Disease Using Naïve Bayes Algorithm R.Karthiyayini 1, S.Chithaara 2 Assistant Professor, Department of computer Applications, Anna University, BIT campus, Tiruchirapalli, Tamilnadu,

More information

University of Chicago Graduate School of Business. Business 41000: Business Statistics Solution Key

University of Chicago Graduate School of Business. Business 41000: Business Statistics Solution Key Name: OUTLINE SOLUTIONS University of Chicago Graduate School of Business Business 41000: Business Statistics Solution Key Special Notes: 1. This is a closed-book exam. You may use an 8 11 piece of paper

More information

Learning from Data: Naive Bayes

Learning from Data: Naive Bayes Semester 1 http://www.anc.ed.ac.uk/ amos/lfd/ Naive Bayes Typical example: Bayesian Spam Filter. Naive means naive. Bayesian methods can be much more sophisticated. Basic assumption: conditional independence.

More information

Attribution. Modified from Stuart Russell s slides (Berkeley) Parts of the slides are inspired by Dan Klein s lecture material for CS 188 (Berkeley)

Attribution. Modified from Stuart Russell s slides (Berkeley) Parts of the slides are inspired by Dan Klein s lecture material for CS 188 (Berkeley) Machine Learning 1 Attribution Modified from Stuart Russell s slides (Berkeley) Parts of the slides are inspired by Dan Klein s lecture material for CS 188 (Berkeley) 2 Outline Inductive learning Decision

More information

Tweaking Naïve Bayes classifier for intelligent spam detection

Tweaking Naïve Bayes classifier for intelligent spam detection 682 Tweaking Naïve Bayes classifier for intelligent spam detection Ankita Raturi 1 and Sunil Pranit Lal 2 1 University of California, Irvine, CA 92697, USA. araturi@uci.edu 2 School of Computing, Information

More information

Week 2: Conditional Probability and Bayes formula

Week 2: Conditional Probability and Bayes formula Week 2: Conditional Probability and Bayes formula We ask the following question: suppose we know that a certain event B has occurred. How does this impact the probability of some other A. This question

More information

Bayesian Tutorial (Sheet Updated 20 March)

Bayesian Tutorial (Sheet Updated 20 March) Bayesian Tutorial (Sheet Updated 20 March) Practice Questions (for discussing in Class) Week starting 21 March 2016 1. What is the probability that the total of two dice will be greater than 8, given that

More information

STA 371G: Statistics and Modeling

STA 371G: Statistics and Modeling STA 371G: Statistics and Modeling Decision Making Under Uncertainty: Probability, Betting Odds and Bayes Theorem Mingyuan Zhou McCombs School of Business The University of Texas at Austin http://mingyuanzhou.github.io/sta371g

More information

Causality and the Doomsday Argument

Causality and the Doomsday Argument Causality and the Doomsday Argument Ivan Phillips The Futurists Guild January 24 th 2005 Abstract Using the Autodialer thought experiment, we show that the Self-Sampling Assumption (SSA) is too general,

More information

Projektgruppe. Categorization of text documents via classification

Projektgruppe. Categorization of text documents via classification Projektgruppe Steffen Beringer Categorization of text documents via classification 4. Juni 2010 Content Motivation Text categorization Classification in the machine learning Document indexing Construction

More information

BUILDING A SPAM FILTER USING NAÏVE BAYES. CIS 391- Intro to AI 1

BUILDING A SPAM FILTER USING NAÏVE BAYES. CIS 391- Intro to AI 1 BUILDING A SPAM FILTER USING NAÏVE BAYES 1 Spam or not Spam: that is the question. From: "" Subjet: real estate is the only way... gem oalvgkay Anyone an buy real estate with no

More information

E3: PROBABILITY AND STATISTICS lecture notes

E3: PROBABILITY AND STATISTICS lecture notes E3: PROBABILITY AND STATISTICS lecture notes 2 Contents 1 PROBABILITY THEORY 7 1.1 Experiments and random events............................ 7 1.2 Certain event. Impossible event............................

More information

Performance Analysis of Naive Bayes and J48 Classification Algorithm for Data Classification

Performance Analysis of Naive Bayes and J48 Classification Algorithm for Data Classification Performance Analysis of Naive Bayes and J48 Classification Algorithm for Data Classification Tina R. Patil, Mrs. S. S. Sherekar Sant Gadgebaba Amravati University, Amravati tnpatil2@gmail.com, ss_sherekar@rediffmail.com

More information

Lecture 5/6. Enrico Bini

Lecture 5/6. Enrico Bini Advanced Lecture 5/6 November 8, 212 Outline 1 2 3 Theorem (Lemma 3 in [?]) Space of feasible C i N is EDF-schedulable if and only if i U i 1 and: t D n i=1 { max, t + Ti D i T i } C i t with D = {d i,k

More information

Discrete Structures for Computer Science

Discrete Structures for Computer Science Discrete Structures for Computer Science Adam J. Lee adamlee@cs.pitt.edu 6111 Sennott Square Lecture #20: Bayes Theorem November 5, 2013 How can we incorporate prior knowledge? Sometimes we want to know

More information

Free Human Memory Experiments For College Students

Free Human Memory Experiments For College Students UNIVERSITY OF EDINBURGH COLLEGE OF SCIENCE AND ENGINEERING SCHOOL OF INFORMATICS COGNITIVE MODELLING (LEVEL 10) COGNITIVE MODELLING (LEVEL 11) Friday 1 April 2005 00:00 to 00:00 Year 4 Courses Convener:

More information

Decision Trees and Networks

Decision Trees and Networks Lecture 21: Uncertainty 6 Today s Lecture Victor R. Lesser CMPSCI 683 Fall 2010 Decision Trees and Networks Decision Trees A decision tree is an explicit representation of all the possible scenarios from

More information

Machine Learning in Spam Filtering

Machine Learning in Spam Filtering Machine Learning in Spam Filtering A Crash Course in ML Konstantin Tretyakov kt@ut.ee Institute of Computer Science, University of Tartu Overview Spam is Evil ML for Spam Filtering: General Idea, Problems.

More information

Unit 19: Probability Models

Unit 19: Probability Models Unit 19: Probability Models Summary of Video Probability is the language of uncertainty. Using statistics, we can better predict the outcomes of random phenomena over the long term from the very complex,

More information

Full and Complete Binary Trees

Full and Complete Binary Trees Full and Complete Binary Trees Binary Tree Theorems 1 Here are two important types of binary trees. Note that the definitions, while similar, are logically independent. Definition: a binary tree T is full

More information

ISAAC LEVI JAAKKO HINTIKKA

ISAAC LEVI JAAKKO HINTIKKA ISAAC LEVI JAAKKO HINTIKKA I agree with Jaakko Hintikka that the so-called conjunction fallacy of Kahneman and Tversky is no fallacy. I prefer a different explanation of the mistake made these authors

More information

Lecture Note 1 Set and Probability Theory. MIT 14.30 Spring 2006 Herman Bennett

Lecture Note 1 Set and Probability Theory. MIT 14.30 Spring 2006 Herman Bennett Lecture Note 1 Set and Probability Theory MIT 14.30 Spring 2006 Herman Bennett 1 Set Theory 1.1 Definitions and Theorems 1. Experiment: any action or process whose outcome is subject to uncertainty. 2.

More information

Math 58. Rumbos Fall 2008 1. Solutions to Assignment #3

Math 58. Rumbos Fall 2008 1. Solutions to Assignment #3 Math 58. Rumbos Fall 2008 1 Solutions to Assignment #3 1. Use randomization to test the null hypothesis that there is no difference between calcium supplementation and a placebo for the experimental data

More information

Statistical Machine Learning from Data

Statistical Machine Learning from Data Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Gaussian Mixture Models Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique

More information

Economics 1011a: Intermediate Microeconomics

Economics 1011a: Intermediate Microeconomics Lecture 12: More Uncertainty Economics 1011a: Intermediate Microeconomics Lecture 12: More on Uncertainty Thursday, October 23, 2008 Last class we introduced choice under uncertainty. Today we will explore

More information

An evolutionary learning spam filter system

An evolutionary learning spam filter system An evolutionary learning spam filter system Catalin Stoean 1, Ruxandra Gorunescu 2, Mike Preuss 3, D. Dumitrescu 4 1 University of Craiova, Romania, catalin.stoean@inf.ucv.ro 2 University of Craiova, Romania,

More information

Less naive Bayes spam detection

Less naive Bayes spam detection Less naive Bayes spam detection Hongming Yang Eindhoven University of Technology Dept. EE, Rm PT 3.27, P.O.Box 53, 5600MB Eindhoven The Netherlands. E-mail:h.m.yang@tue.nl also CoSiNe Connectivity Systems

More information

Homework 8 Solutions

Homework 8 Solutions CSE 21 - Winter 2014 Homework Homework 8 Solutions 1 Of 330 male and 270 female employees at the Flagstaff Mall, 210 of the men and 180 of the women are on flex-time (flexible working hours). Given that

More information

Spam Filter Optimality Based on Signal Detection Theory

Spam Filter Optimality Based on Signal Detection Theory Spam Filter Optimality Based on Signal Detection Theory ABSTRACT Singh Kuldeep NTNU, Norway HUT, Finland kuldeep@unik.no Md. Sadek Ferdous NTNU, Norway University of Tartu, Estonia sadek@unik.no Unsolicited

More information

THE FUNDAMENTAL THEOREM OF ARBITRAGE PRICING

THE FUNDAMENTAL THEOREM OF ARBITRAGE PRICING THE FUNDAMENTAL THEOREM OF ARBITRAGE PRICING 1. Introduction The Black-Scholes theory, which is the main subject of this course and its sequel, is based on the Efficient Market Hypothesis, that arbitrages

More information

The Economic Value of Certainty By Les McGuire, MBA August 19, 2003

The Economic Value of Certainty By Les McGuire, MBA August 19, 2003 The Economic Value of Certainty By Les McGuire, MBA August 19, 2003 This is supposed to be an article about Whole Life Insurance, and for those who have eyes to see it is. However, before discussing product,

More information

BAYES' THEOREM IN DECISION MAKING Reasoning from Effect to Cause

BAYES' THEOREM IN DECISION MAKING Reasoning from Effect to Cause BAYES' THEOREM IN DECISION MAKING Reasoning from Effect to Cause by Jack V. Michaels, PE, CVS Value Engineering Department Martin Marietta Orlando Aerospace ABSTRACT Thomas Bayes (1702-1761) was an English

More information

Experimental Uncertainty and Probability

Experimental Uncertainty and Probability 02/04/07 PHY310: Statistical Data Analysis 1 PHY310: Lecture 03 Experimental Uncertainty and Probability Road Map The meaning of experimental uncertainty The fundamental concepts of probability 02/04/07

More information

R Simulations: Monty Hall problem

R Simulations: Monty Hall problem R Simulations: Monty Hall problem Monte Carlo Simulations Monty Hall Problem Statistical Analysis Simulation in R Exercise 1: A Gift Giving Puzzle Exercise 2: Gambling Problem R Simulations: Monty Hall

More information

Exception Handling. development.

Exception Handling. development. Exception Handling Name: ID: These questions do not have a formal, definitive answer. They are meant to be food for thoughts. Feel free to seek answers on browsing the Internet, talking to other software

More information

Email Spam Detection A Machine Learning Approach

Email Spam Detection A Machine Learning Approach Email Spam Detection A Machine Learning Approach Ge Song, Lauren Steimle ABSTRACT Machine learning is a branch of artificial intelligence concerned with the creation and study of systems that can learn

More information

Log-Linear Models a.k.a. Logistic Regression, Maximum Entropy Models

Log-Linear Models a.k.a. Logistic Regression, Maximum Entropy Models Log-Linear Models a.k.a. Logistic Regression, Maximum Entropy Models Natural Language Processing CS 6120 Spring 2014 Northeastern University David Smith (some slides from Jason Eisner and Dan Klein) summary

More information

The "Dutch Book" argument, tracing back to independent work by. F.Ramsey (1926) and B.deFinetti (1937), offers prudential grounds for

The Dutch Book argument, tracing back to independent work by. F.Ramsey (1926) and B.deFinetti (1937), offers prudential grounds for The "Dutch Book" argument, tracing back to independent work by F.Ramsey (1926) and B.deFinetti (1937), offers prudential grounds for action in conformity with personal probability. Under several structural

More information

How To Value Cash Flow

How To Value Cash Flow Lecture: II 1 Time Value of Money (TVM) A dollar today is more valuable than a dollar sometime in the future...! The intuitive basis for present value what determines the effect of timing on the value

More information

MAS108 Probability I

MAS108 Probability I 1 QUEEN MARY UNIVERSITY OF LONDON 2:30 pm, Thursday 3 May, 2007 Duration: 2 hours MAS108 Probability I Do not start reading the question paper until you are instructed to by the invigilators. The paper

More information

Linear Classification. Volker Tresp Summer 2015

Linear Classification. Volker Tresp Summer 2015 Linear Classification Volker Tresp Summer 2015 1 Classification Classification is the central task of pattern recognition Sensors supply information about an object: to which class do the object belong

More information

ASAV Configuration Advanced Spam Filtering

ASAV Configuration Advanced Spam Filtering ASAV Configuration Advanced Spam Filtering Step 1: Login to http://asav.mediaring.sg/ using the login credentials supplied in the Spam, Virus (ASAV) activation email. Step 2: Configuring Protection Level

More information

Customer Classification And Prediction Based On Data Mining Technique

Customer Classification And Prediction Based On Data Mining Technique Customer Classification And Prediction Based On Data Mining Technique Ms. Neethu Baby 1, Mrs. Priyanka L.T 2 1 M.E CSE, Sri Shakthi Institute of Engineering and Technology, Coimbatore 2 Assistant Professor

More information

Email Filters that use Spammy Words Only

Email Filters that use Spammy Words Only Email Filters that use Spammy Words Only Vasanth Elavarasan Department of Computer Science University of Texas at Austin Advisors: Mohamed Gouda Department of Computer Science University of Texas at Austin

More information

Data Modeling & Analysis Techniques. Probability & Statistics. Manfred Huber 2011 1

Data Modeling & Analysis Techniques. Probability & Statistics. Manfred Huber 2011 1 Data Modeling & Analysis Techniques Probability & Statistics Manfred Huber 2011 1 Probability and Statistics Probability and statistics are often used interchangeably but are different, related fields

More information

Conditional Probability, Hypothesis Testing, and the Monty Hall Problem

Conditional Probability, Hypothesis Testing, and the Monty Hall Problem Conditional Probability, Hypothesis Testing, and the Monty Hall Problem Ernie Croot September 17, 2008 On more than one occasion I have heard the comment Probability does not exist in the real world, and

More information

Data Mining. Nonlinear Classification

Data Mining. Nonlinear Classification Data Mining Unit # 6 Sajjad Haider Fall 2014 1 Nonlinear Classification Classes may not be separable by a linear boundary Suppose we randomly generate a data set as follows: X has range between 0 to 15

More information

Lecture 14: Section 3.3

Lecture 14: Section 3.3 Lecture 14: Section 3.3 Shuanglin Shao October 23, 2013 Definition. Two nonzero vectors u and v in R n are said to be orthogonal (or perpendicular) if u v = 0. We will also agree that the zero vector in

More information

3818 - Introduction to Statistics (Online) Syllabus/Course Information

3818 - Introduction to Statistics (Online) Syllabus/Course Information 3818 - Introduction to Statistics (Online) Syllabus/Course Information Course Description Econ 3818 is a first course in probability and statistical methods, with an introduction to econometrics. This

More information

Evaluation of Candidates for Norwegian Doctoral Degrees

Evaluation of Candidates for Norwegian Doctoral Degrees Evaluation of Candidates for Norwegian Doctoral Degrees Guidelines for the evaluation of candidates for doctoral degrees at the University of Tromsø Recommended by the Norwegian Association of Higher education

More information

Lecture: Mon 13:30 14:50 Fri 9:00-10:20 ( LTH, Lift 27-28) Lab: Fri 12:00-12:50 (Rm. 4116)

Lecture: Mon 13:30 14:50 Fri 9:00-10:20 ( LTH, Lift 27-28) Lab: Fri 12:00-12:50 (Rm. 4116) Business Intelligence and Data Mining ISOM 3360: Spring 203 Instructor Contact Office Hours Course Schedule and Classroom Course Webpage Jia Jia, ISOM Email: justinjia@ust.hk Office: Rm 336 (Lift 3-) Begin

More information

Introduction to Machine Learning Lecture 1. Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.edu

Introduction to Machine Learning Lecture 1. Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.edu Introduction to Machine Learning Lecture 1 Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.edu Introduction Logistics Prerequisites: basics concepts needed in probability and statistics

More information

Bayesian Spam Filtering

Bayesian Spam Filtering Bayesian Spam Filtering Ahmed Obied Department of Computer Science University of Calgary amaobied@ucalgary.ca http://www.cpsc.ucalgary.ca/~amaobied Abstract. With the enormous amount of spam messages propagating

More information

Basics of Statistical Machine Learning

Basics of Statistical Machine Learning CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu Modern machine learning is rooted in statistics. You will find many familiar

More information

MACHINE LEARNING & INTRUSION DETECTION: HYPE OR REALITY?

MACHINE LEARNING & INTRUSION DETECTION: HYPE OR REALITY? MACHINE LEARNING & INTRUSION DETECTION: 1 SUMMARY The potential use of machine learning techniques for intrusion detection is widely discussed amongst security experts. At Kudelski Security, we looked

More information

controlled assessment Biology Chemistry Physics Science: Double Award Science: Single Award

controlled assessment Biology Chemistry Physics Science: Double Award Science: Single Award (REVISED SPECIFICATION) GCSE SCIENCES Controlled Assessment: Guidance Biology Chemistry Physics Science: Double Award Science: Single Award Issued September 2011 controlled assessment SECTION A This guidance

More information

Knowledge Discovery and Data Mining

Knowledge Discovery and Data Mining Knowledge Discovery and Data Mining Unit # 10 Sajjad Haider Fall 2012 1 Supervised Learning Process Data Collection/Preparation Data Cleaning Discretization Supervised/Unuspervised Identification of right

More information

How To Understand The Relation Between Simplicity And Probability In Computer Science

How To Understand The Relation Between Simplicity And Probability In Computer Science Chapter 6 Computation 6.1 Introduction In the last two chapters we saw that both the logical and the cognitive models of scientific discovery include a condition to prefer simple or minimal explanations.

More information

Definition and Calculus of Probability

Definition and Calculus of Probability In experiments with multivariate outcome variable, knowledge of the value of one variable may help predict another. For now, the word prediction will mean update the probabilities of events regarding the

More information

If n is odd, then 3n + 7 is even.

If n is odd, then 3n + 7 is even. Proof: Proof: We suppose... that 3n + 7 is even. that 3n + 7 is even. Since n is odd, there exists an integer k so that n = 2k + 1. that 3n + 7 is even. Since n is odd, there exists an integer k so that

More information

Bayes Theorem. Bayes Theorem- Example. Evaluation of Medical Screening Procedure. Evaluation of Medical Screening Procedure

Bayes Theorem. Bayes Theorem- Example. Evaluation of Medical Screening Procedure. Evaluation of Medical Screening Procedure Bayes Theorem P(C A) P(A) P(A C) = P(C A) P(A) + P(C B) P(B) P(E B) P(B) P(B E) = P(E B) P(B) + P(E A) P(A) P(D A) P(A) P(A D) = P(D A) P(A) + P(D B) P(B) Cost of procedure is $1,000,000 Data regarding

More information

FACTORING AFTER DEDEKIND

FACTORING AFTER DEDEKIND FACTORING AFTER DEDEKIND KEITH CONRAD Let K be a number field and p be a prime number. When we factor (p) = po K into prime ideals, say (p) = p e 1 1 peg g, we refer to the data of the e i s, the exponents

More information

Knowledge Discovery and Data Mining

Knowledge Discovery and Data Mining Knowledge Discovery and Data Mining Unit # 11 Sajjad Haider Fall 2013 1 Supervised Learning Process Data Collection/Preparation Data Cleaning Discretization Supervised/Unuspervised Identification of right

More information

Page 331, 38.4 Suppose a is a positive integer and p is a prime. Prove that p a if and only if the prime factorization of a contains p.

Page 331, 38.4 Suppose a is a positive integer and p is a prime. Prove that p a if and only if the prime factorization of a contains p. Page 331, 38.2 Assignment #11 Solutions Factor the following positive integers into primes. a. 25 = 5 2. b. 4200 = 2 3 3 5 2 7. c. 10 10 = 2 10 5 10. d. 19 = 19. e. 1 = 1. Page 331, 38.4 Suppose a is a

More information

Machine Learning using MapReduce

Machine Learning using MapReduce Machine Learning using MapReduce What is Machine Learning Machine learning is a subfield of artificial intelligence concerned with techniques that allow computers to improve their outputs based on previous

More information

Linear regression methods for large n and streaming data

Linear regression methods for large n and streaming data Linear regression methods for large n and streaming data Large n and small or moderate p is a fairly simple problem. The sufficient statistic for β in OLS (and ridge) is: The concept of sufficiency is

More information

Knowledge-Based Probabilistic Reasoning from Expert Systems to Graphical Models

Knowledge-Based Probabilistic Reasoning from Expert Systems to Graphical Models Knowledge-Based Probabilistic Reasoning from Expert Systems to Graphical Models By George F. Luger & Chayan Chakrabarti {luger cc} @cs.unm.edu Department of Computer Science University of New Mexico Albuquerque

More information

The Calculus of Probability

The Calculus of Probability The Calculus of Probability Let A and B be events in a sample space S. Partition rule: P(A) = P(A B) + P(A B ) Example: Roll a pair of fair dice P(Total of 10) = P(Total of 10 and double) + P(Total of

More information

Insurance Analytics - analýza dat a prediktivní modelování v pojišťovnictví. Pavel Kříž. Seminář z aktuárských věd MFF 4.

Insurance Analytics - analýza dat a prediktivní modelování v pojišťovnictví. Pavel Kříž. Seminář z aktuárských věd MFF 4. Insurance Analytics - analýza dat a prediktivní modelování v pojišťovnictví Pavel Kříž Seminář z aktuárských věd MFF 4. dubna 2014 Summary 1. Application areas of Insurance Analytics 2. Insurance Analytics

More information

Decision Support System For A Customer Relationship Management Case Study

Decision Support System For A Customer Relationship Management Case Study 61 Decision Support System For A Customer Relationship Management Case Study Ozge Kart 1, Alp Kut 1, and Vladimir Radevski 2 1 Dokuz Eylul University, Izmir, Turkey {ozge, alp}@cs.deu.edu.tr 2 SEE University,

More information

Address for Correspondence

Address for Correspondence Research Article IMPROVED BAYESIAN ANTI-SPAM FILTER IMPLEMENTATION AND ANALYSIS OF INDEPENDENT SPAM CORPUSES 1 P. U. Anitha, 2 Dr. C. V. Guru Rao Address for Correspondence 1 Assistant Professor, Path

More information

Anti-Spam Filter Based on Naïve Bayes, SVM, and KNN model

Anti-Spam Filter Based on Naïve Bayes, SVM, and KNN model AI TERM PROJECT GROUP 14 1 Anti-Spam Filter Based on,, and model Yun-Nung Chen, Che-An Lu, Chao-Yu Huang Abstract spam email filters are a well-known and powerful type of filters. We construct different

More information