Simple Language Models for Spam Detection
|
|
|
- Willa Hunt
- 10 years ago
- Views:
Transcription
1 Simple Language Models for Spam Detection Egidio Terra Faculty of Informatics PUC/RS - Brazil Abstract For this year s Spam track we used classifiers based on language models. These models are used to compute the log-likelihood for each individual message and then classify them as either ham or spam. Different data sets were used to train these language models. Our approach is simple, we initially create simple unigram language models and smooth the probabilities of unseen tokens by means of the expected likelihood estimator with a small discount probability tuned in a training corpus. 1 Introduction The statistical approaches for spam filtering are often Bayesian and assume a multivariate Bernoulli distribution for tokens based on the number of messages they appear[9, 6, 1, 4, 8]. Each word in each corpus, ham and spam, is assigned a conditional probability for each given class, i.e. P (w C = ham) and P (w C = spam). For a new message M, the goal is to compute P (C = ham M) and P (C = spam M), which will then be used to decide to what class the message belongs. Robinson [8] proposes the use of the χ 2 distribution confidence intervals to decide if the message is spam or ham. In the most common approach the naive bayes every word is considered independent from each other, i.e. for a given class C, P (C M) = w M P (C w). To convert P (w C) into the desired P (C w), the bayes formula is used: P (C w) = [P (w C)P (C)]/P (w). Language Models on the other hand normally take a multinomial approach within a single distribution for all tokens, regardless of how many messages they occur. One benefit of the multinomial approach is the number of available discounting and smoothing methods to handle unseen tokens, whereas some Bayesian approaches often handle this situation heuristically (e.g. unseen tokens are assigned a constant probability regardless of other tokens probabilities). Our submissions to this year s TREC are based on language models. While spam classification can be seen as a text categorization with only two classes, it has some distinct characteristics from normal text, among others: a) massages have some structure; b) spam is written to look like a legit message; c) the arrival order of messages maybe important; d) spam messages mutate to avoid filters. It is our intention to evaluate the language models in this context since it has yielded good results in other text classification tasks, outperforming state-ofart classifiers such as SVMs in many of them [7]. Section 2 gives a brief introduction to the language models used in our submissions. The following section, 3, describes the training data we used for our models and section 4 presents our 1
2 results and discussion. 2 Simple Language Models Language models are generative with some order approximation to the language [10]. We start described the general form of the model: P (M) = P (w 1 ) P (w i w 1..w i 1) (1) i=2..n where w 1..w i 1 is called the history. The effect of history is more important in nearby neighbors, i.e. the outcome of the current word is more influenced by recently occurring ones. This is explored in the high-order approximations and it is equivalent to the Markov assumptions, where older history is not taken into account P (M) = P (w i w i k+1..w i 1 ) (2) i=1..n for a k-order approximation. A first-order word approximation (unigram model) is simply P (M) = P (w i ) (3) i=1..n where M = n and P (w i ) are observed from actual text. Equation 3 is also used to described the zero-order approximation. Shannon makes the distinction between zero and first based only on the estimates of P (w i ). In the zero-order, the estimates are sampled from an uniform distribution. The model is called generative since we can, by sampling with replacement, generate sequences that will have similar distribution to the model. When used in text classification, a model can be built for each class and the most likely class is chosen [7]. In the case of a dichotomous test, such as the one found in spam filtering, there are other alternatives. Dunning proposes the use of log-likelihood test to validate the hypothesis that the classes are distinguishable [3]. For that he explored the fact that a log-likelihood is asymptotically χ 2 distributed and as such we can use confidence intervals from the distribution to decide if a message is spam or not. Another important aspect in language models is the sparseness problem. Specially in higherorder approximations, the number of potential combinations is high. For a vocabulary V, containing v = V distinct words, a k-order approximation will have v k combinations. Even for very large corpora the number of possible combinations is greater than then number of seen ones, and even for modest values of k because the vocabulary tend to be large. For this problem several approaches exists [2, 5], the so-called smoothing and discounting techniques. The basic idea is to reserve some probability mass for the unseen events in the training data. For simplicity, we have used the expected likelihood estimate: P ele (w i ) = f(w i) + λ (4) N + Bλ where N is the size of the corpus (tokens) and B is the number of unique words (types). The frequency of a word is given by f(w i ). The value of λ can be trained and if its value is set to 1 then it is equivalent to the Laplace s law [5]. 3 Training For each class, ham and spam, we created distinct unigram language models. For every new message the likelihood is computed in both ham 2
3 Corpus % Ham % Spam 1-ROCA% Full mrx sb tm Table 1: Results for run pucrs0 and spam language models and the log-likelihood odds is used to classify the message. In two submissions, pucrs0 and pucrs1, we use the result of the equation 5 to classify the message. If the outcome is higher than a threshold the message is considered spam, otherwise it is ham. log LL = log P s(m) P h (M) = w M log P s (w) log P h (w) (5) where P s (w) is the probability of generating the word w according to the spam language model and P h (w) the analogous for the ham language model. We use a train-on-everything approach, for all incoming messages we update the corresponding language model. In the first run, labeled pucrs0, the token frequencies from the incoming messages are accumulated to those originated from the training data set, in this case the spamassassin corpus. The same approach was used in the second run, labeled pucrs1, but, instead of the spamassassin corpus, we used messages from a spam archive and for ham messages we used documents extracted from the AQUAINT corpus. In the third run, pucrs2, the token frequencies from the incoming messages are kept separate from those originated from the training Corpus % Ham % Spam 1-ROCA% Full mrx sb tm Table 2: Results for run pucrs1 Corpus % Ham % Spam 1-ROCA% Full mrx sb tm Table 3: Results for run pucrs2 data, once again the spamassassin corpus. result, four distinct models are kept: As incoming ham (P h ), incoming spam (P s ), background ham (P hb ) and background spam (P sb ). The two pairs of language models are then linearly interpolated, as depicted in equation 6. The mixture parameters are not fixed and, after a certain number of messages are collected, the weights assigned to background probabilities are reduced linearly. The initial weights λ s for spam language models and λ h for the ham language models are trained on spamassassin corpus. These weights are reduced as the number of incoming messages increases. log LL = log λ sp s (M) + (1 λ s )P sb (M) λ h P h (M) + (1 λ h )P hb (M) = w M log[λ s P s (M) + (1 λ s )P sb (M)] log[λ h P h (M) + (1 λ h )P hb (M)] (6) 3
4 Figure 1: ROC for the Full corpus Figure 3: ROC for the sb corpus Figure 2: ROC for the mrx corpus To tune the decision process we use a threshold for the likelihood odds and actively change its value based on the misclassifications performed so far. Since ham misclassifications is considered to be worse than spam misclassification, the threshold tuning is more aggressive when the former occur. In our case, this tuning is particularly necessary when the classifiers have not being exposed to enough data to create the model. In spam detection this can be the case when adapting the filter to a new user that does not have enough training data. This tuning process has an effect on the number of messages correctly or incorrectly classified but does not make a difference on the area under the ROC curve. Figure 4: ROC for the tm corpus 4 Results and Discussion For our run pucrs0, the results are shown in the table 1 for the several corpora used. Table 2 presents the results for pucrs1 and table 3 contain the results for pucrs2. The overall misclassification rates are high and on two corpora, sb and tm, the spam misclassification is very high. These two corpora have number of spam messages small compared to the total number of messages: only around 11% of messages on sb and tm corpora is spam. In these two corpora the classification thresholds are biased to ham scores, i.e., the threshold to classify a message as spam is high. The two other corpora are more balanced, only 18% of messages in mrx corpus are 4
5 ham and 42% of messages in the full corpus is ham. This poor performance indicates that the current tweaking of the threshold is not robust for corpus with unbalanced number of messages on the two classes. The area under the ROC is a more robust measure to determine the usefulness of the scores assigned by the classifiers. The learning rates are also influenced by the active threshold tuning. Unfortunately, setting the threshold is a quick change but running the filters on the private data is not yet possible if ever. The results for pucrs2 are worse than the other two runs, as shown in the ROC graphs in Figures 1,2,3, and 4. With the exception of the full corpus, pucrs2 performs statistically worse the pucrs0 and pucrs1 in terms of area under the ROC. Only in the full corpus the performance is similar. This suggests that either the mixture with a background model is ineffective or that the initial weight and decaying rate need to be trained for each corpus. We have tried Robinson s chi-square Bayesian, adopted in many popular filters such as Bogofilter and Spamassassing, in our framework and the results are not better than the language models. This suggests that the high misclassification rates may be also due to the preprocessing of the messages. Our tokenization process is too simple, we use standard stopword list and all headers are ignored with the exception of the to:, subject: and from: entries. No special handling is made for attachments and multipart messages, all data is considered to be text. All words with less than three characters is discarded. Some regular patterns are used to split URL s and addresses. Acknowledgments We thank the University of Waterloo for letting us use their computer resources. References [1] I. Androutsopoulos, J. Koutsias, K.V. Chandrinos, G. Paliouras, and C.D. Spyropoulus. An evaluation of naive bayesian anti-spam filtering. In Proceedings of the Workshop on Machine Learning in the New Information Age, 11th European Conference on Machine Learning (ECML 2000), [2] S. Chen and J. Goodman. An empirical study of smoothing techniques for language modeling. Technical report, Harvard University, [3] Ted Dunning. Accurate methods for the statistics of surprise and coincidence. Computational Linguistics, 19(1), [4] Paul Graham. A plan for spam. [5] Christopher D. Manning and Hinrich Schutze. Foundations of Statistical Natural Language Processing. MIT Press, [6] Patrick Pantel and Dekang Lin. Spamcop a spam classification and organization program. In Proceedings of AAAI-98 Workshop on Learning for Text Categorization, [7] Fuchun Peng, Dale Schuurmans, and Shaojun Wang. Language and task independent text categorization with simple language models. In Proceedings of the 2003 Human Language Technology Conference of 5
6 the North American Chapter of the Association for Computational Linguistics, [8] Gary Robinson. A statistical approach to the spam problem. Linux Journal, (107), [9] Mehran Sahami, Susan Dumais, David Heckerman, and Eric Horvitz. A bayesian approach to filtering junk . In Proceedings of AAAI-98 Workshop on Learning for Text Categorization, [10] Claude E. Shannon. A mathematical theory of communication. Bell System Technical Journal, 27: , July and October
CAS-ICT at TREC 2005 SPAM Track: Using Non-Textual Information to Improve Spam Filtering Performance
CAS-ICT at TREC 2005 SPAM Track: Using Non-Textual Information to Improve Spam Filtering Performance Shen Wang, Bin Wang and Hao Lang, Xueqi Cheng Institute of Computing Technology, Chinese Academy of
A MACHINE LEARNING APPROACH TO SERVER-SIDE ANTI-SPAM E-MAIL FILTERING 1 2
UDC 004.75 A MACHINE LEARNING APPROACH TO SERVER-SIDE ANTI-SPAM E-MAIL FILTERING 1 2 I. Mashechkin, M. Petrovskiy, A. Rozinkin, S. Gerasimov Computer Science Department, Lomonosov Moscow State University,
Combining Global and Personal Anti-Spam Filtering
Combining Global and Personal Anti-Spam Filtering Richard Segal IBM Research Hawthorne, NY 10532 Abstract Many of the first successful applications of statistical learning to anti-spam filtering were personalized
On Attacking Statistical Spam Filters
On Attacking Statistical Spam Filters Gregory L. Wittel and S. Felix Wu Department of Computer Science University of California, Davis One Shields Avenue, Davis, CA 95616 USA Abstract. The efforts of anti-spammers
Naive Bayes Spam Filtering Using Word-Position-Based Attributes
Naive Bayes Spam Filtering Using Word-Position-Based Attributes Johan Hovold Department of Computer Science Lund University Box 118, 221 00 Lund, Sweden [email protected] Abstract This paper
Bayesian Spam Filtering
Bayesian Spam Filtering Ahmed Obied Department of Computer Science University of Calgary [email protected] http://www.cpsc.ucalgary.ca/~amaobied Abstract. With the enormous amount of spam messages propagating
Spam Filtering with Naive Bayesian Classification
Spam Filtering with Naive Bayesian Classification Khuong An Nguyen Queens College University of Cambridge L101: Machine Learning for Language Processing MPhil in Advanced Computer Science 09-April-2011
A Two-Pass Statistical Approach for Automatic Personalized Spam Filtering
A Two-Pass Statistical Approach for Automatic Personalized Spam Filtering Khurum Nazir Junejo, Mirza Muhammad Yousaf, and Asim Karim Dept. of Computer Science, Lahore University of Management Sciences
Not So Naïve Online Bayesian Spam Filter
Not So Naïve Online Bayesian Spam Filter Baojun Su Institute of Artificial Intelligence College of Computer Science Zhejiang University Hangzhou 310027, China [email protected] Congfu Xu Institute of Artificial
Abstract. Find out if your mortgage rate is too high, NOW. Free Search
Statistics and The War on Spam David Madigan Rutgers University Abstract Text categorization algorithms assign texts to predefined categories. The study of such algorithms has a rich history dating back
Immunity from spam: an analysis of an artificial immune system for junk email detection
Immunity from spam: an analysis of an artificial immune system for junk email detection Terri Oda and Tony White Carleton University, Ottawa ON, Canada [email protected], [email protected] Abstract.
PSSF: A Novel Statistical Approach for Personalized Service-side Spam Filtering
2007 IEEE/WIC/ACM International Conference on Web Intelligence PSSF: A Novel Statistical Approach for Personalized Service-side Spam Filtering Khurum Nazir Juneo Dept. of Computer Science Lahore University
1 Maximum likelihood estimation
COS 424: Interacting with Data Lecturer: David Blei Lecture #4 Scribes: Wei Ho, Michael Ye February 14, 2008 1 Maximum likelihood estimation 1.1 MLE of a Bernoulli random variable (coin flips) Given N
Adaptive Filtering of SPAM
Adaptive Filtering of SPAM L. Pelletier, J. Almhana, V. Choulakian GRETI, University of Moncton Moncton, N.B.,Canada E1A 3E9 {elp6880, almhanaj, choulav}@umoncton.ca Abstract In this paper, we present
Sentiment analysis using emoticons
Sentiment analysis using emoticons Royden Kayhan Lewis Moharreri Steven Royden Ware Lewis Kayhan Steven Moharreri Ware Department of Computer Science, Ohio State University Problem definition Our aim was
Detecting E-mail Spam Using Spam Word Associations
Detecting E-mail Spam Using Spam Word Associations N.S. Kumar 1, D.P. Rana 2, R.G.Mehta 3 Sardar Vallabhbhai National Institute of Technology, Surat, India 1 [email protected] 2 [email protected]
Spam Filters: Bayes vs. Chi-squared; Letters vs. Words
Spam Filters: Bayes vs. Chi-squared; Letters vs. Words Cormac O Brien & Carl Vogel Abstract We compare two statistical methods for identifying spam or junk electronic mail. The proliferation of spam email
Tweaking Naïve Bayes classifier for intelligent spam detection
682 Tweaking Naïve Bayes classifier for intelligent spam detection Ankita Raturi 1 and Sunil Pranit Lal 2 1 University of California, Irvine, CA 92697, USA. [email protected] 2 School of Computing, Information
Adaption of Statistical Email Filtering Techniques
Adaption of Statistical Email Filtering Techniques David Kohlbrenner IT.com Thomas Jefferson High School for Science and Technology January 25, 2007 Abstract With the rise of the levels of spam, new techniques
Increasing the Accuracy of a Spam-Detecting Artificial Immune System
Increasing the Accuracy of a Spam-Detecting Artificial Immune System Terri Oda Carleton University [email protected] Tony White Carleton University [email protected] Abstract- Spam, the electronic
Introduction to Bayesian Classification (A Practical Discussion) Todd Holloway Lecture for B551 Nov. 27, 2007
Introduction to Bayesian Classification (A Practical Discussion) Todd Holloway Lecture for B551 Nov. 27, 2007 Naïve Bayes Components ML vs. MAP Benefits Feature Preparation Filtering Decay Extended Examples
2. NAIVE BAYES CLASSIFIERS
Spam Filtering with Naive Bayes Which Naive Bayes? Vangelis Metsis Institute of Informatics and Telecommunications, N.C.S.R. Demokritos, Athens, Greece Ion Androutsopoulos Department of Informatics, Athens
A Three-Way Decision Approach to Email Spam Filtering
A Three-Way Decision Approach to Email Spam Filtering Bing Zhou, Yiyu Yao, and Jigang Luo Department of Computer Science, University of Regina Regina, Saskatchewan, Canada S4S 0A2 {zhou200b,yyao,luo226}@cs.uregina.ca
Three-Way Decisions Solution to Filter Spam Email: An Empirical Study
Three-Way Decisions Solution to Filter Spam Email: An Empirical Study Xiuyi Jia 1,4, Kan Zheng 2,WeiweiLi 3, Tingting Liu 2, and Lin Shang 4 1 School of Computer Science and Technology, Nanjing University
A Case-Based Approach to Spam Filtering that Can Track Concept Drift
A Case-Based Approach to Spam Filtering that Can Track Concept Drift Pádraig Cunningham 1, Niamh Nowlan 1, Sarah Jane Delany 2, Mads Haahr 1 1 Department of Computer Science, Trinity College Dublin 2 School
Filtering Junk Mail with A Maximum Entropy Model
Filtering Junk Mail with A Maximum Entropy Model ZHANG Le and YAO Tian-shun Institute of Computer Software & Theory. School of Information Science & Engineering, Northeastern University Shenyang, 110004
Spam Filtering based on Naive Bayes Classification. Tianhao Sun
Spam Filtering based on Naive Bayes Classification Tianhao Sun May 1, 2009 Abstract This project discusses about the popular statistical spam filtering process: naive Bayes classification. A fairly famous
Search and Data Mining: Techniques. Text Mining Anya Yarygina Boris Novikov
Search and Data Mining: Techniques Text Mining Anya Yarygina Boris Novikov Introduction Generally used to denote any system that analyzes large quantities of natural language text and detects lexical or
A Comparison of Event Models for Naive Bayes Text Classification
A Comparison of Event Models for Naive Bayes Text Classification Andrew McCallum [email protected] Just Research 4616 Henry Street Pittsburgh, PA 15213 Kamal Nigam [email protected] School of Computer
Bayesian Spam Detection
Scholarly Horizons: University of Minnesota, Morris Undergraduate Journal Volume 2 Issue 1 Article 2 2015 Bayesian Spam Detection Jeremy J. Eberhardt University or Minnesota, Morris Follow this and additional
Email Filters that use Spammy Words Only
Email Filters that use Spammy Words Only Vasanth Elavarasan Department of Computer Science University of Texas at Austin Advisors: Mohamed Gouda Department of Computer Science University of Texas at Austin
SpamNet Spam Detection Using PCA and Neural Networks
SpamNet Spam Detection Using PCA and Neural Networks Abhimanyu Lad B.Tech. (I.T.) 4 th year student Indian Institute of Information Technology, Allahabad Deoghat, Jhalwa, Allahabad, India [email protected]
[email protected] ** The High Institute of Zahra for Comperhensive Professions, Zahra-Libya
AN ANTI-SPAM SYSTEM USING ARTIFICIAL NEURAL NETWORKS AND GENETIC ALGORITHMS ABDUELBASET M. GOWEDER *, TARIK RASHED **, ALI S. ELBEKAIE ***, and HUSIEN A. ALHAMMI **** * The High Institute of Surman for
Spam Filters: Bayes vs. Chi-squared; Letters vs. Words
Spam Filters: Bayes vs. Chi-squared; Letters vs. Words Cormac O Brien & Carl Vogel Computational Linguistics Group & Centre for Computing and Language Studies Trinity College University of Dublin obriencevogel
1 Introductory Comments. 2 Bayesian Probability
Introductory Comments First, I would like to point out that I got this material from two sources: The first was a page from Paul Graham s website at www.paulgraham.com/ffb.html, and the second was a paper
An Efficient Spam Filtering Techniques for Email Account
American Journal of Engineering Research (AJER) e-issn : 2320-0847 p-issn : 2320-0936 Volume-02, Issue-10, pp-63-73 www.ajer.org Research Paper Open Access An Efficient Spam Filtering Techniques for Email
Email Spam Detection A Machine Learning Approach
Email Spam Detection A Machine Learning Approach Ge Song, Lauren Steimle ABSTRACT Machine learning is a branch of artificial intelligence concerned with the creation and study of systems that can learn
Part III: Machine Learning. CS 188: Artificial Intelligence. Machine Learning This Set of Slides. Parameter Estimation. Estimation: Smoothing
CS 188: Artificial Intelligence Lecture 20: Dynamic Bayes Nets, Naïve Bayes Pieter Abbeel UC Berkeley Slides adapted from Dan Klein. Part III: Machine Learning Up until now: how to reason in a model and
A Personalized Spam Filtering Approach Utilizing Two Separately Trained Filters
2008 IEEE/WIC/ACM International Conference on Web Intelligence and Intelligent Agent Technology A Personalized Spam Filtering Approach Utilizing Two Separately Trained Filters Wei-Lun Teng, Wei-Chung Teng
Bayes and Naïve Bayes. cs534-machine Learning
Bayes and aïve Bayes cs534-machine Learning Bayes Classifier Generative model learns Prediction is made by and where This is often referred to as the Bayes Classifier, because of the use of the Bayes rule
Using Biased Discriminant Analysis for Email Filtering
Using Biased Discriminant Analysis for Email Filtering Juan Carlos Gomez 1 and Marie-Francine Moens 2 1 ITESM, Eugenio Garza Sada 2501, Monterrey NL 64849, Mexico [email protected] 2
Machine Learning Final Project Spam Email Filtering
Machine Learning Final Project Spam Email Filtering March 2013 Shahar Yifrah Guy Lev Table of Content 1. OVERVIEW... 3 2. DATASET... 3 2.1 SOURCE... 3 2.2 CREATION OF TRAINING AND TEST SETS... 4 2.3 FEATURE
Spam or Not Spam That is the question
Spam or Not Spam That is the question Ravi Kiran S S and Indriyati Atmosukarto {kiran,indria}@cs.washington.edu Abstract Unsolicited commercial email, commonly known as spam has been known to pollute the
Partitioned Logistic Regression for Spam Filtering
Partitioned Logistic Regression for Spam Filtering Ming-wei Chang University of Illinois 201 N Goodwin Ave Urbana, IL, USA [email protected] Wen-tau Yih Microsoft Research One Microsoft Way Redmond, WA,
Content-Based Spam Filtering and Detection Algorithms- An Efficient Analysis & Comparison
Content-Based Spam Filtering and Detection Algorithms- An Efficient Analysis & Comparison 1 R.Malarvizhi, 2 K.Saraswathi 1 Research scholar, PG & Research Department of Computer Science, Government Arts
A Proposed Algorithm for Spam Filtering Emails by Hash Table Approach
International Research Journal of Applied and Basic Sciences 2013 Available online at www.irjabs.com ISSN 2251-838X / Vol, 4 (9): 2436-2441 Science Explorer Publications A Proposed Algorithm for Spam Filtering
Question 2 Naïve Bayes (16 points)
Question 2 Naïve Bayes (16 points) About 2/3 of your email is spam so you downloaded an open source spam filter based on word occurrences that uses the Naive Bayes classifier. Assume you collected the
Anti Spamming Techniques
Anti Spamming Techniques Written by Sumit Siddharth In this article will we first look at some of the existing methods to identify an email as a spam? We look at the pros and cons of the existing methods
An Efficient Three-phase Email Spam Filtering Technique
An Efficient Three-phase Email Filtering Technique Tarek M. Mahmoud 1 *, Alaa Ismail El-Nashar 2 *, Tarek Abd-El-Hafeez 3 *, Marwa Khairy 4 * 1, 2, 3 Faculty of science, Computer Sci. Dept., Minia University,
IMPROVING THE PERFORMANCE OF ANTI-SPAM FILTERS USING OUT-OF-VOCABULARY STATISTICS
Ingeniare. Revista chilena de ingeniería, vol. 17 Nº 3, 2009, pp. 386-392 IMPROVING THE PERFORMANCE OF ANTI-SPAM FILTERS USING OUT-OF-VOCABULARY STATISTICS MEJORA DEL DESEMPEÑO DE FILTROS ANTI-SPAM USANDO
Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
IMPROVING SPAM EMAIL FILTERING EFFICIENCY USING BAYESIAN BACKWARD APPROACH PROJECT
IMPROVING SPAM EMAIL FILTERING EFFICIENCY USING BAYESIAN BACKWARD APPROACH PROJECT M.SHESHIKALA Assistant Professor, SREC Engineering College,Warangal Email: [email protected], Abstract- Unethical
Non-Parametric Spam Filtering based on knn and LSA
Non-Parametric Spam Filtering based on knn and LSA Preslav Ivanov Nakov Panayot Markov Dobrikov Abstract. The paper proposes a non-parametric approach to filtering of unsolicited commercial e-mail messages,
Detecting Spam Bots in Online Social Networking Sites: A Machine Learning Approach
Detecting Spam Bots in Online Social Networking Sites: A Machine Learning Approach Alex Hai Wang College of Information Sciences and Technology, The Pennsylvania State University, Dunmore, PA 18512, USA
A Bayesian Topic Model for Spam Filtering
Journal of Information & Computational Science 1:12 (213) 3719 3727 August 1, 213 Available at http://www.joics.com A Bayesian Topic Model for Spam Filtering Zhiying Zhang, Xu Yu, Lixiang Shi, Li Peng,
Spam Filtering Based On The Analysis Of Text Information Embedded Into Images
Journal of Machine Learning Research 7 (2006) 2699-2720 Submitted 3/06; Revised 9/06; Published 12/06 Spam Filtering Based On The Analysis Of Text Information Embedded Into Images Giorgio Fumera Ignazio
Accelerating Techniques for Rapid Mitigation of Phishing and Spam Emails
Accelerating Techniques for Rapid Mitigation of Phishing and Spam Emails Pranil Gupta, Ajay Nagrale and Shambhu Upadhyaya Computer Science and Engineering University at Buffalo Buffalo, NY 14260 {pagupta,
CS 348: Introduction to Artificial Intelligence Lab 2: Spam Filtering
THE PROBLEM Spam is e-mail that is both unsolicited by the recipient and sent in substantively identical form to many recipients. In 2004, MSNBC reported that spam accounted for 66% of all electronic mail.
CS 533: Natural Language. Word Prediction
CS 533: Natural Language Processing Lecture 03 N-Gram Models and Algorithms CS 533: Natural Language Processing Lecture 01 1 Word Prediction Suppose you read the following sequence of words: Sue swallowed
Lan, Mingjun and Zhou, Wanlei 2005, Spam filtering based on preference ranking, in Fifth International Conference on Computer and Information
Lan, Mingjun and Zhou, Wanlei 2005, Spam filtering based on preference ranking, in Fifth International Conference on Computer and Information Technology : CIT 2005 : proceedings : 21-23 September, 2005,
Machine Learning. CS 188: Artificial Intelligence Naïve Bayes. Example: Digit Recognition. Other Classification Tasks
CS 188: Artificial Intelligence Naïve Bayes Machine Learning Up until now: how use a model to make optimal decisions Machine learning: how to acquire a model from data / experience Learning parameters
Filtering Spam Using Search Engines
Filtering Spam Using Search Engines Oleg Kolesnikov, Wenke Lee, and Richard Lipton ok,wenke,rjl @cc.gatech.edu College of Computing Georgia Institute of Technology Atlanta, GA 30332 Abstract Spam filtering
Anti-Spam Filter Based on Naïve Bayes, SVM, and KNN model
AI TERM PROJECT GROUP 14 1 Anti-Spam Filter Based on,, and model Yun-Nung Chen, Che-An Lu, Chao-Yu Huang Abstract spam email filters are a well-known and powerful type of filters. We construct different
Three types of messages: A, B, C. Assume A is the oldest type, and C is the most recent type.
Chronological Sampling for Email Filtering Ching-Lung Fu 2, Daniel Silver 1, and James Blustein 2 1 Acadia University, Wolfville, Nova Scotia, Canada 2 Dalhousie University, Halifax, Nova Scotia, Canada
Author Gender Identification of English Novels
Author Gender Identification of English Novels Joseph Baena and Catherine Chen December 13, 2013 1 Introduction Machine learning algorithms have long been used in studies of authorship, particularly in
Enrique Puertas Sánz Francisco Carrero García Universidad Europea de Madrid Villaviciosa de Odón 28670 Madrid, SPAIN 34 91 211 5611
José María Gómez Hidalgo Universidad Europea de Madrid Villaviciosa de Odón 28670 Madrid, SPAIN 34 91 211 5670 [email protected] Content Based SMS Spam Filtering Guillermo Cajigas Bringas Group R&D - Vodafone
Email Classification Using Data Reduction Method
Email Classification Using Data Reduction Method Rafiqul Islam and Yang Xiang, member IEEE School of Information Technology Deakin University, Burwood 3125, Victoria, Australia Abstract Classifying user
Chapter 7. Language models. Statistical Machine Translation
Chapter 7 Language models Statistical Machine Translation Language models Language models answer the question: How likely is a string of English words good English? Help with reordering p lm (the house
Homework 4 Statistics W4240: Data Mining Columbia University Due Tuesday, October 29 in Class
Problem 1. (10 Points) James 6.1 Problem 2. (10 Points) James 6.3 Problem 3. (10 Points) James 6.5 Problem 4. (15 Points) James 6.7 Problem 5. (15 Points) James 6.10 Homework 4 Statistics W4240: Data Mining
Language Modeling. Chapter 1. 1.1 Introduction
Chapter 1 Language Modeling (Course notes for NLP by Michael Collins, Columbia University) 1.1 Introduction In this chapter we will consider the the problem of constructing a language model from a set
Journal of Information Technology Impact
Journal of Information Technology Impact Vol. 8, No., pp. -0, 2008 Probability Modeling for Improving Spam Filtering Parameters S. C. Chiemeke University of Benin Nigeria O. B. Longe 2 University of Ibadan
Spam Filtering Based on Latent Semantic Indexing
Spam Filtering Based on Latent Semantic Indexing Wilfried N. Gansterer Andreas G. K. Janecek Robert Neumayer Abstract In this paper, a study on the classification performance of a vector space model (VSM)
Impact of Feature Selection Technique on Email Classification
Impact of Feature Selection Technique on Email Classification Aakanksha Sharaff, Naresh Kumar Nagwani, and Kunal Swami Abstract Being one of the most powerful and fastest way of communication, the popularity
T-61.3050 : Email Classification as Spam or Ham using Naive Bayes Classifier. Santosh Tirunagari : 245577
T-61.3050 : Email Classification as Spam or Ham using Naive Bayes Classifier Santosh Tirunagari : 245577 January 20, 2011 Abstract This term project gives a solution how to classify an email as spam or
An Approach to Detect Spam Emails by Using Majority Voting
An Approach to Detect Spam Emails by Using Majority Voting Roohi Hussain Department of Computer Engineering, National University of Science and Technology, H-12 Islamabad, Pakistan Usman Qamar Faculty,
Analysis of Spam Filter Methods on SMTP Servers Category: Trends in Anti-Spam Development
Analysis of Spam Filter Methods on SMTP Servers Category: Trends in Anti-Spam Development Author André Tschentscher Address Fachhochschule Erfurt - University of Applied Sciences Applied Computer Science
OPINION MINING IN PRODUCT REVIEW SYSTEM USING BIG DATA TECHNOLOGY HADOOP
OPINION MINING IN PRODUCT REVIEW SYSTEM USING BIG DATA TECHNOLOGY HADOOP 1 KALYANKUMAR B WADDAR, 2 K SRINIVASA 1 P G Student, S.I.T Tumkur, 2 Assistant Professor S.I.T Tumkur Abstract- Product Review System
Differential Voting in Case Based Spam Filtering
Differential Voting in Case Based Spam Filtering Deepak P, Delip Rao, Deepak Khemani Department of Computer Science and Engineering Indian Institute of Technology Madras, India [email protected],
Using News Articles to Predict Stock Price Movements
Using News Articles to Predict Stock Price Movements Győző Gidófalvi Department of Computer Science and Engineering University of California, San Diego La Jolla, CA 9237 [email protected] 21, June 15,
Predicting the Stock Market with News Articles
Predicting the Stock Market with News Articles Kari Lee and Ryan Timmons CS224N Final Project Introduction Stock market prediction is an area of extreme importance to an entire industry. Stock price is
Controlling Spam E-mail at the Routers
Controlling Spam E-mail at the Routers Banit Agrawal Nitin Kumar Mart Molle Department of Computer Science & Engineering University of California, Riverside, CA, 92521, USA email: bagrawal, nkumar, mart
