Probabilistic topic models for sentiment analysis on the Web



Similar documents
Data Mining Yelp Data - Predicting rating stars from review text

Topic models for Sentiment analysis: A Literature Survey

Latent Dirichlet Markov Allocation for Sentiment Analysis

Improving Twitter Sentiment Analysis with Topic-Based Mixture Modeling and Semi-Supervised Training

Joint Sentiment/Topic Model for Sentiment Analysis

Exploring English Lexicon Knowledge for Chinese Sentiment Analysis

II. RELATED WORK. Sentiment Mining

A Comparative Study on Sentiment Classification and Ranking on Product Reviews

Identifying Focus, Techniques and Domain of Scientific Papers

Final Project Report

A Survey on Product Aspect Ranking

SUIT: A Supervised User-Item Based Topic Model for Sentiment Analysis

VCU-TSA at Semeval-2016 Task 4: Sentiment Analysis in Twitter

Literature Survey on Data Mining and Statistical Report for Drugs Reviews

Table of Contents. Chapter No. 1 Introduction 1. iii. xiv. xviii. xix. Page No.

Exploiting Topic based Twitter Sentiment for Stock Prediction

Sentiment analysis on tweets in a financial domain

Semantic Sentiment Analysis of Twitter

Network Big Data: Facing and Tackling the Complexities Xiaolong Jin

SENTIMENT ANALYSIS: A STUDY ON PRODUCT FEATURES

An Introduction to Data Mining

IT services for analyses of various data samples

Subordinating to the Majority: Factoid Question Answering over CQA Sites

Microblog Sentiment Analysis with Emoticon Space Model

Part III: Machine Learning. CS 188: Artificial Intelligence. Machine Learning This Set of Slides. Parameter Estimation. Estimation: Smoothing

Machine Learning. Chapter 18, 21. Some material adopted from notes by Chuck Dyer

Online Courses Recommendation based on LDA

Learning is a very general term denoting the way in which agents:

A Survey on Product Aspect Ranking Techniques

International Journal of Computer Science Trends and Technology (IJCST) Volume 2 Issue 3, May-Jun 2014

FINDING EXPERT USERS IN COMMUNITY QUESTION ANSWERING SERVICES USING TOPIC MODELS

Bing Liu. Web Data Mining. Exploring Hyperlinks, Contents, and Usage Data. With 177 Figures. ~ Spring~r

Text Analytics. A business guide

Chapter ML:XI (continued)

Sentiment Analysis on Big Data

Classifying Manipulation Primitives from Visual Data

Micro blogs Oriented Word Segmentation System

Topical Authority Identification in Community Question Answering

An Automatic and Accurate Segmentation for High Resolution Satellite Image S.Saumya 1, D.V.Jiji Thanka Ligoshia 2

A Tri-Role Topic Model for Domain-Specific Question Answering

Semi-Supervised Learning for Blog Classification

Sentiment Analysis of Microblogs

Introduction to Data Mining

Towards SoMEST Combining Social Media Monitoring with Event Extraction and Timeline Analysis

FEATURE SELECTION AND CLASSIFICATION APPROACH FOR SENTIMENT ANALYSIS

Data Mining & Data Stream Mining Open Source Tools

On the Feasibility of Answer Suggestion for Advice-seeking Community Questions about Government Services

Bayesian Machine Learning (ML): Modeling And Inference in Big Data. Zhuhua Cai Google, Rice University

NAVIGATING SCIENTIFIC LITERATURE A HOLISTIC PERSPECTIVE. Venu Govindaraju

Data Mining on Social Networks. Dionysios Sotiropoulos Ph.D.

Jointly Modeling Aspects and Opinions with a MaxEnt-LDA Hybrid

Fraud Detection in Online Reviews using Machine Learning Techniques

A Systematic Cross-Comparison of Sequence Classifiers

A Statistical Text Mining Method for Patent Analysis

Sentiment analysis of Twitter microblogging posts. Jasmina Smailović Jožef Stefan Institute Department of Knowledge Technologies

Evidence-based Software Process Recovery

Using Artificial Intelligence to Manage Big Data for Litigation

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data

Sentiment Analysis Tool using Machine Learning Algorithms

Chapter 8. Final Results on Dutch Senseval-2 Test Data

ONLINE RESUME PARSING SYSTEM USING TEXT ANALYTICS

Predicting the Stock Market with News Articles

Practical Graph Mining with R. 5. Link Analysis

Document Classification with Latent Dirichlet Allocation

A Sentiment Analysis Model Integrating Multiple Algorithms and Diverse. Features. Thesis

Analyzing survey text: a brief overview

Mobile Apps Jovian Lin, Ph.D.

How To Classify Data Stream Mining

Sentiment Classification on Polarity Reviews: An Empirical Study Using Rating-based Features

MODELLING OF COORDINATING PRODUCTION AND INVENTORY CYCLES IN A MANUFACTURING SUPPLY CHAIN INVOLVING REVERSE LOGISTICS

A CRF-based approach to find stock price correlation with company-related Twitter sentiment

The Open University s repository of research publications and other research outputs

Globally Optimal Crowdsourcing Quality Management

Sentiment Analysis and Topic Classification: Case study over Spanish tweets

SENTIMENT ANALYSIS: TEXT PRE-PROCESSING, READER VIEWS AND CROSS DOMAINS EMMA HADDI BRUNEL UNIVERSITY LONDON

Robust Sentiment Detection on Twitter from Biased and Noisy Data

A Unified Probabilistic Model for Aspect-Level Sentiment Analysis

Review Topic Discovery with Phrases using the Pólya Urn Model

Automatic Identification of Arabic Language Varieties and Dialects in Social Media

How To Create A Text Classification System For Spam Filtering

Bagged Ensemble Classifiers for Sentiment Classification of Movie Reviews

Twitter Sentiment Analysis of Movie Reviews using Machine Learning Techniques.

Semi-Supervised and Unsupervised Machine Learning. Novel Strategies

Behavior Analysis in Crowded Environments. XiaogangWang Department of Electronic Engineering The Chinese University of Hong Kong June 25, 2011

THE digital age, also referred to as the information

Towards an Evaluation Framework for Topic Extraction Systems for Online Reputation Management

Making Sense of the Mayhem: Machine Learning and March Madness

Analysis Tools and Libraries for BigData

Summarizing Online Forum Discussions Can Dialog Acts of Individual Messages Help?

An Adaptive Approach to Named Entity Extraction for Meeting Applications

A Hierarchical Aspect-Sentiment Model for Online Reviews

Learning to Suggest Questions in Online Forums

A Survey on Intrusion Detection System with Data Mining Techniques

Maschinelles Lernen mit MATLAB

Natural Language to Relational Query by Using Parsing Compiler

Ordinance governing the Degrees of Doctor of Philosophy, Master of Philosophy and Professional Doctorates

International Journal of Research in Advent Technology Available Online at:

PREDICTIVE TECHNIQUES IN SOFTWARE ENGINEERING : APPLICATION IN SOFTWARE TESTING

Transcription:

University of Exeter Department of Computer Science Probabilistic topic models for sentiment analysis on the Web Chenghua Lin September 2011 Submitted by Chenghua Lin, to the the University of Exeter as a thesis for the degree of Doctor of Philosophy in Computer Science, September 2011. This thesis is available for Library use on the understanding that it is copyright material and that no quotation from the thesis may be published without proper acknowledgement. I certify that all material in this thesis which is not my own work has been identified and that no material has previously been submitted and approved for the award of a degree by this or any other University. (signature)...

Dedicated to my parents, Xingwen Lin and Yuanfang Chen.

Acknowledgements This thesis would not have been possible without the help and support from the following people. First and foremost, I would like to thank my supervisor Yulan He, who has been an exemplary teacher and mentor in the past three years of my PhD studies. I am extremely grateful to Yulan s encouragement, support and patience during the difficult time of my research. My gratitude also goes to my second supervisor Richard Everson, whose input and advice on my work were very useful. Richard is also very generous with his time for discussing research, for which I am grateful. I also thank Alex Gong, without whose kind help I would not have been able to pursue the research I really enjoy. I am fortunate to have a number of great friends who have supported me both research wise and non-research wise, particularly Naihui He, Junaid Khan, and Bing Qiao. Without them, my PhD would have been a less enjoyable one. Deepest thanks to my parents and family. Their lifetime love and care help me regain my confidence and make me once again believe in myself. Finally, I thank Yuan. Her kindness, patience and care sustain me. 2

Abstract Sentiment analysis aims to use automated tools to detect subjective information such as opinions, attitudes, and feelings expressed in text, and has received a rapid growth of interest in natural language processing in recent years. Probabilistic topic models, on the other hand, are capable of discovering hidden thematic structure in large archives of documents, and have been an active research area in the field of information retrieval. The work in this thesis focuses on developing topic models for automatic sentiment analysis of web data, by combining the ideas from both research domains. One noticeable issue of most previous work in sentiment analysis is that the trained classifier is domain dependent, and the labelled corpora required for training could be difficult to acquire in real world applications. Another issue is that the dependencies between sentiment/subjectivity and topics are not taken into consideration. The main contribution of this thesis is therefore the introduction of three probabilistic topic models, which address the above concerns by modelling sentiment/subjectivity and topic simultaneously. The first model is called the joint sentiment-topic (JST) model based on latent Dirichlet allocation (LDA), which detects sentiment and topic simultaneously from text. Unlike supervised approaches to sentiment classification which often fail to produce satisfactory performance when applied to new domains, the weakly-supervised nature of JST makes it highly portable to other domains, where the only supervision information required is a domain-independent sentiment lexicon. Apart from document-level sentiment classification results, JST can also extract sentiment-bearing topics automatically, which is a distinct feature compared to the existing sentiment analysis approaches. The second model is a dynamic version of JST called the dynamic joint sentimenttopic (djst) model. djst respects the ordering of documents, and allows the analysis of topic and sentiment evolution of document archives that are collected over a long time span. By accounting for the historical dependencies of documents from the past epochs in the generative process, djst gives a richer posterior topical structure than JST, and can better respond to the permutations of topic prominence. We also derive online inference procedures based on a stochastic EM algorithm for efficiently updating the model parameters. 3

The third model is called the subjectivity detection LDA (subjlda) model for sentencelevel subjectivity detection. Two sets of latent variables were introduced in subjlda. One is the subjectivity label for each sentence; another is the sentiment label for each word token. By viewing the subjectivity detection problem as weakly-supervised generative model learning, subjlda significantly outperforms the baseline and is comparable to the supervised approach which relies on much larger amounts of data for training. These models have been evaluated on real world datasets, demonstrating that joint sentiment topic modelling is indeed an important and useful research area with much to offer in the way of good results. 4

Publications Chapter 3 is based on the previously published work: Lin, C., He, Y., Everson, R. and Rüger, S. Weakly-supervised Joint Sentiment-Topic Detection from Text, IEEE Transactions on Knowledge and Data Engineering (TKDE), to appear. Lin, C., He, Y., and Everson, R. A Comparative Study of Bayesian Models for Unsupervised Sentiment Detection, In Proceedings of the 14th Conference on Computational Natural Language Learning (CoNLL), Uppsala, Sweden, 2010. Lin, C. and He, Y. Joint Sentiment/Topic Model for Sentiment Analysis, In Proceedings of the 18th ACM Conference on Information and Knowledge Management (CIKM), Hong Kong, China, 2009. Chapter 5 is based on the previously published work: Lin, C., He, Y. and Everson, R. Sentence Subjectivity Detection with Weakly-Supervised Learning, In Proceedings of the 5th International Joint Conference on Natural Language Processing (IJCNLP), Chiang Mai, Thailand, 2011. Other publications: He, Y., Lin, C. and Alani, H. Automatically Extracting Polarity-Bearing Topics for Cross-Domain Sentiment Classification, In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics (ACL), Portland, Oregon, 2011. He, Y. and Lin, C. Protein-Protein Interactions Classification from Text via Local Learning with Class Priors, In Proceedings of the 14th International Conference on Applications of Natural Language to Information Systems (NLDB), Saabrucken, Germany, 2009. 5

Contents 1 Introduction 12 1.1 Sentiment Analysis................................ 12 1.2 Probabilistic Topic Models............................ 13 1.3 Research Challenges............................... 14 1.3.1 Training Sentiment Classifier without Labelled Data......... 14 1.3.2 Simultaneously Detecting Topic and Sentiment............ 15 1.3.3 Capturing Sentiment and Topic Dynamics............... 15 1.3.4 Subjectivity Detection without Labelled Data............. 17 1.4 Thesis Contribution............................... 17 1.5 Thesis Overview................................. 18 2 Related Work 20 2.1 Sentiment Analysis................................ 20 2.1.1 Supervised Approaches......................... 20 2.1.2 Semi-supervised Approaches....................... 22 2.1.2.1 Monolingual Domain Adaptation............... 22 2.1.2.2 Cross-lingual Sentiment Classification............ 25 2.1.3 Unsupervised and Weakly-supervised Approaches........... 26 2.1.4 Sentiment Dynamics and Subjectivity Detection........... 28 2.1.4.1 Sentiment Dynamics...................... 28 2.1.4.2 Subjectivity Detection..................... 29 2.1.5 Discussion................................. 31 2.2 Generative Topic Models............................. 32 2.2.1 Latent Dirichlet Allocation....................... 34 2.2.1.1 Model Inference........................ 36 2.2.2 Static Topic Models........................... 39 2.2.3 Joint Sentiment and Aspect Models.................. 41 2.2.4 Dynamic Topic Models.......................... 44 6

CONTENTS 2.2.5 Discussion................................. 46 3 Weakly-Supervised Joint Sentiment-Topic Model 48 3.1 Introduction.................................... 48 3.2 Joint Sentiment-Topic (JST) Model....................... 49 3.2.1 Incorporating Model Priors....................... 52 3.2.2 Model Inference............................. 53 3.3 Reverse Joint Sentiment-Topic (Reverse-JST) Model............. 56 3.4 Experimental Setup............................... 57 3.4.1 Datasets.................................. 57 3.4.2 Defining Model Priors.......................... 58 3.4.3 Hyperparameter Settings........................ 59 3.4.4 Classifying Document Sentiment.................... 60 3.5 Experimental Results............................... 60 3.5.1 Sentiment Classification Results vs. Different Number of Topics... 60 3.5.2 Comparison with Existing Models................... 64 3.5.3 Sentiment Classification Results with Different Features....... 65 3.5.4 Topic Extraction............................. 68 3.6 Discussion..................................... 69 4 Dynamic Joint Sentiment-Topic Model 70 4.1 Introduction.................................... 70 4.2 Dynamic JST (djst) Model.......................... 71 4.2.1 Online Inference............................. 76 4.2.1.1 Deriving the Gibbs Sampler................. 76 4.2.2 Evolutionary Parameters Estimation.................. 78 4.2.2.1 Estimating the Evolutionary Matrix E t........... 78 4.2.2.2 Estimating the Weight Vector µ t............... 79 4.2.3 Hyperparameter Settings........................ 80 4.3 Experimental Setup............................... 82 4.3.1 Dataset.................................. 82 4.3.2 Evaluation Metrics............................ 84 4.4 Experimental Results............................... 85 4.4.1 Performance vs. Number of Time Slices................ 85 4.4.2 Comparison with Other Models..................... 88 4.4.3 Exploring Different Input Features................... 92 4.4.4 Topic Evolution.............................. 94 7

CONTENTS 4.5 Discussion..................................... 97 5 SubjLDA: a Weakly Supervised Topic Model for Subjectivity Detection 98 5.1 Introduction.................................... 98 5.2 The SubjLDA Model............................... 99 5.2.1 Model Inference............................. 101 5.3 Experimental Setup............................... 104 5.3.1 Dataset.................................. 104 5.3.2 Lexical Prior Knowledge......................... 105 5.3.3 Hyperparameter Setting......................... 106 5.4 Experimental Results............................... 107 5.4.1 Subjectivity Classification Results................... 107 5.4.2 Performance vs. Incorporating Different Priors............ 109 5.4.3 Sentiment Topics............................. 113 5.5 Discussions.................................... 113 6 Conclusion and Future Work 115 6.1 Future Work................................... 117 6.1.1 Modelling Linguistic Knowledge.................... 117 6.1.2 Incorporating Other Types of Prior Information........... 118 6.1.3 Automatic Estimation of Topic Number................ 118 6.1.4 Visualization and User Interfaces.................... 118 A The Fixed-Point Iteration Algorithm for Updating α 119 B Estimating the Weight Vector µ t of the djst Model 122 References 134 8

List of Figures 1.1 Amazon Kindle cover reviews. Text highlighted in green and red indicate the pros and cons of the product respectively.................. 16 2.1 LDA model..................................... 34 3.1 JST model..................................... 50 3.2 (a) Reverse-JST model; (b) JST model..................... 55 3.3 Sentiment classification accuracy vs. different topic number settings..... 61 3.3 Sentiment classification accuracy vs. different topic number settings..... 62 4.1 Dynamic JST model................................ 72 4.2 Three different time slice models with a total number of historical time slices S=3. The variable t 1 denotes epoch t 1............... 73 4.3 Exponential decay function with different decay rates............. 80 4.4 Dataset statistics. (a) number of reviews of each add-on over the epochs; (b) average user rating for each add-on over the epochs............ 83 4.5 Perplexity vs. number of time slices....................... 86 4.6 Accuracy vs. number of time slices........................ 87 4.7 Perplexity for each epoch............................. 89 4.8 Perplexity and sentiment classification accuracy versus number of topics... 90 4.9 Average training time per epoch with different number of time slices..... 91 4.10 Performance vs. different input features. Left panel: perplexity; right panel: sentiment classification accuracy..................... 93 4.11 Example topics evolved over time. Topic labels shown above the boxed topics were derived manually from the coloured words and the number denotes epoch number. Topics in upper and lower panels are the positive and negative sentiment topics respectively.................... 95 9

LIST OF FIGURES 4.12 Occurrence probability of topics with time. Positive and negative sentiment topics correspond to the topics listed in the upper and lower panel of Figure 4.11 respectively.............................. 96 5.1 subjlda model.................................. 100 5.2 Sentiment lexicon statistics............................ 110 5.3 Subjectivity classification performance vs. different prior information by gradually adding the subjective and neutral words. The vertical dashed line denotes the point where all the positive and negative words have been incorporated into the model; 200P+400N denotes adding the least frequent 200 positive and 400 negative words. For example, 500Neu denotes adding the least frequent 500 neutral words in addition to all the positive and negative words................................... 111 5.3 Subjectivity classification performance vs. different prior information by gradually adding the subjective and neutral words. The vertical dashed line denotes the point where all the positive and negative words have been incorporated into the model; 200P+400N denotes adding the least frequent 200 positive and 400 negative words. For example, 500Neu denotes adding the least frequent 500 neutral words in addition to all the positive and negative words................................... 112 5.4 Sentiment topics extracted by subjlda..................... 113 10

List of Tables 2.1 An illustration of four topic examples extracted from The American Political Science Review by LDA........................... 33 3.1 Parameter notations used in the JST model................... 51 3.2 Dataset statistics. Note: denotes before preprocessing and * denotes after preprocessing.................................... 57 3.3 Prior information statistics............................ 59 3.4 Significant test results. Note: blank denotes the performance of JST and Reverse-JST is significantly undistinguishable; * denotes JST significantly outperforms Reverse-JST............................. 63 3.5 Performance comparison with existing models on sentiment classification accuracy. (Note: boldface denotes the best results.).............. 64 3.6 Unigram and bigram features statistics..................... 66 3.7 Sentiment classification results with different features. Note: boldface denotes the best results............................... 66 3.8 Topic examples extracted by JST under different sentiment labels...... 67 4.1 Parameter notations for the djst model.................... 75 5.1 Notations used for the subjlda model..................... 101 5.2 Subjectivity classification results. (Boldface indicates the best results.)... 108 11