RNN for Sentiment Analysis: Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank

Similar documents
Package syuzhet. February 22, 2015

Distributed Representations of Sentences and Documents

The multilayer sentiment analysis model based on Random forest Wei Liu1, Jie Zhang2

RECURSIVE DEEP LEARNING FOR NATURAL LANGUAGE PROCESSING AND COMPUTER VISION

SZTE-NLP: Aspect Level Opinion Mining Exploiting Syntactic Cues

A Look Into the World of Reddit with Neural Networks

Semantic Compositionality through Recursive Matrix-Vector Spaces

Sentiment Classification on Polarity Reviews: An Empirical Study Using Rating-based Features

Recurrent Convolutional Neural Networks for Text Classification

Data Mining Yelp Data - Predicting rating stars from review text

SENTIMENT ANALYSIS: A STUDY ON PRODUCT FEATURES

RRSS - Rating Reviews Support System purpose built for movies recommendation

Bagged Ensemble Classifiers for Sentiment Classification of Movie Reviews

Sentiment analysis: towards a tool for analysing real-time students feedback

A Sentiment Analysis Model Integrating Multiple Algorithms and Diverse. Features. Thesis

Identifying Focus, Techniques and Domain of Scientific Papers

Sentiment analysis for news articles

NATURAL LANGUAGE QUERY PROCESSING USING PROBABILISTIC CONTEXT FREE GRAMMAR

Author Gender Identification of English Novels

A Distributed Sentiment Analysis Development Environment

Building a Question Classifier for a TREC-Style Question Answering System

VCU-TSA at Semeval-2016 Task 4: Sentiment Analysis in Twitter

Machine Learning for natural language processing

Compositional Sentiment Analysis

End-to-End Sentiment Analysis of Twitter Data

Computational Linguistics and Learning from Big Data. Gabriel Doyle UCSD Linguistics

Semi-Supervised Learning for Blog Classification

Robust Sentiment Detection on Twitter from Biased and Noisy Data

Gated Neural Networks for Targeted Sentiment Analysis

Sentiment Analysis for Movie Reviews

Knowledge Base Completion using Compositional Vector Space Models

A Decision Support Approach based on Sentiment Analysis Combined with Data Mining for Customer Satisfaction Research

Sentiment analysis on news articles using Natural Language Processing and Machine Learning Approach.

Sentiment Analysis of Twitter Data

DEPENDENCY PARSING JOAKIM NIVRE

Outline of today s lecture

TOOL OF THE INTELLIGENCE ECONOMIC: RECOGNITION FUNCTION OF REVIEWS CRITICS. Extraction and linguistic analysis of sentiments

Sentence paraphrase detection: When determiners and word order make the difference

Sentiment Analysis Tool using Machine Learning Algorithms

Towards SoMEST Combining Social Media Monitoring with Event Extraction and Timeline Analysis

Sentiment Analysis of Movie Reviews and Twitter Statuses. Introduction

Facilitating Business Process Discovery using Analysis

The Role of Sentence Structure in Recognizing Textual Entailment

Sentiment Analysis. D. Skrepetos 1. University of Waterloo. NLP Presenation, 06/17/2015

Reasoning With Neural Tensor Networks for Knowledge Base Completion

Sentiment analysis on tweets in a financial domain

Sentiment Analysis and Topic Classification: Case study over Spanish tweets

A Survey on Product Aspect Ranking

Web Mining. Margherita Berardi LACAM. Dipartimento di Informatica Università degli Studi di Bari

Fine-grained German Sentiment Analysis on Social Media

Clustering Connectionist and Statistical Language Processing

Automatic Text Analysis Using Drupal

A Sentiment Detection Engine for Internet Stock Message Boards

CIKM 2015 Melbourne Australia Oct. 22, 2015 Building a Better Connected World with Data Mining and Artificial Intelligence Technologies

Impact of Financial News Headline and Content to Market Sentiment

FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM

Active Learning SVM for Blogs recommendation

OPTIMIZATION OF NEURAL NETWORK LANGUAGE MODELS FOR KEYWORD SEARCH. Ankur Gandhe, Florian Metze, Alex Waibel, Ian Lane

II. RELATED WORK. Sentiment Mining

BLOG COMMENTS SENTIMENT ANALYSIS FOR ESTIMATING FILIPINO ISP CUSTOMER SATISFACTION

Data Mining Algorithms Part 1. Dejan Sarka

Supervised Feature Selection & Unsupervised Dimensionality Reduction

EFFICIENTLY PROVIDE SENTIMENT ANALYSIS DATA SETS USING EXPRESSIONS SUPPORT METHOD

CS 229, Autumn 2011 Modeling the Stock Market Using Twitter Sentiment Analysis

Bridging CAQDAS with text mining: Text analyst s toolbox for Big Data: Science in the Media Project

Interest Rate Prediction using Sentiment Analysis of News Information

Natural Language to Relational Query by Using Parsing Compiler

User Modeling with Neural Network for Review Rating Prediction

Evaluating Sentiment Analysis Methods and Identifying Scope of Negation in Newspaper Articles

Module Catalogue for the Bachelor Program in Computational Linguistics at the University of Heidelberg

A Content based Spam Filtering Using Optical Back Propagation Technique

Chapter 2 Automatic Expansion of a Social Network Using Sentiment Analysis

How To Write A Summary Of A Review

Neuro-Fuzzy Classification Techniques for Sentiment Analysis using Intelligent Agents on Twitter Data

ARABIC SENTENCE LEVEL SENTIMENT ANALYSIS

Open Domain Information Extraction. Günter Neumann, DFKI, 2012

FEATURE SELECTION AND CLASSIFICATION APPROACH FOR SENTIMENT ANALYSIS

Text Mining - Scope and Applications

Performance Analysis of Naive Bayes and J48 Classification Algorithm for Data Classification

Data Mining. Nonlinear Classification

Classification algorithm in Data mining: An Overview

Recurrent Neural Networks

Recognition. Sanja Fidler CSC420: Intro to Image Understanding 1 / 28

Web Document Clustering

An Insight Of Sentiment Analysis In The Financial News

Transcription:

RNN for Sentiment Analysis: Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank Borui(Athena) Ye University of Waterloo borui.ye@uwaterloo.ca July 15, 2015 1 / 26

Overview 1 Introduction 2 Related Work 3 Stanford Sentiment Treebank 4 Recursive Neural Models RNN: Recursive Neural Network MV-RNN: Matrix-Vector RNN RNTN: Recursive Neural Tensor Network Tensor Backprop through Structure 5 Experiments Borui(Athena) Ye July 15, 2015 2 / 26

Paper Information Richard Socher, Alex Perelygin, Jean Y.Wu, Jason Chuang, Christopher D. Manning, Andrew Y. Ng and Christopher Potts, Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank, In Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing, pp 1631-1642. 2013. Introduction Borui(Athena) Ye July 15, 2015 3 / 26

Introduction Semantic Compositionality : to calculate in a systematic way the polarity values of larger syntactic constituents as some function of the polarities of their subconstituents[ 1]. Corpus (Sentiment Treebank) 11,855 sentences based on extracted from movie reviews [ 2] 215,154 phrases parsed from sentences using Stanford Parser[ 3], each annotated by 3 annotators. [1] K. Moilanen and S. Pulman, Sentiment composition, in Proceedings of RANLP, vol. 7, 2007, pp. 378 382. [2] B. Pang and L. Lee, Seeing stars: Exploiting class relationships for sentiment categorization with respect to rating scales, in Proceedings of the 43rd Annual Meeting on Association for Computational Linguistics, Association for Computational Linguistics, 2005, pp. 115 124. [3] D. Klein and C. D. Manning, Accurate unlexicalized parsing, in Proceedings of the 41st Annual Meeting on Association for Computational Linguistics-Volume 1, Association for Computational Linguistics, 2003, pp. 423 430. Introduction Borui(Athena) Ye July 15, 2015 4 / 26

Introduction (Cont.) Introduction Borui(Athena) Ye July 15, 2015 5 / 26

Introduction (Cont.) Experiments 1 Fine-grained Sentiment For All Phrases 2 Full Sentence Binary Sentiment 3 Model Analysis: Contrastive Conjunction 4 Model Analysis: High Level Negation Negating Positive Sentences Negating Negative Sentences 5 Model Analysis: Most Positive and Negative Phrases Introduction Borui(Athena) Ye July 15, 2015 6 / 26

Related Work Semantic Vector Spaces. Distributed similarities of single words. But often fail to distinguish antonyms. Co-occurence of a word and its context [ 4]. How often a word appears in a certan syntactic context [ 5]. Compositionality in Vector Spaces. Most of them capture two word compositions. Word vector addition, multiplication, etc. [ 6] Represent phrases as matrixes and define composition method as matrix multiplication. [ 7] [4] P. D. Turney, P. Pantel, et al., From frequency to meaning: Vector space models of semantics, Journal of artificial intelligence research, vol. 37, no. 1, pp. 141 188, 2010. [5] S. Padó and M. Lapata, Dependency-based construction of semantic space models, Computational Linguistics, vol. 33, no. 2, pp. 161 199, 2007. [6] J. Mitchell and M. Lapata, Composition in distributional models of semantics, Cognitive science, vol. 34, no. 8, pp. 1388 1429, 2010. [7] E. Grefenstette and M. Sadrzadeh, Experimental support for a categorical compositional distributional model of meaning, in Proceedings of the Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics, 2011, pp. 1394 1404. Related Work Borui(Athena) Ye July 15, 2015 7 / 26

Related Work (Cont.) Sentiment Analysis. Bag-of-words representations [ 8]. Extracting features or polarity shifting rules on syntactic structures [ 9] Recursive Neural Models Will be covered later. [8] B. Pang and L. Lee, Opinion mining and sentiment analysis, Foundations and trends in information retrieval, vol. 2, no. 1-2, pp. 1 135, 2008. [9] L. Polanyi and A. Zaenen, Contextual valence shifters, in Computing attitude and affect in text: Theory and applications, 2006, pp. 1 10. Related Work Borui(Athena) Ye July 15, 2015 8 / 26

Stanford Sentiment Treebank Data retrieval and processing: Get movie review excerpts from the rottentomatoes.com, which includes 10,662 sentences, half positive, half negative. Parse sentences using the Stanford Parser. Using Amazon Mechanical Turk to label the resulting 215,154 phrases. Stanford Sentiment Treebank Borui(Athena) Ye July 15, 2015 9 / 26

Statistics Findings: 1 Most of the short n-grams are neural; 2 Longer n-grams are evenly distributed; 3 Extreme sentiment degrees rarely happen. Stanford Sentiment Treebank Borui(Athena) Ye July 15, 2015 10 / 26

Recursive Neural Models Tri-gram example of bottom up fashion: Initialization Initialize each word vector using uniform distribution: U( r, r), where r = 0.0001. Stack word vectors into matrix L R d V, where d is vector dimension, V is vocabulary size. Recursive Neural Models Borui(Athena) Ye July 15, 2015 11 / 26

RNN: Recursive Neural Network[ 10] where f = tanh, W R d 2d [10] R. Socher, J. Pennington, E. H. Huang, et al., Semi-supervised recursive autoencoders for predicting sentiment distributions, in Proceedings of the Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics, 2011, pp. 151 161. Recursive Neural Models Borui(Athena) Ye July 15, 2015 12 / 26

MV-RNN: Matrix-Vector RNN[ 11] Main Idea: represent every node in the parse tree both as a vector and a matrix. where W, W M R d 2d [11] R. Socher, B. Huval, C. D. Manning, et al., Semantic compositionality through recursive matrix-vector spaces, in Proceedings of the 2012 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning, Association for Computational Linguistics, 2012, pp. 1201 1211. Recursive Neural Models Borui(Athena) Ye July 15, 2015 13 / 26

MV-RNN: Matrix-Vector RNN (Cont.) Problem: size of parameters becomes very large and depends on the size of the vocabulary. Solution: use a simple powerful composition function with a fixed number of parameters. Recursive Neural Models Borui(Athena) Ye July 15, 2015 14 / 26

RNTN: Recursive Neural Tensor Network Main Idea: use the same, tensor-based compostition function for all nodes. Definition h R d : output of the tensor product V [1:d] R 2d 2d d : tensor that defines multiple bilinear forms. V [i] R 2d 2d : each slice of V [1:d]. Recursive Neural Models Borui(Athena) Ye July 15, 2015 15 / 26

RNTN: Recursive Neural Tensor Network (Cont.) Intuitively, we can interpret each slice of the tensor as capturing a specific type of composition. Recursive Neural Models Borui(Athena) Ye July 15, 2015 16 / 26

Tensor Backprop through Structure Each node is assigned a label via: y a = softmax(w s a) where W s R 5 d is the sentiment classification matrix. Recursive Neural Models Borui(Athena) Ye July 15, 2015 17 / 26

Tensor Backprop through Structure (Cont.) Goal: minimize the KL-divergence between the predicted distribution y i R C 1 at node i and the target distribution t i R C 1. The error function of a sentence is: where θ = (V, W, W s, L). Recursive Neural Models Borui(Athena) Ye July 15, 2015 18 / 26

Experiments Two kinds of experiment: Large quantitative evaluations on the test set. Linguistic phenomena: contrastive conjunction and negation. Baselines: Bag-of-words features + Naive Bayes (NB) Bag-of-words features + SVM (SVM) Bag-of-bigram features + Naive Bayes (BiNB) Averages of neural word vectors (VecAvg) RNN MV-RNN Experiments Borui(Athena) Ye July 15, 2015 19 / 26

Sentiment Classification 1 Exp. 1: Fine-grained Sentiment For All Phrases 2 Exp. 2: Full Sentence Binary Sentiment Experiments Borui(Athena) Ye July 15, 2015 20 / 26

Accuracy Recursive models work better on shorter grams. RNTN upper bounds other models at most n-gram lengths. Experiments Borui(Athena) Ye July 15, 2015 21 / 26

Exp. 3: Model Analysis: Contrastive Conjunction X but Y Structure : two phrases, X and Y, connect by but. Experiment result: the test set includes 131 cases (subset of the original test set), RNTN achieve a accuracy of 41%, compared to MV-RNN (37), RNN (36) and binb(27). Experiments Borui(Athena) Ye July 15, 2015 22 / 26

Exp. 4: Model Analysis: High Level Negation Set 1: Negating Positive Sentences Experiments Borui(Athena) Ye July 15, 2015 23 / 26

Exp. 4: Model Analysis: High Level Negation (Cont.) Set 2: Negating Negative Sentences Experiments Borui(Athena) Ye July 15, 2015 24 / 26

Exp. 4: Model Analysis: High Level Negation (Cont.) Experiments Borui(Athena) Ye July 15, 2015 25 / 26

Thank you! Borui(Athena) Ye July 15, 2015 26 / 26