Chinese Event Extraction Using Deep Neural Network with Word Embedding

Similar documents
The multilayer sentiment analysis model based on Random forest Wei Liu1, Jie Zhang2

Deep Learning for Chinese Word Segmentation and POS Tagging

A Look Into the World of Reddit with Neural Networks

Identifying Focus, Techniques and Domain of Scientific Papers

Micro blogs Oriented Word Segmentation System

Ngram Search Engine with Patterns Combining Token, POS, Chunk and NE Information

POSBIOTM-NER: A Machine Learning Approach for. Bio-Named Entity Recognition

Package syuzhet. February 22, 2015

Open Domain Information Extraction. Günter Neumann, DFKI, 2012

CIKM 2015 Melbourne Australia Oct. 22, 2015 Building a Better Connected World with Data Mining and Artificial Intelligence Technologies

Recurrent Convolutional Neural Networks for Text Classification

Machine Learning for natural language processing

Sentiment analysis: towards a tool for analysing real-time students feedback

Stanford s Distantly-Supervised Slot-Filling System

Building a Question Classifier for a TREC-Style Question Answering System

Architecture of an Ontology-Based Domain- Specific Natural Language Question Answering System

Automatic Detection and Correction of Errors in Dependency Treebanks

Ming-Wei Chang. Machine learning and its applications to natural language processing, information retrieval and data mining.

Learning to Process Natural Language in Big Data Environment

Network Machine Learning Research Group. Intended status: Informational October 19, 2015 Expires: April 21, 2016

Parsing Software Requirements with an Ontology-based Semantic Role Labeler

Microblog Sentiment Analysis with Emoticon Space Model

Transition-Based Dependency Parsing with Long Distance Collocations

SYSTRAN Chinese-English and English-Chinese Hybrid Machine Translation Systems for CWMT2011 SYSTRAN 混 合 策 略 汉 英 和 英 汉 机 器 翻 译 系 CWMT2011 技 术 报 告

31 Case Studies: Java Natural Language Tools Available on the Web

Opinion Mining with Deep Recurrent Neural Networks

Applications of Deep Learning to the GEOINT mission. June 2015

Bing Liu. Web Data Mining. Exploring Hyperlinks, Contents, and Usage Data. With 177 Figures. ~ Spring~r

Data Mining Algorithms Part 1. Dejan Sarka

Web Information Mining and Decision Support Platform for the Modern Service Industry

Recurrent Neural Networks

ANALYSIS OF LEXICO-SYNTACTIC PATTERNS FOR ANTONYM PAIR EXTRACTION FROM A TURKISH CORPUS

Machine Learning. CUNY Graduate Center, Spring Professor Liang Huang.

End-to-End Sentiment Analysis of Twitter Data

Domain Classification of Technical Terms Using the Web

Interactive Dynamic Information Extraction

Emotion Detection in Code-switching Texts via Bilingual and Sentimental Information

A Unified Architecture for Natural Language Processing: Deep Neural Networks with Multitask Learning

Machine Learning: Overview

Robust Sentiment Detection on Twitter from Biased and Noisy Data

Extracting Events from Web Documents for Social Media Monitoring using Structured SVM

Chinese Open Relation Extraction for Knowledge Acquisition

CIRGIRDISCO at RepLab2014 Reputation Dimension Task: Using Wikipedia Graph Structure for Classifying the Reputation Dimension of a Tweet

Towards SoMEST Combining Social Media Monitoring with Event Extraction and Timeline Analysis

Effective Self-Training for Parsing

Joint POS Tagging and Text Normalization for Informal Text

SYSTRAN 混 合 策 略 汉 英 和 英 汉 机 器 翻 译 系 统

3 Paraphrase Acquisition. 3.1 Overview. 2 Prior Work

Distributed forests for MapReduce-based machine learning

Protein-protein Interaction Passage Extraction Using the Interaction Pattern Kernel Approach for the BioCreative 2015 BioC Track

Joint Word Segmentation and POS Tagging using a Single Perceptron

ETL Ensembles for Chunking, NER and SRL

DEPENDENCY PARSING JOAKIM NIVRE

International Journal of Scientific & Engineering Research, Volume 4, Issue 11, November ISSN

Chapter 8. Final Results on Dutch Senseval-2 Test Data

SWIFT: A Text-mining Workbench for Systematic Review

Search Result Optimization using Annotators

Automated Content Analysis of Discussion Transcripts

Special Topics in Computer Science

Learning Character-level Representations for Part-of-Speech Tagging

Applying Co-Training Methods to Statistical Parsing. Anoop Sarkar anoop/

Pedestrian Detection with RCNN

Classification of Natural Language Interfaces to Databases based on the Architectures

Domain Independent Knowledge Base Population From Structured and Unstructured Data Sources

Benefits of HPC for NLP besides big data

A Survey on Product Aspect Ranking Techniques

Tibetan-Chinese Bilingual Sentences Alignment Method based on Multiple Features

Natural Language to Relational Query by Using Parsing Compiler

Semi-Supervised Learning for Blog Classification

Sentiment Analysis of Twitter Data

Integrating NLTK with the Hadoop Map Reduce Framework Human Language Technology Project

Learning is a very general term denoting the way in which agents:

Automatic Text Analysis Using Drupal

PULLING OUT OPINION TARGETS AND OPINION WORDS FROM REVIEWS BASED ON THE WORD ALIGNMENT MODEL AND USING TOPICAL WORD TRIGGER MODEL

A Knowledge-Poor Approach to BioCreative V DNER and CID Tasks

EFFICIENT DATA PRE-PROCESSING FOR DATA MINING

Lecture 6: Classification & Localization. boris. ginzburg@intel.com

Tibetan Number Identification Based on Classification of Number Components in Tibetan Word Segmentation

: Introduction to Machine Learning Dr. Rita Osadchy

Search and Information Retrieval

Doctoral Consortium 2013 Dept. Lenguajes y Sistemas Informáticos UNED

Sentiment Analysis Tool using Machine Learning Algorithms

Predict Influencers in the Social Network

A Comparative Study on Sentiment Classification and Ranking on Product Reviews

Extracting Clinical entities and their assertions from Chinese Electronic Medical Records Based on Machine Learning

Web Document Clustering

Transfer Learning for Latin and Chinese Characters with Deep Neural Networks

SENTIMENT ANALYSIS: A STUDY ON PRODUCT FEATURES

Search and Data Mining: Techniques. Text Mining Anya Yarygina Boris Novikov

Summarizing Online Forum Discussions Can Dialog Acts of Individual Messages Help?

Deep learning applications and challenges in big data analytics

Distance Metric Learning in Data Mining (Part I) Fei Wang and Jimeng Sun IBM TJ Watson Research Center

Learning Translation Rules from Bilingual English Filipino Corpus

Supporting Online Material for

Transcription:

Chinese Event Extraction Using Deep Neural Network with Word Embedding Yandi Xia and Yang Liu The University of Texas at Dallas Computer Science Department yandixia,yangl@hlt.utdallas.edu on Chinese. Event trigger identification is often the first step of an event extraction system. It identifies the words that most explicitly indicate the occurrence of events. For example, in this sentence: Yahoo announced its acquisition of KIMO website. Here acquisition is an event trigger, and it triggers one of the subtypes of BUSINESS event type Merge-Org event. There are some trends in the research for event extraction in the past years. First, much prior work has focused on exploiting rich language features like syntax, part-of-speech (POS), parsing structure, named entities, synonym dictionaries, etc. For example, Chen and Ji (2009) and Chen and NG (2012) both used a variety of such features, achieving the stateof-the-art performance for Chinese event extraction. In English corpus, the earliest work on ACE corpus (Ahn, 2006) also designed a lot of features for this task and set the baseline of English event extraction. However, these systems have some common drawbacks: a) those features vary from corpus to corpus, and language to language. One almost needs to setup the whole system for new data. For example, (Chen and Ji, 2009) designed a character based system with character based features for Chinese event extraction, in order to reduce the errors caused by word segmentation. However, such features make no sense for English corpora. On the other hand, in the English systems, features are needed to represent phenomena such as tenses, whereas in Chinese, this is not a problem. Same thing happens if one wants to migrate an event extraction system from news article domain to biomedical domains. Most of the features designed previously are not very useful for the new data. b) those features depend on some fundamenarxiv:1610.00842v1 [cs.cl] 4 Oct 2016 Abstract A lot of prior work on event extraction has exploited a variety of features to represent events. Such methods have several drawbacks: 1) the features are often specific for a particular domain and do not generalize well; 2) the features are derived from various linguistic analyses and are error-prone; and 3) some features may be expensive and require domain expert. In this paper, we develop a Chinese event extraction system that uses word embedding vectors to represent language, and deep neural networks to learn the abstract feature representation in order to greatly reduce the effort of feature engineering. In addition, in this framework, we leverage large amount of unlabeled data, which can address the problem of limited labeled corpus for this task. Our experiments show that our proposed method performs better compared to the system using rich language features, and using unlabeled data benefits the word embeddings. This study suggests the potential of DNN and word embedding for the event extraction task. 1 Introduction Event extraction is a task of information extraction. It is a complicated task including many sub-tasks. In ACE, event extraction includes four sub-tasks (Ahn, 2006; Grishman et al., 2005): event trigger identification, trigger type classification, argument identification, and argument role classification. In this paper, we focus on trigger identification

tal NLP modules (e.g., POS tagging, parsing, named entity recognition), which are not perfect and introduce a lot of noise in the derived features. This is especially true for Chinese (and languages with low resources), in which many NLP tools do not perform well. c) some of features are expensive to acquire, because expert knowledge or resources are required, for example, synonym dictionaries. Second, due to the lack of labeled data for event extraction, more and more research started focusing on semi-supervised learning (Liao and Grishman, 2010; Ali, 2014; Li et al., 2014), unsupervised learning (Rusu et al., 2014), and distantly supervised learning (Reschke et al., ). These studies show that additional unlabeled data is very useful to acquire more information or understand about language. For example, large unlabeled data can be used to learn the patterns for event extraction (Liao and Grishman, 2010; Huang and Riloff, 2012). Motivated by the need to overcome the problem with designing features and the potential benefit of unlabeled data, in this pilot study, we set to answer three questions: (1) can we get around feature engineering in event trigger detection using the deep neural networks and word embedding? (2) is word embedding a better representation for this task than the large set of carefully crafted discrete features? (3) can we effectively leverage unlabeled data for word embedding in the trigger detection task? Word embedding has been proved to be very successful combining with deep learning, the increasingly popular learning method. After Collobert et al. (2011) and Socher et al. (2012) brought up a unified deep structure for NLP tasks, much work using this combination has emerged to challenge the traditional feature based methods. Collobert et al. (2011) showed great performance in tasks such as part-ofspeech-tagging, chunking, named entity recognition, and semantic role labeling with one unified deep structure, which is comparable to those feature based methods. Li and Liu (2014) s work applied word embedding on text normalization task which use the similarity between word vector to represent the sematic relationship between two words. Qi et al. (2014) and Zheng et al. (2013) adopted this structure for Chinese NLP tasks, and beat the state-of-the-art performance in word segmentation, part-of-speech-tagging and named entity recognition tasks. In addition, Li et al. (2015) applied word embedding on two large corpus, one is a set of news articles and the other is their corresponding summaries. Then for each token in two word embedding, they design additional features to help estimate their importance for final summary. The experiments showed that the feature based on word embedding are very useful. Inspired by these successful efforts, in this work we design a deep structure event extraction system which takes word embedding representation of Chinese data. We expect that deep learning can learn abstract feature representation and word embedding can effectively represent semantic and syntactic similarity between words, and thus can help identify new event trigger words that are not in the training examples. For example, if beat is a trigger word in the training set, it is hard to use synonym information to determine word attack is a trigger in the test data. However, word embedding may be able to find such semantic similar words, in either a supervised or unsupervised fashion, and improve system s recall. To our knowledge, there is no prior work that has explored the use of word embedding and deep learning for Chinese event extraction. In this work, we build a deep neural network model which represents Chinese characters with word embedding vectors. To leverage unlabeled data, we generate word embeddings from that and use them for pre-training in DNN. We evaluate our methods on the Chinese ACE corpus. Different from previous work that used ground truth information such as named entities, time and value labels, we use a more realistic setup with such information automatically generated. Our results show that we can achieve better performance using DNN than a feature-based maximum entropy classifier for event trigger detection, and then using unlabeled data for word embedding pretraining has additional benefit. The rest of the paper is structured as follows. Section 2 describes the event trigger identification methods, including the maxi-

mum entropy classifier that uses a set of rich features, and our DNN and word embedding system. Section 3 shows the experimental results. Conclusions and future work appear in Section 4. 2.2 DNN Model 2 Event Trigger Identification The follow introduces the two methods we use for event trigger word identification. 2.1 Baseline Following the works of (Chen and Ji, 2009; Chen and NG, 2012), we build the feature based baseline system. As observed in (Chen and Ji, 2009), there is a considerable portion of segmentation errors because of the low performance of the Chinese word segmentation tool and the ambiguity in human labels. Therefore, we model it as a sequence-labeling task and use the BIO encoding method (each character either begins, is inside, or outside of a trigger word) As for the features, both papers did a lot in feature analysis, therefore we adopt most of their features, listed below: Lexical features: current character; the characters that surround it; current POS; and the POS that surround it; the combination of current character and its POS. Syntactic features: depth of the current character in the parse tree; path from the current node to the root; the phrase structure expanded from the parent node; phrase type. Semantic dictionary: whether it is in the predicate list from (Xue and Palmer, 2009); synonym entry number of the current character. Nearest entity: the type of the left nearest entity to the current character; the type of the right nearest entity to the current character. Here, distance is measured by text length. In (Chen and Ji, 2009; Chen and NG, 2012), human labeled named entity information was used. In this study, we use a more realistic setup we use automatically identified named entities in our features. In our experiments, we will evaluate the effect of using such imperfect features. Figure 1: Structure of DNN model. We followed the work of (Collobert et al., 2011) and designed a similar DNN architecture, as shown in Figure 1. Considering the word segmentation problem in Chinese, we also choose the window based character stream as input. Here we model the input character as its index in the character dictionary. The first layer is a lookup table whose entries are word embedding vectors. Then we concatenate those selected vectors and pass it to the hidden layers. In the hidden layers we use nonlinear Tanh as the activation function. At the top of the structure, we put softmax to output the probabilities of the character being a part of a trigger. Note in this method, we do not use any linguistically motivated features. The input used in this DNN is just word embedding vectors. All the weights in the DNN including the word embedding vectors are trained automatically in the back-propagation process using the gradient descent algorithm. During testing, the DNN system gives the probabilities of each character as BIO tag, we treat them as emission probabilities, together with the transition probability from baseline CRF system, we use Viterbi decoding process to conduct the final prediction. Li and Liu (2015a) and Li and Liu (2015b) showed that using separated but well trained weight in Viterbi decoding can give improvement in certain conditions. The DNN tool we use is python based software Theano (Bastien et al., 2012).

2.3 Using Unlabeled Data Unlabeled data contains a lot of abstract syntax and semantic information that can be very useful to NLP tasks. In order to take advantage of unlabeled data (simply Chinese sentences, no event labels), we first use the RNN based word2vec toolkit from to generate the initial word vector dictionary as pre-training. These vectors then will act like part of the DNN weights and change their values in the supervised learning progress via back propagation. 3 Experiments The corpus we use is ACE 2005 Chinese corpus. There are 633 documents in the corpus. We randomly choose 66 documents as the test data, and 567 documents as training data, which is similar to (Chen and Ji, 2009). For performance metric, we use precision, recall, and F1 score. If the offset and word length of the identified character chunk exactly match the gold value, we consider that the corresponding trigger is correctly identified. The unlabeled data we use is from (Graff and Chen, 2005). It contains 2,466,840 Chinese newswire documents, totaling 3.9 GB. We use 100K documents of them. Table 1 shows the trigger identification results for different methods. For the baseline system using ME classifier, we show two results. One is using the named entities obtained from the Stanford corenlp tool, and the other one uses ground truth NER labels. It is clear that ground truth NER information can boost the performance considerably. However, such accurate information is very hard to get in real world cases. For the DNN model, we also report two results, without using the unlabeled data to pretrain the network, vs. the one that takes advantage of large amount of unlabeled data for pretraining. We can see that pretraining improves the system performance significantly. In addition, during the experiments, we noticed that DNN without pretraining converges far more slower than that with pretraining. Comparing the results using the DNN and the ME classifier, we can see even without using the unlabeled data, the DNN results are better than the feature-based ME classifier (using automatically generated NE information). This suggests that no feature engineering is required in the DNN model it simply uses the character embedding to generate more effective abstract feature representation for this classification task. When unlabeled data is incorporated, the DNN performance is much better, even outperforming using the reference NE information in the ME classifier. In addition, the improved DNN results over the ME method are because of the higher recall, which is consistent with our expectation using word embedding vectors can find semantically similar words. recall precision F1-score CRF baseline auto NE 48.09 71.37 57.47 CRF baseline ref NE 52.17 74.71 61.44 DNN 51.36 66.78 58.06 DNN with unlabeled data 61.14 67.77 64.29 Table 1: Event trigger word identification results using different systems. Figure 2 shows the F1 score when varying the vector size of the character embeddings. This pattern is similar to that for other tasks. When the size is too small, it cannot represent detailed information of the language characteristics; when it is too big, the system has too many parameters and loses the power of the abstract representation. Figure 2: Trigger word identification results using DNN with different character vector sizes.

4 Conclusions and Future Work In this paper, we applied word embeddings to represent Chinese characters. Our result on the event trigger identification task shows that it is a better representation compared to human designed language specific features. We also show that the combination of word embedding and DNN outperform the classifier that relies on a large set of linguistic features, and that this framework can effectively leverage unlabeled data to improve system performance. This is the first study exploring deep learning and word embedding for Chinese event extraction. In our current work, we use a relatively small window of characters as the input of the DNN. For future work, we plan to find a way to model longer context in DNN for event extraction. Furthermore, we will move on CNN and RNN architecture for this task. Acknowledgments References [Ahn2006] David Ahn. 2006. The stages of event extraction. In Proceedings of the Workshop on Annotating and Reasoning about Time and Events, pages 1 8. Association for Computational Linguistics. [Ali2014] H Ali. 2014. Event extraction for balkan languages. EACL 2014, page 65. [Bastien et al.2012] Frédéric Bastien, Pascal Lamblin, Razvan Pascanu, James Bergstra, Ian J. Goodfellow, Arnaud Bergeron, Nicolas Bouchard, and Yoshua Bengio. 2012. Theano: new features and speed improvements. [Chen and Ji2009] Zheng Chen and Heng Ji. 2009. Language specific issue and feature exploration in chinese event extraction. In Proceedings of Human Language Technologies: The 2009 Annual Conference of the North American Chapter of the Association for Computational Linguistics, Companion Volume: Short Papers, pages 209 212. Association for Computational Linguistics. [Chen and NG2012] Chen Chen and V Incent NG. 2012. Joint modeling for chinese event extraction with rich linguistic features. In In COL- ING. Citeseer. [Collobert et al.2011] Ronan Collobert, Jason Weston, Léon Bottou, Michael Karlen, Koray Kavukcuoglu, and Pavel Kuksa. 2011. Natural language processing (almost) from scratch. The Journal of Machine Learning Research, 12:2493 2537. [Graff and Chen2005] David Graff and Ke Chen. 2005. Chinese gigaword. LDC Catalog No.: LDC2003T09, ISBN, 1:58563 230. [Grishman et al.2005] Ralph Grishman, David Westbrook, and Adam Meyers. 2005. Nyus english ace 2005 system description. ACE, 5. [Huang and Riloff2012] Ruihong Huang and Ellen Riloff. 2012. Bootstrapped training of event extraction classifiers. In Proceedings of the 13th Conference of the European Chapter of the Association for Computational Linguistics, pages 286 295. Association for Computational Linguistics. [Li and Liu2014] Chen Li and Yang Liu. 2014. Improving text normalization via unsupervised model and discriminative reranking. In Proceedings of ACL. [Li and Liu2015a] Chen Li and Yang Liu. 2015a. Improving named entity recognition in tweets via detecting non-standard words. In Proceedings of ACL. [Li and Liu2015b] Chen Li and Yang Liu. 2015b. Joint pos tagging and text normalization for informal text. In Proceedings of IJCAI. [Li et al.2014] Peifeng Li, Qiaoming Zhu, and Guodong Zhou. 2014. Employing event inference to improve semi-supervised chinese event extraction. Proceedings of COLING 2014, the 25th International Conference on Computational Linguistics: Technical Papers, page 21612171. [Li et al.2015] Chen Li, Yang Liu, and Lin Zhao. 2015. Using external resources and joint learning for bigram weighting in ilp-based multidocument summarization. In Proceedings of NAACL-HLT, pages 778 787. [Liao and Grishman2010] Shasha Liao and Ralph Grishman. 2010. Filtered ranking for bootstrapping in event extraction. In Proceedings of the 23rd International Conference on Computational Linguistics, pages 680 688. Association for Computational Linguistics. [Qi et al.2014] Yanjun Qi, Sujatha G Das, Ronan Collobert, and Jason Weston. 2014. Deep learning for character-based information extraction. In Advances in Information Retrieval, pages 668 674. Springer. [Reschke et al.] Kevin Reschke, Martin Jankowiak, Mihai Surdeanu, Christopher D Manning, and Daniel Jurafsky. Event extraction using distant supervision. [Rusu et al.2014] Delia Rusu, James Hodson, and Anthony Kimball. 2014. Unsupervised techniques for extracting and clustering complex events in news. ACL 2014, page 26.

[Socher et al.2012] Richard Socher, Yoshua Bengio, and Christopher D Manning. 2012. Deep learning for nlp (without magic). In Tutorial Abstracts of ACL 2012, pages 5 5. Association for Computational Linguistics. [Xue and Palmer2009] Nianwen Xue and Martha Palmer. 2009. Adding semantic roles to the chinese treebank. Natural Language Engineering, 15(01):143 172. [Zheng et al.2013] Xiaoqing Zheng, Hanyang Chen, and Tianyu Xu. 2013. Deep learning for chinese word segmentation and pos tagging. In EMNLP, pages 647 657.