Part of Speech Tagging - A solved problem?

Size: px
Start display at page:

Download "Part of Speech Tagging - A solved problem?"

Transcription

1 Part of Speech Tagging - A solved problem? Wolfgang Fischl December 8, 2009 Abstract Since 100 B.C. humans are aware that the language consists of several distinct parts, called parts-of-speech. Identifying those parts-of-speech plays a crucial role in many fields of linguistics. Since TAGGIT, the first large-scale part-of-speech tagger, many algorithms and methods have been developed. Such include rule-based, probabilistic and hybrid taggers. When tagging large text corpora some problems might arise and finally the question is asked if part-of-speech tagging is a solved problem. 1 Introduction Part-of-speech (also known as: POS, tagsets, world classes, morphological classes, or lexical tags) play an important role in nearly every human language. What are part-of-speech? Already 100 B.C. someone identified eight parts-of-speech in Greek language: noun, verb, pronoun, preposition, adverb, conjunction, participle, and article. These eight are the basis for most part-of-speech descriptions and are also included in the more recent lists of parts-of-speech (see Table 1). Name year of appearance No. of classes Brown Corpus Penn Treebank C5 tagset C7 tagset Table 1: Recent part-of-speech lists The POS tag of a word is not only influenced by the word itself, but also by the words neighbors. In fact one word can have different POS tags depending on its location. For example the word object can be used either as noun (the object) or as verb (to object). Knowing the exact POS can be very helpful in many language processing applications. The POS of the word object changes the pronunciation (as noun the first syllable is pronounced as verb the second) [1]. Other applications include text indexing or the linguistic analysis of large tagged text corpora [2]. Part-of-speech tagging is the automatic assignment of such tags to entire texts. In this paper the goal is to understand the part-of-speech tags, the methods used by all part-of-speech tagging algorithms, some part-of-speech taggers and their problems and to explore if part-of-speech tagging is an already solved computational problem. 1

2 2 Part-of-speech tagging 2.1 Tags in English The eight part-of-speech tags identified in 100 B.C. are just a broad description of the English language. Actually there is more interest in identifying more specific POS tags. The smallest now used POS tagset is 45 tags big, the largest has 146 tags (see Table 1). What are now these 45 tags of the Penn Treebank tagset? For example the nouns themselves are separated into four classes: Singular, or mass nouns (llama, snow); Plural nouns (llamas); Proper nouns, singular (IBM); Proper nouns, plural (Austrians). The verbs consist of 6 classes: base form (eat), past tense (ate), gerund (eating), past participle (eaten), non- 3sg pres (eat), 3g pres (eats). A detailed list of the Penn Treebank tags can be found in [3]. The classes can be categorized in open and closed classes. Closed classes are those that have a fixed number of members, for example articles are a closed class (in English articles are only: the, a and an). Open classes don t have a fixed number of members, because occasionally new words are found, or borrowed from other languages. In the English language there are four open classes: nouns, verbs, adjectives and adverbs [1]. 2.2 The tagging process The input for every tagging algorithm is a string of words and a specified tagset. The output is a single best tag for each word. Most algorithms have 3 steps in common to find the best tag: ([1], [2]): 1. Tokenization is part of most tagging algorithms or is done as preprocessing step. The text is divided into tokens. These include end-of-sentence punctuation marks and word-like units. 2. Ambiguity look-up uses a lexicon and the tagset to assign each word a list of possible part-of-speech tags. Is a word not in the lexicon a guesser can make assumptions using the word neighborhood. The guesser only needs to consider the open class tag types, because most lexicons contain all closed-class words. 3. Disambiguation is the last and also the hardest part of the 3 steps. After step 2 every word has multiple possible tags. Disambiguation now tries to eliminate all but one tag. There are two classes of methods that can be used for disambiguation. Rule-based tries to eliminate tags by using specific rules, and one such tagger is described in section 3. Stochastic taggers use a tagged training set to compute a probability of a given word in a specific context. The tag with the highest probability is chosen. A stochastic tagger is described in section 4. Some methods use a combination of both. One such method is described in section 5. 2

3 3 Rule-based POS taggers One of the first POS taggers was called TAGGIT and proposed in 1971 [4]. It was based on context-pattern rules and used a 71-item tagset and a disambiguation grammar of 3,300 rules [1]. TAGGIT was correct in 77 per cent of all words in the million word Brown University corpus. Because all rules need to be hand-written, a rule-based tagger requires lots of work and knowledge of the language. In 1992 a new tagger was proposed by Brill, et. al., which learns the rules itself [5]. The motivation behind the simple rule-based part-of-speech tagger from Eric Brill is that stochastic or probabilistic taggers (see section 4) achieve substantially better results in tagging. On the other hand earlier rule-based taggers needed a wide variety of rules and knowledge of the language. Therefore he created a tagger which generates rules from a set of templates and a training text. All words are automatically assigned an initial tag with a basic lexical tagger. Then patches are generated out of eight patch templates (excerpted from [5]): Change tag a to tag b when: 1. The preceding (following) word is tagged z. 2. The word two before (after) is tagged z. 3. One of the two preceding (following) words is tagged z. 4. One of the three preceding (following) words is tagged z. 5. The preceding word is tagged z and the following word is tagged w. 6. The preceding (following) word is tagged z and the word two before (after) is tagged w. 7. The current word is (is not) capitalized. 8. The previous word is (is not) capitalized. For each error triple < tag a, tag b, number > and patch the reduction in error is computed. The patch with the highest reduction of error is added to the list of patches. The patch acquisition procedure continues until a specific threshold is reached (e.g. no more reduction in error). A new text is tagged by first assigning an initial tag with the same lexical tagger the training corpus has been tagged. Then each patch from the list of patches is applied, hopefully decreasing the error rate. [5] showed that an accuracy of 95%-99% is possible. Furthermore only 71 patches were generated achieving these high accuracies, which showed that rulebased taggers can compete with stochastic methods. This tagger is very easy to understand and can be used with many different text genres, just by training a tagger for that specific genre. This kind of tagging is sometimes also called transformation-based tagging. 3

4 The advantages of rule-based taggers are that rules can be hand-written and easy comprehended, further rule-based taggers can achieve good results with just a few rules. The disadvantages are that the rules are language and corpus specific and that programming a really good tagger takes a large amount of work and needs lots of linguistic knowledge. 4 Stochastic POS taggers Stochastic taggers make use of probabilities. The goal is to search the most probable sequence of tags for a given set of words. This can be modeled as a special case of Bayesian inference, called Hidden Markov Model. ˆt n 1 = argmax ˆt n 1 P (t n 1 w n 1 ) argmax ˆt n 1 n P (w i t i ) P (t i t i 1 ) (1) ˆt n 1 of eqn. 1 gives us the most probable sequence of tags. It can be calculated with formula 1, allthough several assumptions are made: 1. The probability (P (t n 1 w1 n )) of a sequence of words w1 n receiving the sequence of tags t n 1 can be transformed with Bayes rule to P (wn 1 tn 1 )P (tn 1 ). i=1 P(w n 1 ) The denominator of the fraction can be dropped since the probability of a sequence of words w n 1 doesn t change for a different tag sequence. Leaving us with a prior P (t n 1 ) and a likelihood P (w n 1 t n 1 ). The likelihood is now the probability of seeing a sequence of words given a sequence of tags. 2. The words are independent of each other. Therefore P (w n 1 t n 1 ) n i=1 P (w i t i ). 3. The probability of a tag is only dependent on its predecessor: P (t n 1 ) n i=1 P (t i t i 1 ). Such taggers are called bigram taggers. The algorithm can be improved by also using trigrams (tag is dependent on two predecessors) and also a combination of bigram and trigram probabilities is possible. Combining all assumptions gives us eqn. 1. [1] 4.1 Calculation probabilities In eqn. 1 are 2 unknown probabilities but those are easy to train. An estimate of probabilities can be calculated from an annotated corpus, e.g. the 1 million word Brown corpus. To calculate an estimate for the probability of a tag t i coming after a tag t i 1 the ratio of counts in eqn. 2 is calculated for every pair < t i, t i 1 > in the corpus. P (t i t i 1 ) = C (t i 1, t i ) C (t i 1 ) An estimate for the word likelihoods P (w i t i ) can also be calculated from a ratio of counts. (eqn. 3) (2) P (w i t i ) = C (t i, w i ) C (t i ) (3) 4

5 The computation of ˆt n 1 (eqn. 1) over all possible tags t i and all possible words w j is exponential. As formalized Hidden Markov Model ˆt n 1 can be calculated in linear time. 4.2 Hidden Markov Models [1] To calculate the most probable sequence of tag ˆt n 1 we use a Hidden Markov Model. It s called Hidden because only the observations are given and the states (or tags in our case) are not known and can t be determined, because our equation (1) depends on tags of previous words. A Hidden Markov Model consist of the tuple M = (Q, A, B, S, E) (4) Q is the set of states q 1, q 2,..., q n. Here the states are all possible tags. A is the set of transition probabilities a 01, a 02,..., a n1,..., a nn and a ij represents the transition of state q i to state q j. This set can be written as a transition probability matrix and corresponds to the prior P (t i t i 1 ). Each element in the matrix tells us the probability of having a tag j after the tag i. B is the set of observation likelihoods b i (o t ) and are the probability of an observation o t being generated from a state i. The observation likelihoods B can also be written as matrix and correspond to the likelihood P (w i t i ), therefore the element in the matrix tells us the probability of a word w i having tag t i. S is the start state and E the end state. Further there is a set of observation symbols O, which are the words w 1,..., w n. The most likely tag sequence or transitions have to be found given this sequence of observation symbols. The Viterbi algorithm is the most common decoding algorithm for HMMs that gives the most likely tag sequence given a set of states, transition probabilities and observation likelihoods. Algorithm 1 Viterbi 1: num states #tags 2: Create path probability matrix viterbi[num states + 2, n + 2] 3: viterbi[0, 0] 1.0 4: for word w = 1 to n do 5: for tag t = 1 to num states do 6: viterbi[t, w] max 1 t num states[viterbi[t, w 1] a t,t] b t (o w ) 7: back pointer[t, w] argmax 1 t num states[viterbi[t, w 1] a t,t] 8: end for 9: end for 10: Backtrace from highest probability state in final column of viterbi[] and 11: return path. The algorithm 1 works now as following: 5

6 1. A path probability matrix is created. A start and end tag and a start and end observation is added. 2. The cell with the start tag and the start observation (viterbi[0, 0]) is initialized with 1. All other values in the first column are Now the algorithm moves from column to column. 4. Filling every row in the current column with the maximum of the previous probability of the tag times the transition probability from tag t to t. This maximum is then multiplied with the observation likelihood of the current word o given the current tag t. 5. Also a pointer is remembered to backtrace the path of the maximum probabilities. 6. Step 3-5 are repeated until the last column is reached. When the path is back-traced from the largest value in the last column to the first column, the sequence of most probable tags can be reconstructed. Stochastic taggers can be easily adapted to new languages and new corpora by just re-training the model. On the other hand stochastic taggers can be hard to program. 5 Hybrid Taggers Both stochastic and probabilistic taggers are available, but there are also taggers that combine both methods. CLAWS4 is such a hybrid tagger [6]. CLAWS4 uses in a first step a sequence of eight tests to assign initial tags. These tests also include rules for spoken English, multipart words, tags for specific word-endings and words containing hyphens. After the first step each word has one or more tags and in the second step a HMM eliminates all but one tag. [6] claims an accuracy of percent. 6 Problems in POS Tagging Unknown words The largest problem for POS taggers are unknown words. As mentioned earlier those can only appear with open-class types, but this is still a fair amount of tags to choose from. Several approaches are available to tag unknown words: Guess the tag from the context. This means that depending on the previous and following tags a tag for the word is chosen. This further implies that the tag is ambiguous among all possible tags with equal probability. Some algorithms use the morphology of the word to guess its tag. Morphology means that the prefix and suffix of a word is analyzed (e.g.: Capitalization means that the word is likely to be a proper noun). Other algorithms consider probability of all the suffixes of all words in the training corpus. 6

7 Other algorithms use a combination of all features (prefixes, suffixes, capitalization and context) of a word together with a linear regression model, or another classification task, to predict a tag for the unknown word. Spelling errors So far only text without spelling errors are considered. But what if we start to tag web pages or magazines that have many type- and spelling-errors? Correcting these text beforehand is very crucial and often the correction is as ambiguous as assigning a tag to a word. For example the misspelled word acress can have at least 6 candidate corrections: actress, cress, caress, access, across, acres. Choosing the unintended candidate can change the tags of a whole sequence of words. Tokenization In section 2.2 the first step of tagging is described as tokenization. It depends on the tagging algorithm, but starting with wrong tokens can also change the sequence of tags. For example some algorithms split contractions and the s-genitive from the word stems (e.g.: children s) or separate proper nouns like New York. 7 Conclusion In many classification tasks one would be very happy to achieve results near the 96% as shown in the tagging algorithms presented in this paper and also in [1]. Therefore the question arises: Is part-of-speech tagging a solved problem? If one looks at the results of Marcus et. al. [3], for example, where they found out that human annotators agreed in just 96% percent of the cases, the answer would be yes, since the last 3% can be of the ambiguity in the language itself. Some special cases may not be clearly defined and therefore lead to different tags. The accuracy can also be limited by the error in the training data. On the other hand we have just looked at algorithms for the English language. All though there are taggers for different languages and also some taggers can be easily modified for another language than English, not every language is as suitable for tagging with the same methods used in the English language. Further research could investigate how to tag different languages with just one tagger, and with that tagger it might be possible to find common structures in all languages. Or some languages turn out to be not tagable. Apparently achieving an accuracy of 100% might not even be possible, it would be a challenging task to find such tagger. The computational linguistic community and research in part-of-speech tagging proposed many algorithms that are now used in many different fields (e.g.: Hidden Markov Models). The accuracy of any part-of-speech tagger depends to the main part on a welltagged training corpus. As we have seen, current algorithms perform very well in part-of-speech tagging. Taggers for other languages and also other research domains can benefit from the ongoing research in part-of-speech tagging. 7

8 References [1] D. Jurafsky and J. H. Martin, Speech and Language Processing: An introduction to speech recognition, computational linguistics and natural language processing., ch. 5: Part-of-Speech Tagging, pp Prentice Hall, [2] A. Voutilainen, The Oxford handbook of computational linguistics, ch. 11: Part-of-Speech Tagging, pp Oxford University Press, [3] M. P. Marcus, B. Santorini, and M. A. Marcinkiewicz, Building a large annotated corpus of english: The penn treebank, Computational Linguistics, vol. 19, no. 2, pp , [4] B. B. Greene and G. M. Rubin, Automatic grammatical tagging of english, tech. rep., Department of Linguistics, Brown University, [5] E. Brill, A simple rule-based part of speech tagger, in HLT 91: Proceedings of the workshop on Speech and Natural Language, (Morristown, NJ, USA), pp , Association for Computational Linguistics, [6] R. Garside and N. Smith, A hybrid grammatical tagger: Claws4, in Corpus Annotation: Linguistic Information from Computer Text Corpora, pp , Longman,

Testing Data-Driven Learning Algorithms for PoS Tagging of Icelandic

Testing Data-Driven Learning Algorithms for PoS Tagging of Icelandic Testing Data-Driven Learning Algorithms for PoS Tagging of Icelandic by Sigrún Helgadóttir Abstract This paper gives the results of an experiment concerned with training three different taggers on tagged

More information

POS Tagsets and POS Tagging. Definition. Tokenization. Tagset Design. Automatic POS Tagging Bigram tagging. Maximum Likelihood Estimation 1 / 23

POS Tagsets and POS Tagging. Definition. Tokenization. Tagset Design. Automatic POS Tagging Bigram tagging. Maximum Likelihood Estimation 1 / 23 POS Def. Part of Speech POS POS L645 POS = Assigning word class information to words Dept. of Linguistics, Indiana University Fall 2009 ex: the man bought a book determiner noun verb determiner noun 1

More information

UNKNOWN WORDS ANALYSIS IN POS TAGGING OF SINHALA LANGUAGE

UNKNOWN WORDS ANALYSIS IN POS TAGGING OF SINHALA LANGUAGE UNKNOWN WORDS ANALYSIS IN POS TAGGING OF SINHALA LANGUAGE A.J.P.M.P. Jayaweera #1, N.G.J. Dias *2 # Virtusa Pvt. Ltd. No 752, Dr. Danister De Silva Mawatha, Colombo 09, Sri Lanka * Department of Statistics

More information

POS Tagging 1. POS Tagging. Rule-based taggers Statistical taggers Hybrid approaches

POS Tagging 1. POS Tagging. Rule-based taggers Statistical taggers Hybrid approaches POS Tagging 1 POS Tagging Rule-based taggers Statistical taggers Hybrid approaches POS Tagging 1 POS Tagging 2 Words taken isolatedly are ambiguous regarding its POS Yo bajo con el hombre bajo a PP AQ

More information

A Mixed Trigrams Approach for Context Sensitive Spell Checking

A Mixed Trigrams Approach for Context Sensitive Spell Checking A Mixed Trigrams Approach for Context Sensitive Spell Checking Davide Fossati and Barbara Di Eugenio Department of Computer Science University of Illinois at Chicago Chicago, IL, USA dfossa1@uic.edu, bdieugen@cs.uic.edu

More information

Tagging with Hidden Markov Models

Tagging with Hidden Markov Models Tagging with Hidden Markov Models Michael Collins 1 Tagging Problems In many NLP problems, we would like to model pairs of sequences. Part-of-speech (POS) tagging is perhaps the earliest, and most famous,

More information

Search and Data Mining: Techniques. Text Mining Anya Yarygina Boris Novikov

Search and Data Mining: Techniques. Text Mining Anya Yarygina Boris Novikov Search and Data Mining: Techniques Text Mining Anya Yarygina Boris Novikov Introduction Generally used to denote any system that analyzes large quantities of natural language text and detects lexical or

More information

PoS-tagging Italian texts with CORISTagger

PoS-tagging Italian texts with CORISTagger PoS-tagging Italian texts with CORISTagger Fabio Tamburini DSLO, University of Bologna, Italy fabio.tamburini@unibo.it Abstract. This paper presents an evolution of CORISTagger [1], an high-performance

More information

Context Grammar and POS Tagging

Context Grammar and POS Tagging Context Grammar and POS Tagging Shian-jung Dick Chen Don Loritz New Technology and Research New Technology and Research LexisNexis LexisNexis Ohio, 45342 Ohio, 45342 dick.chen@lexisnexis.com don.loritz@lexisnexis.com

More information

Training and evaluation of POS taggers on the French MULTITAG corpus

Training and evaluation of POS taggers on the French MULTITAG corpus Training and evaluation of POS taggers on the French MULTITAG corpus A. Allauzen, H. Bonneau-Maynard LIMSI/CNRS; Univ Paris-Sud, Orsay, F-91405 {allauzen,maynard}@limsi.fr Abstract The explicit introduction

More information

Brill s rule-based PoS tagger

Brill s rule-based PoS tagger Beáta Megyesi Department of Linguistics University of Stockholm Extract from D-level thesis (section 3) Brill s rule-based PoS tagger Beáta Megyesi Eric Brill introduced a PoS tagger in 1992 that was based

More information

Grammars and introduction to machine learning. Computers Playing Jeopardy! Course Stony Brook University

Grammars and introduction to machine learning. Computers Playing Jeopardy! Course Stony Brook University Grammars and introduction to machine learning Computers Playing Jeopardy! Course Stony Brook University Last class: grammars and parsing in Prolog Noun -> roller Verb thrills VP Verb NP S NP VP NP S VP

More information

Efficient Techniques for Improved Data Classification and POS Tagging by Monitoring Extraction, Pruning and Updating of Unknown Foreign Words

Efficient Techniques for Improved Data Classification and POS Tagging by Monitoring Extraction, Pruning and Updating of Unknown Foreign Words , pp.290-295 http://dx.doi.org/10.14257/astl.2015.111.55 Efficient Techniques for Improved Data Classification and POS Tagging by Monitoring Extraction, Pruning and Updating of Unknown Foreign Words Irfan

More information

Improving Data Driven Part-of-Speech Tagging by Morphologic Knowledge Induction

Improving Data Driven Part-of-Speech Tagging by Morphologic Knowledge Induction Improving Data Driven Part-of-Speech Tagging by Morphologic Knowledge Induction Uwe D. Reichel Department of Phonetics and Speech Communication University of Munich reichelu@phonetik.uni-muenchen.de Abstract

More information

LANGUAGE! 4 th Edition, Levels A C, correlated to the South Carolina College and Career Readiness Standards, Grades 3 5

LANGUAGE! 4 th Edition, Levels A C, correlated to the South Carolina College and Career Readiness Standards, Grades 3 5 Page 1 of 57 Grade 3 Reading Literary Text Principles of Reading (P) Standard 1: Demonstrate understanding of the organization and basic features of print. Standard 2: Demonstrate understanding of spoken

More information

31 Case Studies: Java Natural Language Tools Available on the Web

31 Case Studies: Java Natural Language Tools Available on the Web 31 Case Studies: Java Natural Language Tools Available on the Web Chapter Objectives Chapter Contents This chapter provides a number of sources for open source and free atural language understanding software

More information

Correlation: ELLIS. English language Learning and Instruction System. and the TOEFL. Test Of English as a Foreign Language

Correlation: ELLIS. English language Learning and Instruction System. and the TOEFL. Test Of English as a Foreign Language Correlation: English language Learning and Instruction System and the TOEFL Test Of English as a Foreign Language Structure (Grammar) A major aspect of the ability to succeed on the TOEFL examination is

More information

Improving statistical POS tagging using Linguistic feature for Hindi and Telugu

Improving statistical POS tagging using Linguistic feature for Hindi and Telugu Improving statistical POS tagging using Linguistic feature for Hindi and Telugu by Phani Gadde, Meher Vijay Yeleti in ICON-2008: International Conference on Natural Language Processing (ICON-2008) Report

More information

Albert Pye and Ravensmere Schools Grammar Curriculum

Albert Pye and Ravensmere Schools Grammar Curriculum Albert Pye and Ravensmere Schools Grammar Curriculum Introduction The aim of our schools own grammar curriculum is to ensure that all relevant grammar content is introduced within the primary years in

More information

Symbiosis of Evolutionary Techniques and Statistical Natural Language Processing

Symbiosis of Evolutionary Techniques and Statistical Natural Language Processing 1 Symbiosis of Evolutionary Techniques and Statistical Natural Language Processing Lourdes Araujo Dpto. Sistemas Informáticos y Programación, Univ. Complutense, Madrid 28040, SPAIN (email: lurdes@sip.ucm.es)

More information

Comma checking in Danish Daniel Hardt Copenhagen Business School & Villanova University

Comma checking in Danish Daniel Hardt Copenhagen Business School & Villanova University Comma checking in Danish Daniel Hardt Copenhagen Business School & Villanova University 1. Introduction This paper describes research in using the Brill tagger (Brill 94,95) to learn to identify incorrect

More information

Word Completion and Prediction in Hebrew

Word Completion and Prediction in Hebrew Experiments with Language Models for בס"ד Word Completion and Prediction in Hebrew 1 Yaakov HaCohen-Kerner, Asaf Applebaum, Jacob Bitterman Department of Computer Science Jerusalem College of Technology

More information

Third Grade Language Arts Learning Targets - Common Core

Third Grade Language Arts Learning Targets - Common Core Third Grade Language Arts Learning Targets - Common Core Strand Standard Statement Learning Target Reading: 1 I can ask and answer questions, using the text for support, to show my understanding. RL 1-1

More information

Why language is hard. And what Linguistics has to say about it. Natalia Silveira Participation code: eagles

Why language is hard. And what Linguistics has to say about it. Natalia Silveira Participation code: eagles Why language is hard And what Linguistics has to say about it Natalia Silveira Participation code: eagles Christopher Natalia Silveira Manning Language processing is so easy for humans that it is like

More information

ONLINE ENGLISH LANGUAGE RESOURCES

ONLINE ENGLISH LANGUAGE RESOURCES ONLINE ENGLISH LANGUAGE RESOURCES Developed and updated by C. Samuel for students taking courses at the English and French Language Centre, Faculty of Arts (Links live as at November 2, 2009) Dictionaries

More information

KINDGERGARTEN. Listen to a story for a particular reason

KINDGERGARTEN. Listen to a story for a particular reason KINDGERGARTEN READING FOUNDATIONAL SKILLS Print Concepts Follow words from left to right in a text Follow words from top to bottom in a text Know when to turn the page in a book Show spaces between words

More information

English. Universidad Virtual. Curso de sensibilización a la PAEP (Prueba de Admisión a Estudios de Posgrado) Parts of Speech. Nouns.

English. Universidad Virtual. Curso de sensibilización a la PAEP (Prueba de Admisión a Estudios de Posgrado) Parts of Speech. Nouns. English Parts of speech Parts of Speech There are eight parts of speech. Here are some of their highlights. Nouns Pronouns Adjectives Articles Verbs Adverbs Prepositions Conjunctions Click on any of the

More information

Livingston Public Schools Scope and Sequence K 6 Grammar and Mechanics

Livingston Public Schools Scope and Sequence K 6 Grammar and Mechanics Grade and Unit Timeframe Grammar Mechanics K Unit 1 6 weeks Oral grammar naming words K Unit 2 6 weeks Oral grammar Capitalization of a Name action words K Unit 3 6 weeks Oral grammar sentences Sentence

More information

PTE Academic Preparation Course Outline

PTE Academic Preparation Course Outline PTE Academic Preparation Course Outline August 2011 V2 Pearson Education Ltd 2011. No part of this publication may be reproduced without the prior permission of Pearson Education Ltd. Introduction The

More information

Pupil SPAG Card 1. Terminology for pupils. I Can Date Word

Pupil SPAG Card 1. Terminology for pupils. I Can Date Word Pupil SPAG Card 1 1 I know about regular plural noun endings s or es and what they mean (for example, dog, dogs; wish, wishes) 2 I know the regular endings that can be added to verbs (e.g. helping, helped,

More information

Ling 201 Syntax 1. Jirka Hana April 10, 2006

Ling 201 Syntax 1. Jirka Hana April 10, 2006 Overview of topics What is Syntax? Word Classes What to remember and understand: Ling 201 Syntax 1 Jirka Hana April 10, 2006 Syntax, difference between syntax and semantics, open/closed class words, all

More information

English Appendix 2: Vocabulary, grammar and punctuation

English Appendix 2: Vocabulary, grammar and punctuation English Appendix 2: Vocabulary, grammar and punctuation The grammar of our first language is learnt naturally and implicitly through interactions with other speakers and from reading. Explicit knowledge

More information

Turkish Radiology Dictation System

Turkish Radiology Dictation System Turkish Radiology Dictation System Ebru Arısoy, Levent M. Arslan Boaziçi University, Electrical and Electronic Engineering Department, 34342, Bebek, stanbul, Turkey arisoyeb@boun.edu.tr, arslanle@boun.edu.tr

More information

CS 533: Natural Language. Word Prediction

CS 533: Natural Language. Word Prediction CS 533: Natural Language Processing Lecture 03 N-Gram Models and Algorithms CS 533: Natural Language Processing Lecture 01 1 Word Prediction Suppose you read the following sequence of words: Sue swallowed

More information

Shallow Parsing with PoS Taggers and Linguistic Features

Shallow Parsing with PoS Taggers and Linguistic Features Journal of Machine Learning Research 2 (2002) 639 668 Submitted 9/01; Published 3/02 Shallow Parsing with PoS Taggers and Linguistic Features Beáta Megyesi Centre for Speech Technology (CTT) Department

More information

Reliable and Cost-Effective PoS-Tagging

Reliable and Cost-Effective PoS-Tagging Reliable and Cost-Effective PoS-Tagging Yu-Fang Tsai Keh-Jiann Chen Institute of Information Science, Academia Sinica Nanang, Taipei, Taiwan 5 eddie,chen@iis.sinica.edu.tw Abstract In order to achieve

More information

Speech Recognition on Cell Broadband Engine UCRL-PRES-223890

Speech Recognition on Cell Broadband Engine UCRL-PRES-223890 Speech Recognition on Cell Broadband Engine UCRL-PRES-223890 Yang Liu, Holger Jones, John Johnson, Sheila Vaidya (Lawrence Livermore National Laboratory) Michael Perrone, Borivoj Tydlitat, Ashwini Nanda

More information

NLP Programming Tutorial 5 - Part of Speech Tagging with Hidden Markov Models

NLP Programming Tutorial 5 - Part of Speech Tagging with Hidden Markov Models NLP Programming Tutorial 5 - Part of Speech Tagging with Hidden Markov Models Graham Neubig Nara Institute of Science and Technology (NAIST) 1 Part of Speech (POS) Tagging Given a sentence X, predict its

More information

Building a Question Classifier for a TREC-Style Question Answering System

Building a Question Classifier for a TREC-Style Question Answering System Building a Question Classifier for a TREC-Style Question Answering System Richard May & Ari Steinberg Topic: Question Classification We define Question Classification (QC) here to be the task that, given

More information

Statistical Machine Translation: IBM Models 1 and 2

Statistical Machine Translation: IBM Models 1 and 2 Statistical Machine Translation: IBM Models 1 and 2 Michael Collins 1 Introduction The next few lectures of the course will be focused on machine translation, and in particular on statistical machine translation

More information

Evaluation of Bayesian Spam Filter and SVM Spam Filter

Evaluation of Bayesian Spam Filter and SVM Spam Filter Evaluation of Bayesian Spam Filter and SVM Spam Filter Ayahiko Niimi, Hirofumi Inomata, Masaki Miyamoto and Osamu Konishi School of Systems Information Science, Future University-Hakodate 116 2 Kamedanakano-cho,

More information

Morphology. Morphology is the study of word formation, of the structure of words. 1. some words can be divided into parts which still have meaning

Morphology. Morphology is the study of word formation, of the structure of words. 1. some words can be divided into parts which still have meaning Morphology Morphology is the study of word formation, of the structure of words. Some observations about words and their structure: 1. some words can be divided into parts which still have meaning 2. many

More information

Building A Vocabulary Self-Learning Speech Recognition System

Building A Vocabulary Self-Learning Speech Recognition System INTERSPEECH 2014 Building A Vocabulary Self-Learning Speech Recognition System Long Qin 1, Alexander Rudnicky 2 1 M*Modal, 1710 Murray Ave, Pittsburgh, PA, USA 2 Carnegie Mellon University, 5000 Forbes

More information

Part-of-Speech Tagging for Bengali

Part-of-Speech Tagging for Bengali Part-of-Speech Tagging for Bengali Thesis submitted to Indian Institute of Technology, Kharagpur for the award of the degree of Master of Science by Sandipan Dandapat Under the guidance of Prof. Sudeshna

More information

Strand: Reading Literature Topics Standard I can statements Vocabulary Key Ideas and Details

Strand: Reading Literature Topics Standard I can statements Vocabulary Key Ideas and Details Strand: Reading Literature Key Ideas and Details Craft and Structure RL.3.1 Ask and answer questions to demonstrate understanding of a text, referring explicitly to the text as the basis for the answers.

More information

An online semi automated POS tagger for. Presented By: Pallav Kumar Dutta Indian Institute of Technology Guwahati

An online semi automated POS tagger for. Presented By: Pallav Kumar Dutta Indian Institute of Technology Guwahati An online semi automated POS tagger for Assamese Presented By: Pallav Kumar Dutta Indian Institute of Technology Guwahati Topics Overview, Existing Taggers & Methods Basic requirements of POS Tagger Our

More information

Sentence Blocks. Sentence Focus Activity. Contents

Sentence Blocks. Sentence Focus Activity. Contents Sentence Focus Activity Sentence Blocks Contents Instructions 2.1 Activity Template (Blank) 2.7 Sentence Blocks Q & A 2.8 Sentence Blocks Six Great Tips for Students 2.9 Designed specifically for the Talk

More information

Minnesota K-12 Academic Standards in Language Arts Curriculum and Assessment Alignment Form Rewards Intermediate Grades 4-6

Minnesota K-12 Academic Standards in Language Arts Curriculum and Assessment Alignment Form Rewards Intermediate Grades 4-6 Minnesota K-12 Academic Standards in Language Arts Curriculum and Assessment Alignment Form Rewards Intermediate Grades 4-6 4 I. READING AND LITERATURE A. Word Recognition, Analysis, and Fluency The student

More information

C o p yr i g ht 2015, S A S I nstitute Inc. A l l r i g hts r eser v ed. INTRODUCTION TO SAS TEXT MINER

C o p yr i g ht 2015, S A S I nstitute Inc. A l l r i g hts r eser v ed. INTRODUCTION TO SAS TEXT MINER INTRODUCTION TO SAS TEXT MINER TODAY S AGENDA INTRODUCTION TO SAS TEXT MINER Define data mining Overview of SAS Enterprise Miner Describe text analytics and define text data mining Text Mining Process

More information

According to the Argentine writer Jorge Luis Borges, in the Celestial Emporium of Benevolent Knowledge, animals are divided

According to the Argentine writer Jorge Luis Borges, in the Celestial Emporium of Benevolent Knowledge, animals are divided Categories Categories According to the Argentine writer Jorge Luis Borges, in the Celestial Emporium of Benevolent Knowledge, animals are divided into 1 2 Categories those that belong to the Emperor embalmed

More information

Glossary of literacy terms

Glossary of literacy terms Glossary of literacy terms These terms are used in literacy. You can use them as part of your preparation for the literacy professional skills test. You will not be assessed on definitions of terms during

More information

Sentiment analysis on news articles using Natural Language Processing and Machine Learning Approach.

Sentiment analysis on news articles using Natural Language Processing and Machine Learning Approach. Sentiment analysis on news articles using Natural Language Processing and Machine Learning Approach. Pranali Chilekar 1, Swati Ubale 2, Pragati Sonkambale 3, Reema Panarkar 4, Gopal Upadhye 5 1 2 3 4 5

More information

Speech and Language Processing

Speech and Language Processing Speech and Language Processing An Introduction to Natural Language Processing, Computational Linguistics, and Speech Recognition Second Edition Daniel Jurafsky Stanford University James H. Martin University

More information

10th Grade Language. Goal ISAT% Objective Description (with content limits) Vocabulary Words

10th Grade Language. Goal ISAT% Objective Description (with content limits) Vocabulary Words Standard 3: Writing Process 3.1: Prewrite 58-69% 10.LA.3.1.2 Generate a main idea or thesis appropriate to a type of writing. (753.02.b) Items may include a specified purpose, audience, and writing outline.

More information

English Grammar Checker

English Grammar Checker International l Journal of Computer Sciences and Engineering Open Access Review Paper Volume-4, Issue-3 E-ISSN: 2347-2693 English Grammar Checker Pratik Ghosalkar 1*, Sarvesh Malagi 2, Vatsal Nagda 3,

More information

Hidden Markov Models in Bioinformatics. By Máthé Zoltán Kőrösi Zoltán 2006

Hidden Markov Models in Bioinformatics. By Máthé Zoltán Kőrösi Zoltán 2006 Hidden Markov Models in Bioinformatics By Máthé Zoltán Kőrösi Zoltán 2006 Outline Markov Chain HMM (Hidden Markov Model) Hidden Markov Models in Bioinformatics Gene Finding Gene Finding Model Viterbi algorithm

More information

Year 1 reading expectations (New Curriculum) Year 1 writing expectations (New Curriculum)

Year 1 reading expectations (New Curriculum) Year 1 writing expectations (New Curriculum) Year 1 reading expectations Year 1 writing expectations Responds speedily with the correct sound to graphemes (letters or groups of letters) for all 40+ phonemes, including, where applicable, alternative

More information

Course Syllabus My TOEFL ibt Preparation Course Online sessions: M, W, F 15:00-16:30 PST

Course Syllabus My TOEFL ibt Preparation Course Online sessions: M, W, F 15:00-16:30 PST Course Syllabus My TOEFL ibt Preparation Course Online sessions: M, W, F Instructor Contact Information Office Location Virtual Office Hours Course Announcements Email Technical support Anastasiia V. Mixcoatl-Martinez

More information

Nouns may show possession or ownership. Use an apostrophe with a noun to show something belongs to someone or to something.

Nouns may show possession or ownership. Use an apostrophe with a noun to show something belongs to someone or to something. Nouns Section 1.4 Possessive Nouns Nouns may show possession or ownership. Use an apostrophe with a noun to show something belongs to someone or to something. Jane s dress is red. The table s legs were

More information

SEPTEMBER Unit 1 Page Learning Goals 1 Short a 2 b 3-5 blends 6-7 c as in cat 8-11 t 12-13 p

SEPTEMBER Unit 1 Page Learning Goals 1 Short a 2 b 3-5 blends 6-7 c as in cat 8-11 t 12-13 p The McRuffy Kindergarten Reading/Phonics year- long program is divided into 4 units and 180 pages of reading/phonics instruction. Pages and learning goals covered are noted below: SEPTEMBER Unit 1 1 Short

More information

2016-2017 Curriculum Catalog

2016-2017 Curriculum Catalog 2016-2017 Curriculum Catalog 2016 Glynlyon, Inc. Table of Contents LANGUAGE ARTS 400 COURSE OVERVIEW... 1 UNIT 1: JESUS, OUR EXAMPLE... 3 UNIT 2: WORKING WITH INFORMATION... 3 UNIT 3: THE STORY OF OUR

More information

LINGSTAT: AN INTERACTIVE, MACHINE-AIDED TRANSLATION SYSTEM*

LINGSTAT: AN INTERACTIVE, MACHINE-AIDED TRANSLATION SYSTEM* LINGSTAT: AN INTERACTIVE, MACHINE-AIDED TRANSLATION SYSTEM* Jonathan Yamron, James Baker, Paul Bamberg, Haakon Chevalier, Taiko Dietzel, John Elder, Frank Kampmann, Mark Mandel, Linda Manganaro, Todd Margolis,

More information

UNIVERSITÀ DEGLI STUDI DELL AQUILA CENTRO LINGUISTICO DI ATENEO

UNIVERSITÀ DEGLI STUDI DELL AQUILA CENTRO LINGUISTICO DI ATENEO TESTING DI LINGUA INGLESE: PROGRAMMA DI TUTTI I LIVELLI - a.a. 2010/2011 Collaboratori e Esperti Linguistici di Lingua Inglese: Dott.ssa Fatima Bassi e-mail: fatimacarla.bassi@fastwebnet.it Dott.ssa Liliana

More information

GMAT.cz www.gmat.cz info@gmat.cz. GMAT.cz KET (Key English Test) Preparating Course Syllabus

GMAT.cz www.gmat.cz info@gmat.cz. GMAT.cz KET (Key English Test) Preparating Course Syllabus Lesson Overview of Lesson Plan Numbers 1&2 Introduction to Cambridge KET Handing Over of GMAT.cz KET General Preparation Package Introduce Methodology for Vocabulary Log Introduce Methodology for Grammar

More information

Applying Co-Training Methods to Statistical Parsing. Anoop Sarkar http://www.cis.upenn.edu/ anoop/ anoop@linc.cis.upenn.edu

Applying Co-Training Methods to Statistical Parsing. Anoop Sarkar http://www.cis.upenn.edu/ anoop/ anoop@linc.cis.upenn.edu Applying Co-Training Methods to Statistical Parsing Anoop Sarkar http://www.cis.upenn.edu/ anoop/ anoop@linc.cis.upenn.edu 1 Statistical Parsing: the company s clinical trials of both its animal and human-based

More information

Strand: Reading Literature Topics Standard I can statements Vocabulary Key Ideas and Details

Strand: Reading Literature Topics Standard I can statements Vocabulary Key Ideas and Details Strand: Reading Literature Topics Standard I can statements Vocabulary Key Ideas and Details Craft and Structure RL.5.1 Quote accurately from a text when explaining what the text says explicitly and when

More information

Index. 344 Grammar and Language Workbook, Grade 8

Index. 344 Grammar and Language Workbook, Grade 8 Index Index 343 Index A A, an (usage), 8, 123 A, an, the (articles), 8, 123 diagraming, 205 Abbreviations, correct use of, 18 19, 273 Abstract nouns, defined, 4, 63 Accept, except, 12, 227 Action verbs,

More information

CINTIL-PropBank. CINTIL-PropBank Sub-corpus id Sentences Tokens Domain Sentences for regression atsts 779 5,654 Test

CINTIL-PropBank. CINTIL-PropBank Sub-corpus id Sentences Tokens Domain Sentences for regression atsts 779 5,654 Test CINTIL-PropBank I. Basic Information 1.1. Corpus information The CINTIL-PropBank (Branco et al., 2012) is a set of sentences annotated with their constituency structure and semantic role tags, composed

More information

Learning Morphological Disambiguation Rules for Turkish

Learning Morphological Disambiguation Rules for Turkish Learning Morphological Disambiguation Rules for Turkish Deniz Yuret Dept. of Computer Engineering Koç University İstanbul, Turkey dyuret@ku.edu.tr Ferhan Türe Dept. of Computer Engineering Koç University

More information

Coupling an annotated corpus and a morphosyntactic lexicon for state-of-the-art POS tagging with less human effort

Coupling an annotated corpus and a morphosyntactic lexicon for state-of-the-art POS tagging with less human effort Coupling an annotated corpus and a morphosyntactic lexicon for state-of-the-art POS tagging with less human effort Pascal Denis and Benoît Sagot Equipe-project ALPAGE INRIA and Université Paris 7 30, rue

More information

The Basics of Graphical Models

The Basics of Graphical Models The Basics of Graphical Models David M. Blei Columbia University October 3, 2015 Introduction These notes follow Chapter 2 of An Introduction to Probabilistic Graphical Models by Michael Jordan. Many figures

More information

Unit: Fever, Fire and Fashion Term: Spring 1 Year: 5

Unit: Fever, Fire and Fashion Term: Spring 1 Year: 5 Unit: Fever, Fire and Fashion Term: Spring 1 Year: 5 English Fever, Fire and Fashion Unit Summary In this historical Unit pupils learn about everyday life in London during the 17 th Century. Frost fairs,

More information

Customizing an English-Korean Machine Translation System for Patent Translation *

Customizing an English-Korean Machine Translation System for Patent Translation * Customizing an English-Korean Machine Translation System for Patent Translation * Sung-Kwon Choi, Young-Gil Kim Natural Language Processing Team, Electronics and Telecommunications Research Institute,

More information

Categorizing and Tagging Words

Categorizing and Tagging Words Chapter 4 Categorizing and Tagging Words 4.1 Introduction In the last chapter we dealt with words in their own right. We saw that some distinctions can be collapsed using normalization, but we did not

More information

VCU-TSA at Semeval-2016 Task 4: Sentiment Analysis in Twitter

VCU-TSA at Semeval-2016 Task 4: Sentiment Analysis in Twitter VCU-TSA at Semeval-2016 Task 4: Sentiment Analysis in Twitter Gerard Briones and Kasun Amarasinghe and Bridget T. McInnes, PhD. Department of Computer Science Virginia Commonwealth University Richmond,

More information

Hybrid algorithms for preprocessing agglutinative languages and less-resourced domains effectively

Hybrid algorithms for preprocessing agglutinative languages and less-resourced domains effectively Hybrid algorithms for preprocessing agglutinative languages and less-resourced domains effectively Doctor of Philosophy Dissertation György Orosz Roska Tamás Doctoral School of Sciences and Technology

More information

Stock Market Prediction Using Data Mining

Stock Market Prediction Using Data Mining Stock Market Prediction Using Data Mining 1 Ruchi Desai, 2 Prof.Snehal Gandhi 1 M.E., 2 M.Tech. 1 Computer Department 1 Sarvajanik College of Engineering and Technology, Surat, Gujarat, India Abstract

More information

Points of Interference in Learning English as a Second Language

Points of Interference in Learning English as a Second Language Points of Interference in Learning English as a Second Language Tone Spanish: In both English and Spanish there are four tone levels, but Spanish speaker use only the three lower pitch tones, except when

More information

Towards a RB-SMT Hybrid System for Translating Patent Claims Results and Perspectives

Towards a RB-SMT Hybrid System for Translating Patent Claims Results and Perspectives Towards a RB-SMT Hybrid System for Translating Patent Claims Results and Perspectives Ramona Enache and Adam Slaski Department of Computer Science and Engineering Chalmers University of Technology and

More information

Strand: Reading Literature Topics Standard I can statements Vocabulary Key Ideas and Details

Strand: Reading Literature Topics Standard I can statements Vocabulary Key Ideas and Details Strand: Reading Literature Key Ideas and Craft and Structure Integration of Knowledge and Ideas RL.K.1. With prompting and support, ask and answer questions about key details in a text RL.K.2. With prompting

More information

Collecting Polish German Parallel Corpora in the Internet

Collecting Polish German Parallel Corpora in the Internet Proceedings of the International Multiconference on ISSN 1896 7094 Computer Science and Information Technology, pp. 285 292 2007 PIPS Collecting Polish German Parallel Corpora in the Internet Monika Rosińska

More information

Experiments in Cross-Language Morphological Annotation Transfer

Experiments in Cross-Language Morphological Annotation Transfer Experiments in Cross-Language Morphological Annotation Transfer Anna Feldman, Jirka Hana, and Chris Brew Ohio State University, Department of Linguistics, Columbus, OH 43210-1298, USA Abstract. Annotated

More information

An Approach to Handle Idioms and Phrasal Verbs in English-Tamil Machine Translation System

An Approach to Handle Idioms and Phrasal Verbs in English-Tamil Machine Translation System An Approach to Handle Idioms and Phrasal Verbs in English-Tamil Machine Translation System Thiruumeni P G, Anand Kumar M Computational Engineering & Networking, Amrita Vishwa Vidyapeetham, Coimbatore,

More information

Learning Translation Rules from Bilingual English Filipino Corpus

Learning Translation Rules from Bilingual English Filipino Corpus Proceedings of PACLIC 19, the 19 th Asia-Pacific Conference on Language, Information and Computation. Learning Translation s from Bilingual English Filipino Corpus Michelle Wendy Tan, Raymond Joseph Ang,

More information

Introduction to Algorithmic Trading Strategies Lecture 2

Introduction to Algorithmic Trading Strategies Lecture 2 Introduction to Algorithmic Trading Strategies Lecture 2 Hidden Markov Trading Model Haksun Li haksun.li@numericalmethod.com www.numericalmethod.com Outline Carry trade Momentum Valuation CAPM Markov chain

More information

EAP 1161 1660 Grammar Competencies Levels 1 6

EAP 1161 1660 Grammar Competencies Levels 1 6 EAP 1161 1660 Grammar Competencies Levels 1 6 Grammar Committee Representatives: Marcia Captan, Maria Fallon, Ira Fernandez, Myra Redman, Geraldine Walker Developmental Editor: Cynthia M. Schuemann Approved:

More information

The XMU Phrase-Based Statistical Machine Translation System for IWSLT 2006

The XMU Phrase-Based Statistical Machine Translation System for IWSLT 2006 The XMU Phrase-Based Statistical Machine Translation System for IWSLT 2006 Yidong Chen, Xiaodong Shi Institute of Artificial Intelligence Xiamen University P. R. China November 28, 2006 - Kyoto 13:46 1

More information

Mining Online Diaries for Blogger Identification

Mining Online Diaries for Blogger Identification Mining Online Diaries for Blogger Identification Haytham Mohtasseb Department of Computing and Informatics University of Lincoln Lincoln, UK hmohtasseb@lincoln.ac.uk Amr Ahmed Department of Computing and

More information

Ngram Search Engine with Patterns Combining Token, POS, Chunk and NE Information

Ngram Search Engine with Patterns Combining Token, POS, Chunk and NE Information Ngram Search Engine with Patterns Combining Token, POS, Chunk and NE Information Satoshi Sekine Computer Science Department New York University sekine@cs.nyu.edu Kapil Dalwani Computer Science Department

More information

openmind 1 Practice Online

openmind 1 Practice Online Macmillan Practice Online is the easy way to get all the benefits of online learning and with over 100 courses to choose from, covering all competence levels and ranging from business English to exam practice

More information

Modern Natural Language Interfaces to Databases: Composing Statistical Parsing with Semantic Tractability

Modern Natural Language Interfaces to Databases: Composing Statistical Parsing with Semantic Tractability Modern Natural Language Interfaces to Databases: Composing Statistical Parsing with Semantic Tractability Ana-Maria Popescu Alex Armanasu Oren Etzioni University of Washington David Ko {amp, alexarm, etzioni,

More information

A Primer on Text Analytics

A Primer on Text Analytics A Primer on Text Analytics From the Perspective of a Statistics Graduate Student Bradley Turnbull Statistical Learning Group North Carolina State University March 27, 2015 Bradley Turnbull Text Analytics

More information

Cambridge English: First (FCE) Frequently Asked Questions (FAQs)

Cambridge English: First (FCE) Frequently Asked Questions (FAQs) Cambridge English: First (FCE) Frequently Asked Questions (FAQs) Is there a wordlist for Cambridge English: First exams? No. Examinations that are at CEFR Level B2 (independent user), or above such as

More information

Morphemes, roots and affixes. 28 October 2011

Morphemes, roots and affixes. 28 October 2011 Morphemes, roots and affixes 28 October 2011 Previously said We think of words as being the most basic, the most fundamental, units through which meaning is represented in language. Words are the smallest

More information

Non-exam Assessment Tasks

Non-exam Assessment Tasks SPECIMEN MATERIAL ENTRY LEVEL CERTIFICATE STEP UP TO ENGLISH Silver Step 5972/1 Component 1 Literacy Topics Planning the Prom Non-exam Assessment Task and Teachers Notes Specimen 2015 Time allowed: 1 hour

More information

The Design of a Proofreading Software Service

The Design of a Proofreading Software Service The Design of a Proofreading Software Service Raphael Mudge Automattic Washington, DC 20036 raffi@automattic.com Abstract Web applications have the opportunity to check spelling, style, and grammar using

More information

CST and CAHSEE Academic Vocabulary

CST and CAHSEE Academic Vocabulary CST and CAHSEE Academic Vocabulary Grades K 12 Math and ELA This document references Academic Language used in the Released Test Questions from the 2008 posted CAHSEE Released Test Questions (RTQs) and

More information

Texas Success Initiative (TSI) Assessment

Texas Success Initiative (TSI) Assessment Texas Success Initiative (TSI) Assessment Interpreting Your Score 1 Congratulations on taking the TSI Assessment! The TSI Assessment measures your strengths and weaknesses in mathematics and statistics,

More information

MODULE 15 Diagram the organizational structure of your company.

MODULE 15 Diagram the organizational structure of your company. Student name: Date: MODULE 15 Diagram the organizational structure of your company. Objectives: A. Diagram the organizational chart for your place of business. B. Determine the importance of organization

More information

EAS Basic Outline. Overview

EAS Basic Outline. Overview EAS Basic Outline Overview This is the course outline for your English Language Basic Course. This course is delivered at pre intermediate level of English, and the course book that you will be using is

More information