Predicting lexical proficiency in language learner texts using computational indices
|
|
|
- Arron Hicks
- 9 years ago
- Views:
Transcription
1 Predicting lexical proficiency in language learner texts using computational indices Language Testing XX(X) 1 20 The Author(s) 2010 Reprints and permission: sagepub.co.uk/journalspermissions.nav DOI: / Scott A. Crossley Georgia State University, USA Tom Salsbury Washington State University, USA Danielle S. McNamara University of Memphis, USA Scott Jarvis Ohio University, USA Abstract The authors present a model of lexical proficiency based on lexical indices related to vocabulary size, depth of lexical knowledge, and accessibility to core lexical items. The lexical indices used in this study come from the computational tool Coh-Metrix and include word length scores, lexical diversity values, word frequency counts, hypernymy values, polysemy values, semantic coreferentiality, word meaningfulness, word concreteness, word imagability, and word familiarity. Human raters evaluated a corpus of 240 written texts using a standardized rubric of lexical proficiency. To ensure a variety of text levels, the corpus comprised 60 texts each from beginning, intermediate, and advanced second language (L2) adult English learners. The L2 texts were collected longitudinally from 10 English learners. In addition, 60 texts from native English speakers were collected. The holistic scores from the trained human raters were then correlated to a variety of lexical indices. The researchers found that lexical diversity, word hypernymy values and content word frequency explain 44% of the variance of the human evaluations of lexical proficiency in the examined writing samples. The findings represent an important step in the development of a model of lexical proficiency that incorporates both vocabulary size and depth of lexical knowledge features. Keywords computational linguistics, corpus linguistics, depth of lexical knowledge, hypernymy, lexical diversity, lexical frequency, lexical proficiency, vocabulary size Corresponding author: Scott A. Crossley, 34 Peachtree St. Suite 1200, One Park Tower Building, Atlanta, GA [email protected]
2 2 Language Testing XX(X) Introduction Although a clear definition of lexical proficiency is lacking, most research in second language (L2) studies indicates that lexical proficiency refers to vocabulary size, depth of vocabulary knowledge and the accessibility of core lexical items (Meara, 2005). The lexical proficiency of speakers, especially L2 learners, is of crucial interest to language acquisition and linguistic competence for three main reasons. First, misinterpretations of lexical items produced by learners are key elements in communication errors (Ellis, 1995); second, lexical proficiency strongly correlates with academic achievement (Daller et al., 2003); and third, understanding lexical acquisition in relation to its deeper, cognitive functions can lead to increased awareness of how learners process and produce language (Crossley et al., 2009; Crossley et al., 2010b; Salsbury et al., in press). Measuring lexical proficiency benefits from using assessments that evaluate more than just simple lexical knowledge, such as the number or variety of words used, the frequency of words used, or the ability to match dictionary definitions with strings of letters (Nation, 2005). Nevertheless, many studies of lexical proficiency have relied on these types of performance assessments (Polio, 2001), which provide only a partial and relatively surface level representation of lexical knowledge. A fuller understanding of lexical proficiency, along with a practical, computational model with which to assess lexical proficiency, would prove beneficial to language researchers, language learners, language teachers, and educational institutions. The purpose of this study is to examine the potential for lexical indices to predict human evaluations of lexical proficiency based on lexical features related to vocabulary size, depth of knowledge, and access to core lexical items. Vocabulary size relates to how many words a learner knows (e.g. lexical features such as diversity or vocabulary size). Depth of knowledge features, in contrast, relate to how well a word is known (e.g. lexical features such as semantic relatedness, word sense relations, and word associations). Finally, access to core lexical items relates to how quickly words can be retrieved or processed (e.g. lexical features such as word concreteness and familiarity). Our study focuses on one simple question: what lexical indices are predictive of human ratings of lexical proficiency? To address this question, we test specific computerized indices against holistic human judgments of lexical proficiency in written texts. Our aim is to provide a formal model that helps to explain human judgments of lexical proficiency. We are better able to address this question than our predecessors because developments in computational algorithms and models related to lexical processing now allow for the quick and accurate assessment of a variety of lexical features. These features can be measured using automated lexicons, pattern classifiers, part-of-speech taggers, syntactic parsers, shallow semantic interpreters, psycholinguistic word measurement indices, and other lexical components that have been developed in the fields of cognitive science, natural language processing, and computational linguistics (Jurafsky and Martin, 2008). These computational measures have the capability not only to measure vocabulary size features such as lexical diversity, but also depth of knowledge features such as semantic relatedness, word sense relations, and word associations, and access to core lexical items through psycholinguistic word properties. However, the examination of lexical knowledge, lexical output, and lexical development using these computational indices is still in
3 Crossley et al. 3 its infancy. Orchestrating these measures into a practical model from which to measure lexical proficiency has not occurred. L2 lexical assessment Most L2 assessment tests measure receptive skills (Pearson et al., 2007). Those tests that examine productive lexical knowledge have generally focused on breadth and depth of knowledge features. For instance, Lexical Frequency Profiles (LFP; Nation and Heatley, 1996) and P-Lex (Meara and Bell, 2001) both assess L2 lexical development using indices of lexical frequency (i.e. breadth of knowledge). Laufer and Nation (1995) argued that LFP could discriminate between L2 proficiency levels and that it correlated with independent measures of vocabulary knowledge. However, Meara (2005) criticized LFP for being less predictive, especially for shorter texts, than Laufer and Nation claimed. From a depth of knowledge perspective, Meara s (2005) automated tool, V_Links, tests a learner s vocabulary organization using a small number of words. Survey instruments are also common in assessing L2 learners depth of knowledge. Many of these instruments ask participants to generate word derivatives (Schmitt and Meara, 1997) or identify the word associates of a stimulus word (Read, 1993). One problem with survey instruments such as these is that testing depth of knowledge requires the extensive testing of individual words. Testing individual words is not only extremely time consuming, but also does not allow for the testing of a representative sample of words (Meara and Wolter, 2004). For instance, if a language learner was tested on only 50 words and information was extracted on 10 features of those words, a survey would need to be designed that tested 500 items. Measures of lexical proficiency Contemporary notions of lexical proficiency are premised not on the number of words a learner knows, but also on psycholinguistic models of word knowledge and lexical networks. Psycholinguistic word information relates to linguistic properties of words that affect word processing and learnability such as word concreteness, imagability, and familiarity. Lexical networks are the result of connections between conceptual levels, sense relations, semantic co-referentiality, and word associations (Crossley et al., 2009, 2010a, 2010b; Haastrup and Henriksen, 2000). In lexical networks, words interrelate with other words to form clusters of words that act categorically. These clusters connect to other clusters and other words, until entire lexicons are developed based on interconnections (Haastrup and Henriksen, 2000). Common lexical relationships that help form lexical networks include heirarchical relations such as hypernymy, sense relations such as those found in polysemy, semantic similarity connections, and word frequency effects. Measures related to lexical network models and psycholinguistic properties along with frequency and lexical diversity are discussed in detail below. Lexical diversity. Historically, indices related to lexical diversity concentrated on type-token ratios (TTR), which are simple formulas that divide the number of different words (types) by the total number of words (tokens) in a given text. There are various transformations of
4 4 Language Testing XX(X) simple TTR measures such as Corrected TTR (Carrol, 1964), Log TTR (Herdan, 1960), D (Malvern et al., 2004), Advanced TTR (Daller et al., 2003) Guiraud Advanced (Daller et al., 2003), and the Measure of Textual Lexical Diversity (MTLD; McCarthy and Jarvis, in press). The premise behind lexical diversity indices is that more diverse vocabularies are indicative of more proficient and larger lexicons. Although lexical diversity indices are broad, versatile measures of lexical proficiency, they fail to account for important qualitative information about words, namely word difficulty (Vermeer, 2000; Daller et al., 2003). Lexical frequency. Lexical frequency is often equated with lexical richness. Most frequency indices depend on frequency lists and are based on the hypothesis that a higher lexical proficiency results in the use of less frequent words (Meara and Bell, 2001). Word frequency has traditionally been assigned to the breadth of knowledge category, but this categorization is debatable. Ellis (2002), for instance, argues that the production and comprehension of words is a function of their frequency of occurrence in language. Under this approach, word frequency helps determine lexical acquisition because each repetition of a word strengthens the connections between the word and its meaning categorization. As learners are exposed to frequent words, there is a reduction in processing time because the practice time with the word increases. Such a model of lexical acquisition is supported by studies that demonstrate that high frequency words are named more quickly than low frequency words (Balota and Chumbly, 1984), are processed more quickly in reading tasks (Kirsner, 1994), are judged more quickly to be words in lexical decision tasks (Forster, 1976), and have faster response latencies (Glanzer and Ehrenreich, 1979). Most studies concentrating on L2 reading and writing proficiency have found that beginning L2 learners are more likely to comprehend, process, and produce higher frequency words (Crossley and Salsbury, 2010; Ellis, 2002) and that beginning L2 learners use higher frequency words than advanced learners (Bell, 2003; Meara and Bell, 2001). However, recent studies into L2 lexical development that have focused on spoken data have demonstrated that beginning L2 learners might first produce words that are more concrete even if they are less frequent. Only later do learners appear to produce words that are more frequent. This is likely the result of frequent words having more senses and thus being more ambiguous in potential meaning (Crossley et al., 2010b). Word meaningfulness. Word meaningfulness relates to how many associations a word has with other words (Toglia and Battig, 1978). Words with high meaningfulness include words like food, music, and people while words with low meaningfulness include acumen, cowl, and oblique. Words in the first list invoke multiple word associations, while those in the second list have fewer associations. It is argued that L2 learners have fewer and more varied word associations than native speakers (Schmitt and Meara, 1997; Zareva, 2007) because L2 learners have not developed close semantic links between words (Ellis and Beaton, 1993). Semantic links between words mediate the organization and storage of words. In addition, words with more associations are argued to be acquired first (Ellis and Beaton, 1993).
5 Crossley et al. 5 The link between word associations and lexical acquisition is supported in several recent studies. For instance, Zareva (2007) found that higher proficiency learners provide significantly more word associations than intermediate and beginning level learners. She argued that larger vocabularies allow for a greater number of word associations. A second, recent study (Salsbury et al., in press) explored word meaningfulness values provided by the MRC psycholinguistic database and compared the changes in the values as a function of time spent studying English. Salsbury et al. found that meaningfulness scores decreased as time spent studying English increased. The study suggests that higher proficiency learners use more difficult words that have fewer word associations. This finding was supported in Crossley and Salsbury (2010), who used computational indices to predict produced and not produced verbs and nouns in beginning L2 data. The results of this study demonstrated that word meaningfulness was a significant predictor of early verb and noun production in L2 learners. Overall, these studies support the notion that as a language learner s network of word associations develops, a greater number of word associations become available to the learner. The strength of these associations allows learners to extend their lexical production to include peripheral words that have fewer associations. Word concreteness. Word concreteness refers to here-and-now concepts, ideas, and things constituting core lexical knowledge (Toglia and Battig, 1978). On a continuum, very concrete words would be placed on one end and very abstract words on the opposite end. The concreteness of a word has implications for a word s learnability because concrete words, as compared to abstract words, have advantages in tasks involving recall, word recognition, lexical decision, pronunciation, and comprehension (Gee et al., 1999). Regarding L2 lexical acquisition, studies have demonstrated that concrete words are learned earlier (Crossley et al., 2009; Salsbury et al., in press) and more easily than abstract words (Carter, 1987). Word familiarity. Word familiarity is related to word frequency (Schmitt and Meara, 1997) in that words that are rated as more familiar tend to occur more frequently in text and discourse. However, familiar words are salient whereas frequent words may not be (e.g. determiners or prepositions). Few studies have investigated word familiarity directly. A recent study (Salsbury et al., in press) investigated the development of familiar words in L2 learners and found that L2 learners productive vocabulary did not significantly change in terms of word familiarity. Nonetheless, word familiarity remains an important feature of lexical proficiency to investigate because familiar words are likely core lexical items that act as anchors from which to extend word meanings and associations. Word imagability. Imagability scores are important because a word or concept that triggers a mental image quickly and easily is more likely to be recalled. Thus, highly imagable words likely constitute core lexical items in a learner s lexicon. Ellis and Beaton (1993) found that highly imagable words were good candidates for keyword techniques in second language vocabulary learning. Additionally, highly imagable words have more context availability (Schwanenflugel, 1991) because they are experienced and analyzed
6 6 Language Testing XX(X) visually (Ellis and Beaton, 1993). Context availability and visual analysis facilitate the learning of words for L2 learners. Hypernymy. Hypernymy is based on the connection between general and specific lexical items (Chaffin and Glass, 1990). Hypernymic relations are hierarchical associations between hypernyms (superordinate words) and hyponyms (subordinate words). A hypernym is a word that is more general than a related word (car as compared to convertible) and a hyponym is more specific than a related word (convertible as compared to car). Hypernymy is an important organizational system for lexical relations in network models (Haastrup and Henriksen, 2000; Crossley et al., 2009) because it allows for hierarchical categorizations that explain properties that hyponyms inherit from hypernyms. Thus, hypernymy affords the economical representation of lexical properties (Chaffin and Glass, 1990; Murphy, 2004) and lexical generalization (Murphy, 2004). From a developmental perspective, hypernymic relations are more likely acquired as the learners advance cognitively (Anglin, 1993), as they increase their levels of education (Snow, 1990), and as they acquire more specific lexical knowledge (Wolter, 2001). L2 studies have demonstrated that learners use more words of general than of specific meanings (Levenston and Blum, 1977) and that L2 learners use of general terms often results in inappropriate overgeneralizations (Ijaz, 1986). More recent studies have demonstrated that hypernymic relations by L2 learners increase (become less specific) as time is spent studying English (Crossley et al., 2009). Hypernymy is also important in early L2 verb production (Crossley and Salsbury, 2010). Polysemy. Polysemous words are words that have more than one related sense. For instance, the word class has at least six related senses, including socio-economic class, a body of students, a course of study, a collection of things sharing similar attributes, a league ranked by quality (usually sports related), and elegance in dress or behavior. Polysemous relations are connected to conceptual organization because words with multiple overlapping senses are thought to be part of the same conceptual structure (Murphy, 2004). Under a network approach, the multiple senses in a polysemous word are located in a single lexical entry. Such an approach suggests that separate entries for related word senses are uneconomical because they fail to capture the sense connections in the word s uses. However, word senses located within a single lexical item allow learners to efficiently recognize meaning relationships between a word s senses (Verspoor and Lowie, 2003). Studies concerning the polysemy knowledge of L2 learners have found that word sense knowledge increases as L2 learners gain proficiency and that word senses are retained more than forgotten (Schmitt, 1998). Crossley et al. (2010b) also found that as learners language proficiency grew, their production of multiple word senses increased. Both studies provide evidence for the development of word senses in L2 learners. In summary, past studies have demonstrated that lexical features related to vocabulary size, depth of knowledge, and access to core lexical items are important indicators of lexical growth and proficiency in L2 learners. Many of these features are important in supporting lexical network theories (e.g. hypernymy, word meaningfulness, and polysemy)
7 Crossley et al. 7 or measure psycholinguistic word properties of language (e.g. concreteness, imagability, and familiarity). Method Our goal is to determine the degree to which automated lexical indices related to vocabulary size, depth of knowledge, and access to core lexical items can predict human ratings of lexical proficiency. To accomplish this, we analyzed a corpus of scored writing samples using lexical indices taken from the computational tool Coh-Metrix (Graesser et al., 2004). To ensure a wide variety of lexical proficiency, we collected writing samples from ten L2 learners at the beginning, intermediate, and advanced levels, as classified through Institutional Test of English as a Foreign Language (TOEFL) scores. We also collected writing samples from a comparison corpus of native speakers. Trained raters scored the writing samples using a holistic lexical proficiency rubric. We first divided the scored writing samples into training and test sets (Whitten and Frank, 2005). To examine which lexical variables were most predictive of lexical proficiency, we conducted correlations and a linear regression model comparing the human lexical proficiency ratings and the Coh-Metrix lexical indices using the training set only. The results of this analysis were later extended using the regression model to the held back, independent test set data, and finally to the complete corpus. Corpus collection Writing samples were collected from participants involved in a longitudinal study in an intensive language program at a large university in the United States. The writing samples were unstructured and unprepared daily written journals (i.e., freewrites) that were completed as part of the learners regular intensive English coursework. The teachers did not edit or grade the journal entries, but they did comment on the content of the entries. These comments modeled for the learners correct word forms and specific vocabulary in the context of the students journal entry. The learners typically responded in writing to their teachers questions and comments in subsequent journal entries. All original texts were handwritten by the participants and later entered electronically by the researchers. The participants ranged in age from 18 to 27 years old and came from a variety of L1 backgrounds (Korean, Japanese, Arabic, French, Bambara, Portuguese, Spanish, and Turkish). Most of the participants began their intensive English studies in the first level of the 6-level program as determined by an in-house placement exam. The students also completed an institutional TOEFL exam every two months. One hundred and eighty L2 writing samples were selected from a corpus of written data collected from 10 L2 learners over a one-year period. The number of texts collected from each learner was not equal because all learners did not submit the same number of writing samples. All texts were corrected for spelling in order for the lexical items to be analyzed by Coh-Metrix. We did not correct lexical mistakes, though, because we are interested in lexical production, not accuracy. To ensure appropriate lexical proficiency variability across the writing samples, Institutional TOEFL scores were used to classify the L2 writers samples into beginning,
8 8 Language Testing XX(X) intermediate, and advanced categories. Although classifications for level are not available for Institutional TOEFL scores, these scores have been compared to other proficiency measures that do include level classification. These include American Council on Teaching Foreign Languages (ACTFL) ratings (Boldt et al., 1992) and the TOEIC (Educational Testing Service). The classification criteria used in these proficiency measures were linked to the participants Institutional TOEFL scores to categorize the learners. Institutional TOEFL scores below 400 were categorized as beginning level; scores between 400 and 499 were categorized as intermediate level, and scores at 500 or above were categorized as advanced level. In total, 60 writing samples were selected for the three proficiency levels (beginning, intermediate, and advanced). However, not all students were represented at each level because not all learners entered the program at the beginning level and not all students progressed to the advanced level. A comparison corpus of 60 native speaker free-writings was selected from the Stream of Consciousness Data Set from the Pennebaker Archive Project (Newman et al., 2008). The stream of consciousness corpus includes over 2300 free writing samples collected from freshman psychology students over a three-year period. Like the L2 writing samples, the free-writing samples were unstructured and unprepared. All writing samples collected from the native speaker corpus were classified under the category native speaker. Thus, our writing sample corpus contained 60 writing samples from each proficiency level as well as 60 writing samples from the native speakers for a total of 240 sample texts. Writing samples were separated at the paragraph level. The samples were controlled for text length by randomly selecting a text segment from each sample that was about 140 words (depending on paragraph constraints; see Table 1 for more details). To ensure that text length differences did not exist across the levels, an Analysis of Variance (ANOVA) test was conducted. The ANOVA demonstrated no significant differences in text length between levels. Survey instrument The survey instrument used in this study was a holistic grading rubric that we adapted from the American Council on the Teaching of Foreign Languages (ACTFL) proficiency guidelines for speaking and writing (ACTFL Inc., 1999) and holistic writing proficiency rubrics produced by American College Testing (ACT) and the College Board (for use in SAT writing evaluation). The proficiency levels in these rubrics along with the evaluative sections related to lexical proficiency were combined to form an Table 1. Mean and standard deviations for text length Learner type Mean SD Beginner Intermediate Advanced Native
9 Crossley et al. 9 outline from which to develop a holistic scoring metric to evaluate overall lexical proficiency. Five experts in language processing with Ph.D.s in either linguistics or cognitive psychology reviewed the rubric outline and made suggestions for the outline s revision. Once these suggestions were implemented, the revised outline was then subjected to two usability tests: first by experts in the field of language processing and second by human raters (native speakers of English). The usability tests led to additional revisions and the final holistic rubric that we use in this study. The final rubric defined lexical proficiency as skillful language use that is accurate and fluent and is characterized by the appropriate use of conceptual categories, coherence, and lexicalsemantic connections (see Appendix 1 for rubric). Human ratings To assess the 240 writing samples that comprise our written corpus, three native speakers of English were trained as expert raters. The three raters were all graduate students in an English Department at a large university in the United States. The raters were trained on an initial selection of 20 writing samples taken from a training corpus not included in the written corpus used in the study. The raters assigned each writing sample a score between 1 (minimum) and 5 (maximum). To assess inter-rater reliability, Pearson correlations were conducted between all possible pairs of rater responses. The resulting three correlations were averaged to provide a mean correlation between the raters. This correlation was then weighted based on the number of raters (Hatch and Lazaraton, 1991). If the correlations within the three raters did not exceed.70 (p <.001), the raters were retrained on a new set of 20 writing samples. Using this approach, the raters needed to be retrained once. After the second training, the average correlation between the three raters was r =.796 (p <.001) with a weighted correlation of r =.921. Variable selection All the lexical indices used in this study were provided by Coh-Metrix (Graesser et al., 2004). Coh-Metrix is a computational tool that reports on over 600 linguistic indices related to conceptual knowledge, cohesion, lexical difficulty, syntactic complexity, and simple incidence scores (see Appendix 2 for a screenshot of Coh-Metrix output). Many of the indices reported by Coh-Metrix are related to lexical proficiency. These indices can be separated into vocabulary size, depth of knowledge, and access to lexical core measures. The measures related to vocabulary size include indices of lexical diversity as reported by MTLD (McCarthy and Jarvis, in press) and D (Malvern et al., 2004). The measures related to depth of knowledge include indices for hypernymy, polysemy (as reported by WordNet; Fellbaum, 1998), semantic co-referentiality (as reported by Latent Semantic Analysis; Landauer et al., 2007), and word frequency (as reported by the CELEX database; Baayen et al., 1995). Those measures related to accessing core lexical units include word concreteness, word familiarity, word imagability, and word meaningfulness (as reported by the MRC Psycholinguistic Database; Wilson, 1988). We also include indices related to word length for comparison purposes. This provided us with a total of 10 measures. For information about the computation of these measures in
10 10 Language Testing XX(X) Coh-Metrix and their relation to theories of lexical proficiency, we refer the reader to Crossley and McNamara (2009), Crossley and Salsbury (2010), Crossley et al. (2009), Crossley et al. (2010a), Crossley et al. (2010b), and Salsbury et al. (in press). To test our results on an independent corpus that was not used in the initial analysis, we used training and test sets. We divided the corpus into two sets: a training set (n = 180) and a testing set (n = 60) based on a 67/33 split. The purpose of the training set was to identify which of the Coh-Metrix variables best correlated with the human scores assigned to each freewrite. These variables were later used to predict the human scores in the training set using a linear regression model. Later, the freewrites in the test set were analyzed using the regression model from the training set to calculate the predictability of the variables in an independent corpus (Whitten and Frank, 2005). In order to allow for a more reliable interpretation of the multiple regression, we ensured that there were at least 20 times more cases (texts) than variables (the lexical indices). A 20 to 1 ratio allows for the interpretation of each variable s individual contribution to the regression model (Barcikowski and Stevens, 1975; Field, 2005). We used Pearson correlations to select a variable from each measure to be used in a multiple regression. Only those variables that demonstrated significant correlations with the human ratings were selected. With 160 ratings in the training set, this allowed us to choose 8 lexical variables out of the 10 selected. To check for multicollinearity, we conducted additional, correlation tests on the selected variables. If the variables did not exhibit collinearity (r <.70), they were then used in the multiple regression analysis. The selected measures and their respective indices are discussed below in reference to their importance in lexical proficiency. Results Pearson correlations training set We selected the Coh-Metrix measures that demonstrated the highest Pearson correlation when compared to the human ratings of the written samples. The 10 selected variables and their measures along with their r values and p values are presented in Table 2, sorted by the strength of the correlation. No indices from either the polysemy measure or the word length measure demonstrated significant correlations. Thus only eight variables demonstrated significant correlations with the human ratings. Collinearity Pearson correlations demonstrated that the word concreteness score was highly correlated (>.70) with the word imagability score (N = 160, r =.935, p <.001). Because the word concreteness value had a lower correlation with the human ratings as compared to the word imagability score, the word concreteness value was dropped from the multiple regression analysis. Thus, there were only seven variables included in the final analysis.
11 Crossley et al. 11 Table 2. Selected variables based on Pearson correlations Variable Measure r value p value D Lexical diversity.583 <.001 CELEX content word frequency Word frequency.386 <.001 Word meaningfulness content words MRC database.372 <.001 Word familiarity content words MRC database.362 <.001 Word imagability all words MRC database.313 <.001 Word hypernymy average Hypernymy.281 <.001 LSA sentence to sentence adjacent Semantic co-referentiality.254 <.001 Word concreteness all words MRC database.235 <.010 Polysemy standard deviation Polysemy.141 >.050 Average syllables per word Word length.007 >.050 Multiple regression training set A linear regression analysis was conducted for the seven remaining variables. These seven variables were regressed onto the raters evaluations for the 160 writing samples in the training set. The variables were checked for outliers and multicollinearity. Coefficients were checked for both variance inflation factors (VIF) values and tolerance. All VIF values and tolerance levels were at about 1, indicating that the model data did not suffer from multicollinearity (Field, 2005). The linear regression using the seven variables yielded a significant model, F(3, 156) = , p <.001, r =.678, r 2 =.460. Three variables were significant predictors in the regression: D, word hypernymy values, and CELEX content word frequency. Four variables were not significant predictors: word meaningfulness, word familiarity, LSA sentence to sentence, and word imagability. The latter variables were left out of the subsequent model; t-test information on these variables from the regression model as well as the amount of variance explained (r 2 ) is presented in Table 3. The results from the linear regression demonstrate that the combination of the three variables accounts for 46% of the variance in the human evaluations of lexical proficiency for the 160 essays examined in the training set (see Table 4 for additional information). Table 3. Statistics (t values, p values, and variance explained) for training set variables Variable t value p value Variance explained (r 2 ) D < Word hypernymy average < CELEX content word frequency < Word meaningfulness content words > Word familiarity content words > LSA sentence to sentence adjacent.543 > Word imagability all words.457 >.050 0
12 12 Language Testing XX(X) Table 4. Linear regression analysis to predict essay ratings training set Entry Variable added Correlation r² B B SE Entry 1 D Entry 2 Word hypernymy average Entry 3 CELEX content word frequency Notes: Estimated constant term is 4.701; B is unstandardized Beta; B is standardized Beta; SE is standard error Test set model To further support the results from the multiple regression conducted on the training set, we used the B weights and the constant from the training set multiple regression analysis to estimate how the model would function on an independent data set (the 80 evaluated writing samples held back in the test set). The model produced an estimated value for each writing sample in the test set. We then conducted a Pearson correlation between the estimated score and the actual score. We used this correlation along with its r 2 to demonstrate the strength of the model on an independent data set. The model for the test set yielded r =.649, r 2 =.421. The results from the test set model demonstrate that the combination of the three variables accounted for 42% of the variance in the evaluation of the 80 writing samples comprising the test set. Total set model To examine how the model from the training set predicted the variance in the human evaluations for the total corpus of writing samples, we used the B weights and the constant from the training set multiple regression analysis on the entire data set (the 240 evaluated writing samples that were contained in both the training and the test set). If the total set model is similar to the training and test set model, we can say with a high degree of confidence that it is a reliable model (Whitten and Frank, 2005). We followed the same methodology for the total set model as for the test set model. The model for the entire data set yielded r =.666, r 2 =.444. The results from the entire data set model demonstrate that the combination of the three variables accounts for 44% of the variance in the evaluation of the 240 writing samples that comprise the lexical proficiency corpus. Discussion The results of this analysis demonstrate that three lexical indices taken from Coh-Metrix predict 44% of the variance of the human evaluations of lexical proficiency in the examined writing samples. The correlation value for the analysis (r =.66) demonstrates a large effect size indicating a strong relationship between the three lexical indices and the human evaluations of lexical proficiency (Cohen, 1992). The three variables, lexical diversity, word hypernymy, and word frequency are linked to both vocabulary size and depth of knowledge lexical features. No indices related to accessing core lexical items contributed significantly to our model of lexical proficiency, although many of these
13 Crossley et al. 13 indices demonstrated significant correlations with the human ratings. This finding has important implications for language testing and theories of lexical acquisition. From a language assessment perspective, the results of this study present a model from which to evaluate the lexical proficiency of language learners. The model is supported through convergent validity, but more importantly, it does not suffer from a lack of construct validity because it is strongly linked to psychological theories of how the mind processes language. In addition, the model is also simple and practical enough that it can be used to inform pedagogical decisions and can be used to measure lexical development in language learners. The model for predicting lexical proficiency is presented below. Predicted lexical proficiency = (.022 lexical diversity: D value) + ( average of word hypernymy value) + (.736 content word frequency value) The above model relies on the use of Coh-Metrix. To apply the model, a user would input a text into the tool and collect the results for each of the indices used in the model. The user would then calculate the results from the model. The model will report a number between 1 and 5, which correspond to the lexical proficiency levels found in the survey instrument used by the human raters (see Appendix 1). Unlike previous research into lexical proficiency, the results of this study allow us to extend them into a practical, computational model because the results do not rely on participant survey responses, which are time consuming, or test a relatively small number of words (Meara, 2005; Wesche and Paribakht, 1996; Schmitt and Meara, 1997). The natural evolution of survey approaches that attempt to measure multiple word features necessarily require researchers to develop more finely tuned tests with fewer words, thus making testing lexical proficiency over a range of words impractical. Such a conclusion led Meara and Wolter (2004) to decide that lexical proficiency could best be tested through a single measure of vocabulary size and not through depth of knowledge or access to core lexical items. However, unlike survey instruments meant to measure lexical proficiency, our approach allows us to examine relatively short texts produced by a learner using a variety of indices that provide adequate lexical coverage. In addition, these indices link to lexical features that are important to lexical proficiency and that go beyond vocabulary size. From a theoretical perspective, this study provides evidence as to the strength of various lexical indices in predicting overall lexical proficiency. Perhaps the most robust finding of this study is that an index of lexical diversity, D, explains almost 34% of the variance in human judgments of written lexical proficiency. This signifies that human evaluations of lexical proficiency can best be predicted by the variety of words that a writer produces. Lexical diversity, as an indicator of vocabulary size, thus best explains proficient lexical use. In the writing samples analyzed, neither depth of word knowledge features nor access to core lexical items were as indicative as lexical diversity. Therefore, in our model of lexical proficiency the number of different words that the writer produces is the greatest contributor with more proficient lexicons characterized by more variety in word use. The second most predictive lexical index for the sampled corpus was word hypernymy average, which explained just over 7% of the variance in our human ratings. Accordingly,
14 14 Language Testing XX(X) we argue that the level of specificity that a writer uses relates to judgments of lexical proficiency. More accurately, the more specific words a writer uses, the more likely the writer is to be judged as being less lexically proficient. Such a prediction is supported by recent research that demonstrates that L2 learners word use becomes less specific as time spent studying a language increases (Crossley et al., 2009). Thus, as writers develop hierarchical knowledge of word relations, allowing for less specific word use, they are evaluated as being more lexically proficient. Lastly, this analysis has demonstrated that word frequency is also an important indicator of lexical proficiency. Our word frequency index accounted for about 5% of the variance in the human evaluations of lexical proficiency. Specifically, writing samples with less frequent words were evaluated as having greater lexical proficiency. This finding might be interpreted as supporting the importance of connectionist networks in lexical proficiency judgments. In such an interpretation, learners exposure to frequent words allows the words to be more easily accessible because of their repetition in language. The accessibility of frequent words, which develops over time, likely affords learners with greater processing resources to retrieve less frequent words. The use of less frequent words is thus associated with writers being more lexically proficient. Of secondary interest are those lexical features that demonstrated significant correlations to the lexical proficiency ratings, but were not included in the final regression analysis. Although these indices were not predictive of lexical proficiency in our final model, they are still theoretically interesting. All of these indices were related to either depth of lexical knowledge features or access to core lexical items. They included the MRC psycholinguistic database indices (meaningfulness, familiarity, imagability, and concreteness) and our index of semantic co-referentiality taken from LSA. These indices demonstrated that writing samples scored as more lexically proficient contained less meaningful words, less familiar words, less imagable words, less concrete words, and less semantic co-referentiality. In general, these correlations support the notion that more lexically proficient writers use more sophisticated words that are more abstract, less familiar, less imagable, and have fewer semantic associations with other words. Additionally, the indices that did not demonstrate significant correlations are of interest. For instance, our measure of polysemy did not demonstrate significant correlations with human judgments of lexical proficiency. This finding is likely the result of the index itself, which does not measure the frequency of the senses used or the number of senses for each word. Rather, the index reports on the number of senses that a word has in general, but not for the specific instance of the word as found in the context of the sample. Thus, we cannot exclude the importance of polysemy in models of lexical proficiency, but we can rule out the effectiveness of the WordNet polysemy measure in explaining human judgments of lexical proficiency. Also, polysemy features might be subsumed under the word frequency index that we used. It is generally accepted that more frequent words are also more polysemous (Crossley et al., 2010b). It is also important to note that traditional indices of word length showed no correlations with human judgments of lexical proficiency. We can therefore discount the notion that humans evaluate lexical proficiency based on word length.
15 Crossley et al. 15 Conclusion This analysis has demonstrated that three lexical indices are predictive of human ratings of lexical proficiency. The indices, D, WordNet hypernymy, and CELEX content word frequency, explain 44% of the variance in human evaluations of lexical proficiency in the writing samples examined. These findings indicate strong relationships between the human evaluations and the lexical indices. The results of this study have important theoretical implications as well as the potential to transfer the resulting algorithm into an exploratory model from which to test lexical proficiency in language learners. Such explorations into lexical proficiency are critical because L2 lexical acquisition is enigmatic (Schmitt, 1998) and an overall theory of lexical acquisition is lacking (Meara, 1997). Additionally, the tools currently available from which to analyze lexical proficiency are time intensive and cannot examine a representative sample of words. The approaches described in this study address these problems by examining the efficacy of using computational indices that can examine larger word samples quickly and automatically. This study also presents an approach that is rooted both in applied linguistics and psychology as called for by Meara (1997). We collected experimental data and analyzed it to produce a formal model that helps explain human judgments of lexical proficiency. We then used held back data to predict how well the data fits the predictions made by the formal model. We argue that the questions we ask and to which we provide answers are not only technically interesting, but also psychologically interesting in that they help to explain the interrelation of various lexical indices in explaining lexical proficiency. There are, of course, elements of this current study that need to be further developed in order to improve the extendibility of the findings. First, our sample population of L2 learners was relatively small. With a smaller sample population, there should be less variance between the writing samples. We are optimistic that a larger sample population might actually increase the amount of variance explained by our selected computational indices. Second, many of the indices that we tested are not representative of the entire English lexicon. For example the MRC meaningfulness index we used contains human ratings for only 2627 words. While this might appear small, past analyses have demonstrated that the MRC indices provide coverage for about 80 85% of the word tokens found in speech (Salsbury et al., in press). We expect that those percentages would be lower for written language in which there is less lexical redundancy. Accordingly, many of the indices used in this study could benefit from additional lexical coverage. Third, some features of lexical proficiency were not included in this study because they have not been computationally implemented. Primary among these is an index of collocational proficiency. A study examining potential models of collocational proficiency has recently been completed (Crossley and Salsbury, in press). We are hopeful that such an index of collocational proficiency will prove an important predictor of lexical proficiency and strengthen the statistical findings reported in this paper. Last, this study only concentrated on written language. Further studies into the predictive validity of these indices for speech data are warranted. While we present a model of lexical proficiency that might be considered rather simple, we have confidence in its predictive power. However, we also recognize that the model
16 16 Language Testing XX(X) we present does not explain all of the variance in the human ratings. This is likely the result of concentrating only on lexical indices related to vocabulary size, depth of knowledge, and access to core lexical items. We hypothesize that additional variance might be explained by indices that measure additional lexical features such as collocation proficiency and syntactic patterns. Additional variance might be explained by words pronounceability, orthography, and morphology (Nation, 2005). We argue, though, that this study forms a strong foundation for moving past exploratory considerations of a lexical organization system (Meara, 2002) and towards functional models of lexical proficiency. This is an important step in understanding what the term lexical proficiency signifies and addressing methods of assessing it. Acknowledgements This research was supported in part by the Institute for Education Sciences (IES R305A and IES R305G ). Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the IES. We also give special thanks to our human raters: Brandi Williams, Jessica Mann, and Abigail Voller. We would also like to acknowledge the assistance of Dr. Phillip McCarthy in helping validate the scoring rubric used in this study. References Anglin, J. M. (1993). Vocabulary development: A morphological analysis. Monographs of the Society for Research in Child Development, 58 (10), Baayen, R. H., Piepenbrock, R., & Gulikers, L. (1995). CELEX. Philadelphia, Linguistic Data Consortium Balota, D. A., & Chumbley, J. L. (1984). Are lexical decisions a good measure of lexical access? The role of word frequency in the neglected decision stage. Journal of Experimental Psychology: Human Perception and Performance, 10, Barcikowski, R., & Stevens, J. P. (1975). A Monte Carlo study of the stability of canonical correlations, canonical weights, and canonical variate-variable correlations. Multivariate Behavioral Research, Vol. 10: Bell, H. (2003). Using Frequency Lists to Assess L2 Texts. University of Wales Swansea: Unpublished Thesis. Boldt, R. F, Larsen-Freeman, D., Reed, M. S., & Courtney, R. G. (1992). Distributions of ACTFL ratings by TOEFL score ranges. TOEFL Research Reports, 41. Carrol, J. B. (1964). Language and Thought. New Jersey: Prentice-Hall. Carter, R. (1987). Vocabu1ary: Applied linguistic perspectives. London: Allen & Unwin. Chaffin, R., & Glass, A. (1990). A comparison of hyponym and synonym decisions. Journal of Psycholinguistic Research, 19 (4), Cohen, J. (1992). A power primer. Psychological Bulletin 112, Crossley, S. A. & McNamara, D. S. (2009). Computationally assessing lexical differences in second language writing. Journal of Second Language Writing, 17 (2), Crossley, S. A., & Salsbury, T. (2010). Using lexical indices to predict produced and not produced words in second language learners. The Mental Lexicon, 5 (1), Crossley, S. A., & Salsbury, T. (in press). The development of collocational accuracy and production in English second language speakers. International Review of Applied Linguistics in Teaching, 49 (1).
17 Crossley et al. 17 Crossley, S. A., Salsbury, T., & McNamara, D. S. (2009). Measuring L2 lexical growth using hypernymic relationships. Language Learning, 59 (2). Crossley, S. A., Salsbury, T., & McNamara, D. S. (2010a). The development of semantic relations in second language speakers: A case for Latent Semantic Analysis. Vigo International Journal of Applied Linguistics, 7, Crossley, S. A., Salsbury, T., & McNamara, D. S. (2010b). The development of polysemy and frequency use in English second language speakers. Language Learning, 60 (3). Daller, H., van Hout, R., & Treffers-Daller, J. (2003). Lexical richness in the spontaneous speech of bilinguals. Applied Linguistics, 24 (2), Ellis, N. (2002). Frequency effects in language processing. Studies in Second Language Acquisition, 24 (2), Ellis, N. C., & Beaton, A. (1993). Psycholinguistic determinants of foreign language vocabulary acquisition. Language Learning, 43 (4), Ellis, R. (1995). Modified oral input and the acquisition of word meanings. Applied Linguistics, 16, Fellbaum, C. (1998). WordNet: An electronic lexical database. Cambridge, MA: MIT Press. Field, A. (2005). Discovering statistics using SPSS. London: Sage Publications. Forster, K. (1976). Accessing the mental lexicon. In R. J. Wales & E. Walker (Eds.), New approaches to language mechanisms (pp ). Amsterdam: North Holland Gee, N. R., Nelson, D. L., & Krawczyk, D. (1999). Is the concreteness effect a result of underlying network interconnectivity? Journal of Memory and Language, 40, Glanzer, M., & Ehrenreich, S. L. (1979). Structure and search of the internal lexicon. Journal of Verbal Learning and Verbal Behavior, 18, Graesser, A.C., McNamara, D.S., Louwerse, M.M., and Cai, Z. (2004). Coh-Metrix: Analysis of text on cohesion and language. Behavioral Research Methods, Instruments, and Computers. 36: Haastrup, K., & Henriksen, B. (2000). Vocabulary acquisition: Acquiring depth of knowledge through network building. International Journal of Applied Linguistics, 10 (2), Hatch, E., & Lazaraton, A. (1991). The research manual: Design and statistics for applied linguistics. Boston, MA: Newbury House. Herdan, G. (1960). Type-token mathematics. Gravenhage: Mouton. Ijaz, I. H. (1986). Linguistic and cognitive determinants of lexical acquisition in a second language. Language Learning, 36 (4), Jurafsky, D., & Martin, J. H. (2008). Speech and language processing: An introduction to natural language processing, computational linguistics, and speech recognition, Upper Saddle River, NJ: Prentice-Hall. Kirsner, K. (1994). Implicit processes in second language learning. In N. Ellis (Ed.), Implicit and explicit learning of languages (pp ). San Diego, CA: Academic Press. Landauer, T. K., McNamara, D. S, Dennis, S., & Kintsch, W. (Eds.). (2007). LSA: A road to meaning. Mahwah, NJ: Erlbaum. Laufer, B., & Nation, P. (1995). Vocabulary size and use: Lexical richness in L2 written production. Applied Linguistics, 16 (3), Levenston, E., & Blum, S. (1977). Aspects of lexical simplification in the speech and writing of advanced adult learners. In P. S. Corder & E. Roulet (Eds.), The notions of simplification, interlanguages and pidgins and their relation to second language pedagogy (pp ). Librairie Droz.
18 18 Language Testing XX(X) Malvern, D. D., Richards, B. J., Chipere, N., & Durán, P. (2004). Lexical diversity and language development: Quantification and assessment. Houndmills, Basingstoke, Hampshire: Palgrave Macmillan. McCarthy, P. M., & Jarvis, S. (in press). MTLD, vocd-d, and HD-D: A validation study of sophisticated approaches to lexical diversity assessment. Behavior Research Methods. Meara, P. (1997). Towards a new approach to modelling vocabulary acquisition. In N. Schmitt & M. McCarthy (Eds.), Vocabulary: Description, acquisition and pedagogy (pp ). Cambridge: Cambridge University Press. Meara, P. (2002). The rediscovery of vocabulary. Second Language Research, 18 (4): Meara, P. (2005). Designing vocabulary tests for English, Spanish and other languages. In C. Butler, S. Christopher, M. Á. Gómez González, & S. M. Doval-Suárez (Eds.) The dynamics of language use (pp ). Amsterdam, John Benjamins Press. Meara, P., & Bell, H. (2001). P_Lex: A simple and effective way of describing the lexical characteristics of short L2 texts. Prospect, 16 (3), Meara, P., & Wolter, B. (2004). V_Links: Beyond vocabulary depth. Angles on the English-Speaking World, 4, Murphy, G. L. (2004). The big book of concepts. Cambridge, MA: MIT Press. Nation, I. S. P. (2005). Teaching and learning vocabulary. In E. Hinkel (Ed.), Handbook of research in second language teaching and learning (pp ). Mahwah, New Jersey: Lawrence Erlbaum Associates, Publishers. Nation, I. S. P., & Heatley, A. (1996). VocabProfile, Word and Range: Programs for processing text. Wellington, New Zealand: School of Linguistics and Applied Language Studies, Victoria University of Wellington.. Newman, M. L., Groom, C. J., Handelman, L. D., Pennebaker, J. W. (2008). Gender difference in language use: An analysis of 14,000 text samples. Discourse Processes, 45, Pearson, P. D., Hiebert, E. H., & Kamil, M. L. (2007). Theory and research into practice: Vocabulary assessment: What we know and what we need to learn. Reading Research Quarterly, 42 (2), Polio, C. (2001). Research methodology in second language writing research: The case textbased studies. In T. Silva & P. K. Matsuda (Eds.), On second language writing (pp ). Mahwah, New Jersey: Lawrence Erlbaum Associates. Read, J. (1993). The development of a new measure of L2 vocabulary knowledge. Language Testing, 10, Salsbury, T., Crossley, S. A., & McNamara, D. S. (in press). Psycholinguistic word information in second language oral discourse. Second Language Research. Schmitt, N. (1998). Tracking the incremental acquisition of a second language vocabulary: A longitudinal study, Language Learning, 48 (2), Schmitt, N., & Meara, P. (1997). Researching vocabulary through a word knowledge framework. Studies in Second Language Acquisition, 19, Schwanenflugel, P. (1991). Why are abstract concepts hard to understand? In P. Schwanenflugel (Ed.), The psychology of word meanings (pp ). Hillsdale, New Jersey: Erlbaum. Snow, C. E. (1990). The development of definitional skill. Journal of Child Language, 17, Toglia, M. P., & Battig, W. F. (1978). Handbook of semantic word norms. Hillsdale, NJ: Lawrence Erlbaum Associates, Publishers.
19 Crossley et al. 19 Vermeer, A. (2000). Coming to grips with lexical richness in spontaneous speech data. Language Testing, 17 (1), Verspoor, M., & Lowie, W. (2003). Making sense of polysemous words. Language Learning, 53 (3), Wesche, M., & Paribakht, T. M. (1996). Assessing vocabulary knowledge: Depth vs. breadth. Canadian Modern Language Review, 53, Whitten, I. A. & Frank, E. (2005). Data Mining. San Francisco: Elsevier. Wilson, M.D. (1988) The MRC Psycholinguistic Database: Machine Readable Dictionary, Version 2. Behavioural Research Methods, Instruments and Computers, 20 (1), Wolter, B. (2001). Comparing the L1 and L2 mental lexicons: A depth of individual word knowledge model. Studies in Second Language Acquisition, 23, Zareva, A. (2007). Structure of the second language mental lexicon: How does it compare to native speakers lexical organization? Second Language Research, 23 (2), Appendix 1: Holistic rating form After reading each writing sample, assign a holistic score based on the rubric below. For the following evaluations you will need to use a grading scale between 1 (minimum) and 5 (maximum). The distance between each grade (e.g. 1 2, 3 4, 4 5) should be considered equal. Score of 5: A sample in this category demonstrates clear and consistent mastery of the English lexicon, although it may have a few minor errors. A typical sample effectively uses appropriate conceptual categories (both concrete and abstract), demonstrates clear coherence between words, lexical-semantic connections, and is lexically diverse (enough to explain complex matters in detail). In general, the sample exhibits a skillful use of language, using a varied, accurate, and apt vocabulary with ease and fluency. Score of 4: A sample in this category demonstrates reasonably consistent mastery of the English lexicon, although it will have occasional errors or lapses in lexical quality. The sample demonstrates the appropriate use of conceptual categories (both concrete and abstract), coherence between words, lexical-semantic connections, and lexical diversity that allows for the discussion of complex matters, but not always. Overall, the sample uses appropriate and precise vocabulary and appears fluent and accurate. Score of 3: A sample in this category demonstrates adequate lexical mastery, although it will have lapses in quality. The sample demonstrates some appropriate uses of conceptual categories (including abstract concepts, but mostly concrete concepts), coherence between words, lexical-semantic connections, and lexical diversity. Overall, the sample uses generally appropriate and precise vocabulary, but demonstrates an inconsistent mastery. Score of 2: A sample in this category demonstrates developing lexical mastery, but is marked by one or more weaknesses in conceptual categories, lexical-semantic connections, cohesion between words, and lexical diversity. However, discourse in the sample
20 20 Language Testing XX(X) is generally connected. By and large, the sample uses weak vocabulary or inappropriate word choice, lacks variety, and depends on concrete words. The lexical problems in the sample are serious enough that meaning is somewhat obscured. Score of 1: A sample in this category demonstrates little lexical mastery, and is flawed by two or more weaknesses in conceptual categories, lexical-semantic connections, cohesion between words, and lexical diversity. The discourse in this category displays very little coherence and facility in the use of language. It also relies heavily on the use and repetition of memorized phrases. Overall, the sample demonstrates limited vocabulary, incorrect word choices, and exhibits frequent lexical problems so serious that meaning is often obscured. Holistic score based on attached rubric (1 5): Appendix 2
Efficient Techniques for Improved Data Classification and POS Tagging by Monitoring Extraction, Pruning and Updating of Unknown Foreign Words
, pp.290-295 http://dx.doi.org/10.14257/astl.2015.111.55 Efficient Techniques for Improved Data Classification and POS Tagging by Monitoring Extraction, Pruning and Updating of Unknown Foreign Words Irfan
An Analysis of the Eleventh Grade Students Monitor Use in Speaking Performance based on Krashen s (1982) Monitor Hypothesis at SMAN 4 Jember
1 An Analysis of the Eleventh Grade Students Monitor Use in Speaking Performance based on Krashen s (1982) Monitor Hypothesis at SMAN 4 Jember Moh. Rofid Fikroni, Musli Ariani, Sugeng Ariyanto Language
Differences in linguistic and discourse features of narrative writing performance. Dr. Bilal Genç 1 Dr. Kağan Büyükkarcı 2 Ali Göksu 3
Yıl/Year: 2012 Cilt/Volume: 1 Sayı/Issue:2 Sayfalar/Pages: 40-47 Differences in linguistic and discourse features of narrative writing performance Abstract Dr. Bilal Genç 1 Dr. Kağan Büyükkarcı 2 Ali Göksu
Chapter Seven. Multiple regression An introduction to multiple regression Performing a multiple regression on SPSS
Chapter Seven Multiple regression An introduction to multiple regression Performing a multiple regression on SPSS Section : An introduction to multiple regression WHAT IS MULTIPLE REGRESSION? Multiple
Some Implications of Controlling Contextual Constraint: Exploring Word Meaning Inference by Using a Cloze Task
Some Implications of Controlling Contextual Constraint: Exploring Word Meaning Inference by Using a Cloze Task Abstract 20 vs Keywords: Lexical Inference, Contextual Constraint, Cloze Task 1. Introduction
Additional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jin-tselink/tselink.htm
Mgt 540 Research Methods Data Analysis 1 Additional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jin-tselink/tselink.htm http://web.utk.edu/~dap/random/order/start.htm
Gallito 2.0: a Natural Language Processing tool to support Research on Discourse
Presented in the Twenty-third Annual Meeting of the Society for Text and Discourse, Valencia from 16 to 18, July 2013 Gallito 2.0: a Natural Language Processing tool to support Research on Discourse Guillermo
MASTER OF PHILOSOPHY IN ENGLISH AND APPLIED LINGUISTICS
University of Cambridge: Programme Specifications Every effort has been made to ensure the accuracy of the information in this programme specification. Programme specifications are produced and then reviewed
Relating the ACT Indicator Understanding Complex Texts to College Course Grades
ACT Research & Policy Technical Brief 2016 Relating the ACT Indicator Understanding Complex Texts to College Course Grades Jeff Allen, PhD; Brad Bolender; Yu Fang, PhD; Dongmei Li, PhD; and Tony Thompson,
How To Teach English To Other People
TESOL / NCATE Program Standards STANDARDS FOR THE ACCREDIATION OF INITIAL PROGRAMS IN P 12 ESL TEACHER EDUCATION Prepared and Developed by the TESOL Task Force on ESL Standards for P 12 Teacher Education
A Survey of Online Tools Used in English-Thai and Thai-English Translation by Thai Students
69 A Survey of Online Tools Used in English-Thai and Thai-English Translation by Thai Students Sarathorn Munpru, Srinakharinwirot University, Thailand Pornpol Wuttikrikunlaya, Srinakharinwirot University,
Building a Question Classifier for a TREC-Style Question Answering System
Building a Question Classifier for a TREC-Style Question Answering System Richard May & Ari Steinberg Topic: Question Classification We define Question Classification (QC) here to be the task that, given
Reading Competencies
Reading Competencies The Third Grade Reading Guarantee legislation within Senate Bill 21 requires reading competencies to be adopted by the State Board no later than January 31, 2014. Reading competencies
CREATING LEARNING OUTCOMES
CREATING LEARNING OUTCOMES What Are Student Learning Outcomes? Learning outcomes are statements of the knowledge, skills and abilities individual students should possess and can demonstrate upon completion
Lexico-Semantic Relations Errors in Senior Secondary School Students Writing ROTIMI TAIWO Obafemi Awolowo University, Ile-Ife, Nigeria
Nordic Journal of African Studies 10(3): 366-373 (2001) Lexico-Semantic Relations Errors in Senior Secondary School Students Writing ROTIMI TAIWO Obafemi Awolowo University, Ile-Ife, Nigeria ABSTRACT The
ELL Considerations for Common Core-Aligned Tasks in English Language Arts
ELL Considerations for Common Core-Aligned Tasks in English Language Arts A substantial body of research clearly indicates that utilizing specific instructional modifications as well as targeted pedagogical
Technical Report. Overview. Revisions in this Edition. Four-Level Assessment Process
Technical Report Overview The Clinical Evaluation of Language Fundamentals Fourth Edition (CELF 4) is an individually administered test for determining if a student (ages 5 through 21 years) has a language
Study Plan for Master of Arts in Applied Linguistics
Study Plan for Master of Arts in Applied Linguistics Master of Arts in Applied Linguistics is awarded by the Faculty of Graduate Studies at Jordan University of Science and Technology (JUST) upon the fulfillment
Silvermine House Steenberg Office Park, Tokai 7945 Cape Town, South Africa Telephone: +27 21 702 4666 www.spss-sa.com
SPSS-SA Silvermine House Steenberg Office Park, Tokai 7945 Cape Town, South Africa Telephone: +27 21 702 4666 www.spss-sa.com SPSS-SA Training Brochure 2009 TABLE OF CONTENTS 1 SPSS TRAINING COURSES FOCUSING
The Word Associates Format: Validation evidence
Article The Word Associates Format: Validation evidence Language Testing 28(1) 105 126 The Author(s) 2011 Reprints and permission: sagepub.co.uk/journalspermissions.nav DOI: 10.1177/0265532210373605 http://ltj.sagepub.com
ANALYSIS OF LEXICO-SYNTACTIC PATTERNS FOR ANTONYM PAIR EXTRACTION FROM A TURKISH CORPUS
ANALYSIS OF LEXICO-SYNTACTIC PATTERNS FOR ANTONYM PAIR EXTRACTION FROM A TURKISH CORPUS Gürkan Şahin 1, Banu Diri 1 and Tuğba Yıldız 2 1 Faculty of Electrical-Electronic, Department of Computer Engineering
Dr. Wei Wei Royal Melbourne Institute of Technology, Vietnam Campus January 2013
Research Summary: Can integrated skills tasks change students use of learning strategies and materials? A case study using PTE Academic integrated skills items Dr. Wei Wei Royal Melbourne Institute of
Natural Language to Relational Query by Using Parsing Compiler
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 4, Issue. 3, March 2015,
Teaching Vocabulary to Young Learners (Linse, 2005, pp. 120-134)
Teaching Vocabulary to Young Learners (Linse, 2005, pp. 120-134) Very young children learn vocabulary items related to the different concepts they are learning. When children learn numbers or colors in
Student Learning Outcomes in Hybrid and Face-to-Face Beginning Spanish Language Courses
Student Learning Outcomes in Hybrid and Face-to-Face Beginning Spanish Language Courses Casilde A. Isabelli University of Nevada, (United States of America) [email protected] Abstract This study investigates
Texas Success Initiative (TSI) Assessment. Interpreting Your Score
Texas Success Initiative (TSI) Assessment Interpreting Your Score 1 Congratulations on taking the TSI Assessment! The TSI Assessment measures your strengths and weaknesses in mathematics and statistics,
Modern foreign languages
Modern foreign languages Programme of study for key stage 3 and attainment targets (This is an extract from The National Curriculum 2007) Crown copyright 2007 Qualifications and Curriculum Authority 2007
Predictability of Vocabulary Size on Learners EFL Proficiency: Taking VST, CET4 and CET6 as Instruments
Studies in Literature and Language Vol. 10, No. 3, 2015, pp. 18-23 DOI:10.3968/6679 ISSN 1923-1555[Print] ISSN 1923-1563[Online] www.cscanada.net www.cscanada.org Predictability of Vocabulary Size on Learners
Effective Data Retrieval Mechanism Using AML within the Web Based Join Framework
Effective Data Retrieval Mechanism Using AML within the Web Based Join Framework Usha Nandini D 1, Anish Gracias J 2 1 [email protected] 2 [email protected] Abstract A vast amount of assorted
Understanding Learners Cognitive Abilities: A Model of Mobilizing Non-English Majors Cognitive Abilities in the Process of Their Writing in English
Understanding Learners Cognitive Abilities: A Model of Mobilizing Non-English Majors Cognitive Abilities in the Process of Their Writing in English Liu Fengming Southwest Normal University Introduction
How To Teach Reading
Florida Reading Endorsement Alignment Matrix Competency 1 The * designates which of the reading endorsement competencies are specific to the competencies for English to Speakers of Languages (ESOL). The
STRATEGIES FOR EFFECTIVE VOCABULARY INSTRUCTION
STRATEGIES FOR EFFECTIVE VOCABULARY INSTRUCTION DONNA C. KESTER PHILLIPS Assistant Professor of Education Niagara University, NY CHANDRA J. FOOTE Professor of Education Niagara University, NY LAURIE J.
Glossary of Terms Ability Accommodation Adjusted validity/reliability coefficient Alternate forms Analysis of work Assessment Battery Bias
Glossary of Terms Ability A defined domain of cognitive, perceptual, psychomotor, or physical functioning. Accommodation A change in the content, format, and/or administration of a selection procedure
Learning Strategies for Vocabulary Development
Learning Strategies for Vocabulary Development 105 Learning Strategies for Vocabulary Development Yongqi Gu Victoria University of Wellington, New Zealand ABSTRACT This paper reports on a study of changes
ACT Research Explains New ACT Test Writing Scores and Their Relationship to Other Test Scores
ACT Research Explains New ACT Test Writing Scores and Their Relationship to Other Test Scores Wayne J. Camara, Dongmei Li, Deborah J. Harris, Benjamin Andrews, Qing Yi, and Yong He ACT Research Explains
AMERICAN COUNCIL ON THE TEACHING OF FOREIGN LANGUAGES (ACTFL)
AMERICAN COUNCIL ON THE TEACHING OF FOREIGN LANGUAGES (ACTFL) PROGRAM STANDARDS FOR THE PREPARATION OF FOREIGN LANGUAGE TEACHERS (INITIAL LEVEL Undergraduate & Graduate) (For K-12 and Secondary Certification
Multi language e Discovery Three Critical Steps for Litigating in a Global Economy
Multi language e Discovery Three Critical Steps for Litigating in a Global Economy 2 3 5 6 7 Introduction e Discovery has become a pressure point in many boardrooms. Companies with international operations
A Performance Comparison of Native and Non-native Speakers of English on an English Language Proficiency Test ...
technical report A Performance Comparison of Native and Non-native Speakers of English on an English Language Proficiency Test.......... Agnes Stephenson, Ph.D. Hong Jiao Nathan Wall This paper was presented
Facilitating Business Process Discovery using Email Analysis
Facilitating Business Process Discovery using Email Analysis Matin Mavaddat [email protected] Stewart Green Stewart.Green Ian Beeson Ian.Beeson Jin Sa Jin.Sa Abstract Extracting business process
SEMINOLE COUNTY PUBLIC SCHOOLS ESOL Strategies For Mainstream Teachers 2011-2012
SEMINOLE COUNTY PUBLIC SCHOOLS ESOL Strategies For Mainstream Teachers 2011-2012 ESOL/World Languages Department For additional resources and strategies, visit the ESOL/ World Languages Department website
CHARACTERISTICS FOR STUDENTS WITH: LIMITED ENGLISH PROFICIENCY (LEP)
CHARACTERISTICS FOR STUDENTS WITH: LIMITED ENGLISH PROFICIENCY (LEP) Research has shown that students acquire a second language in the same way that they acquire the first language. It is an exploratory
French Language and Culture. Curriculum Framework 2011 2012
AP French Language and Culture Curriculum Framework 2011 2012 Contents (click on a topic to jump to that page) Introduction... 3 Structure of the Curriculum Framework...4 Learning Objectives and Achievement
A STATISTICS COURSE FOR ELEMENTARY AND MIDDLE SCHOOL TEACHERS. Gary Kader and Mike Perry Appalachian State University USA
A STATISTICS COURSE FOR ELEMENTARY AND MIDDLE SCHOOL TEACHERS Gary Kader and Mike Perry Appalachian State University USA This paper will describe a content-pedagogy course designed to prepare elementary
Knowledge Discovery from patents using KMX Text Analytics
Knowledge Discovery from patents using KMX Text Analytics Dr. Anton Heijs [email protected] Treparel Abstract In this white paper we discuss how the KMX technology of Treparel can help searchers
COMPUTATIONAL DATA ANALYSIS FOR SYNTAX
COLING 82, J. Horeck~ (ed.j North-Holland Publishing Compa~y Academia, 1982 COMPUTATIONAL DATA ANALYSIS FOR SYNTAX Ludmila UhliFova - Zva Nebeska - Jan Kralik Czech Language Institute Czechoslovak Academy
KSE Comp. support for the writing process 2 1
KSE Comp. support for the writing process 2 1 Flower & Hayes cognitive model of writing A reaction against stage models of the writing process E.g.: Prewriting - Writing - Rewriting They model the growth
Using LSI for Implementing Document Management Systems Turning unstructured data from a liability to an asset.
White Paper Using LSI for Implementing Document Management Systems Turning unstructured data from a liability to an asset. Using LSI for Implementing Document Management Systems By Mike Harrison, Director,
Integrating Reading and Writing for Effective Language Teaching
Integrating Reading and Writing for Effective Language Teaching Ruwaida Abu Rass (Israel) Writing is a difficult skill for native speakers and nonnative speakers alike, because writers must balance multiple
How the Computer Translates. Svetlana Sokolova President and CEO of PROMT, PhD.
Svetlana Sokolova President and CEO of PROMT, PhD. How the Computer Translates Machine translation is a special field of computer application where almost everyone believes that he/she is a specialist.
Search and Information Retrieval
Search and Information Retrieval Search on the Web 1 is a daily activity for many people throughout the world Search and communication are most popular uses of the computer Applications involving search
Automated Essay Scoring with the E-rater System Yigal Attali Educational Testing Service [email protected]
Automated Essay Scoring with the E-rater System Yigal Attali Educational Testing Service [email protected] Abstract This paper provides an overview of e-rater, a state-of-the-art automated essay scoring
Key words related to the foci of the paper: master s degree, essay, admission exam, graders
Assessment on the basis of essay writing in the admission to the master s degree in the Republic of Azerbaijan Natig Aliyev Mahabbat Akbarli Javanshir Orujov The State Students Admission Commission (SSAC),
Why are Organizations Interested?
SAS Text Analytics Mary-Elizabeth ( M-E ) Eddlestone SAS Customer Loyalty [email protected] +1 (607) 256-7929 Why are Organizations Interested? Text Analytics 2009: User Perspectives on Solutions
Descriptive Statistics
Descriptive Statistics Primer Descriptive statistics Central tendency Variation Relative position Relationships Calculating descriptive statistics Descriptive Statistics Purpose to describe or summarize
Moderator and Mediator Analysis
Moderator and Mediator Analysis Seminar General Statistics Marijtje van Duijn October 8, Overview What is moderation and mediation? What is their relation to statistical concepts? Example(s) October 8,
Evaluation of the Impacts of Data Model and Query Language on Query Performance
Evaluation of the Impacts of Data Model and Query Language on Query Performance ABSTRACT Hock Chuan Chan National University of Singapore [email protected] It is important to understand how users
Early Childhood Study of Language and Literacy Development of Spanish-Speaking Children
Early Childhood Study of Language and Literacy Development of Spanish-Speaking Children Subproject 1 of Acquiring Literacy in English: Crosslinguistic, Intralinguistic, and Developmental Factors Project
Teaching Young Children How to Read: Phonics vs. Whole Language. Introduction and Background
Kelly Waldo Senior Capstone Paper Paoze Thao, PhD California State University Monterey Bay Teaching Young Children How to Read: Phonics vs. Whole Language Introduction and Background I am interested in
DISCRIMINANT FUNCTION ANALYSIS (DA)
DISCRIMINANT FUNCTION ANALYSIS (DA) John Poulsen and Aaron French Key words: assumptions, further reading, computations, standardized coefficents, structure matrix, tests of signficance Introduction Discriminant
Running head: SCHOOL COMPUTER USE AND ACADEMIC PERFORMANCE. Using the U.S. PISA results to investigate the relationship between
Computer Use and Academic Performance- PISA 1 Running head: SCHOOL COMPUTER USE AND ACADEMIC PERFORMANCE Using the U.S. PISA results to investigate the relationship between school computer use and student
'JOURNAL SOCIAL' She. John Ishiyama* Pergamon
Pergamon Available online at www.sciencedirect.com Per nscieince c_h DIRECT' SSCIENCE The Social Science Journal 42 (2005) 359-366 T She SOCIAL' 'JOURNAL The structure of an undergraduate major and student
TExMaT I Texas Examinations for Master Teachers. Preparation Manual. 085 Master Reading Teacher
TExMaT I Texas Examinations for Master Teachers Preparation Manual 085 Master Reading Teacher Copyright 2006 by the Texas Education Agency (TEA). All rights reserved. The Texas Education Agency logo and
psychology and its role in comprehension of the text has been explored and employed
2 The role of background knowledge in language comprehension has been formalized as schema theory, any text, either spoken or written, does not by itself carry meaning. Rather, according to schema theory,
Oral language is the foundation on which reading and writing are
If children come to school with welld eveloped oral language, it must be expanded. If children come to school with underdeve l o p e d oral language, it must be developed. Foundations for Learning to Read
Opportunity Document for STEP Literacy Assessment
Opportunity Document for STEP Literacy Assessment Introduction Children, particularly in urban settings, begin school with a variety of strengths and challenges that impact their learning. Some arrive
LEARNING WITH SIGN AND LIPREADING : ONLINE MULTIMEDIA EDUCATIONAL SOFTWARE
LEARNING WITH SIGN AND LIPREADING : ONLINE MULTIMEDIA EDUCATIONAL SOFTWARE Charalampos Alatzias 1, Anatoli Makarona 2, Theofilos Tsanaktsidis 3 1 Informatics Teacher, MΕd, High School and Lyceum for the
The Impact of Management Information Systems on the Performance of Governmental Organizations- Study at Jordanian Ministry of Planning
The Impact of Management Information Systems on the Performance of Governmental Organizations- Study at Jordanian Ministry of Planning Dr. Shehadeh M.A.AL-Gharaibeh Assistant prof. Business Administration
THE USE OF LEXICAL NETWORKS IN EFL VOCABULARY TEACHING
THE USE OF LEXICAL NETWORKS IN EFL VOCABULARY TEACHING Dr. Hasan BAYRAKTAR, Middle East Technical University Department of Foreign Language Education Ankara, TURKEY ABSTRACT This study has evaluated the
COMMUNICATION COMMUNITIES CULTURES COMPARISONS CONNECTIONS. STANDARDS FOR FOREIGN LANGUAGE LEARNING Preparing for the 21st Century
COMMUNICATION COMMUNITIES CULTURES COMPARISONS CONNECTIONS STANDARDS FOR FOREIGN LANGUAGE LEARNING Preparing for the 21st Century Language and communication are at the heart of the human experience. The
Implementation Date Fall 2008. Marketing Pathways
PROGRAM CONCENTRATION: CAREER PATHWAY: COURSE TITLE: Marketing, Sales & Service Optional Course for All Marketing Pathways Marketing Research In this course, high school students will gain an understanding
Utah Comprehensive Counseling and Guidance Program Evaluation Report
Utah Comprehensive Counseling and Guidance Program Evaluation Report John Carey and Karen Harrington Center for School Counseling Outcome Research School of Education University of Massachusetts Amherst
Chapter 8. Final Results on Dutch Senseval-2 Test Data
Chapter 8 Final Results on Dutch Senseval-2 Test Data The general idea of testing is to assess how well a given model works and that can only be done properly on data that has not been seen before. Supervised
Syllabus: a list of items to be covered in a course / a set of headings. Language syllabus: language elements and linguistic or behavioral skills
Lexical Content and Organisation of a Language Course Syllabus: a list of items to be covered in a course / a set of headings Language syllabus: language elements and linguistic or behavioral skills Who
Experimental methods. Elisabeth Ahlsén Linguistic Methods Course
Experimental methods Elisabeth Ahlsén Linguistic Methods Course Experiment Method for empirical investigation of question or hypothesis 2 types a) Lab experiment b) Naturalistic experiment Question ->
Early Childhood Education Draft Board Resolution
SAMPLE RESOLUTION Early Childhood Education Draft Board Resolution How To Use This Resolution California s new school funding law, (the Local Control Funding Formula or LCFF) provides an opportunity for
Difficulties that Arab Students Face in Learning English and the Importance of the Writing Skill Acquisition Key Words:
Difficulties that Arab Students Face in Learning English and the Importance of the Writing Skill Acquisition Key Words: Lexical field academic proficiency syntactic repertoire context lexical categories
Handwriting in the 21st Century?
Handwriting in the 21st Century? AN EDUCATIONAL SUMMIT Handwriting in the 21st Century? Research Shows Why Handwriting Belongs in Today s Classroom A Summary of Research Presented at Handwriting in the
Service courses for graduate students in degree programs other than the MS or PhD programs in Biostatistics.
Course Catalog In order to be assured that all prerequisites are met, students must acquire a permission number from the education coordinator prior to enrolling in any Biostatistics course. Courses are
PTE Academic. Score Guide. November 2012. Version 4
PTE Academic Score Guide November 2012 Version 4 PTE Academic Score Guide Copyright Pearson Education Ltd 2012. All rights reserved; no part of this publication may be reproduced without the prior written
National assessment of foreign languages in Sweden
National assessment of foreign languages in Sweden Gudrun Erickson University of Gothenburg, Sweden [email protected] This text was originally published in 2004. Revisions and additions were made
Students Track Their Own Learning with SAS Data Notebook
Students Track Their Own Learning with SAS Data Notebook Lucy Kosturko SAS Institute, Inc. [email protected] Jennifer Sabourin SAS Institute, Inc. [email protected] Scott McQuiggan SAS Institute,
GEORGIA STANDARDS FOR THE APPROVAL OF PROFESSIONAL EDUCATION UNITS AND EDUCATOR PREPARATION PROGRAMS
GEORGIA STANDARDS FOR THE APPROVAL OF PROFESSIONAL EDUCATION UNITS AND EDUCATOR PREPARATION PROGRAMS (Effective 9/01/08) Kelly Henson Executive Secretary Table of Contents Standard 1: Candidate Knowledge,
Task-based pedagogy in technology mediated writing
Lingnan University Digital Commons @ Lingnan University Staff Publications - Lingnan University 7-6-2015 Task-based pedagogy in technology mediated writing Preet Pankaj HIRADHAR Lingnan University, Hong
Nebraska School Counseling State Evaluation
Nebraska School Counseling State Evaluation John Carey and Karen Harrington Center for School Counseling Outcome Research Spring 2010 RESEARCH S c h o o l o f E d u c a t i o n U n i v e r s i t y o f
Comparing Ontology-based and Corpusbased Domain Annotations in WordNet.
Comparing Ontology-based and Corpusbased Domain Annotations in WordNet. A paper by: Bernardo Magnini Carlo Strapparava Giovanni Pezzulo Alfio Glozzo Presented by: rabee ali alshemali Motive. Domain information
Text Analytics. A business guide
Text Analytics A business guide February 2014 Contents 3 The Business Value of Text Analytics 4 What is Text Analytics? 6 Text Analytics Methods 8 Unstructured Meets Structured Data 9 Business Application
Reading Instruction and Reading Achievement Among ELL Students
Research Into Practice READING Reading Instruction and Reading Achievement Among ELL Students Principles of ELL Reading Instruction Some very straightforward principles, directly supported by research,
The Impact of Input Enhancement through Multimedia on the Improvement of Writing Ability
The Impact of Input Enhancement through Multimedia on the Improvement of Writing Ability Roya Khoii 1, Azin Tabrizi 2 1 Islamic Azad University, Tehran North Branch, Tehran, Iran, 2 Young Researchers Club,
This Master thesis would not have its current shape without the help of several people. I
Acknowledgements This Master thesis would not have its current shape without the help of several people. I would like to thank my supervisor, Dr Marjolijn Verspoor, who guided me through the process of
Issues in Second Language Listening Comprehension and the Pedagogical Implications
Issues in Second Language Listening Comprehension and the Pedagogical Implications Tomoko Kurita 1 Teachers College, Columbia University, Tokyo ABSTRACT Listening is an important language skill to develop
SPECIFIC LEARNING DISABILITIES (SLD)
Together, We Can Make A Difference Office 770-577-7771 Toll Free1-800-322-7065 www.peppinc.org SPECIFIC LEARNING DISABILITIES (SLD) Definition (1) Specific learning disability is defined as a disorder
How to write a technique essay? A student version. Presenter: Wei-Lun Chao Date: May 17, 2012
How to write a technique essay? A student version Presenter: Wei-Lun Chao Date: May 17, 2012 1 Why this topic? Don t expect others to revise / modify your paper! Everyone has his own writing / thinkingstyle.
Trust, Job Satisfaction, Organizational Commitment, and the Volunteer s Psychological Contract
Trust, Job Satisfaction, Commitment, and the Volunteer s Psychological Contract Becky J. Starnes, Ph.D. Austin Peay State University Clarksville, Tennessee, USA [email protected] Abstract Studies indicate
