Pearson s Text Complexity Measure



Similar documents
Any Town Public Schools Specific School Address, City State ZIP

dm106 TEXT MINING FOR CUSTOMER RELATIONSHIP MANAGEMENT: AN APPROACH BASED ON LATENT SEMANTIC ANALYSIS AND FUZZY CLUSTERING

Reading Instruction and Reading Achievement Among ELL Students

ETS Automated Scoring and NLP Technologies

Automated Scoring for the Assessment of Common Core Standards

Interactive Learning Environments 8(2):

Scoring TELL (Test of English Language Learning)

Readability Visualization for Massive Text Data

The test uses age norms (national) and grade norms (national) to calculate scores and compare students of the same age or grade.

Frequently Asked Questions about Achieve3000 s LevelSet TM Assessment

Reading Specialist (151)

Automated Essay Scoring with the E-rater System Yigal Attali Educational Testing Service

There are basically three options available for overcoming barriers to learning:

Linking DIBELS Oral Reading Fluency with The Lexile Framework for Reading

Discourse Markers in English Writing

The Alignment of Common Core and ACT s College and Career Readiness System. June 2010

Linking Singapore s English Language Syllabus 2010 to The Lexile Framework for Reading: STELLAR as an Example

MINUTE TAKING. All material copyright of Lindsay Wright This pack is for sample purposes only, and not for re-use

How to Improve Reading Comprehension

Grade 4 Reading Comprehension Sample Selections and Items Test Information Document

6.4 Normal Distribution

A GUIDE TO THE SHIFTS IN THE ELA COMMON CORE STATE STANDARDS

What Is Singapore Math?

Summarizing and Paraphrasing

Problem of the Month: Fair Games

How to Paraphrase Reading Materials for Successful EFL Reading Comprehension

LevelSet: Online Lexile Assessment

Gallito 2.0: a Natural Language Processing tool to support Research on Discourse

Information Provided by the Qualitative Reading Inventory-5

Revisiting the readability of management information systems journals again

Alignment of the National Standards for Learning Languages with the Common Core State Standards

An Introduction to. Metrics. used during. Software Development

REVIEW OF HANDBOOK OF AUTOMATED ESSAY EVALUATION: CURRENT APPLICATIONS AND NEW DIRECTIONS

* Lab Experience Course Code: (1) General Studies/Core (3) Major ** Field Experience (2) Enhanced General Studies (4) Professional Education

Depth-of-Knowledge Levels for Four Content Areas Norman L. Webb March 28, Reading (based on Wixson, 1999)

Reading Competencies

TESOL Standards for P-12 ESOL Teacher Education = Unacceptable 2 = Acceptable 3 = Target

Correlational Research. Correlational Research. Stephen E. Brock, Ph.D., NCSP EDS 250. Descriptive Research 1. Correlational Research: Scatter Plots

assessment report ... Academic & Social English for ELL Students: Assessing Both with the Stanford English Language Proficiency Test

Test Bias. As we have seen, psychological tests can be well-conceived and well-constructed, but

READING SPECIALIST STANDARDS

Standard Two: Knowledge of Mathematics: The teacher shall be knowledgeable about mathematics and mathematics instruction.

DEFINING COMPREHENSION

Computer and Information Scientists $105, Computer Systems Engineer. Aeronautical & Aerospace Engineer Compensation Administrator

DESCRIPTIVE STATISTICS. The purpose of statistics is to condense raw data to make it easier to answer specific questions; test hypotheses.

Performance Assessment Task Bikes and Trikes Grade 4. Common Core State Standards Math - Content Standards

Reading Street and English Language Learners

Quick and Easy Access to Measures of Text Complexity Using the Lexile Framework for Reading

Activity 4: Planning an Extension Lesson on Energy Using the 5E Model

Standards for Certification in Early Childhood Education [ ]

Units of Study 9th Grade

Biography-Driven Culturally Responsive Teaching

THE NORTH CAROLINA TESTING PROGRAM ELEMENTARY AND MIDDLE SCHOOL GRADES 3 8

What to Expect on the Compass

ESOL Endorsement Program

Achieving Reading Proficiency for All

LITERACY: READING LANGUAGE ARTS

Measurement Information Model

NW COLORADO BOCES ALTERNATIVE LICENSURE PROGRAM

IMPLEMENTATION NOTE. Validating Risk Rating Systems at IRB Institutions

Glossary of Terms Ability Accommodation Adjusted validity/reliability coefficient Alternate forms Analysis of work Assessment Battery Bias

THE NORTH CAROLINA TESTING PROGRAM ELEMENTARY AND MIDDLE SCHOOL GRADES 3 8

PREPARATION MATERIAL FOR THE GRADUATE RECORD EXAMINATION (GRE)

Appendix B Data Quality Dimensions

Reading in a Foreign Language April 2009, Volume 21, No. 1 ISSN pp

Chapter 6: The Information Function 129. CHAPTER 7 Test Calibration

ETS Policy Statement for Documentation of Intellectual Disabilities in Adolescents and Adults

Module 3: Correlation and Covariance

The Effects of Start Prices on the Performance of the Certainty Equivalent Pricing Policy

A logistic approximation to the cumulative normal distribution

Understanding Types of Assessment Within an RTI Framework

Grade 3 Reading Comprehension Sample Selections and Items Test Information Document

FIELD 002: EARLY CHILDHOOD TEST OBJECTIVES

ACT Aspire TM Exemplar Test Items

Credit Scorecards for SME Finance The Process of Improving Risk Measurement and Management

Responsibilities of Users of Standardized Tests (RUST) (3rd Edition) Prepared by the Association for Assessment in Counseling (AAC)

PPVT -4 Publication Summary Form

Big idea: Key generalization or enduring understanding that students will take with them after the completion ofa learning unit.

Investigating the genetic basis for intelligence

Overview. Essential Questions. Grade 2 Mathematics, Quarter 4, Unit 4.4 Representing and Interpreting Data Using Picture and Bar Graphs

Assessment Policy. 1 Introduction. 2 Background

Technical Report. Overview. Revisions in this Edition. Four-Level Assessment Process

STANDARD. Risk Assessment. Supply Chain Risk Management: A Compilation of Best Practices

DOCTOR OF PHILOSOPHY IN EDUCATIONAL PSYCHOLOGY

Language Arts Literacy Areas of Focus: Grade 6

Spectrophotometry and the Beer-Lambert Law: An Important Analytical Technique in Chemistry

Rubrics for AP Histories. + Historical Thinking Skills

INCREASE YOUR PRODUCTIVITY WITH CELF 4 SOFTWARE! SAMPLE REPORTS. To order, call , or visit our Web site at

Test-Retest Reliability and The Birkman Method Frank R. Larkey & Jennifer L. Knight, 2002

Foundation Paper Supporting Special Education Students with READ 180

Online Assessment Systems

The Effects of Read Naturally on Grade 3 Reading: A Study in the Minneapolis Public Schools

Geometry Solve real life and mathematical problems involving angle measure, area, surface area and volume.

Language Arts Literacy Areas of Focus: Grade 5

CHCEDS305A Support the development of reading skills

Opportunity Document for STEP Literacy Assessment

ELL Considerations for Common Core-Aligned Tasks in English Language Arts

Using LSI for Implementing Document Management Systems Turning unstructured data from a liability to an asset.

Correlation key concepts:

Purposes and Processes of Reading Comprehension

Transcription:

Pearson s Text Complexity Measure White Paper Thomas K. Landauer, Ph.D. May 2011

TEXT COMPLEXITY 1 About Pearson Pearson, the global leader in education and education technology, provides innovative print and digital education materials for prek through college, student information systems and learning management systems, teacher licensure testing, teacher professional development, career certification programs, and testing and assessment products that set the standard for the industry. Pearson s other primary businesses include the Financial Times Group and the Penguin Group. For more information about the Assessment & Information group of Pearson, visit http://www.pearsonassessments.com/. About Pearson s White Papers Pearson s white paper series shares the perspectives of our assessment experts on selected topics of interest to educators, researchers, policy makers and other stakeholders involved in assessment and instruction. Pearson s publications in.pdf format may be obtained at: http://www.pearsonassessments.com/research.

TEXT COMPLEXITY 2 Introduction Pearson s Text Complexity Measure Pearson s Knowledge Technologies group has developed a new measure of text complexity that is fundamentally different from current readability measures such as Flesch Kincaid (Kincaid et al., 1975), Coleman-Liau (Coleman et al., 1975), Dale-Chall (Chall et al., 1995) and Lexiles (Stenner, 1996). These all rely on a small number of surface features such as the average number of words or characters per sentence, the log of the common usage frequency of the words in the sentences, and a few selected syntactic and grammatical features. That is not to say that such combinations of features cannot predict readability usefully, especially when averaged over many sentences and passages and combined with good psychometrics. They do. But what we offer here is a new approach that stands on a broader and deeper foundation that enables greater accuracy of measurement and a wider range of analytic capabilities. As an illustration of the weakness of conventional readability measures that the new approach improves upon, our new metric made it possible to estimate that word frequency accounts on average for only about 34% of the variance in how well a single word is known, varying from 60% for 4 th grade readers to only 10% for 1st year college readers. This finding is an example of a language phenomenon of considerable importance for text complexity measurement, that meanings themselves are learned. Word meanings (and thus passage meanings) depend critically on what and how many other words are already known (Anderson & Freebody, 1981). Thus, word (and paragraph) learning is a synergistic and recursive process. In contrast, the conventional readability metrics ignore these changes, instead treating words as stationary constants that need only be counted. Thus, measuring readability by number of words and their common usage frequency (e.g. Thorndike et al., 1944), and doing it the same way for 4th to 12th graders, misses much of the causality of varying comprehensibility (Landauer, Kireyev & Panaccione, 2011). The text complexity metric overcomes this lack by applying a well established and accurate computational language simulation model, Latent Semantic Analysis (LSA) (Deerwester et al., 1990, Landauer, T. K & Dumais, S. T., 1997), to mimic the way that word (and paragraph) meanings are learned through reading. In doing so, it applies many of the same underlying technologies as Pearson's automatic reading, writing, speaking and listening assessment technologies (see Foltz et al., 1999), all of which match human judgments 90-100% as well as human judgments match each other. The new application of the Word Maturity metric to text complexity uses LSA to trace the evolving similarity of words and passages to each other and to those of proficient adult readers as a function of how much and what has been read. Just as for human learners, the evolution is driven by cumulative encounters with new words in new contexts that convey and teach new meanings (e.g. Anderson & Freebody, 1981). By adding new words and contexts in a natural order and quantity, the simulation closely mimics human learning. Just as in human learning, the way that meaning grows for every word (and for every word for every reader) depends on the unique contexts and orders in which each word and passage is encountered.

TEXT COMPLEXITY 3 To follow the same learning paths that human learners do, the simulation uses a very similar number and distribution of words and passages as those encountered by average American readers. Of course every reader reads different things to some extent. The system handles this by creating a large group of independent randomly chosen sets of passages of the same size from a very large text corpus (a collection of texts representing a selected content area, in this case hundreds of thousands of passages and words read by American students and early adults.) This assures that all the words in almost any given reading can be included by substitution. As explained in the next section, it is the representation of words and passages as mathematical vectors (best thought of here simply as long strings of independent numbers) that enables word and passage meanings to change so as to consistently accommodate all the similarities between each word and each other one. A particularly important result of this process is that the same words are learned as much as four times as fast if they occur when the amount of text already read is large than if they occur very early (Landauer & Dumais, 1997, Kireyev & Landauer, 2011). In summary, the most important advance of the new measure lies in the fact it measures how the meanings of words themselves, and thus of passages, change for learners with increasing exposure to language, even though the spelled "word forms" don't change. These developmental changes are not captured at all by the standard methods; in them, words have the same meanings for readers of all ages and reading experience. Thus, although the new measure is as sensitive to frequency of encounter and length of passages as current measures, its estimates of text complexity include the changes of meaning and interrelationships between words with reading experience. Mathematical foundations of the measure The Word Maturity metric exploits a powerful statistical algorithm, Singular Value Decomposition (see Landauer & Dumais, 1997), to closely simulate the dynamic evolution of word and passage meanings 1 encountered through a student s life. The algorithm first finds the similarity of the vector standing for every word (and every paragraph) to the vector for every other word (and paragraph) in a corpus of words representative of those encountered by early readers. Then more paragraphs and their words are added to the corpus in a natural order, and the algorithm is rerun to make all the similarities fit together again. To make the items fit, their vector representations must change, and thus the meanings of the words and passages also change. To measure the development of word knowledge, the vector similarity of each word to that of an average college educated adult is computed at each step. Thus, as more paragraphs and their words are added, the system tracks how close the meaning of each has become to that of an adult. This creates an individual learning curve for each separate word as a function of how much and what typical adults have read. 2 An important component of the process is the ordering of the paragraphs added. In the research stage of this effort, published Lexile levels were used, starting with a roughly estimated number of paragraphs that a beginning reader would have read, and adding a series of 15 progressive sets of 5,000 paragraph-like passages. A sample of individual word maturation trajectories created in this manner is shown in Figure 1.

TEXT COMPLEXITY 4 Note, however, that sufficiently similar results have been obtained with other sources used to order the entries, including other current readability measures. Note carefully, however, that it is the resulting highly varied shapes of the individual word trajectories that are important, not just the rising number of total words encountered, whose average would closely resemble the single trajectory that a standard measure of word frequency would provide. Figure 1. Five contrasting examples of Word Maturity trajectories. The x-axis in Fig. 1 is the number of paragraphs added to the corpus at each step; the y- axis shows how close the individual word meanings are to their simulated meanings for first year college students on a scale of 0 to 1. The word "dog" starts out almost as mature as it will get, "productivity" is unknown until the reader has encountered a total of around 40,000 paragraphs, the others words fall between with markedly different developmental paths--of which current measures are unaware. The method also makes it possible to investigate in some detail how the meanings of a word change as it matures. This analysis is carried out by comparing a word's current vector to that of related words, which could eventually include, for example, the changing relative contributions of different senses to a word's overall meaning It is important to note that the science of word learning has for decades recognized that word learning depends much more strongly on what the reader already knows than on occurrences of the word itself (Landauer & Dumais, 1997) although that is also obviously necessary, and that LSA, and thus Word Maturity, faithfully model that mechanism as demonstrated in many other applications.

TEXT COMPLEXITY 5 Figure 2. Estimating the maturity of a given word or passage The vertical blue line shows an example average maturity, how near on average the words in a given text have become to their adult meanings as a function of the total number of paragraphs added. Although the maturity of a word or passage could be assessed in other ways, for example by fitting logistic curves to each word trajectory, the current and easiest to understand is by just measuring how far each word in a text is from its adult level. From this both the average level of the text, shown as the vertical blue line in Figure 2, and of single maturity levels by where they intersect that line. Thus, for example, one could average all the growth curve heights or pick the least mature word in the text, the one that intersects nearest the bottom of the graph. For a real example, a recent evaluation under the Gates Foundation sponsored Student Achievement Partners collaboration, used an optimally weighted combination of the average maturity and the values of the ten "most difficult" words, the ones farthest from maturity. Currently, a number of other variables are added to the Text Complexity measure, some identical or similar to the standard ones and others reflecting various other linguistic features--at this point most as constants--to account for other factors than word and paragraph maturity. Adding these features increases overall prediction accuracy modestly. However, importantly, because many of their effects on complexity also change with increasing occurrences in ever-changing contexts, in the future maturities of other features may also be tracked and integrated into the overall complexity measure, presumably increasing accuracy still more.

TEXT COMPLEXITY 6 How well the metric assesses text complexity The new measure has been evaluated against a number of text data sets of varying kinds in terms of both text content and criterion measures. The criterion measures include the reported school grades of texts used in NCLB reading comprehension tests and a widely used Item Response Theory (IRT)-based measure (the Stanford9) based on text comprehension questions. Table 1 below lists the data sets, their accuracy criteria and the number of text passages involved. For these five data sets, text complexity measures were compared against the listed criteria. Description of Text data Summary Street: Randomly chosen whole texts from a large set of readings used in Pearson K-12 tutorial reading passages, the vast majority of which are informative text, all used in Pearson s WriteToLearn reading comprehension and iterative summarization formative tutorial exercises Performance criterion Assigned grade level of readings Number of texts 800 State Readings: A Collection of diverse texts used as items in NCLB reading comprehension tests in 27 state and two national assessments. Assigned grade level of readings 1221 SAT9: Texts used in the well established Stanford9 reading comprehension test Grade levels of use and Rasch IRT-based measures of student comprehension 110 Common Core: Example texts from the Common Core State Standards Appendix B. selected by reading experts for use in varying grade levels NAEP: NAEP reading comprehension passages from NAEP 4,8, and 12 th grade assessments Appropriate grade levels according to experts Grade levels of the test takers 243 39 Table 1. Data, criteria, number of texts

TEXT COMPLEXITY 7 Table 2 lists Pearson correlations with the criterion variables for Pearson s Text Complexity Measure and for two standard text complexity measures, Coleman-Liau and Flesch gradelevel, and an optimal combination the two most general and important factors in readability formulas. Readings Summary Street readings n = 800 State readings n = 1221 SAT9 readings (vs. Grade) n =111 SAT9 readings (vs. IRT) n = 110 Common Core readings n =243 NAEP readings n = 39 Pearson s Text Complexity Measure Flesch Grade Coleman- Liau Word freq and sentence length 0.881 0.637 0.636 0.609 0.862 0.591 0.598 0.604 0.803 0.673 0.732 0.566 0.844 0.678 0.763 0.685 0.736 0.485 0.508 0.545 0.783 0.286 0.217 0.086 Table 2. The first column lists the readings used to evaluate Pearson s measure of text complexity, the second the correlations between them and Pearson s Text Complexity measure. The third and fourth columns show correlations with the texts of two widely used traditional readability measures. The last column is for a simple optimized combination of the average word frequency and number of words per sentence. Thus, on average, the text complexity measure correlated with criterion measures about a third better than the traditional measures. Measures of the method's validity Text complexity is not yet a well-defined concept. For example, almost all of the measures thereof that researchers have put forward have assessed difficulties at the word and sentence levels and evaluated results either by correlation with grade levels in which they are used or placed on an abstract IRT scale. However, the ability to understand complex text is surely not limited to the skills tapped by the meanings of words and passages alone, as important as these may be. Other cognitive abilities must surely play a part. For one example, a goal of continuing Pearson s text complexity research is to include difficulties caused by the order in which information is delivered, which is most obviously critical in math and science texts, but can also make narratives hard to follow even if almost all the words are well known and most of the sentences short. Other examples include logic, reasoning, argument resolution, specialized and quotidian world knowledge, and much of what goes under the rubric of intelligence.

TEXT COMPLEXITY 8 Thus we will need more research on the effects on comprehension caused by a larger range of variables before we can measure validity properly. In the interim, however, to shed a modicum of light on the issue, a concurrent validity study was conducted of the correlation between word maturity on the scale described above and four well validated measures of abilities, each with somewhat different face-value conceptual relations to text complexity. Described below are the test names, brief descriptions of the tests, the number of items in the correlations, and correlations of the tests item difficulties. (Spearman rho rather than Pearson s r is used to increase the comparability of the various scales.) Test Task Description Number of items in test Peabody Picture Vocabulary Test, Fourth Edition (PPVT -4) (Maddux, 1999) Kaufman Brief Intelligence Test, Second Edition (KBIT-2) (Kaufman et al., 1990) Receptive (choose correct picture) and expressive (name picture) vocabulary test for ages 2.6-90 years Verbal and non-verbal multiple choice cognitive ability score for ages 4-90 years Spearman correlation with item difficulties 192 0.74 41 0.82 Kaufman Assessment Battery for Children Expressive Vocabulary (KABC-EV) (Kaufman et al., 1985) Expressive vocabulary scale for ages 3-18 years 35 0.72 Kaufman Assessment Battery for Children Verbal Knowledge (KABC- VK) Updated version of KABC-II 45 0.75 Table 3. Correlations with some of Pearson s well-established tests conceptually related to text complexity. Without being specifically designed to assess "text complexity", these tests share a common concept, the variability of comprehensibility of different words. Thus, since they measure a variety of somewhat different concepts, both to each other and to text complexity, their moderately high correlation with Pearson s text complexity measure seems as great as one could expect (or want) and lends considerable evidence of its validity. In progress and planned extensions and revisions Pearson plans to exploit the capabilities offered by the Word Maturity metric in a variety of ways. The following are in early developmental stages. - A small set of well-differentiated components of text complexity that teachers, parents and other educators will find intuitive. For example, a set of four such components might be word difficulty, sentence difficulty, passage difficulty and overall comprehension difficulty. (We note that results of complex measurements too often take the form of equally complex or arbitrary names, which are avoided.)

TEXT COMPLEXITY 9 - Separate Text Complexity measures based on corpora restricted to particular genres, e.g. informative or narrative, or to different school and college subjects, e.g. biology and psychology. - A measure of text comprehension difficulty due to the flow and order in which information is conveyed, which, for example, will be equally important for science and literature. - Maturity trajectories for separate senses of words. Where from here? We think that the next step should be using the new technology to teach in addition to assessing. Given the enlarged ability to pinpoint the comprehension problems that texts present, it will be easier to choose and target the instruction we deliver. Indeed, we have already developed a companion measure of word and passage text complexity growth that assesses individual reader knowledge of an unlimited number of individual words using the same basic technology, as well as an age-independent vocabulary measure that, like Text Complexity, separately estimates knowledge of every word in an unlimited corpus, but here for any individual student. This will make it possible to teach specifically those words that a reader currently does not know and needs to learn. Endnotes 1 These are "meanings" in the sense that "meaning" is used in LSA, which in turn agree with human judgments of the similarity of meaning of passages 90% as well as two humans agree with each other (Landauer, T. K. (2002) 2 This requires the constraint that there are be always the same number of independent dimensions, that is the 300 real numbers for each word and paragraph. Acknowledgements Evaluations and planning of our methods were greatly aided by a Gates Foundation sponsored discussion and collaboration group directed by Meredith and David Liben from Student Achievement Partners and chaired by David Coleman.

TEXT COMPLEXITY 10 References Anderson R. C., & Freebody, P. (1981). Vocabulary Knowledge. In J. T. Guthrie (Ed.), Comprehension and teaching: Research reviews (pp. 77-117). International Reading Association, Newark DE. Chall, J. S., & Dale, E. (1995). Readability Revisited The New Dale-Chall Readability Formula. Cambridge: Brookline Books. Coleman, M., & Liau, T. L. (1975). A computer readability formula designed for machine scoring. Journal of Applied Psychology, 60, 283-284. Deerwester, S., Dumais, S., Furnas, G.W., Landauer, T. K., & Harshman, R. (1990). Indexing by latent semantic analysis. Journal of the American Society for Information Science, 41(6), 391-407. Foltz, P., Laham, D., & Landauer, T. K. (1999). The intelligent essay assessor: Applications to educational technology. Interactive Multimedia Electronic Journal of Computer- Enhanced Learning (1999). Kaufman, A. S., & Kaufman, N. L. (1985). Kaufman Assessment Battery for Children. Journal of Pediatric Psychology, 10, 257-259. Kaufman, A. S., & Kaufman, N. L. (1990). Kaufman Brief Intelligence Test, Second Edition (KBIT-2). Pearson. Kincaid, J. P., Fishburne, R. P., Rogers, R. L., & Chissom, B. S. (1975). Derivation of New Readability Formulas (Automated Readability Index, Fog Count, and Flesch Reading Ease formula) for Navy Enlisted Personnel. Research Branch Report, 8-75. Chief of Naval Technical Training: Naval Air Station Memphis. Kireyev, K., & Landauer, T. K. (2011). Word Maturity: Computational Modeling of Word Knowledge. 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies (ACL-HLT), 2011. Landauer, T. K., & Dumais, S. (1997). A solution to Plato's problem: The Latent Semantic Analysis theory of the acquisition, induction, and representation of knowledge. Psychological Review, 104, 211-240. Landauer, T. K., Kireyev, K., & Panaccione, C. (2011). Word Maturity: A New Metric for Word Knowledge. Scientific Studies of Reading, 15(1), 92-108. Maddux, C. D. (1999). Peabody Picture Vocabulary Test III (PPVT-III). Diagnostique, 24(1-4), 221-28. Stenner, A. J. (1996). Measuring Reading Comprehension with the Lexile Framework. Fourth North American Conference on Adolescent / Adult Literacy. Washington D.C. Thorndike, E. L., & Lorge, I. (1944). The Teacher's Word Book of 30,000 Words. Teachers College, Columbia University.