Standardized Achievement Tests and English Language Learners: Psychometrics Issues

Size: px
Start display at page:

Download "Standardized Achievement Tests and English Language Learners: Psychometrics Issues"

Transcription

1 EDUCATIONAL ASSESSMENT, 8(3), Copyright 2002, Lawrence Erlbaum Associates, Inc. Standardized Achievement Tests and English Language Learners: Psychometrics Issues Jamal Abedi Graduate School of Education and Information Studies CRESST/University of California, Los Angeles Using existing data from several locations across the U.S., this study examined the impact of students language background on the outcome of achievement tests. The results of the analyses indicated that students assessment results might be confounded by their language background variables. English language learners (ELLs) generally perform lower than non-ell students on reading, science, and math a strong indication of the impact of English language proficiency on assessment. Moreover, the level of impact of language proficiency on assessment of ELL students is greater in the content areas with higher language demand. For example, analyses showed that ELL and non-ell students had the greatest performance differences in the language-related subscales of tests in areas such as reading. The gap between the performance of ELL and non-ell students was smaller in science and virtually nonexistent in the math computation subscale, where language presumably has the least impact on item comprehension. The results of our analyses also indicated that test item responses by ELL students, particularly ELL students at the lower end of the English proficiency spectrum, suffered from low reliability. That is, the language background of students may add another dimension to the assessment outcome that may be a source of measurement error in the assessment for English language learners. Further, the correlation between standardized achievement test scores and external criterion measures was significantly larger for the non-ell students than for the ELL students. Analyses of the structural relationships between individual items and between items and the total test scores showed a major difference between ELL and non-ell students. Structural models for ELL students demonstrated lower statistical Requests for reprints should be sent to Jamal Abedi, UCLA CSE/CRESST, 300 Charles E. Young Drive North, GSE&IS Building, 3rd Floor, Los Angeles, CA jabedi@ cse.ucla.edu

2 232 ABEDI fit. The factor loadings were generally lower for ELL students, and the correlations between the latent content-based variables were also weaker for them. We speculate that language factors may be a source of construct-irrelevant variance in standardized achievement tests (Messick, 1994) and may affect their construct validity. Due to the rapidly changing demographics of the U.S. population, fairness and validity issues in assessment are becoming top priorities in the national agenda. Between 1990 and 1997, the number of U.S. residents not born in the United States increased by 30%, from 19.8 million to 25.8 million (Hakuta & Beatty, 2000). According to the Survey of the States Limited English Proficient Students and Available Educational Programs and Services Summary Report, over 4.4 million limited English proficient 1 students were enrolled in public schools (National Clearinghouse for English Language Acquisition and Language Instruction Educational Programs, 2002). To provide fair assessment and uphold standards on instruction for every child in this country, both federal (e.g., No Child Left Behind Act of 2001) and state legislation now require the inclusion of all students, including ELLs, into large-scale assessments (Abedi, Lord, Hofstetter, & Baker, 2000; Mazzeo, Carlson, Voelkl, & Lutkus, 2000). Such inclusion requirements have prompted new interest in modifying assessments to improve the level of English language learners participation and to enhance validity and equitability of inferences drawn from the assessments themselves. Standardized, high-stakes achievement tests are frequently used for assessment and classification of ELL students, as well as for accountability purposes. They shape instruction and student learning (Linn, 1995). About 40% of districts and schools use achievement tests for assigning ELL students to specific instructional services within a school, and over 70% of districts and schools use achievement tests to reclassify students from ELL status (Zehler, Hopstock, Fleischman, & Greniuk, 1994). However, as most standardized, content-based tests (such as science and math tests) are administered in English and normed on native English-speaking test populations, they may inadvertently function as English language proficiency tests. English language learners may be unfamiliar with the linguistically complex structure of test questions, may not recognize vocabulary terms, or may mistakenly in- 1 The term English language learner (ELL) refers to students who are not native speakers of English and are not as proficient in English as native speakers. A subgroup of these students with a lower level of English proficiency is referred to as limited English proficient (LEP). The term LEP is used primarily by government-funded programs to classify students as well as by the National Assessment of Educational Progress (NAEP) for determining inclusion criteria. In this article we use ELL to refer to students who are not native English speakers and who are not reclassified as fluent in English.

3 STANDARDIZED ACHIEVEMENT TESTS FOR ELLS 233 terpret an item literally (Duran, 1989; Garcia, 1991). They may also perform less well on tests because they read more slowly (Mestre, 1988). Thus, language factors are likely to reduce the validity and reliability of inferences drawn about students content-based knowledge, as stated in the Standards for Educational and Psychological Testing (American Educational Research Association, American Psychological Association, & National Council on Measurement in Education [AERA, APA, & NCME], 1999): For all test takers, any test that employs language is, in part, a measure of their language skills. This is of particular concern for test takers whose first language is not the language of the test. Test use with individuals who have not sufficiently acquired the language of the test may introduce construct-irrelevant components to the testing process. In such instances, test results may not reflect accurately the qualities and competencies intended to be measured. Therefore it is important to consider language background in developing, selecting, and administering tests and in interpreting test performance. (p. 91) As indicated earlier, a major criticism of standardized achievement tests is the exclusion of ELL students from the norming group for these tests. Linn (1995) refers to the issues associated with inclusion of all students as one of the three most notable of the new features of this reform effort. The inclusion of all students in its assessments has also been among the major issues for NAEP (see, e.g., Mazzeo et al., 2000). Navarrette and Gustke (1996) expressed several concerns about the exclusion of ELL students from the norming groups of standardized achievement tests: Not including students from linguistically diverse backgrounds in the norming group, not considering the match or mismatch between a student s cultural and school experiences, and not ensuring for English proficiency have led to justified accusations of bias and unfairness in testing. (p. 2) Findings from a series of studies conducted by the National Center for Research on Evaluation, Standards, and Student Testing (CRESST) on the impact of students language background on their performance indicated that(a) student language background affects students performance in content-based areas such as math and science, and(b) the linguistic complexity of test items may threaten the validity and reliability of achievement tests, particularly for ELL students (see Abedi & Leon, 1999; Abedi, Leon, & Mirocha, 2001; Abedi & Lord, 2001; Abedi et al., 2000). Thus, the literature on the assessment of ELLs clearly suggests that language factors confound the test results of English language learners. However, the literature is not clear on the level of impact that language factors may have on different content areas. That is, would the impact level of language on test outcomes differ across the different content areas? Another issue concerns the impact level of lan-

4 234 ABEDI guage factors on the validity and reliability of content-based assessments for ELLs. Available data from four large school sites in the nation enabled us to explore these issues in greater detail. Research Questions METHODOLOGY 1. Could the performance difference between ELL and non-ell students be partly explained by language factors in the assessment? 2. Could the linguistic complexity of test items as a possible source of measurement error influence the reliability of the assessment? 3. Could the linguistic complexity of test items as a possible source of construct-irrelevant variance influence the validity of the assessment? Data Sources The data for this study were obtained from four locations across the U.S. To assure anonymity, these data sites are referred to as Sites 1 to 4. Item-level standardized achievement test data and background information were obtained for participating students. The background variables included gender, ethnicity, free/reduced price lunch participation, parent education, student ELL status, and students with disabilities (SD) status. Table 1 summarizes some of the main characteristics of the four data sites. As data in Table 1 show, there were similarities and differences among the four data sites. All sites used standardized tests for measuring students achievement in English and other content-based areas, but they differed in the type of test administered. Although all sites had an index of students English language proficiency status (ELL or bilingual status), and they all provided some student background information, they differed in the type of language proficiency index used and the type of background variables provided. These differences limited our ability to perform identical analyses at the different sites for cross-validation purposes. However, there were enough similarities in the data structures at the four different sites to allow for meaningful comparisons. The following is a brief description of each of the four data sites. Site 1. Site 1 is a large urban school district. Data on the Iowa Tests of Basic Skills (ITBS) were obtained for Grades 3 through 8 in No information was available on students ELL status; however, students were categorized as to whether or not they were receiving bilingual services. Among the 36,065 students in the Grade 3 population, 7,270 (about one in five) of these students were receiv-

5 STANDARDIZED ACHIEVEMENT TESTS FOR ELLS 235 TABLE 1 Summary of Characteristics of the Four Data Sites Data Site Site 1 Site 2 Site 3 Site 4 Location type Large urban district Entire state Large urban district Entire state Total number of students, 430,914 5,844,111 approx. 200, ,969 K 12 Percent of ELL, K N/A 6.9 Language designation Bilingual/ ELL/non-ELL ELL/non-ELL ELL/non-ELL nonbilingual Grades data available , 11 3, 6, 8, 10 Achievement tests used ITBS SAT9 SAT9 SAT9 Language proficiency N/A LAS N/A LAS tests used Accommodation data N/A N/A N/A N/A Years data available Note. ELL = English language learner; ITBS = Iowa Tests of Basic Skills; SAT9 = Stanford Achievement Test, 9th edition; LAS = Language Assessment Scales; N/A = not available. ing bilingual services. In Grade 6 there were 28,313 students in the population, with 3,341 (11.8%) receiving bilingual services. In Grade 8 there were 25,406 students in the population, and 2,306 fewer than one in ten (9.1%) were receiving bilingual services. Site 2. Site 2 is a state with a very large number of ELL students. There were a total of 414,169 students in the Grade 2 population of the state, and 125,109 (30.2%) of these students were ELLs. In Grade 7 there were 349,581 students, of whom 73,993 (21.2%) were ELL students. In Grade 9 there were 309,930 students, and 57,991 (18.7%) were ELL students. Stanford Achievement Test, 9th edition (Stanford 9) test data were obtained for all students in Grades 2 to 11 who were enrolled in the statewide public schools for the academic year. Site 3. Site 3 is an urban school district. Stanford 9 test data were available for all students in Grades 10 and 11 for the academic year. Accommodation data were obtained from the district and included both the type and number of accommodations received. There were 12,919 students in the Grade 10 population, and 431 (3.3%) of these students were ELLs. In Grade 11 there were 9,803 students in the population, of whom 339 (3.5%) were ELL students. Site 4. Site 4 is a state with a large number of ELL students. Access was provided to Stanford 9 summary test data for all students in Grades 3, 6, 8, and 10 who were enrolled in the state s public schools for the academic year. There were a total of 13,810 students in the Grade 3 population of the state, and 1,065 (7.7%) of these students were ELLs. In Grade 6 there were 12,998 students in the

6 236 ABEDI population, of whom 813 (6.3%) were ELL students. In Grade 8 there were 12,400 students, and 807 (6.5%) were ELL students. Design and Statistical Approach To provide responses to the research questions outlined previously, data from the four sites were analyzed. There were some differences in the type and format of the data across the four sites; however, similar analyses were performed on the four data sets, and the four sites were used as cross-validation samples. The main hypothesis of this study focused on the possible impact of students language background on their performance. Therefore, the focus of the analyses was on the comparison between the level of performance of ELL and non-ell students. However, to develop an understanding about the role of other contributing factors in the assessment of ELL students, comparisons were also made between students with respect to other background variables, such as family income and parent education. Students mean normal-curve equivalent (NCE) scores on different subscales of standardized achievement tests were compared across subgroups using analysis of variance and t tests in a multiple-comparison framework. To examine the impact of language on the reliability of tests and on the level of measurement error, internal consistency coefficients were computed for different tests across categories by students ELL status and other background variables, such as family income and parent education. This approach was based on the assumption that test items within each strand or subscale were measuring the same construct; that is, they were unidimensional (see Cortina, 1993). To study the impact of language factors on the validity of tests, the structural equation approach was used (Bollen, 1989). Through the application of multiple-group factor analyses, the internal structural relationship of test items and the relationships of test scores with external criteria were examined. It must be noted at this point that in some of our data sites, we had access to the data for the entire student population. Therefore, application of inferential statistical techniques was not necessary. However, to be consistent with the analyses for the other sites that provided data for subgroups of the population, as well as the entire population, we report statistical analyses for all four data sites. Findings from these analyses are presented next. RESULTS Three main research questions guided the analyses and reporting of the results. These questions were based on (a) issues concerning content-based performance differences between ELLs and non-ells due to language factors, (b) the impact of language factors on reliability of the tests, and (c) the impact of language factors on the validity of the tests.

7 STANDARDIZED ACHIEVEMENT TESTS FOR ELLS 237 The results of analyses are reported in three sections: (a) performance differences between ELL and non-ell students, (b) impact of language factors on reliability, and (c) validity. Performance Differences Between ELL and Non-ELL Students Due to Possible Impact of Language Factors The results of analyses of data from the four sites consistently suggested that ELL students performed substantially lower than non-ell students. However, the performance gap between ELL and non-ell students was not the same across the content areas. In content areas with a higher level of language demand (e.g., reading and writing), the performance gap between ELL and non-ell students was the highest, whereas in content areas with less language demand (e.g., math and science), the performance gap was much smaller and in some cases was almost nonexistent (e.g., math computation). To present a picture of the performance gap trend between ELL and non-ell students, we report the descriptive statistics on the site with the largest ELL population for two grades, an early elementary grade and a secondary school grade. To conserve space, we have summarized the results of the descriptive analyses for the other three sites. Table 2 presents the number and percentage of students in Grades 2 and 9 in Site 2 who took the Stanford 9 tests in reading, math, and science, by student ELL and disability status. TABLE 2 Site 2 Grades 2 and 9 Stanford 9 Frequencies for Students Students With a Normal Curve Equivalent Score All Students Reading Math Science n % n % n % n % Grade 2 SD only 17, , , NA NA ELL only 120, , , NA NA ELL and SD 4, , , NA NA Non-ELL/Non-SD 271, , , NA NA All students 414, , , NA NA Grade 9 SD only 18, , , , ELL only 53, , , , ELL and SD 4, , , , Non-ELL/Non-SD 233, , , , All students 309, , , , Note. SD = students with disabilities; ELL = English language learner.

8 238 ABEDI As data in Table 2 show, over 29% of all Grade 2 students at Site 2 who participated in Stanford 9 testing were ELL students. This percentage point (29.1%) may not represent the actual percentage of ELL students at Site 2 because some ELL students did not participate in the assessment due to language barriers. The percentage of ELL students who participated in the Stanford 9 testing was 17.2% for Grade 9, which was substantially lower than for Grade 2 (29.1%). There were slight differences between percentages of ELL students across the different content areas in this site. The large number of ELL students in this site provided a unique opportunity to perform analyses at the subgroup level to examine the impact of students background variables on academic achievement. Table 3 presents means, standard deviations, and numbers of students in reading, math, and science for Stanford 9 test scores by subgroups of students. In addition to data by students ELL status, we included subgroup data by school lunch program (a proxy for family income) and parent education, which were highly confounded with students ELL status. In general, the results of analyses reported in Table 3 indicate that: ELL students performed substantially lower than non-ell students, particularly in content areas with more language demand such as reading. For example, the mean reading score for ELL students in Grade 2 was 31.6 (SD = 15.9, N = 97,862) compared with a mean of 49.3 (SD = 19.7, N = 252,696) for non-ell students. This difference was significant beyond the.01 nominal level (t = 250.6, df = 350,556, p <.001). 2 The performance gap between ELL and non-ell students was smaller in the lower grades. For example, there was a 17.7-point difference between ELL and non-ell students in Grade 2 reading mean scores as compared with a 22-point difference for students in Grade 9. The performance gap between ELL and non-ell students decreased when the level of language demand of test items decreased. For example, for Grade 9 students, the performance gap between ELL and non-ell students in reading was 22 points, as compared to 15.4 points in math. The results of analyses also show that other background variables affect test performance. Background variables such as family income (as measured by participation in free/reduced price lunch program) and parent education may not be directly related to students ELL status, but are confounded with it. 2 Since we are working with the population of students in this site, no statistical comparison is needed. Even a minor difference would be real. However, following tradition, we conducted some statistical significance testing. To control for multiple comparisons, we used the Benjamini Hochberg False Discovery Rate procedure. For a description of this procedure see Benjamini and Hochberg (1994).

9 TABLE 3 Site 2 Grade 2 Stanford 9 Subsection Scores Grade 2 Grade 9 Subgroup/Grade Reading Math Science Reading Math Science ELL status ELL M NA SD NA N 97, ,519 NA 48,801 50,666 50,179 Non-ELL M NA SD NA N 252, ,397 NA 224, , ,457 School lunch Free/reduced price M NA SD NA N 106, ,461 NA 56,499 57,961 57,553 No free/reduced price M NA SD NA N 304, ,409 NA 338, , ,663 Parent education Not high school grad M NA SD NA N 54,855 63,960 NA 69,934 71,697 71,183 High school graduate M NA SD NA N 93, ,276 NA 71,986 73,187 72,810 Some college M NA SD NA N 66,530 70,381 NA 70,364 70,971 70,687 College graduate M NA SD NA N 54,391 56,451 NA 87,654 88,241 87,956 Post graduate studies M NA SD NA N 25,571 26,367 NA 34,987 35,087 35,022 Note. ELL = English languge learner. 239

10 240 ABEDI Students who did not participate in the free/reduced price lunch program had higher mean scores in all subject areas than those who did participate in the program. For example, the average NCE score for reading for Grade 2 students who participated in the free/reduced price lunch program was 35.4 (SD = 17.5, N = 106,999), as compared with an average score of 47.0 (SD = 20.6, N = 304,092) for those who did not participate in the program. The difference was statistically significant (t = 177.8, df = 411,089, p <.001). For Grade 9 students participating in the free/reduced price lunch program, the average NCE score for reading was 32.0 (SD = 16.2, N = 56,499), as compared with an average of 42.6 (SD = 19.7, N = 338,285) for those who did not participate in the program. The difference between the performances of the two groups was statistically significant (t = 139.2, df = 394,755, p <.001). The results also indicate that parent education has a substantial impact on the Stanford 9 test scores. For example, the average NCE score for reading for Grade 2 students of parents with low education (not high school graduate) was 30.1 (SD = 15.3, N = 54,855), as compared with an average of 62.1 (SD = 18.7, N = 25,571) for students of parents with high education (post graduate education). This difference was statistically significant (t = 238.8, df = 80,424, p <.001). For Grade 9 students, the average NCE score for reading for the low parent education category was 29.2 (SD = 15.0, N = 69,934). For students with parents in the high education category, the average was 57.6 (SD = 19.6, N = 34,987). This difference was statistically significant (t = 238.4, df = 104,919, p <.001). The results of our analyses also suggest that family income and parent education are confounded with students ELL status. Table 4 presents frequencies and percentages of family income (free/reduced price lunch program) and parent education by ELL status. TABLE 4 Site 2 Free/Reduced Price Lunch Status and Parent Education by ELL Status No Free/Reduced Price Lunch Parent Education Not HS Grad Post Grad Total Free/Reduced Price Lunch Parent Education Not HS Grad Post Grad Total Grand Total Non-ELL 20,738 22,410 43,148 9, ,743 53, % 28.2% 54.4% 12.3% 1.3% 13.6% 68.0% ELL 15, ,360 8, ,006 25, % 1.2% 20.6% 10.9% 0.5% 11.4% 32.0% Total 36,122 23,386 59,508 18,411 1,338 19,749 79, % 29.5% 75.1% 23.2% 1.7% 24.9% 100.0% Note. Percentages reported are based on the total number of students. Not HS Grad = not high school graduate; Post Grad = post graduate education; ELL = English language learner.

11 STANDARDIZED ACHIEVEMENT TESTS FOR ELLS 241 A chi-square of 12,096.72, which shows confounding of these variables, was significant beyond the.01 nominal level (χ 2 = 12,096.72, p <.001). A square contingency coefficient of.132 presents a rough estimate of the proportion of common variance (or confounding) among the three variables. These results suggest that a greater percentage of ELL students are from families with lower income and lower education. For example, 95% of ELL students had parents with low education, whereas only 57% of non-ell students had parents with low education. Thirty-six percent of all ELL students participated in the free/reduced price lunch program as compared with only 20% of non-ell students. However, the results of analyses in this study suggest that among these background variables, language factors show a greater impact on assessment, much greater than family income or parent education. To make a more clear comparison between the performance of subgroups of students (e.g., by ELL status, family income, and parent education) in different content areas, a Disparity Index (DI) was computed. For example, to compute DI by students ELL status, the mean score for ELL students was subtracted from the mean for non-ell students. The difference was then divided by the mean for ELL students, and the result was multiplied by 100. Table 5 shows the DI by student ELL status, as well as by school lunch program and parent education, for Grades 2 and 7, for Site 2, in four content areas. 3 Similar results were obtained for other grades (see Abedi & Leon, 1999). As the data in Table 5 show, the average DI for ELL status over reading, math, language, and spelling for Grade 2 was 48.1 (i.e., over all four subject areas, non-ell students outperformed ELL students by 48.1%). For Grade 7, the DI was We also computed DI by school lunch program and parent education. The DI for school lunch program for Grade 2 students was That is, students who did not participate in the school lunch program outperformed students who participated in the program by 29.6%. For Grade 7, the DI was We also compared the performance of students with the lowest level of parent education with students and the highest level of parent education. The DI for parent education for Grade 2 was 99.3; that is, children of parents with the highest level of education (post graduate education) outperformed children of parents with lower levels of education ( no education or elementary level education ) by 99.3%. The DI for Grade 7 by parent education was By comparing the math DI with the DIs of the language-related subscales (reading, language, and spelling), we can see the impact of language on students performance. The DIs for all categories (ELL status, school lunch, and parent education) were smaller for math and larger for reading. For example, for Grade 2 students, the DI (non-ell vs. ELL) was 55.8 in reading (non-ell students outperformed 3 We have presented the results for Grade 7 rather than Grade 9 to cover a larger range of students in different grades.

12 242 ABEDI TABLE 5 Site 2 Grades 2 and 7 Disparity Indexes (DI) by ELL Status, Free/Reduced Price Lunch, and Parent Education DI Reading Math Language Spelling Average Difference Grade 2 ELL/Non-ELL Free/reduced lunch Parent education Grade 7 ELL/Non-ELL Free/reduced lunch Parent education Note. ELL = English language learner. ELL students by 55.8%), 60.2 in language, and 42.8 in spelling, as compared with a DI of 33.5 in math. For Grade 7 students, the DIs (non-ell vs. ELL) were 96.9 for reading, 70.7 for language, and 81.1 for spelling, compared to 50.4 for math. The DIs for school lunch program (nonparticipant vs. participant in free/reduced price lunch) for Grade 2 students were 32.7 for reading, 35.2 for language, and 25.3 for spelling, as compared with 25.1 for math. However, the difference between DIs for math and language-related subscales was largest across the ELL categories. In Table 5, we included these DI differences underthe Difference column.thedidifferencebyellstatus 4 was19.4forgrade 2 and 32.5 for Grade 7, as compared with the school lunch program DI differences of 6.0 and 7.7, respectively, and the parent-education DI differences of 15.8 and 9.8, respectively. Once again, these data suggest that language factors may have a more profound impact on the assessment outcome than other background variables, such as family income and parent education, particularly for ELL students. To shed light on the impact of language factors on assessment, analyses by math subscales were conducted and will be presented. Standardized achievement tests such as the Stanford 9 and ITBS include in their tests different math subscales that have varying degrees of language demand. These subscales range from testing math analytical skills, concepts and estimation, and problem solving with a relatively higher level of language demand to testing math computation with a minimal level of language demand. If the hypothesis concerning the impact of language on content-based performance is tenable, then the performance difference between ELL and non-ell students should be at the minimum level in content-based tests with a minimal level of language demand, such as math computation. This was exactly what the results of our analyses showed. 4 This DI difference was computed as follows: The three language-related DIs (reading, language, and spelling) were averaged. The result was then subtracted from the DI for math.

13 STANDARDIZED ACHIEVEMENT TESTS FOR ELLS 243 TABLE 6 Site 1 Disparity Indexes of Nonbilingual Over Bilingual Students on Math and Reading Test Level Primary Grade Math Concepts and Estimation Math Problem Solving and Data Interpretation Math Computation Reading Average of all levels/grades Data obtained from some of the sites in this study included different subscale scores including math computation. Table 6 presents the DIs for bilingual students compared with nonbilingual 5 students by level and grade for math concepts and estimation, math problem solving, math computation, and reading in Site 1. The results of the DI analyses shown in Table 6 present several interesting patterns: 1. The DIs indicated that the nonbilingual students generally outperformed the bilingual students. However, the magnitude of the DIs depends, to a greater extent, on the level of language demand of the test items. The DI for test items with less language demand was smaller than for other items. For example, in Grade 3, bilingual students performed better on math computation, which has the lowest level of language demand. 2. Major differences between bilingual and nonbilingual students were found for students in Grades 3 and above. There seemed to be a positive relationship between the mean score differences and grade level, in that the difference increased as the grade level increased, up to Grade 5. Starting with Grade 6, the DI was still positive, but the rate of increase was not as systematic as before. For example, in Grade 3, nonbilingual over bilingual students had DIs of 5.3 in math concepts and estimation, 11.1 in math problem solving and data interpretation, 3.1 in math computation, and 23.4 in reading. In Grade 4, these indexes increased to 26.9 for math concepts and estimation, 19.3 for math problem solving and data interpretation, 6.9 for math computation, and 30.1 for reading. The indexes further increased in Grade 5 to 36.5 for math concepts and estimation, 32.7 for math problem solving and data interpretation, 12.6 for math computation, and 41.1 for reading. 5 This site did not provide information on students ELL status. Instead, we used students bilingual status as a proxy for ELL status.

14 244 ABEDI 3. The largest gap between bilingual and nonbilingual students was in reading. The next largest gaps were in the content areas that appear to have more language demand. For example, the math concepts and estimation and the math problem solving and data interpretation subsections seem to have more language demand than the math computation subsection. Correspondingly, the DIs were higher for those subsections. The average DI for Grades 3 through 8 was 27.7 for math concepts and estimation. That is, the mean of the nonbilingual group in math concepts and estimation was 27.7% higher than the bilingual group mean. A similar trend was observed in math problem solving and data interpretation; the average DI for this subsection was The average DI for math computation, however, was 9.0, which was substantially lower than the corresponding DIs for the other two math subsections. These results were consistent across the different data sites. Table 7 reports the DIs, non-ell versus ELL students, for reading, math total, and the math calculation and math analytical subscales for Grades 3, 6, and 8 at Site 4. Once again, the results of analyses clearly suggest the impact of language factors on students performance, particularly in areas with more language demand. For example, in reading, ELL students had the largest performance gap with non-ell students. The average DI for reading across the three grades was 86.7, as compared with the average performance gap of 33.4 for math total. Among the math subscale scores, those with less language demand showed a smaller performance gap. The average DI was 41.0 for math analytical and 20.1 for math calculation. The math calculation DI was substantially less than the DI for reading (86.7) and for math analytical (41.0). However, it must be indicated at this point that language demand and cognitive complexity of test items may also be confounded. That is, items in the math calculation subscale may not only have less language demand, but they may also be less cognitively demanding than other math subscales, such as math problem solving. This is a caveat in our discussion on the impact of language on content-based assessments. TABLE 7 Site 4 Disparity Indexes of Non-ELL Versus ELL Students in Reading and Subscales of Math Disparity Index Grade Reading Math Total Math Calculation Math Analytical Average over the three grades Note. ELL = English language learner.

15 STANDARDIZED ACHIEVEMENT TESTS FOR ELLS 245 Possible Impact of Language Factors on Reliability of Assessments In classical test theory, reliability is defined as the ratio of the true-score variance (σ 2 T) to observed-score variance (σ 2 X) (Allen & Yen, 1979). This observed score variance (σ 2 X) is the sum of two components, the true-score variance (σ 2 T) and the error variance (σ 2 E). In a perfectly reliable test, the error variance (σ 2 E) would be zero; therefore, the true-score variance (σ 2 T) would be equal to the observed-score variance. However, in measurement with human participants there is always an error component, whether large or small, which is referred to in classical test theory as the measurement error (see Allen & Yen, 1979; Linn & Gronlund, 1995; Salvia & Ysseldyke, 1998). Appropriate evaluation of the measurement error is important in any type of assessment, whether in a traditional, multiple-choice approach or in performance-based assessments (Linn, 1995; see also AERA, APA, & NCME, 1999). Many different sources (e.g., occasion, task, test administration conditions) may contribute to measurement error in traditional, closed-ended assessment instruments. In addition to these sources, the reliability of performance assessment measures suffers from yet another source of measurement error, variation in scoring of open-ended items. More important, in the assessment of ELL students, language factors may be another serious source of measurement error, due to unnecessary linguistic complexity in content-based areas. In the classical approach to estimating reliability of assessment tools, the level of contribution of different sources to measurement error may be indeterminable. Through the generalizability approach, one would be able to determine the extent of the variance each individual source contributes (such as occasion, tasks, items, scorer, and language factors) to the overall measurement error (see Cronbach, Gleser, Nanda, & Rajaratnam, 1972; Shavelson & Webb, 1991). To estimate reliability of the standardized achievement tests used in this study and to investigate their measurement error, we considered different approaches. Since parallel forms or test retest data were not available, we decided to use an internal consistency approach. The main limitation with the internal consistency approach, however, is the assumption of unidimensionality. For example, the literature has indicated that the alpha coefficient, which is a measure of internal consistency, is extremely sensitive to multidimensionality of test items (see, e.g., Abedi, 1996; Cortina, 1993). However, because the test items within each content area are assumed to measure the same construct, we believe this approach may be appropriate for estimating reliability of the achievement tests used in this study. Because different data sites used different tests, and because within the individual sites, different test forms were used in different grades, these analyses were performed separately for each site and each grade. Within each grade, we con-

16 246 ABEDI ducted the internal consistency analyses separately for ELL and non-ell students. The results obtained from analyses at different sites were consistent. Due to space limitations, only the results from Site 2, the site with the largest number of students, are presented. A complete report of the results of analyses can be found in Abedi et al. (2001). Language (and perhaps other variables, such as socioeconomic status and opportunity to learn) may cause a restriction of range in the score distribution that may result in lower internal consistency. Table 8 presents reliability (internal consistency) coefficients for the Stanford 9 data for Grade 2 students in Site 2. As the data in Table 8 show, non-ell students had higher coefficients than the ELL students. There was also a slight difference between the alpha coefficients across the free/reduced price lunch categories. Nonparticipants in the free/reduced price lunch program had slightly higher alphas than the participating students. For example, the average reliability for the reading subscale for the nonparticipant group was.913, as compared with an average reliability of.893 for the participant group (a difference of.021), and for ELL students the average reliability was.856, as compared with an average reliability of.914 for non-ell students, a difference of.058 (non-ells refers to English only). The results of our analyses, which are consistent across the different sites, indicate that the difference in internal consistency coefficients between ELL and non-ell students is significantly larger than the difference between these coefficients across the free/reduced price lunch and parent education categories. Table 9 presents the reliability (internal consistency) coefficients for Grade 9 students. Comparing the internal consistency coefficients for Grade 9 students with those for Grade 2 students (reported in Table 8) once again revealed that reliability coefficients for ELL students were lower than the coefficients for non-ell students. This was particularly true for students in higher grades, where language has more impact on performance. In both Grade 3 and Grade 9, reliabilities were lower for ELL students. However, in Grade 9, the difference between reliability coefficients for ELL and non-ell students was larger. For example, for Grade 2, the difference between reliability coefficients for ELL and non-ell students was.058 in reading,.013 in math, and.062 in language, as compared with the ELL/non-ELL reliability difference of.109 for reading,.096 for math, and.120 for language in Grade 9. The difference between the overall reliability coefficient of ELL students and English-only students for Grade 9 was.167, which was substantially higher than the respective difference of.043 in Grade 2. Thus, the reliability gap between ELL and non-ell students increases with increase in the grade level. This may be due to the use of more complex language structures in higher grades. The results of these analyses strongly suggest that students language background factors have a profound effect on their assessment outcomes, above and beyond other background characteristics such as family income and parent education.

17 TABLE 8 Site 2 Grade 2 Stanford 9 Subscale Reliabilities Non-ELL Students: Free Lunch Participation Subscale (No. of Items) Yes No English Only FEP RFEP ELL Reading N = 209,262 N = 58,485 N = 34,505 N = 29,771 N = 3,471 N = 101,399 Word study (48) Vocabulary (30) Reading comp. (30) Average reliability Math N = 220,971 N = 63,146 N = 249,000 N = 31,444 N = 3,673 N = 118,740 Problem solving (45) Procedures (28) Average reliability Language N = 218,003 N = 62,028 N = 245,384 N = 31,035 N = 3,612 N = 111,752 Total (44) Note. ELL = English language learner; FEP = fluent English proficient; RFEP = redesignated fluent English proficient. 247

18

19 STANDARDIZED ACHIEVEMENT TESTS FOR ELLS 249 Validity Research has indicated that complex language in content-based assessments for nonnative speakers of English may reduce the validity and reliability of inferences drawn about students content-based knowledge. For example, results from earlier CRESST language background studies (Abedi & Lord, 2001; Abedi, Lord, & Hofstetter, 1998; Abedi et al., 2000; Abedi, Lord, & Plummer, 1997) provided support for a strong link between language factors and content-based performance. The linguistic factors in content-based assessments (such as math and science) may be considered a source of construct-irrelevant variance because they are not conceptually related to the content being assessed (Messick, 1994): With respect to distortion of task performance, some aspects of the task may require skills or other attributes having nothing to do with the focal constructs in question, so that deficiencies in the construct-irrelevant skills might prevent some students from demonstrating the focal competencies. (p. 14) To examine the impact of students language background on the validity of standardized achievement tests, analyses were performed to compare criterion validity coefficients for ELL and non-ell students and to examine differences between the structural relationship of ELL and non-ell groups. Linguistic complexity of test items, as a possible source of construct-irrelevant variance, may be a threat to the validity of achievement tests, because it could be a source of measurement error in estimating the reliability of the tests. Intercorrelation between individual test items, the correlation between items and total test score (the internal validity coefficient), and the correlation between item score and total test score with the external criteria (the students other achievement data) were computed. A significant difference across the ELL categories in the relationships between test items, between individual items and total test scores (internal validity), and between overall test scores and external criteria may be indicative of the impact of language on the validity of tests. Since language factors should not influence the performance of non-ell students, these relationships may be stronger for non-ell students. To examine the hypothesis regarding differences between ELL and non-ell students on the structural relationship of the test items, a series of structural equation models were created for Site 2 and Site 3 data. Fit indexes were compared across ELL and non-ell groups. The results generally indicated that the relationships between individual items, items with the total test score, and items with the external criteria were higher for non-ell students than for ELL students. In creating the structural models, test items in each content area (e.g., reading, science, and math) were grouped as parcels. Figure 1 presents item par-

20 250 ABEDI FIGURE 1 Latent variable model for reading, science, and math. cels and latent variables for reading, math, and science for Site 2. As Figure 1 shows, the 54 reading items were grouped into four parcels. Each parcel was constructed to systematically contain items with three degrees of item difficulty: easy, difficult, and moderately difficult items (for a description of the item parcels and ways to create them, see Catell & Burdsal, 1975). A reading latent variable was constructed based on these four parcels. Similarly, item parcels and latent variables for math and science were created from the 48 math items and 40 science items by the same process. The correlations between the reading, math and science latent variables were estimated. Models were tested on randomly selected subsamples to demonstrate the cross-validation of the results. Table 10 shows the results of the structural models for Grade 9 at Site 2. Correlations of item parcels with the latent factors were consistently lower for ELL students than they were for non-ell students. This finding was true for all parcels regardless of which grade or which sample of the population was tested. For example, for Grade 9 ELL students, the correlations for the four reading parcels ranged from a low of.719 to a high of.779 across the two samples (see Table 10). In comparison, for non-ell students, the correlations for the four reading parcels ranged from a low of.832 to a high of.858 across the two samples. The item parcel correlations were also larger for non-ell students than for ELL students in math and science. Again, these results were consistent across the different samples.

21 STANDARDIZED ACHIEVEMENT TESTS FOR ELLS 251 TABLE 10 Site 2 Grade 9 Stanford 9 Reading, Math, and Science Structural Modeling Results (df = 51) Non-ELL (N = 22,782) ELL (N = 4,872) Sample 1 Sample 2 Sample 1 Sample 2 Factor loadings Reading comprehension Parcel Parcel Parcel Parcel Math factor Parcel Parcel Parcel Parcel Science factor Parcel Parcel Parcel Parcel Factor correlation Reading vs. Math Reading vs. Science Science vs. Math Goodness of fit Chi-square NFI NNFI CFI Note. There was significant invariance for all constraints tested with the multiple group model (Non-ELL/ELL). ELL = English language learner; NFI = Normed Fit Index; NNFI = Nonnormed Fit Index; CFI = Comparative Fit Index. The correlations between the latent factors were also larger for non-ell students than they were for ELL students. This gap in latent factor correlations between non-ell and ELL students was especially large when there was more language demand. For example, in Sample 1 for Grade 9, the correlation between latent factors for math and reading for non-ell students was.782 compared to just.645 for ELL students. When comparing the latent factor correlations between reading and science from the same population, the correlation was still larger for non-ell students (.837) than for ELL students (.806), but the gap between the correlations was smaller. This was likely due to language demand differences. Multiple group structural models were used to test whether the differences between

22 252 ABEDI non-ell and ELL students mentioned previously were significant. There were significant differences for all constraints tested at the p <.05 level. The results of simple structure confirmatory factor analyses also showed differences on factor loadings and factor correlations between the ELL and non-ell groups for the Site 3 data. The hypotheses of invariance of factor loadings and factor correlations between the ELL and non-ell groups were tested. Specifically, we tested the following null hypotheses: Correlations between parcel scores and a reading latent variable are the same for the ELL and non-ell groups. Correlations between parcel scores and a science latent variable are the same for the ELL and non-ell groups. Correlations between parcel scores and a math latent variable are the same for the ELL and non-ell groups. Correlations between content-based latent variables are the same for the ELL and non-ell groups. Table 11 summarizes the results of structural models for reading and math tests for Site 3 students in Grade 10. Table 11 includes fit indexes for the ELL and non-ell groups, correlations between parcel scores and content-based latent variables (factor loadings), and correlations between latent variables. Hypotheses regarding the invariance of factor loadings and factor correlations between ELL and non-ell groups were tested. Significant differences between the ELL and non-ell groups at or below.05 nominal levels were identified. These differences are indicated by an asterisk next to each of the constraints. There were several significant differences between the ELL and non-ell groups on the correlations between parcel scores and latent variables. For example, on the math subscale, differences in factor loadings between the ELL and non-ell groups on Parcels 2 and 3 were significant. Table 11 also shows a significant difference between the ELL and non-ell groups on the correlation between reading and math latent variables. These results indicate that: 1. Findings from the two cross-validation samples are very similar and provide evidence on the consistency of the results. 2. Structural models show a better fit for non-ell than for ELL students. 3. Correlations between parcel scores and the content-based latent variables are generally lower for ELL students. 4. Correlations between the content-based latent variables are lower for ELL students. The results suggest that language factors may be a source of construct-irrelevant variance in the assessment of ELL students.

The No Child Left Behind Act (NCLB; Public Law No.

The No Child Left Behind Act (NCLB; Public Law No. The No Child Left Behind Act and English Language Learners: Assessment and Accountability Issues by Jamal Abedi There are major issues involved with the disaggregated No Child Left Behind (NCLB) Act in

More information

Research-Supported Accommodation for English Language Learners in NAEP. CSE Technical Report 586

Research-Supported Accommodation for English Language Learners in NAEP. CSE Technical Report 586 Research-Supported Accommodation for English Language Learners in NAEP CSE Technical Report 586 Jamal Abedi, Mary Courtney, and Seth Leon CRESST/University of California, Los Angeles January 2003 Center

More information

Achievement of Children Identified with Special Needs in Two-way Spanish/English Immersion Programs

Achievement of Children Identified with Special Needs in Two-way Spanish/English Immersion Programs The Bridge: From Research to Practice Achievement of Children Identified with Special Needs in Two-way Spanish/English Immersion Programs Dr. Marjorie L. Myers, Principal, Key School - Escuela Key, Arlington,

More information

Classification System for English Language Learners: Issues and Recommendations

Classification System for English Language Learners: Issues and Recommendations Classification System for English Language Learners: Issues and Recommendations Jamal Abedi, University of California, Davis High-stakes decisions for the instruction and assessment of English language

More information

A Performance Comparison of Native and Non-native Speakers of English on an English Language Proficiency Test ...

A Performance Comparison of Native and Non-native Speakers of English on an English Language Proficiency Test ... technical report A Performance Comparison of Native and Non-native Speakers of English on an English Language Proficiency Test.......... Agnes Stephenson, Ph.D. Hong Jiao Nathan Wall This paper was presented

More information

2013 New Jersey Alternate Proficiency Assessment. Executive Summary

2013 New Jersey Alternate Proficiency Assessment. Executive Summary 2013 New Jersey Alternate Proficiency Assessment Executive Summary The New Jersey Alternate Proficiency Assessment (APA) is a portfolio assessment designed to measure progress toward achieving New Jersey

More information

ACT Research Explains New ACT Test Writing Scores and Their Relationship to Other Test Scores

ACT Research Explains New ACT Test Writing Scores and Their Relationship to Other Test Scores ACT Research Explains New ACT Test Writing Scores and Their Relationship to Other Test Scores Wayne J. Camara, Dongmei Li, Deborah J. Harris, Benjamin Andrews, Qing Yi, and Yong He ACT Research Explains

More information

EXPL ORI NG THE RE LAT ION SHI P ATTE NDA NCE AND COG NIT IVE GAIN S OF LA S BES T STU DEN TS

EXPL ORI NG THE RE LAT ION SHI P ATTE NDA NCE AND COG NIT IVE GAIN S OF LA S BES T STU DEN TS CRESST REPORT 757 Denise Huang Seth Leon Aletha M. Harven Deborah La Torre Sima Mostafavi EXPL ORI NG THE RE LAT ION SHI P BETW EEN LA S BES T PR OGR AM ATTE NDA NCE AND COG NIT IVE GAIN S OF LA S BES

More information

Glossary of Terms Ability Accommodation Adjusted validity/reliability coefficient Alternate forms Analysis of work Assessment Battery Bias

Glossary of Terms Ability Accommodation Adjusted validity/reliability coefficient Alternate forms Analysis of work Assessment Battery Bias Glossary of Terms Ability A defined domain of cognitive, perceptual, psychomotor, or physical functioning. Accommodation A change in the content, format, and/or administration of a selection procedure

More information

CHARTER SCHOOL PERFORMANCE IN PENNSYLVANIA. credo.stanford.edu

CHARTER SCHOOL PERFORMANCE IN PENNSYLVANIA. credo.stanford.edu CHARTER SCHOOL PERFORMANCE IN PENNSYLVANIA credo.stanford.edu April 2011 TABLE OF CONTENTS INTRODUCTION... 3 DISTRIBUTION OF CHARTER SCHOOL PERFORMANCE IN PENNSYLVANIA... 7 CHARTER SCHOOL IMPACT BY DELIVERY

More information

(PART 3 OF 3) CRESST REPORT 737 JULY, 2008. Mikyung Kim Wolf Joan L. Herman Lyle F. Bachman Alison L. Bailey Noelle Griffin

(PART 3 OF 3) CRESST REPORT 737 JULY, 2008. Mikyung Kim Wolf Joan L. Herman Lyle F. Bachman Alison L. Bailey Noelle Griffin CRESST REPORT 737 Mikyung Kim Wolf Joan L. Herman Lyle F. Bachman Alison L. Bailey Noelle Griffin RECOMMENDATIONS FOR ASSESSING ENGLISH LANGUAGE LEARNERS: ENGLISH LANGUAGE PROFICIENCY MEASURES AND ACCOMMODATION

More information

The Impact of Bilingual Education on English Language Acquisition Rates for English Language Learners, Including Exceptionalities. Dr.

The Impact of Bilingual Education on English Language Acquisition Rates for English Language Learners, Including Exceptionalities. Dr. The Impact of Bilingual Education on English Language Acquisition Rates for English Language Learners, Including Exceptionalities Dr. Jennifer Dixon Statement on Labels and Acronyms 1. I believe that each

More information

Connecting English Language Learning and Academic Performance: A Prediction Study

Connecting English Language Learning and Academic Performance: A Prediction Study Connecting English Language Learning and Academic Performance: A Prediction Study American Educational Research Association Vancouver, British Columbia, Canada Jadie Kong Sonya Powers Laura Starr Natasha

More information

Assessment Accommodations for English Language Learners: Implications for Policy-Based Empirical Research

Assessment Accommodations for English Language Learners: Implications for Policy-Based Empirical Research Review of Educational Research Spring 24, Vol. 74, No. 1, pp. 1-28 Assessment Accommodations for English Language Learners: Implications for Policy-Based Empirical Research Jamal Abedi University of California,

More information

Association Between Variables

Association Between Variables Contents 11 Association Between Variables 767 11.1 Introduction............................ 767 11.1.1 Measure of Association................. 768 11.1.2 Chapter Summary.................... 769 11.2 Chi

More information

Descriptive Statistics

Descriptive Statistics Descriptive Statistics Primer Descriptive statistics Central tendency Variation Relative position Relationships Calculating descriptive statistics Descriptive Statistics Purpose to describe or summarize

More information

Section 7: The Five-Step Process for Accommodations for English Language Learners (ELLs)

Section 7: The Five-Step Process for Accommodations for English Language Learners (ELLs) : The Five-Step Process for Accommodations for English Language Learners (ELLs) Step 1: Setting Expectations Expect English Language Learners (ELLs) to Achieve Grade-level Academic Content Standards Federal

More information

CHARTER SCHOOL PERFORMANCE IN INDIANA. credo.stanford.edu

CHARTER SCHOOL PERFORMANCE IN INDIANA. credo.stanford.edu CHARTER SCHOOL PERFORMANCE IN INDIANA credo.stanford.edu March 2011 TABLE OF CONTENTS INTRODUCTION... 3 CHARTER SCHOOL IMPACT BY STUDENTS YEARS OF ENROLLMENT AND AGE OF SCHOOL... 6 DISTRIBUTION OF CHARTER

More information

A National Study of School Effectiveness for Language Minority Students Long-Term Academic Achievement

A National Study of School Effectiveness for Language Minority Students Long-Term Academic Achievement A National Study of School Effectiveness for Language Minority Students Long-Term Academic Achievement Principal Investigators: Wayne P. Thomas, George Mason University Virginia P. Collier, George Mason

More information

The Language Factor in Mathematics Tests

The Language Factor in Mathematics Tests APPLIED MEASUREMENT IN EDUCATION, 14(3), 219 234 The Language Factor in Mathematics Tests Jamal Abedi University of California, Los Angeles/National Center for Research on Evaluation, Standards, and Student

More information

YEAR 3 REPORT: EVOLUTION OF PERFORMANCE MANAGEMENT ALBANY NY CHARTER SCHOO CHARTER SCHOOL PERFORMANCE IN NEW YORK CITY. credo.stanford.

YEAR 3 REPORT: EVOLUTION OF PERFORMANCE MANAGEMENT ALBANY NY CHARTER SCHOO CHARTER SCHOOL PERFORMANCE IN NEW YORK CITY. credo.stanford. YEAR 3 REPORT: EVOLUTION OF PERFORMANCE MANAGEMENT CHARTER SCHOOL PERFORMANCE IN NEW YORK CITY IN credo.stanford.edu ALBANY NY CHARTER SCHOO January 2010 SUMMARY This report supplements the CREDO National

More information

For Immediate Release: Thursday, July 19, 2012 Contact: Jim Polites 860.713.6525

For Immediate Release: Thursday, July 19, 2012 Contact: Jim Polites 860.713.6525 For Immediate Release: Thursday, July 19, 2012 Contact: Jim Polites 860.713.6525 2012 CMT, CAPT RESULTS SHOW SOME INCREASES, WHILE GAPS IN ACHIEVEMENT PERSIST The Connecticut State Department of Education

More information

Academic Achievement of English Language Learners in Post Proposition 203 Arizona

Academic Achievement of English Language Learners in Post Proposition 203 Arizona Academic Achievement of English Language Learners in Post Proposition 203 Arizona by Wayne E. Wright Assistant Professor University of Texas, San Antonio Chang Pu Doctoral Student University of Texas,

More information

2009 CREDO Center for Research on Education Outcomes (CREDO) Stanford University Stanford, CA http://credo.stanford.edu June 2009

2009 CREDO Center for Research on Education Outcomes (CREDO) Stanford University Stanford, CA http://credo.stanford.edu June 2009 Technical Appendix 2009 CREDO Center for Research on Education Outcomes (CREDO) Stanford University Stanford, CA http://credo.stanford.edu June 2009 CREDO gratefully acknowledges the support of the State

More information

Student Mobility and the Impact on Student Assessment Scores

Student Mobility and the Impact on Student Assessment Scores Student Mobility and the Impact on Student Assessment Scores STUDENT MOBILITY AND THE IMPACT ON STUDENT ASSESSMENT SCORES IN HOLYOKE The Holyoke Public School District completed a study on student mobility

More information

Effective Early Literacy Skill Development for English Language Learners: An Experimental Pilot Study of Two Methods*

Effective Early Literacy Skill Development for English Language Learners: An Experimental Pilot Study of Two Methods* Effective Early Literacy Skill Development for English Language Learners: An Experimental Pilot Study of Two Methods* Jo Ann M. Farver, Ph.D. Department of Psychology University of Southern California,

More information

Allen Elementary School

Allen Elementary School Allen Elementary School April 4, 216 Dear Parents and Community Members: We are pleased to present you with the (AER), which provides key information on the 214-15 educational progress for the. The AER

More information

Evaluating Analytical Writing for Admission to Graduate Business Programs

Evaluating Analytical Writing for Admission to Graduate Business Programs Evaluating Analytical Writing for Admission to Graduate Business Programs Eileen Talento-Miller, Kara O. Siegert, Hillary Taliaferro GMAC Research Reports RR-11-03 March 15, 2011 Abstract The assessment

More information

YEAR 3 REPORT: EVOLUTION OF PERFORMANCE MANAGEMENT ALBANY NY CHARTER SCHOO CHARTER SCHOOL PERFORMANCE IN ARIZONA. credo.stanford.edu.

YEAR 3 REPORT: EVOLUTION OF PERFORMANCE MANAGEMENT ALBANY NY CHARTER SCHOO CHARTER SCHOOL PERFORMANCE IN ARIZONA. credo.stanford.edu. YEAR 3 REPORT: EVOLUTION OF PERFORMANCE MANAGEMENT CHARTER SCHOOL PERFORMANCE IN ARIZONA IN credo.stanford.edu ALBANY NY CHARTER SCHOO June 2009 INTRODUCTION This report supplements the CREDO National

More information

Guidelines for the Assessment of English Language Learners

Guidelines for the Assessment of English Language Learners Guidelines for the Assessment of English Language Learners Copyright 2009 by Educational Testing Service. All rights reserved. ETS, the ETS logo and LISTENING. LEARNING. LEADING. are registered trademarks

More information

Surviving the Test: English Language Learners in Public Schools

Surviving the Test: English Language Learners in Public Schools Heather LaChapelle Public Intellectual Essay May 30, 2007 Surviving the Test: English Language Learners in Public Schools Since the passage of the No Child Left Behind Act in 2001, standardized tests and

More information

Stability of School Building Accountability Scores and Gains. CSE Technical Report 561. Robert L. Linn CRESST/University of Colorado at Boulder

Stability of School Building Accountability Scores and Gains. CSE Technical Report 561. Robert L. Linn CRESST/University of Colorado at Boulder Stability of School Building Accountability Scores and Gains CSE Technical Report 561 Robert L. Linn CRESST/University of Colorado at Boulder Carolyn Haug University of Colorado at Boulder April 2002 Center

More information

Transitioning English Language Learners in Massachusetts: An Exploratory Data Review. March 2012

Transitioning English Language Learners in Massachusetts: An Exploratory Data Review. March 2012 Transitioning English Language Learners in Massachusetts: An Exploratory Data Review March 2012 i This document was prepared by the Massachusetts Department of Elementary and Secondary Education Mitchell

More information

Introduction to. Hypothesis Testing CHAPTER LEARNING OBJECTIVES. 1 Identify the four steps of hypothesis testing.

Introduction to. Hypothesis Testing CHAPTER LEARNING OBJECTIVES. 1 Identify the four steps of hypothesis testing. Introduction to Hypothesis Testing CHAPTER 8 LEARNING OBJECTIVES After reading this chapter, you should be able to: 1 Identify the four steps of hypothesis testing. 2 Define null hypothesis, alternative

More information

Hispanic High School Student Graduates of Two-Way Bilingual Programs: Attitudes Toward School, Education and the Two-Way Program

Hispanic High School Student Graduates of Two-Way Bilingual Programs: Attitudes Toward School, Education and the Two-Way Program Hispanic High School Student Graduates of Two-Way Bilingual Programs: Attitudes Toward School, Education and the Two-Way Program Kathryn Lindholm-Leary San Jose State University Nationally, the academic

More information

Measuring Students Level of English Proficiency: Educational Significance and Assessment Requirements

Measuring Students Level of English Proficiency: Educational Significance and Assessment Requirements Educational Assessment, 13:193 214, 2008 Copyright Taylor & Francis Group, LLC ISSN: 1062-7197 print/1532-6977 online DOI: 10.1080/10627190802394404 Measuring Students Level of English Proficiency: Educational

More information

Section Format Day Begin End Building Rm# Instructor. 001 Lecture Tue 6:45 PM 8:40 PM Silver 401 Ballerini

Section Format Day Begin End Building Rm# Instructor. 001 Lecture Tue 6:45 PM 8:40 PM Silver 401 Ballerini NEW YORK UNIVERSITY ROBERT F. WAGNER GRADUATE SCHOOL OF PUBLIC SERVICE Course Syllabus Spring 2016 Statistical Methods for Public, Nonprofit, and Health Management Section Format Day Begin End Building

More information

Chi Square Tests. Chapter 10. 10.1 Introduction

Chi Square Tests. Chapter 10. 10.1 Introduction Contents 10 Chi Square Tests 703 10.1 Introduction............................ 703 10.2 The Chi Square Distribution.................. 704 10.3 Goodness of Fit Test....................... 709 10.4 Chi Square

More information

YEAR 3 REPORT: EVOLUTION OF PERFORMANCE MANAGEMENT ALBANY NY CHARTER SCHOO CHARTER SCHOOL PERFORMANCE IN FLORIDA. credo.stanford.edu.

YEAR 3 REPORT: EVOLUTION OF PERFORMANCE MANAGEMENT ALBANY NY CHARTER SCHOO CHARTER SCHOOL PERFORMANCE IN FLORIDA. credo.stanford.edu. YEAR 3 REPORT: EVOLUTION OF PERFORMANCE MANAGEMENT CHARTER SCHOOL PERFORMANCE IN FLORIDA IN credo.stanford.edu ALBANY NY CHARTER SCHOO June 2009 INTRODUCTION This report supplements the CREDO National

More information

Higher Performing High Schools

Higher Performing High Schools COLLEGE READINESS A First Look at Higher Performing High Schools School Qualities that Educators Believe Contribute Most to College and Career Readiness 2012 by ACT, Inc. All rights reserved. A First Look

More information

Chapter 5: Analysis of The National Education Longitudinal Study (NELS:88)

Chapter 5: Analysis of The National Education Longitudinal Study (NELS:88) Chapter 5: Analysis of The National Education Longitudinal Study (NELS:88) Introduction The National Educational Longitudinal Survey (NELS:88) followed students from 8 th grade in 1988 to 10 th grade in

More information

UNDERSTANDING THE TWO-WAY ANOVA

UNDERSTANDING THE TWO-WAY ANOVA UNDERSTANDING THE e have seen how the one-way ANOVA can be used to compare two or more sample means in studies involving a single independent variable. This can be extended to two independent variables

More information

Nebraska School Counseling State Evaluation

Nebraska School Counseling State Evaluation Nebraska School Counseling State Evaluation John Carey and Karen Harrington Center for School Counseling Outcome Research Spring 2010 RESEARCH S c h o o l o f E d u c a t i o n U n i v e r s i t y o f

More information

Calculating P-Values. Parkland College. Isela Guerra Parkland College. Recommended Citation

Calculating P-Values. Parkland College. Isela Guerra Parkland College. Recommended Citation Parkland College A with Honors Projects Honors Program 2014 Calculating P-Values Isela Guerra Parkland College Recommended Citation Guerra, Isela, "Calculating P-Values" (2014). A with Honors Projects.

More information

California Charter Schools Serving Low-SES Students: An Analysis of the Academic Performance Index

California Charter Schools Serving Low-SES Students: An Analysis of the Academic Performance Index California Charter Schools Serving Low-SES Students: An Analysis of the Academic Performance Index March 11, 2002 By Simeon P. Slovacek, Ph.D. Antony J. Kunnan, Ph.D. Hae-Jin Kim, M.A. This report presents

More information

A study of the Singapore math program, Math in Focus, state test results

A study of the Singapore math program, Math in Focus, state test results Advisory Board: Michael Beck, President Beck Evaluation & Testing Associates, Inc. Jennifer M. Conner, Assistant Professor Indiana University Keith Cruse, Former Managing Director Texas Assessment Program

More information

Similar English Learner Students, Different Results: Why Do Some Schools Do Better?

Similar English Learner Students, Different Results: Why Do Some Schools Do Better? REPORT OF FINDINGS Similar English Learner Students, Different Results: Why Do Some Schools Do Better? A follow-up analysis, based upon a large-scale survey of California elementary schools serving high

More information

X = T + E. Reliability. Reliability. Classical Test Theory 7/18/2012. Refers to the consistency or stability of scores

X = T + E. Reliability. Reliability. Classical Test Theory 7/18/2012. Refers to the consistency or stability of scores Reliability It is the user who must take responsibility for determining whether or not scores are sufficiently trustworthy to justify anticipated uses and interpretations. (AERA et al., 1999) Reliability

More information

Master Plan Evaluation Report for English Learner Programs

Master Plan Evaluation Report for English Learner Programs Master Plan Evaluation Report (2002-03) for English Learner Programs Page i Los Angeles Unified School District Master Plan Evaluation Report for English Learner Programs 2002-03 Prepared by Jesús José

More information

Charter School Performance in Ohio 12/18/2014

Charter School Performance in Ohio 12/18/2014 Charter School Performance in Ohio 12/18/2014 CREDO at Stanford University 434 Galvez Mall Stanford University Stanford, CA 94305-6010 CREDO, the Center for Research on Education Outcomes at Stanford University,

More information

Test Reliability Indicates More than Just Consistency

Test Reliability Indicates More than Just Consistency Assessment Brief 015.03 Test Indicates More than Just Consistency by Dr. Timothy Vansickle April 015 Introduction is the extent to which an experiment, test, or measuring procedure yields the same results

More information

IDENTIFICATION AND ASSESSMENT OF ENGLISH LANGUAGE LEARNERS

IDENTIFICATION AND ASSESSMENT OF ENGLISH LANGUAGE LEARNERS IDENTIFICATION AND ASSESSMENT OF ENGLISH LANGUAGE LEARNERS CHAPTER 3 INDEX 3.1 PURPOSE AND SCOPE... 3 1 3.2 IDENTIFICATION AND REFERRAL OF ENGLISH LEARNERS SUSPECTED OF HAING A DISABILITY... 3 1 3.3 SPECIAL

More information

Relating the ACT Indicator Understanding Complex Texts to College Course Grades

Relating the ACT Indicator Understanding Complex Texts to College Course Grades ACT Research & Policy Technical Brief 2016 Relating the ACT Indicator Understanding Complex Texts to College Course Grades Jeff Allen, PhD; Brad Bolender; Yu Fang, PhD; Dongmei Li, PhD; and Tony Thompson,

More information

Utah Comprehensive Counseling and Guidance Program Evaluation Report

Utah Comprehensive Counseling and Guidance Program Evaluation Report Utah Comprehensive Counseling and Guidance Program Evaluation Report John Carey and Karen Harrington Center for School Counseling Outcome Research School of Education University of Massachusetts Amherst

More information

Professional Development and Self-Efficacy of Texas Educators and the Teaching of English Language Learners

Professional Development and Self-Efficacy of Texas Educators and the Teaching of English Language Learners International Journal of Humanities and Social Science Vol. 3 No. 19; November 2013 Professional Development and Self-Efficacy of Texas Educators and the Teaching of English Language Learners Kathleen

More information

A Hands-On Exercise Improves Understanding of the Standard Error. of the Mean. Robert S. Ryan. Kutztown University

A Hands-On Exercise Improves Understanding of the Standard Error. of the Mean. Robert S. Ryan. Kutztown University A Hands-On Exercise 1 Running head: UNDERSTANDING THE STANDARD ERROR A Hands-On Exercise Improves Understanding of the Standard Error of the Mean Robert S. Ryan Kutztown University A Hands-On Exercise

More information

Characteristics of Colorado s Online Students

Characteristics of Colorado s Online Students Characteristics of Colorado s Online Students By: Amanda Heiney, Dianne Lefly and Amy Anderson October 2012 Office of Online & Blended Learning 201 E. Colfax Ave., Denver, CO 80203 Phone: 303-866-6897

More information

Two steps are necessary to implement this process. 1. Administer the Home Language survey to all students enrolled in the school corporation.

Two steps are necessary to implement this process. 1. Administer the Home Language survey to all students enrolled in the school corporation. INDIANA DEPARTMENT OF EDUCATION Office of English Language Learning & Migrant Education Guidelines to Satisfy Legal Requirements of Title VI of the Civil Rights Act of 1964 (42 USCS, 2000d) Lau v. Nichols

More information

GMAC. Predicting Success in Graduate Management Doctoral Programs

GMAC. Predicting Success in Graduate Management Doctoral Programs GMAC Predicting Success in Graduate Management Doctoral Programs Kara O. Siegert GMAC Research Reports RR-07-10 July 12, 2007 Abstract An integral part of the test evaluation and improvement process involves

More information

How To Use The College Learning Assessment (Cla) To Plan A College

How To Use The College Learning Assessment (Cla) To Plan A College Using the Collegiate Learning Assessment (CLA) to Inform Campus Planning Anne L. Hafner University Assessment Coordinator CSULA April 2007 1 Using the Collegiate Learning Assessment (CLA) to Inform Campus

More information

The Effects of Read Naturally on Grade 3 Reading: A Study in the Minneapolis Public Schools

The Effects of Read Naturally on Grade 3 Reading: A Study in the Minneapolis Public Schools The Effects of Read Naturally on Grade 3 Reading: A Study in the Minneapolis Public Schools David Heistad, Ph.D. Director of Research, Evaluation, and Assessment Minneapolis Public Schools Introduction

More information

School Performance Framework: Technical Guide

School Performance Framework: Technical Guide School Performance Framework: Technical Guide Version 1.6 August 2010 This technical guide provides information about the following topics as they related to interpreting the school performance framework

More information

!"#$%&'"()"!*+,(-./0"12.342*0"()"" 5-67+04"8.-6%.62"82.*-2*0"!

!#$%&'()!*+,(-./012.342*0() 5-67+048.-6%.6282.*-2*0! "#$%&'"()"*+,(-./0"12.342*0"()"" 5-67+04"8.-6%.62"82.*-2*0" 923+7+.":+(0;6%+7.*(-,?72,;9.-342

More information

" Y. Notation and Equations for Regression Lecture 11/4. Notation:

 Y. Notation and Equations for Regression Lecture 11/4. Notation: Notation: Notation and Equations for Regression Lecture 11/4 m: The number of predictor variables in a regression Xi: One of multiple predictor variables. The subscript i represents any number from 1 through

More information

Non-Parametric Tests (I)

Non-Parametric Tests (I) Lecture 5: Non-Parametric Tests (I) KimHuat LIM lim@stats.ox.ac.uk http://www.stats.ox.ac.uk/~lim/teaching.html Slide 1 5.1 Outline (i) Overview of Distribution-Free Tests (ii) Median Test for Two Independent

More information

Chapter 5 English Language Learners (ELLs) and the State of Texas Assessments of Academic Readiness (STAAR) Program

Chapter 5 English Language Learners (ELLs) and the State of Texas Assessments of Academic Readiness (STAAR) Program Chapter 5 English Language Learners (ELLs) and the State of Texas Assessments of Academic Readiness (STAAR) Program Demographic projections indicate that the nation s English language learner (ELL) student

More information

Factorial Invariance in Student Ratings of Instruction

Factorial Invariance in Student Ratings of Instruction Factorial Invariance in Student Ratings of Instruction Isaac I. Bejar Educational Testing Service Kenneth O. Doyle University of Minnesota The factorial invariance of student ratings of instruction across

More information

Review of Special Education in the Commonwealth of Massachusetts

Review of Special Education in the Commonwealth of Massachusetts Review of Special Education in the Commonwealth of Massachusetts Thomas Hehir and Associates Thomas Hehir, Todd Grindal and Hadas Eidelman Boston, Massachusetts April 2012 Report commissioned by the Massachusetts

More information

The Personal Learning Insights Profile Research Report

The Personal Learning Insights Profile Research Report The Personal Learning Insights Profile Research Report The Personal Learning Insights Profile Research Report Item Number: O-22 995 by Inscape Publishing, Inc. All rights reserved. Copyright secured in

More information

Elementary Statistics

Elementary Statistics Elementary Statistics Chapter 1 Dr. Ghamsary Page 1 Elementary Statistics M. Ghamsary, Ph.D. Chap 01 1 Elementary Statistics Chapter 1 Dr. Ghamsary Page 2 Statistics: Statistics is the science of collecting,

More information

There are three kinds of people in the world those who are good at math and those who are not. PSY 511: Advanced Statistics for Psychological and Behavioral Research 1 Positive Views The record of a month

More information

Has Progress Been Made in Raising Achievement for English Language Learners?

Has Progress Been Made in Raising Achievement for English Language Learners? State Test Score Trends Through 2007-08, Part 6 Has Progress Been Made in Raising Achievement for English Language Learners? Center on Education Policy April 2010 Has Progress Been Made in Raising Achievement

More information

Center for Advanced Studies in Measurement and Assessment. CASMA Research Report

Center for Advanced Studies in Measurement and Assessment. CASMA Research Report Center for Advanced Studies in Measurement and Assessment CASMA Research Report Number 13 and Accuracy Under the Compound Multinomial Model Won-Chan Lee November 2005 Revised April 2007 Revised April 2008

More information

Charter School Performance in Ohio 12/9/2014

Charter School Performance in Ohio 12/9/2014 Charter School Performance in Ohio 12/9/2014 CREDO at Stanford University 434 Galvez Mall Stanford University Stanford, CA 94305-6010 CREDO, the Center for Research on Education Outcomes at Stanford University,

More information

St. Louis Area School District - Trends in the Study of Educational Behavior

St. Louis Area School District - Trends in the Study of Educational Behavior Geospatial Perspectives of School, Teacher, and Student Variables Focus on Recently De-accredited St. Louis City School District Mark C. Hogrebe Lydia Kyei-Blankson Li Zou Center for Inquiry in Science

More information

Transadaptation: Publishing Assessments in World Languages

Transadaptation: Publishing Assessments in World Languages assessment report. : Publishing Assessments in World Languages........ Sasha Zucker Margarita Miska Linda G. Alaniz Luis Guzmán September 2005 : Publishing Assessments in World Languages Introduction In

More information

Score Comparability of Online and Paper Administrations of the Texas Assessment of Knowledge and Skills. Walter D. Way. Laurie Laughlin Davis

Score Comparability of Online and Paper Administrations of the Texas Assessment of Knowledge and Skills. Walter D. Way. Laurie Laughlin Davis Score Comparability of Online and Paper Administrations of the Texas Assessment of Knowledge and Skills Walter D. Way Laurie Laughlin Davis Steven Fitzpatrick Pearson Educational Measurement Paper presented

More information

Pragmatic Perspectives on the Measurement of Information Systems Service Quality

Pragmatic Perspectives on the Measurement of Information Systems Service Quality Pragmatic Perspectives on the Measurement of Information Systems Service Quality Analysis with LISREL: An Appendix to Pragmatic Perspectives on the Measurement of Information Systems Service Quality William

More information

Calculating, Interpreting, and Reporting Estimates of Effect Size (Magnitude of an Effect or the Strength of a Relationship)

Calculating, Interpreting, and Reporting Estimates of Effect Size (Magnitude of an Effect or the Strength of a Relationship) 1 Calculating, Interpreting, and Reporting Estimates of Effect Size (Magnitude of an Effect or the Strength of a Relationship) I. Authors should report effect sizes in the manuscript and tables when reporting

More information

Study Guide for the Final Exam

Study Guide for the Final Exam Study Guide for the Final Exam When studying, remember that the computational portion of the exam will only involve new material (covered after the second midterm), that material from Exam 1 will make

More information

English Learner Program Description White Bear Lake Area Schools

English Learner Program Description White Bear Lake Area Schools English Learner Program Description White Bear Lake Area Schools March, 2012 Please direct questions to: Kathleen Daniels Director of Special Services kathleen.daniels@isd624.org 1 The purpose of this

More information

Principals Use of Computer Technology

Principals Use of Computer Technology Principals Use of Computer Technology 85 Lantry L. Brockmeier James L. Pate Don Leech Abstract: The Principal s Computer Technology Survey, a 40-item instrument, was employed to collect data on Georgia

More information

II. DISTRIBUTIONS distribution normal distribution. standard scores

II. DISTRIBUTIONS distribution normal distribution. standard scores Appendix D Basic Measurement And Statistics The following information was developed by Steven Rothke, PhD, Department of Psychology, Rehabilitation Institute of Chicago (RIC) and expanded by Mary F. Schmidt,

More information

Top-to-Bottom Ranking, Priority, Focus and Rewards Schools Identification Business Rules. Overview

Top-to-Bottom Ranking, Priority, Focus and Rewards Schools Identification Business Rules. Overview Top-to-Bottom Ranking, Priority, Focus and Rewards Schools Identification Business Rules Overview Top-to-Bottom Ranking: List of schools and ranked by their performance. The ranking is based on student

More information

2014-2015 MCA Data Discoveries

2014-2015 MCA Data Discoveries 2014-2015 MCA Data Discoveries What does Minnesota think is important? What do we want kids to do? Pass important tests Be Proficient Grow in their knowledge Growth Have equal opportunity to learn Close

More information

Using Eggen & Kauchak, Educational Psychology: Windows on Classrooms for the New York State Teacher Certification Examinations

Using Eggen & Kauchak, Educational Psychology: Windows on Classrooms for the New York State Teacher Certification Examinations Using Eggen & Kauchak, Educational Psychology: Windows on Classrooms for the New York State Teacher Certification Examinations Introduction The New York State Education Department requires that prospective

More information

Predicting Successful Completion of the Nursing Program: An Analysis of Prerequisites and Demographic Variables

Predicting Successful Completion of the Nursing Program: An Analysis of Prerequisites and Demographic Variables Predicting Successful Completion of the Nursing Program: An Analysis of Prerequisites and Demographic Variables Introduction In the summer of 2002, a research study commissioned by the Center for Student

More information

ILLINOIS DISTRICT REPORT CARD

ILLINOIS DISTRICT REPORT CARD -63-56-6- McHenry CHSD 56 McHenry CHSD 56 McHenry, ILLINOIS 2 ILLINOIS DISTRICT REPORT CARD and federal laws require public school districts to release report cards to the public each year. Starting in

More information

Mapping State Proficiency Standards Onto the NAEP Scales:

Mapping State Proficiency Standards Onto the NAEP Scales: Mapping State Proficiency Standards Onto the NAEP Scales: Variation and Change in State Standards for Reading and Mathematics, 2005 2009 NCES 2011-458 U.S. DEPARTMENT OF EDUCATION Contents 1 Executive

More information

Statistics Review PSY379

Statistics Review PSY379 Statistics Review PSY379 Basic concepts Measurement scales Populations vs. samples Continuous vs. discrete variable Independent vs. dependent variable Descriptive vs. inferential stats Common analyses

More information

Abstract Title: Identifying and measuring factors related to student learning: the promise and pitfalls of teacher instructional logs

Abstract Title: Identifying and measuring factors related to student learning: the promise and pitfalls of teacher instructional logs Abstract Title: Identifying and measuring factors related to student learning: the promise and pitfalls of teacher instructional logs MSP Project Name: Assessing Teacher Learning About Science Teaching

More information

Benchmark Assessment in Standards-Based Education:

Benchmark Assessment in Standards-Based Education: Research Paper Benchmark Assessment in : The Galileo K-12 Online Educational Management System by John Richard Bergan, Ph.D. John Robert Bergan, Ph.D. and Christine Guerrera Burnham, Ph.D. Submitted by:

More information

Top-to-Bottom Ranking, Priority, Focus and Reward Schools Identification Business Rules. Overview

Top-to-Bottom Ranking, Priority, Focus and Reward Schools Identification Business Rules. Overview Top-to-Bottom Ranking, Priority, Focus and Reward Schools Identification Business Rules Overview Top-to-Bottom Ranking: List of schools, rank ordered by their performance. The ranking is based on student

More information

State of New Jersey 2012-13 41-5460-050 OVERVIEW WARREN COUNTY VOCATIONAL TECHNICAL SCHOOL WARREN 1500 ROUTE 57 WARREN COUNTY VOCATIONAL

State of New Jersey 2012-13 41-5460-050 OVERVIEW WARREN COUNTY VOCATIONAL TECHNICAL SCHOOL WARREN 1500 ROUTE 57 WARREN COUNTY VOCATIONAL 1 415465 OVERVIEW TECHNICAL SCHOOL 15 ROUTE 57 GRADE SPAN 912 WASHINGTON, NEW JERSEY 78829618 1. This school's academic performance is high when compared to schools across the state. Additionally, its

More information

Additional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jin-tselink/tselink.htm

Additional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jin-tselink/tselink.htm Mgt 540 Research Methods Data Analysis 1 Additional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jin-tselink/tselink.htm http://web.utk.edu/~dap/random/order/start.htm

More information

2013 A-F Letter Grade Accountability System TECHNICAL MANUAL

2013 A-F Letter Grade Accountability System TECHNICAL MANUAL 2013 A-F Letter Grade Accountability System TECHNICAL MANUAL Arizona Department of Education John Huppenthal, Superintendent For more information, please contact: Research & Evaluation Section (602) 542-5151

More information

Questions and Answers Regarding English Language Learners (ELLs) with Disabilities. Volume 10

Questions and Answers Regarding English Language Learners (ELLs) with Disabilities. Volume 10 Questions and Answers Regarding English Language Learners (ELLs) with Disabilities Volume 10 1) What factors should be considered prior to referral for evaluation? A: Local Education Agencies (LEAs) must

More information

Keeping Kids in School: An LA s BEST Example A Study Examining the Long-Term Impact of LA s BEST on Students Dropout Rates

Keeping Kids in School: An LA s BEST Example A Study Examining the Long-Term Impact of LA s BEST on Students Dropout Rates Keeping Kids in School: An LA s BEST Example A Study Examining the Long-Term Impact of LA s BEST on Students Dropout Rates Final Deliverable December 2005 Denise Huang, Kyung Sung Kim, Anne Marshall, and

More information

Statistics. Measurement. Scales of Measurement 7/18/2012

Statistics. Measurement. Scales of Measurement 7/18/2012 Statistics Measurement Measurement is defined as a set of rules for assigning numbers to represent objects, traits, attributes, or behaviors A variableis something that varies (eye color), a constant does

More information

Academic Performance of IB Students Entering the University of California System from 2000 2002

Academic Performance of IB Students Entering the University of California System from 2000 2002 RESEARCH SUMMARY Academic Performance of IB Students Entering the University of California System from 2000 2002 IB Global Policy & Research Department August 2010 Abstract This report documents the college

More information