Acommon question asked by policymakers, teachers, parents, and some test takers
|
|
|
- Dominick Johnston
- 9 years ago
- Views:
Transcription
1 No. 16 September 2011 Why Do Standardized Testing Programs Report Scaled Scores? Why Not Just Report the Raw or Percent-Correct Scores? By Xuan Tan and Rochelle Michel Key concepts To understand why testing programs report scaled scores, it is useful to understand these ideas: Raw score Total number of raw points a test taker receives based on the number of questions answered correctly; typically, for example, 20 correct answers means a raw score of 20 Scaled score Scores that have been mathematically transformed from one set of numbers (i.e., the raw score) to another set of numbers in order to make them comparable in some way for example, across different editions, or forms, of the same test continued on p. 2 Acommon question asked by policymakers, teachers, parents, and some test takers is: Why is it not appropriate to use raw or percent-correct scores for comparing test takers across different test editions? Why do standardized testing programs go through complicated statistical processes to come up with scaled scores? Standardized tests are widely used in K 12, higher education, and in some professions for such purposes as accountability, admissions, certification, and licensure. Standardized tests provide a common basis for evaluating and comparing test-takers abilities in a specific content area. They are administered and scored in a predetermined, set manner that is consistent for all test takers (e.g., test questions, time allowed for each administration, scoring procedures). However, in order for standardized testing programs to have consistency in score interpretation when there are different editions of the test, programs often transform test scores (summed raw score points assigned to different questions) into a set of values different from the raw score points obtained directly from a test. Further, testing programs often report these transformed test scores, which are called scaled scores, rather than reporting percent-correct scores derived from the raw score points. This standardization allows scores reported from a test to have consistent meaning for all test takers. This article highlights the reasons why percent-correct scores are generally not used as the primary reported scores and provides additional details on what a scaled score is, how scaled scores are obtained, and the reasons for providing scaled scores, as well as their usefulness in interpreting test scores. The Raw and Percent-Correct Score A raw score is the total number of score points a test taker obtains by answering questions correctly on a test. A percent-correct score represents the percentage of questions a test taker answered correctly on a test. For example, if a test taker answered 20 out of 50 questions on a test correctly, then his or her percent-correct score would be 40%. The raw score, in this example, is 20. The percent-correct score can be considered Editor s note: Xuan Tan is a Psychometrician in the ETS Research & Development division s Center for Statistical Analysis. Rochelle Michel, also a doctoral-level psychometrician by training, is a Product Manager in ETS s Higher Education division. 1
2 Key concepts continued from p. 1 Equating Process used to place all forms of the same test on the same scale and make scores comparable across forms Anchor items Set of questions that is common to different forms of the same test in order to facilitate the statistical comparison of group ability and form difficulty that takes place during equating Percent-correct score Percentage of questions a test taker answered correctly on a test; often used in classroom tests an adjusted raw score to account for differences in the lengths of different tests. The percent-correct score is easy to calculate and understand, and is often used in classroom tests for score reporting. Why are percent-correct scores not used as the primary reported scores? In the case of many standardized testing programs, it is necessary to develop multiple editions of a test as a way of dealing with the issue of content exposure. Test questions from standardized tests are usually secure, but when the same test is repeatedly administered to a large number of test takers, the questions can become exposed to the public and jeopardize the testing process. Test takers can remember the test questions and share them with future test takers through different media. To address this concern, test takers taking a test at different times may receive different editions of the test. Also, test takers within the same administration may be administered different editions of the test to address security concerns. In some cases, multiple editions of a test are developed in response to state laws requesting the disclosure of test questions after each administration. Standardized testing programs often develop different editions of the same test that contain different sets of test questions conforming to predefined content and statistical specifications. These different editions are commonly called forms. Although strict adherence to common test specifications or blueprints allows test developers to create multiple forms that are remarkably similar in difficulty, they are rarely, if ever, exactly equal in difficulty (Holland & Dorans, 2006). This makes it hard to use the percent-correct score for fair comparisons of test takers performances on different forms of the same test. For example, getting 50% correct on a hard form may mean the test taker has more knowledge and skill than another test taker getting 60% correct on a relatively easier form. For the same reason, the raw scores cannot be used to compare test takers performances on different forms. When two test takers get the same raw score on two different forms, the test taker who took the more difficult form has demonstrated a higher level of performance than the test taker who took the relatively easier form. Most standardized testing programs require scores that can be compared across different forms. In order for different stakeholders (states, schools, etc.) to make consistent and fair decisions based on assessment results, the scores reported from standardized tests need to be comparable that is, scores must carry the same meaning regardless of which form was administered. Simply put, scores on different forms of a test should indicate the same level of performance no matter which form the test taker received. Most standardized testing programs do not use percent-correct scores as the primary scale for reporting assessment results because such scores are not comparable across forms. The raw scores are not comparable across forms either. However, they are often reported to the test takers along with the scaled scores as a direct indicator of how many points a test taker obtained from the set of questions on a test form. 2
3 For an easier form, a test taker needs to answer slightly more questions correctly to get a particular scaled score. For a more difficult form, a test taker can get the same scaled score answering slightly fewer questions correctly. Table 1. Scaled Scores for Form A and Form B Scaled Score Raw Score Form A Form B Table 1 shows an example of scaled scores associated with different raw scores for two different forms. In this hypothetical example, Form A is the more difficult form. To achieve the same scaled score of 195, a test taker needs to answer 96 out of the 100 questions correctly on Form A, but needs to answer 97 questions correctly on Form B. In some cases, however, percent-correct scores are used as auxiliary scores as a way of providing score users with additional information to assist in understanding their performance. In such cases, the professional guidelines used at ETS 1 and in the testing industry 2 as a whole typically call on testing programs to state the limitations of the percent-correct scores and provide guidelines as to the appropriate use of the test scores. The Scaled Score To achieve comparability, standardized testing programs report scaled scores. The reported scaled scores are obtained by statistically adjusting and converting raw scores onto a common scale to account for differences in difficulty across different forms. For an easier form, a test taker needs to answer slightly more questions correctly to get a particular scaled score. For a more difficult form, a test taker can get the same scaled score answering slightly fewer questions correctly. Table 1 shows an example of scaled scores associated with different raw scores for two different forms, A and B. As illustrated in Table 1, Form A is the more difficult form. To achieve the same scaled score of 195, a test taker needs to answer 96 out of the 100 questions correctly on Form A, but needs to answer 97 questions correctly on Form B. In order to obtain comparable scaled scores across different forms of a test, testing programs use processes known as scaling and equating. Scaling, sometimes referred to as setting the scale, is the process by which raw scores are transformed, either linearly 1 ETS Standards for Quality and Fairness (ETS, 2002) 2 Standards for Educational and Psychological Testing (American Educational Research Association, American Psychological Association, & National Council of Measurement in Education, 1999) 3
4 Under most circumstances, having test takers take two forms of the same test at the same time is not practical due to issues such as increased testing time and test-taker fatigue. or nonlinearly, to a scale with a range of numbers that are usually different from the possible range of raw scores. The transformed scores, which are called scaled scores, are reported to the test score users. During the scaling process, the first form administered (or one form, if more than one is administered at the same administration) is considered the base form and the initial scale is set using this base form. For example, a scale of 100 to 200 could be selected as the scaled score range for a test with a possible range of raw scores from 0 to 100. Scores on all subsequent forms are placed on the same scale (100 to 200) as the base form through another process known as equating. Equating is the process by which raw scores on a new form are adjusted to account for the differences in form difficulty from a base or reference form. In order to quantify and adjust for differences in form difficulty, it would be desirable to have the same group of test takers take the two forms (the new form and the reference form) at the same time. The difference in average performance on the two forms is a direct indication of the difference in form difficulty. Then, scores on the new form can be statistically adjusted to make average performances on both forms equal. However, under most circumstances, having test takers take two forms of the same test at the same time is not practical due to issues such as increased testing time and testtaker fatigue. Another option is to have two different groups of test takers take the two forms at the same administration or at two different administrations. However, because these two groups of test takers could have different average abilities, the difference in average performance on the two forms could indicate the existence of both group ability differences and form difficulty differences. Anchor Items In order to isolate and quantify the difference in form difficulty, a common practice for standardized testing programs is to embed a common set of test questions, called an anchor, in both the new form and the reference form. Since both groups of test takers answer the same set of anchor questions, the difference in average performance on the anchor questions provides an indication of group ability differences. When the group ability difference is quantified and removed from the difference in average performance on the two forms, what is left in the average performance difference is an indication of the difference in form difficulty. With the difference in form difficulty identified and quantified, scores on the new form can then be statistically adjusted to remove the impact of the form difficulty difference. Figure 1 illustrates the scaling and equating process to obtain the scaled scores for two forms. Form A is the first form and the base form on which the initial scaling is done. Form B is the second form that is equated to Form A. In this illustration, Form A is the harder form. Two steps are involved in obtaining the raw-to-scale score relationship of Form B. First, the raw scores on Form B are equated to raw scores on Form A and, as a result, higher scores on Form B correspond to lower scores on Form A. Second, the rawto-scale relationship for Form A is applied to the equated scores on Form B. Once the scaling and equating processes are completed, scaled scores from different test forms are considered interchangeable, which means the scaled scores indicate the same levels of performance across forms of the test. 4
5 Application Scaled scores can provide useful information about the test taker. For example, they may, if designed well: Facilitate meaningful comparisons of scores from test takers who took different editions of the test at different times Help score users to form meaningful inferences about what test takers know and can do, but discourage them from making misinterpretations and inappropriate inferences Allow for measurement precision while avoiding overemphasis on minor differences between points along the scale Form A Raw Score Scaled Score Raw Score Adjusted Score Scaled Score STEP 1: Scaling STEP 2: Equating Form B Figure 1. The scaling and equating process for obtaining scaled scores. Forms A and B are from the hypothetical example introduced in Table 1, found on page 3. As shown in Figure 1, after the scaling and equating processes, scaled scores obtained from Form A and Form B are equivalent and interchangeable. If two test takers taking Forms A and B respectively got the same scaled score of 194 (corresponding to raw scores of 95 on Form A and 96 on Form B), we know these two test takers exhibited the same level of performance on this test. One might ask: Why not report the adjusted scores for Form B instead of the scaled scores? This is because the adjusted scores would be on the same scale as the raw scores and could be easily misinterpreted as the raw scores. Thus, the scaled scores are used and are commonly set to a range of values different from the raw score values. Regardless of the scaling procedure used, a norm or reference group is often used, and data collected on the norm group is often reported to accompany the scale for score interpretation purposes. This initial sample of test takers is usually selected to be representative and reflect the demographics of the intended testing population (ETS, 2010). When the group is representative of the population of test takers, this allows for better interpretation of the scaled scores. The reference group s performance can serve as a benchmark against which a new group of test takers can compare their subsequent performances. Usefulness of the Scaled Score The utility of the scaled score comes from allowing for meaningful score interpretations and, at the same time, minimizing misinterpretations and inappropriate inferences. Test-score users frequently want additional information to assist in the interpretation of scaled scores. Providing information related to content, norm or reference groups, and precision of scores helps with the meaningful interpretation of these scores (Kolen & Brennan, 2004; Petersen, Kolen, & Hoover, 1989). 5
6 Information about the precision of the scaled scores assists test users with making appropriate interpretations based on the reported scaled scores. The number of distinct scores on the scale should allow for measurement precision but avoid encouraging too much emphasis on differences between points along the score scale (e.g., Kolen & Brennan, 2004). For example, the scaled score can be reported in various increments (1-point increments, 5-point increments, 10-point increments, etc.). Usually, we want each additional correct answer to make a difference in the test takers scaled score, but not such a large difference that people exaggerate its importance. The selection of a score scale, with appropriate increments, aids in the usefulness of the reported scaled scores to the test-score users. R&D Connections is published by ETS Research & Development Educational Testing Service Rosedale Road, 19-T Princeton, NJ [email protected] Editor: Jeff Johnson Copy Editor: Eileen Kerrigan Layout Design: Tom Hutchinson Visit ETS Research & Development on the web at Follow ETS Research on Twitter (@ETSresearch) Copyright 2011 by Educational Testing Service. All rights reserved. ETS, the ETS logo and LISTENING. LEARNING. LEADING. are registered trademarks of Educational Testing Service (ETS) Thus, although percent-correct scores are easy to calculate and easy to understand, they are often misinterpreted, especially in circumstances where more than one edition of a test exists. Alternatively, scaled scores should be the primary scores provided to test-score users so that reported scores are more appropriately used and correctly interpreted. References American Educational Research Association, American Psychological Association, & National Council of Measurement in Education. (1999). Standards for educational and psychological testing. Washington, DC: AERA. Educational Testing Service. (2010). Glossary of standardized testing terms. Retrieved from Educational Testing Service. (2002). ETS standards for quality and fairness. Princeton, NJ: Author. Holland, P. W., & Dorans, N. J. (2006). Linking and equating. In R. L. Brennan (Ed.), Educational measurement (4th ed., pp ). Westport, CT: Praeger. Kolen, M. J., & Brennan, R. L. (2004). Test equating, scaling, and linking (2nd ed.). New York: Springer. Petersen, N. S., Kolen, M. J., & Hoover, H. D. (1989). Scaling, norming, and equating. In R.L. Linn (Ed.), Educational measurement (3rd ed., pp ). New York: Macmillan.
Understanding Your Praxis Scores
2013 14 Understanding Your Praxis Scores The Praxis Series Assessments are developed and administered by Educational Testing Service (E T S ). Praxis Core Academic Skills for Educators (Core) tests measure
The Official Study Guide
The Praxis Series ebooks The Official Study Guide French: World Language Test Code: 5174 Study Topics Practice Questions Directly From the Test Makers Test-Taking Strategies www.ets.org/praxis Study Guide
EQUATING TEST SCORES
EQUATING TEST SCORES (Without IRT) Samuel A. Livingston Listening. Learning. Leading. Equating Test Scores (Without IRT) Samuel A. Livingston Copyright 2004 Educational Testing Service. All rights reserved.
The new Verbal Reasoning and Quantitative Reasoning score scales:
Better by Design The new Verbal Reasoning and Quantitative Reasoning score scales: A helpful overview of what you need to know Inside: An in-depth look at the new 130 170 score scales Guidance on making
To many people, standardized testing means multiple choice testing. However,
No. 11 September 2009 Examples Constructed-response questions are a way of measuring complex skills. These are examples of tasks test takers might encounter: Literature Writing an essay comparing and contrasting
ETS Research Spotlight. Assessment and Gaming
ETS Research Spotlight Assessment and Gaming ISSUE 8, APRIL 2013 Foreword ETS Research Spotlight ISSUE 8 APRIL 2013 Senior Editor: Jeff Johnson Contributing Editor: Hans Sandberg Copy Editor: Eileen Kerrigan
A BRIEF GUIDE TO SELECTING AND USING PRE-POST ASSESSMENTS
A BRIEF GUIDE TO SELECTING AND USING PRE-POST ASSESSMENTS Prepared by the National Evaluation and Technical Assistance Center for the Education of Children and Youth who are Neglected, Delinquent, and
Understanding Your Praxis Scores
2014 15 Understanding Your Praxis Scores The Praxis Series Assessments are developed and administered by Educational Testing Service (E T S ). Praxis Core Academic Skills for Educators (Core) tests measure
Validity, Fairness, and Testing
Validity, Fairness, and Testing Michael Kane Educational Testing Service Conference on Conversations on Validity Around the World Teachers College, New York March 2012 Unpublished Work Copyright 2010 by
The Official Study Guide
The Praxis Series ebooks The Official Study Guide Middle School Mathematics Test Code: 5169 Study Topics Practice Questions Directly from the Test Makers Test-Taking Strategies www.ets.org/praxis Study
Study Guide for the Mathematics: Proofs, Models, and Problems, Part I, Test
Study Guide for the Mathematics: Proofs, Models, and Problems, Part I, Test A PUBLICATION OF ETS Table of Contents Study Guide for the Mathematics: Proofs, Models, and Problems, Part I, Test TABLE OF CONTENTS
Assessment Centres and Psychometric Testing. Procedure
Assessment Centres and Psychometric Testing Procedure INDEX INTRODUCTION 3 DEFINITION 3 Assessment & Development Centres 3 Psychometric Tests 3 SCOPE 4 RESPONSIBILITIES 5 WHEN SHOULD TESTS BE USED 5 CHOOSING
Responsibilities of Users of Standardized Tests (RUST) (3rd Edition) Prepared by the Association for Assessment in Counseling (AAC)
Responsibilities of Users of Standardized Tests (RUST) (3rd Edition) Prepared by the Association for Assessment in Counseling (AAC) Many recent events have influenced the use of tests and assessment in
A s h o r t g u i d e t o s ta n d A r d i s e d t e s t s
A short guide to standardised tests Copyright 2013 GL Assessment Limited Published by GL Assessment Limited 389 Chiswick High Road, 9th Floor East, London W4 4AL www.gl-assessment.co.uk GL Assessment is
As a reader of ETS s R&D Connections, you have doubtless graduated from high school
No. 18 February 2012 Key Concepts Dropping Out of High School: Prevalence, Risk Factors, and Remediation Strategies Dropping out To quit a course or school without achieving a diploma. Dropout A student
Study Guide for the Physical Education: Content and Design Test
Study Guide for the Physical Education: Content and Design Test A PUBLICATION OF ETS Copyright 2011 by Educational Testing Service. All rights reserved. ETS, the ETS logo, GRE, and LISTENING. LEARNING.
ETS Research Spotlight. Measuring Learning Outcomes in Higher Education: The Role of Motivation
ETS Research Spotlight Measuring Learning Outcomes in Higher Education: The Role of Motivation ISSUE 9, AUGUST 2013 Foreword ETS Research Spotlight ISSUE 9 AUGUST 2013 Senior Editor: Jeff Johnson Contributing
The test uses age norms (national) and grade norms (national) to calculate scores and compare students of the same age or grade.
Reading the CogAT Report for Parents The CogAT Test measures the level and pattern of cognitive development of a student compared to age mates and grade mates. These general reasoning abilities, which
Calculator Use on Stanford Series Mathematics Tests
assessment report Calculator Use on Stanford Series Mathematics Tests Thomas E Brooks, PhD Betsy J Case, PhD Tracy Cerrillo, PhD Nancy Severance Nathan Wall Michael J Young, PhD May 2003 (Revision 1, November
ACT Research Explains New ACT Test Writing Scores and Their Relationship to Other Test Scores
ACT Research Explains New ACT Test Writing Scores and Their Relationship to Other Test Scores Wayne J. Camara, Dongmei Li, Deborah J. Harris, Benjamin Andrews, Qing Yi, and Yong He ACT Research Explains
Code of Professional Responsibilities in Educational Measurement
Code of Professional Responsibilities in Educational Measurement Prepared by the NCME Ad Hoc Committee on the Development of a Code of Ethics: Cynthia B. Schmeiser, ACT--Chair Kurt F. Geisinger, State
Assessing Adult English Language Learners
PART IV: TOPICS IN ADULT ESL EDUCATION & FAMILY LITERACY Learner assessments are used in adult basic education (ABE), adult English as a Second Language (ESL), and family literacy programs for many different
The Official Study Guide
The Praxis Series ebooks The Official Study Guide Pennsylvania Grades 4 8 Subject Concentration: Science Test Code: 5159 Study Topics Practice Questions Directly from the Test Makers Test-Taking Strategies
Tom wants to find two real numbers, a and b, that have a sum of 10 and have a product of 10. He makes this table.
Sum and Product This problem gives you the chance to: use arithmetic and algebra to represent and analyze a mathematical situation solve a quadratic equation by trial and improvement Tom wants to find
Understanding Your Test Record and Profile Chart for the PSB-Nursing School Aptitude Examination (RN)
Understanding Your Test Record and Profile Chart for the PSB-Nursing School Aptitude Examination (RN) This data sheet has been prepared to help you understand the meaning of the scores you earned on the
Using the ETS Proficiency Profile Test to Meet VSA Requirements Frequently Asked Questions
Using the ETS Proficiency Profile Test to Meet VSA Requirements Frequently Asked Questions What is the ETS Proficiency Profile? The ETS Proficiency Profile from ETS is an outcomes assessment of four core
Feifei Ye, PhD Assistant Professor School of Education University of Pittsburgh [email protected]
Feifei Ye, PhD Assistant Professor School of Education University of Pittsburgh [email protected] Validity, reliability, and concordance of the Duolingo English Test Ye (2014), p. 1 Duolingo has developed
Study Guide for the Elementary Education: Content Knowledge Test
Study Guide for the Elementary Education: Content Knowledge Test A PUBLICATION OF ETS Copyright 2011 by Educational Testing Service. All rights reserved. ETS, the ETS logo, GRE, and LISTENING. LEARNING.
Study Guide for the Middle School Science Test
Study Guide for the Middle School Science Test A PUBLICATION OF ETS Copyright 2008 by Educational Testing Service EDUCATIONAL TESTING SERVICE, ETS, the ETS logo, LISTENING., LEARNING., LEADING., and GRE,
Mathematics Concepts and Applications Administration and Scoring Guide
Mathematics Concepts and Applications Administration and Scoring Guide Pearson Executive Office 560 Green Valley Drive Blmington, MN 55437 800.627.727 www.psychcorp.com Copyright 20 2 NCS Pearson, Inc.
Validity, reliability, and concordance of the Duolingo English Test
Validity, reliability, and concordance of the Duolingo English Test May 2014 Feifei Ye, PhD Assistant Professor University of Pittsburgh School of Education [email protected] Validity, reliability, and
Click on the links below to jump directly to the relevant section
Click on the links below to jump directly to the relevant section What is algebra? Operations with algebraic terms Mathematical properties of real numbers Order of operations What is Algebra? Algebra is
Horizontal and Vertical Alignment ...
policy report........ Betsy Case, Ph.D. Sasha Zucker Presentation by Betsy Case, Ph.D., in Beijing, China at the China US Conference on Alignment of Assessments and Instruction July 2005 Introduction Alignment
Graphical Integration Exercises Part Four: Reverse Graphical Integration
D-4603 1 Graphical Integration Exercises Part Four: Reverse Graphical Integration Prepared for the MIT System Dynamics in Education Project Under the Supervision of Dr. Jay W. Forrester by Laughton Stanley
Transadaptation: Publishing Assessments in World Languages
assessment report. : Publishing Assessments in World Languages........ Sasha Zucker Margarita Miska Linda G. Alaniz Luis Guzmán September 2005 : Publishing Assessments in World Languages Introduction In
ETS Automated Scoring and NLP Technologies
ETS Automated Scoring and NLP Technologies Using natural language processing (NLP) and psychometric methods to develop innovative scoring technologies The growth of the use of constructed-response tasks
Principles of Data-Driven Instruction
Education in our times must try to find whatever there is in students that might yearn for completion, and to reconstruct the learning that would enable them autonomously to seek that completion. Allan
Information for teachers about online TOEIC Listening and Reading practice tests from
oxford english testing.com Information for teachers about online TOEIC Listening and Reading practice tests from www.oxford english testing.com oxford english testing.com What do the online TOEIC Listening
GMAT SYLLABI. Types of Assignments - 1 -
GMAT SYLLABI The syllabi on the following pages list the math and verbal assignments for each class. Your homework assignments depend on your current math and verbal scores. Be sure to read How to Use
THE FAIR TESTING IMPERATIVE IN NURSING EDUCATION A Living Document from the National League for Nursing
THE FAIR TESTING IMPERATIVE IN NURSING EDUCATION A Living Document from the National League for Nursing NLN Board of Governors February 2012 INTRODUCTION The National League for Nursing recognizes the
Automated Scoring for the Assessment of Common Core Standards
Automated Scoring for the Assessment of Common Core Standards David M. Williamson, Senior Research Director, Applied Research and Development, ETS Randy E. Bennett, Frederiksen Chair, Assessment Innovation,
Mathematics Computation Administration and Scoring Guide
Mathematics Computation Administration and Scoring Guide Pearson Executive Office 560 Green Valley Drive Blmington, MN 55437 800.627.727 www.psychcorp.com Copyright 20 2 NCS Pearson, Inc. All rights reserved.
Study Guide for the Music: Content Knowledge Test
Study Guide for the Music: Content Knowledge Test A PUBLICATION OF ETS Acknowledgements Educational Testing Service and The Praxis Series wish to extend special thanks to the following for helping with
Fractions as Numbers INTENSIVE INTERVENTION. National Center on. at American Institutes for Research
National Center on INTENSIVE INTERVENTION at American Institutes for Research Fractions as Numbers 000 Thomas Jefferson Street, NW Washington, DC 0007 E-mail: [email protected] While permission to reprint this
Teacher Effectiveness Webinar Series
Teacher Effectiveness Webinar Series Using Teacher Evaluation to Change Teaching Welcome Elizabeth Greninger, Ph.D. Webinar facilitator, edcount, LLC Courtney A. Bell, Ph.D. Presenter, Understanding Teaching
Evaluation: Designs and Approaches
Evaluation: Designs and Approaches Publication Year: 2004 The choice of a design for an outcome evaluation is often influenced by the need to compromise between cost and certainty. Generally, the more
oxford english testing.com
oxford english testing.com The Oxford Online Placement Test: The Meaning of OOPT Scores Introduction oxford english testing.com The Oxford Online Placement Test is a tool designed to measure test takers
Independent samples t-test. Dr. Tom Pierce Radford University
Independent samples t-test Dr. Tom Pierce Radford University The logic behind drawing causal conclusions from experiments The sampling distribution of the difference between means The standard error of
Solving Algebra and Other Story Problems with Simple Diagrams: a Method Demonstrated in Grade 4 6 Texts Used in Singapore
The Mathematics Educator 2004, Vol. 14, No. 1, 42 46 Solving Algebra and Other Story Problems with Simple Diagrams: a Method Demonstrated in Grade 4 6 Texts Used in Singapore Sybilla Beckmann Out of the
2.6 Exponents and Order of Operations
2.6 Exponents and Order of Operations We begin this section with exponents applied to negative numbers. The idea of applying an exponent to a negative number is identical to that of a positive number (repeated
2014 School Counselor Performance Appraisal Rubric SAU 21
Program Planning & Implementation Standard 1: Rating 4 Highly Effective Rating 3 Effective Rating 2 Improvement Comprehensive School Counseling Program School counselors collaboratively plan, implement,
Understanding Your GACE Scores
Understanding Your GACE Scores July 2015 Georgia educator certification is governed by the Georgia Professional Standards Commission (GaPSC). The assessments required for educator certification are administered
Raw Score to Scaled Score Conversions
Jon S Twing, PhD Vice President, Psychometric Services NCS Pearson - Iowa City Slide 1 of 22 Personal Background Doctorate in Educational Measurement and Statistics, University of Iowa Responsible for
Test Bias. As we have seen, psychological tests can be well-conceived and well-constructed, but
Test Bias As we have seen, psychological tests can be well-conceived and well-constructed, but none are perfect. The reliability of test scores can be compromised by random measurement error (unsystematic
Test Item Analysis & Decision Making Offered by the Measurement and Evaluation Center
Test Item Analysis & Decision Making Offered by the Measurement and Evaluation Center 1 Analyzing Multiple-Choice Item Responses Understanding how to interpret and use information based on student test
Development and Validation of the National Home Inspector Examination
Development and Validation of the National Home Inspector Examination Examination Board of Professional Home Inspectors 800 E. Northwest Hwy. Suite 708 Palatine Illinois 60067 847-298-7750 [email protected]
Research Rationale for the Criterion SM Online Writing Evaluation Service
Research Rationale for the Criterion SM Online Writing Evaluation Service ETS invests substantial resources in designing products to improve student learning and in evaluating their effects. One such product
PRELIMINARY ITEM STATISTICS USING POINT-BISERIAL CORRELATION AND P-VALUES
PRELIMINARY ITEM STATISTICS USING POINT-BISERIAL CORRELATION AND P-VALUES BY SEEMA VARMA, PH.D. EDUCATIONAL DATA SYSTEMS, INC. 15850 CONCORD CIRCLE, SUITE A MORGAN HILL CA 95037 WWW.EDDATA.COM Overview
Recent Developments Affecting the Disclosure of Test Data and Materials: Comments Regarding the 1996 Statement on the Disclosure of Test Data 1
Recent Developments Affecting the Disclosure of Test Data and Materials: Comments Regarding the 1996 Statement on the Disclosure of Test Data 1, American Psychological Association In t r o d u c t i o
TEST-DRIVEN accountability is now the
Ten Big Effects of the No Child Left Behind Act on Public Schools The Center on Education Policy has been carefully monitoring the implementation of NCLB for four years. Now Mr. Jennings and Ms. Rentner
SYSTEMS OF EQUATIONS AND MATRICES WITH THE TI-89. by Joseph Collison
SYSTEMS OF EQUATIONS AND MATRICES WITH THE TI-89 by Joseph Collison Copyright 2000 by Joseph Collison All rights reserved Reproduction or translation of any part of this work beyond that permitted by Sections
RE: PCAOB Rulemaking Docket Matter No. 041: Concept Release on Audit Quality Indicators
Office of the Secretary Public Company Accounting Oversight Board 1666 K Street, N.W. Washington, DC 20006-2803 September 29, 2015 RE: PCAOB Rulemaking Docket Matter No. 041: Concept Release on Audit Quality
TEXAS RISING STAR WEBINAR SERIES: CURRICULUM AND EARLY LEARNING GUIDELINES RECORDED OCTOBER 29, 2015 NOTES
TEXAS RISING STAR WEBINAR SERIES: CURRICULUM AND EARLY LEARNING GUIDELINES RECORDED OCTOBER 29, 2015 NOTES The topics that will be addressed during these webinars include: 1. The General Overview: Curriculum
Using Excel 2000 to Create a Weighted-Grade Grade Book
Using Excel 2000 to Create a Weighted-Grade Grade Book This handout assumes that you already have familiarity with creating and copying formulas in Excel 2000. If you do not, you should consult our handout
Information. Security Series. Evaluate Your. Information. Security Program. Survey to find out if you are making progress
Information Security Series 1 of 5 booklets in the series 1 Evaluate Your Information Security Program Survey to find out if you are making progress Date: 01/29/2008 Version: 3.0 County of Sacramento Evaluate
Welcome back to EDFR 6700. I m Jeff Oescher, and I ll be discussing quantitative research design with you for the next several lessons.
Welcome back to EDFR 6700. I m Jeff Oescher, and I ll be discussing quantitative research design with you for the next several lessons. I ll follow the text somewhat loosely, discussing some chapters out
Test Registration Time Format Code
Introduction As you know, teaching English to middle or secondary school students is a rewarding and challenging profession. One way you will demonstrate that you are ready for your teaching license is
What are some effective standards-based classroom assessment practices?
How does classroom assessment help teachers and students? Classroom assessments can help teachers plan and implement effective instruction and can help students learn at deeper and higher levels. Assessments
Communication Process
Welcome and Introductions Lesson 7 Communication Process Overview: This lesson teaches learners to define the elements of effective communication and its process. It will focus on communication as the
Communicating with the EXADEP Program. EXADEP Score Reporting Schedule: 2015 16
1 Communicating with the EXADEP Program Inquiries from Educators By Email [email protected] By Phone 1-787-753-6363 By Fax 1-787-250-7426 By TDD 1-787-758-4598 By Mail EXADEP Program ETS Puerto Rico Office 250
Assessment That Drives Instruction
Improving Instruction Through Assessment Assessment That Drives Instruction Pokey Stanford Stacy Reeves TEACHING Exceptional Children, Vol. 37, No. 4, pp. 18-22. Copyright 2005 CEC. Two challenging aspects
INTRODUCING LANGUAGE TEACHER COGNITION
INTRODUCING LANGUAGE TEACHER COGNITION Simon Borg, Centre for Language Education Research, School of Education, University of Leeds The Origins of Teacher Cognition Research Teacher cognition research
Accountability Brief
Accountability Brief Public Schools of North Carolina State Board of Education North Carolina Department of Public Instruction Michael E. Ward, Superintendent March, 2003 Setting Annual Growth Standards:
Georgia s Technology Literacy Challenge Fund Grant Application. Bremen City Schools. Section 3: Grant Proposal
1 Georgia s Technology Literacy Challenge Fund Grant Application Bremen City Schools Section 3: Grant Proposal Narrative: Due to our system size (total enrollment: 1,430 pre-k to 12) and our belief that
The Role and Responsibilities of the Managing Partner Timothy I. Michel, CPA
The Role and Responsibilities of the Managing Partner Timothy I. Michel, CPA I became the managing partner (MP) in my prior firm after having spent 25 years serving clients as a practice partner. During
Considerations for Using STAR Data with Educator Evaluation: Ohio
Considerations for Using STAR Data with Educator Evaluation: Ohio Purpose Renaissance Learning has developed this document in response to customer requests for information on how to use the data generated
COMMONLY ASKED QUESTIONS ABOUT THE ASI
COMMONLY ASKED QUESTIONS ABOUT THE ASI The Interview Format - Does It Have to be an Interview? This is perhaps the most often asked question regarding the ASI. In the search for faster and easier methods
INTRODUCTION TO READING AND LITERATURE SAMPLE TESTS
INTRODUCTION TO READING AND LITERATURE SAMPLE TESTS The Oregon Department of Education provides sample tests to demonstrate the types of reading selections and questions students at grades 3, 5, 8 and
Guidance paper - The use of calculators in the teaching and learning of mathematics
Guidance paper - The use of calculators in the teaching and learning of mathematics Background and context In mathematics, the calculator can be an effective teaching and learning resource in the primary
Microsoft Office Live Meeting Events User s Guide
Microsoft Office Live Meeting Events User s Guide Information in this document, including URL and other Internet Web site references, is subject to change without notice. Unless otherwise noted, the companies,
Study Guide for the Special Education: Core Knowledge Tests
Study Guide for the Special Education: Core Knowledge Tests A PUBLICATION OF ETS Table of Contents Study Guide for the Praxis Special Education: Core Knowledge Tests TABLE OF CONTENTS Chapter 1 Introduction
Study Guide for the English Language, Literature, and Composition: Content Knowledge Test
Study Guide for the English Language, Literature, and Composition: Content Knowledge Test A PUBLICATION OF ETS Table of Contents Study Guide for the English Language, Literature, and Composition: Content
Understanding Your Texas Educator Certification Program Test Scores
Understanding Your Texas Educator Certification Program Test Scores April 2015 Texas educator certification is governed by the State Board for Educator Certification (SBEC). The tests required for educator
Sample Fraction Addition and Subtraction Concepts Activities 1 3
Sample Fraction Addition and Subtraction Concepts Activities 1 3 College- and Career-Ready Standard Addressed: Build fractions from unit fractions by applying and extending previous understandings of operations
U.S. Department of Education NCES 2011-460 NAEP. Tools on the Web
U.S. Department of Education NCES 2011-460 NAEP Tools on the Web Whether you re an educator, a member of the media, a parent, a student, a policymaker, or a researcher, there are many resources available
