How Do Teacher Value-Added Measures Compare to Other Measures of Teacher Effectiveness?

Similar documents
Value-Added Measures of Educator Performance: Clearing Away the Smoke and Mirrors

Teacher Value-Added and Comparisons with Other Measures of Teacher Effectiveness. Summary

Teacher Value-Added and Credentials as Tools for School Improvement. Douglas N. Harris University of Wisconsin at Madison

Recommendations to Improve the Louisiana System of Accountability for Teachers, Leaders, Schools, and Districts: Second Report

American Statistical Association

Tulsa Public Schools Teacher Observation and Evaluation System: Its Research Base and Validation Studies

Combining Multiple Performance Measures: Do Common Approaches Undermine Districts Personnel Evaluation Systems?

Evaluating Teachers with Classroom Observations Lessons Learned in Four Districts

Doctoral Program in Mathematics Education Outcomes Assessment

SIGNIFICANT CHANGES IN THE NEW ORLEANS TEACHER WORKFORCE

Development of CEL s Teacher Evaluation Rubric

Open Letter to the College Board What an AP Virtual Lab Must Do How Smart Science Labs Match AP Lab Criteria

RATING A TEACHER OBSERVATION TOOL. Five ways to ensure classroom observations are focused and rigorous

Measuring Principal Effectiveness Full Day PIL Session: Framework for Leadership

Supporting State Efforts to Design and Implement Teacher Evaluation Systems. Glossary

Massachusetts Department of Elementary and Secondary Education. Professional Development Self- Assessment Guidebook

Linda Blanton Professor, Florida International University

Overview of NYS Annual Professional Performance Review (aka APPR) LCSD Board of Education Meeting November 26, 2012

QUALITY EVALUATION ROLLOUT DATA ANALYTICS CONVENING OCTOBER 11, 2013

User s Guide for PARCC Assessment Administration Capacity Planning Tool - February 2014

Using Value Added Models to Evaluate Teacher Preparation Programs

Principal Appraisal Overview

Table 1: Number of students enrolled in the program in Fall, 2011 (approximate numbers)

Incorporating the MARS Sales Management Simulation into a Sales Management Course

Transcript: What Is Progress Monitoring?

Department of Political Science. College of Social Science. Undergraduate Bachelor s Degree in Political Science

PUBLIC Law, Chapter 330 LD 1446, item 1, 124th Maine State Legislature An Act To Create the Maine Online Learning Program

Analyzing Marketing Cases

A Guide to Understanding and Using Data for Effective Advocacy

Fact Sheet UPDATED January 2016

South Carolina College- and Career-Ready (SCCCR) Probability and Statistics

Summary of the Case American Intercontinental Universit y1 Master of Education Program

EVALUATING PRINCIPALS. Balancing accountability with professional growth

2. Starting Points--naive (strawperson) analyses, spurious associations?

ASSESSMENT GLOSSARY OF TERMS

Faculty Member Completing this Form: Melissa Cummings

ACT Research Explains New ACT Test Writing Scores and Their Relationship to Other Test Scores

WHAT SCHOOLS DO FAMILIES WANT (AND WHY)?

CSM. Psychology, BS. 3. Students can demonstrate knowledge and understanding in history and systems, research methods, and statistics.

Emerging consensus regarding the goal of preparing all students to

INITIAL PROGRAM APPROVAL TRAINING 2015 TRANSCRIPT

Essential Principles of Effective Evaluation

Stability of School Building Accountability Scores and Gains. CSE Technical Report 561. Robert L. Linn CRESST/University of Colorado at Boulder

Kurt F. Geisinger. Changes in Outcomes Assessment for Accreditation: The Case of Teacher Education in the United States

Professional School Counselor Effectiveness Rubric 2011

National assessment of foreign languages in Sweden

Validation of incentive pay. business results

A Practitioner s Guide to Statistical Sampling in E-Discovery. October 16, 2012

An Analysis of the Essential Abilities and Work Art of Educational Administration Staff in Chinese Universities

DOCTOR OF PHILOSOPHY DEGREE. Educational Leadership Doctor of Philosophy Degree Major Course Requirements. EDU721 (3.

GLOSSARY of Assessment Terms

Assessing MEdT Teacher Candidates Leadership and Collaboration Scott Robinson Masters of Education in Teaching*

Psy 212- Educational Psychology Practice Test - Ch. 1

Understanding Types of Assessment Within an RTI Framework

Teacher Education Comes on Stage

Bangor School Department Grades 9-Diploma Visual Arts Standards

Military History, B.A.

IAB Evaluation Study of Methods Used to Assess the Effectiveness of Advertising on the Internet

Formal and informal Assessment

What is the purpose of this document? What is in the document? How do I send Feedback?

International School Leadership Certificate Module Core Content

Opinion on Summary Judgment

Ten Principles for successful E-learning

A New Series of Papers on Teacher Compensation from the University of Wisconsin CPRE Group

Integrating Technology in the Classroom

Holistic Rating Training Requirements. Texas Education Agency Student Assessment Division

Basic Assessment Concepts for Teachers and School Administrators.

Ohlone College Program Review Report

Causes of Conflicts for Local Information Technology Managers in Multinational Companies

Recipes for Rational Government from the Independent Women s Forum. Carrie Lukas, Managing Director, Independent Women s Forum

COMMISSIONER OF EDUCATION PRESIDENT OF THE UNIVERSITY OF THE STATE OF NEW YORK

University Mission School Mission Department Mission Degree Program Mission

Assessment Target Reports Frequently Asked Questions

THE EFFECT OF SETTLEMENT IN KAPLOW S MULTISTAGE ADJUDICATION

User Manual 03/12/2014. A collaborative effort by

Teacher and Leader Evaluation Requirements An Overview

General Approaches to evaluating student learning Qatar University, 2010

A Guide to Implementing Principal Performance Evaluation in Illinois

Assessment of Student Learning Plan (ASLP): ART Program

A LMS is a software application for the administration, documentation, tracking, reporting and delivery of educational content online.

February 2, Ms. Sophia McArdle U.S. Department of Education 1990 K Street NW Room 8017 Washington, DC Re: Docket ID ED-2014-OPE-0057

How To Know If An Instructional Program Works

Chapter 5, Learning to Think

Dunlap Community Unit School District 323

Frequently Asked Questions Contact us:

Teacher Effectiveness Webinar Series

The Application of Statistics Education Research in My Classroom

Stronge Teacher Effectiveness Performance Evaluation System

Fridley Alternative Compensation Plan Executive Summary

Texas Teacher Evaluation and Support System FAQ

6. Additional Requirements - Teachers Created Tuesday, April 30, 2013 Updated Wednesday, September 18, 2013

Chapter 3 - Alternative Assessments

13 th Annual Texas A & M University Assessment Conference College Station, Texas February 2013

Peer Reviews of Teaching A Best Practices Guide California State University, Northridge

Automated Text Analytics. Testing Manual Processing against Automated Listening

QMB Business Analytics CRN Fall 2015 W 6:30-9:15 PM -- Lutgert Hall 2209

NETS for Teachers: Achievement Rubric

Dixie State College of Utah

Exit Survey Results For Preliminary Education Specialist Teaching Credential

School leadership and student outcomes: Identifying what works and why

Transcription:

How Do Teacher Value-Added Measures Compare to Other Measures of Teacher Effectiveness? Douglas N. Harris Associate Professor of Economics University Endowed Chair in Public Education Carnegie Foundation webinar October 31, 2012

New Teacher Evaluations and New Menu of Measures Value-added to student achievement Classroom observations (structured) Principal evaluations (unstructured) Student evaluations Student Learning Objectives (SLOs)

Objective A key goal for policymakers and practitioners is to develop a system that yields valid and reliable conclusions about teacher performance Put differently, we want to minimize misclassification of teachers, e.g., putting highperforming teachers in the low-performing category Comparing the various measures helps us identify the optimal mix of measures for making accurate performance judgments

Questions What do we know about how alternative teacher effectiveness measures compare? What more needs to be known on this issue? What can t be resolved by empirical evidence on this issue? How, and under what circumstances, does this issue impact the decisions and actions that districts can make on teacher evaluation?

How correlated are value-added measures with other measures? Value-added measures (VA) are weakly correlated with teacher credentials VA measures are correlated with almost all measures now on the table for the new breed of teacher evaluation +0.12 to +0.34 correlation between value-added measures and classroom observation rubrics +0.2 to +0.3 correlation between VA principal eval. (unstructured)

Importance Issue about Interpretation A common reaction: these are weak correlations: and suggest value-added is not a very good measure BUT it is important to be precise what is a weak correlation? Random error reduces the maximum possible correlation, possibly well below +1.0 Specifically, maximum is probably no higher than +0.7.

Other Reasons They Differ Reliability See prior slide Refers to the degree to which the measure is consistent when repeated. If either measure is imperfectly reliable, then this reduces correlations Validity Refers to the degree to which something measures what it claims to measure, at least on average If two measures try to capture the same element of performance, and either is invalid, then correlation reduced Less obvious answer: The two chosen measures may not intend to measure the same thing Implication: Differences between VA and other measures does not necessarily mean that either measure is wrong

Difficulty Determining Which Factors are Most Important Reliability is easier to measure; only requires multiple observations over time under similar conditions Direct test of validity requires a true measure of teacher performance as a comparison However, no true measure exists, so we have to test validity indirectly Some studies implicitly assume that value-added measures are correct

Other Approaches to Assessing Validity in VA Simulations (evidence fairly positive for VA) Experiments (evidence hard to interpret) Testing the assumptions (evidence indicates problems with VA) Other statistical tests (evidence is more supportive of VA)

Summary of Evidence Validity Reliability Value-added Mixed evidence Low-Modest Rel. Classroom obs N.A. Modest Rel. Principal evals N.A. N.A. Student surveys N.A. N.A. SLOs N.A. N.A.

What more needs to be known on this issue? Two main problems with existing evidence Limited evidence about validity and reliability of most measures Easy to (legitimately) criticize VA because there is so much more evidence Evidence is detached from practice 1.Evidence is about measures outside of highstakes settings 2.Numerous practical issues (e.g., applying common framework to all teachers)

What can t be resolved by empirical evidence on this issue? It is up to policymakers to decide what student outcomes we value and therefore what mix of measures of teacher effectiveness is optimal There does not seem to be complete agreement on what good teaching is For example, some emphasize teacher performance in raising students academic skills and others more concerned about motivating and engaging students These skills are related but not the same Classroom observations and student surveys are likely to be more effective measures of motivation/engagement

How does the evidence and discussion here impact decisions? Wide agreement on multiple measures, but idea only gets us so far It s probably not very useful to use measures that add no new information Once we have 6 measures, is it likely that adding a 7 th will lead to different performance classifications? Probably not. Each measure is costly However, different measures may be useful for different purposes Formative versus summative assessments Classroom observations might be considered essential for formative feedback even if they classified teacher performance exactly the same way as other measures

Summary While policymakers should consider the validity and reliability of all their measures, we know more about value-added than others. Value-added measures are positively related to almost all other commonly accepted measures of teacher performance such as principal evaluations and classroom observations. The correlations appear fairly weak, but this is due primarily to lack of reliability in essentially all measures.

Summary The measures should yield different performance results because they are trying to measure different aspects of teaching Using multiple measures can increase reliability; validity is also improved so long as the additional measures capture aspects of teaching we value. Once we have two or three performance measures, the costs of more measures for accountability may not be justified. But additional formative assessments of teachers may still be worthwhile to help these teachers improve.