Dr. Deborah A. Brady Sheila Muir DESE



Similar documents
Understanding District-Determined Measures

Technical Guide B: Measuring Student Growth & Piloting District-Determined Measures

Math Science Partnership (MSP) Program: Title II, Part B

Massachusetts District-Determined Measures. The Arts

INTER-OFFICE CORRESPONDENCE LOS ANGELES UNFIED SCHOOL DISTRICT LOCAL DISTRICT EAST

Revisioning Graduate Teacher Education in North Carolina Master of Arts in Elementary Education Appalachian State University

Migrant Education Program Evaluation Toolkit A Tool for State Migrant Directors. Summer 2012

Massachusetts Department of Elementary and Secondary Education. Professional Development Self- Assessment Guidebook

International Baccalaureate Diploma Assessment Policy Holy Heart of Mary High School

Colorado High School Graduation Guidelines

Research Findings on the Transition to Algebra series

District-Determined Measures: Overview and Guidance

State Assessments 2015: MCAS or PARCC? William H. Lupini Superintendent of Schools Thursday, May 22, 2014

Advancing Professional Excellence Guide Table of Contents

Springfield Public Schools English Language Learner Recommended Actions and Implementation Plans

Program Models. proficiency and content skills. After school tutoring and summer school available

Teacher Education Portfolio Guidelines and Rubric

Nutrition and Culinary Arts I. Example SLO

Advantages and Disadvantages of Various Assessment Methods

Integrating Technology in the Classroom

Teacher Performance Evaluation System

TITLE 23: EDUCATION AND CULTURAL RESOURCES SUBTITLE A: EDUCATION CHAPTER I: STATE BOARD OF EDUCATION SUBCHAPTER b: PERSONNEL

Texas Principal Evaluation and Support System FAQ

Delaware Performance Appraisal System Second Edition (DPAS II)

Higher Performing High Schools

Oregon Framework for Teacher and Administrator Evaluation and Support Systems

PARCC Recommendation for Presentation to School Committee 09/17/2014

Tackling the NEW Teacher Evaluation Guidelines

Leader Keys Effectiveness System Implementation Handbook

Teacher Evaluation. Missouri s Educator Evaluation System

Massachusetts District-Determined Measures. World Languages and Foreign Languages

NJTESOL/NJBE Spring Conference May 27 th and 28 th, 2015 Hot Topics in ESL Secondary Education

Manchester Essex Regional School District District Curriculum Accommodation Plan (DCAP)

TITLE 23: EDUCATION AND CULTURAL RESOURCES SUBTITLE A: EDUCATION CHAPTER I: STATE BOARD OF EDUCATION SUBCHAPTER b: PERSONNEL

Methodology Teaching in a Digital Age Phase 1

Georgia s Technology Literacy Challenge Fund Grant Application. Bremen City Schools. Section 3: Grant Proposal

Texas Teacher Evaluation and Support System FAQ

IDAHO CONNECTS ONLINE SCHOOL

SCHOOL IMPROVEMENT GRANT (SIG) PRACTICE:

Illinois State Board of Education

Fridley Alternative Compensation Plan Executive Summary

Department of Secondary Education Kutztown University of Pennsylvania. Master s Degree Portfolio Project

Mississippi Statewide Teacher Appraisal Rubric M-STAR

Master s Degree Portfolio Project

Please use the guidance provided in addition to this template to develop components of the SLO and populate each component in the space below.

Career and Technical Education Policy Proposal David W. Gordon, Superintendent Sacramento County Office of Education.

Masters Comprehensive Exam and Rubric (Rev. July 17, 2014)

Chapter 3 - Alternative Assessments

Education Research Brief

A Guide to Curriculum Development: Purposes, Practices, Procedures

Field Guide for the Principal Performance Review. A Guide for Principals and Evaluators

Master Scheduling for Inclusive Practices Voices Over

Annual Report on Curriculum, Instruction, and Student Achievement Independent School District 700, Hermantown Community Schools

BUILDING CURRICULUM ACCOMMODATION PLAN

Practicum/Internship Handbook. Office of Educational Field Experiences

James Rumsey Technical Institute Employee Performance and Effectiveness Evaluation Procedure

St. Charles School District. Counselor Growth Guide and. Evaluation Documents

Overview of NYS Annual Professional Performance Review (aka APPR) LCSD Board of Education Meeting November 26, 2012

Essential Principles of Effective Evaluation

WORLD S BEST WORKFORCE PLAN

Teacher and Leader Evaluation Requirements An Overview

Orange County Public Schools. State Accountability System September 2, 2014

Illinois State Board of Education

THE SCHOOL BOARD OF ST. LUCIE COUNTY, FLORIDA TEACHER PERFORMANCE APPRAISAL SYSTEM, TABLE OF CONTENTS

Appendix E. Role-Specific Indicators

MARZANO SCHOOL LEADERSHIP EVALUATION MODEL

Student Learning Objective (SLO) Template

Hiawatha Academies School District #4170

Fair dismissal district means any common or union high school district or education service district.

How To Get An Individualized Technical Studies Associate Degree At Northcentral Technical College

The performance assessment shall measure the extent to which the teacher s planning:

Section 7: The Five-Step Process for Accommodations for English Language Learners (ELLs)

TESTING AND THE NEW FLORIDA STANDARDS

Walpole Public Schools Leadership Council Goals Date: 7/1/ Plan # 8

SOCIAL WORK PROGRAM ASSESSMENT REPORT

Frequently Asked Questions Contact us:

Technical Review Coversheet

QUESTIONS AND ANSWERS MassCore Updated October 16, 2015

Proficient 2 (80-89%)

Middle Grades Action Kit How To Use the Survey Tools!

Missouri School Counselor Performance Assessment (MoSCPA) Task Requirements

ANNUAL REPORT ON CURRICULUM, INSTRUCTION AND STUDENT ACHIEVEMENT

Wisconsin Educator Effectiveness System. Principal Evaluation Process Manual

CITY OF ST. CHARLES SCHOOL DISTRICT PERFORMANCE-BASED GUIDANCE COUNSELOR EVALUATION

08X540. School For Community Research and Learning 1980 Lafayette Avenue School Address: Bronx, NY 10473

Independent Study Model Application for Professional Development Credit

NORFOLK PUBLIC SCHOOLS SUMMATIVE SPECIAL EDUCATION TEACHER EVALUATION. Summative Evaluation Directions

TENNESSEE STATE BOARD OF EDUCATION

writing standards aligned IEPs.

S=Specific, M=Measurable, A=Appropriate, R=Realistic & Rigorous, T=Timebound

Æ Please note that we are continuing to make adjustments to our program to

Frequently Asked Questions. Spring 2013 Smarter Balanced Pilot

Name of Grant Program: Financial Literacy Pilot Program Fund Code: 740 PART III REQUIRED PROGRAM INFORMATION

State of Texas Assessments of Academic Readiness. Parent & Student Information

Indiana s Department of Education STEM Education Implementation Rubric

The RETELL Initiative: Rethinking Equity & Teaching for English Language Learners

2. Explain how the school will collect and analyze student academic achievement data.

FY15 DANVERS PUBLIC SCHOOLS DISTRICT GOALS THEME I ACHIEVEMENT

CPM High Schools California Standards Test (CST) Results for

Transcription:

Dr. Deborah A. Brady Sheila Muir DESE

By the end of today, participants will: 1. Understand the impact of DDMs on educator evaluation 2. Understand quality expectations and assessment criteria for DDM assessments 3. Begin to develop DDMs that will be used in 2014-2015

District Determined Measures DEFINITION Measures of student learning, growth, and achievement related to the Curriculum Frameworks, that are comparable across grade or subject level district-wide

The Educator Evaluation Framework Everyone earns two ratings Summative Performance Rating Student Impact Rating Exemplary Proficient Needs Improvement Unsatisfactory High Moderate Low Massachusetts Department of Elementary & Secondary Education

The Educator Evaluation Framework Everyone earns two ratings Ratings are obtained through data collected from observations, walk-throughs and artifacts Ratings are based on trends and patterns in student learning, growth and achievement over a period of at least 2 years. Data gathered from DDM s and State-wide testing Summative Performance Rating Student Impact Rating Massachusetts Department of Elementary & Secondary Education

Summative Rating Exemplary Proficient 1-yr Self- Directed Growth Plan 2-yr Self-Directed Growth Plan Needs Improvement Directed Growth Plan Unsatisfactory Improvement Plan Impact Rating on Student Performance Low Moderate High Rating of Impact on Student Learning Massachusetts Department of Elementary and Secondary Education 6

The purpose of DDMs is to assess Teacher Impact The student scores, the Low, Moderate, and High growth rankings are totally internal DESE (in two years) will see: Summative Rating (Exemplary, Proficient, Needs Improvement or Unsatisfactory) AND Impact on Student Learning overall L, M or H for each educator (based on two year trend for at least two measures)

Year Measure Student Results Year 1 MCAS SGP, grade 8 English language arts Low growth Year 1 Writing assessment Moderate growth Year 2 MCAS SGP, grade 8 English language arts Moderate growth Year 2 Portfolio assessment Moderate growth

Key Messages 9 Massachusetts Department of Elementary & Secondary Education

The Role of DDMs To provide educators with an opportunity to: Understand student knowledge and learning patterns more clearly Broaden the range of what knowledge and skills are assessed and how learning is assessed Improve educator practice and student learning Provide educators with feedback about their performance with respect to professional practice and student achievement Provide evidence of an educators impact on student learning

District Determined Measures Regulations: Every educator will need data from at least 2 different measures (different courses or different assessments within the same course) Trends must be measured over a course of at least 2 years One measure must be taken from State-wide testing data such as MCAS if available One measure must be taken from at least one District Determined Measure

The Development of DDMs Timeline 2013-2014 District-wide training, development of assessments and pilot 2014-2015 All educators must have 2 DDMs in place and collect the first year s data 2015-2016 Second year data is collected and all educators receive an impact rating that is sent to DESE

Is the measure aligned to content? Does it assess what is most important for students to learn and be able to do? Does it assess what the educators intend to teach?

Is the measure informative? Do the results of the measure inform educators about curriculum, instruction, and practice? Does it provide valuable information to educators about their students? Does it provide valuable information to schools and districts about their educators?

Does it measure growth? Student growth scores provide greater insight into student learning than is possible through the sole use of single-point-in-time student achievement measures. DDMs that measure growth help to even the playing field for educators allowing educators who teach students who start out behind have a similar opportunity to demonstrate their impact on student learning as educators who teacher students who start out ahead.

GROWTH SCORES for Educators Will Need to Be Tabulated for All Locally Developed Assessments MCAS SGP 244/ 25 SGP 4503699 230/ 35 SGP 225/ 92 SGP

Comparable within a grade, subject, or course across schools within a district. Identical measures are recommended for educators with the same job, e.g., all 5 th grade teachers in a district. Impact Ratings should have a consistent meaning across educators; therefore, DDMs should not have significantly different levels of rigor

Exceptions: When might assessments not be identical? Different content (different sections of Algebra I) Differences in untested skills (reading and writing on math test for ELL students) Other accommodations (fewer questions to students who need more time) NOTE: Roster Verification will allow teachers to verify that all of these students were in her/his supervision The number of students that comprises a class that is too small has not yet been determined by DESE.

District Determined Measures Direct Measures Indirect Measures

Direct Measures Assess student learning, growth, or achievement with respect to specific content Strongly preferred for evaluation because they measure the most immediately relevant outcomes from the education process

Indirect Measures Provide information about students from means other than student work May include student record information (e.g., grades, attendance or tardiness records, or other data related to student growth or achievement such as high school graduation or college enrollment rates) To be considered for use as DDMs, a link (relationship) between indirect measures and student growth or achievement must be established

Direct Measures Portfolio assessments Repeated measures Holistic Evaluation Approved commercial assessments District developed pre- and post-unit and course assessments Capstone projects

Direct Measures If a portfolio is to be used as a DDM that measures growth, it must be designed to capture progress rather than to showcase accomplishments. 23

Direct Measures Description: Multiple assessments given throughout the year Example: Running records, mile run Measuring Growth: Graphically Ranging from the sophisticated to simple Considerations: Authentic Tasks Avoid repeated measures on which students may demonstrate improvement over time simply due to familiarity with the assessment

# of errors Direct Measures Running Record Error Rate Low Growth 70 60 High Growth Mod Growth 50 40 30 20 10 0 Date of Administration 25

Description: Assess growth across student work collected throughout the year. Example: Tennessee Arts Growth Measure System Direct Measures Measuring Growth: Growth rubric should include detailed descriptions of what growth looks like across the examples and not the quality at any individual point. Considerations: Option for multifaceted performance assessments Rating can be challenging & time consuming The approach would also be valuable when the students in a particular grade, subject, or course have quite varied levels of ability but are working on a common task.

Direct Measures 1 2 3 4 Details No improvement in the level of detail. One is true * No new details across versions * New details are added, but not included in future versions. * A few new details are added that are not relevant, accurate or meaningful Modest improvement in the level of detail One is true * There are a few details included across all versions * There are many added details are included, but they are not included consistently, or none are improved or elaborated upon. * There are many added details, but several are not relevant, accurate or meaningful Considerable Improvement in the level of detail All are true * There are many examples of added details across all versions, * At least one example of a detail that is improved or elaborated in future versions *Details are consistently included in future versions *The added details reflect relevant and meaningful additions Outstanding Improvement in the level of detail All are true * On average there are multiple details added across every version * There are multiple examples of details that build and elaborate on previous versions * The added details reflect the most relevant and meaningful additions Example taken from Austin, a first grader from Anser Charter School in Boise, Idaho. Used with permission from Expeditionary Learning. Learn more about this and other examples at http://elschools.org/studentwork/butterfly-drafts 27

Direct Measures While some state and commercial assessments use sophisticated statistical methods for calculating growth with a post-test only (such as the MCAS), this approach is unlikely to be feasible or appropriate for a locally-developed academic measure.

Direct Measures Description: The same or similar assessments administered at the beginning and at the end of the course or year Measuring Growth: Difference between pre- and post-test. Considerations: Students should not take a pre-test that is not valuable to their learning. We don t want this to be an exercise in measuring growth for DDMs but a useful learning experience in its own right. Do all students have an equal chance of demonstrating growth?

Direct Measures Advantages: Authentic Closely tied to curriculum Drawbacks: Must assess only the knowledge and skills explicitly taught as part of the curriculum Difficult to measure growth

Description: A single assessment or data that is paired with other information Measuring Growth, where possible: Use a baseline Assume equal beginning Considerations: May be only option for some indirect measures What is the quality of the baseline information? Indirect Measures

Measure growth Instrument items collectively show a range of difficulty from easy to hard so that the instrument shows no floor or ceiling effects as either a pretest or a posttest Alignment to curriculum Frameworks CCOs (Core Curriculum Objectives from DESE) http://www.doe.mass.edu/edeval/ddm/example/ Includes writing to text Assessment security

Development of a common administration procedure addressing Clarifying prompts Responses to student queries Students not following directions Length of time to complete Time of year administered Development of a common scoring process Common rubric Exemplars Assure comparability of assessments Reliability Validity Non-Bias

Ensure that Assessments Inform instruction primary goal Meet requirements of DDMs secondary goal Consider the minimum number of assessments necessary Each educator must have 2 One should measure a year s growth The other may be a unit assessment Graduation requirement courses

Assessments must reflect a range of cognitive demands and a range of item difficulties (from easier to harder items) so that they discriminate appropriately between strong and weak performers. Existing Sources of DDMs: Consider modifying assessments that you already use Consider CEPAs (Curriculum Embedded Performance Assessments) in DESE Model Curriculum Units http://www.doe.mass.edu/candi/model/default.html Consider Commercial Assessments Galileo for mathematics and science

The important part of this process needs to be the focus: Your discussions about student learning with colleagues Your use of assessment data to inform instruction Your discussions about student learning with your evaluator An ongoing process

Checklist Tracker

Please complete to the best of your ability before the end of the day today Return to Office Will be used to Identify groups who need assistance Ensure rigor Ensure comparability

Sch ool MEPID Last Name First Name HS 07350 Smith Abagail 1 0 G r a d e Subject Course Cours e ID ELA Grade 10 ELA HS 07351 Smith Abagail 9 ELA World Studies HS 07351 Smith Abagail 9 ELA Grade 9 ELA HS 07352 Smith Brent 1 Math IMM 2 0 HS 07352 Smith Brent 1 0 HS 07353 Smith Cathy 1 1 Potential DDM1 01051 MCAS ELA 10 NO 01058 Writing to text 01051 Math IMM 1 MCAS MATH10 NO Social Science World Studies 04053 Potential DDM2 Potential DDM3 HS 07354 Smith Deb 1 2 HS 07355 Smith Emily 1 2 Enginee ring Scienc e Engineerin g Science

Please complete to the best of your ability before the end of the day today Return to office Will be used to ensure that all educators have 2 DDMs for 2014-2015

Perfect is the enemy of good. Voltaire

DESE DDM Technical Guides http://www.doe.mass.edu/edeval/ddm/ Model Curriculum Units for Curriculum Embedded Performance Assessments (CEPAs) with rubrics http://www.doe.mass.edu/candi/model/ DESE Core Curriculum Objectives http://www.doe.mass.edu/edeval/ddm/example/ Delaware Rubrics http://www.doe.k12.de.us/aab/english_language_arts/writin g_rubrics.shtml