Reliability and Validity of Methods Used to Gather Evidence

Similar documents
X = T + E. Reliability. Reliability. Classical Test Theory 7/18/2012. Refers to the consistency or stability of scores

Chapter 3 Psychometrics: Reliability & Validity

RESEARCH METHODS IN I/O PSYCHOLOGY

RESEARCH METHODS IN I/O PSYCHOLOGY

Glossary of Terms Ability Accommodation Adjusted validity/reliability coefficient Alternate forms Analysis of work Assessment Battery Bias

Measurement: Reliability and Validity

Sources of Validity Evidence

Reliability and validity, the topics of this and the next chapter, are twins and

Advantages and Disadvantages of Various Assessment Methods

A PARADIGM FOR DEVELOPING BETTER MEASURES OF MARKETING CONSTRUCTS

DATA COLLECTION AND ANALYSIS

Chapter 6 Appraising and Managing Performance. True/False Questions

An Instructor s Guide to Understanding Test Reliability. Craig S. Wells James A. Wollack

Assessment, Case Conceptualization, Diagnosis, and Treatment Planning Overview

Test Reliability Indicates More than Just Consistency

Measurement: Reliability and Validity Measures. Jonathan Weiner, DrPH Johns Hopkins University

Guided Reading 9 th Edition. informed consent, protection from harm, deception, confidentiality, and anonymity.

Validity and Reliability in Social Science Research

Test-Retest Reliability and The Birkman Method Frank R. Larkey & Jennifer L. Knight, 2002

Glossary of Testing, Measurement, and Statistical Terms

MAY Legal Risks of Applicant Selection and Assessment

a. Will the measure employed repeatedly on the same individuals yield similar results? (stability)

Reliability Overview

Perception and Individual Decision Making

PEER REVIEW HISTORY ARTICLE DETAILS VERSION 1 - REVIEW. Tatyana A Shamliyan. I do not have COI. 30-May-2012

Standard Deviation Estimator

Technical Report. Overview. Revisions in this Edition. Four-Level Assessment Process

Development and Validation of the National Home Inspector Examination

THE ACT INTEREST INVENTORY AND THE WORLD-OF-WORK MAP

Designing a Questionnaire

Assessment of children s educational achievements in early childhood education

WHAT IS A JOURNAL CLUB?

Using Personality to Predict Outbound Call Center Job Performance

Statistics. Measurement. Scales of Measurement 7/18/2012

Chapter Five: RELATIONSHIP DYNAMICS

Outcomes of Preservice Teacher s Technology Use

Developing and Using a Computer Self-Efficacy Scale for Adults. James H. Brown Doctoral Student, Urban Education University of Wisconsin-Milwaukee

Measuring Teacher Technology Integration: Going Beyond Teacher Use

National assessment of foreign languages in Sweden

Assurance of Learning Assessment Process

oxford english testing.com

IMPLEMENTATION NOTE. Validating Risk Rating Systems at IRB Institutions

KNOWLEDGE AND EDUCATION NURSING PRACTICE EXECUTIVE SUMMARY AT ENTRY TO IN ALBERTA

Best Practices. Create a Better VoC Report. Three Data Visualization Techniques to Get Your Reports Noticed

Partial Estimates of Reliability: Parallel Form Reliability in the Key Stage 2 Science Tests

Test Bias. As we have seen, psychological tests can be well-conceived and well-constructed, but

Validity, Fairness, and Testing

This chapter will demonstrate how to perform multiple linear regression with IBM SPSS

Dividend Yield and Stock Return in Different Economic Environment: Evidence from Malaysia

standardized tests used to assess mental ability & development, in an educational setting.

Assessment Terminology for Gallaudet University. A Glossary of Useful Assessment Terms

Theoretical Biophysics Group

Chapter. Understanding Measurement. Chapter. Outline. Key Terms

Technical Information

INDIVIDUAL MASTERY for: St#: Test: CH 9 Acceleration Test on 09/06/2015 Grade: A Score: % (38.00 of 41.00)

INDIVIDUAL MASTERY for: St#: Test: CH 9 Acceleration Test on 29/07/2015 Grade: B Score: % (35.00 of 41.00)

The Evolution of Leadership Theory

Integrating HR & Talent Management Processes

Research Methods & Experimental Design

Social Perception and Attribution

2016 Edenred-Ipsos Barometer Understand and improve Wellbeing At Work. May 2016

Governors State University

True Score Theory Measurement Error Theory of Reliability Types of Reliability Reliability & Validity

Role Expectations Report for Sample Employee and Receptionist

Understanding Your Test Record and Profile Chart for the PSB-Nursing School Aptitude Examination (RN)

classroom Tool Part 3 of a 5 Part Series: How to Select The Right

Scale Construction and Psychometrics for Social and Personality Psychology. R. Michael Furr

Department of. Curriculum, Foundations, and Reading. Curriculum, Foundations, and Reading. Degrees. Endorsement. Doctoral Degrees

Employee Performance Management Training

Program Quality Assessment. William Wiener

3. From Student Learning Assessment: Options and Resources, (Philadelphia, PA: Middle States Commission on Higher Education, 2003: p.

Performance Management at

Reliability Analysis

Canonical Correlation Analysis

How to Get More Value from Your Survey Data

CHAPTER 7 PRESENTATION AND ANALYSIS OF THE RESEARCH FINDINGS

North Dakota Human Resource Management Services Performance Evaluation

How Professionally Skeptical are Certified Fraud Examiners? Joel Pike Gerald Smith

Basic Assessment Concepts for Teachers and School Administrators.

Graduate Student Perceptions of the Use of Online Course Tools to Support Engagement

MAT Reliability and Validity

Effectiveness of Customer Relationship Management Programs

Procedures for Estimating Internal Consistency Reliability. Prepared by the Iowa Technical Adequacy Project (ITAP)

CHAPTER - V SUMMARY, CONCLUSION AND RECOMMENDATIONS

Coni Francis, PhD, RD Nutrition and Dietetics Program University of Northern Colorado

The Influence of Trust and Commitment on Customer Relationship Management Performance in Mobile Phone Services

A Basic Guide to Analyzing Individual Scores Data with SPSS

Research Report for Adaptive Testing Assessment

Validation of the Chally Assessment System with a Sample of Sales Representatives

CHAPTER 8 PERFORMANCE APPRAISAL CHAPTER DESCRIPTION

History and Purpose of the Standards for Educational and Psychological Testing

Bachelor's Degree in Business Administration and Master's Degree course description

Prospect Theory Ayelet Gneezy & Nicholas Epley

Autism Spectrum Disorder Performance Standards and Evaluation Criteria Rubric

Reporting and Interpreting Scores Derived from Likert-type Scales

ADJUSTING SERVQUAL MODEL IN A HIGH EDUCATION LIBRARY SERVICE

Transcription:

Reliability and Validity of Methods Used to Gather Evidence Reliability and Validity narrow down the pool of possible summative and formative methods: Reliability refers to the extent to which trial tests of a method with representative student populations fairly and consistently assess the expected traits or dimensions of student learning within the construct of that method. Validity refers to the extent to which a method prompts students to represent the dimensions of learning desired. A valid method enables direct and accurate assessment of the learning described in outcome statements. (Assessing for Learning: Building a sustainable commitment across the institution by Maki 2004) Reliability Reliable measures can be counted on to produce consistent responses over time: Reliable data variance in scores is attributable to actual differences in what is being measured, such as knowledge, performance or attitudes Unreliable data score variance is due to measurement error; which can include such things as the individuals responding to the instrument, the administration and scoring of the instrument, and the instrument itself (Assessment Essentials: planning, implementing, and improving assessment in higher education by Palomba and Banta 1999) Sources of error in regard to reliability: (i) from the student lack of motivation to take assessment seriously prior experience with being evaluated test anxiety, coaching, and physiological variables (ii) from the assessment instrument test items ambiguously worded rating scale is confusing or vague

representativeness of test items; e.g., if two test forms differ in their emphasis on program content, inconsistent scores may result (iii) from the conditions of assessment administration varying the style of test administration procedures produces varying scores The greater the error in any assessment information, the less reliable it is, and the less likely it is to be useful. Types of reliability : (i) (ii) (iii) (iv) Stability usually described in a test manual as test-retest reliability; if the same test is readministered to the same students within a short period of time, their scores should be highly similar, or stable Equivalence the degree of similarity of results among alternate forms of the same test; tests should have high levels of equivalence if different forms are offere Homogeneity or internal consistency the interrelatedness of the test items used to measure a given dimension of learning and development Interrater reliability the consistency with which raters evaluate a single performance of a given group of students Barriers to establishing reliability include rater bias the tendency to rate individuals or objects in an idiosyncratic way: central tendency error in which an individual rates people or objects by using the middle of the scale leniency error in which an individual rates people or objects by using the positive end of the scale severity error in which an individual rates people or objects by using the negative end of the scale halo error when a rater s evaluation on one dimension of a scale (such as work quality) is influenced by his or her perceptions from another dimension (such as punctuality)

Major Types of Reliability (Assessing Academic Programs in Higher Education by Allen 2004) Test-retest reliability Parallel forms reliability (or alternate forms reliability) A reliability estimate based on assessing a group of people twice and correlating the two scores. This coefficient measures score stability. A reliability estimate based on correlating scores collected using two versions of the procedure. This coefficient indicates score consistency across the alternative versions. Inter-rater reliability How well two or more raters agree when decisions are based on subjective judgments. Internal consistency reliability Coefficient alpha A reliability estimate based on how highly parts of a test correlate with each other. An internal consistency reliability estimate based on correlations among all items on a test. Split-half reliability Validity An internal consistency reliability estimate based on correlating two scores, each calculated on half of a test. Valid measures on ones in which the instrument measures what we want it to measure: Construct-related validity refers to the congruence between the meaning of the underlying construct and the items on the test or survey; i.e., do results correlate with other instruments examining the same construct? Criterion-related validity includes predictive validity: how dependable is the relationship between the scores or answers on an instrument and a particular future outcome? Content-related validity refers to the match between the content of the instrument and the content of the curriculum or other domain of interest (Assessment Essentials: planning, implementing, and improving assessment in higher education by Palomba and Banta 1999) Validity must be judged according to the application of each use of the method. The validity of an assessment method is never proved absolutely; it can only be supported by an accumulation of evidence from several categories. For any assessment methods to be used in decision making, the following categories should be considered: Content relevance and representativeness o The selected test should be a representative sample from those educational objectives which the test is supposed to measure o The test should cover what the program covered and should place emphasis in proportion to the program s emphases

o Tests may be reliable but not valid for a particular program Internal test structure o Typically demonstrated through intercorrelations among items covering the same content domain External test structure o Necessary when the educator wishes to compare test scores or ratings with other measures or related variables Process of probing responses o Typically sought at two points during any test or scale construction: initially in the test construction to determine whether the students interpretations are consistent with the intent of the test designer; and at the point of probing the process to see if a pattern might be discovered on those students who scored very high or very low Test s similarities and differences over time and across groups and settings o In studying validity evidence over time, some outcome measures should increase over time Value implications and social consequences o If a test or rating scale discriminates against certain groups of people, that test or scale should be considered suspect. Validity: involves establishing that an assessment measures what it is supposed to measure can be thought of as the extent of the relationship between an assessment and the construct the assessment is supposed to predict Major Types of Validity (Assessing Academic Programs in Higher Education by Allen 2004) Construct validity Criterion-related validity Construct validity is examined by testing predictions based on the theory (or construct) underlying the procedure. For example, faculty might predict that scores on a test that assesses knowledge of anthropological terms will increase as anthropology students progress in their major. We have more confidence in the test s construct validity if predictions are empirically supported. Criterion-related validity indicates how well results predict a phenomenon of interest, and it is based on correlating assessment results with this criterion. For example, scores on an admissions test can be correlated with college GPA to demonstrate criterion-related validity.

Face validity Formative validity Sampling validity Face validity is assessed by subjective evaluation of the measurement procedure. This evaluation may be made by test takers or by experts in what is being assessed. Formative validity is how well an assessment procedure provides information that is useful for improving what is being assessed. Sampling validity is how well the procedure s components, such as test items, reflect the full range of what is being assessed. For example, a valid test of content mastery should assess information across the entire content area, not just isolated segments.