Usability of Montana PCMH 2014 Quality of Care and Utilization Data

Similar documents
HEDIS/CAHPS 101. August 13, 2012 Minnesota Measurement and Reporting Workgroup

The Promise of Regional Data Aggregation

Proven Innovations in Primary Care Practice

Atrius Health ACO Initiative. Agenda

Centers for Medicare & Medicaid Services Quality Measurement and Program Alignment

Maureen Mangotich, MD, MPH Medical Director

ESSENTIA HEALTH AS AN ACO (ACCOUNTABLE CARE ORGANIZATION)

Chapter Three Accountable Care Organizations

LEARNING WHAT WORKS AND INCREASING KNOWLEDGE

CMS Quality Measurement and Value Based Purchasing Programs Kate Goodrich, MD MHS Director, Quality Measurement and Health Assessment Group, CMS

Medicare Shared Savings Program Quality Measure Benchmarks for the 2015 Reporting Year

ACO Operational Innovations Featuring the Winners of NAACOS Call for Innovation

HEALTH PRICE TRANSPARENCY IMPROVEMENTS FOR TEXAS CONSUMERS APRIL 15, 2014 TEXAS DEPARTMENT OF INSURANCE STAKEHOLDER MEETING

The Montefiore ACO and Behavioral Health Integration: A Work in Progress. Henry Chung, MD Bruce Schwartz, MD

Department of Health Services. Behavioral Health Integrated Care. Health Home Certification Application

Performance Measurement in CMS Programs Kate Goodrich, MD MHS Director, Quality Measurement and Health Assessment Group, CMS

Meaningful Use Qualification Plan

Finding Common Ground: Vermont s Blueprint for Health and ACO Shared Savings Programs

2013 ACO Quality Measures

CCO Incentive Metrics: Requirements for Reporting on EHR- Based Measures in 2015 GUIDANCE DOCUMENTATION

1. How are you using health IT enabled clinical quality measures for internal quality improvement efforts and patients care?

A Primer on Ratio Analysis and the CAH Financial Indicators Report

HEDIS 2012 Results

Oregon Standards for Certified Community Behavioral Health Clinics (CCBHCs)

See page 331 of HEDIS 2013 Tech Specs Vol 2. HEDIS specs apply to plans. RARE applies to hospitals. Plan All-Cause Readmissions (PCR) *++

MaineCare Value Based Purchasing Initiative

What is an Accountable Care Organization & Why is it Important to Your Home Infusion Company?

Arkansas Blue Cross and Blue Shield Patient Centered Medical Home Provider Manual

A COMPARISON OF MEDI-CAL MANAGED CARE P4P MEASURE SETS

Vermont ACO Shared Savings Program Quality Measures: Recommendations for Year 2 Measures from the VHCIP Quality and Performance Measures Work Group

A predictive analytics platform powered by non-medical staff reduces cost of care among high-utilizing Medicare fee-for-service beneficiaries

ACO Program: Quality Reporting Requirements. Jennifer Faerberg Mary Wheatley April 28, 2011

CQMs. Clinical Quality Measures 101

The role of t he Depart ment of Veterans Affairs (VA) as

Health Information Technology in the United States: Information Base for Progress. Executive Summary

CHAPTER 114. AN ACT establishing a Medicaid Accountable Care Organization Demonstration Project and supplementing Title 30 of the Revised Statutes.

Improving Quality of Care for Medicare Patients: Accountable Care Organizations

Demonstrating Meaningful Use Stage 1 Requirements for Eligible Providers Using Certified EMR Technology

Improving Quality of Care for Medicare Patients: Accountable Care Organizations

Appendix 2. PCMH 2014 and CMS Stage 2 Meaningful Use Requirements

Using Predictive Analytics to Reduce COPD Readmissions

ACO Name and Location Allina Health Minneapolis, Minnesota

Explanation of care coordination payments as described in Section of the PCMH provider manual

Assertive Community Treatment (ACT) Providing Health Home Care Management Interim Instruction: February 19, 2014

Accountable Care Organizations (ACO) Proposed Rule Summary March 31, 2011

US Hospital Information Systems Overview and Outlook, Managing Information in an Era of Reform

Vermont ACO Shared Savings Program: Recommendations for Year 2 Quality Measures

Ophthalmology Meaningful Use Attestation Guide Stage Edition

Main Section of the Proposal

8/14/2012 California Dual Demonstration DRAFT Quality Metrics

Frequently Asked Questions (FAQs) about the Home Health Compare (HHC) Star Ratings

2015 Data Validation Strategy

DSRIP QUARTERLY REVIEW PROCESS: Project Requirement - Timeframe. Project Requirement - Unit Level Reporting

Contact: Barbara J Stout RN, BSC Implementation Specialist University of Kentucky Regional Extension Center

Health Care Reform Update January 2012 MG LILLY USA, LLC. ALL RIGHTS RESERVED

HCUP KIDS INPATIENT DATABASE FILE COMPOSITION BY STATE

UB-04 Claim Form Instructions

Welcome The AAMC, UHC and FPSC Web Conference on 2014 PQRS Proposed Changes will begin shortly.

THE EVOLUTION OF CMS PAYMENT MODELS

Value-Based Payment and Health System Transformation

VIII. Dentist Crosswalk

March 12, Attention: HIT Policy Committee Meaningful Use Comments. CMS-0033-P, Proposed Rules, Electronic Health Record (EHR) Incentive Program

CMS Next Generation ACO Model. Payment Models Work Group April 20 th, 2015

An Introduction to HealthInfoNet s HIE Reporting & Analytics. 6th Annual APS Healthcare Maine Conference May 14, 2015

Total Cost of Care and Resource Use Frequently Asked Questions (FAQ)

Medical Records Law: Regulatory Issues Meaningful Use? EHR v. EMR

ACO s as Private Label Insurance Products

Small Physician Groups Aim High

Inpatient Psychiatric Facility Quality Reporting Program Manual

Iowa Medicaid Integrated Health Home Provider Agreement General Terms

QUESTIONS AND RESPONSES FOR: RFP: Patient Centered Medical Home Consultant

Mental Health and Substance Abuse Services in Medicaid and SCHIP in Colorado

Transcription:

Usability of Montana PCMH 2014 Quality of Care and Utilization Data Presentation at the Patient-Centered Medical Home Stakeholder Council Meeting Helena, Montana May 20, 2015 Nancy McCall, Sc.D. Juliet Rubini, R.N. Kristin Geonnotti Brianna Sullivan, B.A.

Agenda Introduction Usability of quality of care data and recommendations for future data collection Usability of utilization data and recommendations for future data collection Strategies for incorporating data into a meaningful evaluation and public report This study was conducted with the support from the Robert Wood Johnson Foundation s State Health and Value Strategies Program 2

Goal for Performance Measurement Our assessment of usability focuses upon validity and integrity of data submitted in first reporting period of 2014 Measuring what is intended to be measured Fidelity to measure owner s specifications Free of systematic errors Cannot directly assess the degree to which patients with targeted clinical conditions were systematically and correctly identified for quality of care measures What we can assess varies between patient-level and aggregate data reported Without validity and integrity of data you cannot achieve your aims of measuring PCMH performance 3

Quality of Care 4

2014 Quality of Care Reporting Requirements Four measures Blood pressure control Tobacco use and intervention A1c control Age-appropriate immunization for children Guidance Patient-level (Option 1) data or attested aggregate statistics (Option 2) Specifications Physician Quality Reporting System (PQRS) National Immunization Survey (NIS) 69 PCMHs reported 5

Evaluation of Guidance Evaluated fidelity to PQRS and CHIPRA specifications Identified inconsistencies that might lead to differences in reporting across PCMHs Minor differences between guidance and PQRS specifications for blood pressure control and A1c Greater specificity in PQRS for all three measures Major difference for tobacco cessation and intervention Denominator included only users versus all adults with clinic visits Numerator includes only patients who received intervention versus patients screened for tobacco use and users who received intervention Differences in childhood immunization relative to CHIPRA No numerator or denominator definition in NIS Refusals/contraindications in denominator versus not in denominator Montana guidance excluded seasonal flu and 2-dose Rotavirus 6

PCMH Survey and Analysis of Data Integrity Survey asked PCMHs to Identify the methods used to collect the data they reported Explain reporting criteria used(pqrs versus Montana guidance versus other) 39 entities completed surveys for the universe of 69 clinics Analysis of Data Integrity Reviewed findings from Montana Department of Public Health s review of initial submission Analyzed face validity of data High level findings on missing data or anomalous data patterns National prevalence of diabetes, hypertension, tobacco use and control Comparison between aggregate and patient-level PCMHs Not intended as a thumbs up or down but identify areas for improvement 7

Clinic Reporting Method Source: Mathematica analysis of Montana PCMH Quality of Care Reporting Survey, 2015. 8

Blood Pressure Control Definition Source: Mathematica analysis of Montana PCMH Quality of Care Reporting Survey, 2015. 9

Tobacco Use and Intervention Definition Source: Mathematica analysis of Montana PCMH Quality of Care Reporting Survey, 2015. 10

Patients Included in Tobacco Screening Source: Mathematica analysis of Montana PCMH Quality of Care Reporting Survey, 2015. 11

A1c Control Definition Source: Mathematica analysis of Montana PCMH Quality of Care Reporting Survey, 2015. 12

Immunization Refusal or Contraindication Documentation Source: Mathematica analysis of Montana PCMH Quality of Care Reporting Survey, 2015. 13

Missing Data Epidemiologists did a lot of data cleaning! Missing data Patient identifiers ranging from none to 1/3 of clinics affecting up to 25% of patients Age and sex limited missing data Receipt of recommended service Considerable variation in missing data across measures Immunization was most problematic with a lot of missing data for the 7 individual immunizations versus the composite measure No children were documented in the numerator as a refusal or with a medical contraindication 14

Anomalous Data Patterns Dates of service Out-of-range dates included in submission 70% of clinics had A1c dates not in 2014 or missing affecting 3% of patients Tobacco non-users 6 clinics reported patient-level data only on tobacco users so prevalence of use could not be estimated Aggregate data reporting for immunization required resubmission by 25 clinics Summation of individual immunizations > composite measure Inconsistent identification of control when A1c = 9.0 15

Anomalous Data Patterns Formatting led to the need for a lot of data cleaning Dates in non-date formats Combined systolic and diastolic readings A1c levels reported as numbers and percentages None done, N or 0 inserted when not done Immunizations requested Y, N, MC, R Individual vaccine name, number of vaccines received, 100% Tobacco cessation intervention request Y, N Name of intervention often provided with some dates embedded Data editing rules applied Greater than 95% of patient-level data are used to assess face validity 16

Prevalence of Tobacco Use by Option Source: Mathematica analysis of Montana PCMH Quality of Care Data, 2014. 17

Prevalence of Hypertension by Option Source: Mathematica analysis of Montana PCMH Quality of Care Data, 2014. 18

Prevalence of Diabetes by Option Source: Mathematica analysis of Montana PCMH Quality of Care Data, 2014. 19

Smoking cessation by Option Source: Mathematica analysis of Montana PCMH Quality of Care Data, 2014. 20

Blood Pressure Control by Option Source: Mathematica analysis of Montana PCMH Quality of Care Data, 2014. 21

A1c Control by Option Source: Mathematica analysis of Montana PCMH Quality of Care Data, 2014 22

Recommendations Increase fidelity to PQRS measures and provide greater specificity in guidance Most notable for tobacco cessation denominator definition Develop decision rule for A1c control when A1c = 9.0 Revisit the childhood immunization measure NIS does not provide numerator or denominator definitions Seasonal flu and Rotavirus not included in guidance but in NIS Standardized reporting tool to prompt user when invalid values are added Increases burden on clinics and costs of development but increases data validity Develop consistent data recoding and cleaning edits and rules 23

Recommendations Conduct a limited medical record review To directly assess accuracy of prevalence and measurement Examine how the clinics reporting aggregate data compare with known data problems from the patient-level data Target clinics for technical assistance in reporting If prevalence or measure achievement out of range of other clinics and there is no clinical reason for the discrepancy 24

Acute Care Utilization Measures 25

2014 Utilization Reporting Requirements Two measures Emergency Room (ER) visits Hospitalizations Aggregate statistics from four payers Fully-insured book of business PCMH-attributed population Evaluated guidance provided to payers Held discussions with payers to understand how they were constructing the two measures Revised guidance provided to payers Requested clarification on their approach to constructing the metrics 26

Evaluation of Guidance ER Visits per 1,000 members with medical coverage per year. Report Observation Stays as a separate category and exclude them from ER visits and hospitalizations. Report ER visits that lead to hospitalization in a separate category from ER visits that do not lead to hospitalization. Report multiple ER visits on same day as separate events. 27

Evaluation of Guidance Hospitalizations per 1,000 members with medical coverage per year Include all acute care facilities. Exclude non-acute facilities such as swing bed designations, long-term care hospitals, and rehabilitation hospitals. Include hospitalizations that occur outside of Montana. Roll multiple components of care during a continuous episode into a single admission count, as long as they are all inpatient care. 28

Payer Definitions for Utilization Measures Payers Numerator components definitions Payer A Payer B Payer C Payer D Emergency Room Visits ER visit reporting by disposition All ER visits reported as a single rate - x - - ER visits reported separately by disposition (lead to hospitalization vs. did not lead to hospitalization) x - - x Only reported ER visits that did not lead to a hospitalization - - x - Multiple ER visits on the same day Reported as a single ER visit x - - - Reported as multiple ER visits - x x x Observation Bed Stays Reported separately from ER visits and hospitalizations x - - x Not reported - x x - Source: Mathematica analysis of Montana Payer responses to Questionnaire from the Montana Commissioner of Securities and Insurance, 2015. 29

Payer Definitions for Utilization Measures Numerator components definitions Hospitalizations Payer A Payer B Payers Payer C Included hospitalizations that occurred outside of Montana x x x x Rates reported included the following facilities: Short-term general and specialty hospitals x x x x Critical access hospitals x x x x Psychiatric hospitals and units x x x x Birthing centers x - - x Rehabilitation hospitals and units - x - x Long-term care hospitals - x - x Swing beds - x - - Skilled nursing facilities - - x x Hospitalizations reported as: Separate admissions if change in facility or transfer x x - - A single admission for a continuous inpatient episode - - x x Payer D 30

Recommendations Consider further modifications of measure definitions Increase performance comparability across PCMHs Increase linkage to actionability Categorizing ER visits Reporting multiple versus single events Reporting by disposition Categorizing hospitalizations Types of hospitals Continuous episodes of acute care Categorizing observation bed stays Reporting out separately 31

Recommendations Additional considerations Report final action claims Roll-up interim and final bills into a single hospitalization count Clarify reporting of newborn and delivery hospitalizations Consider removing from hospitalization measure or report only delivery due to lack of actionability by PCMH Ambulatory Care Sensitive Conditions AHRQ Prevention Quality Indicators Low-birthweight or premature neonates Case-mix adjustment Utilization can vary across patients with different sociodemographic and health status characteristics Aggregate statistics do not allow for standardizing across payers or over time 32

Strategies for Incorporating Data into a Meaningful Evaluation and Public Report 33

Evaluation Designs Major challenge in evaluating the effects of practice transformation into medical homes is identifying what would have happened absent transformation Need to isolate the medical home effect from all other changes that are occurring simultaneously during the reporting period ACOs, doc fix, economy, CVS Health, other factors Evaluation designs have varying degrees of strength of evidence Stronger evaluation designs have greater data requirements; benefit from a robust intervention and comparison group Patient level evaluations allow for greater control of differences in case-mix over time 34

Evaluation Designs Descending order of strength of evaluation approach Randomized controlled trial Not currently practical in Montana Difference-in-differences Strongest quasi-experimental design Requires comparison group and data for pre- and post time periods Pre-post differences for PCMH group only Strong assumption that all changes are due to practice transformation Multiple pre-intervention periods can strengthen this approach Quality-of-care measures often aim for meeting a benchmark If comparison of cross-sectional samples, must control for changes in case-mix (i.e., Medicaid expansion) Cross-sectional differences in post-period Does not account for any pre-pcmh transformation trends in outcomes 35

Identifying a Robust Intervention Group For whom do insurers want to provide additional payments? For whom within the practice should the medical home be held accountable for outcomes? Intent-to-treat all patients attributed to practice based upon specific criteria Minimum dosage level of intervention by practice Actively receiving enhanced services chronic conditions Main concern: dilution of effect on outcomes by including patients not receiving enhanced services 36

Attribution: Identifying a Robust Intervention Group Claims-based algorithm Requires decisions about type of services, type of providers, and rules for determining responsible physicians Often based upon a patient s prior number of E&M visits (1-2 years) with primary care providers Requires previous patient utilization Can produce attribution lists that are only 50% of practice s perceived active patients, and may not identify patients in most need Clinic-based roster Practices identify patients for whom they believe they are the primary care provider may identify a non-representative group of patients Multiple practices may claim the same patient Patient-identified clinics Patients select or identify a practice as their primary care provider Most prevalent in an HMO model 37

Identifying a Robust Comparison Group Practice-level interventions with claims-based algorithm Identify a pool of potential comparison group practices with similar characteristics but without recognition as a PCMH Most desirable to stay within same geographic area Need to affirm that there are not systematic differences between practices that are PCMHs and those that are not PCMHs Apply propensity score matching at the practice level Identify relevant practice characteristics and data sources Practice ownership, % of practice meeting meaningful use criteria Apply claims-based algorithm to assign patients Use same algorithm as intervention group assignment 38