Michigan Statewide HMIS: Performance Improvement Using ...
Upcoming SlideShare
Loading in...5
×
 

Michigan Statewide HMIS: Performance Improvement Using ...

on

  • 555 views

 

Statistics

Views

Total Views
555
Views on SlideShare
555
Embed Views
0

Actions

Likes
0
Downloads
0
Comments
0

0 Embeds 0

No embeds

Accessibility

Upload Details

Uploaded via as Microsoft PowerPoint

Usage Rights

© All Rights Reserved

Report content

Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

Cancel
  • Full Name Full Name Comment goes here.
    Are you sure you want to
    Your message goes here
    Processing…
Post Comment
Edit your comment

Michigan Statewide HMIS: Performance Improvement Using ... Michigan Statewide HMIS: Performance Improvement Using ... Presentation Transcript

  • Michigan Statewide HMIS Performance Improvement using Outcome & Indicator Measures
  • About Michigan
    • Michigan Statewide HMIS
    • 2.5 years old
    • 57 CoCs including Detroit
    • 370+ agencies with 1,322 programs
    • 3 State Departments (DHS,DCH,DOE)
    • Over 120,000 unique clients
  • Michigan Vertical Configuration
  • Measurement: An Ongoing Collaboration
    • Funding Organizations
    • Program Leadership
    • Staff who collect information and enter data
  • Selecting What to Measure
    • Mix of measures
      • Process Indicators
        • Time to referral or placement in housing
        • Percentage returning for ongoing services
      • Short-term or Intermediate objectives - measure critical processes
        • Positive Housing placement at Discharge
        • Improved income at discharge
        • Linked with needed supportive services
        • Percent re-admissions
      • Outcomes – measures sustained change
        • Stable Housing
        • Stable Employment
  • Coordinating the Process of Measurement
    • How do we know what success looks like?
      • Client Characteristics / Who is Served
      • When measurement occurs
      • Establishing Base Lines
      • Benchmarking / Comparing Performance
      • Improving Performance
      • The Environment / Contextual Variables
      • Evaluate Outcomes
  • Defining the Population Who is Served
    • Developing shared screening questions. Data Standards plus:
      • Have they had a lease in their own name?
      • Were they homeless as children?
      • What is their housing history?
      • What is their education history?
      • What is their employment history?
      • What is their health history?
      • Are there disabilities of long duration?
      • Using a Self Sufficiency Score
  • When Measurement Occurs
    • Intermediate Outcomes Processes/Indicators
      • Measured during care or at discharge.
      • Reflect stable and optimized program processes.
    • Examples
      • Length of time to critical referrals
      • % of staff turnover
      • # of re-admissions
      • Session compliance
      • Medication Errors
      • Critical Incidents
      • Customer Satisfaction
    • Outcomes
      • Measured after discharge for some period of time.
      • Were changes achieved in care sustained?
      • Did we impact other long-term primary outcomes?
    • Examples
      • Positive Housing Placement
      • Improved Income
      • Employment/School
  • Base Lines & Targets
    • Critical for evaluating change and optimizing performance
      • Realistic / honest targets
      • Stabilized Processes.
        • Percentiles are consistent across time
      • Systematic and Routine Measurement
        • Before,
        • During, and
        • After program change.
  • Benchmarking
    • Working with “like” programs to:
      • Identify realistic targets
      • Identify performance that is off norm
      • Share ideas to improve
      • Address measurement problems
      • Build transparency
  • Example Control Chart: School Attendance
  • Defining the Measure
    • Purpose: To measure the Programs ability to support the educational plan of children in care.
    • A client is considered to have met their individual education plan for the day if they participated for 75% or more of the time they were scheduled to be involved in education activities.
    • Available education days reflect only days when it was possible for the client to participate in their individual education plan. For example, if the resident’s educational plan is to attend school, only those days that the school was open during the reporting period are counted. Excused absences do not count as possible days. If a client was excused due to illness, pass or an appointment, these days are not considered available education days.
  • Creating Management Information
    • Cutting the performance rate to guide data driven management decisions.
      • Example 1: Employment: Overall is avg 50%, but for women the avg is 65%, for men it is 30%. Do you revise services to men or serve only women?
      • Example 2: Retention in housing: Overall is avg 72%, but for those w. substance abuse problems that number is only 15%. You conduct a study to determine barrier for those clients.
  • The Context of Services
    • Lack of affordable housing
    • Vacancy rates
    • Unemployment rates
    • Lack of access to supporting services
      • Long Waiting Lists
      • No Insurance or Pay Source
    • Barriers to success or community assets
  • Supporting Performance Improvement?
    • Support the development of initial measures (indicators, objectives and outcomes) and targets through meetings with funders, program leadership and staff.
    • Provide a menu of measures for program leadership to select from.
  • Project Activities Continued
    • Provide reports that may be run monthly with a click of the button on all selected measures. The reports will aggregate data that is viewable through the HMIS for Programs, Agencies, CoCs, and Statewide.
    • Convene routine Benchmarking meetings to discuss measurement issues as well as performance ideas.
  • Project Activities Continued
    • Maintain, with local support , a database of contextual variables.
    • Share/problem-solve measurement issues. Support the evolution of measures.
    • Provide data for statewide analysis as defined in contracts.
  • Contact Information
    • Barbara Ritter
    • 517-230-7146
    • [email_address]
    • MCAH WEB SITE:
    • www.mihomeless.org