Presentation of the proliferation of measurement in health care and how organizations should redesign indicator selection processes to engage and motivate managers to improve performance. Presentation to eHealth students based on 2023 PhD dissertation.
M Heenan_PhD Dissertation Lecture_eHealth Lecture_Engaging Leaders in KPI Selection_Final_July 2023
1. Health care indicator selection processes:
How to select engaging metrics to motivate front-line managers.
Dr. Michael A. Heenan, PhD
McMaster University, eHealth Summer Program
PhD Dissertation Based Presentation
July 26, 2023
2. Mike Heenan, Ph.D.
ā¢ Hospital Executive Vice-President
ā¢ Sessional Lecturer, McMaster University ā DeGroote School
of Business
ā¢ PhD, Business Administration, Health Policy Management,
2023
ā¢ Thesis research: how health care organizations selected key
performance metrics and how the process of indicator
selection impacts front line managers motivation to
improve performance.
2
3. Session Format
3
eHealth ā Why this Lecture?
The Literature and Our Problem ā Metric Madness
Hospital Executive Perception of Front-Line Managerās Knowledge
What Indicators Motivate Front-Line Clinical Managers?
What did the pandemic teach us?
The 5-P Indicator Selection Process Framework
The Politics of Improvement: Your Challenge as eHealth Leaders
5. eHealth ā Why this Lecture?
Your program objectives:
1. Understand eHealth concepts
2. Communicate effectively in eHealth contexts
3. Apply eHealth Knowledge in practical contexts
5
6. Key Readings We will Reference
ā¢ Damschroder LJ, Robinson CH, Francis J, et al. Effects of Performance Measure Implementation on Clinical
Manager and Provider Motivation. J Gen Intern Med. 2014; 29(S4):877-884. DOI: 10.1007/s11606-014-
3020-9.
ā¢ Elliot C, Mcullagh C, Brydon M, Zwi K. Developing key performance indicators for a tertiary childrenās
hospital network. Australian Health Review. 2018; 42(5):491-500. DOI: 10.1071/AH17263.
ā¢ Ginsburg LS. Factors that Influence Line Managersā Perceptions of Hospital Performance Data. Health
Services Research. 2003; 38(1):261-286. DOI: 10.1111/1475-6773.00115.
ā¢ Heenan MA, Randall GE, Evans JM. Selecting Performance Indicators and Targets in Health Care: An
International Scoping Review and Standardized Process Framework. Risk Management and Healthcare
Policy. 2022; 15:747-764. DOI: 10.2147/RMHP.S357561.
ā¢ Mannion R, Braithwaite J. Unintended consequences of performance measurement in health care: 20 salutary
lessons from the English National Health Service. Internal Medicine Journal. 2012; 42(5):569ā574. DOI:
10.1111/j.1445-5994.2012.02766.x
6
7. To Start ā¦
Two Key Definitions
Performance Measurement is defined as the collection, use and public
reporting of data for the purpose of quality improvement, accountability,
and transparency.
America Academy of Family Physicians Performance Measurement Criteria. [Internet] 2022. [Cited 27 Nov 2022].
Available from: https://www.aafp.org/about/policies/all/performance-measures.html
Indicators are defined as measurable elements of practice performance that
relate to clinical, population health, financial, or organizational performance.
Lawrence M, Olesen F. Indicators of Quality in Health Care. The European Journal of General Practice.
1997; 3(3):103-108. DOI: 10.3109/13814789709160336
7
8. To Start ā¦
Indicators are Metrics, Metrics are Indicators
Indicators also known as
ā¢ Key Performance Indicators
ā¢ Performance Indicators
ā¢ Quality Indicators
ā¢ Performance Metrics
ā¢ Performance Measures
ā¢ Quality Metrics
8
9. To Start ā¦
Donabedian Framework
Framework for evaluating health care service quality.
Information on quality of care can be seen through three types of indicators:
Structure: factors that affect the context in which care is delivered
Process: actions the describe how care is delivered
Outcome: effects of healthcare delivery on patients or populations, including
changes to health status, behavior, or knowledge as well as patient satisfaction
and health-related quality of life
9
Donabedian A. The quality of care: how can it be assessed? JAMA. 1988 Sep 23; 260(12):1743-8. DOI: 10.1001/jama.1988.03410120089033.
11. Needed Measurement Led to Too Much Measurement
ā¢ USA, 1999 - Approximately 98,000 Americans lost their lives each year due to medical error.
ā¢ Canada, 2004 - Average of 7.5% of hospital admissions resulting in an adverse event.
ā¢ Multiple governments, regulatory agencies, and funding bodies have mandated the
collection and monitoring of hundreds of KPIs by hospitals
11
Kohn LT, Corrigan JM, Donaldson MS. To Err is Human: Building a Safer Health System. Washington, DC: National Academy Press, 2000.
Baker GR, Norton P, Flintoff V, et al. The Canadian Adverse Events Study: the incidence of adverse events among hospital patients in Canada. Canadian Medical Association Journal.
2004; 170(11):1678-1686. DOI: 10.1503/cmaj.1040498
12. Needed Measurement has Led to Too Much Measurement
12
USA:
ā¢ Centre for Medicare and Medicaid Services (CMS) monitors over 1700 KPIs
ā¢ National Quality Forum (NQF) approved KPIs grew from 200 in 2005 to over 700 in 2011
ā¢ Study concluded over 550 distinct KPIs were being collected across 48 states
Canada
Ontario Hospital Association and Health Quality Ontario argue there are over 300 quality KPIs reported
by hospitals.
International
Over-measurement has been echoed in the United Kingdom and Australia (Mannion and Braithwaite 2011).
Panzer RJ 2013; Wilensky G 2018; Meyer GS 2012; Greenburg A 2019; Mannion R 2012;
13. The Management Problem:
Overabundance of KPIs has resulted in unintended consequences
13
ā¢ Data mandated by state bodies do not necessarily reflect every local context and may be
used inappropriately
ā¢ Public reporting of KPIs without local input has resulted in a lack of trust between
providers and political bodies, and gaming
ā¢ Reporting mandates have also increased financial costs to organizations
ā¢ Building of the information technology and data infrastructure required to support
measurement has amplified the amount of data available to health care organizations
Mannion R and Braithwaite 2011; Safavi 2006; Teare 2014; Meyer et al 201, Berwick 2016
14. Perla: The U Curve
Paralyzed decision making (Cassel et al 2014, Perla 2013).
14
Perla R. Commentary: Health Systems Must Strive for Data Maturity. American Journal of Medical Quality. 2013; 28(3):263-264. DOI:10.1177/1062860612465000.
15. Call for Balance and Parsimony
1. Measure process quality
2. Measure value
3. Design data systems for internal quality needs and spinoff external
quality measures
4. Use return on investment
5. Establish ongoing process to refine and select core measures
15
Meyer GS, Nelson EC, Pryor DB, et al. More quality measures versus measuring what matters: a call for balance and parsimony. BMJ Qual Saf. 2012; 21(11):964-968.
DOI: 10.1136/bmjqs-2012-001081.
16. Question # 1
How does eHealth further hinder the phenomenon of over-
measurement?
16
17. A Call for Balance, butā¦
Not Directly Linked to Goal Theory
Some agencies have heeded these calls.
ā¢ Institute of Medicine (2015) recommends 15 national KPIs that systems should monitor (Blumenthal and
McGinnis 2015).
ā¢ National Quality Forum (2019) instituted annual review process that endorses certain KPIs for adoption to
ensure their relevancy to overall system goals. (NQF 2019)
ā¢ World Health Organization started examining processes in 2008 and offered criteria to narrow KPI selection
(Smith et al 2008).
ā¢ Research papers describe Delphi methods used to select KPIs in clinical areas (Schull et al 2011, Madsen et al
2016, Harvey 2016, Murphy 2016).
These reports describe criteria and methods to select an appropriate number of quality and patient safety
KPIs at the system or clinical service level, butā¦
a) They do not clarify who should be involved in KPI and Target selection, or,
b) how health service provider organizations like hospitals should use selection processes as an
instrument to motivate leaders to achieve agreed upon goals.
17
19. What processes do hospitals use to select performance indicators and do they align with best
practices? A multiple-case study of four hospitals in Ontario, Canada
Qualitative, multiple-case study of four acute care hospitals and how they select indicators
Results
ā¢ Process elements largely absent include:
ā¢ adopting evidence-based selection criteria
ā¢ considering finance and human resources indicators;
ā¢ considering if indicators measure structure, process or outcomes,
ā¢ engaging a broader set of end-users in the selection process.
ā¢ Revealed senior management teams had little confidence in clinical unit managersā
ability to use indicators to improve performance.
19
21. Who are and why do we care about
Front-Line Managers?
Who
ā¢ Clinical unit / Front-Line managers are employees who are both supervised by an
organizationās top managers and who themselves supervise front-line employees.10
Why
ā¢ Managersā commitment and ability to implement innovative change at the unit level
has been linked to strategy realization, efficiency of operations, cost control, and overall
achievement of quality outcomes.10-12
ā¢ If positively motivated and engaged, Managers can effectively support innovation and
change. If managers are disengaged and not motivated, it can impede improvement in
the units they oversee.13-15
ā¢ Managersā motivation and commitment increases when their senior managers support
them with the infrastructure and resources necessary for change.14,16
21
22. Managers and Goal-Setting Theory of Motivation
Goal-Setting Theory of Motivation emphasizes the relationship between goals and
performance.
Task performance is influenced by the managerās ability; the stated goalās content, intensity,
and duration; the managerās self-efficacy; and the personal goals of the manager.
Key Attributes:
1. Benefit of participating in goal setting processes is cognition on whatās important and why;
2. Specific goals with harder to reach improvements result in a greater effort by managers;
3. Regular feedback increases performance effort by improving managerās self-efficacy.
4. Managers that feel supported by their leadership have improved self-efficacy;
5. Managers that feel part of a larger team improve effort through both a shared understanding of the
goalās direction and shared-accountability; and,
6. Managers increased confidence leads to greater motivation to act and perform tasks.
22
Locke EA, Latham GP. Building a practically useful theory of goal setting and task motivation: A 35-year odyssey. American Psychologist. 2002; 57(9): 705ā717. DOI:
10.1037//0003-066x.57.9.705
23. Need to understand base motivations:
Why clinicians become managers
ā¢ Clinical managers never thought of becoming a manager when completing
their initial clinical education.
ā¢ Front-line clinicians move into management to help improve quality and
the working conditions of their fellow staff.
ā¢ When asked what indicators they prefer clinical unit managers referenced
process indicators that help advance clinical practice, and front-line staff
engagement as criteria by which they would choose indicators.
23
Kleinman CS. Leadership Roles, Competencies, and Education: How Prepared Are Our Nurse Managers? Journal of Nursing Administration. 2003; 33(9):451-455.
DOI: 10.1097/00005110-200309000-00005.
White J. Motivation for Seeking a Nurse manager Position. [dissertation]. San Diego, CA: University of San Diego, 2015.
Available from: https://digital.sandiego.edu/dissertations/17
24. Why do managers use performance data?
24
Ivankovic D, Poldrugovac M, Garel P, Klazinga NS, Kringos DS (2020) Why, what and how do European healthcare managers use performance data? Results of a survey and workshop
among members of the European Hospital and Healthcare Federation. PLoS ONE 15 (4): e0231345. https://doi.org/10.1371/journal.pone.0231345
25. What barriers do managers face in using data?
ā¢ Managers gave three barriers: lack of training
and knowledge on measurement, workload,
and job experience.
ā¢ Managers need to obtain is the understanding,
interpretation, and use of performance data for
evidence-based decision making.
ā¢ Clinicians are often appointed to managerial
roles without business skills, and they therefore
need to learn them through trial and error.
25
āMy education is based on how to be a
nurse, not how to decipher data and know
what it means. So, some of this stuff is
foreign to me and itās been trial by fire for
me.ā
Case B-Manager 3
Birken SA, Lee SYD, Weiner BJ. 2012; DOI: 10.1186/1748-5908-7-28.
Naranjo-Gil D. 2009; DOI: 10.1097/HMR.0b013e31819e8fd0.
Kleinman CS. 2003; DOI: 10.1097/00005110-200309000-00005.
McCallin AM, Frankson C. 2010; 1DOI: 10.1111/j.1365-2834.2010.01067.x
Heenan, MA. 2023. (Doctoral dissertation).
āI have not received any training or
education on measurement here. When I
started in the role, it was kind of just sink or
swim.ā
Case A-Manager 3
26. How clinical unit managersā roles in selecting hospital indicators impacts their motivation and
self-efficacy to improve performance: A qualitative multiple-case study.
Qualitative multiple-case study of four hospitals involving 22 managers. Semi-structured
Interviews. Thematic Analysis.
Results:
ā¢ Managers reported being unaware of the criteria hospitals used to select indicators, are
not trained in measurement, feel they have little control over the indicators they were
accountable for, and were fearful of reporting on performance.
ā¢ Managers motivated by process indicators focused on clinical quality and patient
experience vs. outcome or business indicators.
ā¢ Managers gained confidence over time and by learning how to communicate the utility
of indicators to front-line staff.
26
Heenan, M. A. (2023). Health Care Indicator Selection Processes and their Impact on Clinical Unit Managersā Motivation and Self-Efficacy to Improve
Performance (Doctoral dissertation).
27. Ginsburg: Factors that Influence Line Managersā
Perceptions of Hospital Performance Data
27
Ginsburg LS. Factors that Influence Line Managersā Perceptions of Hospital Performance Data. Health Services Research. 2003; 38(1):261-286. DOI: 10.1111/1475-6773.00115
āSimply generating performance data will not, in and of itself, lead to improvement.ā
Factors Significant Correlation to
Perceived Usefulness of Data
ā¢ Data quality
ā¢ Data relevance
ā¢ Report complexity
ā¢ Intensity of dissemination
28. Damschroder: Effects of Performance Measure
Implementation on Manager Motivation
ā¢ Indicators need to be
designed to both incent
individuals and build
coalitions.
ā¢ Indicators sit within a QI
foundation but only effective
if seen positively by fully
engaged staff and
management.
28
Damschroder LJ, Robinson CH, Francis J, et al. Effects of Performance Measure Implementation on Clinical Manager and Provider Motivation. J Gen Intern Med. 2014;
29(S4):877-884. DOI: 10.1007/s11606-014-3020-9.
29. Question # 2
Now knowing the motivations and challenges end-users have
in using data, how can eHealth help the phenomena of over-
measurement?
29
31. What did the pandemic teach us?
Executive Perspectives
ā¢ Senior management leaders
mentioned that some clinical
practice, supply chain, and human
resource indicators, once thought
not important to regular
monitoring, had surfaced as
priorities.
31
Heenan MA. Health Care Indicator Selection Processes and their Impact on Clinical Unit Managersā Motivation and Self-Efficacy to Improve Performance (Doctoral dissertation).
āCOVID taught us that after years of trying to get the
organization to make data-driven decisions that often fell
on deaf ears, the organization was thirsty for data about
this new disease. This demand for data was likely a result
of operating in a new unknown environment and the need
to base operational decisions on emerging data compared
to traditional lagging indicators.
COVID also taught us that we likely had been
providing the wrong data before the pandemic. COVID
forced us to prioritize our efforts and stopped us from
chasing too many indicators. It led us to provide more
leading process-based indicators that clinical units found
useful in planning their day versus lagging outcome data
that government wanted.ā
Case B-Executive 2
32. What did the pandemic teach us?
Front-Line Manager Perspectives
Clinical unit managers uniformly shared that the
pandemic:
ā¢ created a single priority for their organizations;
ā¢ involved greater engagement in, and lead to,
faster decision making;
ā¢ Focus on process indicators related to practice
and staff safety.
ā¢ Example, new indicator monitored more
closely ā personal protective equipment (PPE)
Inventory Levels
32
āCOVID has changed everything. I've seen
senior leader sentiment sort of evolve from
when people are being told to do things versus
asking us why we're doing them. Senior leaders
have become clinically focused. We see them
maybe a little bit better for who they are now
because the data is not being shoved down our
throats. We are being asked if certain data
makes sense to us. My VP is just, yep, whatever
you need, so I get that support, which makes a
huge difference.ā
Case A-Manager 4.
Heenan MA. Health Care Indicator Selection Processes and their Impact on Clinical Unit Managersā Motivation and Self-Efficacy to Improve Performance (Doctoral dissertation).
33. Lessons
1. Crises create opportunities to see new needs and new ways
2. Decision making structures get flatter, relationships gets stronger,
because you engage the end-user deliberately and quickly
3. Process metrics matter and help inform priority actions
4. Managers want to be engaged to better understand their accountabilities
and help select process indicators that measure quality, clinical practice,
patient experience and staff engagement
33
34. In designing or selecting indicators, what framework
should guide us?
The 5-P Indicator Selection Process Framework
34
35. Selecting performance indicators and targets in health care:
An international scoping review and standardized process framework
Purpose: Synthesize international approaches to indicator selection and develop a standardized process
framework.
Methods:
ā¢ PRISMA-ScR Scoping review using Pub Med & Web of Science.
ā¢ English-language papers from 11 countries, 2010-2020.
ā¢ Reviewed 33 peer-reviewed papers and 11 grey-literature documents.
ā¢ Looked at processes with a lens to: Aim; Governance; Preparation; Process Methods; Validation
Results:
ā¢ Some processes used Appraisal of Indicators through Research and Evaluation (AIRE) Instrumentand the
Quality Indicator Critical Appraisal (QICA) tool to guide which individual indicators should be considered.
ā¢ No standardized indicator selection process framework exists that focuses on engagement that solicits buy
in and drives motivation.
35
Heenan MA, Randall GE, Evans JM. Selecting Performance Indicators and Targets in Health Care: An International Scoping Review and Standardized Process Framework.
Risk Management and Healthcare Policy. 2022; 15:747-764.
36. The 5-P Indicator Selection Process Framework
Incomplete indicator selection processes
risk:
o over-measurement,
o lack of prioritizing strategic and
operational goals,
o lack of support by end-users
o paralyzed decision-making ability.
ā¢ Standardized framework developed to
assist organizations complete a
comprehensive indicator and target
selection process.
36
Domain Element
Clarify Aim
Develop Guiding Principles
Identify Level of Use
Build Governance Structures
Recruit Participants
Seek End-User Input
Research Evidence-Based Literature
Build an Inventory of Potential Indicators
Categorize Potential Indicators into Strategic
Orient and Train Participants
Utilize a Consensus Building Method
Identify a Facilitator
Indicator Selection Criteria
Analytical Assess Indicators
Set Indicator Targets
Assess Data Quality
Validate with End-Users
5-P Indicator Selection Process Framework
Prepare
Procedure
Prove
Purpose
Polity
Heenan, MA, Randall GE, Evans JA. Selecting performance indicators and targets in health care: An international scoping review and standardized process framework. Risk Management and
Healthcare Policy. 2022; 15:747-764. DOI: 10.2147/RMHP.S357561.
37. Domain: Purpose
37
āPurposeā sets out the reasons why an indicator selection and target setting process is undertaken
ā facilitating a shared understanding of what the organization is setting out to achieve.
Elements
ā¢ Clarify Aim
ā¢ Set Guiding Principles
ā¢ Identify Level of Use
38. Purpose: Three Elements
38
Watkins: Setting clear aims and guiding principles allows participants to understand the desired goals a process is
aiming to achieve and the actions to which they are expected to contribute.
Watkins MD. Demystifying Strategy: The What, Who How and Why. Harvard Business Review. September 10, 2007. Available from: https://hbr.org/2007/09/demystifying-strategy-the-what
39. Elliot, 2018
39
Elliot C, Mcullagh C, Brydon M, Zwi K. Developing key performance indicators for a tertiary childrenās hospital network. Australian Health Review. 2018; 42(5):491-500. DOI: 10.1071/AH17263.
Aim:
Following merger ad strategic plan, to address executive level
performance indicators for pediatric tertiary care services across two
campus
Guiding Principles:
KPIs must address the mission statement, āworking in partnership to
improve heath and well being of children through clinical care, research
education and advocacyā.
Drive quality improvement in child health outcomes
Level of Use:
Executive Level / Board
40. Domain: Polity
Elements
ā¢ Build Governance Structures
ā¢ Recruit Participants
40
āPolityā identifies the governance structures that manage the selection process, how the process
will be resourced, and who participates.
41. Polity: Two Elements
41
Aktaa: Expert panels have inherent bias. Therefore, expansion of participants is important mitigation. Indicator
selection processes that have broader participation are more likely than others to have a more inclusive view of
front-line operations
Aktaa S, Batra G, Wallentin L, et al. European Society of Cardiology methodology for the development of quality indicators for the quantification of cardiovascular care and outcomes. European
Heart Journal Quality of Care and Clinical Outcomes. 2022; 8(1):4-13. DOI: 10.1093/ehjqcco/qcaa069.
42. Example: Elliot, 2018
42
Elliot C, Mcullagh C, Brydon M, Zwi K. Developing key performance indicators for a tertiary childrenās hospital network. Australian Health Review. 2018; 42(5):491-500. DOI: 10.1071/AH17263.
Governance Structure
Single body - Executive Group
Participants / Inputs:
Network Executives;
Clinical Directors;
Medical, Allied Health and Nurse Managers.
43. Domain: Prepare
Elements
ā¢ Seek End-User Input
ā¢ Research Evidence Base Literature
ā¢ Build an Inventory of Indicators
ā¢ Categorize Indicators into Strategic Themes
ā¢ Orient and Train Participants
43
āPrepareā addresses how to plan for selection.
44. Prepare: Five Elements
44
Heenan
ā¢ Final list of indicators built from comprehensive sources improves their relevancy to end-users while enabling future
comparability and benchmarking.
ā¢ End-user input upfront on indicator knowledge and user requirements and issued orientation materials increased
participant engagement and improved understanding of the process among participants.
49. Domain: Procedure
Elements
ā¢ Utilize a Consensus Building Methodology
ā¢ Identify a Facilitator
ā¢ Set Indicator Selection Criteria
ā¢ Analytically Assess Indicators
ā¢ Set Indicator Targets
49
āProcedureā describes the how to steps used to assess indicators and targets and gain consensus.
50. Procedure: Five Elements
50
Stetler: Internal facilitation may bias processes whereas external facilitators have an ability to play both a problem
solving and supportive role that might otherwise be difficult when only internal parties are engaged. External
facilitation is valued when a process is focused on an interactive problem or distinct activity.
Stetler CB, Legro MW, Rycroft-Malone J, Bowman C, Curran G, Guihan M, Hagedorn H, Pineros S, Wallace CM. Role of external facilitation in implementation of research findings: a qualitative
evaluation of facilitation experiences in the Veterans Health Administration. Implementation Science. 2006; 1(1):1-5. DOI: 10.1186/1748-5908-1-23.
52. Domain: Prove
Elements
ā¢ Assess Data Quality (Quantitatively)
ā¢ Validate with End-Users (Qualitatively)
52
āProveā describes the validation processes used to test any final set of indicators for data quality
and relevance with end-users.
53. Prove: Two Elements
53
Heenan
ā¢ Processes that statistically test indicators increase scientific soundness and better informs target setting.
ā¢ Processes that validate a final list of indicators with end-users improve relevance and usability by users, especially in
cases where the expert panels did not include front-line directors, managers, or patients.
Heenan, MA, Randall GE, Evans JA. Selecting performance indicators and targets in health care: An international scoping review and standardized process framework. Risk Management and
Healthcare Policy. 2022; 15:747-764. DOI: 10.2147/RMHP.S357561.
54. Examples: Elliot, 2018 and Heenan, 2023
54
QUALITATIVE: Noted Limitation (Elliot)
āThere was little or no consultation with less senior staff members, which risks losing the
support of staff who will be relied upon to provide data and ensure accurate reportingā
QUANTITATIVE: 2022 Case Study of 4 Ontario Hospitals (Heenan)
ā¢ None quantitatively validated KPIs - reasoned most of the indicators they monitor are
mandated by external funding agencies, and it is therefore assumed these indicators
were previously validated.
ā¢ This is a false assumption.
ā¢ While the technical formula of an indicator may have been validated by an external
agency, the data that is generated from the different local hospital information systems
may be of different quality than the data by which an indicator was tested by the
external agency. (e-heath)
55. Key Conclusion
Recall
ā¢ Overabundance of indicators has paralyzed decision making, and eroded trust
between those who ask for indicators and those who are expected to use them to
make change.
ā¢ Many policy institutes and academics have called for a more appropriate, lower
number of indicators.
Therefore
ā¢ Indicator selection or reduction processes cannot occur by happenstance.
ā¢ The 5-P Indicator Selection Process Framework to ensure a systematic, evidence-
based, and inclusive approach that engages measurement experts and those who
use indicators to monitor and improve performance in both selection and
validation. 55
56. Question # 3
What domain and element of the 5-P Indicator Selection
Process Framework can eHealth most help with? Why?
56
57. The politics of improvement:
Your challenge as eHealth Leaders
57
Process improvement is not only about adopting evidence but includes the process of
engaging those responsible for implementing the prescribed change
58. Beyond the evidence (or technology):
Youāre not just an eHealth Leader but a Change Leader
58
ā¢ Quality Improvement initiatives are not simple
ā¢ Evidence is insufficient in itself to ensure implementation
Distributed Nature of Costs
and Benefits
Variety of Value Systems
Underlying Interventions
Whatās in it ?
For Whom ?
What ideologies and
ethics are involved ?
Langley A, Denis J. Beyond evidence: the micropolitics of improvement. BMJ Quality and Safety. 2014; 20:i43-i46. DOI: 10.1136/bmjqs.2010.046482.
59. Hard Core and Soft Periphery
Hard core = Evidence that is
irreducible and carries key benefit.
(The Technical What)
Soft Periphery = Complex delivery
systems that take different forms and
if poorly organized may destroy the
benefit of the hard core.
(The Cultural Why & People)
59
60. Navigate:
The Micro-Political / Change Leader Check List
1. Understand the evidence
2. Identify who it will impact
3. Understand how it will change practice
4. Donāt make change alone: find a champion(s)
5. Solicit input on how to adopt locally
6. Sell the why (involve end-users)
7. Teach: Be a coach not a consultant
60
62. Your Program Objectives:
You hold knowledge and power, Use it Wisely
1. Understand eHealth Concepts
ā¢ The technology and systems you build and promote will be used to
make change by producing information and data.
ā¢ This information and data (indicators) can overwhelm the end-
users not aware of eHealth concepts.
2. Communicate Effectively in eHealth Contexts
ā¢ Ask why are we producing the information?
3. Apply eHealth Knowledge in practical contexts
ā¢ What will it be used for? And by Whom?
62
63. Final Conclusions
Answering ā¦ āHow to select engaging performance metrics?ā
1. Have a Structured, Open, Transparent Process
2. Include End-Users as Participants
3. Educate and Train participants on concepts
4. Allow Input on Past Experience before Selection
5. People Process Things to get to Outcomes so Process Metrics Matter
6. Validate Data in Real Environments
63
64. Abd then donāt forget ā¦The U Curve
64
Perla R. Commentary: Health Systems Must Strive for Data Maturity. American Journal of Medical Quality. 2013; 28(3):263-264. DOI:10.1177/1062860612465000.
66. Dissertation Reference
66
Heenan, M. A. (2023). Health Care Indicator Selection Processes and their Impact on Clinical Unit Managersā
Motivation and Self-Efficacy to Improve Performance (Doctoral dissertation).
https://macsphere.mcmaster.ca/bitstream/11375/28348/2/Heenan_Michael_A_20
2302_PhD.pdf