Published on in Vol 27 (2025)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/68757, first published .
How to Refine and Prioritize Key Performance Indicators for Digital Health Interventions: Tutorial on Using Consensus Methodology to Enable Meaningful Evaluation of Novel Digital Health Interventions

How to Refine and Prioritize Key Performance Indicators for Digital Health Interventions: Tutorial on Using Consensus Methodology to Enable Meaningful Evaluation of Novel Digital Health Interventions

How to Refine and Prioritize Key Performance Indicators for Digital Health Interventions: Tutorial on Using Consensus Methodology to Enable Meaningful Evaluation of Novel Digital Health Interventions

Tutorial

1School of Nursing and Midwifery, Trinity College Dublin, Dublin, Ireland

2Department of Clinical Informatics, Beth Israel Deaconess Medical Centre, Boston, MA, United States

3Department of Medicine, Harvard Medical School, Boston, MA, United States

4Institute for Health and Society, Faculty of Medicine, University of Oslo, Oslo, Norway

5Department of Clinical Neuroscience, Karolinska Institutet, Stockholm, Sweden

6School of Nursing, Midwifery and Health Systems, University College Dublin, Dublin, Ireland

Corresponding Author:

Catherine McCabe, Prof Dr

School of Nursing and Midwifery

Trinity College Dublin

24 D`Olier At

Dublin 2

Dublin, D02 T283

Ireland

Phone: 353 1 8933019

Email: camccabe@tcd.ie


Digital health interventions (DHIs) have the potential to improve health care and health promotion. However, there is a lack of guidance in the literature for the development, refinement, and prioritization of key performance indicators (KPIs) for the evaluation of DHIs. This paper presents a 4-stage process used in the Gravitate Health project based on stakeholder consultation and consensus for this purpose. The Gravitate Health consortium, which comprises private and public partners from across Europe and the United States, is developing innovative digital health solutions in the form of Federated Open-Source Platform and G-lens to present users with individualized digital information about their medicines. The first stage of this was the consultative process for the development of KPIs involving stakeholder (Gravitate Health project leads) consultations at the planning stages of the project. This resulted in the formation of an extensive list of KPIs organized into 7 categories. The second stage was conducting a scoping review, which confirmed the need for extensive stakeholder consultation in all stages of the KPI development, refinement, and prioritization process. The third stage was a period of further consultation with all consortium members, which resulted in the elimination of 1 category of KPIs. The fourth stage involved using the Delphi technique for refining and prioritizing the remaining 6 categories of KPIs. It is unusual to use this methodology in a nonresearch exercise, but it provided a clear consultative framework and structure that facilitated the achievement of consensus within a large consortium of 250 members on a substantial list of KPIs for the project. Consortium members ranked the relevance and importance of each KPI. The final list of KPIs provides substantial indicators sensitive to the needs of a broad group of stakeholders that are being used to capture real-world data in developing and evaluating DHIs.

J Med Internet Res 2025;27:e68757

doi:10.2196/68757

Keywords



Recent advancements in digital technology have enhanced many aspects of health care, allowing for digital monitoring, supporting health-related behaviors, and enabling communication [Zeeb H, Pigeot I, Schüz B. Digital public health-rapid technological progress, but many open public health questions. Bundesgesundheitsblatt Gesundheitsforschung Gesundheitsschutz. 2020;63(2):135-136. [FREE Full text] [CrossRef] [Medline]1]. One way this has been achieved is through the use of digital health interventions (DHIs), which have the potential to improve health care through illness prevention and health promotion [Wienert J, Jahnel T, Maaß L. What are digital public health interventions? First steps toward a definition and an intervention classification framework. J Med Internet Res. 2022;24(6):e31921. [FREE Full text] [CrossRef] [Medline]2,Schoeppe S, Alley S, Van Lippevelde W, Bray NA, Williams SL, Duncan MJ, et al. Efficacy of interventions that use apps to improve diet, physical activity and sedentary behaviour: a systematic review. Int J Behav Nutr Phys Act. 2016;13(1):127. [FREE Full text] [CrossRef] [Medline]3]. DHIs are defined as a “discrete functionality of digital technology designed to improve health system processes in order to achieve health sector objectives” [Digital health platform handbook: building a digital information infrastructure (infostructure) for health. World Health Organization. 2020. URL: https://iris.who.int/handle/10665/337449 [accessed 2025-03-06] 4]. For example, DHIs may help facilitate tailored health solutions such as providing targeted support and education and enhancing treatment adherence [Sieben A, Bredie SJHB, van Laarhoven CJHMK, Schoonhoven L, Burger DM, van Onzenoort HAW. e-Health interventions and improvement in treatment adherence. Ned Tijdschr Geneeskd. 2014;158:A8460. [Medline]5] while reducing the burden on health care providers where time and resources are already limited [Pouls BPH, Vriezekolk JE, Bekker CL, Linn AJ, van Onzenoort HAW, Vervloet M, et al. Effect of interactive eHealth interventions on improving medication adherence in adults with long-term medication: systematic review. J Med Internet Res. 2021;23(1):e18901. [FREE Full text] [CrossRef] [Medline]6]. However, evidence regarding the efficacy of DHIs to improve health outcomes has been mixed [Wang Z, Zhu Y, Cui L, Qu B. Electronic health interventions to improve adherence to antiretroviral therapy in people living with HIV: systematic review and meta-analysis. JMIR Mhealth Uhealth. 2019;7(10):e14404. [CrossRef] [Medline]7,Jeminiwa R, Hohmann L, Qian J, Garza K, Hansen R, Fox BI. Impact of eHealth on medication adherence among patients with asthma: a systematic review and meta-analysis. Respir Med. 2019;149:59-68. [FREE Full text] [CrossRef] [Medline]8].

The Gravitate Health consortium comprises private and public partners from across Europe and the United States, drawing on expertise from within academia and research institutes, knowledge transfer experts, European Federation of Pharmaceutical Industries and Associations and the Innovative Medicines Initiative 2 partners, health care providers and payers, patient organizations and consumer groups, regulators and product information providers, and the digital technology industries. Gravitate Health is developing innovative solutions in the form of Federated Open-Source Platform and G-lens, which is an application, to present users with individualized digital information about their medicines, which is adapted to their personal needs and context. These solutions will access the electronic product information for each medicine that users are taking. It will use additional health data such as a personalized patient summary to focus and highlight the most personally relevant content from the electronic product information. The intended use of the Gravitate Health solutions is to enable users to understand their medicines in an individualized way, for example, understand why they are taking each medicine, what effects to anticipate, what precautions to take while consuming the medicine, and how best to follow recommendations and adhere to their ongoing treatment. It is anticipated that by being informed, users will better adhere to their medication schedule, which will lead to better health outcomes, fewer complications, and reduced health care use.

The Gravitate Health project work plan will see these solutions deployed and scaled across 4 demonstration sites using key performance indicators (KPIs) as a baseline for evaluation. KPIs are measures or variables that provide a means of observing and optimizing achievement and reflect changes connected to performance [Handbook on health inequality monitoring: with a special focus on low-and middle-income countries. World Health Organization. 2013. URL: https:/​/www.​who.int/​docs/​default-source/​gho-documents/​health-equity/​handbook-on-health-inequality-monitoring/​handbook-on-health-inequality-monitoring.​pdf [accessed 2025-03-11] 9,Khalil H, Kynoch K. Implementation of sustainable complex interventions in health care services: the triple C model. BMC Health Serv Res. 2021;21(1):143. [CrossRef] [Medline]10]. They deliver relevant information to stakeholders regarding the progress and sustainability of complex interventions [Varela T, Zamorano P, Muñoz P, Rain C, Irazoqui E, Sapag JC, et al. Evaluation of the implementation progress through key performance indicators in a new multimorbidity patient-centered care model in Chile. BMC Health Serv Res. 2023;23(1):439. [FREE Full text] [CrossRef] [Medline]11]. One issue with selecting KPIs for the evaluation of innovative DHIs is that there is a lack of evidence for relevant performance indicators. This is because DHIs are by their very nature new and therefore have not previously been assessed. A recent scoping review by Brenner et al [Brenner M, Weir A, McCann M, Doyle C, Hughes M, Moen A, et al. Development of the key performance indicators for digital health interventions: a scoping review. Digit Health. 2023;9:20552076231152160. [FREE Full text] [CrossRef] [Medline]12] concluded that there is a lack of evidence for a standardized approach to the development and prioritization of KPIs for DHIs. This paper highlighted that although most studies recognized and included the input of key stakeholders as a starting point for identifying potential KPIs, a common, rigorous process for the development and prioritization of KPIs for innovative DHIs is needed [Brenner M, Weir A, McCann M, Doyle C, Hughes M, Moen A, et al. Development of the key performance indicators for digital health interventions: a scoping review. Digit Health. 2023;9:20552076231152160. [FREE Full text] [CrossRef] [Medline]12]. This paper addresses this lack of guidance within the literature for developing relevant KPIs for meaningful evaluation by presenting a novel use of the well-established Delphi technique as a process for refining and prioritizing KPIs for novel DHIs.


Process for Refinement and Prioritization of KPIs

Three potential approaches were considered as potentially suitable processes for refining and prioritizing the KPIs for evaluation in this project. The main criterion for consideration was the need for a process that would include all key stakeholders in coming to a consensus in refining and prioritizing KPIs. These were (1) the Sheffield elicitation framework [Oakley J, O'Hagan A. SHELF: the Sheffield ellicitation framework (version 2.0). 2014. URL: https://shelf.sites.sheffield.ac.uk/ [accessed 2025-01-29] 13], (2) a factorial survey approach [Rossi PA, Anderson A. The factorial survey approach: an introduction. In: Rossi P, Nock S, editors. Measuring Social Judgments: The Factorial Survey Approach. Beverly Hills, CA. Sage; 1982. 14], or (3) the Delphi technique [Dalkey N, Helmer O. An experimental application of the DELPHI method to the use of experts. Manag Sci. 1963;9(3):458-467. [CrossRef]15]. The Sheffield elicitation framework and factorial survey approach rely on having a clearly defined group and a limited, strong evidence-based list of KPIs; therefore, neither were regarded as appropriate due to the uniqueness of the Gravitate Health solutions and the subsequent absence of supporting evidence. The list of KPIs for this project was produced at the outset of the project by key stakeholders based on their expertise or role in the project and reflected their aspirations for the digital solutions developed by the Gravitate Health project. Due to the novel nature of the solutions, the list of KPIs identified was lengthy and was not evidence-based. This meant that further refinement and prioritization of the KPIs was required at a more advanced stage of the project and needed to include all consortium members.

The Delphi technique was regarded as the most appropriate method to achieve consensus across all stakeholders in terms of refining and prioritizing the KPIs for evaluation of the Gravitate Health digital solutions. This technique, developed by Dalkey and Helmer [Dalkey N, Helmer O. An experimental application of the DELPHI method to the use of experts. Manag Sci. 1963;9(3):458-467. [CrossRef]15], is a widely accepted process commonly used in research studies for converging opinions on a specific topic from experts within the discipline [Allen J, Brenner M, Hauer J, Molloy E, McDonald D. Severe neurological impairment: a Delphi consensus-based definition. Eur J Paediatr Neurol. 2020;29:81-86. [FREE Full text] [CrossRef] [Medline]16]. The Delphi technique is a systematic and interactive approach in research that involves a panel of experts coming together to share information and experience in order to achieve consensus as an outcome [Nasa P, Jain R, Juneja D. Delphi methodology in healthcare research: how to decide its appropriateness. World J Methodol. 2021;11(4):116-129. [CrossRef] [Medline]17,Steel M, Seaton P, Christie D, Dallas J, Absalom I. Nurse perspectives of nurse-sensitive indicators for positive patient outcomes: a Delphi study. Collegian. 2021;28(2):145-156. [CrossRef]18]. Consensus was an essential feature of refining and prioritizing the existing KPI list in the Gravitate Health project; therefore, using the Delphi technique as a consultative process rather than a research methodology provided a structured and rigorous framework to achieve this.

Development of Initial KPIs and Evaluation

Based on the initial anticipated impact of the Gravitate Health digital solutions, a preliminary list of KPIs was created and classified according to anticipated areas of impact during a face-to-face informal brainstorming session by key stakeholders when the initial grant proposal was being prepared. This key stakeholder group comprised early consortium members representing medicine, patient representative groups, academia, technology, and pharmaceutical industry. The areas of expected impact of the KPIs included access and use, understanding, user experience, patient compliance or adherence, 2-way communication, and risk minimization. This classification of the KPIs and allocation of percentage thresholds for the achievement of these KPIs were based on expert opinion only and conducted using an informal consultative process at a very early stage in the proposal development. There was no specific evidence to support any other particular process for developing relevant KPIs in DHIs [Brenner M, Weir A, McCann M, Doyle C, Hughes M, Moen A, et al. Development of the key performance indicators for digital health interventions: a scoping review. Digit Health. 2023;9:20552076231152160. [FREE Full text] [CrossRef] [Medline]12].

Measurement of KPIs throughout the project is an iterative process (3 intervals) using relevant tools at key stages throughout the project (minimal viable products 1, 2, and 3) to measure digital health literacy (of educational material), patient empowerment, user trust of the platform, usability, accessibility, and satisfaction with the platform. For example, systems data will be used to measure the use of the platform and to gauge user engagement. The Systems Usability Scale will be used to measure usability and the Medication Adherence Report Scale-10 to measure medication adherence [Horne R, Weinman J. Self-regulation and self-management in asthma: exploring the role of illness perceptions and treatment beliefs in explaining non-adherence to preventer medication. Psychol Health. 2002;17(1):17-32. [CrossRef]19,Brooke J. SUS: a 'quick and dirty' usability scale. In: Usability Evaluation in Industry. London. CRC Press; 1996. 20]. Qualitative data will be collected via focus group interviews to obtain subjective accounts of user experiences related to trust, accessibility, usability, and user experience with digital solutions. The findings from each interval are used by the technical teams for product refinement and development throughout the project.

Multimedia Appendix 1

Process of mapping key performance indicators to evaluation tools and time points.

DOCX File , 26 KBMultimedia Appendix 1 describes the process of mapping KPIs to evaluation tools and evaluation time points.

Stakeholder Consultation

Stakeholder consultation included a presentation of the initial KPIs at several internal meetings (monthly) within the Gravitate Health consortium, which had expanded in numbers significantly (n=250) since the initial list of KPIs was created. Each consortium member consulted during these meetings represented a key expert group involved in the ongoing design and evaluation of the Gravitate Health digital health solutions, for example, health care users, health care practitioners, academic institutes, and pharmaceutical and technology industries, and all were from high-income European Union countries. These meetings aimed to seek expert advice and perceptions on the initial KPIs and explain and discuss the use of the Delphi technique to refine and prioritize the KPIs. Overall, this consultative process demonstrated support for the preliminary list of KPIs with the exception of the category on “value for the industry and society.” These are specific deliverables for the project; therefore, it was perhaps not surprising that these KPIs were regarded as not applicable, as they would not be measurable by evaluation tasks in the project. This was an essential and supportive exercise in communicating the work of the consortium evaluation team to the broader stakeholder group and subsequently getting consensus on the process to be used.

Delphi Survey and Consensus

Value statements were created for each of the KPIs, and the list was revised to ensure that the wording was clear in terms of what the KPI was anticipated to measure in order to meet the SMART criteria defined in the World Health Organization guide for indicators: Specific, Measurable, Attainable, Relevant, and Time-Bound [Monitoring and evaluating digital health interventions: a practical guide to conducting research and assessment. World Health Organization. 2016. URL: https://www.who.int/publications/i/item/9789241511766 [accessed 2025-03-11] 21]. All members of the Gravitate Health consortium were then invited by email to contribute to the Delphi survey (n=250). The decision for inclusion of all members was made on consideration of the results from a previous scoping review, which highlighted a significant gap in the evidence-based knowledge in this domain, and strongly emphasized the involvement of stakeholders in developing KPIs for DHIs [Brenner M, Weir A, McCann M, Doyle C, Hughes M, Moen A, et al. Development of the key performance indicators for digital health interventions: a scoping review. Digit Health. 2023;9:20552076231152160. [FREE Full text] [CrossRef] [Medline]12]. A 2-round ranking style Delphi was performed using the Qualtrics platform (version 082024) to identify and rank the final KPIs for use in the project (Figure 1) [Strasser A. Design and evaluation of ranking-type Delphi studies using best-worst-scaling. Technol Anal Strateg Manag. 2018;31(4):492-501. [CrossRef]22]. In round 1, all potential consortium members were sent a cover letter advising them of the purpose of the exercise. Survey items were presented as 2 questions for each of the 6 KPI categories. In the first question, consortium members were presented with the list of proposed KPIs for the category and asked to choose 1 KPI from the list they felt most represented the category and 1 KPI they felt least represented the category. This is in line with the commonly used and reliable approach of developing rank order in Delphi survey responses [Beiderbeck D, Frevel N, von der Gracht HA, Schmidt S, Schweitzer V. Preparing, conducting, and analyzing Delphi surveys: cross-disciplinary practices, new directions, and advancements. MethodsX. 2021;8:101401. [FREE Full text] [CrossRef] [Medline]23-Del Grande C, Kaczorowski J. Rating versus ranking in a Delphi survey: a randomized controlled trial. Trials. 2023;24(1):543. [CrossRef] [Medline]25]. The second question was a free-text box asking consortium members if they would like to suggest an additional KPI that could be reflective of the category. In the first round of the survey, KPIs were listed in random order for each of the categories. Consortium members had 10 days to respond to round 1 and a further 10 days to respond to round 2 of the survey.

In round 2, consortium members were presented with the KPIs in ranked order based on results from the first round and again asked to rank the list in the order of most to least relevant. There were no differences in how consortium members ranked the KPIs in terms of relevance and priority in round 2. Minor wording edits were made during data analysis by study authors (MB and AW) to some of the KPIs in round 2 based on a small number of free-text feedback to increase the clarity and precision of the KPIs.

Figure 1. Flow diagram for consensus process for KPI development, refinement and prioritisation for evaluation of the Gravitate - Health project.

Ethical Considerations

The conduct of this consultative process was solely intended for decision-making in relation to refining and prioritizing a predetermined list of KPIs. Therefore, ethics approval was not required. Participation in this process was voluntary, and participant engagement in this consultative process was regarded as consent. The Qualtrics platform was used to ensure all responses were anonymous. No compensation was given for engagement in the consultative process.


There were 40 respondents in the first round of the project and 41 respondents in the second round, which represents a response rate of 16%. As a research methodology, the Delphi technique has been criticized in health research for methodological issues such as low response rates [Shang Z. Use of Delphi in health sciences research: a narrative review. Medicine (Baltimore). 2023;102(7):e32829. [FREE Full text] [CrossRef] [Medline]26]. However, as a process for consensus achievement in refining and prioritizing KPIs, the characteristics of the consortium members who responded in this process are representative of all key stakeholders in the Gravitate Health consortium. Just over 41% (n=17) from the pharmaceutical industry and 32% (n=13) from digital technology companies provided feedback in round 1 and 2. Other consortium members represented academic or research institutes (n=14, 34%; Table 1). Respondents listing their role as other (n=13, 32%) included the following: medical doctor, patient advocate, medicine agency, health professional working in academia, regulatory authority, user advisory group member, member organization in digital and telehealth, patient representative, and gerontologist or dementia specialist.

The results of the class ranks (the percentage of responses in which each KPI was listed as first in the order of importance) and mean ranks (average number of ranks for each KPI) for each of the 6 categories for the first and second Delphi rounds are presented in

Multimedia Appendix 2

Class and mean rank for rounds 1 and 2 of key performance indicator refinement and prioritization.

PDF File (Adobe PDF File), 124 KBMultimedia Appendix 2.

The mean rank was calculated for each KPI within each category. These show the rankings for each KPI and the changes that occurred in the prioritization of individual KPIs between rounds 1 and 2. No additional KPIs emerged that were to be added to the list during this process. The overall final list of prioritized KPIs for the project is presented in Table 2.

Table 1. Characteristics of respondents.
RoleRound 1 (n=40), n (%)Round 2 (n=41), n (%)
Industry16 (40)17 (42)
Academic or research institute14 (35)14 (34)
Health care providers5 (13)3 (7)
Patient representatives6 (15)5 (12)
Private partners4 (10)7 (17)
Digital technical expertise14 (35)13 (32)
Other9 (23)13 (32)
Table 2. Final list of refined and prioritized key performance indicators using consensus methodology.
Class rankKey performance indicator
Access and use

1Measure how usable and accessible the Gravitate Health platform is.

2Measure citizens’ use of Gravitate Health to manage health information.

3Measure if the platform is accessible on the web and personal devices.

4Measure citizen awareness of the importance of self-management.

5Measure European Union citizens’ awareness of Gravitate Health.
Understanding

1Measure if the Gravitate Health platform has educational material at a digital literacy level that is understandable.

2Measure if the Gravitate Health platform provides users with an understanding of the medication benefits and how and why to take medication.

3Measure if the Gravitate Health platform addresses physically, auditorily, or visually challenged users.

4Measure if the Gravitate Health platform has the multilingual capability.

5Measure if the Gravitate Health digital solution provides notifications and updates on prescription or over-the-counter electronic product information.

6Measure if the Gravitate Health platform has reached the maturity of the technology platform at the end of the project.
User experience

1Measure if the Gravitate Health platform increases patient empowerment and activation.

2Measure if the Gravitate Health platform facilitates patient empowerment and activation.

3Measure if users trust the Gravitate Health platform.

4Measure if the Gravitate Health platform improves the patient-provider interaction using G-lens.

5Measure if the Gravitate Health platform is assessed by users as not providing information overload or missing relevant data.

6Measure if the Gravitate Health platform provides user awareness through digital solution features.

7Measure if the Gravitate Health platform provides caregivers (informal or nonprofessional) with a satisfactory experience.

8Measure if the Gravitate Health platform provides health providers with a satisfactory experience.
Patient compliance or adherence

1Measure if the Gravitate Health platform leads to better medication compliance or adherence by users.

2Measure if the Gravitate Health platform leads to better patient health outcomes.

3Measure if the Gravitate Health platform leads to a safer use of medication or therapy administration.

4Measure if the Gravitate Health platform leads to an increased partnership between the patient and the health team.
Two-way communication

1Measure if the Gravitate Health platform leads to increased provider awareness of own bias and understanding of users’ attitudes toward medication compliance or adherence.

2Measure if the Gravitate Health platform will meet legal and privacy requirements while balancing a need to know, usability, and accessibility and will have a no-harm policy.

3Measure if the Gravitate Health platform leads to an increase in patient opt-ins for providing real-world data.
Risk minimization

1Measure if the Gravitate Health platform leads to a risk minimization function.

2Measure if the Gravitate Health platform provides alerts to users to avoid prescribed and over-the-counter drug interactions.

3Measure if the Gravitate Health platform achieves lower risks across populations.

Principal Findings

Consortium members ranked the usability, accessibility of the G-lens, digital literacy, and patient empowerment as fundamental issues in this project. This reflects the drive for more accessible technology in health care, which had gained traction before the COVID-19 pandemic [Ahern DK, Woods SS, Lightowler MC, Finley SW, Houston TK. Promise of and potential for patient-facing technologies to enable meaningful use. Am J Prev Med. 2011;40(5 Suppl 2):S162-S172. [CrossRef] [Medline]27-Whitehead L, Seaton P. The effectiveness of self-management mobile phone and tablet apps in long-term condition management: a systematic review. J Med Internet Res. 2016;18(5):e97. [FREE Full text] [CrossRef] [Medline]29]. The most important KPI identified, risk minimization, relating to alerts for adverse interfering effects, is not surprising, given the previous work completed by the European Union’s Innovative Medicines Initiative WEB-RADR (Recognizing Adverse Drug Reactions), which highlighted the importance of detecting adverse effects early to improve patient safety [WEB RADR: recognising adverse drug reactions. Innovative Health Initiative. 2017. URL: https://www.ihi.europa.eu/projects-results/project-factsheets/web-radr [accessed 2023-09-15] 30]. However, the third-ranked most important item in this category was the need to achieve lower risks across populations. This reflects the importance of developing a health solution that enhances equity of care in medication management. Changes in KPIs were not observed during the refinement process in round 1 and 2 of the Delphi technique. This is not surprising, as the Gravitate Health digital solution did not exist prior to this exercise; therefore, consortium members had no prior experience of them.

Comparison to Prior Work

The Delphi technique framework provides a complementary approach to the World Health Organization framework for monitoring and evaluating DHIs and to existing health system frameworks such as the Donabedian’s framework on quality of care, the Reach, Effectiveness, Adoption, Implementation, and Maintenance framework for public health, the logic model, and the balanced scorecard [Monitoring and evaluating digital health interventions: a practical guide to conducting research and assessment. World Health Organization. 2016. URL: https://www.who.int/publications/i/item/9789241511766 [accessed 2025-03-11] 21,Donabedian A. The quality of care: how can it be assessed? JAMA. 1988;260(12):1743. [CrossRef]31-Kaplan RS, Norton DP. Using the balanced scorecard as a strategic management system. Harvard Bus Rev. 1996:75-85. [FREE Full text]34]. While these frameworks agree that setting indicators for planning and evaluation in DHI projects is important, they do not offer guidance on how this should be achieved. The process used for KPI development, refinement, and prioritization in the Gravitate Health project provides such guidance. It provides a standardized approach to achieving consensus between large groups of stakeholders and therefore can be used in many types of digital health solution projects in identifying and prioritizing KPIs.

The use of a Delphi technique in the Gravitate Health project provided a well-established framework for a structured approach to achieving consensus in developing, refining, and prioritizing KPIs in DHI initiatives involving large groups of diverse stakeholders where there is limited supporting evidence [Brenner M, Weir A, McCann M, Doyle C, Hughes M, Moen A, et al. Development of the key performance indicators for digital health interventions: a scoping review. Digit Health. 2023;9:20552076231152160. [FREE Full text] [CrossRef] [Medline]12]. Current literature in this area shows that most previous initiatives for DHI evaluation used health assessment frameworks that were adapted for DHI evaluation focused on the usability and clinical outcomes of the interventions [Brenner M, Weir A, McCann M, Doyle C, Hughes M, Moen A, et al. Development of the key performance indicators for digital health interventions: a scoping review. Digit Health. 2023;9:20552076231152160. [FREE Full text] [CrossRef] [Medline]12]. However, as Bradway et al [Bradway M, Carrion C, Vallespin B, Saadatfard O, Puigdomènech E, Espallargues M, et al. mHealth assessment: conceptualization of a global framework. JMIR Mhealth Uhealth. 2017;5(5):e60. [FREE Full text] [CrossRef] [Medline]35] noted, new initiatives should be created with the unique properties of DHIs in mind. Though the specific KPIs identified by this project may not be generalizable to all DHI evaluations, this paper offers a useful first step toward a common methodology for the development of KPIs for DHIs.

Limitations

This consultative process has some limitations. First, although the response rate was low (41 consortium members of 250 contributed), it has been noted that a minimum of 20 participants can achieve response stability in Delphi research studies [Shang Z. Use of Delphi in health sciences research: a narrative review. Medicine (Baltimore). 2023;102(7):e32829. [FREE Full text] [CrossRef] [Medline]26]. The list of KPIs was mapped against feedback from health care users in the broader project; however, the project may have benefitted from wider engagement with a greater number of health care professionals, health care users, and health care organizations. This limitation is mitigated somewhat by all KPIs being measured regardless of the ranking, as the purpose of this project was to refine the initial KPI list and to develop an order of priority. For example, no additional KPIs were identified for inclusion. However, having a greater number of health care professionals and health care users may have led to a different outcome. In addition, limitations to this methodology include a limited number of items that can be included in a list, with some suggesting that the number of items be restricted to 3 to 5, with a maximum of 7, resulting in potential bias due to ties among items, response bias, and issues with standardization [Strasser A. Design and evaluation of ranking-type Delphi studies using best-worst-scaling. Technol Anal Strateg Manag. 2018;31(4):492-501. [CrossRef]22,Azzara C. Qualitatively speaking: the focus group vs. in-depth interview debate. Quirk’s Media. 2010;24(6):16. [FREE Full text]36-Okoli C, Pawlowski SD. The Delphi method as a research tool: an example, design considerations and applications. Inform Manag. Dec 2004;42(1):15-29. [FREE Full text] [CrossRef]39].

Future Directions

Future work will conduct an external validation to measure the feasibility, reliability, or validity to demonstrate the robustness and applicability of the developed KPIs. This could be done by conducting more pilot testing of the indicators in real-world settings, assessing interrater reliability, and exploring the relationship between the KPIs and relevant outcomes.

Conclusions

The objective of this paper was to present a consensus approach to refining and prioritizing candidate KPIs to measure the success of novel digital health solutions produced by the Gravitate Health project. The Delphi technique provided a structured framework for this, and albeit, a very novel context, it proved to be an effective approach. The final list of KPIs provides a list of substantial indicators sensitive to the needs of a broad group of stakeholders that can be used to capture real-world data in developing and evaluating DHIs. This consultative approach offers a suitable 4-stage process to enable the setting of evaluation indicators for the assessment of DHIs (Figure 1).

Acknowledgments

The authors disclosed receipt of the following financial support for the research, authorship, and publication of this paper: this project has received funding from the Innovative Medicines Initiative 2 Joint Undertaking (grant agreement 945334). The Joint Undertaking receives support from the European Union’s Horizon 2020 research and innovation program, the European Federation of Pharmaceutical Industries and Associations, and Datapharm Limited. This communication reflects the views of the authors and neither the Innovative Medicines Initiative nor the European Union, European Federation of Pharmaceutical Industries and Associations, or Datapharm Limited are liable for any use that may be made of the information contained herein. The authors wish to acknowledge the contribution of Koen Nauwelaerts to discussions and decisions about using the Delphi technique and the conduct of the process of refining and prioritizing key performance indicators, and also in reviewing manuscript drafts.

Data Availability

The datasets generated or analyzed during this study are available from the corresponding author on reasonable request.

Authors' Contributions

MB and AW conducted the Delphi project. MB, CM, and AW reviewed and interpreted Delphi project findings. MB and CM wrote the first draft of the manuscript. AM, MI, MB, CM, YQ, MH, CD, MM, and LC reviewed and edited the manuscript.

Conflicts of Interest

MI is a board member of Context Vision AB and Inify Laboratories AB. MI has a collaboration with Florio GMBH, Germany, that develops applications for patient support. These are unrelated to the present work.

Multimedia Appendix 1

Process of mapping key performance indicators to evaluation tools and time points.

DOCX File , 26 KB

Multimedia Appendix 2

Class and mean rank for rounds 1 and 2 of key performance indicator refinement and prioritization.

PDF File (Adobe PDF File), 124 KB

  1. Zeeb H, Pigeot I, Schüz B. Digital public health-rapid technological progress, but many open public health questions. Bundesgesundheitsblatt Gesundheitsforschung Gesundheitsschutz. 2020;63(2):135-136. [FREE Full text] [CrossRef] [Medline]
  2. Wienert J, Jahnel T, Maaß L. What are digital public health interventions? First steps toward a definition and an intervention classification framework. J Med Internet Res. 2022;24(6):e31921. [FREE Full text] [CrossRef] [Medline]
  3. Schoeppe S, Alley S, Van Lippevelde W, Bray NA, Williams SL, Duncan MJ, et al. Efficacy of interventions that use apps to improve diet, physical activity and sedentary behaviour: a systematic review. Int J Behav Nutr Phys Act. 2016;13(1):127. [FREE Full text] [CrossRef] [Medline]
  4. Digital health platform handbook: building a digital information infrastructure (infostructure) for health. World Health Organization. 2020. URL: https://iris.who.int/handle/10665/337449 [accessed 2025-03-06]
  5. Sieben A, Bredie SJHB, van Laarhoven CJHMK, Schoonhoven L, Burger DM, van Onzenoort HAW. e-Health interventions and improvement in treatment adherence. Ned Tijdschr Geneeskd. 2014;158:A8460. [Medline]
  6. Pouls BPH, Vriezekolk JE, Bekker CL, Linn AJ, van Onzenoort HAW, Vervloet M, et al. Effect of interactive eHealth interventions on improving medication adherence in adults with long-term medication: systematic review. J Med Internet Res. 2021;23(1):e18901. [FREE Full text] [CrossRef] [Medline]
  7. Wang Z, Zhu Y, Cui L, Qu B. Electronic health interventions to improve adherence to antiretroviral therapy in people living with HIV: systematic review and meta-analysis. JMIR Mhealth Uhealth. 2019;7(10):e14404. [CrossRef] [Medline]
  8. Jeminiwa R, Hohmann L, Qian J, Garza K, Hansen R, Fox BI. Impact of eHealth on medication adherence among patients with asthma: a systematic review and meta-analysis. Respir Med. 2019;149:59-68. [FREE Full text] [CrossRef] [Medline]
  9. Handbook on health inequality monitoring: with a special focus on low-and middle-income countries. World Health Organization. 2013. URL: https:/​/www.​who.int/​docs/​default-source/​gho-documents/​health-equity/​handbook-on-health-inequality-monitoring/​handbook-on-health-inequality-monitoring.​pdf [accessed 2025-03-11]
  10. Khalil H, Kynoch K. Implementation of sustainable complex interventions in health care services: the triple C model. BMC Health Serv Res. 2021;21(1):143. [CrossRef] [Medline]
  11. Varela T, Zamorano P, Muñoz P, Rain C, Irazoqui E, Sapag JC, et al. Evaluation of the implementation progress through key performance indicators in a new multimorbidity patient-centered care model in Chile. BMC Health Serv Res. 2023;23(1):439. [FREE Full text] [CrossRef] [Medline]
  12. Brenner M, Weir A, McCann M, Doyle C, Hughes M, Moen A, et al. Development of the key performance indicators for digital health interventions: a scoping review. Digit Health. 2023;9:20552076231152160. [FREE Full text] [CrossRef] [Medline]
  13. Oakley J, O'Hagan A. SHELF: the Sheffield ellicitation framework (version 2.0). 2014. URL: https://shelf.sites.sheffield.ac.uk/ [accessed 2025-01-29]
  14. Rossi PA, Anderson A. The factorial survey approach: an introduction. In: Rossi P, Nock S, editors. Measuring Social Judgments: The Factorial Survey Approach. Beverly Hills, CA. Sage; 1982.
  15. Dalkey N, Helmer O. An experimental application of the DELPHI method to the use of experts. Manag Sci. 1963;9(3):458-467. [CrossRef]
  16. Allen J, Brenner M, Hauer J, Molloy E, McDonald D. Severe neurological impairment: a Delphi consensus-based definition. Eur J Paediatr Neurol. 2020;29:81-86. [FREE Full text] [CrossRef] [Medline]
  17. Nasa P, Jain R, Juneja D. Delphi methodology in healthcare research: how to decide its appropriateness. World J Methodol. 2021;11(4):116-129. [CrossRef] [Medline]
  18. Steel M, Seaton P, Christie D, Dallas J, Absalom I. Nurse perspectives of nurse-sensitive indicators for positive patient outcomes: a Delphi study. Collegian. 2021;28(2):145-156. [CrossRef]
  19. Horne R, Weinman J. Self-regulation and self-management in asthma: exploring the role of illness perceptions and treatment beliefs in explaining non-adherence to preventer medication. Psychol Health. 2002;17(1):17-32. [CrossRef]
  20. Brooke J. SUS: a 'quick and dirty' usability scale. In: Usability Evaluation in Industry. London. CRC Press; 1996.
  21. Monitoring and evaluating digital health interventions: a practical guide to conducting research and assessment. World Health Organization. 2016. URL: https://www.who.int/publications/i/item/9789241511766 [accessed 2025-03-11]
  22. Strasser A. Design and evaluation of ranking-type Delphi studies using best-worst-scaling. Technol Anal Strateg Manag. 2018;31(4):492-501. [CrossRef]
  23. Beiderbeck D, Frevel N, von der Gracht HA, Schmidt S, Schweitzer V. Preparing, conducting, and analyzing Delphi surveys: cross-disciplinary practices, new directions, and advancements. MethodsX. 2021;8:101401. [FREE Full text] [CrossRef] [Medline]
  24. Robertson M, Line M, Jones S, Thomas S. International students, learning environments and perceptions: a case study using the Delphi technique. Higher Educ Res Dev. 2010;19(1):89-102. [CrossRef]
  25. Del Grande C, Kaczorowski J. Rating versus ranking in a Delphi survey: a randomized controlled trial. Trials. 2023;24(1):543. [CrossRef] [Medline]
  26. Shang Z. Use of Delphi in health sciences research: a narrative review. Medicine (Baltimore). 2023;102(7):e32829. [FREE Full text] [CrossRef] [Medline]
  27. Ahern DK, Woods SS, Lightowler MC, Finley SW, Houston TK. Promise of and potential for patient-facing technologies to enable meaningful use. Am J Prev Med. 2011;40(5 Suppl 2):S162-S172. [CrossRef] [Medline]
  28. Alessa T, Abdi S, Hawley MS, de Witte L. Mobile apps to support the self-management of hypertension: systematic review of effectiveness, usability, and user satisfaction. JMIR Mhealth Uhealth. 2018;6(7):e10723. [CrossRef] [Medline]
  29. Whitehead L, Seaton P. The effectiveness of self-management mobile phone and tablet apps in long-term condition management: a systematic review. J Med Internet Res. 2016;18(5):e97. [FREE Full text] [CrossRef] [Medline]
  30. WEB RADR: recognising adverse drug reactions. Innovative Health Initiative. 2017. URL: https://www.ihi.europa.eu/projects-results/project-factsheets/web-radr [accessed 2023-09-15]
  31. Donabedian A. The quality of care: how can it be assessed? JAMA. 1988;260(12):1743. [CrossRef]
  32. Glasgow RE, Vogt TM, Boles SM. Evaluating the public health impact of health promotion interventions: the RE-AIM framework. Am J Public Health. 1999;89(9):1322-1327. [CrossRef] [Medline]
  33. Kellogg W. Logic Model Development Guide. Battle Creek, MI. WK Kellogg Foundation; 2004.
  34. Kaplan RS, Norton DP. Using the balanced scorecard as a strategic management system. Harvard Bus Rev. 1996:75-85. [FREE Full text]
  35. Bradway M, Carrion C, Vallespin B, Saadatfard O, Puigdomènech E, Espallargues M, et al. mHealth assessment: conceptualization of a global framework. JMIR Mhealth Uhealth. 2017;5(5):e60. [FREE Full text] [CrossRef] [Medline]
  36. Azzara C. Qualitatively speaking: the focus group vs. in-depth interview debate. Quirk’s Media. 2010;24(6):16. [FREE Full text]
  37. Cohen SS, Di Minin A, Motoyama Y, Palmberg C. The persistence of home bias for important R&D in wireless telecom and automobiles. Rev Policy Res. 2009;26(1-2):55-76. [CrossRef]
  38. Cohen S, Orme B. What's your preference? Asking survey respondents about their preferences creates new scaling decisions. Mark Res. 2004;16:32-37. [FREE Full text]
  39. Okoli C, Pawlowski SD. The Delphi method as a research tool: an example, design considerations and applications. Inform Manag. Dec 2004;42(1):15-29. [FREE Full text] [CrossRef]


DHI: digital health intervention
KPI: key performance indicator
RADR: Recognizing Adverse Drug Reactions


Edited by A Mavragani; submitted 13.11.24; peer-reviewed by R Katiri, W Evans; comments to author 28.01.25; revised version received 13.02.25; accepted 21.02.25; published 16.04.25.

Copyright

©Catherine McCabe, Leona Connolly, Yuri Quintana, Arielle Weir, Anne Moen, Martin Ingvar, Margaret McCann, Carmel Doyle, Mary Hughes, Maria Brenner. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 16.04.2025.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.