Original Paper
1Department of Information Systems, W. P. Carey School of Business, Arizona State University, Tempe, AZ, United States
2Accounting Information Systems, Broad College of Business, Michigan State University, Lansing, MI, United States
3Heinz College of Information Systems and Public Policy, Carnegie Mellon University, Pittsburgh, PA, United States
Corresponding Author:
Rema Padman, PhD
Heinz College of Information Systems and Public Policy
Carnegie Mellon University
5000 Forbes Avenue
Pittsburgh, PA, 15213
United States
Phone: 1 412 268 2180
Email: rpadman@andrew.cmu.edu
Abstract
Background: An estimated 93% of adults in the United States access the internet, with up to 80% looking for health information. However, only 12% of US adults are proficient enough in health literacy to interpret health information and make informed health care decisions meaningfully. With the vast amount of health information available in multimedia formats on social media platforms such as YouTube and Facebook, there is an urgent need and a unique opportunity to design an automated approach to curate online health information using multiple criteria to meet the health literacy needs of a diverse population.
Objective: This study aimed to develop an automated approach to assessing the understandability of patient educational videos according to the Patient Education Materials Assessment Tool (PEMAT) guidelines and evaluating the impact of video understandability on viewer engagement. We also offer insights for content creators and health care organizations on how to improve engagement with these educational videos on user-generated content platforms.
Methods: We developed a human-in-the-loop, augmented intelligence approach that explicitly focused on the human-algorithm interaction, combining PEMAT-based patient education constructs mapped to features extracted from the videos, annotations of the videos by domain experts, and cotraining methods from machine learning to assess the understandability of videos on diabetes and classify them. We further examined the impact of understandability on several dimensions of viewer engagement with the videos.
Results: We collected 9873 YouTube videos on diabetes using search keywords extracted from a patient-oriented forum and reviewed by a medical expert. Our machine learning methods achieved a weighted precision of 0.84, a weighted recall of 0.79, and an F1-score of 0.81 in classifying video understandability and could effectively identify patient educational videos that medical experts would like to recommend for patients. Videos rated as highly understandable had an average higher view count (average treatment effect [ATE]=2.55; P<.001), like count (ATE=2.95; P<.001), and comment count (ATE=3.10; P<.001) than less understandable videos. In addition, in a user study, 4 medical experts recommended 72% (144/200) of the top 10 videos ranked by understandability compared to 40% (80/200) of the top 10 videos ranked by YouTube’s default algorithm for 20 ramdomly selected search keywords.
Conclusions: We developed a human-in-the-loop, scalable algorithm to assess the understandability of health information on YouTube. Our method optimally combines expert input with algorithmic support, enhancing engagement and aiding medical experts in recommending educational content. This solution also guides health care organizations in creating effective patient education materials for underserved health topics.
doi:10.2196/56080
Keywords
Introduction
Research Background
Limited health literacy is a worldwide challenge [Alemayehu G, Chang P, Delaney M, Dietscher C, Dowse R, Khorshid O, et al. Health literacy around the world: policy approaches to wellbeing through knowledge and empowerment. The Economist Intelligence Unit. 2021. URL: https://impact.economist.com/perspectives/sites/default/files/lon_-_es_-_health_literacy_paper_v8_0.pdf [accessed 2025-03-17] 1]. The World Health Organization defines health literacy as “the cognitive and social skills which determine the motivation and ability of individuals to gain access, to understand, and use information in ways which promote and maintain good health.” It is estimated that almost 80% of the US adult population seek online health information [Finney Rutten LJ, Blake KD, Greenberg-Worisek AJ, Allen SV, Moser RP, Hesse BW. Online health information seeking among US adults: measuring progress toward a healthy people 2020 objective. Public Health Rep. 2019;134(6):617-625. [FREE Full text] [CrossRef] [Medline]2]. However, only 12% of adults in the United States are considered proficient in their ability to meaningfully interpret health information [Kutner M, Greenburg E, Jin Y, Paulsen C. The health literacy of America’s adults; results from the 2003 National Assessment of Adult Literacy. National Center for Education Statistics. Sep 2006. URL: https://nces.ed.gov/pubs2006/2006483.pdf [accessed 2025-03-14] 3]. Recently, a large-scale study assessing COVID-19–related health communications from state and federal agencies found that most information, including those from the US public health agency, the Centers for Disease Control and Prevention (CDC), exceeded recommended reading levels [Mishra V, Dexter JP. Comparison of readability of official public health information about COVID-19 on websites of international agencies and the governments of 15 countries. JAMA Netw Open. Aug 03, 2020;3(8):e2018033. [FREE Full text] [CrossRef] [Medline]4]. Health literacy is well recognized as a challenge for both individual and public health, with many adults lacking the requisite skills to engage successfully in the management of their health and health care [Moorhead SA, Hazlett DE, Harrison L, Carroll JK, Irwin A, Hoving C. A new dimension of health care: systematic review of the uses, benefits, and limitations of social media for health communication. J Med Internet Res. Apr 23, 2013;15(4):e85. [FREE Full text] [CrossRef] [Medline]5]. Therefore, providing access to high-quality health information and patient education materials is essential for empowering patients, improving health and cost outcomes, and building societal resilience.
The internet has reduced much of the information asymmetry between health care providers and consumers by offering multiple avenues whereby patients can educate themselves with both user-generated and domain expert–generated content. However, for health consumers who search for health information on digital media platforms, health literacy divides can be exacerbated both by their own lack of knowledge and by algorithmic recommendations, with results that disproportionately impact minority groups and low health literacy populations [Agarwal R, Bjarnadottir M, Rhue L, Dugas M, Crowley K, Clark J, et al. Addressing algorithmic bias and the perpetuation of health inequities: an AI bias aware framework. Health Policy Technol. Mar 2023;12(1):100702. [CrossRef]6]. Health consumers with higher health literacy levels seek online health information more frequently than those with lower health literacy levels [Lee HY, Jin SW, Henning-Smith C, Lee J, Lee J. Role of health literacy in health-related information-seeking behavior online: cross-sectional study. J Med Internet Res. Jan 27, 2021;23(1):e14088. [FREE Full text] [CrossRef] [Medline]7] and report fewer difficulties in accessing high-quality, understandable health information.
Social media platforms such as YouTube (Google LLC), Instagram (Meta Platforms), and Facebook (Meta Platforms) have gained popularity among those searching for online health information due to the ease of posting and disseminating health information in multimedia format [Chou WY, Gaysynsky A, Trivedi N, Vanderpool RC. Using social media for health: national data from HINTS 2019. J Health Commun. Mar 04, 2021;26(3):184-193. [CrossRef] [Medline]8]. With patients regularly turning to social media platforms for health information and advice, pointing patients toward understandable and trustworthy video materials when needed is one mechanism to bridge the vast divide in health literacy and to enhance a patient’s ability to self-manage their medical conditions. While several patient education guidelines promote understandability and clear communication [Shoemaker SJ, Wolf MS, Brach C. Development of the Patient Education Materials Assessment Tool (PEMAT): a new measure of understandability and actionability for print and audiovisual patient information. Patient Educ Couns. Sep 2014;96(3):395-403. [FREE Full text] [CrossRef] [Medline]9], particularly, it is not clear whether social media platforms are following such evidence-based guidelines for online information dissemination [National action plan to improve health literacy. U.S. Department of Health and Human Services, Office of Disease Prevention and Health Promotion. 2010. URL: https://odphp.health.gov/sites/default/files/2019-09/Health_Literacy_Action_Plan.pdf [accessed 2025-03-14] 10]. Health information from social media sources needs to be assessed on multiple criteria, such as understandability, accuracy and timeliness of the medical content, production quality, and credibility and trustworthiness of the content creators, among others.
The US Agency for Healthcare Research and Quality (AHRQ), a division of the federal Department of Health and Human Services, defines patient educational materials as understandable when consumers of diverse backgrounds and varying levels of health literacy can process and explain key health-related messages [Shoemaker SJ, Wolf MS, Brach C. Development of the Patient Education Materials Assessment Tool (PEMAT): a new measure of understandability and actionability for print and audiovisual patient information. Patient Educ Couns. Sep 2014;96(3):395-403. [FREE Full text] [CrossRef] [Medline]9]. The Patient Education Materials Assessment Tool (PEMAT) is a systematic approach developed by AHRQ to evaluate and compare the understandability and actionability of patient educational materials in audiovisual format [Shoemaker SJ, Wolf MS, Brach C. Development of the Patient Education Materials Assessment Tool (PEMAT): a new measure of understandability and actionability for print and audiovisual patient information. Patient Educ Couns. Sep 2014;96(3):395-403. [FREE Full text] [CrossRef] [Medline]9]. It is designed to be used by health care professionals to help determine whether patients will be able to understand and act on the information presented in the educational materials. PEMAT is the only guideline that includes a measure of objective assessment of audiovisual materials. Therefore, it has been widely adopted in evaluating patient educational materials in video format [Vishnevetsky J, Walters CB, Tan KS. Interrater reliability of the Patient Education Materials Assessment Tool (PEMAT). Patient Educ Counsel. Mar 2018;101(3):490-496. [CrossRef] [Medline]11].
With the vast number of user-generated videos available on social media, there is both an urgent need and a unique opportunity to devise an automated approach to evaluate multimedia health information using multiple criteria. While patient educational guidelines such as PEMAT offer critical insights on how the materials should be evaluated for understandability and actionability, relying on health care professionals to manually annotate all the videos is not sustainable or scalable.
Research Objective
In this study, we aimed to address these gaps via 2 main objectives. First, we developed a human-in-the-loop augmented intelligence approach to assess the understandability of patient educational videos deployed on the YouTube platform. We gathered a diverse range of diabetes-related educational videos from YouTube, leveraging its position as the largest video-sharing platform, to create a research test bed. Our second aim was to highlight the importance of understandability for various stakeholders, including content creators and health care providers, as defined by the PEMAT guidelines. Extracting specific multimedia features from these videos, we performed a computational evaluation of understandability in accordance with the PEMAT guidelines for audiovisual materials developed by AHRQ. The selected features were aligned with the PEMAT criteria, and we used a cotraining classification method that incorporated human feedback to ensure accurate assessments. We evaluated the impact of video understandability on viewer engagement, offering insights for content creators on how to improve their educational videos. In addition, we conducted a user study to examine the perceived effectiveness of understandable content from the perspective of health care professionals.
Related Work
Evaluating Online Health Information in Promoting Health Literacy
In this section, we review recent studies that evaluate online health information and their suitability for promoting health literacy based on existing guidelines. To summarize the related literature, we developed a taxonomy in Table 1 that specifically focused on the type of health information, the guidelines or assessment tools used, the criteria used in the evaluation, and the key findings.
Study | Data | Assessment tool | Focus | Finding |
Kang and Lee [Kang SJ, Lee MS. Assessing of the audiovisual patient educational materials on diabetes care with PEMAT. Public Health Nurs. May 2019;36(3):379-387. [CrossRef] [Medline]12], 2019 | A total of 85 videos from hospital websites | PEMATa | Understandability, actionability, and usefulness | The average understandability rating was 49.5%, and the actionability rating was 31.4%. The average usefulness score was 4.3 on a 7-point scale. |
McClure et al [McClure E, Ng J, Vitzthum K, Rudd R. A mismatch between patient education materials about sickle cell disease and the literacy level of their intended audience. Prev Chronic Dis. May 12, 2016;13:E64. [FREE Full text] [CrossRef] [Medline]13], 2016 | A total of 9 print and 4 online patient education materials | SMOGb, PMOSE/IKIRSCHc, PEMAT, and CDCd Clear Communication Index | The reading level of the publicly available patient education materials | Reading levels of available patient education materials exceed the documented average literacy level of the US adult population. |
Johnson et al [Johnson R, Edwards R, Rivers A, Patil C, Walsh S. Evaluating literacy levels of patient education materials for a sickle cell transition group education programme. Health Educ J. Sep 23, 2019;79(3):253-265. [CrossRef]14], 2020 | A total of 4 text-based patient education materials for sickle cell disease | The Flesch Reading Ease Formula, the Flesch-Kincaid Reading Tool, SMOG Readability Formula, the PEMAT, CDC Clear Communication Index, and PMOSE/IKIRSCH tool | Readability, grade level, understandability, and actionability | Literacy levels of the patient education materials were higher than the recommended standards. |
Rooney et al [Rooney MK, Golden DW, Byun J, Lukas RV, Sonabend AM, Lesniak MS, et al. Evaluation of patient education materials for stereotactic radiosurgery from high-performing neurosurgery hospitals and professional societies. Neurooncol Pract. Jan 2020;7(1):59-67. [FREE Full text] [CrossRef] [Medline]15], 2020 | A total of 54 patient education materials from high-performing neurosurgery hospitals and professional societies | Six readability indices | Readability | Publicly available online patient educational materials for stereotactic radiosurgery were written at reading levels above the national recommendation. Furthermore, many lacked information identified as important by patients. |
Williams et al [Williams AM, Muir KW, Rosdahl JA. Readability of patient education materials in ophthalmology: a single-institution study and systematic review. BMC Ophthalmol. Aug 03, 2016;16:133. [FREE Full text] [CrossRef] [Medline]16], 2016 | A total of 950 written patient educational materials | Three guidelines from the AMAe, CDC, and NIHf for written materials | Readability, structure, and presentation | Materials were consistently written at a readability level that was poorly suited for patients with low health literacy. |
Kunze et al [Kunze KN, Krivicich LM, Verma NN, Chahla J. Quality of online video resources concerning patient education for the meniscus: a YouTube-based quality-control study. Arthroscopy. Jan 2020;36(1):233-238. [CrossRef] [Medline]17], 2020 | A total of 50 YouTube videos of meniscus | JAMAg benchmark criteria and Global Quality Score | Quality and reliability | Information on the meniscus found in YouTube videos was of low quality and reliability. |
Sanderson et al [Sanderson SC, Suckiel SA, Zweig M, Bottinger EP, Jabs EW, Richardson LD. Development and preliminary evaluation of an online educational video about whole-genome sequencing for research participants, patients, and the general public. Genet Med. May 2016;18(5):501-512. [FREE Full text] [CrossRef] [Medline]18], 2016 | One YouTube video about genome sequencing | —h | Understandability and knowledge increased | A total of 79% reported that the video was easy to understand, satisfaction scores were high, and knowledge increased significantly. |
Salama et al [Salama A, Panoch J, Bandali E, Carroll A, Wiehe S, Downs S, et al. Consulting "Dr. YouTube": an objective evaluation of hypospadias videos on a popular video-sharing website. J Pediatr Urol. Feb 2020;16(1):70.e1-70.e9. [FREE Full text] [CrossRef] [Medline]19], 2020 | A total of 53 YouTube videos about hypospadias | PEMAT | Understandability and actionability | Only 5.6% of videos were understandable, and 15.1% were actionable. The vast majority of hypospadias-related YouTube content was not appropriate for users with low health literacy. |
Desai et al [Desai T, Shariff A, Dhingra V, Minhas D, Eure M, Kats M. Is content really king? An objective analysis of the public's response to medical videos on YouTube. PLoS One. Dec 18, 2013;8(12):e82469. [FREE Full text] [CrossRef] [Medline]20], 2013 | A total of 607 videos from Mayo Clinic’s social media health network | SAMi | Suitability and user engagement | Health care organizations produce very few videos with high SAM scores. An optimal video is no more likely to engage users than less optimal videos. |
aPEMAT: Patient Education Materials Assessment Tool.
bSMOG: Simple Measure of Gobbledygook.
cPMOSE/IKIRSCH: the Peter Mosenthal and Irwin Kirsch readability formula.
dCDC: Centers for Disease Control and Prevention.
eAMA: American Medical Association.
fNIH: National Institutes of Health.
gJAMA: Journal of The American Medical Association.
hNot applicable.
iSAM: Suitability Assessment of Materials.
Studies have examined online patient educational materials ranging from those created by professionals, such as hospitals and health systems, to user-generated content by the layperson [Rooney MK, Golden DW, Byun J, Lukas RV, Sonabend AM, Lesniak MS, et al. Evaluation of patient education materials for stereotactic radiosurgery from high-performing neurosurgery hospitals and professional societies. Neurooncol Pract. Jan 2020;7(1):59-67. [FREE Full text] [CrossRef] [Medline]15,Kunze KN, Krivicich LM, Verma NN, Chahla J. Quality of online video resources concerning patient education for the meniscus: a YouTube-based quality-control study. Arthroscopy. Jan 2020;36(1):233-238. [CrossRef] [Medline]17]. Several studies have focused on health information in text format [Kang SJ, Lee MS. Assessing of the audiovisual patient educational materials on diabetes care with PEMAT. Public Health Nurs. May 2019;36(3):379-387. [CrossRef] [Medline]12-Johnson R, Edwards R, Rivers A, Patil C, Walsh S. Evaluating literacy levels of patient education materials for a sickle cell transition group education programme. Health Educ J. Sep 23, 2019;79(3):253-265. [CrossRef]14,Williams AM, Muir KW, Rosdahl JA. Readability of patient education materials in ophthalmology: a single-institution study and systematic review. BMC Ophthalmol. Aug 03, 2016;16:133. [FREE Full text] [CrossRef] [Medline]16], while others focused on video content [Kunze KN, Krivicich LM, Verma NN, Chahla J. Quality of online video resources concerning patient education for the meniscus: a YouTube-based quality-control study. Arthroscopy. Jan 2020;36(1):233-238. [CrossRef] [Medline]17-Salama A, Panoch J, Bandali E, Carroll A, Wiehe S, Downs S, et al. Consulting "Dr. YouTube": an objective evaluation of hypospadias videos on a popular video-sharing website. J Pediatr Urol. Feb 2020;16(1):70.e1-70.e9. [FREE Full text] [CrossRef] [Medline]19]. Evaluation guidelines used in prior studies include the Clear Communication Index from the CDC [Johnson R, Edwards R, Rivers A, Patil C, Walsh S. Evaluating literacy levels of patient education materials for a sickle cell transition group education programme. Health Educ J. Sep 23, 2019;79(3):253-265. [CrossRef]14], PEMAT from AHRQ [Salama A, Panoch J, Bandali E, Carroll A, Wiehe S, Downs S, et al. Consulting "Dr. YouTube": an objective evaluation of hypospadias videos on a popular video-sharing website. J Pediatr Urol. Feb 2020;16(1):70.e1-70.e9. [FREE Full text] [CrossRef] [Medline]19], Benchmark criteria from the Journal of the American Medical Association [Kunze KN, Krivicich LM, Verma NN, Chahla J. Quality of online video resources concerning patient education for the meniscus: a YouTube-based quality-control study. Arthroscopy. Jan 2020;36(1):233-238. [CrossRef] [Medline]17], Suitability Assessment of Materials [Desai T, Shariff A, Dhingra V, Minhas D, Eure M, Kats M. Is content really king? An objective analysis of the public's response to medical videos on YouTube. PLoS One. Dec 18, 2013;8(12):e82469. [FREE Full text] [CrossRef] [Medline]20], Global Quality Score [Kunze KN, Krivicich LM, Verma NN, Chahla J. Quality of online video resources concerning patient education for the meniscus: a YouTube-based quality-control study. Arthroscopy. Jan 2020;36(1):233-238. [CrossRef] [Medline]17], and readability indices [Rooney MK, Golden DW, Byun J, Lukas RV, Sonabend AM, Lesniak MS, et al. Evaluation of patient education materials for stereotactic radiosurgery from high-performing neurosurgery hospitals and professional societies. Neurooncol Pract. Jan 2020;7(1):59-67. [FREE Full text] [CrossRef] [Medline]15]. Readability, content organization, and presentation are critical to health care consumers. These factors impact how patients consume educational materials and whether the medical information can be delivered effectively. A host of studies assessing these topics suggest that most education materials may be too complicated for patients to comprehend [Moorhead SA, Hazlett DE, Harrison L, Carroll JK, Irwin A, Hoving C. A new dimension of health care: systematic review of the uses, benefits, and limitations of social media for health communication. J Med Internet Res. Apr 23, 2013;15(4):e85. [FREE Full text] [CrossRef] [Medline]5,Sørensen K, Van den Broucke S, Fullam J, Doyle G, Pelikan J, Slonska Z, et al. Health literacy and public health: a systematic review and integration of definitions and models. BMC Public Health. Jan 25, 2012;12(1):80. [FREE Full text] [CrossRef] [Medline]21], especially for those with low health literacy [Smith SK. Decision‐making and participation in bowel cancer screening: challenges and interventions for low health literacy. In: Proceedings of the 12th International Conference on Communication in Healthcare. 2014. Presented at: ICCH 2014; September 28-October 1, 2014; Amsterdam, The Netherlands.22].
Prior studies have relied on the judgment of domain experts, such as health professionals, to evaluate online health information. Content rated by an expert (such as medical or health professional staff) is the most common approach to assessing videos focused on health education. Health and medical websites are increasingly encouraged to apply for quality certificate assessments as proof of evidence that they are reliable sources of information [Burstin H, Curry S, Ranney ML, Arora V, Wachler BB, Chou WY, et al. Identifying credible sources of health information in social media: phase 2-considerations for non-accredited nonprofit organizations, for-profit entities, and individual sources. NAM Perspect. May 23, 2023;2023:10.31478/202305b. [FREE Full text] [CrossRef] [Medline]23]. However, as the volume of online health information grows exponentially, using experts to evaluate content is not a scalable solution.
Augmented Intelligence and Human-in-the-Loop Training Methods
Evaluation of health care video content requires domain expertise. Given the amount of user-generated video content available, it is not feasible to generate a large, labeled dataset for typical stand-alone machine learning (ML) and natural language processing (NLP) methods. Due to the high level of uncertainty and criticality in health care and problem diversity, our objective was to introduce humanlike cognitive capabilities into artificial intelligence (AI) systems to develop an augmented intelligence approach. While AI, ML and other automation technologies have make substantial advances in recent years, many important health care problems are often solved through the collaboration of human beings and machines [An R, Shen J, Xiao Y. Applications of artificial intelligence to obesity research: scoping review of methodologies. J Med Internet Res. Dec 07, 2022;24(12):e40589. [FREE Full text] [CrossRef] [Medline]24-Wang S, Ning H, Huang X, Xiao Y, Zhang M, Yang EF, et al. Public surveillance of social media for suicide using advanced deep learning models in Japan: time series study from 2012 to 2022. J Med Internet Res. Jun 02, 2023;25:e47225. [FREE Full text] [CrossRef] [Medline]27]. Human-in-the-loop augmented intelligence is defined as an intelligent model that requires human interaction. Bott et al [Bott N, Wexler S, Drury L, Pollak C, Wang V, Scher K, et al. A protocol-driven, bedside digital conversational agent to support nurse teams and mitigate risks of hospitalization in older adults: case control pre-post study. J Med Internet Res. Oct 17, 2019;21(10):e13440. [FREE Full text] [CrossRef] [Medline]28] used human-in-the-loop software design to develop a conversational agent to support nurse teams in mitigating risks of hospitalization. Wang et al [Wang S, Ning H, Huang X, Xiao Y, Zhang M, Yang EF, et al. Public surveillance of social media for suicide using advanced deep learning models in Japan: time series study from 2012 to 2022. J Med Internet Res. Jun 02, 2023;25:e47225. [FREE Full text] [CrossRef] [Medline]27] used a human-in-the-loop method to predict suicidal ideation. In this type of intelligence system, the human is always part of the system and consequently influences the outcome in such a way that the human gives further judgment if a low-confidence result is given by an algorithm. This approach readily allows us to address problems and requirements that may not be easily trained or classified by ML.
Cotraining is a multiview learning paradigm that exploits unlabeled data in addition to labeled data to improve learning performance [Blum A, Mitchell T. Combining labeled and unlabeled data with co-training. In: Proceedings of the Eleventh Annual Conference on Computational Learning Theory. 1998. Presented at: COLT' 98; July 24-26, 1998; Madison, WI. [CrossRef]29]. In ML, unlabeled data are often substantially cheaper and more plentiful than labeled data. YouTube contains thousands of health care–related videos. However, annotating these videos requires significant human effort. Given the domain expertise required, it is not feasible to obtain a large amount of annotated video data. Cotraining trains 2 learners from 2 different views and lets the learners label the most confident unlabeled instances to enlarge the training set of the other learner [Platanios EA, Poon H, Mitchell TM, Horvitz E. Estimating accuracy from unlabeled data: a probabilistic logic approach. In: Proceedings of the 31st International Conference on Neural Information Processing Systems. 2017. Presented at: NIPS'17; December 4-9, 2017; Long Beach, CA.30]. When the 2 learners are inconsistent, a human expert will evaluate the performance and decide on the label. Such a process is repeated until a stopping condition is met. Intuitively, each example contains 2 “views,” and each view contains sufficient information to determine the label of the example. This redundancy implies an underlying structure of the unlabeled data (because they need to be “consistent”), and this structure makes the unlabeled data informative. This approach has been used for a variety of learning problems, including recommender systems [Zhang M, Tang J, Zhang X, Xue X. Addressing cold start in recommender systems: a semi-supervised co-training algorithm. In: Proceedings of the 37th International ACM SIGIR Conference on Research & Development in Information Retrieval. 2014. Presented at: SIGIR '14; July 6-11, 2014; Gold Coast, Australia. [CrossRef]31], text classification [Ma F, Meng D, Xie Q, Li Z, Dong X. Self-paced co-training. In: Proceedings of the 34th International Conference on Machine Learning. 2017. Presented at: ICML'17; August 6-11, 2017; Sydney, Australia.32], NLP [Pierce D, Cardie C. Limitations of co-training for natural language learning from large datasets. In: Proceedings of the 2001 Conference on Empirical Methods in Natural Language Processing. 2001. Presented at: EMNLP 2001; June 3-4, 2001; Pittsburgh, PA.33], and image recognition [Ma F, Meng D, Xie Q, Li Z, Dong X. Self-paced co-training. In: Proceedings of the 34th International Conference on Machine Learning. 2017. Presented at: ICML'17; August 6-11, 2017; Sydney, Australia.32]. The cotraining process is viewed as a combinative label propagation over 2 views. Obtaining labels can be expensive or time consuming because of the involvement of human experts in this research context. Most learning tasks can be made more efficient in terms of labeling cost by intelligently integrating specific unlabeled instances to be labeled by experts.
Research Question
Despite the growing attention of policy makers and health care providers, it is evident that health educational materials remain too complicated for patients to comprehend [Mishra V, Dexter JP. Comparison of readability of official public health information about COVID-19 on websites of international agencies and the governments of 15 countries. JAMA Netw Open. Aug 03, 2020;3(8):e2018033. [FREE Full text] [CrossRef] [Medline]4]. Evaluation of online health information is an urgent issue and is amplified when considering that 80% of Americans search for health information online, and only 12% have proficient health literacy to correctly interpret and use it. In this study, we seek to address the following research question: How can we design a scalable, generalizable, and sustainable approach to combine human cognitive power with ML and NLP methods to evaluate the understandability of user-generated video content for patient education and health literacy promotion?
Methods
Overview
We propose a scalable, generalizable, and augmented-intelligence-based cotraining approach to assess the understandability of YouTube videos for patient education. Our method can be used outside patient education in the broader context of understandability of health information. Figure 1 illustrates our approach, which consists of 5 components: video collection, video analysis, text analysis, cotraining approach for video understandability assessment, and understandability evaluation.

Research Context
YouTube, the largest video-sharing social media platform, hosts >100 million videos providing information on the pathogenesis, diagnosis, treatments, and prevention of various medical conditions [Madathil KC, Rivera-Rodriguez AJ, Greenstein JS, Gramopadhye AK. Healthcare information on YouTube: a systematic review. Health Informatics J. Sep 2015;21(3):173-194. [FREE Full text] [CrossRef] [Medline]34]. Patient educational materials in visual format may be easier to comprehend and adhere to. For patients who need complex medical information, health care advice in a video format may make it more convenient, understandable, and actionable and improve the outcomes and efficiency of care. While this plethora of user-generated content can be leveraged by patients to improve health literacy and adherence to treatments, criticisms of social media use in health care have also raised serious concerns [Guni A, Normahani P, Davies A, Jaffer U. Harnessing machine learning to personalize web-based health care content. J Med Internet Res. Oct 19, 2021;23(10):e25497. [FREE Full text] [CrossRef] [Medline]35]. Our study can offer a path toward patient education and empowerment and improved health literacy of the population by providing clinicians and patients with the ability to easily retrieve understandable and relevant video-based health education content.
This study focuses on the content assessment for diabetes, as it is among the most prevalent chronic condition in the United States and many other parts of the world. More than 100 million US adults are now living with diabetes or prediabetes, according to the CDC 2020 [National diabetes statistics report 2020: estimates of diabetes and its burden in the United States. Centers for Disease Control and Prevention. 2020. URL: https://diabetesresearch.org/wp-content/uploads/2022/05/national-diabetes-statistics-report-2020.pdf [accessed 2023-12-06] 36]. Diabetes is a contributing factor to many other serious health conditions, such as heart disease, stroke, nerve and kidney diseases, and vision loss [National diabetes statistics report 2020: estimates of diabetes and its burden in the United States. Centers for Disease Control and Prevention. 2020. URL: https://diabetesresearch.org/wp-content/uploads/2022/05/national-diabetes-statistics-report-2020.pdf [accessed 2023-12-06] 36-Deshpande AD, Harris-Hayes M, Schootman M. Epidemiology of diabetes and diabetes-related complications. Phys Ther. Nov 2008;88(11):1254-1264. [FREE Full text] [CrossRef] [Medline]38]. To reduce the impact of prediabetes and type 2 diabetes, health care institutions and medical professionals are applying a multipronged approach to increase awareness of diabetes and its consequences and promote patient education on self-management and lifestyle behavior change programs to improve healthy eating habits and increase physical activity [Shrivastava SR, Shrivastava PS, Ramasamy J. Role of self-care in management of diabetes mellitus. J Diabetes Metab Disord. Mar 05, 2013;12(1):14. [FREE Full text] [CrossRef] [Medline]39]. Clear, multimedia-rich, and trustworthy videos can complement and support clinician and public health efforts.
Video Collection
On the basis of inputs from health care professionals and patient searches on diabetes-related discussion forums, we identified 235 search keywords related to the education of patients with diabetes. These keywords covered various aspects of the education of patients with diabetes, including general information about the disease, treatments, laboratory tests, prevention, self-management procedures, and lifestyle management. These keywords are available in Search terms for creating a corpus of videos. Video attributes collected using YouTube data application programming interface.Multimedia Appendix 1
Multimedia Appendix 2
YouTube offers a diverse range of content and perspectives, featuring contributions from health care professionals, patients, caregivers, and the general public. Our dataset included professionally produced videos by health care organizations and individuals on the basics of diabetes, its complications, and treatments. It also contained research presentations from renowned researchers and medical experts on the latest research developments and scientific findings about the disease. However, low-quality content or inaccurate videos produced by both individuals and organizations also contributed to the diversity, which presented a challenge for assessing video understandability.
Main Outcome Variable: Video Understandability
To obtain the ground truth for our outcome variable, video understandability, we relied on experts’ consensus perspective based on the PEMAT for audio and video materials [Shoemaker SJ, Wolf MS, Brach C. Development of the Patient Education Materials Assessment Tool (PEMAT): a new measure of understandability and actionability for print and audiovisual patient information. Patient Educ Couns. Sep 2014;96(3):395-403. [FREE Full text] [CrossRef] [Medline]9], as it is the only systematic method developed to assess video content. Table 2 lists our adaptation of PEMAT, which focuses on 4 aspects of video materials, specifically: content, word choice and style, organization, and layout and design, with multiple criteria within each aspect. The understandability score of a video was calculated based on the scores for each criterion with the following equation. When a video was scored >50%, it was considered to have high understandability.
(1)
Given the volume and scope of health care videos on YouTube, manual evaluation as well as annotation of a large number of videos by domain experts can be time consuming and costly, hence impractical. Our approach used a semisupervised method called cotraining, which not only learns from the labeled observations but also leverages the unlabeled instances to improve model performance. A total of 600 diabetes-related videos were randomly selected from our corpus of 9873 unique videos as the initial labeled dataset for cotraining. Another 100 videos were sampled for evaluation. Sample size calculation indicated that <500 videos were needed to achieve high interrater reliability (κ>0.80) with multiple raters [Donner A, Rotondi MA. Sample size requirements for interval estimation of the kappa statistic for interobserver agreement studies with a binary outcome and multiple raters. Int J Biostat. 2010;6(1):Article 31. [FREE Full text] [CrossRef] [Medline]40]. The remaining videos were used as unlabeled data to evaluate the effectiveness of cotraining for semisupervision. When the ML models yielded inconsistent results, the medical experts reviewed the videos and provided supervision, according to PEMAT. Four physicians, trained to use these guidelines, labeled these videos for video understandability according to the PEMAT guidelines in Table 2. They watched a video; assessed the video according to the criteria within content, word choice and style, organization, layout, and design; and assigned them 0, 1, or not applicable (N/A).
Figure 2 demonstrates the expert evaluation measures and results. Four domain experts watch a videoo [Type 1 diabetes and viral infections. YouTube. URL: https://www.youtube.com/watch?v=4JLnkpdjoU8 [accessed 2025-03-15]
41] and assess the video according to its content, word choice and style, organization, and layout and design. They assign scores from 0, 1, or N/A to items
Table 2. The video in
Figure 2 is considered to have high understandability.
The PEMAT is designed to be completed by health care professionals, including health care providers, health librarians, and other clinical practitioners. The selected raters fall into the targeted user group who are qualified to use the PEMAT tool to rate the videos. Before they started working on annotation, all of them carefully studied the PEMAT user guide [Shoemaker SJ, Wolf WS, Brach C. The Patient Education Materials Assessment Tool (PEMAT) and user’s guide. Agency for Healthcare Research and Quality. URL: https://www.ahrq.gov/health-literacy/patient-education/pemat.html [accessed 2023-12-06]
42]. To maximize the consistency among these raters, we had each rater independently rate the same 10 videos. A study session was held with these 4 raters to discuss items with discrepancies. Each rater provided their rationale for the rating provided. The group reviewed the PEMAT user guide to clarify how each item was intended to be rated and come to a consensus. Then, they rated the rest of the videos based on the consensus. We used the intraclass correlation coefficient [Bartko JJ. The intraclass correlation coefficient as a measure of reliability. Psychol Rep. Aug 01, 1966;19(1):3-11. [CrossRef] [Medline]43] to assess the interrater reliability of the annotation at the video level. To ensure there is an agreement on every video, a fifth rater reviewed and consolidated the videos with discrepancies. Each video took approximately 10 minutes to review. The interrater reliability of the video understandability score was 87%. Table 3 summarizes video understandability scores (according to the PEMAT guidelines).
Content | Scorea | |
The material makes its purpose completely evident. | 0, 1 | |
Word choice and style | ||
The material uses common, everyday language. | 0, 1 | |
Medical terms are used only to familiarize the audience with the terms. When used, medical terms are defined. | 0, 1 | |
The material uses the active voice. | 0, 1 | |
Organization | ||
The material breaks or “chunks” information into short sections. | 0, 1, N/Ab | |
The material’s sections have informative headers. | 0, 1, N/A | |
The material presents information in a logical sequence. | 0, 1 | |
The material provides a summary. | 0, 1, N/A | |
Layout and design | ||
The text on the screen is easy to read. | 0, 1, N/A | |
The material allows the user to hear the words clearly (eg, not too fast, not garbled). | 0, 1, N/A | |
The material uses illustrations and photographs that are clear and uncluttered. | 0, 1, N/A | |
The material uses simple tables with short and clear row and column headings. | 0, 1, N/A |
aScoring: 0=disagree, 1=agree.
bN/A: not applicable.

Variables | Number of “0” scores (no), n (%) | Number of “1” scores (yes), n (%) | Number of N/Aa, n (%) |
The material makes its purpose completely evident. | 175 (25) | 525 (75) | 0 (0) |
The material uses common, everyday language. | 183 (26.1) | 517 (73.9) | 0 (0) |
Medical terms are used only to familiarize the audience with the terms. When used, medical terms are defined. | 241 (34.4) | 459 (65.6) | 0 (0) |
The material uses the active voice. | 174 (24.9) | 526 (75.1) | 0 (0) |
The material breaks or “chunks” information into short sections. | 548 (78.3) | 143 (20.4) | 9 (1.3) |
The material’s sections have informative headers. | 601 (85.9) | 90 (12.9) | 9 (1.3) |
The material presents information in a logical sequence. | 164 (23.4) | 536 (76.6) | 0 (0) |
The material provides a summary. | 458 (65.4) | 233 (33.3) | 9 (1.3) |
The text on the screen is easy to read. | 137 (19.6) | 294 (42) | 269 (38.4) |
The material allows the user to hear the words clearly. | 97 (13.9) | 539 (77) | 64 (9.1) |
The material uses illustrations and photographs that are clear and uncluttered. | 111 (15.9) | 338 (48.3) | 251 (35.9) |
The material uses simple tables with short and clear row and column headings. | 192 (27.4) | 57 (8.1) | 451 (64.4) |
Understandability | 315 (45) | 385 (55) | 0 (0) |
aN/A: not applicable.
Data Analysis
Video data analysis forms the building blocks for designing our ML approach for evaluating patient educational videos. In processing the video data, we extracted the features according to PEMAT criteria in Table 2 with video analysis techniques from the Google Cloud platform.
Table 4 below summarizes the features we extracted from video data processing results.
The PEMAT guidelines suggests that breaking the information into small chunks or sections is positively related to video understandability. We used scene detection methods to detect the number of scenes in a video as an indicator of whether the videos were organized in small sections. We built on prior work that defined a scene as 1 of the subdivisions of a movie or a play, in which the setting is fixed or when it presents continuous action in 1 place [Rasheed Z, Shah M. Scene detection in Hollywood movies and TV shows. In: Proceedings of the 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. 2003. Presented at: CVPR 2003; June 18-20, 2003; Madison, WI. [CrossRef]44]. A scene comprises a single, complete, and unified event or segment of a movie. A scene normally occurs in 1 location and deals with 1 action; the end of a scene is often indicated by a change in time, action, or location. Scene detection is a widely adopted method in computer vision and video analytics for video classification, video understanding, and management [Xiong W, Lee JC. Efficient scene change detection and camera motion annotation for video classification. Comput Vis Image Underst. Aug 1998;71(2):166-181. [CrossRef]45]. Video content analysis relies on scene detection to extract story units and segments. Scene change detection estimates the subsections in a video [Shahraray B. Scene change detection and content-based sampling of video sequences. In: Proceedings of the SPIE Digtial Video Compression: Algorithms and Technologies 1995. 1995. Presented at: SPIE 1995; February 5-10, 1995; San Jose, CA. [CrossRef]46]. Scene change detection is important in a number of video applications, including video indexing and semantic feature extraction. Video content analysis relies on scene detection to extract story units and segments.
A video transcript is a text version of a video’s audio track. Video transcription techniques can extract video narratives, which convey a significant portion of the information in the videos [Liao H, McDermott E, Senior A. Large scale deep neural network acoustic modeling with semi-supervised training data for YouTube video transcription. In: Proceedings of the 2013 IEEE Workshop on Automatic Speech Recognition and Understanding. 2013. Presented at: ASRU 2013; December 8-12, 2013; Olomouc, Czech Republic. [CrossRef]47]. The quality of narratives also may affect viewers’ understanding of the videos. PEMAT evaluates if the material allows the user to hear the words clearly. We conduct video transcription to perform in-depth content analysis and assess the clarity of narratives. Optical character recognition (OCR) is used to detect and extract text, tables, or illustrations in the videos. Layout and design are an important aspect of patient education video evaluation. Text on the screen should be easy to read. The illustrations and tables should have clear headings. OCR can extract features related to the clarity of text, tables, and illustrations and enable the evaluation of video layout and design.
PEMAT expects a video or multimedia material with narration to allow the viewer to hear the words clearly. The narrator or voiceover should speak at a clear and moderate pace, avoiding speech that is too fast, garbled, or difficult to understand. The video transcription algorithm returns not only the transcript but also the confidence score of the predicted transcript. The confidence score reflects whether the speech is clear. We also assessed whether the text on the screen was easy to read with OCR. Audiovisual materials that were overcrowded with words or had text that flashed briefly on the screen were difficult to read and understand. This item was N/A if no text appeared in the material or a narrator read all the text out loud because the material did not rely on the viewer to read the text. We used OCR to detect text in the videos and the confidence score of OCR as a proxy for whether the text was easy to recognize.
Tasks and features | Description | |
Detect shot changes | ||
Number of scenes in a video | The total number of scenes throughout the video | |
Optical character recognition | ||
Text on screen | A string of text detected in the video | |
Text confidence score | The confidence score of a detected text | |
Video transcription | ||
Transcribed text | The automated video transcription results | |
Transcription confidence score | The confidence score of a transcribed text |
Text Analysis for Videos
Readability
PEMAT requires that the material should use common, everyday language that would be easy to understand for most consumers or patients nearly all the time. To assess this criterion, we conducted a readability analysis and used the Flesch-Kincaid readability test to determine how easily the video’s description and transcription could be understood. The Flesch-Kincaid readability test was developed under contract to the US Navy research in 1975 [Kincaid JP, Fishburne RPJ, Rogers RL, Chissom BS. Derivation of new readability formulas (automated readability index, fog count and flesch reading ease formula) for navy enlisted personnel. Institute for Simulation and Training. 1975. URL: https://stars.library.ucf.edu/cgi/viewcontent.cgi?article=1055&context=istlibrary [accessed 2025-03-14] 48]. It has been widely adopted, especially in the public health domain, to assess how easy it is to read a given material [Basch CH, Mohlman J, Hillyer GC, Garcia P. Public health communication in time of crisis: readability of on-line COVID-19 information. Disaster Med Public Health Prep. Oct 11, 2020;14(5):635-637. [FREE Full text] [CrossRef] [Medline]49,Basch E, Deal AM, Dueck AC, Scher HI, Kris MG, Hudis C, et al. Overall survival results of a trial assessing patient-reported outcomes for symptom monitoring during routine cancer treatment. JAMA. Jul 11, 2017;318(2):197-198. [FREE Full text] [CrossRef] [Medline]50].
Syntactic Analysis
PEMAT assesses whether the material uses active voice. It is often argued that the passive voice will result in a structure that is more verbose than the active voice and, therefore, harder to understand and that the meaning of the passive voice is indirect and less forceful than the active voice [Millar N, Budgell B, Fuller K. ‘Use the active voice whenever possible’: the impact of style guidelines in medical journals. Appl Linguist. Sep 2013;34(4):393-414. [CrossRef]51]. Therefore, the use of active voice is highly advocated in patient educational materials, medical writings, and related areas. According to PEMAT, if the material overall uses active voice, this criterion is met. To automatically assess this criterion, we used part-of-speech tagging, a common linguistic technique, to detect the category of verbs in the video description and narratives and compute the number of verbs in active voice. The number of verbs in active voice was extracted with part-of-speech tagging. The verbs in active voice belong to the following tag set: verb, base form, verb, past tense, verb, gerund or present participle, verb non-3rd person singular present, and verb, 3rd person singular present.
Medical Entity Recognition
We adopted a bidirectional long short-term memory model from prior work to extract 6 types of medical terms from the text data [Xie J, Liu X, Dajun Zeng D. Mining e-cigarette adverse events in social media using Bi-LSTM recurrent neural network with word embedding representation. J Am Med Inform Assoc. Jan 01, 2018;25(1):72-80. [FREE Full text] [CrossRef] [Medline]52]. Table 5 lists the medical term categories and provides explanations. These 6 categories cover most medical terminologies used in patient educational materials and communications [Fage-Butler AM, Nisbeth Jensen M. Medical terminology in online patient-patient communication: evidence of high health literacy? Health Expect. Jun 2016;19(3):643-653. [FREE Full text] [CrossRef] [Medline]53].
A total of 5000 sentences were randomly selected from the video description and transcription test bed, with 4000 in the training set and 1000 in the test set. Two expert annotators independently labeled the sentences for semantic types. We used Cohen κ to measure interannotator reliability. The κ value was 0.90 for the medical terminology annotation. A third annotator reviewed the disagreements and made the final judgments. Finally, the ground truth was generated, containing 4000 training sentences and 1000 test sentences. The statistics of the training and test sets are shown in Table 6.
We trained an embedding model using the skip-gram method in Word2vec and implemented a bidirectional long short-term memory model to extract medical terms from video descriptions and transcriptions at the sentence level. Overall, the model achieved a precision of 87.4%, a recall of 87.8%, and an F1-score of 87.3%. We also conducted several experiments to evaluate the classification performance of our method in comparison to dictionary-based approaches and state-of-the-art methods, such as conditional random fields. Performance is reported in Medical entity extraction performance.Multimedia Appendix 3
Medical term category | UMLSa semantic type | Examples |
Body part | bdsy (body system), blor (body location or region), and bpoc (body part, organ, or organ component) | Liver, foot, and pancreas |
Chemicals or drugs | chem (chemical), chvf (chemical viewed functionally), chvs (chemical viewed structurally), clnd (clinical drug), elii (element, ion, or isotope), enzy (enzyme), hops (hazardous or poisonous substance), inch (inorganic chemical), orch (organic chemical), and phsu (pharmacologic substance) | Insulin, metformin, and lantus |
Medical devices | drdd (drug delivery device) and medd (medical device) | Insulin pen and glucometer |
Medical events | acab (acquired abnormality), dsyn (disease or syndrome), inpo (injury or poisoning), mobd (mental or behavioral dysfunction), patf (pathologic function), and sosy (sign or symptom) | Nausea, ketosis, and diabetes |
Medical professionals | humn (human) and famg (family group) | Physician, diabetes educators, and nurses |
Medical procedures | lbpr (laboratory procedure), lbtr (laboratory or test result), and topp (therapeutic or preventive procedure) | HbA1c and creatinine |
aUMLS: Unified Medical Language System.
Training set | Test set | |
Number of sentences | 4000 | 1000 |
Number of mentions of body part | 227 | 101 |
Number of mentions of chemicals and drugs | 2181 | 538 |
Number of mentions of medical devices | 545 | 126 |
Number of mentions of medical events | 784 | 245 |
Number of mentions of medical professionals | 67 | 18 |
Number of mentions of medical procedures | 197 | 53 |
Semantic Analysis
Multiple PEMAT criteria were evaluated using semantic analysis methods. First, PEMAT expects materials to have a summary of the key points or a review of the key points of the material, either in writing or orally. The summary usually comes at the end of the material and starts with summary words. Therefore, we curated a comprehensive list of summary words and phrases from multiple sources and used them to detect whether a material provided a summary.
Second, PEMAT suggests that information in the material should be presented in an order that makes sense to the user. Main messages or the most important ideas should be at the beginning of sections or in bulleted lists because users tend to pay more attention to them [McClure E, Ng J, Vitzthum K, Rudd R. A mismatch between patient education materials about sickle cell disease and the literacy level of their intended audience. Prev Chronic Dis. May 12, 2016;13:E64. [FREE Full text] [CrossRef] [Medline]13]. To measure whether the material presented information in a logical sequence, we evaluated the use of transitional words and phrases in the material. A transition is a change from one idea to another in writing or speaking and can be achieved using transition terms or phrases, which are most often placed at the beginning of sentences, independent clauses, and paragraphs, to create a clear connection between ideas or groups of ideas. Transitions are used to create “flow” in writing or speaking and make its logical development clearer to the audience. The use of transition words and phrases can improve the logical connections in writing and speech [Oliu WE, Brusaw CT, Alred GJ. Writing That Works: Communicating Effectively on the Job. Boston, MA. Bedford/St. Martin's; 2016. 54]. Transition words and phrases can be grouped into categories such as causation, chronology, combinations, contrast, example, clarification, summary, and more. We collected common transitional terms and phrases under these categories as a proxy to measure whether the material presents information in a logical sequence. Textbox 1 [Summary/conclusion words. University of Waterloo. URL: https://uwaterloo.ca/writing-and-communication-centre/tran sition-words-and-phrases [accessed 2023-12-06]
55,Using transitional words and phrases. University of Wisconsin-Madison. URL: https://writing.wisc.edu/handbook/transitions/ [accessed 2023-12-06]
56] lists all the words and phrases used to identify transitions and summaries.
Third, we evaluated whether the material makes its purpose evident. According to the PEMAT user guide, this criterion refers to whether the material uses a title or upfront text that tells the reader what the material is about. We implemented this criterion by checking whether each video had a title, tags, and description. YouTube suggests that tags are descriptive keywords that content creators can add to the video to help viewers find the content. The video’s title, tags, and description are important pieces of metadata for the video’s discovery and should provide critical information about the purpose of the video so that viewers can find the video and decide whether to watch it.
- Summary or conclusion: Finally, in a word, in brief, briefly, in conclusion, in the end, in the final analysis, on the whole, thus, to conclude, to summarize, in sum, to sum up, in summary, lastly, in short, by and large, consequently, as a result, hence, overall, in conclusion, and after all
- Transition: Accordingly, as a result, and so, because, consequently, for that reason, hence, on account of, since, therefore, thus, after, afterwards, always, at length, during, earlier, following, immediately, in the meantime, later, never, next, once, simultaneously, so far, sometimes, soon, subsequently, then, this time, until now, when, whenever, while, additionally, again, also, and, or, not, besides, even more, finally, first, firstly, further, furthermore, in addition, in the first place, in the second place, last, lastly, moreover, next, second, secondly, after all, although, and yet, at the same time, but, despite, however, in contrast, nevertheless, notwithstanding, on the contrary, on the other hand, otherwise, thought, yet, as an illustration, eg, for example, for instance, specifically, to demonstrate, to illustrate, briefly, critically, foundationally, more importantly, of less importance, primarily, above, centrally, opposite to, adjacent to, below, peripherally, below, nearby, beyond, in similar fashion, in the same way, likewise, in like manner, ie, in other word, that is, to clarify, to explain, in fact, of course, undoubtedly, without doubt, surely, indeed, for this purpose, so that, to this end, in order that, and to that end
Cotraining Approach for Video Understandability Assessment
Overview
We defined the video understandability classification in the context of patient education as a multiview learning and binary classification problem. Due to the vast amount of user-generated videos available and the cost to annotate the videos manually, it is essential to deploy an augmented intelligence approach in this context. The cotraining approach enabled us to accomplish this task with limited human effort and incorporate domain experts’ assessment when the results from ML models were insufficient for unambiguous classification. Our dataset included video metadata and video content data. We developed classifiers from 2 sufficient and conditionally independent views (ie, video metadata and video content) to assess video understandability. Following a feature-based design, we engineered the features based on the evaluation criteria in the PEMAT guideline. The classification model we used in the cotraining approach was logistic regression due to its high interpretability, which can benefit health care organizations and health content creators in their future content creation process.
Cotraining-Based Understandability Classification
Figure 3 illustrates the steps in the cotraining approach for video understandability classification. This consists of the following components: a set of L-labeled videos and a set of U-unlabeled videos, a classifier F1 trained with features from video metadata view, a classifier F2 trained with features from video content view, and a hyperparameter confidence threshold. The video metadata contains the video title, video description, video tags, and video use information. It represents how the content creators would like the viewers to perceive the video. The video content view captures the information delivered by the video. Combining video content and video metadata gives us a comprehensive view of the videos on YouTube. An initial labeled dataset L and an unlabeled dataset U are given. The cotraining process in this study is presented in
Textbox 2.

Input
- A set L of labeled video examples
- A set U of unlabeled video examples
Output
- A set L′ of labeled video examples
Procedures
- Loop for K iterations
- Train a classifier F1 on L that considers only video metadata features and train a classifier F2 on L that considers only features from video content features
- Use the trained classifiers to make predictions on videos in U
- Extract p1 positive and n1 negative examples from U on which F1 has the most confident predictions, specified by a predefined confidence threshold
- Extract p2 positive and n2 negative examples from U on which F2 has the most confident predictions, specified by a predefined confidence threshold
- Compare p1 with p2 and n1 with n2
- If a video appears in both p1 and p2 or in both n1 and n2, move the video and its label from U to L’
- If a video appears in both p1 and n2 or both p2 and n1, expert reviewers annotate inconsistent labels and add the final label of this video to L’
- Halt when U is empty or no new videos are added to the labeled set L’
Features From Video Metadata View
In the video metadata view classifier, we leveraged the features generated from video metadata to classify video understandability. As described earlier, each video’s metadata contains the video title, description, and tags, which are submitted by the content creator, suggesting the purpose of a given video. A video with good understandability for patients uses common daily language, so text preprocessing techniques are used to identify the total number of words, sentences, and unique words from the video description. A bidirectional long short-term memory named entity recognition model is used to extract the number of medical terms, of which those in the video description are included in the metadata view classification. Table 7 summarizes the features we extracted from the video metadata view, the methods used to derive the measure, and the PEMAT criteria they fall under.
Name | Description | Method | PEMATa criterion |
Has title | Whether the video has a title | Metadata collection | The material makes its purpose evident |
Has description | Whether the video has a text description | Metadata collection | The material makes its purpose evident |
Has tags | Whether the video has tags | Metadata collection | The material makes its purpose evident |
Description readability | The automated readability index of the video description | Readability analysis | The material uses common everyday language |
Active word count | The number of verbs in active voice in the video description | Syntactic analysis | The material uses active voice |
Summary word count | The number of summary words in the video description | Semantic analysis | The material provides a summary |
Transition word count | The number of transition words in the video description | Semantic analysis | The material presents information in a logical sequence |
Video duration | The total length of the video in seconds | Metadata collection | Medical information encoded in the video metadata |
Description word count | The total number of words in the video description | Metadata collection | Medical information encoded in the video metadata |
Sentence count | The total number of sentences in the video description | Metadata collection | Medical information encoded in the video metadata |
Description unique words | The total number of unique words in the video description | Metadata collection | Medical information encoded in the video metadata |
Description medical term count | The total number of medical terms in the video description | Medical entity recognition | Medical information encoded in the video metadata |
aPEMAT: Patient Education Materials Assessment Tool.
Features From Video Content View
In the video content view, we derived features from the video narratives, video shots, and associated confidence scores. We generated a narrative readability score to examine whether the narrative material uses common everyday language. Part-of-speech tagging was used to extract verbs in active voice in the transcript. The numbers of transition words and summary words were identified according to the transition word list. We used the video transcription confidence score as a proxy to determine whether the users could hear the words clearly in narratives. Videos are often broken into different chunks by scenes. We used Google Video Intelligence to detect the number of scenes in the video as an indicator to determine if the video had short sections and used text processing methods to generate features from the transcript. Table 8 summarizes the features we extracted from the video content view, the methods to derive the measure, and the PEMAT criteria they fall under.
Feature name | Feature description | Method | PEMATa criterion |
Narrative readability | The automated readability index for narrative | Readability analysis | The material uses common everyday language |
Active word count | The number of verbs in active voice in the video transcript | Syntactic analysis | The material uses active voice |
Summary word count | The number of summary words in the video transcript | Semantic analysis | The material provides a summary |
Transition word count | The number of transition words in the video transcript | Semantic analysis | The material presents information in a logical sequence |
Video transcription confidence | The video transcription confidence score | Auto transcription | The material allows users to hear the words clearly |
Text detection confidence | Text recognition confidence score | Optical character recognition | The text on the screen is easy to read |
Scene count | The total number of scenes in the video | Scene detection | The material breaks or “chunks” information into short sections |
Transcript word count | The total number of words in the video transcript | Auto transcription | Medical information encoded in the video |
Transcript unique word | The total number of unique words in the transcript | Auto transcription | Medical information encoded in the video |
Transcript sentence count | The number of unique words in a video | Auto transcription | Medical information encoded in the video |
Transcript medical term | The total number of medical terms in the video narrative | Medical entity recognition | Medical information encoded in the video |
Video object | The total number of unique objects in the video | Object detection | Medical information encoded in the video |
aPEMAT: Patient Education Materials Assessment Tool.
Evaluating the Impact of Video Understandability on User Engagement
Building on prior studies that examined collective engagement on YouTube, we identify 3 measures of user engagement: video view count, comment count, and like count, which can be obtained from publicly available YouTube metadata [Liu X, Zhang B, Susarlia A, Padman R. Go to YouTube and call me in the morning: use of social media for chronic conditions. MIS Q. 2020;44(1b):257-283. [FREE Full text] [CrossRef]57]. To assess the causal impact of understandability on user engagement, we used a method called coarsened exact matching that reduces the impact of confounding in observational causal inference. Because our data were observational, we could not conduct randomized experiments to vary the level of understandability of videos across users and assess the resultant impact on engagement. Because user engagement may be influenced by a whole host of factors external to the content of a video, we used coarsened exact matching that produces a matched sample of videos according to the covariate distributions in the treatment and control groups (ie, videos classified as understandable or not). To achieve covariate balance, we controlled for a host of heuristic measures of video quality, such as the duration of the video, a good description or a comprehensive narrative, technical quality, credentials, and the number of days since being published.
Ethical Considerations
Our study includes an expert evaluation process, for which we obtained approval from the institutional review board (STUDY00015114: Leveraging YouTube Video Analytics for Patient Education). The evaluation process adheres to the required ethical standards for research involving human participants. Our study did not involve the use of medical records or patient information, as the evaluation was conducted with 4 medical experts to assess the proposed approach. A formal informed consent process was implemented, and the consent form is provided in Consent form.Multimedia Appendix 4
Results
Video Understandability Classification
We collected 9873 videos using the search keywords extracted from a patient-oriented forum and reviewed by a medical expert. Among the 9873 videos, 8963 (90.78%) had descriptions, 8719 (88.31%) had narratives, and 4327 (43.83%) had text embedded in the videos. Videos with both descriptions and narratives were included in the subsequent analyses. We applied text and video analytics techniques to extract metadata view features and video content view features. Tables S1 and S2 in Descriptive statistics of features for video understandability classification. Performance metrics and hyperparameters of the proposed approach.Multimedia Appendix 5
Multimedia Appendix 6
Table 9 shows the coefficients of the logistic regression classifiers for each view. The active word count and summary count had a significant and positive impact on understandability. The transition word count in narratives was significant, but that of description was not. Transcription confidence and text detection confidence had a positive impact on video understandability. Video duration as well as medical terms count in descriptions and transcriptions negatively affected the video understandability. The readability scores of both the description and the narratives had a significant and positive impact on video understandability.
The most significant variables were consistent with PEMAT. Low understandability videos were associated with longer duration, lengthier narratives, and a larger number of medical terminologies. For model performance, we compared our predicted results for the 100 videos included in the evaluation set. Although cross-validation is commonly used in evaluating ML models, it is not feasible to collect a large repository of labeled data to evaluate cotraining models. Therefore, we adopted a holdout evaluation that was usually used for cotraining methods. We compared our model with 3 benchmark models: logistic regression, support vector machines, and random forest. To ensure a fair comparison, we carefully tuned the model hyperparameters to obtain the best performance of the benchmark models and the proposed method. For the logistic regression, we experimented with different solvers and regularization methods. Our best performance model used liblinear solver and L2 regularization. The best performance of support vector machines was achieved by radial basis function kernel and a penalty score of 0.1. The best performance of the random forest model was achieved by max_features = log2, and N_estimator =100. Table 10 summarizes the classification performance of our proposed method and benchmarks. Our approach achieved a weighted precision of 0.84, a weighted recall of 0.79, and an F1-score of 0.81 in classifying videos. The results showed that the cotraining method significantly improved the video understandability classification performance. The classifiers trained on 2 views “teach” each other with the additional examples whose labels are given by the other classifier or human experts and hence improve the classification performance [Du J, Ling CX, Zhou ZH. When does cotraining work in real data? IEEE Trans Knowl Data Eng. May 2011;23(5):788-799. [CrossRef]58].
Variable name | Estimate | P value | |
F1: video metadata view | |||
Has title | −0.335 | .35 | |
Has description | −0.217 | .15 | |
Has tags | −0.184 | .18 | |
Description readability | 0.367a | .07 | |
Active word count | 0.029 | .09 | |
Summary word count | 0.152 | .049 | |
Transition word count | 0.096 | .10 | |
Video duration | −0.071 | .09 | |
Description word count | 0.038 | .14 | |
Sentence count | 0.157 | .12 | |
Description unique words | 0.085 | .14 | |
Description medical term | −0.020 | .07 | |
Constant | −0.319 | .11 | |
F2: video content view | |||
Narrative readability | 0.132 | .03 | |
Active word count | 0.117 | .045 | |
Summary word count | 0.045 | .09 | |
Transition word count | 0.028 | .04 | |
Transcription confidence | 0.021 | .04 | |
Shot count | −0.254 | .20 | |
Transcript word count | −0.036 | .14 | |
Transcript unique word | −0.085 | .07 | |
Transcript sentence count | −0.074 | .14 | |
Transcript medical term | −0.009 | .045 | |
Video object | −0.104 | .06 | |
Constant | −0.272 | .12 |
aItalicization indicates a significance level of P<.10.
Precision | Recall | F1-score | AUCa | |
Cotraining with logistic regression | 0.84 | 0.79 | 0.81 | 0.91 |
Logistic regression | 0.63 | 0.60 | 0.61 | 0.63 |
Support vector machines | 0.77 | 0.75 | 0.76 | 0.78 |
Random forest | 0.80 | 0.74 | 0.77 | 0.81 |
aAUC: area under the curve.
Impact of the Cotraining Process on Classification Performance
The cotraining process combines expert efforts and ML methods to classify the video understandability according to the guidelines in the patient educational domain. One critical issue in the human-algorithm connection is to understand how this collaboration between human experts and ML algorithms improves performance. Figure 4 shows the classification error rate on the test set over iterations of training. In each iteration, new instances were added to the training process, lowering the classification error of the metadata view classifier, video content view classifier, and cotraining classifier. The reduction in classification error shows that this iterative process improves the overall performance. Furthermore, by combining human intelligence and machine intelligence from classifiers of 2 different views, the cotraining approach achieved the best performance.

Impact of Expert Involvement on Classification Performance
The human-algorithm interaction in the cotraining process happens in two stages: (1) medical professionals provide a set of labeled examples to initialize the model training and (2) medical professionals are also involved in the cotraining process when there are inconsistent high-confidence labels predicted by 2 different classifiers. Obtaining inputs from domain experts through a human-in-the-loop algorithm design is essential to our chosen task of assessing the understandability of videos from a patient education perspective. Our design also seeks to minimize human involvement while not compromising performance. To this end, we evaluated the impact of human involvement at different stages of model learning.
Devising a model to classify the understandability of patient educational videos requires high-quality training data. The process of creating training data involves medical professionals reviewing and categorizing videos based on guidelines. Figure 5 shows the classification error by the number of labeled training examples. As ML algorithms are dependent on the quality and quantity of their training data, we observed that increasing the number of labeled examples led to an improvement in performance. However, as for cotraining, the benefit of adding more training examples diminished when we accumulated a significant number of labeled training examples (ie, 500) for video understandability classification.
The confidence threshold determined how many predicted labels we included in the label comparison. Its purpose was to prevent the unlabeled samples from being labeled with the wrong labels, thus decreasing the ability of the learner. On the basis of the label confidence threshold, unlabeled data in each iteration was divided into 3 categories: videos with low-confidence labels, videos with consistent and high-confidence labels, and those with inconsistent but high-confidence labels. The lower the confidence threshold, the higher was the number of videos compared and evaluated in each iteration.
A lower threshold can possibly lead to a faster convergence but, at the same time, needs more human involvement during the cotraining process. A higher confidence threshold may lead to the early stopping of the training process because no new labels meet the confidence threshold. When the confidence threshold is high, the training process stops before assigning labels to all the unlabeled data. We followed the majority rule to assign the predicted labels for these unlabeled samples. We observed a negative impact on the classification performance due to early stopping from a high-confidence threshold. When the confidence threshold is too low, too many unlabeled examples are misclassified, which affects the ability of the cotraining model. As a result, we see a performance decrease when the confidence threshold is too low.

The Importance of Understandability of Video Recommendations by Experts
We conducted a small user study to assess the impact of video understandability on experts’ decisions to recommend a YouTube video for patient education. Our automated methods reranked the search results from 20 randomly selected queries according to their video understandability classification. Four medical experts reviewed the top 10 videos according to our reranked results for each query and reported whether they would recommend the videos to patients. Precision at K is a common information retrieval measure used in modern (web-scale) information retrieval systems [Manning CD, Raghavan P, Schutze H. An Introduction to Information Retrieval. Cambridge, UK. Cambridge University Press; 2009. 59]. In web-scale retrieval, queries have thousands of relevant documents, and few users will be interested in reading all of them. Precision at K (P@K) assesses how many of the top K results are relevant (eg, P@10 or “precision at 10” corresponds to the number of relevant results among the top 10 documents). We measured the average precision at K with K varying from 1 to 10 for the 20 randomly selected queries.
Figure 6 shows a chart comparing the significance of video understandability ranking. In total, 30% (6/20) of the top-ranked videos (videos ranked 1 or top 1) in understandability were recommended by an expert. None of the videos that were top ranked by YouTube’s default ranking received this recommendation. In total, 72% (144/200) of the top 10 videos were recommended by experts ranking by understandability, while only 40% (80/200) of the top 10 YouTube-ranked videos were recommended. We concluded that our understandability classification approach had considerable promise in effectively identifying patient education videos and will be evaluated further in future studies.

Impact of Understandability on Engagement
Table 11 shows the logistic regression results for propensity score matching. After matching, there were 365 videos in the treated group and 597 videos in the control group. The summary of the balance on the entire dataset and the matched dataset is available in the
Table 12. The differences in the variables between the 2 groups were significantly reduced after matching.
The matching provided a way to identify the impact of understandability on a video’s view count, like count, and comment count. The estimated treatment effect on the 3 measures of collective engagement is shown in Table 13. The results suggest that video understandability had a significant and positive impact on all 3 dimensions of engagement. Increasing the understandability of the videos can bring more viewership, more comments, and more likes. Our results highlight the importance of understandability for health care organizations and health practitioners on social media platforms.
Coefficients | Estimate | P value |
Intercept | −9.7470 | <.001 |
Log(ChannelViewCount+1) | 0.5162 | <.001 |
Log(ChannelSubscriberCount+1) | 0.2295 | <.001 |
Log(ChannelVideoCount) | −0.7669 | <.001 |
ContentDefinitionSD | −0.4352 | <.001 |
Duration | 0.0002 | <.001 |
Description word count | −0.0015 | .009 |
Description unique word | 0.0069 | <.001 |
Log(publishedDays+1) | 0.2874 | .002 |
Variables | All data | Matched data | ||||
Treated, mean | Control, mean | Standardized mean difference | Treated, mean | Control, mean | Standardized mean difference | |
Log(channelViewCount+1) | 16.0245 | 12.1291 | 1.5060 | 14.9886 | 14.9739 | 0.0057 |
Log(channelSubscriberCount+1) | 10.6054 | 6.3278 | 1.3476 | 9.3214 | 9.2441 | 0.0243 |
Log(channelVideoCount+1) | 5.3956 | 5.1962 | 0.1114 | 5.3068 | 5.3070 | −0.0001 |
contentDefinitionhd | 0.7960 | 0.5683 | 0.5650 | 0.6986 | 0.6986 | 0 |
contentDefinitionsd | 0.2040 | 0.4317 | −0.5650 | 0.3014 | 0.3014 | 0 |
Duration | 676.7022 | 413.2235 | 0.3008 | 503.0027 | 395.4571 | 0.1228 |
word_count_description | 242.9178 | 146.3287 | 0.3735 | 126.7836 | 129.3711 | −0.01 |
unique_word_description | 127.6597 | 76.5907 | 0.4522 | 72.8219 | 73.8849 | −0.0094 |
Log(publishedDays+1) | 7.4026 | 7.3888 | 0.0295 | 7.5097 | 7.5065 | 0.0069 |
Measure | Estimate | P value | |
Log(view count) | |||
Intercept | 8.8946 | <.001 | |
ATEa | 2.5523 | <.001 | |
Log(like count) | |||
Intercept | 3.2320 | <.001 | |
ATE | 2.9494 | <.001 | |
Log(comment count) | |||
Intercept | 1.6211 | <.001 | |
ATE | 3.0981 | <.001 |
aATE: average treatment effect.
Discussion
Principal Findings
In this study, we developed a human-in-the-loop augmented intelligence approach to assess the understandability of 9873 diabetes education videos on the YouTube social media platform in accordance with PEMAT guidelines. The cotraining classification model, which combined ML with expert input, achieved strong performance (precision=0.84, recall=0.79, and F1-score=0.81). We further examined the impact of understandability and found that higher understandability positively impacted viewer engagement (more views, likes, and comments) and increased the likelihood of expert recommendations for patient education. The findings highlight the importance of improving video understandability for enhancing patient engagement with educational materials on contextually relevant health-related topics, potentially advancing the health literacy of individuals and populations.
Implications for Research and Practice
With complex and very large-scale data hosted by digital platforms, billions of people worldwide are accessing health care information through social media channels without any means of verifying the accuracy, understandability, relevance, and other critical criteria associated with the content being disseminated. Hence, there is an urgent need for an evidence-based approach with replicable, scalable, and generalizable AI-based methods for health literacy promotion and patient education. To the best of our knowledge, our research is the first to attempt a validated guideline-driven consolidation of rich multimedia data sources spanning video metadata and content data in text, audio, video, and structured data formats combined with a human-in-the-loop learning strategy to assess video understandability in the health care domain.
Advocates of social media in medicine highlight social media’s potential to enable patient education and empowerment [Househ M, Borycki E, Kushniruk A. Empowering patients through social media: the benefits and challenges. Health Informatics J. Mar 2014;20(1):50-58. [FREE Full text] [CrossRef] [Medline]60], offering the possibility of improving health outcomes [Moorhead SA, Hazlett DE, Harrison L, Carroll JK, Irwin A, Hoving C. A new dimension of health care: systematic review of the uses, benefits, and limitations of social media for health communication. J Med Internet Res. Apr 23, 2013;15(4):e85. [FREE Full text] [CrossRef] [Medline]5]. Health care organizations lack resources to create video content on the wide range of symptoms, diseases, and their progression that are treated by clinicians on a daily basis; offer easily understandable advice that can be integrated into patients’ self-care routines; or provide advice on topics that are outside the physician-patient interaction in a clinical setting. Improving clinicians’, patients’, and the public’s access to usable health information via curated and trusted video recommendations by domain experts can elevate population health literacy, empower patients, and build societal resilience.
As we demonstrate in this study, identifying, curating, and recommending relevant video materials leveraging the vast corpora of publicly available user-generated content is a feasible way to deliver personalized and contextualized health information for patient education. The adaptability of the content found on social media has enabled a variety of applications that were hitherto unthinkable. Well-designed user-generated content videos, in tandem with evidence from rigorous field experiments, could serve as part of a holistic system of care encompassing disease prevention and lifestyle changes along with resources for emotional support, better patient-physician interactions, and providing current and scientifically valid medical information to patients. The approach taken in this study, while evaluated with diabetes videos, has the potential for broader applicability across various health domains. The methods and principles developed in this research could be adapted to other chronic, acute, and infectious health conditions, such as cardiovascular disease, hypertension, pneumonia, and influenza, and broader patient education contexts, such as medication adherence and patient safety. Furthermore, the same approach can also be applied to curate videos for upskilling clinical professionals, such as surgical residents and nursing staff.
Currently, digital technologies for public health literacy and patient education are limited, lack scalability, and do not fully use the vast amount of publicly available health information found online and on social media platforms. Providing a strong open platform will provide a credible alternative to the vested interests of private organizations with proprietary technologies, which will lead to future innovations in novel data collection devices, digital platforms, and technologies in the context of health literacy initiatives.
Our methodology to develop a patient educational video system for understandability by integrating human efforts, that is, the perspectives of clinical practitioners and health care consumers, with ML algorithms is an innovative approach to a societally challenging problem. Patient empowerment and engagement are essential for appropriate disease management. For health organizations that are producing patient educational materials, our approach could be used as an educational tool for enhancing understandability in patient educational video content design. When designing educational materials, the insights from feature analysis of the ML algorithms have the potential to provide best practice guidelines regarding how organizations should engage health consumers with educational videos for varying levels of health literacy. Understandability can be further improved with the use of visual aids, summaries, and tangible tools such as personalized charts. Our study could add to patient communication and education literature and practice by enabling clinical practitioners to identify the most understandable, medically informative, and engaging videos for their patients as digital therapy. By combining algorithmic approaches with impact evaluation, this approach seeks to identify effective intervention methods that enable platform designers and clinicians to retrieve the most appropriate videos as digital therapeutic tools. Our approach can be extended to incorporate recent efforts by digital platforms and reputed national and international health organizations to identify authoritative sources of health information on social media channels and amplify credible content [Burstin H, Curry S, Ranney ML, Arora V, Wachler BB, Chou WY, et al. Identifying credible sources of health information in social media: phase 2-considerations for non-accredited nonprofit organizations, for-profit entities, and individual sources. NAM Perspect. May 23, 2023;2023:10.31478/202305b. [FREE Full text] [CrossRef] [Medline]23].
Limitations and Future Directions
This study has some limitations. Our study is built on the PEMAT guidelines developed by AHRQ. Although it is the most prevalent evaluation tool on patient education materials, PEMAT is not designed for user-generated content but for materials produced by health care organizations. The PEMAT criteria may require adaptation or extension to YouTube videos in evaluating subcriteria, such as whether the materials used for illustration were uncluttered, the technical quality of the video was satisfactory, and so on. In future work, we would like to explore alternative assessment tools or develop one that is more suitable for user-generated videos.
We also relied heavily on the evaluation of patient education materials from 4 physician evaluators, which poses the risk of evaluator bias. The calculated κ score indicates that there was variability in the reviewers’ use of the tools. However, we minimized this limitation by using the adjudication process for each item with a discrepancy, which is the accepted method to achieve consensus scores [O'Connor K, Sarker A, Perrone J, Gonzalez Hernandez G. Promoting reproducible research for characterizing nonmedical use of medications through data annotation: description of a Twitter corpus and guidelines. J Med Internet Res. Feb 26, 2020;22(2):e15861. [FREE Full text] [CrossRef] [Medline]61]. Additional video features that focus on esthetics, production qualities, whether the video contains a human, and so on were not used in this study due to our restricted definition of video understandability following the PEMAT guidelines. In addition to patient educational guidelines, it may also be necessary to examine factors such as concordance, which is the similarity or shared identity between physicians and patients based on a demographic attribute, such as race, gender, ethnicity, or age [Street RLJ, O'Malley KJ, Cooper LA, Haidet P. Understanding concordance in patient-physician relationships: personal and ethnic dimensions of shared identity. Ann Fam Med. 2008;6(3):198-205. [FREE Full text] [CrossRef] [Medline]62]. While understandability of content is an important criterion, other requirements such as accuracy of content, inclusivity and representativeness of content and its narrators, credibility and trustworthiness of the sources producing the videos, and others are equally critical. Finally, the challenges associated with the logistics (such as when, how, who will, and who to) of the delivery of the video recommendations have to be investigated using rigorous implementation science theories and frameworks. Future work will address these issues to potentially improve the reach and educational value of the recommended videos.
Conclusions
This study makes two contributions to the literature in the multidisciplinary area of digital therapeutics for health literacy and human-algorithm collaboration. The first contribution is the development of a human-in-the-loop augmented intelligence method that incorporates human judgment and expertise into an ML-driven computational approach that characterizes the search for cognitively demanding information on social media by combining human cognitive capabilities and AI systems. Our approach uses inputs from domain experts and PEMAT-based patient education constructs combined with ML and NLP methods to design and implement an automated tool that analyzes the understandability of YouTube videos from the perspective of patient education. The second contribution is to enable a better understanding of how patients assimilate health care information by assessing the impact of video understandability on viewer engagement with the videos. Our proposed solution can also provide health organizations with actionable guidance in designing and creating patient educational videos. Our findings can offer policy implications for the design and deployment of such digital artifacts for health literacy improvement, use of health care resources, and the quality of delivered care.
Future work can build on our approach to create a method of automated video retrieval that will accommodate viewers’ varying levels of both health and functional literacy and engagement. Future work can also build on the methods developed in this paper to develop multicriteria recommendations for a range of video content on topics such as education, climate change, financial literacy, and virtual communities based on metadata and video features from large social media platforms, such as YouTube.
Acknowledgments
The authors extend their gratitude to participants at various conferences and symposiums, including the 2024 Institute for Operations Research and the Management Sciences (INFORMS) Seattle Health Applications Society Distinguished Lecture, the 2024 Pittsburgh AI and Healthcare Symposium, and the 2024 Special Interest Group on Health Information Technology (SIG-Health) sponsored Health Information Technology Symposium, affiliated with the International Conference on Information Systems (ICIS 2024).
Additional thanks go to attendees of the 2023 Operational Research Society of India (ORSI)/International Conference on Business Analytics and Intelligence (ICBAI) Conference keynote at the Indian Institute of Science, Bangalore, the 2023 United Nations Generative AI Workshop, Medinfo 2023, the 2019 Artificial Intelligence for Data Discovery and Reuse (AIDR) Workshop, the 2019 American Medical Informatics Association (AMIA) Symposium, the 2022 Conference on Machine Learning, Optimization, and Data Science (LOD), the 2022 INFORMS Data Science Workshop, the 2023 INFORMS Healthcare Conference, the 2022 IEEE International Conference on Healthcare Informatics (ICHI) keynote, the 2022 AMIA Knowledge Discovery and Data Mining (KDDM) webinar, the Virtual Digital Economy(VIDE) Seminar Series, the Neural Information Processing Systems (NeurIPS) Machine Learning for Public Health (MLPH) Workshop, and the Statistical Challenges in Electronic Commerce Research (SCECR) Conferences (2020, 2022, and 2023) for their valuable feedback on earlier drafts of this paper.
The authors also appreciate seminar participants from numerous institutions, including the University of Bologna, University of Chicago, Harvard Medical School, Mayo Clinic-Jacksonville, Weill-Cornell Medicine, University of Warwick, University of Sussex, Cardiff University, Johns Hopkins University, Boston University, McGill University, Massachusetts Institute of Technology (MIT), Michigan State University (MSU) Outreach, University of Maryland, New York University (NYU), Temple University, University of Minnesota, Texas A&M University, University of Illinois at Chicago, Indian Institute of Technology (IIT) Madras, and Indian Institute of Management (IIM) Lucknow, for their insightful comments.
Special thanks are given to E Bioh, S Nair, M Nakhate, M Bowman, and N Navge for their research assistance. The authors also gratefully acknowledge research support from the National Institutes of Health (NIH)/National Library of Medicine (NLM) (R01LM013443).
Data Availability
The datasets generated or analyzed during this study are available from the corresponding author on reasonable request.
Authors' Contributions
XL contributed to conceptualization, data curation, formal analysis, funding acquisition, investigation, methodology, software, visualization, validation, writing—original draft, and writing—review and editing. AS contributed to conceptualization, funding acquisition, investigation, methodology, project administration, resources, supervision, validation, writing—review and editing. RP contributed to conceptualization, funding acquisition, investigation, methodology, project administration, resources, supervision, validation, writing—review and editing.
Conflicts of Interest
None declared.
Multimedia Appendix 2
Video attributes collected using YouTube data application programming interface.
DOCX File , 49 KBMultimedia Appendix 5
Descriptive statistics of features for video understandability classification.
DOCX File , 110 KBMultimedia Appendix 6
Performance metrics and hyperparameters of the proposed approach.
DOCX File , 25 KBReferences
- Alemayehu G, Chang P, Delaney M, Dietscher C, Dowse R, Khorshid O, et al. Health literacy around the world: policy approaches to wellbeing through knowledge and empowerment. The Economist Intelligence Unit. 2021. URL: https://impact.economist.com/perspectives/sites/default/files/lon_-_es_-_health_literacy_paper_v8_0.pdf [accessed 2025-03-17]
- Finney Rutten LJ, Blake KD, Greenberg-Worisek AJ, Allen SV, Moser RP, Hesse BW. Online health information seeking among US adults: measuring progress toward a healthy people 2020 objective. Public Health Rep. 2019;134(6):617-625. [FREE Full text] [CrossRef] [Medline]
- Kutner M, Greenburg E, Jin Y, Paulsen C. The health literacy of America’s adults; results from the 2003 National Assessment of Adult Literacy. National Center for Education Statistics. Sep 2006. URL: https://nces.ed.gov/pubs2006/2006483.pdf [accessed 2025-03-14]
- Mishra V, Dexter JP. Comparison of readability of official public health information about COVID-19 on websites of international agencies and the governments of 15 countries. JAMA Netw Open. Aug 03, 2020;3(8):e2018033. [FREE Full text] [CrossRef] [Medline]
- Moorhead SA, Hazlett DE, Harrison L, Carroll JK, Irwin A, Hoving C. A new dimension of health care: systematic review of the uses, benefits, and limitations of social media for health communication. J Med Internet Res. Apr 23, 2013;15(4):e85. [FREE Full text] [CrossRef] [Medline]
- Agarwal R, Bjarnadottir M, Rhue L, Dugas M, Crowley K, Clark J, et al. Addressing algorithmic bias and the perpetuation of health inequities: an AI bias aware framework. Health Policy Technol. Mar 2023;12(1):100702. [CrossRef]
- Lee HY, Jin SW, Henning-Smith C, Lee J, Lee J. Role of health literacy in health-related information-seeking behavior online: cross-sectional study. J Med Internet Res. Jan 27, 2021;23(1):e14088. [FREE Full text] [CrossRef] [Medline]
- Chou WY, Gaysynsky A, Trivedi N, Vanderpool RC. Using social media for health: national data from HINTS 2019. J Health Commun. Mar 04, 2021;26(3):184-193. [CrossRef] [Medline]
- Shoemaker SJ, Wolf MS, Brach C. Development of the Patient Education Materials Assessment Tool (PEMAT): a new measure of understandability and actionability for print and audiovisual patient information. Patient Educ Couns. Sep 2014;96(3):395-403. [FREE Full text] [CrossRef] [Medline]
- National action plan to improve health literacy. U.S. Department of Health and Human Services, Office of Disease Prevention and Health Promotion. 2010. URL: https://odphp.health.gov/sites/default/files/2019-09/Health_Literacy_Action_Plan.pdf [accessed 2025-03-14]
- Vishnevetsky J, Walters CB, Tan KS. Interrater reliability of the Patient Education Materials Assessment Tool (PEMAT). Patient Educ Counsel. Mar 2018;101(3):490-496. [CrossRef] [Medline]
- Kang SJ, Lee MS. Assessing of the audiovisual patient educational materials on diabetes care with PEMAT. Public Health Nurs. May 2019;36(3):379-387. [CrossRef] [Medline]
- McClure E, Ng J, Vitzthum K, Rudd R. A mismatch between patient education materials about sickle cell disease and the literacy level of their intended audience. Prev Chronic Dis. May 12, 2016;13:E64. [FREE Full text] [CrossRef] [Medline]
- Johnson R, Edwards R, Rivers A, Patil C, Walsh S. Evaluating literacy levels of patient education materials for a sickle cell transition group education programme. Health Educ J. Sep 23, 2019;79(3):253-265. [CrossRef]
- Rooney MK, Golden DW, Byun J, Lukas RV, Sonabend AM, Lesniak MS, et al. Evaluation of patient education materials for stereotactic radiosurgery from high-performing neurosurgery hospitals and professional societies. Neurooncol Pract. Jan 2020;7(1):59-67. [FREE Full text] [CrossRef] [Medline]
- Williams AM, Muir KW, Rosdahl JA. Readability of patient education materials in ophthalmology: a single-institution study and systematic review. BMC Ophthalmol. Aug 03, 2016;16:133. [FREE Full text] [CrossRef] [Medline]
- Kunze KN, Krivicich LM, Verma NN, Chahla J. Quality of online video resources concerning patient education for the meniscus: a YouTube-based quality-control study. Arthroscopy. Jan 2020;36(1):233-238. [CrossRef] [Medline]
- Sanderson SC, Suckiel SA, Zweig M, Bottinger EP, Jabs EW, Richardson LD. Development and preliminary evaluation of an online educational video about whole-genome sequencing for research participants, patients, and the general public. Genet Med. May 2016;18(5):501-512. [FREE Full text] [CrossRef] [Medline]
- Salama A, Panoch J, Bandali E, Carroll A, Wiehe S, Downs S, et al. Consulting "Dr. YouTube": an objective evaluation of hypospadias videos on a popular video-sharing website. J Pediatr Urol. Feb 2020;16(1):70.e1-70.e9. [FREE Full text] [CrossRef] [Medline]
- Desai T, Shariff A, Dhingra V, Minhas D, Eure M, Kats M. Is content really king? An objective analysis of the public's response to medical videos on YouTube. PLoS One. Dec 18, 2013;8(12):e82469. [FREE Full text] [CrossRef] [Medline]
- Sørensen K, Van den Broucke S, Fullam J, Doyle G, Pelikan J, Slonska Z, et al. Health literacy and public health: a systematic review and integration of definitions and models. BMC Public Health. Jan 25, 2012;12(1):80. [FREE Full text] [CrossRef] [Medline]
- Smith SK. Decision‐making and participation in bowel cancer screening: challenges and interventions for low health literacy. In: Proceedings of the 12th International Conference on Communication in Healthcare. 2014. Presented at: ICCH 2014; September 28-October 1, 2014; Amsterdam, The Netherlands.
- Burstin H, Curry S, Ranney ML, Arora V, Wachler BB, Chou WY, et al. Identifying credible sources of health information in social media: phase 2-considerations for non-accredited nonprofit organizations, for-profit entities, and individual sources. NAM Perspect. May 23, 2023;2023:10.31478/202305b. [FREE Full text] [CrossRef] [Medline]
- An R, Shen J, Xiao Y. Applications of artificial intelligence to obesity research: scoping review of methodologies. J Med Internet Res. Dec 07, 2022;24(12):e40589. [FREE Full text] [CrossRef] [Medline]
- Ben-Assuli O, Heart T, Yin N, Klempfner R, Padman R. On expert-machine partnership to predict mortality of congestive heart failure patients. Inf Syst Manag. Feb 09, 2024;41(4):413-431. [CrossRef]
- Ben-Assuli O, Heart T, Klempfner R, Padman R. Human-machine collaboration for feature selection and integration to improve congestive heart failure risk prediction. Decis Support Syst. Sep 2023;172:113982. [CrossRef]
- Wang S, Ning H, Huang X, Xiao Y, Zhang M, Yang EF, et al. Public surveillance of social media for suicide using advanced deep learning models in Japan: time series study from 2012 to 2022. J Med Internet Res. Jun 02, 2023;25:e47225. [FREE Full text] [CrossRef] [Medline]
- Bott N, Wexler S, Drury L, Pollak C, Wang V, Scher K, et al. A protocol-driven, bedside digital conversational agent to support nurse teams and mitigate risks of hospitalization in older adults: case control pre-post study. J Med Internet Res. Oct 17, 2019;21(10):e13440. [FREE Full text] [CrossRef] [Medline]
- Blum A, Mitchell T. Combining labeled and unlabeled data with co-training. In: Proceedings of the Eleventh Annual Conference on Computational Learning Theory. 1998. Presented at: COLT' 98; July 24-26, 1998; Madison, WI. [CrossRef]
- Platanios EA, Poon H, Mitchell TM, Horvitz E. Estimating accuracy from unlabeled data: a probabilistic logic approach. In: Proceedings of the 31st International Conference on Neural Information Processing Systems. 2017. Presented at: NIPS'17; December 4-9, 2017; Long Beach, CA.
- Zhang M, Tang J, Zhang X, Xue X. Addressing cold start in recommender systems: a semi-supervised co-training algorithm. In: Proceedings of the 37th International ACM SIGIR Conference on Research & Development in Information Retrieval. 2014. Presented at: SIGIR '14; July 6-11, 2014; Gold Coast, Australia. [CrossRef]
- Ma F, Meng D, Xie Q, Li Z, Dong X. Self-paced co-training. In: Proceedings of the 34th International Conference on Machine Learning. 2017. Presented at: ICML'17; August 6-11, 2017; Sydney, Australia.
- Pierce D, Cardie C. Limitations of co-training for natural language learning from large datasets. In: Proceedings of the 2001 Conference on Empirical Methods in Natural Language Processing. 2001. Presented at: EMNLP 2001; June 3-4, 2001; Pittsburgh, PA.
- Madathil KC, Rivera-Rodriguez AJ, Greenstein JS, Gramopadhye AK. Healthcare information on YouTube: a systematic review. Health Informatics J. Sep 2015;21(3):173-194. [FREE Full text] [CrossRef] [Medline]
- Guni A, Normahani P, Davies A, Jaffer U. Harnessing machine learning to personalize web-based health care content. J Med Internet Res. Oct 19, 2021;23(10):e25497. [FREE Full text] [CrossRef] [Medline]
- National diabetes statistics report 2020: estimates of diabetes and its burden in the United States. Centers for Disease Control and Prevention. 2020. URL: https://diabetesresearch.org/wp-content/uploads/2022/05/national-diabetes-statistics-report-2020.pdf [accessed 2023-12-06]
- Dal Canto E, Ceriello A, Rydén L, Ferrini M, Hansen TB, Schnell O, et al. Diabetes as a cardiovascular risk factor: an overview of global trends of macro and micro vascular complications. Eur J Prev Cardiol. Dec 2019;26(2_suppl):25-32. [CrossRef] [Medline]
- Deshpande AD, Harris-Hayes M, Schootman M. Epidemiology of diabetes and diabetes-related complications. Phys Ther. Nov 2008;88(11):1254-1264. [FREE Full text] [CrossRef] [Medline]
- Shrivastava SR, Shrivastava PS, Ramasamy J. Role of self-care in management of diabetes mellitus. J Diabetes Metab Disord. Mar 05, 2013;12(1):14. [FREE Full text] [CrossRef] [Medline]
- Donner A, Rotondi MA. Sample size requirements for interval estimation of the kappa statistic for interobserver agreement studies with a binary outcome and multiple raters. Int J Biostat. 2010;6(1):Article 31. [FREE Full text] [CrossRef] [Medline]
- Type 1 diabetes and viral infections. YouTube. URL: https://www.youtube.com/watch?v=4JLnkpdjoU8 [accessed 2025-03-15]
- Shoemaker SJ, Wolf WS, Brach C. The Patient Education Materials Assessment Tool (PEMAT) and user’s guide. Agency for Healthcare Research and Quality. URL: https://www.ahrq.gov/health-literacy/patient-education/pemat.html [accessed 2023-12-06]
- Bartko JJ. The intraclass correlation coefficient as a measure of reliability. Psychol Rep. Aug 01, 1966;19(1):3-11. [CrossRef] [Medline]
- Rasheed Z, Shah M. Scene detection in Hollywood movies and TV shows. In: Proceedings of the 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. 2003. Presented at: CVPR 2003; June 18-20, 2003; Madison, WI. [CrossRef]
- Xiong W, Lee JC. Efficient scene change detection and camera motion annotation for video classification. Comput Vis Image Underst. Aug 1998;71(2):166-181. [CrossRef]
- Shahraray B. Scene change detection and content-based sampling of video sequences. In: Proceedings of the SPIE Digtial Video Compression: Algorithms and Technologies 1995. 1995. Presented at: SPIE 1995; February 5-10, 1995; San Jose, CA. [CrossRef]
- Liao H, McDermott E, Senior A. Large scale deep neural network acoustic modeling with semi-supervised training data for YouTube video transcription. In: Proceedings of the 2013 IEEE Workshop on Automatic Speech Recognition and Understanding. 2013. Presented at: ASRU 2013; December 8-12, 2013; Olomouc, Czech Republic. [CrossRef]
- Kincaid JP, Fishburne RPJ, Rogers RL, Chissom BS. Derivation of new readability formulas (automated readability index, fog count and flesch reading ease formula) for navy enlisted personnel. Institute for Simulation and Training. 1975. URL: https://stars.library.ucf.edu/cgi/viewcontent.cgi?article=1055&context=istlibrary [accessed 2025-03-14]
- Basch CH, Mohlman J, Hillyer GC, Garcia P. Public health communication in time of crisis: readability of on-line COVID-19 information. Disaster Med Public Health Prep. Oct 11, 2020;14(5):635-637. [FREE Full text] [CrossRef] [Medline]
- Basch E, Deal AM, Dueck AC, Scher HI, Kris MG, Hudis C, et al. Overall survival results of a trial assessing patient-reported outcomes for symptom monitoring during routine cancer treatment. JAMA. Jul 11, 2017;318(2):197-198. [FREE Full text] [CrossRef] [Medline]
- Millar N, Budgell B, Fuller K. ‘Use the active voice whenever possible’: the impact of style guidelines in medical journals. Appl Linguist. Sep 2013;34(4):393-414. [CrossRef]
- Xie J, Liu X, Dajun Zeng D. Mining e-cigarette adverse events in social media using Bi-LSTM recurrent neural network with word embedding representation. J Am Med Inform Assoc. Jan 01, 2018;25(1):72-80. [FREE Full text] [CrossRef] [Medline]
- Fage-Butler AM, Nisbeth Jensen M. Medical terminology in online patient-patient communication: evidence of high health literacy? Health Expect. Jun 2016;19(3):643-653. [FREE Full text] [CrossRef] [Medline]
- Oliu WE, Brusaw CT, Alred GJ. Writing That Works: Communicating Effectively on the Job. Boston, MA. Bedford/St. Martin's; 2016.
- Summary/conclusion words. University of Waterloo. URL: https://uwaterloo.ca/writing-and-communication-centre/tran sition-words-and-phrases [accessed 2023-12-06]
- Using transitional words and phrases. University of Wisconsin-Madison. URL: https://writing.wisc.edu/handbook/transitions/ [accessed 2023-12-06]
- Liu X, Zhang B, Susarlia A, Padman R. Go to YouTube and call me in the morning: use of social media for chronic conditions. MIS Q. 2020;44(1b):257-283. [FREE Full text] [CrossRef]
- Du J, Ling CX, Zhou ZH. When does cotraining work in real data? IEEE Trans Knowl Data Eng. May 2011;23(5):788-799. [CrossRef]
- Manning CD, Raghavan P, Schutze H. An Introduction to Information Retrieval. Cambridge, UK. Cambridge University Press; 2009.
- Househ M, Borycki E, Kushniruk A. Empowering patients through social media: the benefits and challenges. Health Informatics J. Mar 2014;20(1):50-58. [FREE Full text] [CrossRef] [Medline]
- O'Connor K, Sarker A, Perrone J, Gonzalez Hernandez G. Promoting reproducible research for characterizing nonmedical use of medications through data annotation: description of a Twitter corpus and guidelines. J Med Internet Res. Feb 26, 2020;22(2):e15861. [FREE Full text] [CrossRef] [Medline]
- Street RLJ, O'Malley KJ, Cooper LA, Haidet P. Understanding concordance in patient-physician relationships: personal and ethnic dimensions of shared identity. Ann Fam Med. 2008;6(3):198-205. [FREE Full text] [CrossRef] [Medline]
Abbreviations
AHRQ: Agency for Healthcare Research and Quality |
AI: artificial intelligence |
API: application programming interface |
CDC: Centers for Disease Control and Prevention |
ML: machine learning |
N/A: not applicable |
NLP: natural language processing |
OCR: optical character recognition |
PEMAT: Patient Education Materials Assessment Tool |
Edited by T de Azevedo Cardoso; submitted 04.01.24; peer-reviewed by K Yousaf, M Machal; comments to author 14.06.24; revised version received 09.10.24; accepted 13.01.25; published 08.04.25.
Copyright©Xiao Liu, Anjana Susarla, Rema Padman. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 08.04.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.