Review
Abstract
Background: New media have become vital sources of cancer-related health information. However, concerns about the quality of that information persist.
Objective: This study aims to identify characteristics of studies considering cancer-related information on new media (including social media and artificial intelligence chatbots); analyze patterns in information quality across different platforms, cancer types, and evaluation tools; and synthesize the quality levels of the information.
Methods: We systematically searched PubMed, Web of Science, Scopus, and Medline databases for peer-reviewed studies published in English between 2014 and 2023. The validity of the included studies was assessed based on risk of bias, reporting quality, and ethical approval, using the Joanna Briggs Institute Critical Appraisal and the STROBE (Strengthening the Reporting of Observational Studies in Epidemiology) checklists. Features of platforms, cancer types, evaluation tools, and trends were summarized. Ordinal logistic regression was used to estimate the associations between the conclusion of quality assessments and study features. A random-effects meta-analysis of proportions was conducted to synthesize the overall levels of information quality and corresponding 95% CIs for each assessment indicator.
Results: A total of 75 studies were included, encompassing 297,519 posts related to 17 cancer types across 15 media platforms. Studies focusing on video-based media (odds ratio [OR] 0.02, 95% CI 0.01-0.12), rare cancers (OR 0.32, 95% CI 0.16-0.65), and combined cancer types (OR 0.04, 95% CI 0.01-0.14) were statistically less likely to yield higher quality conclusions compared to those on text-based media and common cancers. The pooled estimates reported moderate overall quality (DISCERN 43.58, 95% CI 37.80-49.35; Global Quality Score 49.91, 95% CI 43.31-56.50), moderate technical quality (Journal of American Medical Association Benchmark Criteria 46.13, 95% CI 38.87-53.39; Health on the Net Foundation Code of Conduct 49.68, 95% CI 19.68-79.68), moderate-high understandability (Patient Education Material Assessment Tool for Understandability 66.92, 95% CI 59.86-73.99), moderate-low actionability (Patient Education Materials Assessment Tool for Actionability 37.24, 95% CI 18.08-58.68; usefulness 48.86, 95% CI 26.24-71.48), and moderate-low completeness (34.22, 95% CI 27.96-40.48). Furthermore, 27.15% (95% CI 21.36-33.35) of posts contained misinformation, 21.15% (95% CI 8.96-36.50) contained harmful information, and 12.46% (95% CI 7.52-17.39) contained commercial bias. Publication bias was detected only in misinformation studies (Egger test: bias –5.67, 95% CI –9.63 to –1.71; P=.006), with high heterogeneity across most outcomes (I²>75%).
Conclusions: Meta-analysis results revealed that the overall quality of cancer-related information on social media and artificial intelligence chatbots was moderate, with relatively higher scores for understandability but lower scores for actionability and completeness. A notable proportion of content contained misleading, harmful, or commercially biased information, posing potential risks to users. To support informed decision-making in cancer care, it is essential to improve the quality of information delivered through these media platforms.
Trial Registration: PROSPERO CRD420251058032; https://www.crd.york.ac.uk/PROSPERO/view/CRD420251058032
doi:10.2196/73185
Keywords
Introduction
Background
Digital new media have become indispensable tools for health information and network support, notably among individuals experiencing chronic and terminal health conditions, such as many forms of cancer [,]. New media platforms, represented by social media and artificial intelligence (AI) chatbots, offer collaborative and participatory environments that serve as valuable resources for patients with cancers, survivors of cancer, and caregivers of patients [,]. These platforms facilitate the exchange of information [], provide opportunities to seek external support [], and share personal experiences of cancer []. The rise of AI chatbots has attracted considerable interest in their use for health information [], with studies indicating that they can provide higher-quality information and more empathetic responses than physicians []. Further, among the US adults who report being on the internet almost constantly, specifically on social media [], 72% report leveraging their preferred platform, and increasingly generative AI, to learn about health conditions diagnosed in themselves and their friends and family [].
Research shows that turning to web-based outlets can improve psychosocial outcomes across various acute and chronic conditions []. Access to high-quality information further enhances health literacy, decision-making, health behaviors, outcomes, and health care experiences [-]. Conversely, low-quality information can promote trust in unproven, ineffective, and harmful treatments, as well as fake news, conspiracy theories, and misleading testimonials about “natural” or alternative cures []. Kington et al [] described “high-quality information” as science-based and aligned with the best available evidence. Huang et al [] further identified 15 key attributes of information quality, including accuracy, objectivity, accessibility, completeness, and clarity.
As cancer remains a leading cause of death globally, with rising incidence rates [], related discussions are widespread on social media and AI chatbots [,], making the topic a target for commercial and political interests [,]. Substantial financial incentives drive the promotion of treatments [], and patients, often desperate for hope, are particularly susceptible to misleading information circulated in digital spaces []. Limitations in models, databases, and algorithms pose challenges to the accuracy and credibility of the information provided by AI chatbots [,]. Survivors also face cognitive challenges such as fatigue, memory deficits, and reduced executive function [,], which may impair their ability to critically evaluate the information quality they received []. Consequently, low-quality information may lure patients into submitting to costly and harmful treatments, undermining the potential benefits of digital support networks [].
To date, systematic, scoping, and other literature reviews generally report inconclusive and nondefinitive findings on the quality of cancer-related information on social media [,,]. Most cross-sectional studies indicate considerable variability in information quality across platforms, content formats (ie, text-based or video-based), and cancer types []. This inconsistency may stem from differences in cancer types studied, platforms analyzed, and quality assessment approaches used. A critical gap remains the breadth of literature on this topic that directly accounts for (1) a wide variety of cancer types, (2) a multitude of platforms, and (3) differences in findings over time. Social media, in particular, has evolved considerably in the past two decades, giving rise to AI chatbots that now represent a new source of health information. Accounting for factors such as users’ shift in medium preference (text-based vs video-based), and platform changes (Twitter to X, and at the time of writing, the potential closure of TikTok) may reveal temporal trends that implicate improvement or worsening of information quality over time.
Purpose
This paper presents a systematic review of empirical studies published between 2014 and 2023 that address cancer-related information on new media. It explores the growing reliance on these platforms for cancer health information, trends in platform use, and myriad tools used in the included studies to determine the information quality. Specifically, we addressed the following research questions (RQs): (RQ1) What are the key characteristics of studies evaluating the quality of cancer-related information on social media and AI chatbots, and how have these characteristics evolved? (RQ2) What factors influence the conclusion of quality assessments, and how do they vary across platforms, assessment tools, and cancer types? (RQ3) What patterns emerge in the assessment findings of new media cancer-related information quality?
Findings from this review contribute to further supporting our understanding of new media’s role as a source of cancer-related health information. By adopting a longitudinal perspective, the study identifies trends in quality assessment studies using data from social media and AI chatbots and offers evidence-based recommendations for future research on this vital topic.
Methods
Guidelines and Ethical Considerations
This review was reported according to the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines (). Our protocol was retrospectively registered on PROSPERO (CRD420251058032) on May 23, 2025.
As a review study of published studies, our research is considered nonhuman participant research and is exempt from review by an Institutional Review Board.
Search Strategy
Between January and September 2024, we conducted a systematic literature search on four electronic databases (PubMed, Web of Science, Scopus, and Medline) for studies. Full strategies for the databases are available in Table S1 in . We only included studies that underwent peer review and publication between 2014 and 2023 and presented empirical qualitative, quantitative, or computational findings in English. We also reviewed the reference lists of included studies to identify any additional publications that met the inclusion criteria but were not retrieved in the original search. Three researchers conducted independent examinations of the selection criteria, search terms, and the records returned by the search. Disagreements were resolved by discussion among the three researchers until a consensus was reached.
Selection Criteria
We initially included studies if they met three broad categories of inclusion: (1) they contained a synthesis of cancer-related new media data, (2) they used an information quality assessment to rate the quality of data along varying criteria, and (3) they were published in English between 2014 and 2023. Cancer types were defined according to the International Classification of Diseases for Oncology. Likewise, we considered three broad categories of new media for inclusion: (1) text and graph-based media, such as Facebook, Instagram, Pinterest, Reddit, Twitter (now X), WeChat, and Weibo; (2) video-based media such as YouTube, TikTok, and Xigua; and (3) generative AI-based media, such as ChatGPT, Chatsonic, Microsoft Bing AI, and Perplexity.
We excluded studies that did not present original findings, unpublished manuscripts from preprint repositories such as arXiv, conference abstracts, narratives, literature reviews, and editorials. Studies were also excluded if they discussed the quality of information but did not elaborate on how it was assessed, for example, without conducting a comprehensive and systematic data collection and analysis, without referring to any quality criteria.
Assessments of Risk of Bias, Quality, and Certainty
Two researchers independently assessed the quality and risk of bias of the included studies using the Joanna Briggs Institute Critical Appraisal Checklist for Analytical Cross-Sectional Studies [] and the STROBE (Strengthening the Reporting of Observational Studies in Epidemiology) statement checklist []. Studies evaluated by the Joanna Briggs Institute were considered to have a yes or no risk of bias or to have some concerns in study design, conduct, and analysis. The STROBE checklist assessed and illustrated the quality of reporting in key sections of studies. Additionally, we recorded whether each study reported an ethical approval process, including compliance with human and animal rights standards, informed consent procedures, and data privacy protections.
Two researchers independently rated the certainty of evidence of outcome using GRADE (Grading of Recommendations Assessment, Development and Evaluation) via the GRADEpro Guideline Development Tool online platform, including assessment domains of risk of bias, inconsistency, indirectness, imprecision, and publication bias.
Disagreements between the two raters were discussed until a consensus was reached.
Data Extraction
We extracted the following data from the included studies: (1) authors, (2) year of publication, (3) journal name, (4) study design (ie, quantitative, qualitative, or computational), (5) sample size, (6) sample genders, (7) study location, (8) authors’ identities (individual or constitute, nonmedical professional or medical professional), (7) social media platform, (8) search strategies, (9) cancer types, (10) the language of the original data, (11) the profession of the raters, (12) contents of the posts, and (13) study results.
We extracted the information assessment tools used in the included studies, such as the DISCERN [] tool, along with the outcomes measured by each. Despite the variation in tools, most studies assessed at least one of the following: (1) overall quality: a holistic assessment of the posts; (2) technical criteria: an assessment of the layout and functionality of the data; (3) readability and understandability: an assessment of reading and comprehension level; (4) accuracy and misinformation: an evaluation of scientific validity of the content; (5) completeness and coverage: the scope of topics; (6) actionability and usefulness: an assessment of the practical nature of the content; (7) harmfulness: whether the content can cause or promote harm, and (8) commercial bias: evidence of the direct influence of commercial interests. We categorized the 75 studies based on the tools and metrics used to evaluate the quality of information and draw comparisons across studies. Data were extracted by one researcher and independently verified by a second researcher. All analyses were performed using Stata/SE (version 18.0; StataCorp LLC) and RStudio (version 4.3; Posit PBC).
Analysis
Descriptive Analysis
First, we illustrated several granular details for all studies about the search process, rating process, and review process. Second, we documented the frequencies and percentages of all mentioned media platforms, cancer types, and assessment tools for information quality, as well as changes in those patterns over the study period.
Statistical Analysis
For a single study reporting information quality for stratified groups, such as different posters (eg, individual or institute or with or without medical professional), content types (eg, prevention or treatment), or other specified categories, we calculated the pooled mean and pooled standard deviation. For a study that reported a median instead of a mean for a quality indicator, we estimated the mean and SD from the median [].
We standardized quality indicators across different scales by mapping them onto a proportional rate system. For example, a score of 2 in a 1-5 scale system would be standardized as 40% in the proportional rating system. Then those quality scores are classified as follows: “low quality” for scores within 0%-30%, “median quality” for scores within 30%-70%, and “high quality” for scores within 70%-100%. For negative indicators (misinformation, harmfulness, and commercial bias), the coding system was the opposite. To categorize the conclusion of each study based on quality indicators used, we adopted the following criteria: The overall conclusion of information in a study was deemed “positive” if it reported more “high-quality” than “low-quality” results. It was considered “negative” if it reported more “low quality” results than “high quality” results. Studies reporting exclusively “medium quality” results, or equal numbers of “high” and “low quality” results, were classified as “neutral.”
We then used ordinal logistic regression to estimate the associations between study conclusions and media type and cancer type, as well as characteristics of studies based on the search, rating, and report processes []. Regression models were weighted by the natural log of each study’s sample size [], since studies on some kinds of platforms, such as Twitter and Reddit, included larger sample sizes than others, such as YouTube. This analysis reported odds ratios (ORs), where estimates greater than 1 indicate a positive association between the predictor variables and higher quality conclusions.
Meta-Analysis
We used a meta-analysis of proportions to test the homogeneity of results obtained using the same quality assessment tool. Only results reported by three or more studies using the same tool were included in the analysis. Statistical heterogeneity was assessed via I2 and Cochran Q test values, where an I2 value of 25% represents low heterogeneity, 50% represents moderate heterogeneity, and 75% represents high heterogeneity. Analysis of proportions was pooled with a random-effect model with DerSimonian-Laird intervals when the test of heterogeneity was moderate or high []. For studies reporting indicators with 0% or 100% estimates, we used the Freeman-Tukey double arcsine transformation to stabilize the variance []. We calculated 95% CIs with an α level of .05 to estimate statistical significance.
Assessment of publication bias was done via an Egger test and a visual inspection of funnel plot asymmetry, respectively, for each indicator. Because publication bias was suspected, a sensitivity analysis was conducted using leave-one-out analysis and evaluate the influence of each study on the overall estimate.
Results
Overall Characteristics of Included Studies
This review summarizes the body of literature on the quality of cancer-related information on new media, published from 2014 to 2023. Our initial search identified 7841 studies, including (1) 3001 records from PubMed, (2) 2032 records from Web of Science, (3) 1293 records from Scopus, and (4) 1515 records from Medline. After screening papers by title and abstract, 140 potentially eligible papers remained for full-text review. Through full-text review, we excluded 65 additional studies because they lacked analytic rigor or did not include quality of information as a component of results. Our final sample size comprised 75 individual studies. depicts our filtering process according to the PRISMA guidelines.

We present our results below in relation to each of our RQs.
(RQ1) What are the key characteristics of studies evaluating the quality of cancer-related information on social media and AI chatbots, and how have these characteristics evolved?
summarizes characteristics of included studies, including binary classifications of whether certain features were present or not. More than 80% of studies followed similar practices in the search process, including documenting their search period (70/75, 93%), search tool (69/75, 92%), query terms (73/75, 97%), and only considering one search tool (62/75, 83%), single platform (67/75, 89%). A total of 20 studies assessed content in languages other than English, including Chinese (n=6), German (n=5), Arabic (n=3), Japanese (n=2), French (n=2), and Spanish (n=2), and several other languages were examined in single studies, such as Korean, Italian, Turkish, Swedish, and Danish. During the rating process, 73 (97%) studies used a nonblinded rating process, 66 (88%) studies documented the number of raters, and 69 (92%) studies relied on more than one rater. However, across studies that reported content distributor information, 14 (19%) studies defined their rating criteria a priori, and 42 (56%) studies included a rater with a medical background. In the reporting process, more than 60% of studies reported “views,” “like,” or “unlike,” and “comment” behaviors, while a few tracked posts sharing behavior (11/75, 15%). About 70% of studies reported the posters’ identity, but few provided information on their demographics (17/75, 23%). Among the included studies, 67 (89%) reported the contents or topics discussed in the posts.
| Study characteristics | Yes, n (%) | No, n (%) | Other, n (%) | |
| Search process | ||||
| Date or period mentioned | 70 (93) | 5 (7) | N/Aa | |
| Search tools mentioned | 69 (92) | 6 (8) | N/A | |
| More than one search tool used | 13 (17) | 62 (83) | N/A | |
| Search terms mentioned | 73 (97) | 2 (3) | N/A | |
| Initial hits reported | 41 (55) | 34 (45) | N/A | |
| Assessed language other than English | 20 (27) | 55 (73) | N/A | |
| More than one social media platform examined | 8 (11) | 67 (89) | N/A | |
| Cancer type more than one | 30 (40) | 45 (60) | N/A | |
| Rating process | ||||
| Raters blinded for the source | 2 (3) | 73 (97) | N/A | |
| Number of raters reported | 66 (88) | 9 (12) | N/A | |
| More than one rater | 69 (92) | 6 (8) | N/A | |
| Rater independently | 38 (51) | 37 (49) | N/A | |
| Interrater reliability figures for evaluation determined | 34 (45) | 41 (55) | N/A | |
| Process graph contained | 34 (45) | 41 (55) | N/A | |
| Medical professional background for rater | 42 (56) | 17 (23) | 16 (21)b | |
| A priori criteria defined for quality | 14 (19) | 52 (69) | 9 (12)c | |
| Reporting process | ||||
| Engagement: view | 46 (61) | 29 (39) | N/A | |
| Engagement: like or unlike | 56 (75) | 19 (25) | N/A | |
| Engagement: forward or share | 11 (15) | 64 (85) | N/A | |
| Engagement: comment | 47 (63) | 28 (37) | N/A | |
| Poster characteristics reported (gender, or age, or ethnicity, or country) | 17 (23) | 58 (77) | N/A | |
| Poster identity reported (personal, or institute, or medical professional) | 53 (71) | 22 (29) | N/A | |
| Contents or topics mentioned | 67 (89) | 8 (11) | N/A | |
aNot applicable.
bRaters are authors of the study, and it was unclear if they were medical professionals.
cGenerally mentioned using literature or publication as criteria.
Results of Risk of Bias, Quality, and Certainty Assessments
A majority of studies demonstrated strong methodological quality, with 74 (98%) studies appropriately describing subjects and setting, 74 (98%) using standard condition measurement, all using valid outcome measurements, and using appropriate statistical analyses. However, we identified a considerable risk of bias in the handling of confounding factors; only 43 (57%) studies identified potential confounders and implemented strategies to address them (A). Most studies reported key elements, while 22 (29%) studies did not include other analyses (eg, analysis of subgroup and interactions and sensitivity analysis), and 23 (31%) studies did not list funding sources (B). Of the included studies, 34 (45%) reported ethical approval consideration, whereas 41 (55%) did not mention ethical approval. Among those that did report ethical approval, most stated that it was not required as the research only involved publicly available data, excluded human or animal participants, and removed identifiable information (eg, user IDs, links, and contact details), as is common with social media research. Only 3 studies obtained formal approval from research ethics committees [-]. Tables S2-S152 in [,-] present details of the risk of bias and quality evaluation for individual studies.

Funnel plot and Egger test revealed significant asymmetry (bias coefficient –5.67, 95% CI –9.63 to –1.71; P=.006) only for studies evaluating misinformation, suggesting that studies could be missing in the literature that reported null or negative findings of misinformation or small study effects. No bias was detected for other indicators (Figures S1-S11 in ).
Since all included studies were observational in design, they were initially rated as low certainty according to the GRADE framework. Subsequently, the certainty of evidence for all evaluated outcomes was downgraded to very low due to one or more of the following reasons: high inconsistency (I²>75% in most outcomes), risk of indirectness (due to heterogeneous platforms or measurement definitions and tools), and, in some cases, imprecision (wide CIs and small sample sizes). These limitations significantly reduce confidence in the pooled estimates based on estimates. Details for each assessment and the reason for downgrading are in Table S153 in .
Individual Study Characteristics
Media Platforms
A illustrates the frequency of considered media platforms and changes in included platforms over time. A total of 20 studies investigated content from various text and graph platforms, including Twitter (n=11) [,,-], Facebook (n=6) [,-,,], Pinterest (n=5) [-,,], Reddit (n=4) [-,], and Instagram (n=2) [,], and 1 study each on WeChat [] and Weibo []. A total of 51 records studied video-based media; of these, 44 examined YouTube [,-], 6 focused on TikTok [-], and 1 study each on Bilibili [] and Xigua []. Overall, 4 records studied data from generative AI-chatbot media, all of which specifically included ChatGPT [-]; 2 studies also included Perplexity, Chatsonic, and Bing AI [,].
We observed changes in the studied platforms over time. From 2014 to 2016, the number of included social media platforms was limited, with only 1 study investigating content from YouTube and 2 on Twitter. The scope expanded in 2018 and 2019, with increased research on other text- and graphic-based media, including 3 Facebook studies, 1 Pinterest study, 1 Reddit study, and 1 Weibo study. From 2020 to 2022, the number of studies increased, with 26 YouTube studies, while each of the other text- and graphic-based media platforms maintained approximately 3-4 publications. In 2023, this trend persisted, with video-based media, especially YouTube, continuing to be the primary focus of cancer information quality research, as evidenced by 14 studies. This period also marked the emergence of research on generative AI chatbot media, which was the subject of 4 studies.

Cancer Types
We captured 17 unique types of cancer within our data with varying levels of emphasis (B). There were 15 research studied information about prostate cancer [,,,,,-,,,,,,,]; 13 about breast cancer; 9 about skin cancer [,,,,,,,,]; 9 about colorectal cancer [,,,,,,,,]; 7 about bladder cancer [,,,,,,]; 5 about testicular cancer [,,,,]; 4 each about lung cancer [,,,], thyroid cancer [,,,], and kidney cancer [,,,]; 2 each about liver cancer [,], leukemia [,], and brain cancer [,]; and 1 each about cervical [], gastric [], larynx [], pancreatic [], and spine [] cancers. Another 15 studies were about cancer information in general or did not specify the cancer type [,,,,,,,,,,,,,,]. Several studies were not limited to a single type of cancer but examined multiple types. Therefore, those studies were counted more than once.
Quality Assessment Tools
C highlights the presence of formal and informal information quality tools used to evaluate data in a given study. We observed that a majority of studies used DISCERN (44/75, 59%) or evaluated misinformation (44/75, 59%), followed by Global Quality Score (GQS; n=26) and Patient Education Materials Assessment Tool (PEMAT; n=16). Tool use evolved over time. In 2014-2019, misinformation was the most common instrument, used in 10 studies. From 2020 to 2022, both misinformation and DISCERN saw increased use, each featuring in 24 studies. This trend persisted in 2023, with DISCERN remaining prominent in 20 studies published that year.
Content Distributors
Across 54 studies that reported content distributor information, medical professionals, such as surgeons, physicians, endocrinologists, sonographers, radiologists, and dietitians, were frequently identified as key contributors. Video-based studies consistently reported that the most-watched or -engaged videos were published by such medical professionals. Notably, they produced 80% of top TikTok videos on thyroid cancer [,] and 80% of leading YouTube posts on spine tumors [] and hepatocellular carcinoma []. They also contributed over half of the top YouTube posts on rectal cancer surgery [], nutrition [], the mental health of patients with prostate cancer [], radioactive iodine therapy [], and larynx cancer [], as well as on liver cancer [] and gastric cancer [] in Chinese on TikTok.
Institutional contributors, such as hospitals, clinics, academic centers, and universities, also played a role. For instance, they produced 92% of the top YouTube videos on pediatric cancer clinical trials [] and 54% on Merkel cell carcinoma [], though their contributions were lower across other topics.
In contrast, two studies of TikTok revealed that nonmedical individuals, including patients and their families or friends, accounted for over 93% of top videos on gastric cancer in English and Japanese [] and 83% of prostate cancer videos []. Nonmedical contributors, such as media agencies, for-profit companies, herbalists, health websites, or lifestyle vloggers, contributed less than 61.1% of all topics.
Engagement Metrics
Studies analyzing video-based platforms frequently reported engagement metrics such as views, likes or dislikes, shares, and comments. Among 51 such studies, view counts ranged as low as 3 [] to as high as 11 million [], the latter for a thyroid cancer video posted by a medical professional on TikTok. YouTube videos on breast cancer and leukemia also reached high viewership, each exceeding 7 million views [,,]. In contrast, videos on prostate cancer (TikTok) and pediatric cancer clinical trials (YouTube) averaged around 2000 views [,].
Likes and comments were lower than view counts, as has been noted previously. The most-liked video was a TikTok post on thyroid cancer, which garnered 308,000 likes [,], followed by a breast cancer video on YouTube with approximately 226,000 likes []. The most “disliked” video addressed herbal cancer treatments in Arabic, receiving an average of 994 dislikes on YouTube []. Comment activity varied widely. Thyroid cancer videos on TikTok averaged 1252 comments, with the most-commented video amassing over 73,000 comments [,]. Nutrition and lung cancer videos each had over 450 comments [,].
Of nonvideo, text-based entries, Facebook demonstrated the highest overall engagement, particularly in posts related to common cancers, including breast, prostate, colorectal, and lung cancers, as well as dermatological and genitourinary cancers [-].
Factors Impacting Quality Conclusion
(RQ2) What factors influence the conclusion of quality assessments, and how do they vary across platforms, assessment tools, and cancer types?
The ordinal logistic regression analysis identified several factors associated with reporting higher quality conclusions (). Video-based media (OR 0.02, 95% CI 0.01-0.12), studies on rare cancers (OR 0.32, 95% CI 0.16-0.65), and studies on combined cancer types (OR 0.04, 95% CI 0.01-0.14) were less likely to yield high-quality conclusions than text-based media and studies on common cancers.
During the search process, studies using multiple search tools (OR 0.30, 95% CI 0.13-0.73), mentioning search terms (OR 0.13, 95% CI 0.02-0.81), reporting initial hits (OR 0.14, 95% CI 0.07-0.28), sourcing content in languages other than English (OR 0.35, 95% CI 0.16-0.76), and analyzing multiple media platforms (OR 0.06, 95% CI 0.02-0.27) were less likely to report higher quality conclusions.
During the rating process, studies disclosing the number of raters (OR 0.02, 95% CI 0.00-0.14) were less likely to report high-quality conclusions. In contrast, studies including process graphs (OR 3.06, 95% CI 1.61-5.79) and using literature-based assessments (OR 2.93, 95% CI 1.04-8.20) were more likely to report higher quality conclusions.
During the report process, studies reporting engagement metrics such as likes or dislikes (OR 3.35, 95% CI 1.20-9.38), forwards or shares (OR 6.17, 95% CI 1.61-23.65), and mentioning content or topics (OR 3.77, 95% CI 1.43-9.94), were associated with higher odds of reporting high-quality conclusions.
| Characteristics of the study | ORa (95% CI) | P value | |
| Media type (text-based media as reference) | |||
| Video-based | 0.02 (0.01-0.12) | <.001 | |
| AIb-based | 0.70 (0.10-5.01) | .72 | |
| Cancer typec (common cancer as reference) | |||
| Rare cancer | 0.32 (0.16-0.65) | .002 | |
| Combined cancer | 0.04 (0.01-0.14) | <.001 | |
| Search process | |||
| Date or period mentioned | 1.75 (0.32-9.47) | .52 | |
| Search tools mentioned | 2.01 (0.46-8.89) | .36 | |
| More than one search tool used | 0.30 (0.13-0.73) | .008 | |
| Search terms mentioned | 0.13 (0.02-0.81) | .03 | |
| Initial hits reported | 0.14 (0.07-0.28) | <.001 | |
| Assessed language other than English | 0.35 (0.16-0.76) | .009 | |
| Sites on more than one social media platform | 0.06 (0.02-0.27) | <.001 | |
| Cancer type more than one | 1.41 (0.55-3.63) | .47 | |
| Rating process | |||
| Raters blinded for the source | 11.16 (0.60-207.53) | .11 | |
| Number of raters reported | 0.02 (0.00-0.14) | <.001 | |
| More than one rater | 5.21 (0.76-35.75) | .09 | |
| Rater working independently | 0.88 (0.46-1.67) | .69 | |
| Interrater reliability figures for evaluation determined | 1.22 (0.62-2.43) | .56 | |
| Process graph contained | 3.06 (1.61-5.79) | <.001 | |
| Medical professional background for rater (no professional as reference) | |||
| Author as raters | 0.43 (0.14-1.30) | .14 | |
| Professional | 0.52 (0.17-1.60) | .26 | |
| A priori criteria defined for quality (no criteria mentioned as reference) | |||
| Based on literature | 2.93 (1.04-8.20) | .04 | |
| Specific criteria mentioned | 0.94 (0.41-2.14) | .88 | |
| Reporting process | |||
| Engagement: view | 1.64 (0.40-6.70) | .49 | |
| Engagement: like or dislike | 3.35 (1.20-9.38) | .02 | |
| Engagement: forward or share | 6.17 (1.61-23.65) | .008 | |
| Engagement: comment | 1.97 (0.91-4.29) | .09 | |
| Poster characteristics reported (gender, or age, or ethnicity, or country) | 1.51 (0.63-3.64) | .36 | |
| Poster identity reported (personal, or institute, or medical professional) | 1.81 (0.77-4.28) | .18 | |
| Contents or topics mentioned | 3.77 (1.43-9.94) | .007 | |
aOR: odds ratio.
bAI: artificial intelligence.
cUsing GLOBOCAN 2020 statistics, we coded the top 10 most common cancers as “common cancer,” and others as “rare cancer” [].
Information Quality Criteria
(RQ3) What patterns emerge in the assessment findings of new media cancer-related information quality?
Overall Quality
Overall quality refers to a holistic assessment of content, considering its alignment with current scientific standards and whether content achieves educational aims. The most commonly used tools were DISCERN (44/75, 59%) and GQS (26/75, 35%). DISCERN is a 16-item tool evaluating publication reliability, quality of information on treatment choices, and a singular overall quality rating. GQS uses a 5-point Likert scale from 1=poor to 5=excellent.
Overall, content provided by medical individuals and institutions, such as hospitals, physicians, and dietitians, received higher DISCERN scores than that from nonprofessional sources [,,,,,,,,,,,-]. Among these medical groups, hospitals provided higher-quality information than health organizations [], and doctors specializing in modern medicine consistently scored higher than those in traditional medicine []. However, an exception was noted in one study on bladder cancer information on YouTube, where content from medical professionals scored lower []. Interestingly, TikTok videos by news agencies sometimes outperformed medical providers in quality, attributed to the absence of confusing jargon []. Comparisons between for-profit and nonprofit sources yielded mixed results: some studies reported higher DISCERN scores for for-profit sources [,], whereas others, particularly on colorectal cancer on YouTube, found no significant differences [,].
Studies using the DISCERN tool identified varying scores along different criteria. A total of 8 studies reported the highest scores for “explicit aims” [,,,,,,,], 6 for “aims achieved” [,,,,,], and 4 for “benefits of treatments” [,,,]. The most common reason for score deductions was the lack of “additional sources of information,” reported in 7 studies [,,,,,,]. In total, 4 studies identified the lowest scores for failing to “describe what would happen if any treatment is not used” [,,,], and 3 studies noted deficiencies in “providing information source” [,,]. Additionally, 2 studies each noted the lowest scores for reporting “currency of information” [,], “reference to areas of uncertainty” [,], “risks of treatment” [,], and “quality of life” [,].
DISCERN-based assessments also revealed regional and linguistic variations in content quality. A study of gastric cancer TikTok videos found that Chinese-language content was of higher quality than English and Japanese videos []. However, no significant quality differences were observed across prostate and thyroid cancer videos in English, French, German, Italian, and Turkish on YouTube [,]. Geographic comparisons of English-language videos showed that content from the United States consistently ranked higher in quality than that from other locations [,].
Three studies comparing AI-chatbot media on a range of platforms found that they generally provided moderate to high-quality information [,,]. These chatbots frequently cited reputable sources such as the American Cancer Society and the Mayo Clinic.
Studies using GQS also found that quality was typically higher for videos produced by medical professionals [,,,,,,] and for-profit medical providers [,]. For instance, YouTube content posted by medical providers on pediatric cancer clinical trials [], liver cancer [], and skin cancer [] received GQS ratings of 4 or higher, indicating good quality. While YouTube videos on Merkel cell carcinoma [], breast cancer originating from Australia [], and breast cancer videos uploaded by medical advertisers on Xigua [] received GQS scores below 2, indicating poor quality.
Technical Quality
Technical quality is the evaluation of disclosure ethics. The commonly used Journal of the American Medical Association Benchmark Criteria (JAMA-BC) critically assesses web content based on four key “transparency criteria”: authorship, attribution, disclosure, and currency. This tool, which has a high score of 4.0, was applied in five studies of YouTube content. The highest JAMA-BC score, 2.6, was reported for spine tumor videos [], while the lowest score, 1.0, was for nutrition videos posted by independent users [], indicating minimal adherence to reliability standards. Across included studies, content was rated high for “authorship” when clear information about contributors and their credentials was provided; the “disclosure” and “currency” indicators were rated lowest, reflecting a lack of transparency regarding sponsorship, commercial funding, potential conflicts of interest, and dates of posted and updated information [,,].
Transparency of ethics for content was measured using other tools. Three studies used the Health on the Net Foundation Code of Conduct [], which uses eight constructs: (1) authority, (2) complementarity, (3) privacy, (4) attribution, (5) justification, (6) contact details, (7) financial disclosure, and (8) advertising policy. Studies found that most cancer content videos disclosed authority but few disclosed source information, conflicts of interest, financial sources, or advertisement policy [,,]. The Quality Evaluation Scoring Tool, used in one study, measures six aspects of web-based health information: (1) authorship, (2) attribution, (3) conflicts of interest, (4) currency, (5) complementarity, and (6) tone []. The study using the Quality Evaluation Scoring Tool examined TikTok videos on gastric cancer and found that Chinese-language videos scored higher than Japanese- and English-language videos []. Additionally, the Audiovisual Quality Score, which assesses the viewability, precision, and editing of audiovisual materials, revealed that larynx cancer videos from university sources showed clearer and more professional editing [].
Readability and Understandability
Readability and understandability are metrics used to determine how effectively audiences can process information. PEMAT-U is the first part of the PEMAT toolkit to assess the understandability of print and audiovisual materials, which consists of 13 questions measuring the understandability of content’s language, organization, and visual design []. A total of 18 studies adapted PEMAT-U scores for digital content, with a majority reporting high understandability (above 70%). In general, higher scores were attributed to content with a clear purpose and use of accessible language [,]; and lower scores were attributed to content that lacked summaries or educational visual aids to help people understand the content [,,,,]. The highest PEMAT-U score, 88%, was awarded to research on imaging information about prostate cancer on Instagram [], followed by thyroid cancer videos from TikTok []. The lowest PEMAT-U scores (below 30%) were found in Arabic-language YouTube videos on herbal cancer treatments [] and immunotherapy for renal cell cancer and prostate cancers []. Two studies using PEMAT-U found that content generated by AI chatbots often included medical jargon and concise terminology, making it difficult for lay audiences to understand [,].
Three studies measuring readability of content used the Flesch-Kincaid scale, which determines the average reading level needed to comprehend a written document on a continuum from 5 (signifying a fifth-grade reading level) to 16 (indicating a postgraduate reading level). One of these studies examined prostate cancer information on YouTube, reporting a 12th-grade readability level overall []. The other two focused on AI chatbots’ responses to cancer-related inquiries and found that the content was at a college-level readability [,].
Accuracy and Misinformation
Accuracy refers to the extent to which information aligns with established scientific or medical evidence. Terms such as misinformation, misleading content, false claims, and nonevidence-based claims are sometimes used to describe a lack of accuracy. Most studies did not use a standardized tool to evaluate accuracy. Rather, misinformation was typically assessed by identifying the proportion of content that deviated from scientific standards. Some studies applied predefined criteria, while others relied on expert reviewers to assess and classify content with subject experts.
One study analyzing the top 10 most-viewed YouTube videos on tetrahydrocannabinol oil and skin cancer concluded that all contained misinformation []. Similarly, a study of YouTube videos on prostate cancer flagged 76.25% as containing misinformation, with radiotherapy videos demonstrating less misinformation than surgery videos []. Over half of the videos on prostate cancer in Arabic [], and in English on skin cancer [], breast cancer [], immunotherapy for urological tumors [], and postsurgical exercise for breast cancer [] were also identified to contain misinformation.
The misuse or discrediting of health services was the most common type of misinformation flagged by 17 studies, encompassing inappropriate use or dismissal of treatments [,,,,,,,], vaccines [,,,], screenings [,,,,], and diagnostic tests [,,,]. For example, 25% of Facebook posts on acute lymphoblastic leukemia included disapproved treatment protocols and health services [], and 16.5% of Pinterest pins undermined the accuracy and safety of mammograms, advocating instead for alternatives such as ultrasound or thermography, and spreading false claims about bioidentical hormones and breast tumors []. One study found that 12.5% of AI chatbot-generated cancer treatment responses contained hallucinated therapies, such as immunotherapy, which was not clinically recommended []. Additionally, 42% of Arabic-language videos on breast cancer called for inadequate screening and treatment protocols [].
Another common type of misinformation mentioned in 15 studies involved unproven prevention and treatment modalities. For example, a study on Facebook pages about acute lymphoblastic leukemia found that all references to alternative and complementary therapies were related to unproven treatment modalities []. Similarly, 74.2% of alternative medicine content in dermatology-related posts was found to be misleading []. On Twitter, mentions of “alternative treatments” are often linked to external sources, such as hyperlinks, books, videos, or movies, without assessing the credibility of those materials []. Cannabis was frequently portrayed on Facebook and Twitter as an alternative cancer treatment, with 43.8% of such posts relying on anecdotal patient stories and over half using invalid scientific reasoning to support these claims [,,]. One TikTok study found that about 5% of videos featured supernatural or heroic powers as potential cancer treatments []. Other content touting laetrile and colloidal silver on Twitter, despite their potential toxicity and lack of cancer-fighting benefits [,], as well as spiritual healing [], acupuncture, chiropractic care, yoga [], and escharotic black salve [] as cancer treatments on social media.
Eight studies reported misinformation that overstated the effectiveness of certain foods and supplements in preventing or curing cancer. Information on social media falsely claimed some diet [,,,], various supplements [,,] (eg, flaxseed, turmeric, IGF-1, vitamin D, slippery elm, probiotics, and coconut oil), specific fruits [,] (eg, pomegranates and mushroom), herbs [,,] (eg, dandelions, curcumin, slippery elm, blood root, and gumby gumby), drinks [,,] (eg, green tea, miracle beverage, and apple cider vinegar) as “natural remedies,” “cancer pills,” or “cures” for cancers.
Scope and Completeness
Scope refers to the range of topics covered. illustrates the topics identified across the included studies. The most frequently covered topic was treatment (50 studies), encompassing content related to surgery, medication, technique, side effects, and alternative therapies. Background knowledge was the focus of 36 studies and included cancer definitions, pathology, etiology, anatomy, epidemiology, and prognosis. Prevention, mentioned in 35 studies, focused on strategies to reduce cancer risk, such as screening, lifestyle changes, dietary modifications, vaccination, and raising awareness. Diagnosis, which was addressed in 30 studies, refers to identifying cancer through symptoms, tests, staging, and clinical manifestations. Personal experiences and others were the focus of 25 studies, featuring patient stories, psychological support, relationships, news coverage, and emotional responses such as fear, anxiety, and depression.

Completeness refers to how thoroughly health-related content is presented on new media platforms. Nine studies assessed completeness using predefined criteria within each study. One YouTube study used the Sahin critical appraisal tool, which consists of 12 questions covering primary, secondary, and tertiary prevention levels for pancreatic cancer []. Another study used Hexagonal Radar Charts to illustrate content balance and found TikTok videos on genitourinary cancer content adequately covered symptoms and examinations but lacked information on definitions and outcomes []. Several studies suggested the need to include adverse outcomes to ensure completeness. An evaluation of ChatGPT responses applied an informed consent measurement [] found frequent omissions regarding treatments, risks, complications, quality-of-life impacts, and consequences of forgoing treatment for urological cancer []. Similarly, a study of YouTube videos on surgical treatments for spine tumors reported a greater emphasis on benefits than on complications or posttreatment sequelae, potentially biasing patients’ perceptions [].
A study of completeness based on creators found that academic YouTube channels provide more complete information in their videos about colorectal cancer screening than other publisher types []. Completeness also varied by language. In a study of gastric cancer videos on TikTok, Chinese-language content from educational and health professionals was more comprehensive, while English-language videos by individual creators were more complete than their Chinese or Japanese counterparts [].
Actionability and Usefulness
Actionability measures how well information enables individuals to take informed action, and usefulness indicates the extent to which the information benefits personal decision-making processes. The PEMAT-Actionability (PEMAT-A) tool, part of the PEMAT assessment, includes four strategies to evaluate this aspect, with higher scores indicating increased actionability []. PEMAT-A was used in 16 studies, 11 of which reported poor actionability (scores below 50%). Common reasons for low scores included missing figure interpretations and unclear or overly complex instructions [,]. There was some variation by cancer type. YouTube videos on urological and breast cancer [,] and TikTok videos on prostate cancer [] found no actionable content. In contrast, a study of YouTube videos on testicular cancer yielded 100% actionability []. Further, prostate cancer videos on YouTube reported 75% [], outperforming their TikTok counterparts, citing little or no actionability [].
Further, 14 studies assessed the usefulness of new media information, with 4 studies reporting that over 80% of videos on hepatocellular carcinoma [], Merkel cell carcinoma [], Arabic-language prostate cancer [], and Chinese-language gastric cancer [] were deemed ‘useful’. In contrast, only 20.51% of English and 17.46% of Japanese TikTok videos on gastric cancer were rated as ‘useful’ [], with similarly low rates (around 20%) for English and Japanese videos on skin, larynx, and gastric cancers [,,]. Usefulness also varied by content source. Among YouTube videos on rectal cancer surgery and colorectal cancer screening, approximately 15% of videos by nonprofit posters and only 5.39% of videos by for-profit posters were rated as ‘useful’, according to the criteria [,].
Harmfulness
Studies evaluating the harmfulness of media content usually consider four harm-related constructs: (1) harmful actions, (2) harmful inaction, (3) harmful interactions, and (4) economic harm. Five studies used an “informative harm” assessment, calculating the ratio of harmful to positive messages. Harmful inaction was the most commonly identified issue. In one study, it accounted for 73% of harmful content in Japanese tweets []. Another study found that 31% of harmful messages on widely shared social media studies promoted rejecting conventional cancer treatments in favor of unproven alternatives []. The same study reported that economic harm (eg, out-of-pocket costs for unproven treatments or travel) comprised 27.7% of harmful content, while harmful actions (eg, suggesting potentially toxic tests or treatments) accounted for 17% []. In a study on YouTube videos on basal cell carcinoma, all harmful messages originated from laypersons [].
Commercial Bias
Eight studies used commercial bias as a quality criterion by measuring the proportion of content originating from commercial, for-profit, or agenda-driven sources. On Pinterest, commercial bias was most prevalent in prostate cancer posts (14%), followed by bladder (7%) and kidney cancer (1%) []. On video-based platforms, commercial bias ranged from 10% to 27.33% [,,,]. Two studies specifically reported commercial bias in 13.2% and 17% of popular YouTube videos on bladder cancer [,].
Pooled Quality Results
shows the pooled estimated scores for indicators for different quality criteria measurements. The pooled estimates reported moderate overall quality (DISCERN 43.58, 95% CI 37.80-49.35; GQS 49.91, 95% CI 43.31-56.50; I2=93.22%), and moderate technical quality (JAMA-BC 46.13, 95% CI 38.87-53.39; I2=75.29%; Health on the Net Foundation Code of Conduct 49.68, 95% CI 19.68-79.68; I2=94.35%). About 67% of the information was understandable to viewers (PEMAT-U 66.92, 95% CI 59.86-73.99; I2=88.69%). Approximately 27% of posts contained misinformation (27.15, 95% CI 21.36-33.35; I2=99.77%), 34% of essential topics defined by individual study were covered (completeness 34.22, 95% CI 27.96-40.48; I2=77.08%), less than half of the posts provided actionable message and were considered useful (PEMAT-A 37.24, 95% CI 18.08-58.68; I2= 98.53%; usefulness 48.86, 95% CI 26.24-71.48; I2=99.38%). Furthermore, 21% of posts contained harmful messages (harmfulness 21.15, 95% CI 8.96-36.50; I2=88.73%), and 12.46% of posts showed evidence of commercial bias (commercial bias 12.46, 95% CI 7.5-17.39; I2=94.82%). The forest plots for each indicator are in Figures S12-S22 in [,-].
Results of leave-one-out sensitivity analysis were stable across all iterations, with no single study significantly altering the pooled effect. Details of the sensitivity analysis for each indicator are in Figures S23-S33 in [,-].
| Criteria | Indicator | Study count | Sample size | Pooled quality, % (95% CI) | P value | I2 (%) |
| Overall quality | DISCERNa | 44 | 3701 | 43.58 (37.80-49.35) | <.001 | 93.22 |
| Overall quality | GQSb,c | 26 | 1440 | 49.91 (43.31-56.50) | <.001 | 90.14 |
| Technical quality | JAMA-BCd | 12 | 746 | 46.13 (38.87-53.39) | <.001 | 75.29 |
| Technical quality | HONcodee | 3 | 155 | 49.68 (19.68-79.68) | <.001 | 94.35 |
| Understandability | PEMAT-Uf,g | 16 | 1453 | 66.92 (59.86-73.99) | <.001 | 88.69 |
| Accuracy | Misinformationb | 44 | 345,169 | 27.15 (21.36-33.35) | <.001 | 99.77 |
| Completeness | Completenessa | 8 | 1036 | 34.22 (27.96-40.48) | <.001 | 77.08 |
| Actionability | PEMAT-Ag,h | 16 | 1116 | 37.24 (18.08-58.68) | <.001 | 98.53 |
| Actionability | Usefulness | 13 | 1427 | 48.86 (26.24-71.48) | <.001 | 99.38 |
| Harmfulness | Harmfulnessb | 5 | 395 | 21.15 (8.96-36.50) | <.001 | 88.73 |
| Commercial bias | Commercial bias | 11 | 4432 | 12.46 (7.52-17.39) | <.001 | 94.82 |
aThree papers were removed from the meta-analysis because they did not provide the exact score for the indicator.
bContained papers with indicators reporting 0 or 100%.
cGQS: Global Quality Score.
dJAMA-BC: Journal of the American Medical Association Benchmark Criteria.
eHONcode: Health on the Net Foundation Code of Conduct.
fPEMAT-U: Patient Education Material Assessment Tool for Understandability.
gOne paper was removed from the meta-analysis because it provided a combined score for PEMAT-U and PEMAT-A. The quality of information in individual studies in a proportional rating system is shown in Table S154.
hPEMAT-A: Patient Education Material Assessment Tool for Actionability.
Discussion
Principal Results
This systematic review and meta-analysis offer robust insight into new media-driven information quality studies pertaining to cancer. Our findings showcase a complex and diverse body of literature varying across multiple domains, including time, platform, quality assessment types, and findings therein. Several important patterns merited further discussion.
First, evidence of temporal changes highlights the evolution of media platforms over time. The studies evidenced clear changes in used platforms, reflecting their availability and popularity at the time. For example, early in the study period, we principally identified studies analyzing YouTube videos. Prior to 2021, there was also greater use of text-based media, including Twitter, Reddit, Pinterest, Facebook, and others. However, between 2021 and 2023, most studies shifted solely to video-based platforms. The shift in emphasis from text and image to video-based social media largely aligns with US patterns in that a majority of the user-generated social media content created daily since 2021 consists of videos [,]. This period also coincides with changes to data access policies on platforms such as Twitter (now X), which closed its application programming interface in 2022, limiting data collection efforts for research []. Another important point is that the second and third most popular social media platforms in the United States (Facebook and Instagram, respectively) [] represented a small proportion of the total number of studies reviewed (6 studies on Facebook and 2 studies on Instagram). This represents a critical gap that warrants further investigation. The other temporal pattern is that the first cancer-related information quality assessment of generative AI chatbot content occurred in 2023 []. In the coming years, more studies are likely to expand in this area. Future research should compare the quality of information of multiple AI platforms and models—especially in the areas of response consistency and alignment with scientific best practices.
Given the proliferation of video-based social media over time, it is also critical to contextualize information quality patterns by cancer types. Our findings generally highlight that video-based cancer content is engaged but at notable risk for low-quality information. Indeed, our analysis presents high engagement (eg, views, likes, and comments) for video-based platforms, which may be due to their general accessibility, visual appeal, or targeted algorithmic marketing. This finding might suggest that cancer-related content on video-based platforms may have greater reach and visibility compared with text-based content, and empirical evidence supports this [,]. However, our analysis shows that information quality on video platforms is significantly lower than on text-based platforms (OR 0.02, 95% CI 0.01-0.12; P<.001), underscoring challenges in content moderation and infoveillance. Similarly, newly emerging AI chatbots exhibit potential to deliver cancer-related information of higher quality, as evidenced by higher DISCERN scores than video-based media [,,]. However, limitations remain, particularly concerning understandability, actionability, and occasional misinformation [-], which may be driven by hallucinations. Future AI-development efforts should focus on ensuring more comprehensive and representative information, while improving the clarity and factual accuracy of AI-generated responses.
Second, this study highlights shifts in research priorities and interests, as well as changes in cancer types and the quality of information associated with them. Between 2014 and 2018, increased attention was given to certain cancer types, primarily breast, prostate, skin, and colorectal cancers. These cancers are all highly prevalent in the United States, and disparities in health care seeking, treatment, and outcomes affect each [], which may explain their broad emphasis during this period. Studies of these cancers remained prevalent throughout the data collection period. However, between 2018 and 2023, research interest expanded to include rarer and less media-publicized cancers such as gastric, thyroid, and spinal cancers. This shift might reflect increasing interest in underrepresented cancers and the need to address gaps in communication and information quality across a wider range of cancer types; however, more research here is needed.
Our findings likewise demonstrated that information quality varied across the studied cancer types. Specifically, research focusing on rare cancers was of significantly lower quality compared to studies on common, highly prevalent cancers (OR 0.32, 95% CI 0.16-0.65). Indeed, studies of more common cancers such as skin, lung, breast, colorectal, prostate, liver, and thyroid consistently achieved the highest overall quality, understandability, completeness, and usefulness scores [,,,,,,,,,,]. This conclusion may be explained by the relatively late initiation and limited volume of research focused on the quality of information regarding rare cancers.
Third, addressing the quality of cancer-related information on media platforms presents unique challenges and opportunities. Prior research attributes lower information quality on new media to several factors, including algorithmic amplification of popular content regardless of the source [,], brevity of text and short-form limiting content depth [], and increased content from nonexpert creators with no moderation []. The literature highlights various recommendations for addressing these issues, including improving architecture for detecting and eliminating incorrect content [], increasing mobile health (mHealth) and digital health media campaigns to promote health literacy [], encouraging more medical entities to have a social media presence [], and auditing risks and functionalities for AI-generated materials []. The success of these strategies is mixed, suggesting further work is needed to address the confounding issue of the reach and visibility of cancer information relative to information quality risk. Regardless, with the evident growth of health content on new media, timely attention to these issues remains a public health priority.
Future Studies
Findings from this review offer four tangible recommendations for future research. First, while it is clear that new media information quality assessments are growing and popular fields, there is limited consensus on which information quality assessment is needed relative to the task. Based on our findings, we strongly encourage future researchers conducting such assessments to consider robust and validated tools such as DISCERN, PEMAT U/A, GQS, and the JAMA-BC criteria. Although application of these tools can be time-intensive, each provides an evaluation of health content in several domains simultaneously, ensuring thorough analysis of the data. Further, consistently using these tools for information quality assessments will allow for robust and streamlined meta-analyses, which would allow for even greater insights than cross-sectional designs alone. However, we also acknowledge that even when these tools are consistently applied, variations in study objectives, media platforms, and evolving content may still pose challenges in achieving standardized assessments across different research contexts.
Second, when conducting mHealth or digital health interventions, we advise researchers to launch them through video-based apps, such as YouTube and TikTok. Comparing studies consistently revealed greater engagement with data from these sources than text-based sources, regardless of cancer type or account attributes. This directly suggests that the broad access and ease-of-use of these videos may make health-related content more readily accessible to others, regardless of barriers to care. Prior research strongly recommends using innovative strategies such as narrative storytelling or personal testimonials to boost engagement. Regardless of the strategies, it is clear from our findings that the shift away from text-based media is not a trend but part of the shifting social media landscape. At the same time, researchers should also remain abreast of new media trends and adapt accordingly.
Third, mHealth or digital health interventions should be produced by or in partnership with specialists, clinicians, and hospitals. Studies considering content posted by professional entities consistently had better information quality across inventories, cancers, and platforms. Although in some cases, content by medical professionals was not as well-engaged as content from other nonmedical accounts, these cases may reflect content posted on text-based platforms where engagement was consistently lower. Specialists with direct expertise in the topic as video protagonists—including celebrities with lived cancer experience— may help extend the credibility of the content and transcend trust barriers for certain topics, including vaccinations and routine screenings.
Fourth, our results are only starting to show the emergence of AI as a health promotion tool. Professionals in the health and medical fields should continue researching generative AI content for information quality.
Fifth, our results equally highlight the need for licensed care providers to consider the impact of new media on patient health-related decision-making. To our knowledge, few providers, including nurse practitioners, physicians, or insurance representatives, currently inquire about how much new media patients consume during consultations. However, given that our findings indicate a wide array of information quality, coupled with profuse and growing use of new media in the United States and abroad, it may be imperative for providers to inquire about use in controlled settings.
Limitations
First, while we believe we offered a thorough review of the new media cancer information literature, we were not able to consider other facets that may lead to varying information quality, such as the professional level of raters, content language, nation of origin, and different criteria in the searching and rating process. As this review focused only on English-language-based studies, similar reviews focusing on studies in other languages (such as Spanish, Arabic, and Chinese) and media data (such as LinkedIn, Snapchat, and other chatbots) are needed. Second, the inconsistent nomenclature of certain information quality and measurement used may be driving high levels of heterogeneity in our meta-analysis. For example, the measure of “actionability” was sometimes called “usefulness,” depending on the study. This inconsistent nomenclature made it difficult to categorize study findings, which may have resulted in a marginally biased result. Further, even in cases where studies used the same quality assessment, the diversity of platforms, cancer types, and scope of work may have also contributed to higher levels of heterogeneity. Subgroup analyses to explore the sources of heterogeneity were limited by the small number of records included in those subgroups and the poor quality of reporting in the available data. As such, pooled findings should be interpreted as indicative rather than definitive. Third, this study is minimally representative of generative AI as an information source. Given its growing popularity and strong body of literature considering generative AI as a health communication tool, we strongly recommend further systematic reviews specifically including information quality assessments using multiple generative AI tools. Meanwhile, we found a limited number of published studies specifically addressing certain platforms, such as Facebook and Instagram, during our search period. Finally, this review was prone to publication bias among studies on misinformation and other risks of bias.
Conclusions
High-quality cancer information on new media contributed to informed cancer-related decision-making. However, the quality of this information remained at a moderate to moderate-low level, particularly in terms of overall quality, transparency, understandability, accuracy, actionability, harmfulness, and commercial bias. Our findings indicated that poster identity, medium (text-based, video-based, and AI media), and assessments of highly prevalent cancers were associated with higher information quality. Gaps in information quality, which were evident across all studies considered in our review, remain a critical and timely public health issue.
Acknowledgments
We thank Kate Epstein for her editorial services with the language editing and proofreading of this manuscript. We likewise acknowledge that GPT-4o was exclusively used in an editorial capacity to improve the readability of sentences upon revision. All sections of the manuscript were further reviewed and revised by the research team.
Data Availability
The datasets generated or analyzed during this study are available from the corresponding author on reasonable request.
Authors' Contributions
XJL contributed to conceptualization, data curation, formal analysis, methodology, software, validation, visualization, and writing. DV contributed to conceptualization, formal analysis, investigation, methodology, and writing. MAP contributed to formal analysis and methodology. AM contributed to validation and analysis. ERWB contributed to conceptualization, investigation, and writing.
Conflicts of Interest
None declared.
PRISMA 2020 checklist.
PDF File (Adobe PDF File), 136 KBResearch strategy, quality assessment results, risk of bias assessment, quality evaluation, ethical approval process report, certainty of evidence, forest plot, sensitivity analysis, funnel plot, and Egger regression test.
PDF File (Adobe PDF File), 5762 KBReferences
- Chou W-YS, Hunt Y, Folkers A, Augustson E. Cancer survivorship in the age of YouTube and social media: a narrative analysis. J Med Internet Res. Jan 17, 2011;13(1):e7. [FREE Full text] [CrossRef] [Medline]
- Johnson AR, Longfellow GA, Lee CN, Ormseth B, Skolnick GB, Politi MC, et al. Social media as a platform for cancer care decision-making among women: internet survey-based study on trust, engagement, and preferences. JMIR Cancer. Mar 05, 2025;11:e64724. [FREE Full text] [CrossRef] [Medline]
- Brey P, Søraker JH. Philosophy of computing information technology. In: Meijers A, editor. Philosophy of Technology and Engineering Sciences. North-Holland. Handbook of the Philosophy of Science; 2009:1341-1407.
- Hopkins AM, Logan JM, Kichenadasse G, Sorich MJ. Artificial intelligence chatbots will revolutionize how cancer patients access information: ChatGPT represents a paradigm-shift. JNCI Cancer Spectr. 2023;7(2):pkad010. [FREE Full text] [CrossRef] [Medline]
- Gage-Bouchard EA, LaValley S, Warunek M, Beaupin LK, Mollica M. Is cancer information exchanged on social media scientifically accurate? J Cancer Educ. 2018;33(6):1328-1332. [FREE Full text] [CrossRef] [Medline]
- Oh S, Syn SY. Motivations for sharing information and social support in social media: a comparative analysis of acebook, witter, elicious, ou ube, and lickr. Assoc Inf Sci Tech. Apr 21, 2015;66(10):2045-2060. [CrossRef]
- Lazard AJ, Meernik C, Collins MKR, Vereen RN, Benedict C, Valle CG, et al. Social media use for cancer support among young adults with cancer. J Adolesc Young Adult Oncol. 2023;12(5):674-684. [CrossRef] [Medline]
- Li J, Dada A, Puladi B, Kleesiek J, Egger J. ChatGPT in healthcare: a taxonomy and systematic review. Comput Methods Programs Biomed. 2024;245:108013. [CrossRef] [Medline]
- Ayers JW, Poliak A, Dredze M, Leas EC, Zhu Z, Kelley JB, et al. Comparing physician and artificial intelligence chatbot responses to patient questions posted to a public social media forum. JAMA Intern Med. 2023;183(6):589-596. [FREE Full text] [CrossRef] [Medline]
- Atske S, Perrin A. About three-in-ten U.S. adults say they are 'almost constantly' online. Pew Research Center. 2021. URL: https://www.pewresearch.org/short-reads/2021/03/26/about-three-in-ten-u-s-adults-say-they-are-almost-constantly-online/ [accessed 2025-01-23]
- Purcell K, Fox S. Social media and health. Pew Research Center. 2010. URL: https://www.pewresearch.org/internet/2010/03/24/social-media-and-health/ [accessed 2025-01-23]
- Ye Z, Li W, Zhu R. Online psychosocial interventions for improving mental health in people during the COVID-19 pandemic: a systematic review and meta-analysis. J Affect Disord. 2022;316:120-131. [FREE Full text] [CrossRef] [Medline]
- Sarasohn-Kahn J. A role for patients: the argument for self-care. Am J Prev Med. 2013;44(1 Suppl 1):S16-S18. [FREE Full text] [CrossRef] [Medline]
- Rutten LJF, Agunwamba AA, Wilson P, Chawla N, Vieux S, Blanch-Hartigan D, et al. Cancer-related information seeking among cancer survivors: trends over a decade (2003-2013). J Cancer Educ. 2016;31(2):348-357. [CrossRef] [Medline]
- Hill SJ, Sofra TA. How could health information be improved? Recommended actions from the Victorian Consultation on Health Literacy. Aust Health Rev. 2018;42(2):134-139. [CrossRef] [Medline]
- Crafoord M, Ekstrand J, Sundberg K, Nilsson MI, Fjell M, Langius-Eklöf A. Mobile electronic patient-reported outcomes and interactive support during breast and prostate cancer treatment: health economic evaluation from two randomized controlled trials. JMIR Cancer. 2025;11:e53539. [FREE Full text] [CrossRef] [Medline]
- Agrawal L, DaSouza RO, Mulgund P, Chaudhary P. The role of online support, caregiving, and gender in preventative cancer genetic testing participation: cross-sectional study from a national study. JMIR Cancer. 2025;11:e67650. [FREE Full text] [CrossRef] [Medline]
- Gorski DH. Cancer quackery and fake news: targeting the most vulnerable. In: Bernicker EH, editor. Cancer and Society: A Multidisciplinary Assessment and Strategies for Action. Cham, Switzerland. Springer International Publishing; 2019.
- Kington RS, Arnesen S, Chou WS, Curry SJ, Lazer D, Villarruel AM. Identifying credible sources of health information in social media: principles and attributes. NAM Perspect. 2021;2021:10.31478/202107a. [FREE Full text] [CrossRef] [Medline]
- Huang KT, Lee YW, Wang RY. Quality Information and Knowledge. Jersey, United States. Prentice Hall; 1998.
- Cancer key facts. World Health Organisation. 2024. URL: https://www.who.int/news-room/fact-sheets/detail/cancer [accessed 2025-08-13]
- Wang Y, McKee M, Torbica A, Stuckler D. Systematic literature review on the spread of health-related misinformation on social media. Soc Sci Med. 2019;240:112552. [FREE Full text] [CrossRef] [Medline]
- Xu L, Sanders L, Li K, Chow JCL. Chatbot for health care and oncology applications using artificial intelligence and machine learning: systematic review. JMIR Cancer. 2021;7(4):e27850. [FREE Full text] [CrossRef] [Medline]
- Sissung TM, Figg WD. Facebook groups for alternative treatments for cancer: advertising masquerading as community support. Lancet Oncol. 2021;22(1):25-26. [CrossRef] [Medline]
- Ognyanova K, Lazer D, Robertson RE, Wilson C. Misinformation in action: fake news exposure is linked to lower trust in media, higher trust in government when your side is in power. HKS Misinfo Rev. 2020;1(4). [FREE Full text]
- Tascilar M, de Jong FA, Verweij J, Mathijssen RHJ. Complementary and alternative medicine during cancer treatment: beyond innocence. Oncologist. 2006;11(7):732-741. [FREE Full text] [CrossRef] [Medline]
- Volpe NJ, Mirza RG. Chatbots, artificial intelligence, and the future of scientific reporting. JAMA Ophthalmol. 2023;141(9):824-825. [CrossRef] [Medline]
- Naik D, Naik I, Naik N. Imperfectly perfect AI chatbots: limitations of generative AI, large language models and large multimodal models. London, UK. Springer Nature Switzerland; 2024. Presented at: The International Conference on Computing, Communication, Cybersecurity and AI; July 3-4, 2024:43-66; London, UK. [CrossRef]
- Vardy J, Dhillon HM, Pond GR, Rourke SB, Xu W, Dodd A, et al. Cognitive function and fatigue after diagnosis of colorectal cancer. Ann Oncol. 2014;25(12):2404-2412. [FREE Full text] [CrossRef] [Medline]
- Ahles TA, Saykin AJ, McDonald BC, Li Y, Furstenberg CT, Hanscom BS, et al. Longitudinal assessment of cognitive changes associated with adjuvant treatment for breast cancer: impact of age and cognitive reserve. J Clin Oncol. 2010;28(29):4434-4440. [FREE Full text] [CrossRef] [Medline]
- Miovic M, Block S. Psychiatric disorders in advanced cancer. Cancer. 2007;110(8):1665-1676. [FREE Full text] [CrossRef] [Medline]
- Warner EL, Basen-Engquist KM, Badger TA, Crane TE, Raber-Ramsey M. The online cancer nutrition misinformation: a framework of behavior change based on exposure to cancer nutrition misinformation. Cancer. 2022;128(13):2540-2548. [CrossRef] [Medline]
- Zhao Y, Zhang J. Consumer health information seeking in social media: a literature review. Health Info Libr J. 2017;34(4):268-283. [FREE Full text] [CrossRef] [Medline]
- Chen J, Wang Y. Social media use for health purposes: systematic review. J Med Internet Res. 2021;23(5):e17917. [FREE Full text] [CrossRef] [Medline]
- Afful-Dadzie E, Afful-Dadzie A, Egala SB. Social media in health communication: a literature review of information quality. Health Inf Manage. 2023;52(1):3-17. [CrossRef] [Medline]
- Critical appraisal tools. JBI. URL: https://jbi.global/critical-appraisal-tools [accessed 2025-04-20]
- von Elm E, Altman DG, Egger M, Pocock SJ, Gøtzsche PC, Vandenbroucke JP, et al. STROBE Initiative. The Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) statement: guidelines for reporting observational studies. Lancet. 2007;370(9596):1453-1457. [FREE Full text] [CrossRef] [Medline]
- Charnock D, Shepperd S, Needham G, Gann R. DISCERN: an instrument for judging the quality of written consumer health information on treatment choices. J Epidemiol Community Health. 1999;53(2):105-111. [FREE Full text] [CrossRef] [Medline]
- Hozo SP, Djulbegovic B, Hozo I. Estimating the mean and variance from the median, range, and the size of a sample. BMC Med Res Methodol. 2005;5:13. [FREE Full text] [CrossRef] [Medline]
- Eysenbach G, Powell J, Kuss O, Sa E. Empirical studies assessing the quality of health information for consumers on the world wide web: a systematic review. JAMA. 2002;287(20):2691-2700. [CrossRef] [Medline]
- Rodríguez-Barranco M, Tobías A, Redondo D, Molina-Portillo E, Sánchez MJ. Standardizing effect size from linear regression models with log-transformed variables for meta-analysis. BMC Med Res Methodol. 2017;17(1):44. [FREE Full text] [CrossRef] [Medline]
- Higgins JPT, Thompson SG, Deeks JJ, Altman DG. Measuring inconsistency in meta-analyses. BMJ. 2003;327(7414):557-560. [FREE Full text] [CrossRef] [Medline]
- Lin L, Xu C. Arcsine-based transformations for meta-analysis of proportions: pros, cons, and alternatives. Health Sci Rep. 2020;3(3):e178. [FREE Full text] [CrossRef] [Medline]
- Kureyama N, Terada M, Kusudo M, Nozawa K, Wanifuchi-Endo Y, Fujita T, et al. Fact-checking cancer information on social media in Japan: retrospective study using Twitter. JMIR Form Res. 2023;7:e49452. [FREE Full text] [CrossRef] [Medline]
- Lavorgna A, Carr L. Tweets and quacks: network and content analyses of providers of non-science-based anticancer treatments and their supporters on Twitter. Sage Open. 2021;11(1):21582440211003084. [CrossRef]
- Duran MB, Kizilkan Y. Quality analysis of testicular cancer videos on YouTube. Andrologia. 2021;53(8):e14118. [CrossRef] [Medline]
- Johnson SB, Parsons M, Dorff T, Moran MS, Ward JH, Cohen SA, et al. Cancer misinformation and harmful information on Facebook and other social media: a brief report. J Natl Cancer Inst. 2022;114(7):1036-1039. [FREE Full text] [CrossRef] [Medline]
- Iglesias-Puzas Á, Conde-Taboada A, Aranegui-Arteaga B, López-Bran E. "Fake news" in dermatology. Results from an observational, cross-sectional study. Int J Dermatol. 2021;60(3):358-362. [CrossRef] [Medline]
- Alsyouf M, Stokes P, Hur D, Amasyali A, Ruckle H, Hu B. 'Fake news' in urology: evaluating the accuracy of articles shared on social media in genitourinary malignancies. BJU Int. 2019;124(4):701-706. [CrossRef] [Medline]
- Nastasi A, Bryant T, Canner JK, Dredze M, Camp MS, Nagarajan N. Breast cancer screening and social media: a content analysis of evidence use and guideline opinions on twitter. J Cancer Educ. 2018;33(3):695-702. [CrossRef] [Medline]
- Kornides ML, Badlis S, Head KJ, Putt M, Cappella J, Gonzalez-Hernadez G. Exploring content of misinformation about HPV vaccine on twitter. J Behav Med. 2023;46(1-2):239-252. [FREE Full text] [CrossRef] [Medline]
- Tomaszewski T, Morales A, Lourentzou I, Caskey R, Liu B, Schwartz A, et al. Identifying false human papillomavirus (HPV) vaccine information and corresponding risk perceptions from Twitter: advanced predictive models. J Med Internet Res. 2021;23(9):e30451. [FREE Full text] [CrossRef] [Medline]
- Park S, Oh H, Park G, Suh B, Bae WK, Kim JW, et al. The source and credibility of colorectal cancer information on Twitter. Medicine (Baltimore). 2016;95(7):e2775. [FREE Full text] [CrossRef] [Medline]
- Lee KC, Oh H, Park G, Park S, Suh B, Bae WK, et al. Transmissibility of the campaign for colorectal cancer awareness in Korea among twitter users. Ann Coloproctol. 2016;32(5):184-189. [FREE Full text] [CrossRef] [Medline]
- Shi S, Brant AR, Sabolch A, Pollom E. False news of a cannabis cancer cure. Cureus. 2019;11(1):e3918. [FREE Full text] [CrossRef] [Medline]
- Petukhova TA, Wilson BN, Gadjiko M, Lee EH, Wang J, Rossi AM, et al. Utilization of Facebook for support and education by patients with skin cancer. Dermatol Online J. 2020;26(3):13030. [FREE Full text] [Medline]
- Wilner T, Holton A. Breast cancer prevention and treatment: misinformation on Pinterest, 2018. Am J Public Health. 2020;110(S3):S300-S304. [CrossRef] [Medline]
- Herbert AS, Hassan N, Malik RD, Loeb S, Myrie A. Exploring urological malignancies on Pinterest: content analysis. JMIR Cancer. 2022;8(3):e36244. [FREE Full text] [CrossRef] [Medline]
- Du J, Preston S, Sun H, Shegog R, Cunningham R, Boom J, et al. Using machine learning-based approaches for the detection and classification of human papillomavirus vaccine misinformation: infodemiology study of reddit discussions. J Med Internet Res. 2021;23(8):e26478. [FREE Full text] [CrossRef] [Medline]
- Massey PM, Kearney MD, Hauer MK, Selvan P, Koku E, Leader AE. Dimensions of misinformation about the HPV vaccine on Instagram: content and network analysis of social media characteristics. J Med Internet Res. 2020;22(12):e21451. [FREE Full text] [CrossRef] [Medline]
- Xu AJ, Myrie A, Taylor JI, Matulewicz R, Gao T, Pérez-Rosas V, et al. Instagram and prostate cancer: using validated instruments to assess the quality of information on social media. Prostate Cancer Prostatic Dis. 2022;25(4):791-793. [CrossRef] [Medline]
- Yang W, Li B, Liu M, Tong D, Zou Y, Li X, et al. Quality evaluation of health information about breast cancer treatment found on WeChat public accounts. Arch Public Health. 2023;81(1):170. [FREE Full text] [CrossRef] [Medline]
- Chen L, Wang X, Peng T. Nature and diffusion of gynecologic cancer-related misinformation on social media: analysis of Tweets. J Med Internet Res. 2018;20(10):e11515. [FREE Full text] [CrossRef] [Medline]
- Yurdaisik I. Analysis of the most viewed first 50 videos on YouTube about breast cancer. Biomed Res Int. 2020;2020:2750148. [FREE Full text] [CrossRef] [Medline]
- Richardson MA, Park W, Bernstein DN, Mesfin A. Analysis of the quality, reliability, and educational content of YouTube videos concerning spine tumors. Int J Spine Surg. 2022;16(2):278-282. [FREE Full text] [CrossRef] [Medline]
- Çapkınoğlu E, İflazoğlu N. Analysis of YouTube videos on rectal cancer surgery as educational resources. Turk J Colorectal Dis. 2023;33(1):7-12. [CrossRef]
- Yalkin O, Uzunoglu MY, İflazoglu N, Esen E. Are YouTube videos a reliable source for obtaining information about colorectal cancer screening? Cureus. 2022;14(1):e21550. [FREE Full text] [CrossRef] [Medline]
- Reinhardt L, Steeb T, Harlaß M, Brütting J, Meier F, Berking C. Are YouTube videos on cutaneous squamous cell carcinoma a useful and reliable source for patients? J Dtsch Dermatol Ges. 2022;20(12):1641-1644. [CrossRef] [Medline]
- Şallı GA, Egil E. Are YouTube videos useful as a source of information for oral care of leukemia patients? Quintessence Int. 2020;51(1):78-85. [CrossRef] [Medline]
- Kharbat FF, Daabes AA. Assessing Arabic YouTube videos on herbal cancer treatment: absence of health information quality. Health Informatics J. 2023;29(3):14604582231198022. [FREE Full text] [CrossRef] [Medline]
- Chai BS, Ingledew PA. Assessment of lung cancer YouTube videos for patient education. J Cancer Educ. 2023;38(5):1760-1766. [CrossRef] [Medline]
- Steeb T, Reinhardt L, Harlaß M, Heppt MV, Meier F, Berking C. Assessment of the quality, understandability, and reliability of YouTube videos as a source of information on basal cell carcinoma: web-based analysis. JMIR Cancer. 2022;8(1):e29581. [FREE Full text] [CrossRef] [Medline]
- Radadiya D, Gonzalez-Estrada A, Lira-Vera JE, Lizarrga-Torres K, Mahapatra SS, Murguia-Fuentes R, et al. Colonoscopy videos on YouTube: are they a good source of patient education? Endosc Int Open. 2020;8(5):E598-E606. [FREE Full text] [CrossRef] [Medline]
- Morena N, Ahisar Y, Wang X, Nguyen D, Rentschler CA, Meguerditchian AN. Content quality of Youtube videos about metastatic breast cancer in young women: systematic assessment. JMIR Cancer. 2023;9:e45101. [FREE Full text] [CrossRef] [Medline]
- Loeb S, Sengupta S, Butaney M, Macaluso JN, Czarniecki SW, Robbins R, et al. Dissemination of misinformative and biased information about prostate cancer on YouTube. Eur Urol. 2019;75(4):564-567. [CrossRef] [Medline]
- Sütcüoğlu O, Özay Z, Özet A, Yazıcı O, Özdemir N. Evaluation of scientific reliability and quality of YouTube videos on cancer and nutrition. Nutrition. 2023;108:111933. [CrossRef] [Medline]
- Didier AJ, Fang L, Stiene J, Spencer CT, Hamouda DM. Evaluation of the quality and comprehensiveness of YouTube videos discussing pancreatic cancer. J Cancer Educ. 2023;38(6):1894-1900. [CrossRef] [Medline]
- Kaplan K, Solak Y. Evaluation of YouTube videos on hepatocellular carcinoma. J Korean Med Sci. 2023;38(7):e50. [FREE Full text] [CrossRef] [Medline]
- Di Bello F, Di Mauro E, Ruvolo CC, Creta M, La Rocca R, Celentano G, et al. Immunotherapy for urological tumors on YouTube: an information-quality analysis. Vaccines (Basel). 2022;11(1):92. [FREE Full text] [CrossRef] [Medline]
- Bahar-Ozdemir Y, Ozsoy-Unubol T, Akyuz G. Is YouTube a high-quality source of information on cancer rehabilitation? J Cancer Surviv. 2022;16(5):1016-1022. [CrossRef] [Medline]
- Muzii B, Di Bello F, Carraturo F, Di Perna T, Califano G, Morra S, et al. Mental health of prostate cancer patients: content review on YouTube. Int J Environ Res Public Health. 2023;20(6):4721. [FREE Full text] [CrossRef] [Medline]
- Pratsinis M, Babst C, Langenauer J, Schmid H, Piller A, Neuenschwander A, et al. Qualitative assessment of medical information on YouTube: a multilingual comparison of common urological conditions. Urol Int. 2021;105(9-10):757-763. [CrossRef] [Medline]
- Baqain L, Mukherji D, Al-Shamsi HO, Abu-Gheida I, Ibraheem A, Rabii K, et al. Quality and reliability of YouTube videos in Arabic as a source of patient information on prostate cancer. Ecancermedicalscience. 2023;17:1573. [FREE Full text] [CrossRef] [Medline]
- Loeb S, Reines K, Abu-Salha Y, French W, Butaney M, Macaluso JN, et al. Quality of bladder cancer information on YouTube. Eur Urol. 2021;79(1):56-59. [CrossRef] [Medline]
- García-Cano-Fernández AM, Szczesniewski-Dudzik JJ, García-Tello A, Diego-García V, Boronat-Catalá J, Llanes-González L. Quality of bladder cancer information on YouTube. Cent European J Urol. 2022;75(3):248-251. [FREE Full text] [CrossRef] [Medline]
- Aydin MA, Akyol H. Quality of information available on YouTube videos pertaining to thyroid cancer. J Cancer Educ. 2020;35(3):599-605. [CrossRef] [Medline]
- Śledzińska P, Bebyn MG, Furtak J. Quality of YouTube videos on meningioma treatment using the DISCERN instrument. World Neurosurg. 2021;153:e179-e186. [CrossRef] [Medline]
- Shungu N, Haley SP, Berini CR, Foster D, Diaz VA. Quality of YouTube videos on prostate cancer screening for Black men. J Am Board Fam Med. 2021;34(4):724-731. [FREE Full text] [CrossRef] [Medline]
- Rodriguez AMR, Blanco-Díaz M, Diaz PL, de la Fuente Costa M, Dueñas L, Prieto IE, et al. Quality analysis of YouTube videos presenting shoulder exercises after breast cancer surgery. Breast Care (Basel). 2022;17(2):188-198. [FREE Full text] [CrossRef] [Medline]
- Reinhardt L, Steeb T, Mifka A, Berking C, Meier F, German Skin Cancer Council. Quality, understandability and reliability of YouTube videos on skin cancer screening. J Cancer Educ. 2023;38(5):1667-1674. [FREE Full text] [CrossRef] [Medline]
- Vu E, Pratsinis M, Plasswilm L, Schmid H, Panje C, Betschart P. Radiotherapy or surgery? Comparative, qualitative assessment of online patient education materials on prostate cancer. Curr Oncol. 2021;28(5):3420-3429. [FREE Full text] [CrossRef] [Medline]
- Segado-Fernández S, Herrera-Peco I, Jiménez-Gómez B, Núñez CR, Jiménez-Hidalgo PJ, Benítez de Gracia E, et al. Realfood and cancer: analysis of the reliability and quality of YouTube content. Int J Environ Res Public Health. 2023;20(6):5046. [FREE Full text] [CrossRef] [Medline]
- Di Bello F, Ruvolo CC, Cilio S, La Rocca R, Capece M, Creta M, et al. Testicular cancer and YouTube: what do you expect from a social media platform? Int J Urol. 2022;29(7):685-691. [CrossRef] [Medline]
- Mamo A, Szeto MD, Mirhossaini R, Fortugno A, Dellavalle RP. Tetrahydrocannabinol and skin cancer: analysis of YouTube videos. JMIR Dermatol. 2021;4(1):e26564. [FREE Full text] [CrossRef] [Medline]
- Meteran H, Høj S, Sigsgaard T, Diers CS, Remvig C, Meteran H. The usefulness of YouTube videos on lung cancer. J Public Health (Oxf). 2023;45(2):e339-e345. [CrossRef] [Medline]
- Şan H. Use of YouTube as an information source for radioactive iodine therapy: do YouTube videos have high quality? Mol Imaging Radionucl Ther. 2022;31(1):42-48. [FREE Full text] [CrossRef] [Medline]
- Ruppert L, Køster B, Siegert AM, Cop C, Boyers L, Karimkhani C, et al. YouTube as a source of health information: analysis of sun protection and skin cancer prevention related issues. Dermatol Online J. 2017;23(1):13030/qt91401264. [FREE Full text] [Medline]
- Ayoub G, Chalhoub E, Sleilaty G, Kourie HR. YouTube as a source of information on breast cancer in the Arab world. Support Care Cancer. 2021;29(12):8009-8017. [CrossRef] [Medline]
- Godskesen T, Holm SF, Höglund AT, Eriksson S. YouTube as a source of information on clinical trials for paediatric cancer. Inf Commun Soc. 2021;26(4):716-729. [CrossRef]
- Yetişir AE. YouTube as a source of information on Merkel cell carcinoma. TURKDERM. 2023;57(3):83-86. [CrossRef]
- Krakowiak M, Fercho J, Piwowska K, Yuser R, Szmuda T, Słoniewski P. YouTube as a source of patient information for meningiomas: a content quality and audience engagement analysis. Healthcare (Basel). 2022;10(3):506. [FREE Full text] [CrossRef] [Medline]
- Wittenberg-Lyles E, Oliver DP, Demiris G, Swarz J, Rendo M. YouTube as a tool for pain management with informal caregivers of cancer patients: a systematic review. J Pain Symptom Manage. 2014;48(6):1200-1210. [FREE Full text] [CrossRef] [Medline]
- Enver N, Doruk C, Kara H, Gürol E, Incaz S, Mamadova U. YouTube™ as an information source for larynx cancer: a systematic review of video content. Eur Arch Otorhinolaryngol. 2020;277(7):2061-2069. [CrossRef] [Medline]
- Güloğlu S, Özdemir Y, Basim P, Tolu S. YouTube English videos as a source of information on arm and shoulder exercise after breast cancer surgery. Eur J Cancer Care (Engl). 2022;31(6):e13685. [CrossRef] [Medline]
- Sahin AN, Sahin AS, Schwenter F, Sebajang H. YouTube videos as a source of information on colorectal cancer: what do our patients learn? J Cancer Educ. 2019;34(6):1160-1166. [FREE Full text] [CrossRef] [Medline]
- Brachtenbach T, Cardenas R, Pate H, Solis L, Chinnery L, Bowers A, et al. YouTube: searching for answers about breast cancer. Breast Dis. 2020;39(2):85-90. [FREE Full text] [CrossRef] [Medline]
- Wang L, Li Y, Gu J, Xiao L. A quality analysis of thyroid cancer videos available on TikTok. Front Public Health. 2023;11:1049728. [FREE Full text] [CrossRef] [Medline]
- Yang S, Zhan J, Xu X. Is TikTok a high-quality source of information on thyroid cancer? Endocrine. 2023;81(2):270-276. [CrossRef] [Medline]
- Zheng S, Tong X, Wan D, Hu C, Hu Q, Ke Q. Quality and reliability of liver cancer-related short Chinese videos on TikTok and Bilibili: cross-sectional content analysis study. J Med Internet Res. 2023;25:e47210. [FREE Full text] [CrossRef] [Medline]
- Xu AJ, Taylor J, Gao T, Mihalcea R, Perez-Rosas V, Loeb S. TikTok and prostate cancer: misinformation and quality of information using validated questionnaires. BJU Int. 2021;128(4):435-437. [FREE Full text] [CrossRef] [Medline]
- Xue X, Yang X, Xu W, Liu G, Xie Y, Ji Z. TikTok as an information hodgepodge: evaluation of the quality and reliability of genitourinary cancers related content. Front Oncol. 2022;12:789956. [FREE Full text] [CrossRef] [Medline]
- Hu R, Zhang H, Yuan B, Zhang K, Xu J, Cui X, et al. Quality and accuracy of gastric cancer related videos in social media videos platforms. BMC Public Health. 2022;22(1):2025. [FREE Full text] [CrossRef] [Medline]
- Pan P, Yu C, Li T, Zhou X, Dai T, Tian H, et al. Xigua video as a source of information on breast cancer: content analysis. J Med Internet Res. 2020;22(9):e19668. [FREE Full text] [CrossRef] [Medline]
- Szczesniewski JJ, Fouz CT, Alba AR, Goizueta FJD, Tello AG, González LL. ChatGPT and most frequent urological diseases: analysing the quality of information and potential risks for patients. World J Urol. 2023;41(11):3149-3153. [CrossRef] [Medline]
- Chen S, Kann BH, Foote MB, Aerts HJWL, Savova GK, Mak RH, et al. Use of artificial intelligence chatbots for cancer treatment information. JAMA Oncol. 2023;9(10):1459-1462. [FREE Full text] [CrossRef] [Medline]
- Pan A, Musheyev D, Bockelman D, Loeb S, Kabarriti AE. Assessment of artificial intelligence chatbot responses to top searched queries about cancer. JAMA Oncol. 2023;9(10):1437-1440. [CrossRef] [Medline]
- Musheyev D, Pan A, Loeb S, Kabarriti AE. How well do artificial intelligence chatbots respond to the top search queries about urological malignancies? Eur Urol. 2024;85(1):13-16. [CrossRef] [Medline]
- Sung H, Ferlay J, Siegel RL, Laversanne M, Soerjomataram I, Jemal A, et al. Global Cancer Statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin. 2021;71(3):209-249. [FREE Full text] [CrossRef] [Medline]
- Boyer C, Selby M, Scherrer JR, Appel RD. The health on the net code of conduct for medical and health websites. Comput Biol Med. 1998;28(5):603-610. [CrossRef] [Medline]
- Robillard JM, Jun JH, Lai J, Feng TL. The QUEST for quality online health information: validation of a short quantitative tool. BMC Med Inform Decis Mak. 2018;18(1):87. [FREE Full text] [CrossRef] [Medline]
- Shoemaker SJ, Wolf MS, Brach C. Development of the Patient Education Materials Assessment Tool (PEMAT): a new measure of understandability and actionability for print and audiovisual patient information. Patient Educ Couns. 2014;96(3):395-403. [FREE Full text] [CrossRef] [Medline]
- Spatz ES, Suter LG, George E, Perez M, Curry L, Desai V, et al. An instrument for assessing the quality of informed consent documents for elective procedures: development and testing. BMJ Open. 2020;10(5):e033297. [FREE Full text] [CrossRef] [Medline]
- Social media fact sheet. Pew Research Center. 2024. URL: https://www.pewresearch.org/internet/fact-sheet/social-media/ [accessed 2024-01-21]
- Gottfried J. Americans' social media use. Pew Research Center. 2024. URL: https://www.pewresearch.org/internet/2024/01/31/americans-social-media-use/ [accessed 2025-01-21]
- Murtfeldt R, Alterman N, Kahveci I, West JD. RIP Twitter API: a eulogy to its vast research contributions. arXiv. Preprint posted online on April 10, 2024. 2024. [CrossRef]
- Chen Z, Koh PW, Ritter PL, Lorig K, Bantum EO, Saria S. Dissecting an online intervention for cancer survivors: four exploratory analyses of internet engagement and its effects on health status and health behaviors. Health Educ Behav. 2015;42(1):32-45. [CrossRef] [Medline]
- Common cancer types. National Cancer Institute. 2015. URL: https://www.cancer.gov/types/common-cancers [accessed 2025-01-21]
- Xia C, Guha S, Muthukrishnan S. Targeting algorithms for online social advertising markets. 2016. Presented at: InIEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM); August 18-21, 2016:485-492; San Francisco, CA, USA. URL: https://ieeexplore.ieee.org/document/7752279
- Wieringa M. What to account for when accounting for algorithms: a systematic literature review on algorithmic accountability. 2020. Presented at: FAT* '20: Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency; January 27, 2020; New York, NY, United States. URL: https://dl.acm.org/doi/10.1145/3351095.3372833
- Jaidka K, Zhou A, Lelkes Y. Brevity is the soul of twitter: the constraint affordance and political discussion. J Commun. 2019;69(4):345-372. [FREE Full text]
- Kaňková J, Binder A, Matthes J. Helpful or harmful? Navigating the impact of social media influencers' health advice: insights from health expert content creators. BMC Public Health. 2024;24(1):3511. [FREE Full text] [CrossRef] [Medline]
- Wang Z, Yin Z, Argyris YA. Detecting medical misinformation on social media using multimodal deep learning. IEEE J Biomed Health Inform. 2021;25(6):2193-2203. [CrossRef] [Medline]
- Okoro Y, Ayo-Farai O, Maduka CP, Okongwu CC, Sodamade OT. A review of health misinformation on digital platforms: challenges and countermeasures. Int J Appl Res Soc Sci. 2024;6(1):23-36. [CrossRef]
- Sorich MJ, Menz BD, Hopkins AM. Quality and safety of artificial intelligence generated health information. BMJ. 2024;384:q596. [CrossRef] [Medline]
Abbreviations
| AI: artificial intelligence |
| GQS: Global Quality Score |
| GRADE: Grading of Recommendations Assessment, Development and Evaluation |
| JAMA-BC: Journal of the American Medical Association Benchmark Criteria |
| mHealth: mobile health |
| OR: odds ratio |
| PEMAT-U: Patient Education Materials Assessment Tool for Understandability |
| PEMAT-A: Patient Education Materials Assessment Tool for Actionability |
| PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses |
| RQ: research question |
| STROBE: Strengthening the Reporting of Observational Studies in Epidemiology |
Edited by N Cahill; submitted 26.Feb.2025; peer-reviewed by L Agrawal, I Odezuligbo, H Maheshwari; comments to author 03.Apr.2025; revised version received 14.Jun.2025; accepted 06.Aug.2025; published 08.Oct.2025.
Copyright©Xue-Jing Liu, Danny Valdez, Maria A Parker, Andi Mai, Eric R Walsh-Buhi. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 08.Oct.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.

