Published on in Vol 24, No 11 (2022): November

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/39571, first published .
Understanding the Social Mechanism of Cancer Misinformation Spread on YouTube and Lessons Learned: Infodemiological Study

Understanding the Social Mechanism of Cancer Misinformation Spread on YouTube and Lessons Learned: Infodemiological Study

Understanding the Social Mechanism of Cancer Misinformation Spread on YouTube and Lessons Learned: Infodemiological Study

Original Paper

1Division of Communication and Media, Ewha Womans University, Seoul, Republic of Korea

2Department of Media and Communication Studies, Jeonbuk National University, Jeonju, Republic of Korea

3Division of Hemato-Oncology, Department of Internal Medicine, Chungnam National University Sejong Hospital, Sejong-Si, Republic of Korea

4Division of Hematology and Oncology, Department of Internal Medicine, College of Medicine, Chungnam National University, Daejeon, Republic of Korea

5Daejeon Regional Cancer Center, Daejeon, Republic of Korea

6Division of Medical Oncology, Yonsei University College of Medicine, Seoul, Republic of Korea

7Yonsei Cancer Center, Yonsei University Health System, Seoul, Republic of Korea

8Division of Cancer Control and Policy, National Cancer Center Korea, Department of Cancer Control and Population Health, National Cancer Center Graduate School of Cancer Science and Policy, Goyang, Republic of Korea

9Division of Hematology and Oncology, Department of Internal Medicine, Soonchunhyang University Hospital Cheonan, Cheonan, Republic of Korea

*these authors contributed equally

Corresponding Author:

Jung Hye Kwon, MD, PhD

Division of Hemato-Oncology

Department of Internal Medicine

Chungnam National University Sejong Hospital

20, Bodeum 7-Ro

Sejong-Si, 30099

Republic of Korea

Phone: 82 44 995 4781

Email: Kwonjhye.onco@gmail.com


Related ArticleThis is a corrected version. See correction statement in: https://www.jmir.org/2022/11/e44334/

Background: A knowledge gap exists between the list of required actions and the action plan for countering cancer misinformation on social media. Little attention has been paid to a social media strategy for disseminating factual information while also disrupting misinformation on social media networks.

Objective: The aim of this study was to, first, identify the spread structure of cancer misinformation on YouTube. We asked the question, “How do YouTube videos play an important role in spreading information about the self-administration of anthelmintics for dogs as a cancer medicine for humans?” Second, the study aimed to suggest an action strategy for disrupting misinformation diffusion on YouTube by exploiting the network logic of YouTube information flow and the recommendation system. We asked the question, “What would be a feasible and effective strategy to block cancer misinformation diffusion on YouTube?”

Methods: The study used the YouTube case of the self-administration of anthelmintics for dogs as an alternative cancer medicine in South Korea. We gathered Korean YouTube videos about the self-administration of fenbendazole. Using the YouTube application programming interface for the query “fenbendazole,” 702 videos from 227 channels were compiled. Then, videos with at least 50,000 views, uploaded between September 2019 and September 2020, were selected from the collection, resulting in 90 videos. Finally, 10 recommended videos for each of the 90 videos were compiled, totaling 573 videos. Social network visualization for the recommended videos was used to identify three intervention strategies for disrupting the YouTube misinformation network.

Results: The study found evidence of complex contagion by human and machine recommendation systems. By exposing stakeholders to multiple information sources on fenbendazole self-administration and by linking them through a recommendation algorithm, YouTube has become the perfect infrastructure for reinforcing the belief that fenbendazole can cure cancer, despite government warnings about the risks and dangers of self-administration.

Conclusions: Health authorities should upload pertinent information through multiple channels and should exploit the existing YouTube recommendation algorithm to disrupt the misinformation network. Considering the viewing habits of patients and caregivers, the direct use of YouTube hospital channels is more effective than the indirect use of YouTube news media channels or government channels that report public announcements and statements. Reinforcing through multiple channels is the key.

J Med Internet Res 2022;24(11):e39571

doi:10.2196/39571

Keywords



Cancer Misinformation and Social Media

Cancer misinformation on social media has piqued the interest of health care professionals. Misinformation is defined as information that causes people to “hold inaccurate factual beliefs” [1]. Studies found that nearly one-third of cancer content on social media was misinformation [2,3] and that 76.9% of it contained harmful information [3]. Cancer misinformation can significantly reduce a patient’s chances of survival. It may lead to a delay in receiving appropriate treatment, the pursuit of unproven alternative therapies, or the rejection of the currently prescribed treatment [2]. Beyond the disintegration of trust between patients, carers, and physicians, the detrimental impacts of disinformation extend to psychological and mental health [2-5].

Indeed, misinformation is one of the most serious negative consequences of social media. False news spreads much “further, faster, deeper, and more broadly” than the truth on social media [6]. The novelty of misinformation grabs the immediate attention of humans, resulting in human-mediated message delivery. The echo chamber induced by the recommendation algorithm amplifies the spread of inaccurate information among individuals with similar interests [7]. Given that caregivers and patients’ family members frequently associate the information community with social support and develop a strong interest in new treatment information [8-10], social media provides the optimal environment for information dissemination, combining human and nonhuman algorithm-driven exposure and connections. This is of particular concern for physicians, as several studies have shown that people are more receptive to misinformation than accurate information [2,4,10].

Thus, it is imperative that public health agencies and health professionals advocate for proactive measures to offset the detrimental consequences of social media misinformation [10-13]. Numerous solutions have been proposed, including digital literacy education [4], accurate information provision [11,14], and media campaigning [15]. While these solutions seek to provide “true” information, rarely have studies been done on concrete and effective ways to disrupt the flow of misinformation on social media, particularly based on an analysis of the current social media information flow network. As a result, a knowledge gap exists between policy implementation and the list of required actions [16].

This may be attributed to the communication pattern of cancer misinformation. Misinformation is frequently mixed with true information in everyday health communication [17]. This mode of communication has rendered it incapable of scrutinizing new information, because information is not only the mixture of true and false but of old and new. As a result, studying the spread of cancer misinformation requires a clearly visible case that can track and reconstruct the pattern of communication.

Complex Contagion of Health Behaviors in Social Media

Adoption of behaviors frequently manifests as a complex contagion: the spread of collective behaviors requiring contact with multiple sources of activation [18]. Complex contagion occurs when social reinforcement influences the transmission of behaviors, beliefs, and attitudes. Unlike simple contagion, in which exposure results in immediate transmission, complex contagion requires social legitimation, credibility, and behavioral complementarity due to uncertainty [18,19]. Studies found that adopting health-related behaviors, such as smoking, exercise, and antivaccine beliefs, are proven to follow the complex contagion mechanism [20-22]. The key point here is not repeated exposure to a single source, but the exposure to multiple sources for social confirmation and reinforcement [18].

Studies have discovered that social media is one of the perfect environments for observing complex contagion [23-25]. Clustered communities in social media serve as peer-to-peer communication and information dissemination networks [25-27]. Social media research frequently reveals patterns of shared exposure to common stimuli [27], but not all patterns are identical. If complex contagion is viewed as the direct link between social media users, it is social cohesion that binds people together in a networked group, often seen in Facebook [28,29]. If the contagion is based on exposure between individuals, regardless of connection between users, then it is contagion via the network structure that enables access to the same source of exposure, typically found in YouTube through its recommendation algorithm [30,31]. The continuous flow of relevant and engaging YouTube videos, as well as the linking of video content via YouTube’s recommendation system [17,32], creates the environment conducive to social reinforcement, a necessary requirement for complex contagion.

Research Context: Fenbendazole Self-administration on YouTube

On September 3, 2019, a Korean-language YouTube video portraying Joe Tippens, who claimed to be entirely cured of his cancer after self-administration of fenbendazole, was uploaded. This video had a total of 2.4 million view counts and 33,702 likes as of September 21, 2021, and was an instant hit among cancer patients and caregivers in South Korea [33]. On September 20, 2019, news outlets reported that pharmacies were out of stock of fenbendazole tablets [34]. According to the Korean government, fenbendazole sales totaled approximately 229,000 tablets in September 2019, which was 5 times the quantity sold from January to August 2019. In November 2019, around 403,000 tablets were sold [35]. The media also grabbed public attention by reporting about a patient with cancer who was a comedian who publicly proclaimed that he would disclose his status following self-administration of fenbendazole. The Ministry of Food and Drug Safety of South Korea issued a caution on the use of fenbendazole for cancer treatment on October 28, 2019 [36]. The National Cancer Center Korea indicated in January 2020 that no clinical trials with fenbendazole were planned, adding that “it is not worthwhile to pursue” [37]. However, YouTube videos describing personal experiences with self-administration of fenbendazole were continuously released across patient and caregiver communities

Research Questions

Given the research context, this study attempted to address the following research questions: (1) What are the characteristics of the YouTube cancer misinformation network regarding the self-administration of fenbendazole? More specifically, (2) is there evidence to support the complex contagion process in social media? (3) Does the networking pattern of the YouTube misinformation network for fenbendazole provide us with useful insights as to why the conventional campaign of disrupting misinformation through news media is less effective? (4) Is there a structural separation of the network between news media providers and user-generated content?


Study Design

We have undertaken unstructured data analysis by collecting data from YouTube. Following the collection of YouTube video data, the data were analyzed using timeline trend analysis, content analysis, and network analysis. These are common strategies applied to unstructured big data analysis. Indeed, mining unstructured online data and analyzing them to discover patterns is a basic technique for big data analytics [38]. The detailed process of data collection, network data processing, and data analysis is depicted in Multimedia Appendix 1.

YouTube Data Collection

We used Google’s YouTube application programming interface (API) to retrieve a list of search query and recommended videos [39] by using a Python program. The API provides the meta information, such as a video title, channel name, the date and time of upload, and number of views, likes, and comments [40].

Our data collection strategy was as follows. First, we used the term “fenbendazole” in Korean to download the list of the videos from the YouTube API. Second, videos were included in the analysis if the video views were more than 50,000, uploaded between September 2019 and September 2020. The number 50,000 has been frequently used in the literature to filter relevant and popular YouTube videos [41,42]. Third, we compiled a list of the top 10 recommended videos for each video we searched. Figure 1 depicts the recommended video interface. The top 10 is the number that has been used in the literature for examining recommendation effects [43,44]. Note that this list was not the same list as our API search query for fenbendazole. YouTube’s recommendation algorithm analyzes viewers’ data regarding their viewing habits and uses the data to make recommendations. This does not imply that the YouTube API reflects the preferences of the API key holder. The YouTube API returns the data to the developer that matches the query parameter, such as video, channel, and playlist [39]. It is known that the YouTube API offers results based on popularity rather than individual user desires [45].

Figure 1. YouTube recommended video interface. Recommended videos are highlighted on the right side. For privacy protection, the screenshot is blurred.
View this figure

Video Content Analysis

To comprehend the content of the collected fenbendazole YouTube videos, we manually examined the videos that met our selection criteria. Two PhD media researchers participated as coders. The two coders have never viewed fenbendazole YouTube videos. Detailed information about the YouTube videos, such as channel name, video title, and view count, was provided as well as the YouTube video URLs.

The coders coded each video’s channel holders and the viewpoint or attitude toward the self-administration of fenbendazole for cancer treatment. Prior to the coders watching the videos, the channel holders were coded as individuals or institutions. The coders selected subcategory codes for each type of holder after viewing a few videos and engaging in discussion with the research team. We coded individuals as patients, health care professionals, caregivers, and others. The coders further coded whether the holder self-administered fenbendazole if the holder was a patient. Regarding institutions, we categorized the holders as news media, hospitals, and others.

We classified the viewpoint or attitude toward the self-administration of fenbendazole into four subcategories:

  1. The “positive” subcategory code was used when the account holder mentioned that she had self-administered fenbendazole or had recommended or sometimes guided others on how to take fenbendazole without physician consultation.
  2. The “reserved” subcategory code was used when the account holder mentioned that doctor consultation for fenbendazole self-administration was necessary or that fenbendazole may only be used for people with specific health conditions, such as patients with terminal-stage cancer or when no treatment options remained.
  3. The “neutral” subcategory code was used when the video only introduced the fenbendazole case.
  4. The “negative” subcategory code was used when the viewpoint was the opposite of the positive viewpoint (ie, the account holder was against the use of fenbendazole).

Finally, our coders reviewed the comments on the videos in order to understand how viewers responded to the videos and to validate their coding by determining whether viewers interpreted the videos as they had been coded.

Influencer and YouTube Video Recommendation Network

A variety of techniques were used to decipher the misinformation network. First, the timeline of uploaded videos was analyzed, and the influencers, in terms of view count and location in the core of the network, were identified through the network analysis index, k-core. Second, network diagrams were drawn in order to locate the influencers in the network. We used the edge list format to convert the list of video searches and recommendations into a network matrix of relational data. In this process, duplicate videos were counted as the value of network ties. The NetDraw program (Analytic Technologies) was then used to draw and to analyze the network data [46]. Although the analysis unit of the collected data was a video, we depicted the network diagram at the channel level for intuitive understanding, as we were particularly interested in who formed the core of the network rather than the role of each video. As the recommendation algorithm of YouTube reflects the quality of video measured by user appreciation, personal preference, and diversity [40], it is logical that the channels at the network’s core exert the most influence on information flow. Through k-core analysis and multidimensional scaling (MDS), we were able to determine the network core. k-core is an index that identifies a highly cohesive region of the entire graph [47]. When the k-core property is combined with the MDS property that clusters network nodes with comparable relationships to other nodes, the k-core nodes tend to locate at the center of the network (ie, the core of network). As the majority of YouTube videos are viewed as a result of recommendations [44], the core of the recommendation network is the center of the information cascade in the dissemination of misinformation.

Third, we investigated the network by examining the ego networks of institutions’ channels, such as hospitals and news organizations, to see if they were in the core of the network. The ego network or egocentric network refers to “a network based on a particular individual...comprised of all the relations that a focal agent has with others” [47]. We think the connection to news media and hospitals in the network is important, as trust in these institutions is related to the spread of misinformation [9]. In addition, the news media serve as fact-checking institutions, and hospitals provide scientific health information to counter misinformation. Thus, a comparison of both institutions’ ego networks can help us better comprehend and analyze the networking pattern between traditional media and YouTube viewing habits.

Ethical Considerations

This study used publicly available data, which are exempt from Institutional Review Board (IRB) approval as stated in Article 16 of the Enforcement Rule of Bioethics and Safety Act of South Korea and Article 13 of its Enforcement Decree. The legislation specifies that publicly available data are waived from IRB review, unless they include collection and recording of sensitive personal information regulated by the Personal Information Protection Act of South Korea. This study did not collect any personal information.


Searched and Collected Data

The YouTube API search for the term “fenbendazole” in Korean returned 702 videos from 227 channels. In total, 90 videos received over 50,000 views. The videos with more than 50,000 views accounted for 98% of the total view traffic from 702 videos. Regarding recommended videos, 573 videos were collected because of the recommendation overlaps between the recommendations from 90 videos. The total number of overlapping videos was 36.3% (n=327) of the theoretical maximum number of videos (ie, 90 videos × 10 recommendations each = 900) without any overlap between recommended videos. The overlap percentage between the searched results and the suggested videos was 25.6%, as the suggested videos were not restricted to fenbendazole videos but reflected the users’ other viewed subjects.

Self-administered Fenbendazole Videos and Opinions From Health Professionals

The number of YouTube videos uploaded for the day and the total number of views for the day’s videos are depicted in Figure 2. These videos are the result of an API search query. The graph indicates that a significant number of personal videos talking about self-administration of fenbendazole were uploaded during the first 6 months after Joe Tippens’ case was introduced.

Interestingly, individuals who self-administered fenbendazole continued to appear on YouTube following the news reports about fenbendazole selling out. Although the number of personal videos dealing with the body’s reaction after self-administration was not large, the consistency of video uploads was of crucial importance; it was not repeated exposure to a single source that confirmed people’s perspectives but multiple exposure to diverse sources that confirmed their views. Furthermore, these videos have garnered considerable attention. The top three videos showing responses to self-administered fenbendazole received an average of 215,256 views, which is 4 times the average view count of 50,326 for all 573 videos in the analysis data set. As these videos reported a favorable self-assessment of fenbendazole for pain relief and a fall in tumor markers, the information has virtually become reliable information to viewers. The top 20 most-viewed videos are listed in Table 1; 8 of these are personal testimonials about how fenbendazole improved their symptoms.

In addition, some health professionals on YouTube have not strongly criticized fenbendazole use. This signals to viewers that it would be worth trying because at least it would not be toxic. The comments in these videos referred to these health professionals as “true doctor” [48]. Out of 3 professionals, 2 recommended that patients should consult their doctors before taking fenbendazole (Table 1), but they also described the role of other catalysts, such as vitamins, in aiding in the absorption of the fenbendazole components. As these health professionals are active physicians—one is a physician practicing internal medicine and holds a PhD in chemistry, and the other is the director of the Division of Hematology and Medical Oncology in a general hospital—people interested in the self-administration of fenbendazole interpreted this as a positive signal for taking the medication. For example, the comments by users in the video by internal medicine physicians included the following: “an excellent video to know more details. It helped me making decision between confusing opinions” (ID: ***** Kim, anonymized for privacy protection), “...it was hard to trust physicians but by listening your heartfelt comments, my negative trauma is gone today” (ID: ** tree, anonymized), and “thank you for encouragement as a person taking fenbendazole with colorectal cancer at stage 4” (ID: ***flower, anonymized). One other professional—a clinic owner and radiologist—even advocated for the use of fenbendazole.

Figure 2. The timeline of fenbendazole YouTube videos. MFDS: Ministry of Food and Drug Safety; NCC: National Cancer Center Korea.
View this figure
Table 1. The top 20 YouTube fenbendazole videos.
Date (year-month-day)ChannelDescriptionaTitle summaryViewpointbView countc, nLike countc, n
2019-09-03World Village Magazine TVFirst video about fenbendazole effectIntroduction of Joe TippensPositive2,439,44033,702
2020-02-18VIDEOMUGNews mediaNews Report of IssueNeutral1,708,62811,829
2019-11-04Dr. Ezra TVHealth professionalAlbendazole PossibilityReservation1,151,98437,337
2019-10-26JIGUINSelf-administrationFinal-Stage Cancer SymptomPositive869,4339196
2019-12-11Dr. Ezra TVHealth professionalParasites That Transmit CancerReservation660,42026,729
2020-01-13Changbal Testing TVHashtag trollerParasite Shock VisualPositive600,2579529
2019-10-29Dr. Ezra TVHealth professionalPersonal ThoughtReservation521,48716,738
2019-10-07JIGUINSelf-administrationPostfenbendazole Personal ReviewPositive477,38310,320
2020-01-02KimCellPatient taking albendazole (2 tablets)Metastatic Cancer and FenbendazolePositive387,156d
2019-10-09AhnSelf-administrationPostfenbendazole Personal ReviewPositive382,8386694
2019-10-30JIGUINSelf-administrationPostfenbendazole Personal ReviewPositive382,8189652
2020-01-10Sanchae StorySelf-administrationSelf-administration of Fenbendazole and CuredPositive362,2566909
2019-10-01JIGUINSelf-administrationPostfenbendazole Personal ReviewPositive325,4785562
2019-11-21Dr. Ezra TVHealth professionalUS Government is Testing FenbendazolePartially positive or reservation299,93613,376
2019-10-09MitoTV (doctor)Health professionalAnticancer Anthelmintics?Positive278,7068120
2019-10-05JIGUINSelf-administrationReply to CommentsPositive278,3958838
2020-04-21Segye Ilbo (newspaper)News mediaNews Report on Mr. KimNeutral261,3281534
2019-10-24JIGUINSelf-administrationCancer Expert ViewPositive258,6647543
2019-11-28Changbal Testing TVHashtag trollerParasites and CancerNot applicable248,9077009
2019-09-23Dr. Nah’s Medical TalkHealth professionalExpert View on FenbendazoleNegative or reservation239,6433777

aThe description category is based on the classification of the channel holder.

bThe viewpoint category is based on the attitude toward the self-administration of fenbendazole.

cThe count as of September 21, 2021, the date of data collection.

dNot reported.

YouTube Recommendation Network: Creating a Spiral Circle of Positivity for Fenbendazole Self-administration

Figure 3 displays the recommendation video network during the first 6-month period; the network during the whole year is displayed in Multimedia Appendix 1. The network demonstrates that personal videos after the self-administration of fenbendazole were continuously posted and connected through the recommendation algorithm.

Figure 3. Fenbendazole YouTube video recommendation network between September 2019 and February 2020. The default node color is red. The yellow nodes are the channels where fenbendazole self-administration was conducted. The node size reflects the degree of a node, namely, the number of connections with other nodes. The tie strength reflects the number of repeated recommendation linkages.
View this figure

The recommendation network reveals that positive evaluation videos by individual patients and videos by health care professionals synergistically promoted the spread of favorable discourse about fenbendazole. In Figure 3, the first video explaining the positive effect of fenbendazole self-administration on cancer is located outside the network. The central position was taken by one professional video channel—Dr. Ezra TV—which stated that it is preferable for the Korean government to allow people to take fenbendazole in order to collect feasible data rather than warning people, and that it is a sign of capitalism in a capitalist society that a cheap affordable medication would not be invested in for cancer treatment. The Dr. Ezra TV channel was linked to seven video channels about self-administered fenbendazole in a single step, which means that personal videos on the body’s reaction to fenbendazole were directly recommended by the YouTube recommendation system. The channels with the top 20–highest view counts were inextricably linked. For example, the Ahn channel and the JIGUIN channel, along with the Dr. Erza TV channel, were a tightly connected network group, namely, a network clique. These two patient channels reported positive outcomes after the self-administration of fenbendazole in terms of pain reduction and positive blood test results; albeit they did acknowledge that it was unknown whether the effect was primarily due to fenbendazole or was accompanied by other treatments and procedures.

By contrast, the network of recommendations did not include any government or authoritative medical channels. Within the network core, there were two hospital channels; however, it appears that the channels just mirrored users’ YouTube viewing habits. These channels did not include any content on fenbendazole. In fact, these two hospitals are the top two hospitals in terms of the number of cancer patients they treat. News channels played no role in disseminating “true” pertinent information. This becomes obvious when the recommendation network is enlarged to a 1-year time range (Multimedia Appendix 1) in order to visualize the links between the spiral of information and news media, as the initial 6-month recommendation network did not reveal many news media linkages.

The ego network of two hospitals and three news media outlets is depicted in Figure 4. The top two diagrams demonstrate how hospital channels were immediately connected to numerous video channels about self-administration of fenbendazole via a single recommendation, whereas news channels were rarely connected to these. In other words, even if accurate information is distributed, patients and caregivers engaged in self-administration are unlikely to be connected to government and other authentic YouTube medical channels.

Figure 4. Ego networks of the fenbendazole YouTube recommendation network by institutions between September 2019 and February 2020. The default node color is red. The yellow nodes are the channels where fenbendazole self-administration was conducted. The node size reflects the degree of a node, namely the number of connections with other nodes. The tie strength reflects the number of repeated recommendation linkages.
View this figure

Principal Findings

This study delved into one of the cancer misinformation networks on YouTube. By analyzing the data from searched and recommended videos, we found that personal videos about self-administered fenbendazole were continuously uploaded and accumulated over time as if showing promising evidence for the use of fenbendazole as a cancer treatment. In addition, the recommended content network of fenbendazole has become the infrastructure for confirming the audience’s belief and hope in fenbendazole as an alternative cancer medicine. Patients are actively seeking health information over the internet, thereby increasing their self-efficacy in making treatment decisions and altering provider-patient interactions [49-52]. As such, the appearance of supportive professional videos stating that the use of fenbendazole for cancer is scientifically unknown, but possibly helpful, may inspire hope and belief among cancer patients and caregivers who are thinking about self-administering fenbendazole; these videos may also lead patients and caregivers to disregard announcements from the National Cancer Center Korea and the Korean Medical Association.

Unfortunately, the effectiveness of fenbendazole has not been established, and other major adverse effects have been observed [53]. During our investigation, we also found that the YouTube recommendation network was unrelated to credible medical knowledge content. Although some hospital channels were, indeed, connected to the network, it seemed to reflect people’s viewing habits rather than topics related to fenbendazole.

In summary, while the YouTube content and recommendation network served as a substantial information source for complex contagion, medical institutions and government entities were excluded from the network, and no dialogue from them was discovered. This resulted in a breakdown of communication between patients and caregivers, resulting in enormous sales of fenbendazole tablets.

Strategies to Fight Social Media Cancer Misinformation

Given YouTube’s role as a hub for complex contagion, three strategies to fight against social media cancer misinformation networks are recommended. First, health authorities need to upload a variety of pertinent information through multiple channels. This does not necessarily mean the authorities require multiple channels. They can incorporate existing influencers and other channels. The objective is to have numerous sources of exposure in order to disrupt the cascade of misinformation. A single source will not be sufficient to break the complex contagion dynamic.

Second, it is imperative that health authorities take into account YouTube’s recommendation system, current viewing habits, and information flow network between patients and caregivers. As illustrated in Figure 4, prominent hospital channels may be an option, as stakeholders are already engaged in the channels’ content. Adding a new channel would be ineffective, as it would need to build an audience from scratch.

Third, relying on the news media does not resolve the issue: health authorities must take an active role in resolving social media misinformation. The news media, on the other hand, is frequently constrained by mechanical objectivity and is prone to report both sides of an argument. Furthermore, those who follow the news media are not the target audience for the health authorities’ message. For example, social media is not the first preference of individuals who intend to learn from the news.

Limitations

While our investigation exposed YouTube’s cancer misinformation network, it is not without limits. To begin, the study’s data were not collected in real time. The real-time data collection process may have resulted in more concrete information flow dynamics between YouTube videos. However, one of the most difficult areas of health communication is the real-time monitoring system for cancer misinformation. It is difficult to classify misinformation in advance, before it spreads on social media, unless a global censorship system is in place to monitor every conversation. The fenbendazole case in this study, for example, demonstrates how the unexpected introduction of a case from a foreign country, such as that of Joe Tippens, prompted reactions among Korean cancer patients and caregivers.

Second, the study is country specific and focused on fenbendazole as a case study. Although we discovered the insight that the spread of cancer misinformation follows complex contagion logic in social media, more evidence is necessary to augment the study’s conclusions. Interviews with people who have taken fenbendazole, in particular, would be helpful in determining the intensity of social media influence.

Third, we concentrated on the videos that had the highest number of views. This excludes individuals who self-administered fenbendazole but received little attention. Fenbendazole was extensively used to the extent that it was sold out at the national level. A future large-scale national-level investigation will further enrich our understanding of social media misinformation.

Conclusions

This study contributes to the body of knowledge by including practice strategies for combating social media cancer misinformation. By studying the content on YouTube, this study has attempted to close the knowledge gap between what to do and how to do it when it comes to delivery of accurate information via social media. The study proposed a way for campaigning against misinformation and educating people by having health policy authorities use the current network of information flow on YouTube. The study focused on involvement with the actual information flow network rather than implementing a conventional one-way communication strategy, even on social media. This action plan will offer valuable information, especially for people who rely solely on online resources, such as social media, and have limited means of accessing expert health knowledge and information. The study’s recommendation may not be a comprehensive strategy for combating misinformation, but it may be one of the most successful methods for increasing trust between health care practitioners and stakeholders, such as patients and caregivers.

Acknowledgments

This study was supported by the National R&D Program for Cancer Control through the National Cancer Center (NCC) funded by the Ministry of Health & Welfare, Republic of Korea (HA21C0048).

Conflicts of Interest

None declared.

Multimedia Appendix 1

Study design and network diagram.

PDF File (Adobe PDF File), 370 KB

  1. Kuklinski J, Quirk P, Jerit J, Schwieder D, Rich R. Misinformation and the currency of democratic citizenship. J Polit 2000 Aug;62(3):790-816 [FREE Full text] [CrossRef]
  2. Gage-Bouchard EA, LaValley S, Warunek M, Beaupin LK, Mollica M. Is cancer information exchanged on social media scientifically accurate? J Cancer Educ 2018 Dec;33(6):1328-1332 [FREE Full text] [CrossRef] [Medline]
  3. Johnson SB, Parsons M, Dorff T, Moran MS, Ward JH, Cohen SA, et al. Cancer misinformation and harmful information on Facebook and other social media: A brief report. J Natl Cancer Inst 2022 Jul 11;114(7):1036-1039 [FREE Full text] [CrossRef] [Medline]
  4. Warner EL, Waters AR, Cloyes KG, Ellington L, Kirchhoff AC. Young adult cancer caregivers' exposure to cancer misinformation on social media. Cancer 2021 Apr 15;127(8):1318-1324 [FREE Full text] [CrossRef] [Medline]
  5. Miller TT, Maurer SH, Felker JT. Searching for a cure on Facebook: Patterns of social media use amongst caregivers of children with brain tumors. Cancer Med 2022 Sep 28;11(17):3323-3331 [FREE Full text] [CrossRef] [Medline]
  6. Vosoughi S, Roy D, Aral S. The spread of true and false news online. Science 2018 Mar 09;359(6380):1146-1151. [CrossRef] [Medline]
  7. Törnberg P. Echo chambers and viral misinformation: Modeling fake news as complex contagion. PLoS One 2018;13(9):e0203958 [FREE Full text] [CrossRef] [Medline]
  8. Taylor J, Pagliari C. The social dynamics of lung cancer talk on Twitter, Facebook and Macmillan.org.uk. NPJ Digit Med 2019;2:51 [FREE Full text] [CrossRef] [Medline]
  9. Chou WS, Oh A, Klein WMP. Addressing health-related misinformation on social media. JAMA 2018 Dec 18;320(23):2417-2418. [CrossRef] [Medline]
  10. Himelboim I, Han JY. Cancer talk on Twitter: Community structure and information sources in breast and prostate cancer social networks. J Health Commun 2014;19(2):210-225. [CrossRef] [Medline]
  11. Mollica MA, Chou WS, Tonorezos ES, Smith AW. Young adult caregivers' perceptions of cancer misinformation on social media: Response to Warner et al. Cancer 2021 Apr 15;127(8):1177-1179 [FREE Full text] [CrossRef] [Medline]
  12. Peterson JS, Swire-Thompson B, Johnson SB. What is the alternative? Responding strategically to cancer misinformation. Future Oncol 2020 Sep;16(25):1883-1888. [CrossRef] [Medline]
  13. Vraga EK, Bode L. Correction as a solution for health misinformation on social media. Am J Public Health 2020 Oct;110(S3):S278-S280. [CrossRef]
  14. Biancovilli P, Makszin L, Csongor A. Breast cancer on social media: A quali-quantitative study on the credibility and content type of the most shared news stories. BMC Womens Health 2021 May 15;21(1):202 [FREE Full text] [CrossRef] [Medline]
  15. Conley C, Otto A, McDonnell G, Tercyak K. Multiple approaches to enhancing cancer communication in the next decade: Translating research into practice and policy. Transl Behav Med 2021 Nov 30;11(11):2018-2032. [CrossRef] [Medline]
  16. Sylvia Chou W, Gaysynsky A. A prologue to the special issue: Health misinformation on social media. Am J Public Health 2020 Oct;110(S3):S270-S272. [CrossRef]
  17. Tang L, Fujimoto K, Amith M, Cunningham R, Costantini RA, York F, et al. "Down the rabbit hole" of vaccine misinformation on YouTube: Network exposure study. J Med Internet Res 2021 Jan 05;23(1):e23262 [FREE Full text] [CrossRef] [Medline]
  18. Centola D, Macy M. Complex contagions and the weakness of long ties. Am J Sociol 2007 Nov;113(3):702-734. [CrossRef]
  19. Guilbeault D, Becker J, Centola D. Complex contagions: A decade in review. In: Lehmann S, Ahn YY, editors. Complex Spreading Phenomena in Social Systems. Cham, Switzerland: Springer; 2018:3-25.
  20. Christakis NA, Fowler JH. The collective dynamics of smoking in a large social network. N Engl J Med 2008 May 22;358(21):2249-2258 [FREE Full text] [CrossRef] [Medline]
  21. Zhang J, Brackbill D, Yang S, Becker J, Herbert N, Centola D. Support or competition? How online social networks increase physical activity: A randomized controlled trial. Prev Med Rep 2016 Dec;4:453-458 [FREE Full text] [CrossRef] [Medline]
  22. Salathé M, Bonhoeffer S. The effect of opinion clustering on disease outbreaks. J R Soc Interface 2008 Dec 06;5(29):1505-1508 [FREE Full text] [CrossRef] [Medline]
  23. Borge-Holthoefer J, Banos RA, Gonzalez-Bailon S, Moreno Y. Cascading behaviour in complex socio-technical networks. J Complex Netw 2013 Apr 22;1(1):3-24. [CrossRef]
  24. Bakshy E, Rosenn I, Marlow C, Adamic L. The role of social networks in information diffusion. In: Proceedings of the 21st International Conference on World Wide Web. 2012 Presented at: The 21st International Conference on World Wide Web; April 16-20, 2012; Lyon, France p. 519-528. [CrossRef]
  25. Harrigan N, Achananuparp P, Lim E. Influentials, novelty, and social contagion. Soc Networks 2012 Oct;34(4):470-480. [CrossRef]
  26. Centola D. The spread of behavior in an online social network experiment. Science 2010 Sep 03;329(5996):1194-1197. [CrossRef] [Medline]
  27. Centola D. Social media and the science of health behavior. Circulation 2013 May 28;127(21):2135-2144. [CrossRef] [Medline]
  28. Friedkin NE. Social cohesion. Annu Rev Sociol 2004 Aug 01;30(1):409-425. [CrossRef]
  29. Friedkin NE. Structural cohesion and equivalence explanations of social homogeneity. Sociol Methods Res 2016 Jun 30;12(3):235-261. [CrossRef]
  30. Burt RS. Cohesion versus structural equivalence as a basis for network subgroups. Sociol Methods Res 2016 Jun 29;7(2):189-212. [CrossRef]
  31. Burt RS. Social contagion and innovation: Cohesion versus structural equivalence. Am J Sociol 1987 May;92(6):1287-1335. [CrossRef]
  32. Park SJ, Lim YS, Park HW. Comparing Twitter and YouTube networks in information diffusion: The case of the “Occupy Wall Street” movement. Technol Forecast Soc Change 2015 Jun;95:208-217. [CrossRef]
  33. Nam JH. Dog anthelmintic sells out after rumors of cancer treatment...Food and Drug Administration warns "Do not take" [Document in Korean]. SBS News. 2019.   URL: https://news.sbs.co.kr/news/endPage.do?news_id=N1005449578&plink=ORI [accessed 2022-02-20]
  34. Choi H. Unverified YouTube clip: Cancer is cured with dog anthelmintic...sold out [Document in Korean]. JTBC News. 2019.   URL: https://news.jtbc.joins.com/article/article.aspx?news_id=NB11882980 [accessed 2022-02-20]
  35. Kang S. Animal anthelmintic "fenbendazole" has no anticancer effect, but still popular [Document in Korean]. Hit News. 2020.   URL: http://www.hitnews.co.kr/news/articleView.html?idxno=30389 [accessed 2022-02-20]
  36. Anthelmintic for animals. It’s only allowed for animals [Document in Korean]. Ministry of Food and Drug Safety. 2019 Oct 28.   URL: https://www.mfds.go.kr/brd/m_99/down.do?brd_id=ntc0021&seq=43763&data_tp=A&file_seq=2 [accessed 2022-03-02]
  37. Lee E. Anthelmintic for dogs as anticancer medicine – The National Cancer Center says, "it’s not worth it to conduct clinical trials" [Document in Korean]. The JoongAng. 2020.   URL: https://www.joongang.co.kr/article/23678676#home [accessed 2022-03-03]
  38. Das TK, Kumar PM. Big data analytics: A framework for unstructured data analysis. Int J Eng Sci Technol 2013;5(1):155-158 [FREE Full text]
  39. Data API. Search: list. YouTube. 2020.   URL: https://developers.google.com/youtube/v3/docs/search/list [accessed 2021-03-30]
  40. Davidson J, Liebald B, Liu J, Nandy P, Van Vleet T, Gargi U, et al. The YouTube video recommendation system. In: Proceedings of the 4th ACM conference on Recommender Systems. 2010 Presented at: The 4th ACM conference on Recommender Systems; September 26-30, 2010; Barcelona, Spain p. 293-296. [CrossRef]
  41. Burgess J, Matamoros-Fernández A. Mapping sociocultural controversies across digital media platforms: One week of #gamergate on Twitter, YouTube, and Tumblr. Commun Res Pract 2016 Apr 25;2(1):79-96. [CrossRef]
  42. Aggrawal N, Arora A, Anand A, Irshad MS. View-count based modeling for YouTube videos and weighted criteria–based ranking. In: Ram M, Davim JP, editors. Advanced Mathematical Techniques in Engineering Sciences. Boca Raton, FL: CRC Press; 2018.
  43. Papadamou K, Papasavva A, Zannettou S, Blackburn J, Kourtellis N, Leontiadis I, et al. Disturbed YouTube for kids: Characterizing and detecting inappropriate videos targeting young children. In: Proceedings of the 14th International AAAI Conference on Web and Social Media. 2020 Presented at: The 14th International AAAI Conference on Web and Social Media; June 8-11, 2019; Virtual p. 522-533   URL: https://ojs.aaai.org/index.php/ICWSM/article/view/7320
  44. Zhou R, Khemmarat S, Gao L. The impact of YouTube recommendation system on video views. In: Proceedings of the 10th Annual Conference on Internet Measurement. 2010 Presented at: The 10th Annual Conference on Internet Measurement; November 1-30, 2010; Melbourne, Australia p. 404-410. [CrossRef]
  45. Bärtl M. YouTube channels, uploads and views. Convergence 2018 Jan 10;24(1):16-32. [CrossRef]
  46. Borgatti SP, Everett MG, Freeman LC. Ucinet for Windows: Software for Social Network Analysis. Harvard, MA: Analytic Technologies; 2002.   URL: https://sites.google.com/site/ucinetsoftware/home [accessed 2022-10-26]
  47. Scott J. Social Network Analysis. 4th edition. London, UK: SAGE Publications Ltd; 2017.
  48. Lee H. “The controversy of YouTube doctors – they mentioned how to self-administer fenbendazole” [Document in Korean]. Young Doctors. 2019 Nov 18.   URL: https://www.docdocdoc.co.kr/news/articleView.html?idxno=1074269 [accessed 2022-08-15]
  49. Ziebland S, Chapple A, Dumelow C, Evans J, Prinjha S, Rozmovits L. How the internet affects patients' experience of cancer: A qualitative study. BMJ 2004 Mar 06;328(7439):564 [FREE Full text] [CrossRef] [Medline]
  50. McMullan M. Patients using the internet to obtain health information: How this affects the patient-health professional relationship. Patient Educ Couns 2006 Oct;63(1-2):24-28. [CrossRef] [Medline]
  51. Lleras de Frutos M, Casellas-Grau A, Sumalla EC, de Gracia M, Borràs JM, Ochoa Arnedo C. A systematic and comprehensive review of internet use in cancer patients: Psychological factors. Psychooncology 2020 Jan;29(1):6-16. [CrossRef] [Medline]
  52. Bass SB, Ruzek SB, Gordon TF, Fleisher L, McKeown-Conn N, Moore D. Relationship of internet health information use with patient behavior and self-efficacy: Experiences of newly diagnosed cancer patients who contact the National Cancer Institute's Cancer Information Service. J Health Commun 2006 Mar;11(2):219-236. [CrossRef] [Medline]
  53. Oh K. Mr. Kim said "I failed. Never recommend anthelmintics for cancer treatment" [Document in Korean]. Chosun Ilbo. 2020 Sep 22.   URL: https://www.chosun.com/culture-life/health/2020/09/22/II7X43CUAZFEHO4BMLQKYCATUU [accessed 2022-03-30]


API: application programming interface
IRB: Institutional Review Board
MDS: multidimensional scaling


Edited by A Mavragani; submitted 01.06.22; peer-reviewed by L Hao Yen, A Benis, A Singleton; comments to author 12.08.22; revised version received 30.08.22; accepted 20.10.22; published 14.11.22

Copyright

©Ho Young Yoon, Kyung Han You, Jung Hye Kwon, Jung Sun Kim, Sun Young Rha, Yoon Jung Chang, Sang-Cheol Lee. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 14.11.2022.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.