Published on in Vol 24, No 1 (2022): January

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/28152, first published .
Medical and Health-Related Misinformation on Social Media: Bibliometric Study of the Scientific Literature

Medical and Health-Related Misinformation on Social Media: Bibliometric Study of the Scientific Literature

Medical and Health-Related Misinformation on Social Media: Bibliometric Study of the Scientific Literature

Original Paper

1Oral and Maxillofacial Radiology, Applied Oral Sciences and Community Dental Care, Faculty of Dentistry, The University of Hong Kong, Hong Kong, China

2Ludwig Boltzmann Institute for Digital Health and Patient Safety, Medical University of Vienna, Vienna, Austria

3Department of Molecular, Cell and Developmental Biology, University of California Los Angeles, Los Angeles, CA, United States

4Division of Pediatric Nephrology and Gastroenterology, Comprehensive Center for Pediatrics, Medical University of Vienna, Vienna, Austria

5Faculty of Medicine, University of Crete, Heraklion, Greece

6Department of Translational Stem Cell Biology, Medical University of Varna, Varna, Bulgaria

7College of Medicine, Mohammed Bin Rashid University of Medicine and Health Sciences, Dubai, United Arab Emirates

8Department of Anaesthesia, Intensive Care Medicine and Pain Medicine, Medical University of Vienna, Vienna, Austria

9Institute of Genetics and Animal Biotechnology of the Polish Academy of Sciences, Jastrzebiec, Poland

Corresponding Author:

Atanas G Atanasov, PhD

Ludwig Boltzmann Institute for Digital Health and Patient Safety

Medical University of Vienna

Spitalgasse 23

Vienna, 1090

Austria

Phone: 43 664 1929 852

Email: Atanas.Atanasov@dhps.lbg.ac.at


Background: Social media has been extensively used for the communication of health-related information and consecutively for the potential spread of medical misinformation. Conventional systematic reviews have been published on this topic to identify original articles and to summarize their methodological approaches and themes. A bibliometric study could complement their findings, for instance, by evaluating the geographical distribution of the publications and determining if they were well cited and disseminated in high-impact journals.

Objective: The aim of this study was to perform a bibliometric analysis of the current literature to discover the prevalent trends and topics related to medical misinformation on social media.

Methods: The Web of Science Core Collection electronic database was accessed to identify relevant papers with the following search string: ALL=(misinformati* OR “wrong informati*” OR disinformati* OR “misleading informati*” OR “fake news*”) AND ALL=(medic* OR illness* OR disease* OR health* OR pharma* OR drug* OR therap*) AND ALL=(“social media*” OR Facebook* OR Twitter* OR Instagram* OR YouTube* OR Weibo* OR Whatsapp* OR Reddit* OR TikTok* OR WeChat*). Full records were exported to a bibliometric software, VOSviewer, to link bibliographic information with citation data. Term and keyword maps were created to illustrate recurring terms and keywords.

Results: Based on an analysis of 529 papers on medical and health-related misinformation on social media, we found that the most popularly investigated social media platforms were Twitter (n=90), YouTube (n=67), and Facebook (n=57). Articles targeting these 3 platforms had higher citations per paper (>13.7) than articles covering other social media platforms (Instagram, Weibo, WhatsApp, Reddit, and WeChat; citations per paper <8.7). Moreover, social media platform–specific papers accounted for 44.1% (233/529) of all identified publications. Investigations on these platforms had different foci. Twitter-based research explored cyberchondria and hypochondriasis, YouTube-based research explored tobacco smoking, and Facebook-based research studied vaccine hesitancy related to autism. COVID-19 was a common topic investigated across all platforms. Overall, the United States contributed to half of all identified papers, and 80% of the top 10 most productive institutions were based in this country. The identified papers were mostly published in journals of the categories public environmental and occupational health, communication, health care sciences services, medical informatics, and medicine general internal, with the top journal being the Journal of Medical Internet Research.

Conclusions: There is a significant platform-specific topic preference for social media investigations on medical misinformation. With a large population of internet users from China, it may be reasonably expected that Weibo, WeChat, and TikTok (and its Chinese version Douyin) would be more investigated in future studies. Currently, these platforms present research gaps that leave their usage and information dissemination warranting further evaluation. Future studies should also include social platforms targeting non-English users to provide a wider global perspective.

J Med Internet Res 2022;24(1):e28152

doi:10.2196/28152

Keywords



Public health information has been traditionally distributed to the public with the use of printed media, television, or radio. With the rise of participatory web and social media [1] and particularly in the face of recent pandemics, such as the H1N1 influenza pandemic in 2009 and the COVID-19 pandemic [2], the internet plays a major role in information sharing. The general public no longer acts as a passive consumer but plays a critical role in the generation, filtering, and amplification of public health information [1]. Health care–related scientific discoveries are now often condensed into news pieces written in layman’s terms and disseminated to broad and nonexpert audiences via social media, which contributes to not only better visibility of important information, but also better communication between health care professionals and the community [3]. Another major benefit of social media for health care is the potential for patient empowerment by providing a platform where patients can get information about their medical condition, communicate with health care professionals, share their experiences, and support other individuals affected by the same condition [4].

While providing numerous empowerment opportunities, there lies a great potential for miscommunication and misinformation [5] within the social media–based setting of health-related information distribution. While social media has increased and improved the dissemination of scientific results to the community, it has also increased the sensationalist language used to describe scientific findings [6,7]. Often, media articles may report research findings with misinterpretation and overstatement that can lead to confusion, misinformation, and mistrust in scientific reporting [6]. Moreover, social media empowers pseudoexperts and nonexpert influencers in sharing opinions and false information in the area of health care [8]. Very often, important societal figures, such as celebrities, politicians, and activists, without any expert knowledge of a certain topic, but with a large influence, can take part in spreading health-related misinformation [8]. The need for social media to moderate the information shared and increase expert consultation is increasingly evident and could be one way to reduce the spread of misinformation [9].

One of the most polarizing topics in recent years has been vaccination, following a scientific article from 1998 by Wakefield et al, which proposed a causative link between the measles, mumps, and rubella (MMR) vaccine and autism [3,10]. The study by Wakefield et al was later found to be flawed and fraudulent and was retracted [11-14]. Even though the findings in the study by Wakefield et al have since been disproved as numerous subsequent studies found no link between vaccines and autism, the study caused great damage to vaccine programs worldwide, with a considerable increase in the number of people rejecting vaccination in the past decades [9]. Another prominent illustrative example is the case of measles reappearance in the United States [15]. In the United States, there was an immense surge in antivaccine Tweets around 2015 to 2016, closely following the 2014 to 2015 measles outbreak and the release of Wakefield’s antivaccine movie Vaxxed in 2016 [16]. This could be linked to the finding that antivaccine posts on Facebook were often shared and liked more often than provaccine posts [17]. Similarly, individuals exposed to negative opinions on vaccination are more likely to further share the opinions compared to individuals exposed to positive or neutral opinions [18]. This is potentiated by the so called “echo chamber effect,” where many social media users are exposed to curated content that is likely to align to their existing beliefs and exacerbates the strength of the misinformation they receive [19,20].

As medical misinformation is an increasingly relevant topic to study, the amount of available literature is growing. On this background, the aim of this study was to perform a bibliometric analysis of the current literature to discover the prevalent trends and topics related to medical misinformation on social media. Conventional systematic reviews have been published on this topic to identify original articles and summarize their methodological approaches and themes [7,21]. A bibliometric study could complement their findings, for instance, by identifying the most productive authors and institutions, evaluating the geographical distribution of the publications, revealing recurring journals disseminating such research findings, unveiling the most common keywords or concepts reported, and evaluating if the publications were well cited and disseminated in journals with high impact factors. These data can serve as starting points to guide fellow researchers to pinpoint papers relevant to their studies, contact potential collaborators to conduct joint research, and find suitable journals to submit their work. At the same time, these data can help researchers find missing gaps in the literature and missing parties contributing to the field, so that the missing pieces can be filled. Since the most common social media platforms have originated in the United States, it was hypothesized that the United States would have the highest contribution in this area of academic research. This would be an important research question as publication bias toward the United States might shadow or fail to capture the wider spectrum of global developments and experiences regarding medical misinformation on social media.


Data Source and Search Strategy

A bibliometric analysis is a study that applies mathematical and statistical methods to books and other media of communication, such as academic publications [22]. Similar to a previous bibliometric study [23], this work was performed in accordance with the Preferred Reporting Items for Systematic Reviews and Meta-Analysis (PRISMA) statement [24]. The Web of Science (WoS) Core Collection database was accessed on January 13, 2021, via the following search string: ALL=(misinformati* OR “wrong informati*” OR disinformati* OR “misleading informati*” OR “fake news*”) AND ALL=(medic* OR illness* OR disease* OR health* OR pharma* OR drug* OR therap*) AND ALL=(“social media*” OR Facebook* OR Twitter* OR Instagram* OR YouTube* OR Weibo* OR Whatsapp* OR Reddit* OR TikTok* OR WeChat*). The PubMed database was similarly searched for papers mentioning these terms in their titles and abstracts. The search terms about misinformation and its common synonyms were referred from 2 recent systematic reviews [7,25]. No additional filter was placed to restrict the search results, and the indicated search yielded 529 papers in WoS and 285 papers in PubMed. After merging the lists from both databases and removing duplicates, 529 papers remained. Since this was a total-scale analysis of the concerned literature [26], all resultant papers were included without exclusion (Multimedia Appendix 1).

The “Analyze” function of WoS was used to provide initial descriptive statistics regarding the bibliographic data. The numbers of social media platform–specific papers were counted. The approach applied to Facebook is presented here as an example for the used evaluation strategy. In particular, we additionally searched with the following search string: ALL= Facebook* NOT (Twitter* OR Instagram* OR YouTube* OR Weibo* OR Whatsapp* OR Reddit* OR TikTok* OR WeChat*). When the original search string and this new search string were combined with the Boolean operator “AND,” the resulting papers mentioned Facebook but not the other referenced social media.

Outcome Measures

We evaluated the publication and citation counts of contributors in terms of author, institution, country, and journal. We also computed the publication and citation counts of terms and keywords, and identified the top 10 most cited papers. The semantic content of the identified publications was analyzed in the following ways. Citations per paper (CPPs) were computed for terms occurring in the titles, abstracts, and keywords of the identified papers, and n-gram analysis was conducted to identify the most recurring metatext. These analyses aimed to answer the queries listed at the end of the Introduction. Further details are described below.

Data Extraction and Main Analysis

The 529 identified papers were exported in full record with cited references to VOSviewer [27,28] for subsequent bibliometric analyses and visualizations. To visualize the results, a term map was created via VOSviewer to display publication and citation data for terms that appeared in the titles and abstracts of the analyzed papers. We decided to visualize terms that appeared in over 1% of the papers (ie, at least six papers) for improved clarity of the generated image, to avoid a heavily crowded figure [26]. A keyword map was similarly produced with the same frequency threshold, displaying author keywords and keywords added by WoS (KeyWords Plus) altogether. VOSviewer performs text mining by part-of-speech tagging with the aid of Apache OpenNLP and a linguistic filter, and converts plural noun phrases into singular form [29]. Meanwhile, it constructs a map in the following 3 steps based on a co-occurrence matrix: (1) calculation of a similarity index based on association strength (also known as proximity index and probabilistic affinity index), (2) application of the VOS mapping technique to the matrix, and (3) solution transformation to produce consistent results [27]. Besides visualizations, the resultant data from the maps were checked, and the recurring items were presented in tabular format.

In addition, keyword maps were produced for subsets of papers that were specific to Twitter, YouTube, and Facebook. For these maps, keywords with at least two appearances were included.

Exploratory Analysis

Finally, an exploratory n-gram analysis was conducted with the online NGram Analyzer [30] that allows n-gram metatexts to be listed. The abstracts of the publications were pasted into the program and the recurring 5-grams (a contiguous sequence of 5 words) were extracted. After manual checking, meaningful 5-grams with at least four appearances have been reported in the Results.


Our search strategy identified a total of 529 scientific articles addressing medical misinformation on social media. The analysis of these papers revealed that the earliest papers on this subject could be traced back to 2010 and 2011, and the total publication and citation counts increased very rapidly, especially during the last 2 years (Figure 1). Original articles accounted for the majority of the identified publications (n=393, 74.3%), followed by editorial materials (n=50, 9.5%). The article-to-review ratio was 12.7:1 (n=393 vs 31). Proceedings accounted for another 7.2% (n=38). Over 97% of the indexed papers were written in English. The most cited paper among the 529 was also the oldest paper; it involved content analysis of over 5000 relevant Tweets during the 2009 H1N1 outbreak [1]. Within a decade, it has already accumulated 589 citations.

Figure 1. Total publication and citation counts of papers on medical and health-related misinformation on social media. Data are shown until the end of 2020.
View this figure

The most productive author publishing in this subject area was Emily K Vraga from George Mason University (Virginia, USA). She started to publish on this topic in 2015 and accumulated a total of 13 papers, mostly with Leticia Bode and Melissa Tully. Leticia Bode and Joseph A Hill followed second in the list of the most productive researchers, with 9 papers each. Following them were 27 authors with 7 papers each. The top 10 most productive institutions, countries, journals, and WoS categories in which the analyzed works were published are listed in Table 1. The United States contributed to half (265/529, 50.1%) of the identified papers and was the home country of 80% of the top 10 most productive institutions. The identified papers were mostly published in journals belonging to the following categories: public environmental and occupational health, communication, health care sciences services, medical informatics, and medicine general internal.

Social media platform–specific papers accounted for 44.1% (n=233) of all 529 identified papers (Table 2). The most popularly investigated social media were Twitter, YouTube, and Facebook. They also had higher CPPs than other social media.

Table 1. Top 10 most productive institutions, countries, journals, and Web of Science categories publishing papers on medical and health-related misinformation on social media.
VariablePublication count (N=529), n (%)Citations per paper
Institution


Harvard University25 (4.7)13.2

University of Texas System20 (3.8)3.4

University of North Carolina14 (2.6)13.8

University of Pennsylvania14 (2.6)11.9

University of London13 (2.5)30.5

Johns Hopkins University12 (2.3)6.0

University of California System11 (2.1)1.7

University of Minnesota System11 (2.1)2.8

Pennsylvania Commonwealth System of Higher Education10 (1.9)3.6

University System of Sydney10 (1.9)32.1
Country


United States265 (50.1)12.2

United Kingdom53 (9.3)20.0

Italy35 (6.6)9.2

Canada33 (6.2)34.0

Spain30 (5.7)7.2

Australia27 (5.1)19.0

China27 (5.1)13.7

Turkey17 (3.2)5.1

Germany15 (2.8)27.9

India14 (2.6)2.8

Switzerland14 (2.6)14.9
Journal (2019 impact factor)


Journal of Medical Internet Research (5.034)32 (6.0)14.1

American Journal of Public Health (6.464)14 (2.6)3.1

Health Communication (1.965)13 (2.5)9.2

Vaccine (3.143)13 (2.5)28.1

International Journal of Environmental Research and Public Health (2.468)11 (2.1)8.6

PLOS One (2.740)11 (2.1)60.1

Annals of Behavioral Medicine (4.475)8a (1.5)0

Professional de la Informacion (N/Ab)8 (1.5)8.6

Cureus (N/A)6 (1.1)26.7

Journal of Health Communication (1.596)6 (1.1)2.3
Web of Science category


Public environmental and occupational health95 (18.0)12.6

Communication71 (13.4)7.3

Health care sciences services50 (9.5)17.5

Medical informatics48 (9.1)17.4

Medicine general internal38 (7.2)17.2

Computer science information systems33 (6.2)4.1

Information science library science32 (6.0)5.5

Health policy services22 (4.2)13.6

Computer science theory methods21 (4.0)5.4

Immunology21 (4.0)23.6

aAll 8 publications in Annals of Behavioral Medicine were meeting abstracts and received no citation.

bN/A: not applicable.

Table 2. Count of platform-specific papers on medical and health-related misinformation on social media.
Social mediaPublication count, nCitations per paper
Twitter9017.0
YouTube6713.7
Facebook5715.3
WhatsApp64.0
Instagram68.7
Weibo47.5
Reddit22.5
WeChat13.0
TikTok0N/Aa

aN/A: not applicable.

Figure 2 shows the terms extracted from the titles and abstracts of all 529 identified papers. COVID-19 (“covid” at the lower half, n=109, CPP=7.1) and vaccine (upper half, n=62, CPP=15.7) were 2 major health issues identified. Mentioned COVID-19 derivatives included SARS-CoV-2 (“sars cov,” n=9, CPP=11.0), coronavirus (n=22, CPP=15.3), coronavirus disease (n=25, CPP=12.6), and coronavirus pandemic (n=6, CPP=2.3). Mentioned vaccine derivatives included vaccination (n=53, CPP=21.6), vaccination rate (n=7, CPP=5.7), vaccine hesitancy (n=14, CPP=13.8), vaccine misinformation (n=13, CPP=6.9), vaccine preventable disease (n=6, CPP=29.0), and vaccine safety (n=8, CPP=7.8). The top 20 terms with the highest CPPs are listed in Table 3. Notable terms hinting on important issues discussed in the analyzed literature set were public perception, public concern, health authority, peer (related to peer-to-peer support), and policy maker (Table 3).

Figure 2. Term map showing words/phrases extracted from the titles and abstracts of the 529 papers. Circle size is related to the number of papers mentioning the word/phrase. Circle color is related to the citations per paper. The proximity between circles is related to how frequently the terms are co-mentioned in the same papers.
View this figure
Table 3. Top 20 terms with the highest citations per paper.
TermaPublication count (N=529), n (%)Citations per paper
Real time5 (0.9)160.2
Public perception7 (1.3)86.4
Credible source7 (1.3)84.9
Public concern11 (2.1)75.5
Health authority14 (2.6)56.5
Story24 (4.5)54.5
Peer11 (2.1)50.4
Adoption16 (3.0)49.3
Relevant video7 (1.3)48.1
Term34 (6.4)43.3
Sentiment19 (3.6)41.2
Illness6 (1.1)41.0
Zika virus6 (1.1)40.3
Emergency21 (4.0)38.7
Policy maker9 (1.7)38.6
Viewer7 (1.3)37.1
Misperception8 (1.5)36.5
Information source12 (2.3)36.0
Feeling6 (1.1)35.0
Potential risk7 (1.3)35.0

aOnly terms that appeared in at least 1% of papers were considered.

A keyword map is shown in Figure 3. The keyword map showed that several diseases were recurring themes of investigation, such as measles (n=9, CPP=7.7), Ebola (n=22, CPP=11.4), COVID-19 (n=87, CPP=6.4), and cardiovascular diseases (n=9, CPP=1.7). Table 4 presents the top 20 keywords with the highest CPPs, and reveals that risk and safety were among the concepts with the highest CPPs.

Figure 3. Keyword map of the 529 papers. Circle size is related to the number of papers including the word/phrase as a keyword. Circle color is related to the citations per paper. The proximity between circles is related to how frequently the terms are co-mentioned in the same papers.
View this figure
Table 4. Top 20 keywords with the highest citations per paper.
KeywordaPublication count (N=529), n (%)Citations per paper
Risk17 (3.2)51.6
Social network7 (1.3)50.7
Parents7 (1.3)41.1
Hesitancy8 (1.5)39.5
Coverage13 (2.5)37.9
Immunization13 (2.5)34.5
People7 (1.3)31.0
Web 2.014 (2.6)30.4
Knowledge13 (2.5)27.5
Medical information6 (1.1)27.5
Technology8 (1.5)27.3
Public-health8 (1.5)27.1
Attitudes7 (1.3)24.9
Vaccines19 (3.6)24.7
Videos11 (2.1)23.1
Safety7 (1.3)22.7
Care9 (1.7)20.9
Risk communication10 (1.9)20.0
China6 (1.1)19.5
Internet86 (16.3)18.6

aOnly keywords that appeared in at least 1% of papers were considered.

Keyword maps generated for the publication set that investigated Twitter, YouTube, and Facebook are shown in Figure 4A-C. The keyword maps reveal that Twitter research focused on anxiety related to online searching for disease and medical information (cyberchondria and hypochondriasis, cyan), vaccination and Zika virus (red), COVID-19 (blue), Ebola (yellow), cancer (green), and data analysis involving predictive modeling (purple). YouTube research focused on smoking and tobacco (purple), alternative medicine for various diseases, such as rheumatoid arthritis and prostate cancer (green), breast cancer (yellow), COVID-19 (red), and Ebola (blue). Finally, Facebook research focused on online health communities (yellow), vaccine hesitancy related to autism (blue), credibility of health information related to immunization and nutrition (red), cancer (purple), and COVID-19 (green).

The top 10 most cited papers are listed in Table 5. Peer-to-peer support and spread of misinformation were mentioned, and all Twitter, YouTube, and Facebook data were investigated. The themes of these top 10 papers were consistent with the list of highly cited terms listed in Table 3, covering topics such as peer-to-peer support, online videos, and public perception.

The exploratory n-gram analysis resulted in several meaningful 5-gram metatexts with at least four appearances as follows: 6 appearances, “as a source of information” and “the spread of fake news;” 5 appearances, “rumors stigma and conspiracy theories,” “the quality of health information,” and “content reliability and quality of;” 4 appearances, “health anxiety and health literacy,” “the relationship between message factors,” “intentions to trust and share,” “#PedsICU and coronavirus disease 2019,” “in low- and middle-income countries,” “actions for a framework for,” “interacted with perceived message importance,” and “verify and share the message.”

Figure 4. Keyword maps of the papers investigating (A) Twitter, (B) YouTube, and (C) Facebook. Circle size is related to the number of papers mentioning the respective word/phrase as a keyword. Circle color is related to the clustering of the words by the default setting of VOSviewer. The proximity between circles is related to how frequently the terms are co-mentioned in the same papers.
View this figure
Table 5. Top 10 most cited papers on medical and health-related misinformation on social media.
Authors, yearCitations, n
Chew et al, 2010 [1]613
Yaqub et al, 2014 [31]256
Naslund et al, 2016 [32]212
Madathil et al, 2015 [33]195
Kamel Boulos et al, 2011 [34]186
Betsch et al, 2012 [35]168
Syed-Abdul et al, 2013 [36]147
Depoux et al, 2020 [37]136
Singh et al, 2012 [38]123
Bode et al, 2015 [39]121

General Discussion

Using bibliometric analysis, this study identified and quantitatively analyzed 529 papers on medical and health-related misinformation on social media, revealing the most popularly investigated social media platforms, prevailing research themes, most utilized journals, and most productive countries, institutions, and authors.

Findings Concerning the Western World and Its Prevalent Social Media Platforms

This bibliometric analysis on 529 scientific articles concerning medical and health-related misinformation on social media revealed that the most heavily investigated platforms were Twitter, YouTube, and Facebook. This could be related to the finding that most of the top 10 productive countries were from North America and Europe where these social media platforms were dominant. The results also confirmed the hypothesis that the United States had the largest contribution in social media research. The total publication and citation counts increased very rapidly especially during the last 2 years, consistent with the trends identified by previous systematic reviews on this topic [7,21]. On the other hand, this study found that original articles accounted for 74.3% of the analyzed literature set. This implied that one-fourth of the literature was not covered by the abovementioned systematic reviews, which might partly explain some differences in the results. For instance, this study found that Twitter was the most recurring social medium in the literature instead of YouTube, as reported by Wang et al [7]. The strength of the review by Suarez-Lledo and Alvarez-Galvez [21] was that it analyzed the prevalence of health misinformation posts (0%-98%) reported in the original articles. Meanwhile, Wang et al [7] categorized them into predefined theoretical frameworks with the most prevalent ones being public health, health policy, and epidemiology (n=14); health informatics (n=8); communications studies (n=5); vaccines (n=4); cyberpsychology (n=3); and system sciences (n=3). Here, it was found that publications in immunology were on average more frequently cited than communication and computer science, whereas health care sciences and medical informatics papers were in-between. This implies that the more published disciplines do not necessarily warrant more citations. This finding has 2 implications. First, quantity may not necessarily mean quality. Second, field differences in citation rates found in general [40] remained present even when the literature set was confined to a single focus on misinformation. Similar to the current findings, Wang et al [7] also found that the most popular topics were vaccination, Ebola, and Zika virus, with other less popular focus topics being nutrition, cancer, and smoking. In contrast, Wang et al [7] identified fluoridation of water as one of the recurring topics, whereas in this study, COVID-19 emerged as a strong research focus. Moreover, the performed keyword analysis in this work revealed that the research on different social media platforms, such as Twitter, YouTube, and Facebook, focused on different topics. While, at present, we do not have an explanation for this interesting observation, we believe that the reasoning for different topic studies on different social media could be a relevant direction for future research. Such studies may elucidate whether this is due to different prevalences of specific content across the platforms or due to preferential academic interest from research teams with particular interest in specific social media platforms or topics.

Findings Concerning China and Its Prevalent Social Media Platforms

Though some social media platforms were not available in China, China still made it into the top 10 list of the most productive countries (Table 1). With a large population of internet users in China, it could be reasonably expected that Weibo and WeChat, which are popular social media platforms in China, would become more investigated in future studies. One potential barrier to non-Chinese researchers would be content translation, as the majority of their content is written in Chinese. In addition, the fast-growing short video platform TikTok (and its Chinese version Douyin) might also exert significant influence on the health information seeking behavior of internet users in the future. However, TikTok videos might be hard to archive, and video analysis tools might not be as well developed as text mining tools, which might hinder analysis by public health researchers. The same applies to the visual contents posted on Instagram. Current findings seem to suggest that sufficient research on misinformation disseminating through these platforms is missing from the current literature and should be addressed in future research. Readers should be aware that the publication bias toward Europe and North America, especially the United States, indicates that the current body of knowledge might not reflect the wider spectrum of misinformation on global health issues, especially in other parts of the world with large online communities, such as Asia and South America.

The most productive author was found to be Emily K Vraga who is based in the United States. Her studies focused on how to correct health misinformation (in other words, overturn subjects’ misperceptions) dispersed on social media, particularly Facebook and Twitter [39,41-43]. Though China was among the top 10 most productive countries, we found that only 2 of the top 50 most productive authors were based in China. They were King-Wa Fu (n=4) and Chung-Hong Chan (n=3) from the University of Hong Kong, and they focused solely on the Ebola virus [44-47]. This implied that, to grasp the research foci from China, readers need to refer to diverse works from multiple authors instead of that from a few prominent authors. With the continued growth of netizens in China, we anticipate that more productive authors might be based in China in the future.

Elaboration on the Recurring Themes of the Literature

A very important role of social media is to provide peer-to-peer support, as investigated by some publications identified in this study (see Tables 3 and 5), for example, by forming online health communities and support groups, and ensuring stakeholder access to the latest and most relevant scientific information and health interventions [32]. For instance, users could post supportive comments and advice to YouTube videos uploaded by individuals with mental health issues [48]. On the other hand, misinformation spread via social media (especially related to Twitter, see Figure 4A) might lead to cyberchondria (cyberspace-related hypochondria; the unfounded concern escalation about common symptomology based on information found on the internet), with a study revealing that unverified information might be more easily shared by internet users who trust online information and perceived an information overload, and that women were more likely to experience cyberchondria [49]. Cyberchondria could be an important health concern, as a meta-analysis established the correlation between health anxiety and both online information seeking and cyberchondria [50], and another work revealed that it had 5 facets, including reassurance seeking and mistrust of medical professionals [51]. Being flooded by online misinformation would not alleviate the situation but may worsen it. A recent study found that government and professional videos containing solely factual information only accounted for 11% of all YouTube videos on COVID-19 and 10% of views, whereas 28% of the analyzed videos contained misinformation and had up to 62 million views [52]. In this context, the adequacy of funding and resources allocated by governmental bodies to online health literacy campaigns needs to be questioned.

Meanwhile, from the perspective of policy makers, the large amount of information from social media can be monitored and used for the achievement of efficient outcomes. As seen from the results, “health policy services” was among the most recurring journal categories for the analyzed literature set (Table 1) and “policy maker” was one of the recurring terms with the highest CPP (Table 3). For instance, by analyzing tweets related to the COVID-19 pandemic, researchers could identify the top 4 concerns of Twitter users, namely the origin, sources, impact on various levels, and ways to reduce the infection risk [53]. While using similar approaches, keeping these concerns anonymized at an individual level and ensuring that social or ethnic groups expressing specific concerns do not become targets of discrimination are crucial. Authorities could therefore focus on these concerns as they derive measures and disseminate information to the public to contain the pandemic and reduce fears within the community. In this regard, authorities could collaborate with scientific associations and provide incentives to civil society to address ignorance or misinformation on the detected concerns. Future research could compare relevant social media content following interventions, to define the optimal strategies of tackling misinformation on social media.

As mentioned in the Introduction, the fraudulent study linking the MMR vaccine to autism still has a lingering influence on social media, as it is still posted on the Facebook platform by antivaccine advocates, despite its retraction due to fraudulency [54]. Moreover, it was found that the content posted by Facebook users regarding vaccination has been increasingly polarized [20]. One study suggested that the use of angry language could promote the viral spread of the messages, including the misinformation of vaccines causing autism [55], though it was not investigating contents exclusive to vaccines and only binarized words into positive and negative emotions. Summarizing the results from n-gram analysis, netizens might need to be aware of fake news, rumors, stigma, and conspiracy theories circulating on the internet. Content reliability and quality should be assessed, and information should be verified before sharing. One way to cohort authoritative or accurate health care information shared by experts on social media (eg, Twitter) is by the use of a hashtag, so that others can search easily. One example was #PedsICU that promoted pediatric critical care content, as found by n-gram analysis. By sensible collaboration, there may be a chance to mitigate misinformation.

Limitations

Any publications in journals not indexed by WoS were missed in the current analysis. For example, there is a relevant paper investigating misinformation of COVID-19 on TikTok, which is not indexed by WoS [56]. Besides, WoS mainly indexes papers written in English. There may be papers investigating Weibo and WeChat written in Chinese or published in Chinese journals that are not yet indexed by WoS. Preprints are also not indexed in WoS, which could be an important source of preliminary information, but the reliability of such information is debatable due to the lack of peer-review assessment. Moreover, a bibliometric study cannot assess the scientific quality of the content, such as risk of bias, effect size or statistical significance of the results, and whether the conclusions are justified by the respective data reported. The accuracy of data tagging by the literature database could also pose a limitation. For instance, KeyWords Plus are keywords tagged to a paper by an algorithm used by WoS based on the terms from the titles of the cited references [57], and are more broadly descriptive and therefore applicable to analyzing the structure of scientific fields [58]. However, it was unclear how accurate they were compared to other tags such as the National Center for Biotechnology Information’s Medical Subject Headings (“MeSH terms”). Future studies should also incorporate “conspiracy theory” and related terms into their search protocols for more comprehensive results.

Future Research

For potential future research, artificial intelligence (AI) applications for social media content analysis would be an especially promising avenue. With increasing content and misinformation circulating on social media, it becomes practically impossible to manually determine and classify misinformation. AI or machine learning might be employed for such content analysis, which has the potential to achieve high accuracy [59]. Yet, AI could also be exploited to generate and disseminate misinformation to targeted audiences [60]. AI research in health care was most frequently published in journals in computer science and engineering, as reported by Guo et al [61], and indeed, among their identified top 10 most productive journals, only PLOS One was on the list in our study (Table 1). Along this line, with the further development of AI applications for social media content analysis, it might also be of interest to promote the dissemination of such research in mainstream public health journals, in order to reach a broader relevant audience.

Conclusions

Based on an analysis of 529 papers on medical and health-related misinformation on social media, we found that the United States contributed to half of the papers, with 80% of the top 10 most productive institutions being based in this country. The papers were mostly published in journals belonging to the categories public environmental and occupational health, communication, health care sciences services, medical informatics, and medicine general internal. However, they were generally less cited than papers published in immunology, suggesting that more publications did not warrant more citations. Social media platform–specific papers accounted for 44% of all papers. The most popularly investigated social media platforms were Twitter, YouTube, and Facebook. They also had higher CPPs than other social media. Investigations on these platforms had different foci. Twitter-based research investigated cyberchondria and hypochondriasis, YouTube-based research investigated tobacco smoking, and Facebook-based research investigated vaccine hesitancy related to autism. COVID-19 was a common topic investigated across all platforms. An important implication of these findings is that often knowledge on specific themes related to medical misinformation relies on the predominant study of a single social media platform or limited number of platforms, and broader cross-platform studies could be a promising direction for future research. Future studies should also include social platforms aimed at non-English users to provide a wider perspective on global health misinformation.

Conflicts of Interest

None declared.

Multimedia Appendix 1

PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analysis) flow diagram of the literature search.

PNG File , 49 KB

  1. Chew C, Eysenbach G. Pandemics in the age of Twitter: content analysis of Tweets during the 2009 H1N1 outbreak. PLoS One 2010 Nov 29;5(11):e14118 [FREE Full text] [CrossRef] [Medline]
  2. Kouzy R, Abi Jaoude J, Kraitem A, El Alam MB, Karam B, Adib E, et al. Coronavirus Goes Viral: Quantifying the COVID-19 Misinformation Epidemic on Twitter. Cureus 2020 Mar 13;12(3):e7255 [FREE Full text] [CrossRef] [Medline]
  3. Jang SM, Mckeever BW, Mckeever R, Kim JK. From Social Media to Mainstream News: The Information Flow of the Vaccine-Autism Controversy in the US, Canada, and the UK. Health Commun 2019 Jan 13;34(1):110-117. [CrossRef] [Medline]
  4. Househ M, Borycki E, Kushniruk A. Empowering patients through social media: the benefits and challenges. Health Informatics J 2014 Mar 18;20(1):50-58 [FREE Full text] [CrossRef] [Medline]
  5. Hagg E, Dahinten VS, Currie LM. The emerging use of social media for health-related purposes in low and middle-income countries: A scoping review. Int J Med Inform 2018 Jul;115:92-105 [FREE Full text] [CrossRef] [Medline]
  6. Haber N, Smith ER, Moscoe E, Andrews K, Audy R, Bell W, CLAIMS research team. Causal language and strength of inference in academic and media articles shared in social media (CLAIMS): A systematic review. PLoS One 2018 May 30;13(5):e0196346 [FREE Full text] [CrossRef] [Medline]
  7. Wang Y, McKee M, Torbica A, Stuckler D. Systematic Literature Review on the Spread of Health-related Misinformation on Social Media. Soc Sci Med 2019 Nov;240:112552 [FREE Full text] [CrossRef] [Medline]
  8. Racovita M. Lost in translation. Scientists need to adapt to a postmodern world; constructivism can offer a way. EMBO Rep 2013 Aug 16;14(8):675-678 [FREE Full text] [CrossRef] [Medline]
  9. Hill J, Agewall S, Baranchuk A, Booz GW, Borer JS, Camici PG, et al. Medical Misinformation: Vet the Message!. J Am Heart Assoc 2019 Feb 05;8(3):e011838 [FREE Full text] [CrossRef] [Medline]
  10. Wakefield AJ, Murch SH, Anthony A, Linnell J, Casson DM, Malik M, et al. Ileal-lymphoid-nodular hyperplasia, non-specific colitis, and pervasive developmental disorder in children. Lancet 1998 Feb 28;351(9103):637-641. [CrossRef] [Medline]
  11. No Authors. A timeline of the Wakefield retraction. Nat Med 2010 Mar;16(3):248-248. [CrossRef]
  12. Dyer C. Lancet retracts Wakefield's MMR paper. BMJ 2010 Feb 02;340(feb02 4):c696-c696. [CrossRef] [Medline]
  13. Godlee F, Smith J, Marcovitch H. Wakefield's article linking MMR vaccine and autism was fraudulent. BMJ 2011 Jan 05;342(jan05 1):c7452-c7452. [CrossRef] [Medline]
  14. Omer SB. The discredited doctor hailed by the anti-vaccine movement. Nature 2020 Oct 27;586(7831):668-669. [CrossRef]
  15. Hotez PJ, Nuzhath T, Colwell B. Combating vaccine hesitancy and other 21st century social determinants in the global fight against measles. Curr Opin Virol 2020 Apr;41:1-7. [CrossRef] [Medline]
  16. Gunaratne K, Coomes EA, Haghbayan H. Temporal trends in anti-vaccine discourse on Twitter. Vaccine 2019 Aug 14;37(35):4867-4871. [CrossRef] [Medline]
  17. Gandhi CK, Patel J, Zhan X. Trend of influenza vaccine Facebook posts in last 4 years: a content analysis. Am J Infect Control 2020 Apr;48(4):361-367. [CrossRef] [Medline]
  18. Dunn AG, Leask J, Zhou X, Mandl KD, Coiera E. Associations Between Exposure to and Expression of Negative Opinions About Human Papillomavirus Vaccines on Social Media: An Observational Study. J Med Internet Res 2015 Jun 10;17(6):e144 [FREE Full text] [CrossRef] [Medline]
  19. Chou WS, Oh A, Klein WMP. Addressing Health-Related Misinformation on Social Media. JAMA 2018 Dec 18;320(23):2417-2418. [CrossRef] [Medline]
  20. Schmidt AL, Zollo F, Scala A, Betsch C, Quattrociocchi W. Polarization of the vaccination debate on Facebook. Vaccine 2018 Jun 14;36(25):3606-3612. [CrossRef] [Medline]
  21. Suarez-Lledo V, Alvarez-Galvez J. Prevalence of Health Misinformation on Social Media: Systematic Review. J Med Internet Res 2021 Jan 20;23(1):e17187 [FREE Full text] [CrossRef] [Medline]
  22. Pritchard A. Statistical bibliography or bibliometrics. Journal of Documentation 1969;25(4):348-349.
  23. Huang Z, Chen H, Liu Z. The 100 top-cited systematic reviews/meta-analyses in central venous catheter research: A PRISMA-compliant systematic literature review and bibliometric analysis. Intensive Crit Care Nurs 2020 Apr;57:102803. [CrossRef] [Medline]
  24. Moher D, Liberati A, Tetzlaff J, Altman DG, PRISMA Group. Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. PLoS Med 2009 Jul 21;6(7):e1000097 [FREE Full text] [CrossRef] [Medline]
  25. Celliers M, Hattingh M. A Systematic Review on Fake News Themes Reported in Literature. In: Hattingh M, Matthee M, Smuts H, Pappas I, Dwivedi YK, Mäntymäki M, editors. Responsible Design, Implementation and Use of Information and Communication Technology. I3E 2020. Lecture Notes in Computer Science, vol 12067. Cham: Springer; 2020:223-234.
  26. Atanasov AG, Yeung AWK, Klager E, Eibensteiner F, Schaden E, Kletecka-Pulker M, et al. First, Do No Harm (Gone Wrong): Total-Scale Analysis of Medical Errors Scientific Literature. Front Public Health 2020 Oct 16;8:558913 [FREE Full text] [CrossRef] [Medline]
  27. van Eck NJ, Waltman L. Software survey: VOSviewer, a computer program for bibliometric mapping. Scientometrics 2010 Aug 31;84(2):523-538 [FREE Full text] [CrossRef] [Medline]
  28. Waltman L, van Eck NJ, Noyons EC. A unified approach to mapping and clustering of bibliometric networks. Journal of Informetrics 2010 Oct;4(4):629-635. [CrossRef]
  29. Van Eck N, Waltman L. Text mining and visualization using VOSviewer. ISSI Newsletter 2011;7:50-54.
  30. Online NGram Analyzer. Guide to Data Mining.   URL: http://guidetodatamining.com/ngramAnalyzer/ [accessed 2022-01-11]
  31. Yaqub O, Castle-Clarke S, Sevdalis N, Chataway J. Attitudes to vaccination: a critical review. Soc Sci Med 2014 Jul;112:1-11 [FREE Full text] [CrossRef] [Medline]
  32. Naslund JA, Aschbrenner KA, Marsch LA, Bartels SJ. The future of mental health care: peer-to-peer support and social media. Epidemiol Psychiatr Sci 2016 Jan 08;25(2):113-122. [CrossRef]
  33. Madathil KC, Rivera-Rodriguez AJ, Greenstein JS, Gramopadhye AK. Healthcare information on YouTube: A systematic review. Health Informatics J 2015 Sep;21(3):173-194 [FREE Full text] [CrossRef] [Medline]
  34. Kamel Boulos MN, Resch B, Crowley DN, Breslin JG, Sohn G, Burtner R, et al. Crowdsourcing, citizen sensing and sensor web technologies for public and environmental health surveillance and crisis management: trends, OGC standards and application examples. Int J Health Geogr 2011 Dec 21;10:67 [FREE Full text] [CrossRef] [Medline]
  35. Betsch C, Brewer NT, Brocard P, Davies P, Gaissmaier W, Haase N, et al. Opportunities and challenges of Web 2.0 for vaccination decisions. Vaccine 2012 May 28;30(25):3727-3733. [CrossRef] [Medline]
  36. Syed-Abdul S, Fernandez-Luque L, Jian WS, Li YC, Crain S, Hsu MH, et al. Misleading health-related information promoted through video-based social media: anorexia on YouTube. J Med Internet Res 2013 Feb 13;15(2):e30 [FREE Full text] [CrossRef] [Medline]
  37. Depoux A, Martin S, Karafillakis E, Preet R, Wilder-Smith A, Larson H. The pandemic of social media panic travels faster than the COVID-19 outbreak. J Travel Med 2020 May 18;27(3):taaa031 [FREE Full text] [CrossRef] [Medline]
  38. Singh AG, Singh S, Singh PP. YouTube for information on rheumatoid arthritis--a wakeup call? J Rheumatol 2012 May;39(5):899-903. [CrossRef] [Medline]
  39. Bode L, Vraga EK. In Related News, That Was Wrong: The Correction of Misinformation Through Related Stories Functionality in Social Media. J Commun 2015 Jun 23;65(4):619-638. [CrossRef]
  40. Bornmann L, Haunschild R, Mutz R. Should citations be field-normalized in evaluative bibliometrics? An empirical analysis based on propensity score matching. Journal of Informetrics 2020 Nov;14(4):101098. [CrossRef]
  41. Bode L, Vraga EK. See Something, Say Something: Correction of Global Health Misinformation on Social Media. Health Commun 2018 Sep 16;33(9):1131-1140. [CrossRef] [Medline]
  42. Vraga EK, Bode L. Using Expert Sources to Correct Health Misinformation in Social Media. Science Communication 2017 Sep 14;39(5):621-645. [CrossRef]
  43. Vraga EK, Bode L. I do not believe you: how providing a source corrects health misperceptions across social media platforms. Information, Communication & Society 2017 Apr 19;21(10):1337-1353. [CrossRef]
  44. Fung IC, Duke CH, Finch KC, Snook KR, Tseng P, Hernandez AC, et al. Ebola virus disease and social media: A systematic review. Am J Infect Control 2016 Dec 01;44(12):1660-1671. [CrossRef] [Medline]
  45. Fung I, Fu K, Chan C, Chan B, Cheung C, Abraham T, et al. Social Media's Initial Reaction to Information and Misinformation on Ebola, August 2014: Facts and Rumors. Public Health Rep 2016;131(3):461-473 [FREE Full text] [CrossRef] [Medline]
  46. Fung IC, Fu K, Chan C, Chan BSB, Cheung C, Abraham T, et al. Social Media's Initial Reaction to Information and Misinformation on Ebola, August 2014: Facts and Rumors. Public Health Rep 2016 May 01;131(3):461-473 [FREE Full text] [CrossRef] [Medline]
  47. Liang H, Fung IC, Tse ZTH, Yin J, Chan C, Pechta LE, et al. How did Ebola information spread on twitter: broadcasting or viral spreading? BMC Public Health 2019 Apr 25;19(1):438 [FREE Full text] [CrossRef] [Medline]
  48. Naslund JA, Grande SW, Aschbrenner KA, Elwyn G. Naturally occurring peer support through social media: the experiences of individuals with severe mental illness using YouTube. PLoS One 2014 Oct 15;9(10):e110171 [FREE Full text] [CrossRef] [Medline]
  49. Laato S, Islam AKMN, Islam MN, Whelan E. What drives unverified information sharing and cyberchondria during the COVID-19 pandemic? European Journal of Information Systems 2020 Jun 07;29(3):288-305. [CrossRef]
  50. McMullan RD, Berle D, Arnáez S, Starcevic V. The relationships between health anxiety, online health information seeking, and cyberchondria: Systematic review and meta-analysis. J Affect Disord 2019 Feb 15;245:270-278. [CrossRef] [Medline]
  51. McElroy E, Shevlin M. The development and initial validation of the cyberchondria severity scale (CSS). J Anxiety Disord 2014 Mar;28(2):259-265. [CrossRef] [Medline]
  52. Li HO, Bailey A, Huynh D, Chan J. YouTube as a source of information on COVID-19: a pandemic of misinformation? BMJ Glob Health 2020 May 14;5(5):e002604 [FREE Full text] [CrossRef] [Medline]
  53. Abd-Alrazaq A, Alhuwail D, Househ M, Hamdi M, Shah Z. Top Concerns of Tweeters During the COVID-19 Pandemic: Infoveillance Study. J Med Internet Res 2020 Apr 21;22(4):e19016 [FREE Full text] [CrossRef] [Medline]
  54. Shelby A, Ernst K. Story and science: how providers and parents can utilize storytelling to combat anti-vaccine misinformation. Hum Vaccin Immunother 2013 Aug 27;9(8):1795-1801 [FREE Full text] [CrossRef] [Medline]
  55. Bail CA. Emotional Feedback and the Viral Spread of Social Media Messages About Autism Spectrum Disorders. Am J Public Health 2016 Jul;106(7):1173-1180. [CrossRef]
  56. Basch C, Hillyer G, Jaime C. COVID-19 on TikTok: harnessing an emerging social media platform to convey important public health messages. Int J Adolesc Med Health 2020 Aug 10:2020. [CrossRef] [Medline]
  57. Garfield E, Sher IH. KeyWords Plus™—algorithmic derivative indexing. J. Am. Soc. Inf. Sci 1993 Jun;44(5):298-299. [CrossRef]
  58. Zhang J, Yu Q, Zheng F, Long C, Lu Z, Duan Z. Comparing keywords plus of WOS and author keywords: A case study of patient adherence research. J Assn Inf Sci Tec 2015 Jan 08;67(4):967-972. [CrossRef]
  59. Choudrie J, Banerjee S, Kotecha K, Walambe R, Karende H, Ameta J. Machine learning techniques and older adults processing of online information and misinformation: A covid 19 study. Comput Human Behav 2021 Jun;119:106716 [FREE Full text] [CrossRef] [Medline]
  60. Wilner AS. Cybersecurity and its discontents: Artificial intelligence, the Internet of Things, and digital misinformation. International Journal 2018 Jul 26;73(2):308-316. [CrossRef]
  61. Guo Y, Hao Z, Zhao S, Gong J, Yang F. Artificial Intelligence in Health Care: Bibliometric Analysis. J Med Internet Res 2020 Jul 29;22(7):e18228 [FREE Full text] [CrossRef] [Medline]


AI: artificial intelligence
CPPs: citations per paper
MMR: measles, mumps, and rubella
WoS: Web of Science


Edited by A Mavragani; submitted 23.02.21; peer-reviewed by K Reuter, C Weiger, A Roundtree, JP Allem; comments to author 16.03.21; revised version received 30.03.21; accepted 20.12.21; published 25.01.22

Copyright

©Andy Wai Kan Yeung, Anela Tosevska, Elisabeth Klager, Fabian Eibensteiner, Christos Tsagkaris, Emil D Parvanov, Faisal A Nawaz, Sabine Völkl-Kernstock, Eva Schaden, Maria Kletecka-Pulker, Harald Willschke, Atanas G Atanasov. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 25.01.2022.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.