Published on in Vol 26 (2024)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/50353, first published .
Investigation of Public Acceptance of Misinformation Correction in Social Media Based on Sentiment Attributions: Infodemiology Study Using Aspect-Based Sentiment Analysis

Investigation of Public Acceptance of Misinformation Correction in Social Media Based on Sentiment Attributions: Infodemiology Study Using Aspect-Based Sentiment Analysis

Investigation of Public Acceptance of Misinformation Correction in Social Media Based on Sentiment Attributions: Infodemiology Study Using Aspect-Based Sentiment Analysis

Authors of this article:

Ning Ma1 Author Orcid Image ;   Guang Yu1 Author Orcid Image ;   Xin Jin2 Author Orcid Image

Original Paper

1School of Management, Harbin Institute of Technology, Harbin, China

2School of Social Sciences, Harbin Institute of Technology, Harbin, China

Corresponding Author:

Guang Yu, Prof Dr

School of Management

Harbin Institute of Technology

92 West Dazhi Street

Harbin, 150001

China

Phone: 86 139 3649 0774

Email: yug@hit.edu.cn


Background: The proliferation of misinformation on social media is a significant concern due to its frequent occurrence and subsequent adverse social consequences. Effective interventions for and corrections of misinformation have become a focal point of scholarly inquiry. However, exploration of the underlying causes that affect the public acceptance of misinformation correction is still important and not yet sufficient.

Objective: This study aims to identify the critical attributions that influence public acceptance of misinformation correction by using attribution analysis of aspects of public sentiment, as well as investigate the differences and similarities in public sentiment attributions in different types of misinformation correction.

Methods: A theoretical framework was developed for analysis based on attribution theory, and public sentiment attributions were divided into 6 aspects and 11 dimensions. The correction posts for the 31 screened misinformation events comprised 33,422 Weibo posts, and the corresponding Weibo comments amounted to 370,218. A pretraining model was used to assess public acceptance of misinformation correction from these comments, and the aspect-based sentiment analysis method was used to identify the attributions of public sentiment response. Ultimately, this study revealed the causality between public sentiment attributions and public acceptance of misinformation correction through logistic regression analysis.

Results: The findings were as follows: First, public sentiments attributed to external attribution had a greater impact on public acceptance than those attributed to internal attribution. The public associated different aspects with correction depending on the type of misinformation. The accuracy of the correction and the entity responsible for carrying it out had a significant impact on public acceptance of misinformation correction. Second, negative sentiments toward the media significantly increased, and public trust in the media significantly decreased. The collapse of media credibility had a detrimental effect on the actual effectiveness of misinformation correction. Third, there was a significant difference in public attitudes toward the official government and local governments. Public negative sentiments toward local governments were more pronounced.

Conclusions: Our findings imply that public acceptance of misinformation correction requires flexible communication tailored to public sentiment attribution. The media need to rebuild their image and regain public trust. Moreover, the government plays a central role in public acceptance of misinformation correction. Some local governments need to repair trust with the public. Overall, this study offered insights into practical experience and a theoretical foundation for controlling various types of misinformation based on attribution analysis of public sentiment.

J Med Internet Res 2024;26:e50353

doi:10.2196/50353

Keywords



Background

Scholarly interest in researching misinformation dissemination has been extensive [1,2], particularly given the exacerbation of this issue since the COVID-19 pandemic. Such misinformation encompasses a range of forms, including the spread of health misinformation [3], fake news [4], conspiracy theories [5], and other types of COVID-19 misinformation [6]. The spread of misinformation can significantly hinder individuals’ access to accurate information and may exacerbate psychological distress [7]. In addition, the public may lose faith in the government, leading to increased fear and anxiety [8]. The resulting infodemic has heightened concerns about the adverse effects of misinformation [9,10] and prompted greater research attention toward misinformation and infodemic governance [11].

Research on how to mitigate the spread of misinformation has resulted in several recommended interventions [12]. The intervention and prevention strategies for misinformation are also key research points [13]. Fact checking was believed to be an effective approach in reducing the spread of misinformation, but it often requires extensive coordination to improve public resistance [14]. Applying warning tags to social media posts could also potentially reduce sharing behaviors [15], but the credibility of untagged posts could be potentially enhanced [16]. Inaccurate tagging can also aggravate the spread of misinformation, making warning tags ineffective [17]. Another effective approach is to post correction information that directly addresses false facts or claims [18], with authoritative experts playing a prominent role in correction communication [9,19,20]. Using social media platforms and the structure of social networks to block misinformation is also an effective approach [21], and correcting information in real time through social media platforms is another strategy [22].

Preventive measures are also necessary to combat misinformation among the public [12]. The most common framework for implementing preventive interventions is the psychological inoculation theory [23]; it had lasting effects in responding to misinformation about COVID-19 [24]. Furthermore, educational preventive measures are thought to be the most effective [25,26], and personal awareness of misinformation is equally significant [27]. To implement reliable preventive measures on a large scale, we need more professionals with advanced knowledge and convergence of many efforts to yield reliable results [12]. Therefore, government policy support is essential [28], and a clear legal framework on misinformation needs to be established promptly [29]. However, both debunking [30] and prebunking, which refers to prophylactic interventions [23], have limitations and may not be as effective as anticipated [12]. While scholars have proposed suggestions to improve the correction effect [31], sometimes misinformation correction can even backfire [32].

Although researchers have made concerted efforts to correct misinformation on social networks, the results have been suboptimal in many cases [33]. This not only perpetuates misinformation but also risks public resentment [34], undermining the persuasive power of subsequent corrections. Current research on evaluating the effectiveness of misinformation correction through social media has analyzed the impact of various post characteristics on correction outcomes [35]. As emotional information can provide valuable insights into understanding the public perception of misinformation [36-39], more research is needed on how public sentiment can be leveraged to analyze the reasons for poor correction effectiveness. Analyzing public sentiment can help identify public attitudes toward misinformation [34,40], including the effectiveness of correction. Considering the limitations regarding objectivity and timeliness of using questionnaires to assess public attitudes toward misinformation and correction [41], social media data can be used more objectively to analyze public acceptance of misinformation correction [42]. Governments can also use public sentiment analysis to combat misinformation and inappropriate sharing behaviors [43].

Objectives

It is vital to understand how to evaluate public acceptance of misinformation correction using public sentiment analysis on social media. To address this research goal, this study focused on misinformation correction during the COVID-19 pandemic and mining the key factors influencing public acceptance of correction. Specifically, the following three research questions were answered in this study:

  1. How to characterize public acceptance of misinformation correction and different aspects of public sentiment? (Research question 1)
  2. What are the differences in the impact of various public sentiment attributions on the correction of different types of misinformation? (Research question 2)
  3. Which attribution of public sentiment is more critical to shaping public acceptance of correction of different types of misinformation? (Research question 3)

To comprehensively explore the aforementioned questions, this study collected public interaction data on the Weibo platform pertaining to important misinformation dissemination events related to COVID-19 since the outbreak. First, aspect-based sentiment analysis (ABSA) was performed to attribute public sentiments to multiple dimensions drawing on attribution theory. Second, the study explored how public acceptance of misinformation correction related to public sentiment attributions in different events from the public perspective. Finally, the significant degree of attribution results obtained through public sentiment response revealed the specific reasons for public attitudes. Consequently, this study provides theoretical support and practical evidence for developing correction strategies for misinformation, which could positively impact the formulation of correction communication measures.


Research Design and Framework

This study proposed a research framework, as depicted in Figure 1, based on attribution theory to evaluate public acceptance of misinformation correction by analyzing the attribution results of public sentiment response to correction.

Figure 1. Research and analysis framework of public acceptance of misinformation correction based on attribution theory. The framework includes the specific steps of data collection, model training, causality analysis, and other processes.

The underlying reasons for the ineffectiveness of misinformation correction as perceived by the public were elucidated by this framework. By examining various types of misinformation dissemination events, we categorized and analyzed key public sentiment attributions that contribute to public acceptance of misinformation correction in different situations. Furthermore, we considered public views that are not directly related to correction and whose role in analyzing the efficacy of correction cannot be ignored.

Event Selection and Data Collection

Weibo is a prominent social media platform in China that has been extensively studied in relation to public opinion [42]. One noteworthy feature of Weibo is the Weibo Refuting Rumor account, which serves to clarify and correct rumors that circulate on the web. This study used the posts from the Weibo Refuting Rumors account to screen 31 misinformation events related to COVID-19 that occurred between January 26, 2020, and April 14, 2022. The corresponding correction posts and their comments were obtained through data support from Zhiwei Data Sharing Platform. In total, the correction posts for the 31 events comprised 33,422 Weibo posts, and the corresponding Weibo comments amounted to 370,218. This study analyzed these comments to investigate the attributions that influenced public attitudes regarding misinformation correction.

Current research on misinformation, particularly related to COVID-19, focuses on categorizing it based on content. One such classification method through fact checking divided misinformation into 6 categories, which included treatment, conspiracy, government measures, vaccines, number of cases, and others [38]. Drawing from this categorization, this study categorized 31 misinformation events related to COVID-19 into 4 main categories through fact checking and examining the public opinion on social networks in China: prevention and treatment, conspiracy, government measures, and COVID-19 development. (Specific event information and classification can be found in Multimedia Appendix 1.)

Main Components of Sentiment Attributions in Misinformation Correction

Public sentiment responses to misinformation correction can originate from various sources [44]. According to attribution theory, public perception of misinformation correction is influenced by how they attribute the causes of their sentiment response. This attribution process involves 3 dimensions: locus, stability, and controllability [45]. When the public encounters misinformation correction, the attribution of public sentiment response ultimately falls on the issue of locus of causality. At this point, whether the public sentiment responses are attributed to internal or external attributions becomes the key of the research [46,47].

This study categorized the sources of public sentiment based on the key elements of misinformation into 4 categories: the background of the event, the channel of information, the information itself, and the recipient [34]. Regarding actual misinformation correction, the background of the event corresponds to the aspect of misinformation, the channel of information refers to the communication channel aspect, the information itself pertains to the correction aspect, and the recipient is the public. The internal attribution of public sentiment pertains to the public’s own experience as the recipient, whereas other aspects of public sentiment fall under external attribution. Public sentiment responses attributed to these different aspects are associated with public acceptance of misinformation correction. Furthermore, the influence of the government on public sentiment is a crucial factor, and its response to misinformation also affects the public sentiment response [28]. According to the aforementioned theoretical analysis, public sentiments were divided into fine-grained sentiments in 6 aspects and 11 dimensions, which are detailed in Table 1 and reflect the attributions of public sentiment response.

Table 1. The specific description and classification of sentiment attribution into aspects.
Locus, aspect, and dimensionSpecific description
External

Correction


MCBCorrection action against misinformation


MDDebunker of misinformation

Misinformation


MMOThe main object of misinformation events


MSBThe spread and manufacturing of misinformation


DOMDisseminators and producers of misinformation

Communication channels


CMCommunication media and celebrities

Government


GAOfficial government and public authority representative


LGLocal government


TMTreatment and response measures in misinformation events
Internal

Experience


SPPersonal feelings and experience
No attribution

Repost


REPRepost

Ethical Considerations

The comment data used for this study are publicly accessible on the Weibo platform, and none of the data contained any details that could be used to track individual users. The authors consulted with their respective institutional review boards and were informed that no approvals were necessary.

Computation of Public Acceptance and Public Sentiments

To evaluate public attitudes toward misinformation correction in the comments, it is crucial to extract public acceptance of the correction. As the text data used in this study were in Chinese, the Bidirectional Encoder Representations from Transformers (BERT) pretraining model based on the Chinese pretraining model of the Social Computing and Information Retrieval Research Center from the Harbin Institute of Technology was used to assess public opinions [48,49]. Specifically, comments were labeled as tag 1 if the public recognized the correction and as tag 0 if the public considered the correction unreliable or ineffective. In the labeling process, 2 professional researchers randomly selected 10,000 comments from all the processed Weibo comments and labeled them based on whether the public recognized the correction. The tagged text was then cross-validated by 2 professional researchers. After labeling, the data were split into validation and training sets in a 1:9 ratio, with labels of 1 representing public acceptance and 0 representing public nonacceptance. The pretraining model was then adjusted to the use environment and used for model training. The model had a good performance, and it is presented in Table 2. It was used to assess public acceptance of the correction in the remaining comments.

Table 2. Model performance of Bidirectional Encoder Representations from Transformers (BERT) and aspect-based sentiment analysis (ABSA) on different indicators.
Model and effectiveness indicatorPerformance (%)
BERT_Roberta

Accuracy83.21

Recall81.55

F1-score80.71
ABSA

APC_accuracya87.27

APC_f1_scoreb87.18

ATE_accuracyc89.08

ATE_f1_scored88.87

aAPC_accuracy: ABSA development has progressed effectively. The accuracy of aspect polarity category.

bAPC_f1_score: The F1-score of aspect polarity category.

cATE_accuracy: The accuracy of aspect term extraction.

dATE_f1_score: The F1-score of aspect term extraction.

To identify the specific target objects of public sentiments in each text and analyze the attribution of public sentiment response, this study used the aspect-based sentiment computing method. ABSA is a method to analyze the sentiments of different aspects of an object in text. Unlike traditional sentiment classification, which only provides the comprehensive sentiment polarity of the text, ABSA focuses on both target extraction and target sentiment classification, making it a more fine-grained sentiment analysis method [50,51]. ABSA development has progressed effectively and has greatly improved the performance of the model, especially Chinese ABSA [52]. The most common application scenario for ABSA is in content analysis of restaurant and hotel reviews [53], where it completed the task of fine-grained sentiment analysis with 5 aspects and 18 dimensions [54]. Recently, ABSA has also been applied to extract themes and analyze relevant discourse on social media during COVID-19 [55], and other scholars have conducted aspect-based fine-grained sentiment analysis on user interaction data on social media about country image during the COVID-19 pandemic [56].

This study used ABSA involving 2 tasks: aspect term extraction and aspect polarity classification [52]. To adapt to the language environment of misinformation on Weibo, the corpus labeling rules were adjusted, and public sentiments were classified into negative and nonnegative categories. The texts were labeled according to the classification in Table 1. In total, 2 professional researchers labeled the corpus, resulting in a total of 17,629 corpora through cross-validation, which were divided into validation and training sets with a ratio of 1:9 for model training. The model performance is shown in Table 2. This model was then used to predict aspect sentiments in the remaining comments and investigate the specific attributions of public sentiments.

Causal Analysis of Public Acceptance of Correction

This study used logistic regression analysis to investigate the particular attributions that impact public acceptance of misinformation correction. The dependent variable for this analysis was public acceptance of misinformation correction, whereas the independent variables were the number of negative and nonnegative sentiments in the text across 11 dimensions. Moreover, the study controlled for gender, number of fans, number of likes, geographical location, and user authentication identity type by incorporating them as control variables in the following logistic regression model:

Z = ln(p/[1 – p]) = b0 + b1MCB_non +b2MCB_neg + b3MCB_non + b4MD_neg + b5MMO_non + b6MMO_neg + b7MSB_non + b8MSB_neg + b9DOM_non + b10DOM_neg + b11CM_non + b12CM_neg + b13GA_non + b14GA_neg + b15LG_non + b16LG_neg + b17TM_non + b18TM_neg + b19SP_non + b20SP_neg + b21REP + b22GEN + b23FAN + b24LIKE + b25LOC + b26TYPE + ε0,

Where p was the probability that the public thought the correction was valid, 1 – p was the probability that the public thought the correction was invalid, bi was the regression coefficient of each variable, b0 was the constant term, and e0 was the random disturbance term. The specific variable descriptions are listed in Textbox 1. As the aspect sentiment calculation for each comment was based on the number of negative or nonnegative sentiments present within that comment, each variable representing aspect sentiment included both negative and nonnegative dimensions.

Textbox 1. Variable names and description of logistic regression analysis.

Variable and description

  • MCB_non: number of nonnegative sentiments about correction action against misinformation
  • MCB_neg: number of negative sentiments about correction action against misinformation
  • MD_non: number of nonnegative sentiments about debunker of misinformation
  • MD_neg: number of negative sentiments about debunker of misinformation
  • MMO_non: number of nonnegative sentiments about the main object of misinformation events
  • MMO_neg: number of negative sentiments about the main object of misinformation events
  • MSB_non: number of nonnegative sentiments about the spread and manufacturing of misinformation
  • MSB_neg: number of negative sentiments about the spread and manufacturing of misinformation
  • DOM_non: number of nonnegative sentiments about disseminators and producers of misinformation
  • DOM_neg: number of negative sentiments about disseminators and producers of misinformation
  • CM_non: number of nonnegative sentiments about communication media and celebrities
  • CM_neg: number of negative sentiments about communication media and celebrities
  • GA_non: number of nonnegative sentiments about the official government and public authority representatives
  • GA_neg: number of negative sentiments about the official government and public authority representatives
  • LG_non: number of nonnegative sentiments about local government
  • LG_neg: number of negative sentiments about local government
  • TM_non: number of nonnegative sentiments about treatment and response measures in misinformation events
  • TM_neg: number of negative sentiments about treatment and response measures in misinformation events
  • SP_non: number of nonnegative sentiments about personal feelings and experience
  • SP_neg: number of negative sentiments about personal feelings and experience
  • REP: repost
  • GEN: gender
  • FAN: number of fans
  • LIKE: number of likes
  • LOC: the geographical location of the user’s registration
  • TYPE: user authentication identity type

This study examined the primary public sentiment attributes that influenced public acceptance of misinformation correction posted on social media based on theoretical hypotheses and conceptual models. From a public sentiment perspective, this study extracted variables from various aspects while controlling for user-specific attributes such as gender, number of fans, number of likes, and geographical location. The aim was to analyze the attributions of public sentiments related to public acceptance of misinformation correction. Thus, this study sought to identify why the public perceived correction to be ineffective and provided theoretical support for relevant correction communication methods.

Differences in Public Sentiment Attributions Among the Corrections of Different Types of Misinformation

We constructed a regression model to analyze which variables had a significant impact on public acceptance of misinformation correction. However, it remained to be tested whether the influence of the same variable on different types of misinformation could be compared through the regression coefficients. In this study, we categorized the misinformation events into 4 types: prevention and treatment (type P), conspiracy (type C), government measures (type G), and COVID-19 development (type D). The 4 types of misinformation constituted 4 subsample groups, and we tested the differences in coefficients between groups for all variables except for the control variables using coefficient tests based on the seemingly unrelated regression. Existing methods can only test the coefficient difference between 2 groups of samples at a single time, so we had to test the coefficient difference among the 4 subsample groups in pairs. This approach enabled us to assess the difference in the role of a certain aspect of public sentiment in different types of misinformation, providing strategic support for the implementation of correction communication.


Multidimensional Public Sentiments in Misinformation Correction

We conducted statistical analysis on public acceptance of misinformation correction in all comments and public sentiment response to different aspects, including the control variables, for the entire sample data. Table 3 presents the descriptive statistics of the overall sample data. As GEN, LOC, and TYPE were categorical variables and LOC and TYPE had numerous categories, the specific count of each category is not listed in Table 3.

Simple repost aspect was considered as the repost behavior after recognizing the correction in this study. However, Table 3 reveals that the average level of public acceptance of misinformation correction was 0.508, indicating that slightly over half of the public perceived the correction as effective. This finding suggests that the impact of correction on the public was not sufficiently robust. In addition, the statistics of aspect sentiments demonstrated that the number of aspect sentiments expressed in each text varied. Negative sentiments toward the dimensions of government treatment and response measures could appear up to 12 times in a single comment. Some comments did not even express sentiments about a particular dimension and might have solely been reposts.

Upon examining public acceptance of misinformation correction by event type, we identified differences in the level of acceptance across different types of misinformation. As depicted in Figure 2, the correction of type P garnered the highest public acceptance, whereas the correction of types G and D had the lowest acceptance. However, public acceptance of correction of type C fluctuated greatly, suggesting that the effect of correction for type C was not consistent. Moreover, Figure 2 does not reveal any distinct pattern regarding public acceptance of correction based on the sequence of events. Therefore, overall, corrections did not yield fixed and effective outcomes.

The findings presented in Figure 3 indicate that the comments on correction contained more sentiments regarding the aspect of misinformation, creating a noticeable gap between this and the other aspects. Aside from simple reposts, the number of sentiments in the remaining aspects was nearly equal. Furthermore, the composition of sentiments in each aspect of correction significantly varied across different types of misinformation, as evidenced by the proportion of sentiments toward correction depicted in Figure 3. For instance, comments on the correction of type G displayed significantly more sentiment toward the government compared to comments on the correction of other types. These observations indicated that there were variations in public acceptance of the correction of different types of misinformation, with marked differences in the composition of sentiments across specific aspects. Hence, conducting differentiated analyses on the correction of different types of misinformation was necessary to identify the specific reasons for poor public acceptance of correction.

Figure 4 depicts the co-occurrence network of public aspect sentiments in the correction of different types of misinformation, which highlights the disparities in public expression. The co-occurrence network illustrates the co-occurrence of different dimensions of aspect sentiments in public expression. The Jaccard coefficient served as the index for drawing the co-occurrence network, where all edges with a Jaccard coefficient of >0.1 were displayed. Our analysis revealed that the most central aspect sentiment was not necessarily the most frequent among different types of misinformation, and the central aspect sentiment in each type of misinformation network differed. For instance, in type C, the aspect sentiment of the LG dimension was central, whereas the aspect sentiment of MSB was central in type D. The aspect sentiment of MCB was central in type G, and the aspect sentiment of GA was central in type P. These differences in co-occurrence networks suggested the need for differentiated analyses of comments to uncover the reasons why the correction was not more recognized by the public.

Table 3. Descriptive statistics of the variables.
VariableObservations, nValues, mean (SD; range)
Attitudes370,2180.508 (0.5; 0-1)
MCB_nona370,2180.028 (0.199; 0-8)
MCB_negb370,2180.036 (0.213; 0-8)
MD_nonc370,2180.007 (0.1; 0-3)
MD_negd370,2180.002 (0.04; 0-3)
MMO_none370,2180.116 (0.377; 0-5)
MMO_negf370,2180.066 (0.286; 0-7)
MSB_nong370,2180.017 (0.135; 0-5)
MSB_negh370,2180.063 (0.275; 0-5)
DOM_noni370,2180.007 (0.088; 0-3)
DOM_negj370,2180.047 (0.244; 0-6)
CM_nonk370,2180.034 (0.2; 0-7)
CM_negl370,2180.035 (0.22; 0-6)
GA_nonm370,2180.019 (0.154; 0-6)
GA_negn370,2180.014 (0.127; 0-5)
LG_nono370,2180.013 (0.122; 0-6)
LG_negp370,2180.009 (0.101; 0-7)
TM_nonq370,2180.024 (0.167; 0-4)
TM_negr370,2180.029 (0.181; 0-12)
SP_nons370,2180.042 (0.209; 0-3)
SP_negt370,2180.026 (0.164; 0-4)
REPu370,2180.153 (0.362; 0-3)
GENv370,2181.485 (0.516; 1-3)
FANw370,21835,210.49 (1,458,389; 0-2.31×108)
LIKEx370,2187.264 (484.681; 0-173,209)
LOCy370,21820.388 (9.079; 1-36)
TYPEz370,2185.732 (4.674; 1-11)

aMCB_non: number of nonnegative sentiments about correction action against misinformation.

bMCB_neg: number of negative sentiments about correction action against misinformation.

cMD_non: number of nonnegative sentiments about debunker of misinformation.

dMD_neg: number of negative sentiments about debunker of misinformation.

eMMO_non: number of nonnegative sentiments about the main object of misinformation events.

fMMO_neg: number of negative sentiments about the main object of misinformation events.

gMSB_non: number of nonnegative sentiments about the spread and manufacturing of misinformation.

hMSB_neg: number of negative sentiments about the spread and manufacturing of misinformation.

iDOM_non: number of nonnegative sentiments about disseminators and producers of misinformation.

jDOM_neg: number of negative sentiments about disseminators and producers of misinformation.

kCM_non: number of nonnegative sentiments about communication media and celebrities.

lCM_neg: number of negative sentiments about communication media and celebrities.

mGA_non: number of nonnegative sentiments about the official government and public authority representatives.

nGA_neg: number of negative sentiments about the official government and public authority representatives.

oLG_non: number of nonnegative sentiments about local government.

pLG_neg: number of negative sentiments about local government.

qTM_non: number of nonnegative sentiments about treatment and response measures in misinformation events.

rTM_neg: number of negative sentiments about treatment and response measures in misinformation events.

sSP_non: number of nonnegative sentiments about personal feelings and experience.

tSP_neg: number of negative sentiments about personal feelings and experience.

uREP: repost.

vGEN: gender.

qFAN: number of fans.

xLIKE: number of likes.

yLOC: the geographical location of the user’s registration.

zTYPE: user authentication identity type.

Figure 2. The proportion of public acceptance of correction. These 31 misinformation events related to COVID-19 occurred between January 26, 2020, and April 14, 2022. (A) Box plot of the proportion of public acceptance of correction in each type of misinformation. (B) Histogram of the proportion of public acceptance of correction in each misinformation event.
Figure 3. The composition of aspect sentiments in the correction of different types of misinformation. These 31 misinformation events were divided into 4 types: prevention and treatment (type P), conspiracy (type C), government measures (type G), and COVID-19 development (type D). (A) The number of public sentiments in different aspects. (B) The proportion of public sentiments in different aspects.
Figure 4. The attribution co-occurrence network of public aspect sentiment in the correction of different types of misinformation. (A) The co-occurrence network of type C. (B) The co-occurrence network of type D. (C) The co-occurrence network of type G. (D) The co-occurrence network of type P. CM: communication media and celebrities; DOM: disseminators and producers of misinformation; GA: official government and public authority representatives; LG: local government; MCB: correction action against misinformation; MMO: main object of misinformation events; MSB: spread and manufacturing of misinformation; SP: personal feelings and experience; TM: treatment and response measures in misinformation events.

Causality Between Public Acceptance of Misinformation Correction and Attributions of Public Sentiment

The overall sample data were regressed at the event level, and the significance of aspect sentiments for each dimension of all events was calculated, as shown in Figure 5. The significance was normalized to obtain the relative significance intensity of sentiments in each dimension. Figure 5 shows that the repost aspect had a significant relationship with public acceptance of misinformation correction. Moreover, a certain degree of regularity in the relative significance intensity of different types of misinformation was revealed. Although there were some differences in the significance of sentiment in each dimension in each event, the results demonstrated that public sentiments in different dimensions were significantly related to whether the public perceived the correction as effective.

Figure 5. The relative significance intensity of different aspects of public sentiment. The redder the color block, the higher the relative significance intensity. (A) The relative significance intensity in each misinformation event. (B) The relative significance intensity in different types of misinformation. The relative significance intensity in different types of misinformation showed a unique distribution, so the research needed to consider the differences and common characteristics of the different types of misinformation. CM_neg: number of negative sentiments about communication media and celebrities; CM_non: number of nonnegative sentiments about communication media and celebrities; DOM_neg: number of negative sentiments about disseminators and producers of misinformation; DOM_non: number of nonnegative sentiments about disseminators and producers of misinformation; GA_neg: number of negative sentiments about the official government and public authority representatives; GA_non: number of nonnegative sentiments about the official government and public authority representatives; LG_neg: number of negative sentiments about local government; LG_non: number of nonnegative sentiments about local government; MCB_neg: number of negative sentiments about correction action against misinformation; MCB_non: number of nonnegative sentiments about correction action against misinformation; MD_neg: number of negative sentiments about debunker of misinformation; MD_non: number of nonnegative sentiments about debunker of misinformation; MMO_neg: number of negative sentiments about the main object of misinformation events; MMO_non: number of nonnegative sentiments about the main object of misinformation events; MSB_neg: number of negative sentiments about the spread and manufacturing of misinformation; MSB_non: number of nonnegative sentiments about the spread and manufacturing of misinformation; REP: repost; SP_neg: number of negative sentiments about personal feelings and experience; SP_non: number of nonnegative sentiments about personal feelings and experience; TM_neg: number of negative sentiments about treatment and response measures in misinformation events; TM_non: number of nonnegative sentiments about treatment and response measures in misinformation events.

We randomly selected two-thirds of the overall sample data for grouped regression based on 4 different types. Table 4 shows the regression results of public acceptance of misinformation correction for various types of misinformation and each corresponding variable. The variables LIKE, LOC, and TYPE were included as control variables but were found to be insignificant across different types. While FAN was found to be significant, its coefficient was too small to be considered a reference, and only GEN had a certain impact on public acceptance. Therefore, the influence of gender was always controlled in the grouped regression experiment. In addition, we only considered the causality between public aspect sentiment and public acceptance of correction.

Table 4. Regression coefficients for different types of misinformation.
VariableType P regression coefficientType C regression coefficientType G regression coefficientType D regression coefficient
MCB_nona1.232b1.168b1.15b1.542b
MCB_negc–0.495b–1.033b–0.816b–0.17d
MD_none1.855b1.183b1.523b1.766b
MD_negf–0.191–0.652b–0.4810.428
MMO_nong–0.514b–0.209b–0.798b–0.918b
MMO_negh–1.147b–0.596b–1.746b–0.137
MSB_noni–0.352b–0.24b–0.222j–0.339j
MSB_negk2.011b0.934b1.636b1.606b
DOM_nonl–0.48b–0.418b–0.477b–0.223
DOM_negm0.162b0.373b0.103j0.34b
CM_nonn–0.106b0.255b0.347b0.1
CM_nego0.723b0.651b0.886b1.693b
GA_nonp0.574b1.093b0.0750.872b
GA_negq–0.601b–0.982b–0.915b–0.902b
LG_nonr–0.042–0.71b–0.597b0.052
LG_negs–1.219b–1.46b–1.508b–0.243
TM_nont–0.431b–0.221b–0.5b–0.086
TM_negu–0.368b–0.384b–0.939b–2.461b
SP_nonv0.276b1.099b0.347b0.499b
SP_negw–1.064b–0.389b–0.003–0.316b
REPx2.933b2.92b2.729b2.807b
GENy–0.2b–0.017d0.188b0.301b
FANz0b0b0b0b
LIKEaa00d00
LOCab–0.0010.0010–0.002
TYPEac–0.0030–0.0010.001
Constant–0.005–0.113b–1.052b–1.404b

aMCB_non: number of nonnegative sentiments about correction action against misinformation.

bP<.01.

cMCB_neg: number of negative sentiments about correction action against misinformation.

dP<.10.

eMD_non: number of nonnegative sentiments about debunker of misinformation.

fMD_neg: number of negative sentiments about debunker of misinformation.

gMMO_non: number of nonnegative sentiments about the main object of misinformation events.

hMMO_neg: number of negative sentiments about the main object of misinformation events.

iMSB_non: number of nonnegative sentiments about the spread and manufacturing of misinformation.

jP<.05.

kMSB_neg: number of negative sentiments about the spread and manufacturing of misinformation.

lDOM_non: number of nonnegative sentiments about disseminators and producers of misinformation.

mDOM_neg: number of negative sentiments about disseminators and producers of misinformation.

nCM_non: number of nonnegative sentiments about communication media and celebrities.

oCM_neg: number of negative sentiments about communication media and celebrities.

pGA_non: number of nonnegative sentiments about the official government and public authority representatives.

qGA_neg: number of negative sentiments about the official government and public authority representatives.

rLG_non: number of nonnegative sentiments about local government.

sLG_neg: number of negative sentiments about local government.

tTM_non: number of nonnegative sentiments about treatment and response measures in misinformation events.

uTM_neg: number of negative sentiments about treatment and response measures in misinformation events.

vSP_non: number of nonnegative sentiments about personal feelings and experience.

wSP_neg: number of negative sentiments about personal feelings and experience.

xREP: repost.

yGEN: gender.

zFAN: number of fans.

aaLIKE: number of likes.

abLOC: the geographical location of the user’s registration.

acTYPE: user authentication identity type.

All aspects of sentiments in each dimension corresponded to human factors. Therefore, it can be proven that, regardless of the type of misinformation used in this study, public sentiments influenced by human factors affect public acceptance. For SP attributed to internal attribution, it influenced all types. Public acceptance of misinformation correction exhibited a positive correlation with positive public sentiments attributed to personal experience while displaying a negative correlation with negative public sentiments attributed to personal experience. However, its coefficient was not strong in all regression results. On the other hand, for the other dimensions of external attribution, sentiments played a significant role in the regression results. This finding indicated that public sentiments attributed to external attribution had a greater impact on public acceptance than public sentiments attributed to internal attribution regardless of the correction of different types of misinformation.

Regarding correction aspect, public acceptance of misinformation correction was positively correlated with the nonnegative sentiments of MCB, negatively correlated with the negative sentiments of MCB, and positively correlated with the nonnegative sentiments of MD. In addition, the negative sentiments of MD in type C were negatively correlated with public acceptance. For the sentiments of misinformation aspect, the regression coefficients of positive and negative sentiments of the 3 dimensions were integrated. The results supported that public acceptance exhibited a negative correlation with positive sentiments attributed to misinformation while displaying a positive correlation with negative sentiments attributed to misinformation. Regarding the sentiments of communication channels, only in type C and type G public acceptance of misinformation correction was shown to be positively correlated with the nonnegative sentiments of CM. However, in all types, public acceptance was positively correlated with the negative sentiments of CM. Finally, in the government aspect, public acceptance was only negatively correlated with the negative sentiments of GA, positively correlated with the nonnegative sentiments of GA, and negatively correlated with aspect sentiments on both of the other dimensions.

Different Effectiveness of Aspect Sentiments in the Correction of Different Types of Misinformation

To reveal which dimension of sentiments played a greater role in correcting different types of misinformation, it was necessary to compare the regression coefficients of the 4 subsample groups. Therefore, it was necessary to test the difference in coefficients between groups. After pairwise testing among the 4 subsample groups, the results are presented in Table 5. The differences in coefficients between the 2 groups that passed the test are marked with footnotes in the table. Most of these coefficients could be directly compared with the regression coefficients in the grouped regression results.

Table 5. Test for differences in coefficients between groups.
VariableP value (type D and type C)P value (type G and type C)P value (type P and type C)P value (type D and type G)P value (type D and type P)P value (type G and type P)
MCB_nona.01b.84.46.02b.06c.50
MCB_negd<.001e.04b<.001e<.001e.008e.006e
MD_nonf.001e.09c.003e.44.69.30
MD_negg.05c.63.15.21.40.57
MMO_nonh<.001e<.001e<.001e.28<.001e<.001e
MMO_negi<.001e<.001e<.001e<.001e<.001e<.001e
MSB_nonj<.001e<.001e.13.42<.001e<.001e
MSB_negk<.001e<.001e<.001e.72<.001e<.001e
DOM_nonl.13<.001e.51.001e.09c<.001e
DOM_negm.81<.001e<.001e.02b<.001e<.001e
CM_nonn.15.22<.001e.06c.08c<.001e
CM_nego<.001e.001e.12<.001e<.001e.02b
GA_nonp.19<.001e<.001e<.001e.04b<.001e
GA_negq.48.38.001e.93.17.06c
LG_nonr<.001e.17<.001e<.001e.07c<.001e
LG_negs<.001e.81.28<.001e<.001e.27
TM_nont.38<.001e.08c.006e.06c.52
TM_negu<.001e<.001e.82<.001e<.001e<.001e
SP_nonv<.001e<.001e<.001e.18<.001e<.001e
SP_negw.64<.001e<.001e.002e<.001e<.001e
REPx<.001e<.001e.02b.06c<.001e.009e

aMCB_non: number of nonnegative sentiments about correction action against misinformation.

bP<.05.

cP<.10.

dMCB_neg: number of negative sentiments about correction action against misinformation.

eP<.01.

fMD_non: number of nonnegative sentiments about debunker of misinformation.

gMD_neg: number of negative sentiments about debunker of misinformation.

hMMO_non: number of nonnegative sentiments about the main object of misinformation events.

iMMO_neg: number of negative sentiments about the main object of misinformation events.

jMSB_non: number of nonnegative sentiments about the spread and manufacturing of misinformation.

kMSB_neg: number of negative sentiments about the spread and manufacturing of misinformation.

lDOM_non: number of nonnegative sentiments about disseminators and producers of misinformation.

mDOM_neg: number of negative sentiments about disseminators and producers of misinformation.

nCM_non: number of nonnegative sentiments about communication media and celebrities.

oCM_neg: number of negative sentiments about communication media and celebrities.

pGA_non: number of nonnegative sentiments about the official government and public authority representatives.

qGA_neg: number of negative sentiments about the official government and public authority representatives.

rLG_non: number of nonnegative sentiments about local government.

sLG_neg: number of negative sentiments about local government.

tTM_non: number of nonnegative sentiments about treatment and response measures in misinformation events.

uTM_neg: number of negative sentiments about treatment and response measures in misinformation events.

vSP_non: number of nonnegative sentiments about personal feelings and experience.

wSP_neg: number of negative sentiments about personal feelings and experience.

xREP: repost.

Combining the analyses in Tables 4 and 5, we observed that, excluding the effect of repost, in type P, the aspect sentiment with the largest effect was the negative sentiments of MSB followed by the nonnegative sentiments of MD and the nonnegative sentiments of MCB. The negative sentiments of LG and the negative sentiments of MMO also played a significant role in determining whether the public recognized the correction. This finding suggested that, compared with the correction of other types of misinformation, public sentiments about the spread and manufacturing of misinformation in type P played a more significant role. In type C, the negative sentiments of LG were the most influential aspect sentiments, whereas MCB and GA had large regression coefficients, whether positive or negative. This phenomenon indicated that, in type C, public sentiments toward the correction aspect as well as the government aspect played a more significant role. In type G, the negative sentiments of MMO had a more significant effect. Public attention and discussion on the misinformation aspect had a significant impact on the correction of type G. Furthermore, the role of local government in this type of misinformation needed to be considered as public perception of local government also played a significant role. In type D, the most significant impact on misinformation correction was from the negative sentiments of TM, with the negative sentiments of CM also playing a vital role. Notably, the causality between public acceptance of misinformation correction and negative sentiments of TM was apparent. The regression coefficient of negative sentiments of CM was also the highest among the 4 types of misinformation. This showed that, for the correction of type D, public discussions about treatment and response measures of the government and communication channels better reflected public acceptance of correction.

Robustness Test

Finally, to ensure the robustness and reliability of the experimental results and conclusions, the remaining third of the sample data was randomly selected from the entire sample and regrouped according to the 4 different types of misinformation for further testing. The results of the robustness test are shown in Table 6.

The test results showed only minor changes in regression coefficients and significance, which did not alter the overall results and conclusions of the experiment. On the basis of this, it could be concluded that the experimental results of this study were robust and reliable.

Table 6. Robustness test results using different data.
VariableType P regression coefficientType C regression coefficientType G regression coefficientType D regression coefficient
MCB_nona1.123b1.194b1.153b1.493b
MCB_negc–0.567b–1.02b–0.854b–0.212d
MD_none1.813b1.158b1.615b1.773b
MD_negf–0.307–0.629b0.0560.006
MMO_nong–0.51b–0.212b–0.782b–1.024b
MMO_negh–1.148b–0.62b–1.786b–0.065
MSB_noni–0.344b–0.252b–0.375j0.204
MSB_negk1.941b0.919b1.518b1.574b
DOM_nonl–0.492b–0.456b–0.483b–0.039
DOM_negm0.137b0.393b0.1d0.272b
CM_nonn–0.0440.231b0.407b0.1
CM_nego0.705b0.636b0.916b1.638b
GA_nonp0.556b1.121b0.0240.653b
GA_negq–0.411b–0.939b–0.826b–1.045b
LG_nonr–0.096–0.686b–0.424b0.145
LG_negs–1.205b–1.687b–1.326b–0.206
TM_nont–0.496b–0.238b–0.515b0.002
TM_negu–0.398b–0.384b–0.881b–2.421b
SP_nonv0.266b1.086b0.342b0.471b
SP_negw–1.054b–0.373b0.043–0.194d
REPx2.884b2.906b2.735b2.818b
GENy–0.238b–0.0150.203b0.325b
FANz0j0b0b0b
LIKEaa0000
LOCab–0.0010.0010–0.003
TYPEac–0.0020.00100.001
Constant0.038–0.12b–1.058b–1.398b

aMCB_non: number of nonnegative sentiments about correction action against misinformation.

bP<.01.

cMCB_neg: number of negative sentiments about correction action against misinformation.

dP<.10.

eMD_non: number of nonnegative sentiments about debunker of misinformation.

fMD_neg: number of negative sentiments about debunker of misinformation.

gMMO_non: number of nonnegative sentiments about the main object of misinformation events.

hMMO_neg: number of negative sentiments about the main object of misinformation events.

iMSB_non: number of nonnegative sentiments about the spread and manufacturing of misinformation.

jP<.05.

kMSB_neg: number of negative sentiments about the spread and manufacturing of misinformation.

lDOM_non: number of nonnegative sentiments about disseminators and producers of misinformation.

mDOM_neg: number of negative sentiments about disseminators and producers of misinformation.

nCM_non: number of nonnegative sentiments about communication media and celebrities.

oCM_neg: number of negative sentiments about communication media and celebrities.

pGA_non: number of nonnegative sentiments about the official government and public authority representatives.

qGA_neg: number of negative sentiments about the official government and public authority representatives.

rLG_non: number of nonnegative sentiments about local government.

sLG_neg: number of negative sentiments about local government.

tTM_non: number of nonnegative sentiments about treatment and response measures in misinformation events.

uTM_neg: number of negative sentiments about treatment and response measures in misinformation events.

vSP_non: number of nonnegative sentiments about personal feelings and experience.

wSP_neg: number of negative sentiments about personal feelings and experience.

xREP: repost.

yGEN: gender.

zFAN: number of fans.

aaLIKE: number of likes.

abLOC: the geographical location of the user’s registration.

acTYPE: user authentication identity type.


Different Types of Misinformation Require Different Correction Communication

The study findings indicated that there were differences in public acceptance of correction among different types of misinformation and the key attributions that affect public acceptance also differed among them. However, regardless of the number of public discussions or the regression results, external attributions were the most significant among the corrections of 4 types of misinformation. The experimental results indicated that personal feelings and experience had an impact but were not the most important. The attributions in the aspects of correction and misinformation most directly influenced public acceptance of correction. While the media and government were the clear targets of most public sentiment responses, public attitudes toward these 2 aspects of attributions were significantly different from those toward other attributions. Compared with existing literature on misinformation correction effectiveness [2], this study provided a new analytical perspective.

In the correction of prevention and treatment, the regression results indicated that the public disliked misinformation and expected timely and accurate correction, with the government playing a crucial role. Implementers of correction and clarification of misinformation should be chosen carefully. In the correction of conspiracy, the findings illustrated that the public was dissatisfied with both the correction communication method and the implementers. Especially when the government confirmed and corrected misinformation, the public expressed negative sentiments toward local governments more frequently, so the way and timing of government involvement in the correction process needs to be considered prudently. Representatives with greater credibility could play a better role in the correction process [20]. In the correction of government measures, public negative sentiments toward misinformation revealed their attention to specific government measures. This showed that, in addition to the negative impact of misinformation, the public also sought to eliminate various inconveniences caused by the pandemic, reflecting their expectations of the government. However, in the correction of COVID-19 development, the study found that the public paid more attention to government treatment and response measures. The public considered the impact of treatment measures on themselves or whether the measures met their physical or psychological needs. The spread of this type of misinformation also stemmed from public concern about the uncertainty risk [57].

However, previous studies have emphasized that corrections are different depending on the nature [2,58]. The current approach to correction communication is too simplistic and repetitive, leading to public apathy and backlash [33]. It is necessary to adopt different methods of correcting misinformation to prevent stress reactions or aversion to correction and, ultimately, prevent the public from being exposed to misinformation [34]. Public demand for correction measures varies depending on the differences in misinformation types. The investigation into the specific attributions of public acceptance of misinformation correction is valuable as it uses the aspect sentiment analysis method to not only determine public acceptance but also uncover the underlying attributions affecting it. Ultimately, the fundamental issue is that the public must trust the corrections they receive, which requires flexible and diverse methods of communication and correction channels that the public trusts.

The Collapse of Media Credibility

The media play a supporting role in public acceptance of misinformation correction, but the effectiveness of the media in this regard is currently questionable. Public distrust of communication channels, particularly the media, was evident across all types of misinformation. Negative sentiments about communication channels were positively correlated with public acceptance of correction, indicating that the media were not perceived as trustworthy by the public. Media failure to gain public trust had resulted in a negative impact on the effectiveness of corrections. Discussions involving authoritative experts and celebrities were the only aspects of communication channels that generated positive sentiments.

In contrast, the public was more likely to trust authoritative experts. Analysis of comments revealed that one of the main reasons for this distrust was the tendency of some media outlets to prioritize speed over verification, leading to the spread of rumors and misinformation. Therefore, the public expressed their distrust of both traditional media and “we-media.” The collapse of media credibility has reduced the persuasive power of correction spread by the media, thereby diminishing the effectiveness of correction. Enhancing the role of communication channels in promoting public acceptance of corrections might require the support of authoritative experts [20]. Therefore, the media need to work toward rebuilding their image in the public impression to regain lost credibility and effectively support the spread and effect of misinformation correction.

The Crucial Role of the Government

The government plays a crucial role in public acceptance of misinformation correction. Public sentiments toward correction of different types of misinformation also raised significant concerns regarding government involvement. However, experimental findings on public aspect sentiments toward the government demonstrated that only positive sentiments of the GA dimension were positively correlated with public acceptance. Both positive and negative sentiments of the LG and TM dimensions were negatively correlated with public acceptance. This phenomenon suggested that, when assessing public sentiments regarding the government, only the GA dimension had a positive guiding effect.

Although the public still maintained trust in the official government image, expectations of the government persisted. Concerning local governments and treatment and response measures, public trust in some local governments was eroded due to the pandemic, aligning with previous research findings [8]. Many treatment and response measures also failed to meet public expectations. Nevertheless, there is no doubt that the government plays a decisive role in both managing the pandemic and mitigating the impact of misinformation on the public [28,59]. Consequently, to enhance communication efficacy and foster public government engagement, trust between the government and the public must be restored. The pivotal role of the government in shaping public acceptance of misinformation correction cannot be overstated as it also exerts a positive influence on other aspects. Nonetheless, based on the results of this study, efforts are needed to solidify the official government image, rebuild trust in local governments, and implement comprehensive response measures to fully leverage the government’s crucial role.

Limitations and Directions for Future Research

It is important to note that the data used in this study solely originated from user interactions on the Weibo platform. As a result, the analysis of user behaviors and causality can only be applied within the context of Weibo. Therefore, further verification is necessary to generalize the research findings and conclusions to other platforms. Furthermore, the data used in this study specifically pertained to rumors surrounding the COVID-19 pandemic, and they did not encompass rumors unrelated to the pandemic. Consequently, the extension of conclusions and potential variations in other types of misinformation warrant exploration in future research.

Although this study incorporated several control variables and strived to focus on the role of aspect sentiments, it is important to acknowledge that there might be variables associated with personal characteristics that were not accessible within the existing data. This study primarily examined group behavior and might have limitations in accounting for the influence of individual traits. In addition, data on individuals who do not disclose their opinions cannot be obtained, which is a limitation of using social media data for research. As the data were from Chinese social media, more data are needed to verify the expansibility of relevant conclusions to countries with other cultural backgrounds in future research. This study will help identify effective methods for correcting misinformation that are not limited to the pandemic.

Conclusions

Misinformation correction often falls short in effectively reaching and persuading the public, necessitating a thorough investigation into the reasons influencing public acceptance of such efforts. This study identified crucial sentiment attributions that shape public acceptance of correction regarding different types of misinformation. It highlighted the role of aspect sentiments in analyzing public acceptance of misinformation correction.

The findings revealed that a flexible, issue-specific approach is required for correcting different types of misinformation. The correction strategies also require flexible communication tailored to public sentiment attribution. The decline in media credibility negatively impacted the effectiveness of misinformation correction. Regulating media behavior and rebuilding media image are essential for regaining public trust. Furthermore, the government plays a critical role; in particular, some local governments need to regain the trust of the public, enabling government-led correction communication to be more effective.

This study identified key reasons contributing to the difference in public acceptance of misinformation correction, focusing on public sentiment attributions. It provided an empirical and methodological foundation for addressing misinformation governance to some extent.

Acknowledgments

This research was funded by the National Natural Science Foundation of China (grants 72074060 and 72374056).

Data Availability

The data sets generated during and analyzed during this study are available from the corresponding author on reasonable request.

Authors' Contributions

NM proposed the original conceptualization. All authors contributed to methodology, validation, and writing.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Details and types of misinformation-spreading events.

DOCX File , 19 KB

  1. Wang Y, McKee M, Torbica A, Stuckler D. Systematic literature review on the spread of health-related misinformation on social media. Soc Sci Med. Nov 2019;240:112552. [FREE Full text] [CrossRef] [Medline]
  2. Aikin KJ, Betts KR, O'Donoghue AC, Rupert DJ, Lee PK, Amoozegar JB, et al. Correction of overstatement and omission in direct-to-consumer prescription drug advertising. J Commun. Jun 23, 2015;65(4):596-618. [CrossRef]
  3. Suarez-Lledo V, Alvarez-Galvez J. Prevalence of health misinformation on social media: systematic review. J Med Internet Res. Jan 20, 2021;23(1):e17187. [FREE Full text] [CrossRef] [Medline]
  4. Domenico GD, Sit J, Ishizaka A, Nunan D. Fake news, social media and marketing: a systematic review. J Bus Res. Jan 2021;124:329-341. [CrossRef]
  5. van Mulukom V, Pummerer LJ, Alper S, Bai H, Čavojová V, Farias J, et al. Antecedents and consequences of COVID-19 conspiracy beliefs: a systematic review. Soc Sci Med. May 2022;301:114912. [FREE Full text] [CrossRef] [Medline]
  6. Gabarron E, Oyeyemi SO, Wynn R. COVID-19-related misinformation on social media: a systematic review. Bull World Health Organ. Jun 01, 2021;99(6):455-63A. [FREE Full text] [CrossRef] [Medline]
  7. Wang R, Zhang H. Who spread COVID-19 (mis)information online? Differential informedness, psychological mechanisms, and intervention strategies. Comput Human Behav. Jan 2023;138:107486. [FREE Full text] [CrossRef] [Medline]
  8. Pickles K, Cvejic E, Nickel B, Copp T, Bonner C, Leask J, et al. COVID-19 misinformation trends in Australia: prospective longitudinal national survey. J Med Internet Res. Jan 07, 2021;23(1):e23805. [FREE Full text] [CrossRef] [Medline]
  9. Su Y. It doesn't take a village to fall for misinformation: social media use, discussion heterogeneity preference, worry of the virus, faith in scientists, and COVID-19-related misinformation beliefs. Telemat Inform. May 2021;58:101547. [FREE Full text] [CrossRef] [Medline]
  10. Freeman D, Waite F, Rosebrock L, Petit A, Causier C, East A, et al. Coronavirus conspiracy beliefs, mistrust, and compliance with government guidelines in England. Psychol Med. May 21, 2020;52(2):251-263. [CrossRef]
  11. Zarocostas J. How to fight an infodemic. Lancet. Feb 2020;395(10225):676. [CrossRef]
  12. van der Linden S. Misinformation: susceptibility, spread, and interventions to immunize the public. Nat Med. Mar 10, 2022;28(3):460-467. [CrossRef] [Medline]
  13. Jiang M, Gao Q, Zhuang J. Reciprocal spreading and debunking processes of online misinformation: a new rumor spreading–debunking model with a case study. Phys A Stat Mech Appl. Mar 2021;565:125572. [CrossRef]
  14. Chou WY, Gaysynsky A, Vanderpool RC. The COVID-19 misinfodemic: moving beyond fact-checking. Health Educ Behav. Mar 2021;48(1):9-13. [FREE Full text] [CrossRef] [Medline]
  15. Clayton K, Blair S, Busam JA, Forstner S, Glance J, Green G, et al. Real solutions for fake news? Measuring the effectiveness of general warnings and fact-check tags in reducing belief in false stories on social media. Polit Behav. Feb 11, 2019;42(4):1073-1095. [CrossRef]
  16. Pennycook G, Bear A, Collins ET, Rand DG. The implied truth effect: attaching warnings to a subset of fake news headlines increases perceived accuracy of headlines without warnings. Management Science. Nov 2020;66(11):4944-4957. [CrossRef]
  17. Freeze M, Baumgartner M, Bruno P, Gunderson JR, Olin J, Ross MQ, et al. Fake claims of fake news: political misinformation, warnings, and the tainted truth effect. Polit Behav. Feb 05, 2020;43(4):1433-1465. [CrossRef]
  18. Jin Y, van der Meer TG, Lee YI, Lu X. The effects of corrective communication and employee backup on the effectiveness of fighting crisis misinformation. Public Relat Rev. Sep 2020;46(3):101910. [CrossRef]
  19. Hameleers M, van der Meer TG. Misinformation and polarization in a high-choice media environment: how effective are political fact-checkers? Commun Res. Jan 13, 2019;47(2):227-250. [CrossRef]
  20. Vraga EK, Bode L, Tully M. Creating news literacy messages to enhance expert corrections of misinformation on twitter. Commun Res. Jan 30, 2020;49(2):245-267. [CrossRef]
  21. Hosni AI, Li K. Minimizing the influence of rumors during breaking news events in online social networks. Knowl Based Syst. Apr 2020;193:105452. [CrossRef]
  22. Walter N, Brooks JJ, Saucier CJ, Suresh S. Evaluating the impact of attempts to correct health misinformation on social media: a meta-analysis. Health Commun. Aug 06, 2020:1-9. [CrossRef] [Medline]
  23. Lewandowsky S, van der Linden S. Countering misinformation and fake news through inoculation and prebunking. Eur Rev Soc Psychol. Feb 22, 2021;32(2):348-384. [CrossRef]
  24. Basol M, Roozenbeek J, Berriche M, Uenal F, McClanahan WP, van der Linden S. Towards psychological herd immunity: cross-cultural evidence for two prebunking interventions against COVID-19 misinformation. Big Data Soc. May 11, 2021;8(1):205395172110138. [CrossRef]
  25. Dib F, Mayaud P, Chauvin P, Launay O. Online mis/disinformation and vaccine hesitancy in the era of COVID-19: why we need an eHealth literacy revolution. Hum Vaccin Immunother. Dec 31, 2022;18(1):1-3. [FREE Full text] [CrossRef] [Medline]
  26. Maertens R, Roozenbeek J, Basol M, van der Linden S. Long-term effectiveness of inoculation against misinformation: three longitudinal experiments. J Exp Psychol Appl. Mar 2021;27(1):1-16. [CrossRef] [Medline]
  27. Boudewyns V, Betts KR, Johnson M, Paquin RS, O'Donoghue AC, Southwell BG. Experimental evidence of consumer and physician detection and rejection of misleading prescription drug website content. Res Social Adm Pharm. Apr 2021;17(4):733-743. [CrossRef] [Medline]
  28. Rodrigues UM, Xu J. Regulation of COVID-19 fake news infodemic in China and India. Media Int Aust. Aug 17, 2020;177(1):125-131. [CrossRef]
  29. Marsden C, Meyer T, Brown I. Platform values and democratic elections: how can the law regulate digital disinformation? Comput Law Secur Rev. Apr 2020;36:105373. [CrossRef]
  30. Freiling I, Krause NM, Scheufele DA, Brossard D. Believing and sharing misinformation, fact-checks, and accurate information on social media: the role of anxiety during COVID-19. New Media Soc. Jan 22, 2023;25(1):141-162. [FREE Full text] [CrossRef] [Medline]
  31. Wang X, Chao F, Yu G. Evaluating rumor debunking effectiveness during the COVID-19 pandemic crisis: utilizing user stance in comments on Sina Weibo. Front Public Health. 2021;9:770111. [FREE Full text] [CrossRef] [Medline]
  32. Swire-Thompson B, Lazer D. Public health and online misinformation: challenges and recommendations. Annu Rev Public Health. Apr 02, 2020;41:433-451. [CrossRef] [Medline]
  33. Trevors G, Duffy MC. Correcting COVID-19 misconceptions requires caution. Educ Res. Sep 08, 2020;49(7):538-542. [CrossRef]
  34. Chen S, Xiao L, Kumar A. Spread of misinformation on social media: what contributes to it and how to combat it. Comput Human Behav. Apr 2023;141:107643. [CrossRef]
  35. Zhang Y, Guo B, Ding Y, Liu J, Qiu C, Liu S, et al. Investigation of the determinants for misinformation correction effectiveness on social media during COVID-19 pandemic. Inf Process Manag. May 2022;59(3):102935. [FREE Full text] [CrossRef] [Medline]
  36. Petz G, Karpowicz M, Fürschuß H, Auinger A, Stříteský V, Holzinger A. Computational approaches for mining user’s opinions on the Web 2.0. Inf Process Manage. Nov 2014;50(6):899-908. [CrossRef]
  37. Carrasco-Farré C. The fingerprints of misinformation: how deceptive content differs from reliable sources in terms of cognitive effort and appeal to emotions. Humanit Soc Sci Commun. May 09, 2022;9(1):1. [CrossRef]
  38. Charquero-Ballester M, Walter JG, Nissen IA, Bechmann A. Different types of COVID-19 misinformation have different emotional valence on Twitter. Big Data Soc. Sep 22, 2021;8(2):205395172110412. [CrossRef]
  39. Jang H, Rempel E, Roe I, Adu P, Carenini G, Janjua NZ. Tracking public attitudes toward COVID-19 vaccination on tweets in Canada: using aspect-based sentiment analysis. J Med Internet Res. Mar 29, 2022;24(3):e35016. [FREE Full text] [CrossRef] [Medline]
  40. Zhou C, Li K, Lu Y. Linguistic characteristics and the dissemination of misinformation in social media: the moderating effect of information richness. Inf Process Manage. Nov 2021;58(6):102679. [CrossRef]
  41. Tully M, Bode L, Vraga EK. Mobilizing users: does exposure to misinformation and its correction affect users’ responses to a health misinformation post? Soc Media Soc. Dec 10, 2020;6(4):205630512097837. [CrossRef]
  42. Tsao SF, Chen H, Tisseverasinghe T, Yang Y, Li L, Butt ZA. What social media told us in the time of COVID-19: a scoping review. Lancet Digit Health. Mar 2021;3(3):e175-e194. [CrossRef]
  43. Rahmanti AR, Chien CH, Nursetyo AA, Husnayain A, Wiratama BS, Fuad A, et al. Social media sentiment analysis to monitor the performance of vaccination coverage during the early phase of the national COVID-19 vaccine rollout. Comput Meth Prog Bio. Jun 2022;221:106838. [FREE Full text] [CrossRef] [Medline]
  44. Ridderinkhof KR. Emotion in action: a predictive processing perspective and theoretical synthesis. Emot Rev. Oct 2017;9(4):319-325. [FREE Full text] [CrossRef] [Medline]
  45. Weiner B. An attributional theory of achievement motivation and emotion. Psychol Rev. 1985;92(4):548-573. [CrossRef]
  46. Saleh MI. Attribution theory revisited: probing the link among locus of causality theory, destination social responsibility, tourism experience types, and tourist behavior. J Travel Res. Sep 19, 2022;62(6):1309-1327. [CrossRef]
  47. Weiner B. Attributional thoughts about consumer behavior. J Consum Res. Dec 01, 2000;27(3):382-387. [CrossRef]
  48. Che W, Feng Y, Qin L, Liu T. N-LTP: an open-source neural language technology platform for Chinese. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing: System Demonstrations. 2021. Presented at: EMNLP '21; November 7-11, 2021:42-49; Virtual Event. URL: https://aclanthology.org/2021.emnlp-demo.6.pdf [CrossRef]
  49. Cui Y, Che W, Liu T, Qin B, Wang S, Hu G. Revisiting pre-trained models for Chinese natural language processing. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing. 2020. Presented at: EMNLP '20; November 16-20, 2020:657-668; Virtual Event. URL: https://aclanthology.org/2020.findings-emnlp.58.pdf [CrossRef]
  50. Li X, Bing L, Li P, Lam W. A unified model for opinion target extraction and target sentiment prediction. In: Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence and Thirty-First Innovative Applications of Artificial Intelligence Conference and Ninth AAAI Symposium on Educational Advances in Artificial Intelligence. 2019. Presented at: AAAI '19/IAAI '19/EAAI '19; January 27-February 1, 2019:6714-6721; Honolulu, HI. URL: https://dl.acm.org/doi/proceedings/10.5555/3508659 [CrossRef]
  51. Li X, Bing L, Zhang W, Lam W. Exploiting BERT for end-to-end aspect-based sentiment analysis. arXiv. Preprint posted online October 2, 2019.. [FREE Full text] [CrossRef]
  52. Yang H, Zeng B, Yang J, Song Y, Xu R. A multi-task learning model for Chinese-oriented aspect polarity classification and aspect term extraction. Neurocomput. Jan 2021;419:344-356. [CrossRef]
  53. Zhang Y, Lu H, Jiang C, Li X, Tian X. Aspect-based sentiment analysis of user reviews in 5G networks. IEEE Network. Jul 2021;35(4):228-233. [CrossRef]
  54. Zhang Y, Du J, Ma X, Wen H, Fortino G. Aspect-based sentiment analysis for user reviews. Cogn Comput. Jul 13, 2021;13(5):1114-1127. [CrossRef]
  55. Jang H, Rempel E, Roth D, Carenini G, Janjua NZ. Tracking COVID-19 discourse on Twitter in North America: infodemiology study using topic modeling and aspect-based sentiment analysis. J Med Internet Res. Feb 10, 2021;23(2):e25431. [FREE Full text] [CrossRef] [Medline]
  56. Chen H, Zhu Z, Qi F, Ye Y, Liu Z, Sun M, et al. Country image in COVID-19 pandemic: a case study of China. IEEE Trans Big Data. Mar 1, 2021;7(1):81-92. [CrossRef]
  57. Li MH, Chen Z, Rao LL. Emotion, analytic thinking and susceptibility to misinformation during the COVID-19 outbreak. Comput Human Behav. Aug 2022;133:107295. [FREE Full text] [CrossRef] [Medline]
  58. Aikin KJ, Southwell BG, Paquin RS, Rupert DJ, O'Donoghue AC, Betts KR, et al. Correction of misleading information in prescription drug television advertising: the roles of advertisement similarity and time delay. Res Social Adm Pharm. Mar 2017;13(2):378-388. [CrossRef] [Medline]
  59. Yu W, Shen F, Min C. Correcting science misinformation in an authoritarian country: an experiment from China. Telemat Inform. Jan 2022;66:101749. [CrossRef]


ABSA: aspect-based sentiment analysis
BERT: Bidirectional Encoder Representations from Transformers


Edited by A Mavragani; submitted 28.06.23; peer-reviewed by B Southwell, A Holzinger, B Southwell; comments to author 27.05.24; revised version received 16.06.24; accepted 19.06.24; published 16.08.24.

Copyright

©Ning Ma, Guang Yu, Xin Jin. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 16.08.2024.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.