Published on in Vol 24, No 4 (2022): April

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/36489, first published .
Stigmatizing Attitudes Across Cybersuicides and Offline Suicides: Content Analysis of Sina Weibo

Stigmatizing Attitudes Across Cybersuicides and Offline Suicides: Content Analysis of Sina Weibo

Stigmatizing Attitudes Across Cybersuicides and Offline Suicides: Content Analysis of Sina Weibo

Authors of this article:

Ang Li1, 2 Author Orcid Image ;   Dongdong Jiao3 Author Orcid Image ;   Tingshao Zhu2, 4 Author Orcid Image

Original Paper

1Department of Psychology, Beijing Forestry University, Beijing, China

2Key Laboratory of Behavioral Science, Institute of Psychology, Chinese Academy of Sciences, Beijing, China

3National Computer System Engineering Research Institute of China, Beijing, China

4Department of Psychology, University of Chinese Academy of Sciences, Beijing, China

Corresponding Author:

Tingshao Zhu, PhD

Key Laboratory of Behavioral Science

Institute of Psychology

Chinese Academy of Sciences

Hexie Building

No 16 Lincui Rd

Beijing, 100101

China

Phone: 86 150 1096 5509

Email: tszhu@psych.ac.cn


Background: The new reality of cybersuicide raises challenges to ideologies about the traditional form of suicide that does not involve the internet (offline suicide), which may lead to changes in audience’s attitudes. However, knowledge on whether stigmatizing attitudes differ between cybersuicides and offline suicides remains limited.

Objective: This study aims to consider livestreamed suicide as a typical representative of cybersuicide and use social media data (Sina Weibo) to investigate the differences in stigmatizing attitudes across cybersuicides and offline suicides in terms of attitude types and linguistic characteristics.

Methods: A total of 4393 cybersuicide-related and 2843 offline suicide-related Weibo posts were collected and analyzed. First, human coders were recruited and trained to perform a content analysis on the collected posts to determine whether each of them reflected stigma. Second, a text analysis tool was used to automatically extract a number of psycholinguistic features from each post. Subsequently, based on the selected features, a series of classification models were constructed for different purposes: differentiating the general stigma of cybersuicide from that of offline suicide and differentiating the negative stereotypes of cybersuicide from that of offline suicide.

Results: In terms of attitude types, cybersuicide was observed to carry more stigma than offline suicide (χ21=179.8; P<.001). Between cybersuicides and offline suicides, there were significant differences in the proportion of posts associated with five different negative stereotypes, including stupid and shallow (χ21=28.9; P<.001), false representation (χ21=144.4; P<.001), weak and pathetic (χ21=20.4; P<.001), glorified and normalized (χ21=177.6; P<.001), and immoral (χ21=11.8; P=.001). Similar results were also found for different genders and regions. In terms of linguistic characteristics, the F-measure values of the classification models ranged from 0.81 to 0.85.

Conclusions: The way people perceive cybersuicide differs from how they perceive offline suicide. The results of this study have implications for reducing the stigma against suicide.

J Med Internet Res 2022;24(4):e36489

doi:10.2196/36489

Keywords



Background

Suicide remains one of the leading causes of death worldwide according to the latest estimates released by the World Health Organization [1]. Mental illness stigma refers to the negative stereotyping of people with mental illnesses that may lead to discrimination [2]. Although there is often a stigma associated with all mental illnesses, suicide can be especially stigmatized, dismissed as “merely attention-seeking gesturers” [3]. Stigma against suicide can lead to a reduced likelihood of receiving treatment and an increased risk of death by suicide [4,5]. Therefore, lowering stigma can improve mental health outcomes. Stigma reduction efforts are more likely to be effective when targeting specific mental health problems [6-8]. Therefore, it is crucial to understand the negative stereotypes people often associate with suicide and then design antistigma campaigns accordingly.

The internet facilitates self-disclosure and social connection, giving rise to an emerging form of suicide (ie, cybersuicide). Unlike the traditional form of suicide that does not involve the internet (ie, offline suicide), cybersuicide covers a broad range of internet-mediated suicidal behaviors and phenomena, including livestreamed suicide [9,10]. The internet increases people’s willingness to disclose more about themselves and offers highly interactive platforms for interpersonal communication (eg, Twitter and Sina Weibo). Therefore, on the internet, people with suicidal intentions are motivated to share and communicate their thoughts on suicide with others and are able to maintain active contact with their social network in the last moments of life, although members of social networks do not live close to them. In livestreamed suicide, for example, through various media (text messages, pictures, videos, and voice notes), the internet enables people with suicidal intentions to broadcast their suicides to their entire social network before death and allows real time interaction between people with suicidal intentions and their audience. This means that cybersuicide makes a very personal and private act highly public and greatly enhances the interaction between people with suicidal intentions and their audience, which raises challenges to long-held ideas and beliefs about offline suicide [11-13] and facilitates the creation of emerging cultures around suicide [9,10,14,15]. It is well known that social and cultural factors can influence attitudes toward mental health problems, including suicide [16-18]. Therefore, it is not surprising that cybersuicide, which has radically transformed the sociocultural context of suicide, may lead to changes in audience’s attitudes, suggesting the need for further examination of the differences in stigmatizing attitudes across cybersuicides and offline suicides.

The livestreamed suicide is commonly considered as one of the most notable and representative types of cybersuicide, particularly in China [9,19]. From 2003 to 2016, at least 193 livestreamed suicide incidents occurred in China [20]. Therefore, considerable research efforts have been directed toward understanding the specific types of attitudes associated with livestreamed suicide. As social media allow users to freely disclose their feelings and thoughts and include vast quantities of publicly available data, many studies used human coders to analyze relevant social media data (eg, posts with relevant keywords and audience-generated messages related to relevant suicide incidents) and concluded that the public may react strongly against livestreamed suicide [21-24]. Besides, a study found a distinctive type of negative stereotype associated with livestreamed suicide (ie, false representation stigma, a misleading belief that people livestreaming their suicides do not really want to kill themselves) [25], which was not mentioned in previous studies regarding offline suicide. The results of these studies imply that the public may react differently to cybersuicides and offline suicides. However, to our knowledge, no research has been conducted to directly investigate the differences in stigmatizing attitudes toward cybersuicides and offline suicides. Furthermore, attitudes can manifest not only in specific attitude types (what it says exactly) but also in linguistic characteristics of language expressions (how it is said) [26-29]. Recent studies confirmed that health-related stigmatizing attitudes (eg, Alzheimer disease, depression, and livestreamed suicide) can be identified by analyzing linguistic characteristics of expressions in social media posts [30-32]. More importantly, the differences in stigmatizing attitudes across mental health problems can be reflected in different patterns of language use as well [33]. Therefore, apart from the differences in types of stigma, the differences in the linguistic characteristics of stigmatizing expressions also need to be investigated.

Objective

To address these concerns by analyzing social media data (Sina Weibo, a Chinese social media site that is similar to Twitter), this study attempts to directly and systematically investigate the differences in stigmatizing attitudes across cybersuicides and offline suicides in terms of attitude types and linguistic characteristics, respectively.

It is worth noting that cybersuicide is a new and developing form of suicide. The public may not be equally familiar with different types of cybersuicide. In China, because of the prevalence and media coverage of livestreamed suicide, compared with other types of cybersuicide, the public is expected to be more familiar with and more likely to discuss livestreamed suicide on social media. Therefore, to collect sufficient social media data for further analysis, this study aims to consider livestreamed suicide as a typical representative of cybersuicide and compare it with offline suicide.


Research Process

The research process included the following three steps: (1) data collection, (2) data preprocessing, and (3) data analysis. The data collection and preprocessing procedures are shown in Figure 1.

Figure 1. Procedures of data collection and preprocessing.
View this figure
Data Collection

First, a participant pool of active Weibo users was created. According to a previous study, 1,953,485 active Weibo users were identified as potential participants [34]. The application programming interface (API) platform of Sina Weibo enables programmatic access to Weibo data that users choose to share with the public. As Sina Weibo places limits and quotas on API requests, among all potential participants, 1.06 million active users with available data were finally included in the participant pool.

Second, a database of Weibo posts was constructed. In May 2020, using API, a vast amount of publicly available Weibo posts from 1.06 million active users in the participant pool since their registration (the 2020 official numbers of monthly and daily active users: 511 million and 224 million, respectively [35]), were downloaded.

Third, several relevant Weibo posts were identified through database searches. To obtain the posts reflecting attitudes toward cybersuicides and offline suicides, two sets of keywords were used to search the database, including livestreamed suicide (直播自杀 and 自杀直播) and suicide (自杀). It is worth noting that although cybersuicide is a rapidly developing form of suicide, an overwhelming majority of suicide incidents still occur offline. For example, between 2003 and 2016, 193 incidents of livestreamed suicide occurred in China [20], whereas in 2019, the number of suicides in China had reached 116,324 [1]. Therefore, unless otherwise stated, the Chinese people commonly use the term suicide to refer to traditional offline suicide.

The process of database searches included the following three steps: (1) a total of 4460 posts with keywords livestreamed suicide were searched and obtained (cybersuicide group); (2) a total of 933,172 posts with keywords suicide (but without keywords livestreamed suicide) were searched and obtained (offline suicide group); (3) to balance the number of posts in each group, using simple random sampling, 4500 posts were randomly selected from 933,172 posts in the offline suicide group (cybersuicide: 4460 posts and offline suicide: 4500 posts).

Data Preprocessing

After data collection, preprocessing was performed on the raw data to prepare them for further analysis.

First, to exclude irrelevant posts and reclassify misclassified posts, manual scrutiny of the collected data was conducted.

In this study, irrelevant posts were considered as (1) posts that depicted suicides in fictional works (eg, movies, television programs, and novels), (2) posts that focused on suicides in nonhuman animals (eg, dogs), and (3) posts that used suicide-related keywords for nonsuicidal purposes (eg, making a bet). After the removal of irrelevant posts, 7244 posts (cybersuicide: 4460 – 136 = 4324 and offline suicide: 4500 − 1580 = 2920) remained.

Subsequently, 77 posts in the offline suicide group were reclassified as posts related to cybersuicide (livestreamed suicide: n=69, 90% posts; suicide game: n=4, 5% posts; prosuicide website and forum: n=2, 3% posts; and internet suicide pact: n=2, 3% posts). As this study primarily focused on livestreamed suicide rather than other types of cybersuicide, 8 posts related to the other 3 types of cybersuicide were excluded from further analysis.

Therefore, the final sample of this study included 7236 posts (cybersuicide: 4324 + 69 = 4393; offline suicide: 2920 – 69 – 8 = 2843). The demographic characteristics of the participants in the final sample are presented in Table 1.

Table 1. Demographics of participants.
DemographicsAll Weibo posts (N=7236), n (%)Cybersuicide (n=4393), n (%)Offline suicide (n=2843), n (%)
Gender

Male4062 (56.14)2473 (56.29)1589 (55.89)

Female3174 (43.86)1920 (43.71)1254 (44.11)
Regions

North China1312 (18.13)812 (18.48)500 (17.59)

Northeast China299 (4.13)191 (4.35)108 (3.8)

East China2277 (31.47)1330 (30.28)947 (33.31)

Central China374 (5.17)233 (5.3)141 (4.96)

South China1345 (18.59)801 (18.23)544 (19.13)

Southwest China493 (6.81)330 (7.51)163 (5.73)

Northwest China224 (3.1)133 (3.03)91 (3.2)

International and unspecified912 (12.6)563 (12.82)349 (12.28)

Second, to extract psycholinguistic features from each of the 7236 posts automatically, the Simplified Chinese version of Linguistic Inquiry and Word Count software was used. The Simplified Chinese version of Linguistic Inquiry and Word Count is a reliable and valid text analysis tool for the automatic estimation of word frequency in multiple psychologically meaningful categories, including linguistic processes (eg, personal pronouns), psychological processes (eg, affective processes), personal concerns (eg, achievement), spoken categories (eg, assent), and punctuation categories (eg, periods) [36]. After feature extraction, the standardized values of psycholinguistic features were estimated for further analysis.

Data Analysis
Human Coding

To explore the differences in types of stigmatizing attitudes across cybersuicides and offline suicides, a content analysis was performed on all 7236 posts to determine whether each of them reflected stigma. The coding framework was developed based on expert consensus and available evidence. Specifically, in this study, a researcher (AL) reviewed relevant studies [25,37] and then performed an inductive content analysis on all 7236 posts to develop an initial coding framework. Subsequently, another two researchers (DJ and TZ) provided feedback to emend the initial framework. Using the amended framework (Table S1 in Multimedia Appendix 1), 2 independent human coders were recruited and trained to analyze all 7236 posts. The levels of agreement between the 2 coders were measured by Cohen k coefficient, and disagreements were resolved by the decisions of a researcher (AL). All individuals who participated in developing the coding framework and performing manual coding of posts had considerable experience in coding qualitative materials.

Construction of Classification Models

To explore the linguistic differences in stigmatizing expressions across cybersuicides and offline suicides, 2 groups of classification models were built using Waikato Environment for Knowledge Analysis (version 3.9.4; University of Waikato) software. Waikato Environment for Knowledge Analysis provides tools for developing machine learning techniques and applying them to practical data mining problems.

The first group of classification models was built to investigate whether linguistic differences existed in the expression of stigma in general (ie, cybersuicide-related or offline suicide-related stigma as a whole) between cybersuicides and offline suicides. The human coding results were considered as the ground truth for the validation of the classification models. In this study, an imbalanced data problem existed. For example, the number of stigmatizing posts belonging to the offline suicide class (minority class: 588 posts) was obviously lower than those belonging to the cybersuicide class (majority class: 1556 posts). Imbalanced data sets pose a challenge for machine learning modeling, as this problem may result in models with poor predictive performance, especially for the minority class. To handle this problem, using simple random sampling, a certain number of posts were randomly selected from the majority class to obtain a well-balanced data set. Subsequently, to improve classification accuracy, a subset of psycholinguistic features was selected for use in model construction. Specifically, a series of 2-tailed independent t tests were conducted to compare the values of all extracted psycholinguistic features between stigmatizing posts in the cybersuicide and offline suicide groups, and then effect sizes (Cohen d coefficient) were computed from the estimated t values. Features that were statistically significant at .05 and had a Cohen d >0.20 or <−0.20 were considered as key features. Finally, using four different machine learning algorithms (Naïve Bayes, support vector machine, multilayer perceptron neural network, and random forest [RF]), 4 classification models were constructed based on the selected key features. Using a 5-fold cross-validation technique, the classification performance of the established models was evaluated in terms of precision, recall, and F-measure.

It is worth noting that the good classification performance of models in the first group may be attributed to the existence of differences in the amount and distribution of negative stereotypes across cybersuicides and offline suicides rather than the existence of linguistic differences in stigmatizing expressions across cybersuicides and offline suicides. To clarify this issue, the second group of classification models was built to investigate whether linguistic differences existed in the expression of certain negative stereotypes across cybersuicides and offline suicides. To obtain sufficient data for further analysis, in this study, posts reflecting two negative stereotypes (ie, stupid and shallow and glorified and normalized) were examined. Well-balanced data sets, key features, and classification models were obtained using the aforementioned methods.

Ethics Approval

The study protocol was reviewed and approved by the institutional review board of the Institute of Psychology, Chinese Academy of Sciences (protocol number: H15009). Informed consent was not obtained, as this study was based on publicly available data and involved no personally identifiable data collection or analysis.


Human Coding

The Cohen k coefficients for attitudes and negative stereotypes reached 0.88 and 0.81, respectively, indicating almost perfect agreement [38]. The results of human coding are presented in Table 2.

Table 2. Results of human coding (N=7236).
CategoriesCybersuicide, n (%)Offline suicide, n (%)
Attitudes4393 (100)2843 (100)

Stigmatizing1556 (35.4)588 (20.7)

Nonstigmatizing2837 (64.6)2255 (79.3)
Negative stereotypes1556 (100)588 (100)

Weak and pathetic114 (7.3)80 (13.6)

Self-centered97 (6.2)44 (7.5)

Stupid and shallow528 (33.9)129 (21.9)

False representation387 (24.9)13 (2.2)

Glorified and normalized148 (9.5)195 (33.2)

Immoral111 (7.1)69 (11.7)

Strange59 (3.8)14 (2.4)

Embarrassing24 (1.5)7 (1.2)

Vengeful40 (2.6)9 (1.5)

Mad48 (3.1)28 (4.8)

For stigma in general, posts on cybersuicide were more likely than posts on offline suicide to contain stigmatizing expressions (χ21=179.8; P<.001). Similar results were found for different genders and regions, including men (χ21=66.7; P<.001), women (χ21=121.0; P<.001), North China (NC; χ21=37.2; P<.001), East China (EC; χ21=56.4; P<.001), Central China (CC; χ21=10.4; P=.001), South China (SC; χ21=37.6; P<.001), and Southwest China (SWC; χ21=11.8; P=.001).

For negative stereotypes, posts on cybersuicide were often coded as stupid and shallow (528/1556, 33.93%) and false representation (387/1556, 24.87%), whereas posts on offline suicide were often coded as glorified and normalized (195/588, 33.2%) and stupid and shallow (129/588, 21.9%). Furthermore, posts on cybersuicide were more likely than posts on offline suicide to be coded as stupid and shallow (χ21=28.9; P<.001) and false representation (χ21=144.4; P<.001), whereas posts on offline suicide were more likely than posts on cybersuicide to be coded as weak and pathetic (χ21=20.4; P<.001), glorified and normalized (χ21=177.6; P<.001), and immoral (χ21=11.8; P=.001). Similar results were found for different genders and regions. Specifically, significant differences in the proportions of posts coded as weak and pathetic were observed for posts by women (χ21=21.0; P<.001) and from NC (χ21=10.6; P=.001) and EC (χ21=3.9; P=.048). Significant differences in the proportions of posts coded as stupid and shallow were observed for posts by men (χ21=12.6; P<.001) and women (χ21=16.0; P<.001) and from NC (χ21=9.1; P=.003), EC (χ21=7.9; P=.005), SC (χ21=5.4; P=.02), and SWC (χ21=8.0; P=.005). Significant differences in the proportions of posts coded as false representation were observed for posts by men (χ21=86.4; P<.001) and women (χ21=59.5; P<.001) and from NC (χ21=25.5; P<.001), Northeast China (Fisher exact test: P=.004), EC (χ21=48.2; P<.001), CC (Fisher exact test: P=.006), SC (χ21=16.5; P<.001), SWC (χ21=8.1; P=.004), and Northwest China (Fisher exact test: P=.003). Significant differences in the proportions of posts coded as glorification and normalized were observed for posts by men (χ21=127.3; P<.001) and women (χ21=55.1; P<.001) and from NC (χ21=37.5; P<.001), Northeast China (Fisher exact test: P<.001), EC (χ21=60.7; P<.001), CC (Fisher exact test: P=.01), SC (χ21=17.6; P<.001), and SWC (χ21=21.5; P<.001). Significant differences in the proportions of posts coded as immoral were observed for posts by men (χ21=7.8; P=.005) and women (χ21=3.9; P=.048) and from EC (χ21=4.8; P=.03) and Northwest China (Fisher exact test: P=.008). In addition, posts on offline suicide were more likely than posts on cybersuicide to be coded as mad for posts by women (χ21=5.4; P=.02) and from CC (Fisher exact test: P=.04) and SC (Fisher exact test: P=.03).

Construction of Classification Models

Stigma in General

For exploring linguistic differences in the expression of stigma in general between cybersuicides and offline suicides, to achieve a balanced data set, 600 stigmatizing posts were randomly selected from posts in cybersuicide group (cybersuicide: 600 posts and offline suicide: 588 posts). A total of 6 key features were selected for use in the model construction (Table S2 in Multimedia Appendix 1). The RF model exhibited the best classification performance (precision=0.86, recall=0.85, and F-measure=0.85; Table 3).

Table 3. Performance of classification models.
ModelsStigma in generalStupid and shallowGlorified and normalized
Naïve Bayes

Precision0.730.720.72

Recall0.730.720.72

F-measure0.730.720.72
Support vector machine

Precision0.830.800.79

Recall0.830.800.78

F-measure0.830.790.78
Multilayer perceptron neural network

Precision0.850.840.78

Recall0.850.830.77

F-measure0.850.830.77
Random forest

Precision0.860.840.81

Recall0.850.840.81

F-measure0.850.840.81
Stupid and Shallow

For exploring linguistic differences in the expression of stupid and shallow between cybersuicides and offline suicides, to achieve a balanced data set, a total of 150 stupid and shallow–related posts were randomly selected from posts in cybersuicide group (cybersuicide: 150 posts and offline suicide: 129 posts). A total of 4 key features were selected for use in the model construction (Table S2 in Multimedia Appendix 1). The RF model exhibited the best classification performance (precision=0.84, recall=0.84, and F-measure=0.84; Table 3).

Glorified and Normalized

To explore the linguistic differences in the expression of glorified and normalized between cybersuicides and offline suicides (cybersuicide: 148 posts and offline suicide: 195 posts), 28 key features were selected for use in model construction (Table S2 in Multimedia Appendix 1). The RF model exhibited the best classification performance (precision=0.81, recall=0.81, and F-measure=0.81; Table 3).


Principal Findings

To our knowledge, this study provides the first systematic analysis of the differences in stigmatizing attitudes toward cybersuicides and offline suicides. The results of this study have implications for reducing stigma against suicide.

First, it is necessary to confront and reduce stigma against cybersuicide. In this study, a large proportion of cybersuicide-related and offline suicide-related posts were coded as stigmatizing (1556/4393, 35.42% and 588/2843, 20.68%, respectively), although most posts about either type of suicide were not overtly stigmatizing. Cybersuicide allows people with suicidal intentions to interact with their audience in the last moments before death. If the audience responds appropriately to people with suicidal intentions, the likelihood of death would be reduced [39-41]. The prevalence of stigma may hamper audience from effectively responding to people with suicidal intentions [25,42-44]. Therefore, cybersuicide deserves to focus on antistigma campaigns.

Second, the public reacts differently to cybersuicides and offline suicides. In terms of attitude types, in this study, cybersuicide was observed to carry more stigma than offline suicide (cybersuicide: 1556/4393, 35.42% and offline suicide: 588/2843, 20.68%; χ21=179.8; P<.001), implying that the public may react more negatively to cybersuicide than to offline suicide. This is consistent with 2 previous studies that explored audience responses to cybersuicide and offline suicide incidents [24,45]. Furthermore, compared with offline suicide, cybersuicide was more likely to be considered as stupid and shallow (χ21=28.9; P<.001) and false representation (χ21=144.4; P<.001) and was less likely to be considered as weak and pathetic (χ21=20.4; P<.001), glorified and normalized (χ21=177.6; P<.001), and immoral (χ21=11.8; P=.001). Similar results were also found for different genders and regions. This indicates that antistigma campaigns targeting offline suicide may not be effective in changing stigmatizing attitudes toward cybersuicide, suggesting the need for public awareness campaigns that specifically target cybersuicide. In addition, the most notable and obvious difference between the two types of suicide was found in false representation stigma (cybersuicide: 387/1556, 24.87% and offline suicide: 13/588, 2.2%). The misbelief that cybersuicide is not real may influence stigmatizing responses elicited mainly by the value judgment of suicide death itself but not by the perception of suicidal motives, causes, and methods. For example, the glorification and immorality of suicide represent two distinct values: that it is right or wrong to deliberately take one’s own life. However, dismissing cybersuicide as a nonreal condition contradicts the assumption of such 2 stigmas that suicide is a real condition. This might be the reason why cybersuicide was less likely to be considered as glorified and normalized and immoral than offline suicide. It is also worth noting that cybersuicide was commonly considered as stupid and shallow (528/1556, 33.93%) and false representation (387/1556, 24.87%), whereas offline suicide was commonly considered as glorified and normalized (195/588, 33.2%) and stupid and shallow (129/588, 21.9%). According to previous studies, inappropriate audience responses are associated with greater false representation stigma [25], and both suicide ideation and suicide contagion are associated with greater glorification of suicide [46,47]. This indicates that the reduction of stigma against cybersuicide may contribute more in improving audience responses, whereas the reduction of stigma against offline suicide may contribute more in preventing suicide attempts.

Apart from the differences in attitude types, linguistic differences in the expression of stigma between cybersuicides and offline suicides also exist. Such differences existed not only at the level of stigma in general but also at the level of negative stereotypes. In this study, the F-measure values of the classification models ranged from 0.81 to 0.85. Compared with other similar studies [48,49], the classification models achieved satisfying accuracy. These results support the conclusion that the way people perceive cybersuicide is very different from the way people perceive offline suicide, implying that cybersuicide may have a different structure from offline suicide [10].

Third, the use of linguistic analysis methods can facilitate the identification of suicide-related stigma in mass media. Mass media is a major contributor to the dissemination of incorrect information, which may reinforce negative stereotypes surrounding mental illness [50-52]. It would be greatly helpful if mass media campaigns were developed to raise public awareness and challenge the stigma against mental illness. However, because of the sheer volume of information in mass media, it is difficult for human coders to identify and analyze stigmatizing information efficiently, suggesting the need for automatic detection of stigma. Linguistic analysis methods can be used to understand language use patterns of stigmatizing expressions and to construct computational models for the automatic detection of stigma [30,33,53]. For example, in this study, stigmatizing expressions of cybersuicide were associated with more frequent use of words related to leisure (eg, chat) and work (eg, dissemination), which may be attributed to the fact that cybersuicide makes suicidal acts highly public and contradicts the public perception of suicide. By contrast, stigmatizing expressions of offline suicide were associated with more frequent use of words related to achievement (eg, hero), which may be attributed to the higher prevalence of suicide glorification (cybersuicide: 148/1556, 9.51% and offline suicide: 195/588, 33.2%). The constructed machine learning models performed well in classifying stigma toward cybersuicides and offline suicides. This indicates that understanding the linguistic differences in stigmatizing expressions between cybersuicides and offline suicides would make automatic detection more precise. Automatic stigma detection would further improve antistigma efforts. The dissemination of internet messages can be helpful in promoting changes in health-related behaviors [54,55]. However, internet campaigns work best when targeted. With the help of automatic detection and machine learning algorithms, it should be easier to efficiently design and deliver target-specific messages to the population and would promote public mental health awareness.

Limitations

This study has some limitations. First, this study mainly focused on the stigma against livestreamed suicide. Therefore, it is unclear whether these findings are applicable to other types of cybersuicide. Second, social media users are not representative of all people in China. Therefore, the findings may not be applicable to nonusers. Third, the API of Sina Weibo only allowed us to download posts from a certain number of registered users. Therefore, these findings should be further confirmed on a larger scale and in more diverse populations in the future. Fourth, because of the lack of posts obtained from people with suicidal intentions, this study cannot analyze the stigma that people with suicidal intentions put on themselves (ie, self-stigma). Fifth, because of the lack of information about user types (eg, celebrities and general users), this study cannot investigate the differences in attitudes across different types of users and cannot compare attitude responses elicited by the deaths by suicide of different types of users.

Conclusions

This study used a nonintrusive method to directly and systematically examine the differences in stigmatizing attitudes toward cybersuicides and offline suicides. The results of this study support the conclusion that the way people perceive cybersuicide is very different from the way people perceive offline suicide and offer insight into the reduction of stigma against suicide.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Supplementary Tables S1 and S2.

DOCX File , 22 KB

  1. World Health Organization. Suicide worldwide in 2019: global health estimates. Geneva, Switzerland: World Health Organization; Jun 16, 2021.
  2. Corrigan PW, Penn DL. Lessons from social psychology on discrediting psychiatric stigma. Am Psychol 1999 Sep;54(9):765-776. [CrossRef] [Medline]
  3. Sudak H, Maxim K, Carpenter M. Suicide and stigma: a review of the literature and personal reflections. Acad Psychiatry 2008;32(2):136-142. [CrossRef] [Medline]
  4. Carpiniello B, Pinna F. The reciprocal relationship between suicidality and stigma. Front Psychiatry 2017;8:35 [FREE Full text] [CrossRef] [Medline]
  5. Reynders A, Kerkhof AJ, Molenberghs G, Van Audenhove C. Attitudes and stigma in relation to help-seeking intentions for psychological problems in low and high suicide rate regions. Soc Psychiatry Psychiatr Epidemiol 2014 Feb;49(2):231-239. [CrossRef] [Medline]
  6. Carrara BS, Fernandes RH, Bobbili SJ, Ventura CA. Health care providers and people with mental illness: an integrative review on anti-stigma interventions. Int J Soc Psychiatry 2021 Nov;67(7):840-853. [CrossRef] [Medline]
  7. Siegel JT, Flores-Medel E, Martinez DA, Berger DE. Can mental health anti-stigma messages have untoward effects on some people with depression?: an exploratory study. J Health Commun 2019;24(11):821-828. [CrossRef] [Medline]
  8. Corrigan PW. Best practices: strategic stigma change (SSC): five principles for social marketing campaigns to reduce stigma. Psychiatr Serv 2011 Aug;62(8):824-826. [CrossRef] [Medline]
  9. Starcevic V, Aboujaoude E. Cyberchondria, cyberbullying, cybersuicide, cybersex: "new" psychopathologies for the 21st century? World Psychiatry 2015 Feb;14(1):97-100 [FREE Full text] [CrossRef] [Medline]
  10. Fratini A, Hemer SR. Broadcasting your death through livestreaming: understanding cybersuicide through concepts of performance. Cult Med Psychiatry 2020 Dec;44(4):524-543. [CrossRef] [Medline]
  11. La Fontaine J. Explaining suicide: an afterword. Cult Med Psychiatry 2012 Jun;36(2):409-418. [CrossRef] [Medline]
  12. Salem T. Physician-assisted suicide. Promoting autonomy--or medicalizing suicide? Hastings Cent Rep 1999;29(3):30-36. [Medline]
  13. Jamison KR. Suicide and bipolar disorder. J Clin Psychiatry 2000;61 Suppl 9:47-51. [Medline]
  14. Seko Y. On Deathcasting: alone together on the edge of death. In: Nixon PG, Rawal R, Funk A, editors. Digital media across the life course. London, UK: Routledge; 2016:172-189.
  15. Birbal R, Maharajh HD, Birbal R, Clapperton M, Jarvis J, Ragoonath A, et al. Cybersuicide and the adolescent population: challenges of the future? Int J Adolesc Med Health 2009;21(2):151-159. [Medline]
  16. Money TT, Batterham PJ. Sociocultural factors associated with attitudes toward suicide in Australia. Death Stud 2021;45(3):219-225. [CrossRef] [Medline]
  17. Luo A, He H, Mohamed S, Rosenheck R. Medical student attitudes towards people with mental illness in China: a qualitative study. Cult Med Psychiatry 2018 Sep;42(3):535-551. [CrossRef] [Medline]
  18. Chiles C, Stefanovics E, Rosenheck R. Attitudes of students at a US medical school toward mental illness and its causes. Acad Psychiatry 2017 Jun;41(3):320-325. [CrossRef] [Medline]
  19. Chen R, An J, Ou J. Suicidal behaviour among children and adolescents in China. Lancet Child Adolesc Health 2018 Aug;2(8):551-553. [CrossRef] [Medline]
  20. Li A, Huang X, Zhu T. A systematic analysis of online broadcasts of suicidality in China. Asia Pac Psychiatry 2018 Sep;10(3):e12302. [CrossRef] [Medline]
  21. Li A, Jiao D, Liu X, Sun J, Zhu T. A psycholinguistic analysis of responses to live-stream suicides on social media. Int J Environ Res Public Health 2019 Aug 09;16(16):2848 [FREE Full text] [CrossRef] [Medline]
  22. O'Dea B, Achilles MR, Larsen ME, Batterham PJ, Calear AL, Christensen H. The rate of reply and nature of responses to suicide-related posts on Twitter. Internet Interv 2018 Sep;13:105-107 [FREE Full text] [CrossRef] [Medline]
  23. Ma J, Zhang W, Harris K, Chen Q, Xu X. Dying online: live broadcasts of Chinese emerging adult suicides and crisis response behaviors. BMC Public Health 2016 Aug 11;16(1):774 [FREE Full text] [CrossRef] [Medline]
  24. Fu KW, Cheng Q, Wong PW, Yip PS. Responses to a self-presented suicide attempt in social media: a social network analysis. Crisis 2013 Jan 01;34(6):406-412 [FREE Full text] [CrossRef] [Medline]
  25. Li A, Huang X, Hao B, O'Dea B, Christensen H, Zhu T. Attitudes towards suicide attempts broadcast on social media: an exploratory study of Chinese microblogs. PeerJ 2015;3:e1209 [FREE Full text] [CrossRef] [Medline]
  26. Su H, Hunston S. Language patterns and attitude revisited: adjective patterns, attitude and appraisal. Funct Lang 2019 Nov 25;26(3):343-371. [CrossRef]
  27. ŚLebioda A. Stigmatized by the language –linguistic labels of the disabled peoplein Poland and the United States. J Educ Cult Soc 2020 Jan 10;4(2):404-412. [CrossRef]
  28. Puhl R, Peterson JL, Luedicke J. Motivating or stigmatizing? Public perceptions of weight-related language used by health providers. Int J Obes (Lond) 2013 Apr;37(4):612-619. [CrossRef] [Medline]
  29. Pennebaker JW, Mehl MR, Niederhoffer KG. Psychological aspects of natural language. use: our words, our selves. Annu Rev Psychol 2003;54:547-577. [CrossRef] [Medline]
  30. Yu L, Jiang W, Ren Z, Xu S, Zhang L, Hu X. Detecting changes in attitudes toward depression on Chinese social media: a text analysis. J Affect Disord 2021 Feb 01;280(Pt A):354-363. [CrossRef] [Medline]
  31. Li A, Huang X, Jiao D, O'Dea B, Zhu T, Christensen H. An analysis of stigma and suicide literacy in responses to suicides broadcast on social media. Asia Pac Psychiatry 2018 Mar;10(1):e12314. [CrossRef] [Medline]
  32. Oscar N, Fox PA, Croucher R, Wernick R, Keune J, Hooker K. Machine learning, sentiment analysis, and tweets: an examination of Alzheimer's disease stigma on Twitter. J Gerontol B Psychol Sci Soc Sci 2017 Sep 01;72(5):742-751. [CrossRef] [Medline]
  33. Li A, Jiao D, Liu X, Zhu T. A comparison of the psycholinguistic styles of schizophrenia-related stigma and depression-related stigma on social media: content analysis. J Med Internet Res 2020 Apr 21;22(4):e16470 [FREE Full text] [CrossRef] [Medline]
  34. Li L, Li A, Hao B, Guan Z, Zhu T. Predicting active users' personality based on micro-blogging behaviors. PLoS One 2014;9(1):e84997 [FREE Full text] [CrossRef] [Medline]
  35. Weibo reports third quarter 2020 unaudited financial results. Weibo Corporation. 2020.   URL: http:/​/ir.​weibo.com/​news-releases/​news-release-details/​weibo-reports-third-quarter-2020-unaudited-financial-results/​ [accessed 2022-03-25]
  36. Zhao N, Jiao D, Bai S, Zhu T. Evaluating the validity of simplified Chinese version of LIWC in detecting psychological expressions in short texts on social network services. PLoS One 2016;11(6):e0157947 [FREE Full text] [CrossRef] [Medline]
  37. Batterham PJ, Calear AL, Christensen H. The Stigma of Suicide Scale. Psychometric properties and correlates of the stigma of suicide. Crisis 2013 Jan 01;34(1):13-21. [CrossRef] [Medline]
  38. Landis JR, Koch GG. The measurement of observer agreement for categorical data. Biometrics 1977 Mar;33(1):159-174. [Medline]
  39. Armstrong G, Ironfield N, Kelly CM, Dart K, Arabena K, Bond K, et al. Re-development of mental health first aid guidelines for supporting Aboriginal and Torres Strait islanders who are experiencing suicidal thoughts and behaviour. BMC Psychiatry 2018 Jul 16;18(1):228 [FREE Full text] [CrossRef] [Medline]
  40. Ross AM, Kelly CM, Jorm AF. Re-development of mental health first aid guidelines for suicidal ideation and behaviour: a Delphi study. BMC Psychiatry 2014 Sep 13;14:241 [FREE Full text] [CrossRef] [Medline]
  41. Owens D, Horrocks J, House A. Fatal and non-fatal repetition of self-harm. Systematic review. Br J Psychiatry 2002 Sep;181:193-199. [CrossRef] [Medline]
  42. Boukouvalas E, El-Den S, Murphy AL, Salvador-Carulla L, O'Reilly CL. Exploring health care professionals' knowledge of, attitudes towards, and confidence in caring for people at risk of suicide: a systematic review. Arch Suicide Res 2020;24(sup2):S1-31. [CrossRef] [Medline]
  43. Kresovich A. The influence of pop songs referencing anxiety, depression, and suicidal ideation on college students' mental health empathy, stigma, and behavioral intentions. Health Commun 2022 May;37(5):617-627. [CrossRef] [Medline]
  44. Osafo J, Akotia CS, Boakye KE, Dickson E. Between moral infraction and existential crisis: exploring physicians and nurses' attitudes to suicide and the suicidal patient in Ghana. Int J Nurs Stud 2018 Sep;85:118-125. [CrossRef] [Medline]
  45. Larsen ME, Boonstra TW, Batterham PJ, ODea B, Paris C, Christensen H. We feel: mapping emotion on Twitter. IEEE J Biomed Health Inform 2015 Jul;19(4):1246-1252. [CrossRef] [Medline]
  46. Batterham PJ, Han J, Calear AL, Anderson J, Christensen H. Suicide stigma and suicide literacy in a clinical sample. Suicide Life Threat Behav 2019 Aug;49(4):1136-1147. [CrossRef] [Medline]
  47. Beautrais A, Hendin H, Yip P, Takahashi Y, Chia BH, Schmidtke A. Improving portrayal of suicide in the media in Asia. In: Hendin H, Phillips M, Vijayakumar L, Pirkis J, Wang H, Yip P, editors. Suicide Prevention in Asia. Geneva, Switzerland: World Health Organization; Jan 2008:39-49.
  48. Budenz A, Klassen A, Purtle J, Yom Tov E, Yudell M, Massey P. Mental illness and bipolar disorder on Twitter: implications for stigma and social support. J Ment Health 2020 Apr;29(2):191-199. [CrossRef] [Medline]
  49. Straton N, Jang H, Ng R, Vatrapu R, Mukkamala RR. Computational modeling of stigmatized behaviour in pro-vaccination and anti-vaccination discussions on social media. In: IEEE International Conference on Bioinformatics and Biomedicine. 2019 Feb 6 Presented at: BIBM '19; Nov 18-21, 2019; San Diego, CA, USA p. 2673-2681. [CrossRef]
  50. Koike S, Yamaguchi S, Ojio Y, Ohta K, Ando S. Effect of name change of schizophrenia on mass media between 1985 and 2013 in Japan: a text data mining analysis. Schizophr Bull 2016 May;42(3):552-559 [FREE Full text] [CrossRef] [Medline]
  51. Morgan AJ, Jorm AF. Recall of news stories about mental illness by Australian youth: associations with help-seeking attitudes and stigma. Aust N Z J Psychiatry 2009 Sep;43(9):866-872. [CrossRef] [Medline]
  52. Dietrich S, Heider D, Matschinger H, Angermeyer MC. Influence of newspaper reporting on adolescents' attitudes toward people with mental illness. Soc Psychiatry Psychiatr Epidemiol 2006 Apr;41(4):318-322. [CrossRef] [Medline]
  53. Li A, Jiao D, Zhu T. Detecting depression stigma on social media: a linguistic analysis. J Affect Disord 2018 May;232:358-362. [CrossRef] [Medline]
  54. Yom-Tov E, Brunstein-Klomek A, Mandel O, Hadas A, Fennig S. Inducing behavioral change in seekers of pro-anorexia content using internet advertisements: randomized controlled trial. JMIR Ment Health 2018 Feb 22;5(1):e6 [FREE Full text] [CrossRef] [Medline]
  55. Cheng Q, Yom-Tov E. Do search engine helpline notices aid in preventing suicide? Analysis of archival data. J Med Internet Res 2019 Mar 26;21(3):e12235 [FREE Full text] [CrossRef] [Medline]


API: application programming interface
CC: Central China
EC: East China
NC: North China
RF: random forest
SC: South China
SWC: Southwest China


Edited by R Kukafka; submitted 16.01.22; peer-reviewed by L Zhang, A Budenz; comments to author 11.02.22; revised version received 19.02.22; accepted 13.03.22; published 08.04.22

Copyright

©Ang Li, Dongdong Jiao, Tingshao Zhu. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 08.04.2022.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research, is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.