Abstract
Hate is widespread online, hits everyone, and carries negative consequences. Crowd moderation—user-assisted moderation through, e. g., reporting or counter-speech—is heralded as a potential remedy. We explore this potential by linking insights on online bystander interventions to the analogy of crowd moderation as a (lost) public good. We argue that the distribution of costs and benefits of engaging in crowd moderation forecasts a collective action problem. If the individual crowd member has limited incentive to react when witnessing hate, crowd moderation is unlikely to manifest. We explore this argument empirically, investigating several preregistered hypotheses about the distribution of individual-level costs and benefits of response options to online hate using a large, nationally representative survey of Danish social media users (N = 24,996). In line with expectations, we find that bystander reactions, especially costly reactions, are rare. Furthermore, we find a positive correlation between exposure to online hate and withdrawal motivations, and a negative (n-shaped) correlation with bystander reactions.
1 Introduction
Pervasive online hate is a pressing concern in today’s digital landscape (EU Parliament, 2024; UN News, 2023). It knows no bounds and has negative consequences for individuals and society at large (Duggan, 2017; Zuleta and Burkal, 2017; Vidgen et al., 2019; Vogels, 2021; Bor and Petersen, 2022). Despite concerted efforts by social media platforms, current top-down moderation approaches have failed to curb this negative trend (Gillespie, 2018). A significant number of users report witnessing online hate (Duggan, 2017; Vogels, 2021). Exposure triggers negative emotions, including fear, sadness, and anger (Duggan, 2017; Gelber and McNamara, 2016; Vogels, 2021; Woods and Ruscher, 2021), and discourages many from joining online discussions, thus hindering active participation in the digital public sphere and reducing the quality of deliberative democracy (Gurgun et al., 2022; Midtbøen et al., 2017; Nadim and Fladmoe, 2021).
Current approaches to tackling malicious content online mainly involve suspensions and content takedown (Gillespie, 2018), but studies show that, despite widespread agreement that moderation is necessary (Kozyreva et al., 2023; Rasmussen, 2024), social media users do not favor these approaches (Pradel et al., 2024). In response, crowd moderation has emerged as a potential alternative solution (Lampe et al., 2014; Haythornthwaite, 2023; Hettiachchi and Goncalves, 2020). Crowd moderation is a bottom-up, citizen-driven approach where individual users through online bystander reactions (e. g., reporting and counter-speech) counteract hateful postings online and foster a low-hate arena for deliberation,[1] a supplement or alternative to the top-down moderation efforts instated by platform owners and state-level regulation. Compared to professional moderators, users, with their omnipresence and larger reach, possess a unique advantage in monitoring and responding to hate by engaging in real time (Moxey and Bussey, 2020; Porten-Cheé et al., 2020). As native speakers in the discourse of a given platform, users are particularly adept at deciphering subtle nuances of “covert” hate, which algorithms and moderators may struggle to recognize. Moreover, users have a vested interest in fostering a safe and pleasant online environment, as they loathe being targeted by or exposed to hateful content. Throughout the article, we use the term “online hate” to refer to any post containing negative assessment of a person or group primarily for the purpose of expressing one’s negative attitude (Malecki et al., 2021). Online hate does not aim to provide an informed opinion through critical review but merely to express a negative attitude.
We will explore the feasibility of crowd moderation in combating online hate by providing a theoretical argument that links insights on individual bystander interventions to the analogy of crowd moderation as a public goods problem. We argue that the distribution of individual-level costs and benefits of engaging in crowd moderation forecasts a collective action problem, and that the tendency for individual-level bystander passivity situates an enjoyable online debate climate as an undersupplied (if not lost) public good. We theorize that the current lack of effective crowd moderation on social media can be attributed in part to the unfavorable incentive structure, which favors individual-level freeriding in the form of bystander passivity to the extent that crowd moderation in its current form becomes an infeasible strategy.
The article makes two contributions. In the first part, we guide the reader through an alternative, yet complementary, theoretical framework for understanding online bystander passivity to that of the well-known “bystander intervention model” (Darley and Latane, 1968; Fischer et al., 2011). Using the public goods problem as an analogy, we theorize how widespread online bystander passivity in response to online hate can be viewed as undersupply of the public good “crowd moderation.” When individuals abstain from co-producing crowd moderation through online bystander reactions like reporting and counter-speech, the publicly desirable good “crowd moderation of online hate” fails to manifest at the aggregate level. In the second part, we explore self-reported online bystander behavior empirically to gauge the extent to which descriptive and correlational findings predicted[2] by the public goods framework align with empirical patterns identified in a large, nationally representative survey of Danish social media users (N = 24,996). This exemplifies the potential of the theoretical framework.
In the final part, we discuss theoretical and empirical limitations and implications of our findings and how to potentially alter the distribution of individual costs and benefits of engaging in crowd moderation efforts to resolve the public good problem related to undersupply and avoid freeriding. We end the paper with a call for stronger tests of the relations found in this paper following experimental designs (e. g., survey experiments, public goods games).
2 Crowd moderation and the analogy of the public good
Indirect mentions of public goods date back to Loyd and Hume (Laudal, 2019, p. 3) and describe instances where attainment of a mutually preferable situation or product (good) is hampered by the individual’s limited incentive to assume the burden of creating it. This incentive structure can be traced back to features of the good itself. According to its textbook definition, public goods are “those commodities/services that are both non-excludable (it is not possible to stop others who are not actively contributing from enjoying it) and non-rivalrous (one person’s consumption does not reduce the experience of other’s potential consumption)” (Laudal, 2019, 10, italics added). Below, we show that the concept of public goods and associated negative externalities applies to crowd moderation as an undersupplied public good. We acknowledge that crowd moderation may not represent a pure public good given the theoretical option that social media platforms restrict access, i. e., impose excludability (Cornes and Sandler, 1996). However, since this option counters their business model, they are unlikely to use it extensively.
Low-hate online discussions as a public good
The classic examples of public goods in economic theory, like fish in the sea and access to blood transfusions (Abásolo and Tsuchiya, 2014; Cornes and Sandler, 1996, p. 9), are tangible and desirable, but the concept can be equally valuable in fostering understanding of a good like access to low-hate online discussions. While individuals may have diverging views on what constitutes a desirable and pleasant debate environment (Duggan, 2017; Gillespie, 2018), we specify low-hate as resembling a universally desirable trait since social media users generally prefer restrictions on extreme speech on social media (Rasmussen, 2024). Additionally, the public goods analogy does not rely on preference symmetry (Cornes and Sandler, 1996, p. 241).
It can be illustrative to think of low-hate online discussions as we think of voluntary blood banks as a public good given that both feature the core characteristics of public goods. One person receiving a blood transfusion does not (meaningfully) restrict other people’s access to treatment (non-rivalness) (Cornes and Sandler, 1996), and one person enjoying an online debate experience does not take away from other people enjoying it. In addition, access to blood transfusions and low-hate online discourse feature non-excludability, as individuals can enjoy the good without contributing to its production (Abásolo and Tsuchiya, 2014; Cornes and Sandler, 1996, p. 9). Voluntary blood donation schemes rely on many individuals providing a solution to a problem (illness) by donating their private resources (blood). Similarly, the “production” of a pleasant, low-hate online debate environment involves protecting online discussions from “pollution” by hateful rhetoric (problem) through donating time (private resource) to engage in moderation that remedies (solution) the negative influence of hate in online discussions (Gillespie, 2018). Combined, these features of public goods can be expected to produce negative externalities, as they create incentive structures that promote freeriding, leading to an undersupply of valuable, public goods (Cornes and Sandler, 1996, p. 539). If one can enjoy the pleasant online discussion without contributing to moderation, one has limited incentive to help keep online discussions “on track.” Note that in the context of online discussions there are two levels of freeriding: People can be “lurkers” who consume online discussions without contributing to them by posting, and people can enjoy a pleasant online discussion without contributing to crowd moderation. We focus on the latter type of freeriding (which can be considered a “second-order public good,” Fehr and Gächter, 2002; Axelrod, 1986).
3 Costs of engaging in crowd moderation
The underlying assumptions of the concept of “public goods” is that people calculate expected costs and benefits when deciding how to act (Cornes and Sandler, 1996) and that individuals are boundedly rational and (partly) self-interested (Reiss, 2021). If an individual calculates that costs of action outweigh expected benefits, inaction is the likely outcome. We see this in two ways. First, the benefits of crowd moderation, due to its public-good nature, are expected to accrue at a societal level. Second, the very nature of crowd moderation creates a structure in which costs accrue at the individual level. Because the logic of crowd moderation rests on small-scale actions of the many, combining to create large-scale impact at the aggregate level (crowd moderation), most affiliated costs occur at the individual level. The individual user carries the cost of reacting to online hate, yet all spectators and discussion participants enjoy the benefits of the subsequent removal or contestation of hateful statements. This is not unlike the classic examples of freerider problems in the creation of public goods, where “both [or most individuals] will choose not to contribute, meaning that the public good will not be produced, even though the alternative outcome in which both contribute would be Pareto superior” (Reiss, 2021). Given these considerations, we expect crowd moderation efforts (i. e., counter-speech and reporting) to be in short supply and passivity to be the norm. Specifically, we expect that most people abstain from reacting to online hate as discussed in the literature (Carlson and Settle, 2022; Johnson, 2018). This leads us to our first hypothesis:
H1: The most common reaction to witnessing online hate is to remain passive and do nothing (i. e., respond “never react” to all response options in the study design, as described below).
Varied costs of online bystander reactions
Expecting respondents to show reluctance to engage in online bystander reactions because of incentives to freeride begs the question: What are the costs and benefits of reacting when witnessing hate online?
Some costs cut across all types of bystander reactions (Cepollaro et al., 2023), including the time and effort spent evaluating whether content can be deemed hateful and requires reaction, the time and energy it takes to engage in a reaction (press a series of buttons, compose a message, etc.), and the digital literacy required to navigate response options (Crawford and Gillespie, 2016; Obermaier, 2022). Other costs include the negative emotions associated with thoroughly reading a hateful statement and contemplating a reaction (Duggan, 2017; Vogels, 2021; Woods and Ruscher, 2021) and potential damage to social relationships when one engages in heated discussions (Carlson and Settle, 2022).
The literature on online bystander reactions and civil intervention frequently discusses costs in relation to “thresholds” (Hansen et al., 2023; Kunst et al., 2021; Porten-Cheé et al., 2020). Like the least costly response option, do nothing, low threshold reaction options take little effort and are typically enabled by platform tools (Kunst et al., 2021, p. 259; Porten-Cheé et al., 2020, p. 520), e. g., downrating or reporting content to platform moderators. Reporting content as hateful is private and only relays information to platform moderators. High threshold options, e. g., counter-speech in response to hateful content (Kunst et al., 2021, p. 259; Porten-Cheé et al., 2020, p. 521), take more effort and are more costly than reporting because they are publicly visible, also to authors of hateful statements (Kunst et al., 2021, p. 260). The visibility aspect entails potential social repercussions from (un)known spectators (Carlson and Settle, 2022) and makes one a target of future hateful outbursts (Aldamen, 2023; Johnson, 2018, p. 1285; Kutlaca et al., 2020).
We expect higher individually experienced costs to reduce the motivation to assist in crowd moderation, which leads to our second hypothesis:
H2: High-threshold bystander reactions such as “reply” will be less frequent compared to low-threshold reactions such as “report,” such that reporting will be the most frequent behavior, followed by sharing, followed by replying.
One straightforward and high affiliated cost for an individual user is encountering a large amount of online hate. We expect witnessing more online hate to increase motivations to reduce hate through bystander reactions (reporting, counter-speech). As users witness more hate, they also experience the cost of reaction as higher and would be better off remaining passive (Obermaier, 2022, p. 11). However, at some point, the effect should wear off as individuals become too burdened. Two considerations lead us to believe this.
First, up to a point, exposure to online hate indirectly informs individuals of the problem at hand. Witnessing low to medium amounts of hate lets individuals perceive a low-hate online debate environment as a beneficial good worth co-creating and thus to view hateful content as a risk factor.[3] This might lead some to engage in prosocial bystander reactions to attain the good, much like how an increased need for blood transfusions in the aftermath of natural disasters might motivate otherwise passive citizens to temporarily donate. Yet, at high levels of exposure to online hate, the perceived benefits of a pleasant debate environment are likely outweighed by the aggregated negative emotions affiliated with exposure, combined with the increased experienced costs of reacting as the individual observes the harsh debate tone.
Second, exposure to high levels of online hate also raises the individual bystander’s awareness of others’ inactivity, which means limited exposure to counter-speech. This experience may mean that users perceive crowd moderation as a particularly costly act (I will be on my own) and benefits as unlikely to materialize (I cannot overcome this issue alone). We summarize our expectations related to exposure to online hate in our third hypothesis:
H3: A non-linear (inverted u-shape) relationship such that overall (mean) exposure to other-directed online hate correlates positively with overall (mean) prosocial bystander reactions, but there will be a drop at high levels of exposure.
4 Costs of exposure to hateful content online
Considering crowd moderation as an undersupplied public good highlights the negative externalities that stem from the absence of crowd moderation. Hate speech and less severe expressions of hateful content “have consequences beyond the individuals targeted” (Midtbøen et al., 2017, p. 54). Merely observing online hate thus carries negative consequences. For instance, emotional symptoms of observing hate have been identified, such as deep sadness, hurt and fear (Gelber and McNamara, 2016, p. 333). Correspondingly, the Pew Research Center found that observing online hate or harassment carries negative consequences for the individual, like feelings of anxiety (Duggan, 2017; Woods and Ruscher, 2021) and distress (Vogels, 2021). In our fourth hypothesis, we therefore expect witnesses of online hate to express increased negative emotions like sadness, anger, and fear:
H4: Exposure to other-directed online hate is positively related to experiences of negative emotions (sad, angry, fearful), such that higher levels of exposure are associated with higher levels of reported negative emotions.
Research on the consequences of experiencing online hate targeting oneself or a group one identifies with highlights that experiencing online hate is affiliated with a general reluctance to express opinions publicly, and a retreat from public debates (Gurgun et al., 2022; Nadim and Fladmoe, 2021). This echoes findings that individuals are disinclined to engage in heated discussions in fear of damaging social relationships (Carlson and Settle, 2022). We therefore expect individuals facing online hate to express an increased desire to withdraw from online discussions when exposed to higher levels of online hate. As costs accumulate with higher levels of exposure, the motivation to lower experienced costs by withdrawing from discussions increases. This leads to our fifth hypothesis:
H5: Exposure to other-directed online hate is positively related to motivations for withdrawal from political debates on social media, such that higher levels of exposure are associated with higher levels of withdrawal motivations.
For deliberative democracies, reliant on lively exchanges of opinion on topics of societal relevance, withdrawal can pose an important problem (Carlson and Settle, 2022), especially as withdrawal intentions in response to online hate have been found to be particularly high among, e. g., ethnic minorities and women (Vogels, 2021; Andresen et al., 2022).
5 Targeted online hate: Self- vs. other-directed hate
The five hypotheses reflect analyses of online bystander situations, i. e., they refer to situations where individuals experience online hate targeting others (other-directed hate). On the one hand, passivity and intentions to withdraw from future debates in response to online hate should be less pronounced when the hate targets oneself (self-directed). One would expect the benefits of reacting to loom larger when facing self-directed hate, as you gain directly from standing up to it, e. g., through feelings of vindication, while answering other-directed hate only provides benefits at a societal level. On the other hand, inactivity and caution in expressing opinions may be expected to be more prominent in situations of self-directed hate, as the risk of further verbal attacks appears greater, increasing the perceived cost of reacting.
In this article, we focus on bystander situations on social media (i. e., other-directed hate) because of the inherent potential of strength in numbers of the omnipresent silent observers (bystanders). All relevant preregistered hypotheses therefore reflect responses to other-directed hate, but since the public goods framework gives reason to expect differences in the frequency of reacting depending on whether hate is self- or other-directed, we present and discuss results for both scenarios.
6 Method
This study relies on data from a cross-sectional survey study fielded in Denmark among adults (+18) with permanent residence in Denmark in the period of September 27, 2021 to March 23, 2022 (N = 24,996). Below we describe the data and data collection, how we operationalized key variables, and our analysis plan.
Data
Respondents were invited to the survey through e-Boks, a digital mailbox all permanent Danish residents (age 16+) are required to have. The invitation, “Help researchers to understand hate on social media,” contained information about the purpose of the study and was sent to a representative sample of adult Danes based on a sample of 300,000 individuals drawn by Statistics Denmark. Some 16,026 individuals from this sample had no valid account at the time of the study, leaving a net sample of 283,974. We received 24,996 responses, equaling a response rate of approximately 9 %, exceeding our preregistered minimum sample size (10,000 respondents). Sample size was determined not as a function of power to detect a certain effect size, but rather with the goal of using a large-scale, representative sample of the Danish population. No participants met the preregistered exclusion criteria. However, since the survey allowed participants to skip questions they did not want to answer, there are missing values for a sizable portion of participants, and the number of observations differs between statistical tests. The relatively low response rate is not surprising considering that no reminders were sent, and considering expected survey fatigue following the COVID19-pandemic. Still, we are confident that the overall data patterns reported are generalizable to the Danish population. We supplemented the survey data with data on participants’ sex from the Danish national registries.
Measures
The two main variables are respondents’ self-reported exposure to online hate and self-reported reactions to the exposure. To measure exposure to online hate, we presented respondents with three question batteries. The first asked “How often have you personally been the target of hateful comments on …” and listed 9 widely used social media platforms. For each platform, respondents could answer very often, often, once in a while, rarely, and never. The second and third battery used the same list of social media platforms and answer categories but asked how often the respondent had experienced “people you know” and “people you don’t know,” respectively, being targets of hate online. We created two overall indices of how much hate, self- or other-directed, a respondent is exposed to across their social media usage by multiplying experienced hate for each platform by frequency of usage, summing for all platforms the person indicated using. The indices range from 0 to 128, have a mean of 1.38 (self-directed hate) and 7.64 (other-directed hate), and a standard deviation of 4.22 (self-directed hate) and 8.74 (other-directed hate). Since these indices are heavily right-skewed, we followed our preregistered protocol and log-transformed them for most of our analyses (see Appendix D of the supplementary material for model results with untransformed scores). When online hate targets a group, e. g., an ethnic minority, the distinction between self- and other-related hate may be less straightforward, as individuals may infer self-targeting based on group membership, something our operationalizations do not allow us to analyze.
To measure reactions to online hate, including crowd moderation efforts, respondents were asked one or two batteries of questions depending on their indicated exposure to online hate. All respondents who indicated having experienced self- and/or other-directed online hate were asked how often they, in those situations, had “reacted in the following ways.” Possible reactions were: (1) “It made me sad”; (2) “It made me angry”; (3) “It made me scared”; (4) “I responded to the comment”; (5) “I shared the comment with others either on or off social media”;[4] (6) “I reported the comment for violating the guidelines of the social media platform”; (7) “It reduced my desire to participate in online debates”; and (8) “It increased my desire to participate in online debates.” For each item respondents could answer very often, often, once in a while, rarely, and never. Items 1–3 measure emotional reactions; items 4–6 measure crowd moderation efforts; items 7–8 measure intentions to withdraw from online discussions due to exposure to online hate. The distinction between hate targeting “people you know” and “people you do not know” was left out for the measurement of reactions to hate due to the length of the survey. Appendix B lists the range, distribution, mean and standard deviation for all variables, including reaction items.
In addition to the two main variables of exposure and reactions to online hate, we measured several covariates. Following the preregistered expectations, we investigate three covariates: individual personality traits, perceptions of political discussions online, and a measure of “core human values.” A detailed description of these measures and related analyses is available in Appendix D. Full wordings of all questions and answer categories can be found in Appendix A.
Analysis plan
All analyses were carried out using R (version 4.3.2 for Mac OS 14.3). As the distributions of answers to the reaction to online hate items focused on crowd moderation efforts (report, answer, share) were non-normally distributed, we use nonparametric tests where appropriate.
Deviations from preregistration
We deviate from the preregistration in a few instances. We have changed the naming protocol of hypotheses to fit the structure of the article (see Appendix C for an overview of differences in naming protocols). To streamline the terminology used throughout the article, we changed mentions of hate in preregistered hypotheses to “online hate.” Additionally, we had erroneously proposed to analyze differences in reaction frequencies (Hypothesis 4) as between-subject type analyses in the preregistration (Dunn’s all-pairs rank comparison test). Since these differences are within-subject, we instead conduct paired Wilcoxon tests. Finally, we preregistered six additional hypotheses besides the five discussed in the main text. These are presented and analyzed in Appendix D. We have also included two descriptive hypotheses that did not appear in the preregistration (H1 and H4).
7 Results
Descriptive statistics
Overall, participants reported to be exposed to hate on most social media platforms to a considerable degree, although the frequency of exposure varies across platforms and target of hate (see Figure 1).
Also, common responses to hateful content vary widely, depending on whether the experienced hate is self- or other-directed. The most common response to hate regardless of target is to consider withdrawing, followed by anger, and sadness (see Figure 2).
Inferential statistics
In line with H1, we find that for reporting, answering, or sharing hateful content, the most common reaction type is to remain passive with the never option as most frequently chosen by participants both in terms of self- and other-directed hate (see Figure 3 and Table D.3 in Appendix D). Respondents were more prone to abstain from reacting (answer, report, share), i. e., answer never, when the hate was other-directed (answer: 71 %, report: 72 %, share: 80 %) than when it was self-directed (answer: 31 %, report: 58 %, share: 59 %).
H2 predicted that in terms of different prosocial reactions to hate (answering, reporting, sharing), high-threshold reactions (i. e., answering) would be less common than low-threshold reactions (i. e., reporting). For responses to other-directed hate, reflecting classic bystander situations, reporting (M = 0.55, SD = 1.02) was indeed the most frequent action, followed, contrary to our hypothesis, by answering (M = 0.44, SD = 0.79; reporting vs. answering Wilcoxon signed rank test: Z = 8928351.5, p (Holm adjusted) < .001), and only then sharing (M = 0.30, SD = 0.69; answering vs. sharing Wilcoxon signed rank test: Z = 10317004.5, p (Holm adjusted) < .001). For self-directed hate, answering (M = 1.4, SD = 1.25) was more common than reporting (M = 0.93, SD = 1.31; answering vs. reporting, Wilcoxon signed rank test: Z = 3332546, p (Holm adjusted) < .001), followed by sharing (M = 0.79, SD = 1.13; reporting vs. sharing, Wilcoxon signed rank test: Z = 1678001, p (Holm adjusted) < .001). See also Figure 2.
Regarding the relationship between exposure to online hate and prosocial behavior, H3 predicted a non-linear (n-shaped) relationship between exposure to hate and prosocial reactions. Again, we built two models, one for self- and one for other-directed hate, and both showed significant coefficients at moderate to strong levels (according to Acock, 2014) in line with H3. Notably, the relationship between prosocial behavior and self-directed hate was larger than that between prosocial behavior and hate targeting others (see Table 1). Moreover, a graphical inspection of the model results shows that the relationship between exposure to hate and prosocial behavior is not exactly n-shaped. Rather, the positive relationship between exposure to hate and prosocial behavior flattens out towards higher levels of exposure (see Figure 4).
To investigate H4, testing the relationship between exposure to hate and negative emotional responses, we built linear models predicting each of the three negative emotional responses (anger, sadness, fear) separately. For each dependent variable, we built both simple models with only exposure to hate as a predictor variable and robustness check models that added several control variables. All models were in accordance with the hypothesized effect but with small effect sizes (e. g., a 1 % increase in exposure to other-directed hate increased participants’ self-reported anger by about 0.004 points on a 5-point Likert scale), showing that increased exposure to other-directed online hate is weakly associated with increases in negative emotional responses to encountering hate. Table 2 shows the main model results (see Appendix Table D.1 for models with covariates).
|
Prosocial behavior (self-directed hate) |
Prosocial behavior (other-directed hate) |
||||
Predictors |
Beta (std. beta) |
95 % CI (std. CI) |
p |
Beta (std. beta) |
95 % CI (std. CI) |
p |
(Intercept) |
.72 (.02) |
.68 – .76 (-.01 – .05) |
<.001 (.177) |
.05 (.03) |
.04 – .07 (.01 – .04) |
<.001 (<.001) |
Exposure to hate index |
.05 (.38) |
.05 – .06 (.34 – .41) |
<.001 (<.001) |
.04 (.51) |
.04 – .04 (.50 – .53) |
<.001 (<.001) |
Exposure to hate index (squared) |
-.00 (-.02) |
-.00 – -.00 (-.02 – -.01) |
<.001 (<.001) |
-.00 -.03 |
-.00 – -.00 -.03 – -.02 |
<.001 (<.001) |
Observations |
4759 |
18288 |
||||
R2 / R2 adjusted |
.085 / .084 |
.214 / .214 |
Note: Model results for linear models predicting prosocial behavior in relation to self-directed (left) or other-directed (right) hate. Standardized beta and CI values in parentheses.
|
Anger |
Sadness |
Fear |
||||||
Predictors |
Beta |
95 % CI |
p |
Beta |
95 % CI |
p |
Beta |
95 % CI |
p |
(Intercept) |
1.29 |
1.25 – 1.33 |
<.001 |
.93 |
.89 – .97 |
<.001 |
.09 |
.06 – .11 |
<.001 |
Exposure to hate index [log] |
.39 |
.38 – .41 |
<.001 |
.31 |
.29 – .32 |
<.001 |
.15 |
.14 – .17 |
<.001 |
Observations |
18295 |
18292 |
18286 |
||||||
R2 / R2 adjusted |
.089 / .089 |
.052 / .052 |
.034 / .034 |
Note: Linear regression results predicting negative emotional reactions in response to exposure to other-directed hate.
Moving to H5, positing a positive relationship between exposure to hate and motivations to withdraw from social media, we built two linear models predicting withdrawal reactions as a function of exposure to other-directed and self-directed hate respectively (see Table 3). As with H4, we created both models without and with control variables (key demographic variables, personality and political attitudes) as robustness checks.
These robustness models also include an interaction term between encountered hate and prosocial behavior, as we initially preregistered a moderating effect of engaging prosocially with hate on withdrawal (see Appendix table D.4). The results of the models were mixed. Exposure to other-directed hate led to stronger withdrawal reactions (b = 0.16, 95 % CI [0.13, 0.18], p < .001) as predicted in H5, meaning that a 1 % increase in exposure to other-directed hate increased self-reported withdrawal intentions by about 0.002 points. Exposure to self-directed hate showed the opposite relationship (b = -0.15, 95 % CI [-0.23, -0.06], p < .001). The former result was robust to adding covariates or modeling the relationship with non-transformed exposure values, but the latter relationship did not hold when control variables were added (see Appendix table D.4 for full model results of robustness checks). In general, the strength of the relationship between frequency of exposure to hate and withdrawal motivations was weak, with models only explaining 0.2 % (for self-directed hate) to 0.6 % (for other-directed hate) of the observed variance in withdrawal motivations.
|
Withdrawal motivation (self-directed hate) |
Withdrawal motivation (other-directed hate) |
||||
Predictors |
Beta |
95 % CI |
p |
Beta |
95 % CI |
p |
(Intercept) |
1.54 |
1.39 – 1.69 |
<.001 |
1.39 |
1.33 – 1.45 |
<.001 |
Exposure to hate index [log] |
-.15 |
-.23 – -.06 |
<.001 |
.16 |
.13 – .18 |
<.001 |
Observations |
4759 |
18282 |
||||
R2 / R2 adjusted |
.002 / .002 |
.006 / .006 |
Note: Linear model results for the relationship between exposure to self-directed (left) or other-directed (right) hate and expressed motivations to withdraw from social media. Indices for exposure to hate have been log transformed.
8 Discussion
Our findings are in line with the account suggesting that crowd moderation of online hateful content can be viewed as an undersupplied public good. Reflecting the non-excludable and non-rivalrous nature of a low-hate online debate and the concentrated costs and dispersed benefits of obtaining this good via crowd moderation, we find that remaining passive is the most common response to online hate. Furthermore, exposure to online hate—both as victim and as bystander—carries negative emotional costs for the individual. Indeed, wanting to withdraw from social media debates was the most frequent reaction to online hate among all reaction types we assessed. This desire to withdraw from online political discussions further increased with higher exposure to other-directed hate. Yet, the observed effect size for this increase was small, indicating that withdrawal reactions are typical across all levels of exposure to online hate. When people do react to hate, they prefer the low-cost response option of reporting over the high-cost option of counter-speech. Here, the degree of exposure to online hate correlates positively with the self-reported likelihood of engaging in attempts to reestablish pleasant online discussions through reporting and counter-speech at moderate to strong levels, but this tendency drops off at high levels of hate exposure, indicating a turning point at which the perceived costs outweigh the perceived benefits of crowd moderation.
Combined, these findings dampen our hope of crowd moderation as an efficient strategy against online hate. The public good “lens” applied highlights how individual decisions on responding to online hate are linked through the public goods nature of an enjoyable online debate climate sustained by crowd moderation. It further indicates that crowd moderation is unlikely to come about “naturally.” Bystander passivity is especially the norm when online hate is other-directed, which is by far the more common scenario. Engaging in crowd moderation is more likely (while still unlikely) when online hate is self-directed, including costly options like counter-speech.
Overcoming free-riding by incentivizing bystander reactions
Looking at crowd moderation as an undersupplied public good makes it clear that to obtain efficient crowd moderation, we need to remedy the inbuilt incentives to freeride on the crowd moderation efforts of others. We can do so by altering the perceived individual costs or benefits, i. e., by altering how cumbersome it is to react to online hate (individual costs), or how much the individual stands to gain personally from reacting. This can be done in several ways, yet, we limit our discussion to two: (1) encouraging crowd moderation via platform design; and (2) encouraging crowd moderation via interventions.
Studies have shown promising results with regard to boosting bystander reactions by tweaking platform affordances. For example, increasing the visibility of audience size on platforms may reduce perceptions of costs of engaging in crowd moderation by providing “security in numbers” (Buerger, 2021), and may increase benefits by signaling the virtue of moderation to onlookers (DiFranzo et al., 2018; Taylor et al., 2019).
Other platform alterations have been tried to raise the individual’s perceived benefit of reacting to online hate. YouTube’s discontinued Heroes program (Perez, 2016) allowed users, civil society organizations and government agencies that correctly reported content in violation of YouTube Community Guidelines to report multiple videos simultaneously. As compensation for their free labor, “heroes” were rewarded with points that enabled them to access “sneak peeks at new products or even test them out” (Perez, 2016). An alternative reward structure could highlight users who regularly engage in crowd moderation. Facebook’s “Top Fan” badges and Community Awards that allow administrators to single out positive contributions by awarding members with badges like “Insightful” and “Conversation Starter,” could serve as inspiration (Hutchinson, 2019, 2021).
The second strategy of altering perceived costs and benefits of engaging in crowd moderation of online hate entails providing information and training competences. Traditional bystander programs focused on activating bystanders in an offline context in response to, e. g., sexual harassment, have proven effective when delivered as face-to-face training (Kettrey and Marx, 2019; Miller, 2014) and as online training (Ebers and Thomsen, 2022; Salazar et al., 2019), and some studies of interventions targeting online bystanders suggest similar positive results. For instance, findings by Naab et al. indicate that the experienced costs of reporting hate can be lowered by providing “obtrusive intervention information and emphasizing the need for user intervention in comments sections” (Naab et al., 2018, p. 783). Likewise, the cognitive load of contemplating whether something is hateful and necessitates intervention can be reduced by highlighting the “netiquette” of a site (Park et al., 2014). Building on general insights on behavior change interventions, it seems plausible that perceived costs of engaging in crowd moderation of online hate could be reduced by giving individuals advice on how to react, i. e., boost self-efficacy, and by showcasing the effectiveness of bystander intervention, i. e., boost response efficacy (Milne et al., 2000; Banyard et al., 2007; Michie et al., 2014). One opportunity would be to highlight the variety of possible online bystander reactions (reporting, rating, counter-speech, support of victim). Compared to offline bystanders, online bystanders have much greater choice in deciding to react in a visible or invisible way—options that carry different costs. Finally, perceived benefits could be increased by focusing intervention messages on “threat appraisal” (Milne et al. 2000), i. e., explaining how online hate constitutes a threat to democratic deliberation or showcasing how instating a norm of bystander engagement would benefit the individual personally in the future.
Future studies and the need for stronger tests
Our findings indicate that viewing a pleasant online debate environment produced through crowd moderation as an undersupplied public good constitutes a valuable future research agenda. Nevertheless, there are good reasons to advocate for more research of crowd moderation as a public good, especially in designs suited for causal identification of real-life behavior.
First, given that our findings rely on cross-sectional data and correlational analyses, a causal interpretation of our results is ill advised. Our findings need to be corroborated in stronger designs of an experimental nature. Experimental designs would be more suited to disentangle how a change in perceived costs or benefits of engaging in crowd moderation of online hate affects reactions. Whether in the form of survey experiments or classic public goods games, the aim should be of a testing nature.
Second, we rely on self-reported measures of key variables. We asked participants to indicate, e. g., how often they have been exposed to hate on social media and how they responded. Although we cannot rule out that respondents over-report prosocial responses to hate, we believe that the risk of overestimation is small since the electronic survey was anonymous and the average frequency of, e. g., reporting and counter-speech in our study is low (more than half report never having engaged in any kind of crowd moderation).
Third, surveying respondents in only one country limits the opportunities for generalization. Denmark has high levels of internet penetration and social media usage, and a relatively low degree of political polarization, which has been found to correlate with lower levels of online hate speech in political discussions (Bail, 2021). To learn whether our findings travel to countries with higher degrees of political polarization requires further studies. Nevertheless, our study is the most comprehensive examination in Europe of patterns in bystander reactions to online hate.
Lastly, while we have argued that interventions and platform design alterations can potentially mobilize more people to engage in crowd moderation, it remains to be seen whether such strategies are efficient at scale. The assumption that crowd moderation reduces the negative implications of online hate also needs further corroboration. Early indications show promise (Hangartner et al., 2021), but for it to be a public good, we must establish whether crowd moderation is rightly viewed as a good at all.
9 Conclusion
We have argued that by theorizing engagement in crowd moderation of online hate as a public goods problem, we have thrown new light on the feasibility of crowd moderation as an alternative strategy to top-down content moderation. We have shown how the perceived costs and benefits of engagement at the individual level forecast freerider problems in the aggregate. Our empirical exploration of survey data from Denmark has shown that bystanders are prone to remain passive when facing online hate, and favor low-cost reaction options like reporting over the high-cost option of counter-speech. We have also emphasized the negative consequences of exposure to hate for individuals who report experiencing negative emotions (sad, fearful, angry), and for society at large through increased motivations to withdraw from online discussions. Yet, as highlighted, our study has limitations, and we view it as a first guiding step for an emerging research agenda. We hope others will engage in further testing of this proposed theoretical framework on crowd moderation of online hate.
Open Science Practices and Research Ethics statement
Our study was preregistered, and all deviations from it are clearly reported. All measures, data exclusions, as well as sample size calculations are documented. All materials (instructions, data, analysis code, other materials including additional analyses) are posted on OSF. The study received formal IRB approval from the authors’ main institution, and all participants were thoroughly debriefed.
References
Abásolo, I., & Tsuchiya, A. (2014). Blood donation as a public good: An empirical investigation of the free rider problem. The European Journal of Health Economics, 15(3), 313–321. https:///doi.org/10.1007/s10198-013-0496-x10.1007/s10198-013-0496-xSearch in Google Scholar
Acock, A. C. (2014). A gentle introduction to Stata (4th ed.). Stata Press.Search in Google Scholar
Aldamen, Y. (2023). Xenophobia and hate speech towards refugees on social media: Reinforcing causes, negative effects, defense and response mechanisms against that speech. Societies, 13(4), Article 4. https:///doi.org/10.3390/soc1304008310.3390/soc13040083Search in Google Scholar
Andresen, M. J., Karg, S. T. S., Rasmussen, S. H. R., Pradella, L., Rasmussen, J., Lindekilde, L., & Petersen, M. B. (2022). Danskernes oplevelse af had på sociale medier [Danes’ experiences of hate on social media]. Aarhus University Research Output: https://pure.au.dk/portal/en/publications/danskernes-oplevelse-af-had-p%C3%A5-sociale-medierSearch in Google Scholar
Axelrod, R. (1986). An evolutionary approach to norms. American Political Science Review, 80(4), 1095–1111. https:///doi.org/10.2307/196085810.1017/S0003055400185016Search in Google Scholar
Bail, C. (2021). Breaking the social media prism: How to make our platforms less polarizing. Princeton University Press.10.1515/9780691216508Search in Google Scholar
Banyard, V. L., Moynihan, M. M., & Plante, E. G. (2007). Sexual violence prevention through bystander education: An experimental evaluation. Journal of Community Psychology, 35(4), 463–481. https:///doi.org/10.1002/jcop.2015910.1002/jcop.20159Search in Google Scholar
Bor, A., & Petersen, M. B. (2022). The psychology of online political hostility: A comprehensive, cross-national test of the mismatch hypothesis. American Political Science Review, 116(1), 1–18. https:///doi.org/10.1017/S000305542100088510.1017/S0003055421000885Search in Google Scholar
Buerger, C. (2021). #iamhere: Collective counterspeech and the quest to improve online discourse. Social Media + Society, 7(4), 205630512110638. https:///doi.org/10.1177/2056305121106384310.1177/20563051211063843Search in Google Scholar
Carlson, T. N., & Settle, J. E. (2022). What goes without saying: Navigating political discussion in America (1st ed.). Cambridge University Press.10.1017/9781108912495Search in Google Scholar
Cepollaro, B., Lepoutre, M., & Simpson, R. M. (2023). Counterspeech. Philosophy Compass, 18(1), e12890. https:///doi.org/10.1111/phc3.1289010.1111/phc3.12890Search in Google Scholar
Cornes, R., & Sandler, T. (1996). The theory of externalities, public goods, and club goods (2nd ed.). Cambridge University Press. https://doi.org/10.1017/CBO978113917431210.1017/CBO9781139174312Search in Google Scholar
Crawford, K., & Gillespie, T. (2016). What is a flag for? Social media reporting tools and the vocabulary of complaint. New Media & Society, 18(3), 410–428. https:///doi.org/10.1177/146144481454316310.1177/1461444814543163Search in Google Scholar
Darley, J. M., & Latane, B. (1968). Bystander intervention in emergencies: Diffusion of responsibility. Journal of Personality and Social Psychology, 8(4, Pt.1), 377–383. https:///doi.org/10.1037/h002558910.1037/h0025589Search in Google Scholar
DiFranzo, D., Taylor, S. H., Kazerooni, F., Wherry, O. D., & Bazarova, N. N. (2018). Upstanding by design: Bystander intervention in cyberbullying. Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, 1–12.10.1145/3173574.3173785Search in Google Scholar
Duggan, M. (2017, July 11). Witnessing online harassment. Pew Research Center: Internet, Science & Tech. https://www.pewresearch.org/internet/2017/07/11/witnessing-online-harassment/Search in Google Scholar
Ebers, A., & Thomsen, S. L. (2022). Evaluating a gamified bystander program: Evidence from two randomized online field experiments. Hannover Economic Papers (HEP), dp-692. https://ideas.repec.org/p/han/dpaper/dp-692.htmlSearch in Google Scholar
EU Parliament. (2024, January 18). Time to criminalise hate speech and hate crime under EU law [Press release]. https://www.europarl.europa.eu/news/en/press-room/20240112IPR16777/time-to-criminalise-hate-speech-and-hate-crime-under-eu-lawSearch in Google Scholar
Fehr, E., & Gächter, S. (2002) Altruistic punishment in humans. Nature, 415(1), 137–140. https:///doi.org/10.1038/415137a10.1038/415137aSearch in Google Scholar
Fischer, P., Krueger, J. I., Greitemeyer, T., Vogrincic, C., Kastenmüller, A., Frey, D., Heene, M., Wicher, M., & Kainbacher, M. (2011). The bystander-effect: A meta-analytic review on bystander intervention in dangerous and non-dangerous emergencies. Psychological Bulletin, 137(4), 517–537. https:///doi.org/10.1037/a002330410.1037/a0023304Search in Google Scholar
Gelber, K., & McNamara, L. (2016). Evidencing the harms of hate speech. Social Identities, 22(3), 324–341. https:///doi.org/10.1080/13504630.2015.112881010.1080/13504630.2015.1128810Search in Google Scholar
Gillespie, T. (2018). Custodians of the internet: Platforms, content moderation, and the hidden decisions that shape social media. Yale University Press.10.12987/9780300235029Search in Google Scholar
Gurgun, S., Arden-Close, E., Phalp, K., & Ali, R. (2022). Online silence: Why do people not challenge others when posting misinformation? Internet Research, 33(5), 1928–1948. https:///doi.org/10.1108/INTR-06-2022-040710.1108/INTR-06-2022-0407Search in Google Scholar
Hangartner, D., Gennaro, G., Alasiri, S., Bahrich, N., Bornhoft, A., Boucher, J., Demirci, B. B., Derksen, L., Hall, A., Jochum, M., Munoz, M. M., Richter, M., Vogel, F., Wittwer, S., Wüthrich, F., Gilardi, F., & Donnay, K. (2021). Empathy-based counterspeech can reduce racist hate speech in a social media field experiment. Proceedings of the National Academy of Sciences, USA, 118(50), e2116310118. https:///doi.org/10.1073/pnas.211631011810.1073/pnas.2116310118Search in Google Scholar
Hansen, T. M., Lindekilde, L., & Karg, S. T. S. (2023). The devil is in the detail: Reconceptualising bystander reactions to online political hostility. Behaviour & Information Technology, 0(0), 1–14. https:///doi.org/10.1080/0144929X.2023.228265310.1080/0144929X.2023.2282653Search in Google Scholar
Haythornthwaite, C. (2023). Moderation, networks, and anti-social behavior online. Social Media + Society, 9(3), 20563051231196874. https:///doi.org/10.1177/2056305123119687410.1177/20563051231196874Search in Google Scholar
Hettiachchi, D., & Goncalves, J. (2020). Towards effective crowd-powered online content moderation. Proceedings of the 31st Australian Conference on Human-Computer-Interaction, Australia, 342–346. https:///doi.org/10.1145/3369457.336949110.1145/3369457.3369491Search in Google Scholar
Hutchinson, A. (2019, August 24). Facebook adds more page badges to encourage engagement. Social Media Today. https://www.socialmediatoday.com/news/facebook-adds-more-page-badges-to-encourage-engagement/561625/Search in Google Scholar
Hutchinson, A. (2021, September 21). Facebook tests new “community awards” to encourage engagement in groups. Social Media Today. https://www.socialmediatoday.com/news/facebook-tests-new-community-awards-to-encourage-engagement-in-groups/606966/Search in Google Scholar
Johnson, B. G. (2018). Tolerating and managing extreme speech on social media. Internet Research, 28(5), 1275–1291. https:///doi.org/10.1108/IntR-03-2017-010010.1108/IntR-03-2017-0100Search in Google Scholar
Kettrey, H. H., & Marx, R. A. (2019). The effects of bystander programs on the prevention of sexual assault across the college years: A systematic review and meta-analysis. Journal of Youth and Adolescence, 48, 212–227. https:///doi.org/10.1007/s10964-018-0927-110.1007/s10964-018-0927-1Search in Google Scholar
Kozyreva, A., Herzog, S. M., Lewandowsky, S., Hertwig, R., Lorenz-Spreen, P., Leiser, M., & Reifler, J. (2023). Resolving content moderation dilemmas between free speech and harmful misinformation. Proceedings of the National Academy of Sciences, USA, 120(7), e2210666120. https:///doi.org/10.1073/pnas.221066612010.1073/pnas.2210666120Search in Google Scholar
Kunst, M., Porten-Cheé, P., Emmer, M., & Eilders, C. (2021). Do “Good Citizens” fight hate speech online? Effects of solidarity citizenship norms on user responses to hate comments. Journal of Information Technology & Politics, 18(3), 258–273. https:///doi.org/10.1080/19331681.2020.187114910.1080/19331681.2020.1871149Search in Google Scholar
Kutlaca, M., Becker, J., & Radke, H. (2020). A hero for the outgroup, a black sheep for the ingroup: Societal perceptions of those who confront discrimination. Journal of Experimental Social Psychology, 88, 103832. https:///doi.org/10.1016/j.jesp.2019.10383210.1016/j.jesp.2019.103832Search in Google Scholar
Lampe, C., Zube, P., Lee, J., Park, C. H., & Johnston, E. (2014). Crowdsourcing civility: A natural experiment examining the effects of distributed moderation in online forums. Government Information Quarterly, 31(2), 317–326. https:///doi.org/10.1016/j.giq.2013.11.00510.1016/j.giq.2013.11.005Search in Google Scholar
Laudal, T. (2019). A new approach to the economics of public goods. Routledge.10.4324/9780429316616Search in Google Scholar
Malecki, W. P., Kowal, M., Dobrowolska, M., & Sorokowski, P. (2021). Defining online hating and online haters. Frontiers in Psychology, 12, 744614. https:///doi.org/10.3389/fpsyg.2021.74461410.3389/fpsyg.2021.744614Search in Google Scholar
Michie, S., Atkins, L., & West, R. (2014). The behaviour change wheel: A guide to designing interventions. Silverback Publishing.Search in Google Scholar
Midtbøen, A. H., Steen-Johnsen, K., Thorbjørnsrud, K., Fladmoe, A., Nadim, M., Moe, H., Ihlebæk, K. A., Thorseth, I. E., Colbjørnsen, T., & Enjolras, B. (2017). Boundary struggles: Contestations of free speech in the Norwegian public sphere. Nordic Open Access Scholarly Publishing.10.23865/noasp.16Search in Google Scholar
Miller, N. L. (2014). The secret success of nonproliferation sanctions. International Organization, 68(4), 913–944. https:///doi.org/10.1017/S002081831400021610.1017/S0020818314000216Search in Google Scholar
Milne, S., Sheeran, P., & Orbell, S. (2000). Prediction and intervention in health-related behavior: A meta-analytic review of protection motivation theory. Journal of Applied Social Psychology, 30(1), 106–143. https:///doi.org/10.1111/j.1559-1816.2000.tb02308.x10.1111/j.1559-1816.2000.tb02308.xSearch in Google Scholar
Moxey, N., & Bussey, K. (2020). Styles of bystander intervention in cyberbullying incidents. International Journal of Bullying Prevention, 2(1), 6–15. https:///doi.org/10.1007/s42380-019-00039-110.1007/s42380-019-00039-1Search in Google Scholar
Naab, T. K., Kalch, A., & Meitz, T. G. (2018). Flagging uncivil user comments: Effects of intervention information, type of victim, and response comments on bystander behavior. New Media & Society, 20(2), 777–795. https:///doi.org/10.1177/146144481667092310.1177/1461444816670923Search in Google Scholar
Nadim, M., & Fladmoe, A. (2021). Silencing women? Gender and online harassment. Social Science Computer Review, 39(2), 245–258. https:///doi.org/10.1177/089443931986551810.1177/0894439319865518Search in Google Scholar
Obermaier, M. (2022). Youth on standby? Explaining adolescent and young adult bystanders’ intervention against online hate speech. New Media & Society, 0(0) 1,46144482211254. https:///doi.org/10.1177/14614448221125417Search in Google Scholar
Park, S., Na, E.-Y., & Kim, E. (2014). The relationship between online activities, netiquette and cyberbullying. Children and Youth Services Review, 42, 74–81. https:///doi.org/10.1016/j.childyouth.2014.04.00210.1016/j.childyouth.2014.04.002Search in Google Scholar
Perez, S. (2016, September 21). YouTube enlists volunteers to moderate its site via a new “YouTube Heroes” program. TechCrunch. https://techcrunch.com/2016/09/21/youtube-enlists-volunteers-to-moderate-its-site-via-a-new-youtube-heroes-program/Search in Google Scholar
Porten-Cheé, P., Kunst, M., & Emmer, M. (2020). Online civic intervention: A new form of political participation under conditions of a disruptive online discourse. International Journal of Communication, 14(0), Article 0. https://ijoc.org/index.php/ijoc/article/view/10639Search in Google Scholar
Pradel, F., Zilinsky, J., Kosmidis, S., & Theocharis, Y. (2024). Toxic speech and limited demand for content moderation on social media. American Political Science Review. Advance online publication. https:///doi.org/10.1017/S000305542300134X10.1017/S000305542300134XSearch in Google Scholar
Rasmussen, J. (2024). The (limited) effects of target characteristics on public opinion of hate speech laws. OSF. https:///doi.org/10.31234/osf.io/j4nucSearch in Google Scholar
Reiss, J. (2021). Public goods. In E. N. Zalta (Ed.), The Stanford Encyclopedia of Philosophy (Fall 2021). Metaphysics Research Lab, Stanford University. https://plato.stanford.edu/archives/fall2021/entries/public-goods/Search in Google Scholar
Salazar, L. F., Vivolo-Kantor, A., & Schipani-McLaughlin, A. M. (2019). Theoretical mediators of RealConsent: A web-based sexual violence prevention and bystander education program. Health Education & Behavior, 46(1), 79–88. https:///doi.org/10.1177/109019811877912610.1177/1090198118779126Search in Google Scholar
Taylor, S. H., DiFranzo, D., Choi, Y. H., Sannon, S., & Bazarova, N. N. (2019). Accountability and empathy by design: Encouraging bystander intervention to cyberbullying on social media. Proceedings of the ACM on Human-Computer Interaction, 3(CSCW), 1–26. https:///doi.org/10.1145/335922010.1145/3359220Search in Google Scholar
UN News. (2023, January 28). Hate speech: A growing, international threat. UN News. https://news.un.org/en/story/2023/01/1132597Search in Google Scholar
Vidgen, B., Margetts, H., & Harris, A. (2019). How much online abuse is there. Alan Turing Institute. https://www.turing.ac.uk/sites/default/files/2019-11/online_abuse_prevalence_full_24.11.2019_-_formatted_0.pdfSearch in Google Scholar
Vogels, E. A. (2021, January 13). The state of online harassment. Pew Research Center: Internet, Science & Tech. https://www.pewresearch.org/internet/2021/01/13/the-state-of-online-harassment/Search in Google Scholar
Wintterlin, F., Frischlich, L., Boberg, S., Schatto-Eckrodt, T., Reer, F., & Quandt, T. (2021). corrective actions in the information disorder. the role of presumed media influence and hostile media perceptions for the countering of distorted user-generated content. Political Communication, 38(6), 773–791. https:///doi.org/10.1080/10584609.2021.188882910.1080/10584609.2021.1888829Search in Google Scholar
Woods, F. A., & Ruscher, J. B. (2021). Viral sticks, virtual stones: Addressing anonymous hate speech online. Patterns of Prejudice, 55(3), 265–289. https:///doi.org/10.1080/0031322X.2021.196858610.1080/0031322X.2021.1968586Search in Google Scholar
Zuleta, L., & Burkal, R. (2017). Hate speech in the public online debate. The Danish Institute for Human Rights. https://www.humanrights.dk/files/media/migrated/hate_speech_in_the_public_online_debate_eng_2017.pdfSearch in Google Scholar
© 2024 bei den Autoren, publiziert von Walter de Gruyter GmbH, Berlin/Boston
Dieses Werk ist lizensiert unter einer Creative Commons Namensnennung 4.0 International Lizenz.