Nothing Special   »   [go: up one dir, main page]

skip to main content
research-article
Open access

Addressing Interpersonal Harm in Online Gaming Communities: The Opportunities and Challenges for a Restorative Justice Approach

Published: 25 September 2023 Publication History

Abstract

Most social media platforms implement content moderation to address interpersonal harms such as harassment. Content moderation relies on offender-centered, punitive approaches, e.g., bans and content removal. We consider an alternative justice framework, restorative justice, which aids victims in healing, supports offenders in repairing the harm, and engages community members in addressing the harm collectively. To assess the utility of restorative justice in addressing online harm, we interviewed 23 users from Overwatch gaming communities, including moderators, victims, and offenders; such communities are particularly susceptible to harm, with nearly three quarters of all online game players suffering from some form of online abuse. We study how the communities currently handle harm cases through the lens of restorative justice and examine their attitudes toward implementing restorative justice processes. Our analysis reveals that cultural, technical, and resource-related obstacles hinder implementation of restorative justice within the existing punitive framework despite online community needs and existing structures to support it. We discuss how current content moderation systems can embed restorative justice goals and practices and overcome these challenges.

1 Introduction

Social media platforms frequently address online interpersonal harm, such as harassment, through content moderation; this involves reviewing user-submitted content for appropriateness and sanctioning contributors that violate the platform’s rules [42, 94]. However, despite efforts in research and industry to improve moderation practices in recent years, the number of people experiencing severe forms of harassment continues to grow. In 2014, 15% of Americans reported experiencing severe harassment, including physical threats, stalking, sexual harassment, and sustained harassment [33]. That number grew to 18% in 2017 and 25% in 2021. Further, many people report simultaneously experiencing multiple forms of severe harassment [34, 125]. Research shows that online harms are insufficiently addressed by platforms’ and communities’ current approaches [23, 35, 77, 126], and 32% of Americans say that social media companies are doing a poor job at addressing online harassment on their platforms [125]. Alternative approaches are desperately needed, but what principles should guide them, and how would they work in practice?
Restorative justice is a framework that argues for repairing harm and restoring individuals and communities after harm has occurred. In this article, our goal is to draw from restorative justice philosophy and practice to study how an online gaming community currently addresses—and might alternatively address—interpersonal harm. We focus on restorative justice here because it has an established offline practice and has been successfully institutionalized to address harm in other contexts, such as schools and prisons [6, 73, 75]. In recent years, the HCI and CSCW communities have also explored its utility in addressing online harm [12, 69, 106], and we build upon these efforts, as well.
Restorative justice focuses on providing care, support, and in other ways meeting people’s needs after harm has occurred. It has three major principles: (1) identify and address the victim’s needs related to the harm, (2) support the offender in taking accountability and working to repair the harm, and (3) engage the community in the process to support victims and offenders and heal collectively [80, 132]. In practice, restorative justice addresses harm differently than more common punitive models. The main tool for action in a punitive justice model, as embodied in content moderation, is punishing the rule violator. In contrast, in restorative justice, it is communication among the harmed person, the offender, and the community. For instance, in a common restorative justice practice called a victim-offender conference, the victim and offender meet to discuss the harm and how to address it under the guidance of a facilitator; interested community members are also invited to join this conversation since the conference aims at addressing the needs and obligations of all three parties involved. A follow-up process may include apologies or community service by the offenders [132].
This article uses the three restorative justice principles described above and its common practices (e.g., the victim-offender conference) as a vehicle to study the perspectives and practices of victims, offenders, and moderators during instances of online harm. First, we use the principles to evaluate current practices for addressing interpersonal harm and identify the potential need for restorative justice practices. We focus on the experiences of victims, offenders, and moderators, who are key stakeholders and participants in restorative justice conferences (with the moderator acting as facilitator). Second, we use the victim-offender conference to identify the benefits and challenges of practicably implementing restorative justice practices in an online setting.
We study harm cases in the Overwatch gaming community, which spans two major platforms : the Overwatch platform on which the game is played1 and the Discord platform2 on which gaming discussions, teammate selection, and match organization occur. Online gaming communities have long suffered from severe and frequent incidents of online harm [2, 9, 51]. Our analysis of Overwatch, a multi-player game, lets us explore such harm in the context of different types of user relationships, including competition and collaboration. We interviewed self-identified victims (people who have been harmed), offenders (people who have harmed others), and moderators who dealt with the cases being discussed. Our interview protocol resembles the restorative justice practice of pre-conferencing, which is used to learn people’s history and preferences, explain restorative justice to them, and evaluate the appropriateness of holding a victim-offender conference [132]. Additionally, given that restorative justice has been chiefly developed through practice [122], we deepened our understanding of its principles and practices by conducting two interviews with offline restorative justice practitioners.
We find that current, punitive online moderation processes do not effectively stop the perpetuation of harm. First, content moderation is offender-centered and does not address victims’ needs, such as receiving support or healing from harm. Though victims may report individual offenders, they continue to receive harm in a community where abuse is prevalent. Second, content moderation directs offenders’ attention to the punishment they receive instead of the damage they cause. When punishment is ineffective, as is often the case, there are no alternative ways to hold offenders accountable. Finally, community members with a punitive mindset may further perpetuate harm by not acknowledging the harmed person’s experiences or reacting punitively toward perpetrators or victims, particularly when harm cases are complex and layered.
Our findings show that some current moderation practices align with restorative justice, and a few participants have attempted to implement restorative justice practices in their own online communities. Some victims and offenders also expressed needs that align with restorative justice values. However, applying restorative justice online is not straightforward: there are structural, cultural, and resource-related obstacles to implementing a new approach within the existing punitive framework. We elaborate on the potential challenges of implementing restorative justice online and propose ways to design and embed its practices in online communities.
Our work contributes to a growing line of research that applies alternative justice frameworks to address online harm [23, 28, 47, 48, 86, 103, 106]. By evaluating the moderation practice of the Overwatch gaming community through a restorative justice lens and identifying key stakeholders’ preferences for the justice framework, our work sheds light on ways to address online harm that go beyond simply maintaining healthy content and working within a perpetrator-oriented model. We highlight how restorative justice has the potential to reduce the continuance of harm and improve community culture in the long run.

2 Related Work

We now review related work on online harm and content moderation. We also describe the restorative justice framework and its current applications.

2.1 Online Harm and Content Moderation

Online harm can refer to a myriad of toxic behaviors, such as public shaming [98], trolling [22], and invasion of privacy [77]. Our research focuses on interpersonal harm (as opposed to self-directed or collective harm), which is the harm that occurs in the interaction between two or several individuals [72]. Most of the interpersonal harm we study falls under Duggan’s description of online harassment, which includes six categories: offensive name-calling, purposeful embarrassment, physical threats, sustained harassment, stalking, and sexual harassment [34].
Online gaming communities are particularly susceptible to harm: nearly three-quarters of all online game players have suffered some form of online abuse [76]. Within the context of online games, “toxicity” describes abusive communication (e.g., verbal or textual harassment) as well as disruptive gaming behaviors (e.g., exploiting glitches, sabotaging teammates) [2, 93]. Toxic behaviors hurt the gaming experience of involved players; prior research has found that toxicity during game play can negatively affect individual and team performance [83]. Even so, it is pervasive and normalized as part of the competitive game culture [2, 9, 51]. Notably, this toxicity intersects with harm toward certain groups, for example, females [38, 100, 101, 111], gender minorities [99], and people of color [45]. Reid et al. have argued that platforms should offer in-game support tools that immediately address toxicity or help players cope with harm after it occurs [92]. We respond to this call by examining how introducing restorative justice tools and processes may address harm in online gaming communities.
Online social media and gaming platforms currently address harm through content moderation, which usually involves punitive measures such as removing content, muting, or banning offenders [42, 94]. Online platforms’ moderators can be volunteers who are platform users or commercial content moderators hired by social media companies [94, 110]. In recent years, social media platforms have also begun using automated, AI-based tools such as bots to help enact moderation [11, 61, 66].
As the most widely applied model of addressing online harm, content moderation faces many implementation challenges and is criticized for its limits in building healthy communities. One key challenge is the sheer amount of labor required: as social media platforms grow, moderators must address increasing numbers of harm cases. Another challenge is the emotional labor required to moderate potentially upsetting antisocial content [32, 94, 129]. At the same time, moderation is not always efficient and effective in addressing harm: content policies and their implementations have often failed to sufficiently remove disturbing material like fake news (a colloquial term for false or misleading content presented as news) [3], alt-right trolls [88, 97], and revenge porn [24, 124].
Researchers have examined ways to improve the current moderation processes, for example, through setting positive examples and social norms [20, 109], providing explanations for post removals [56, 57], supporting creators in configuring automated moderation filters [59], and placing warning labels on inappropriate content [19, 85]. We add to these efforts by taking a user-centered approach that interrogates how victims, offenders, and community members experience the current content moderation approach in addressing harm through the perspective of restorative justice and identifying opportunities for improvements.
Prior research has found that victims often have needs that content moderation does not address, e.g., the need to make sense of the harm [121, 131], receive emotional support and validation [120], or transform the online environment to prevent harm from happening [131]. Researchers have also developed tools to support victims beyond content moderation, such as tools for social support [13, 78] or collecting and documenting evidence [44, 115]. These tools are often external to the platform on which the harm occurs. We examine how in-house, community-centered restorative processes, which allow a deeper reflection on online harms, may help address victims’ needs.
In recent years, researchers have begun asking questions about the role that social media platforms play in the realization of values important to public discourse, such as freedom of expression, transparency, protection from discrimination, and personal security and dignity [43, 50, 87, 116, 117, 118]. Many scholars have offered ethical frameworks to guide platforms’ content moderation efforts. Perhaps the most prominent framework for ethical conduct in social media moderation is the Santa Clara Principles [105], which outline “minimum levels of transparency and accountability” and propose three principles for providing meaningful due process: (1) publishing data about removed posts, (2) notifying users affected by moderation, and (3) granting moderated users a right to appeal. Shannon Bowen applies Kantian deontology to social media content management decisions, advocating that public relations practitioners rely on universal principles such as dignity, fairness, honesty, transparency, and respect when communicating via social media [16]. A variety of other frameworks, originally developed in offline settings, usefully inform the ethics of content moderation [18]. Yet, restorative justice is an especially informative framework for our narrow focus on addressing online harm because it prioritizes individuals and relationships rather than content; it therefore forms the focal point of our research.
We contribute to an emerging line of research that examines the values of a diverse range of stakeholders of online platforms to inform governance practices. For example, Schoenebeck et al. inquired about various victims’ preferences for moderation measures based on frameworks such as restorative justice, economic justice, and racial justice [106]. Marwick developed an explanatory model of networked harassment by interviewing people who had experienced harassment and Trust & Safety workers at social platforms [79]. Jhaver et al. conceptualized the distinctions between controversial speech and online harassment by talking to both victims and alleged perpetrators of harassment [58, 60]. Supporting such inquiries, Helberger et al. argues that the realization of public values should be a collective effort of platforms, users, and public institutions [50]. Our research builds on this line of work by considering victims, offenders, and community members’ needs and values in harm reparation.

2.2 Restorative Justice

Punitive justice is often most familiar to us since it is the culturally dominant way we deal with harm. Therefore, to explain restorative justice, we first contrast it with punitive justice. Next, we detail common practices in restorative justice and their application in offline scenarios.
In Western cultures, the dominant model for justice when harm occurs involves punishing the offender [119]. This punitive justice model holds that harm violates rules and offenders should suffer in proportion to their offense [39]. The central focus of this model is on punishing and excluding the offender. However, the victim’s concerns about the effects of the offense are rarely taken into account. Further, this model does not help offenders become aware of the negative impact they cause and take accountability to repair the harm [119].
Analyzing how harm is addressed in early MUD (multi-user dungeon) communities, Elizabeth Reid observed that “Punishment on MUDs often shows a return to the medieval. While penal systems in the western nations...have ceased to concentrate upon the body of the condemned as the site for punishment, and have instead turned to “humane” incarceration and social rehabilitation, the exercise of authority on MUDs has revived the old practices of public shaming and torture” [91]. Two decades later, the public spectacle of punishment still prevails on current digital platforms.

2.2.1 Restorative Justice Principles.

Restorative justice provides an alternative way to address harm. It supports the belief that harm is a violation of people and relationships rather than merely a breach of rules [123]. It puts victims at the center of the process and seeks to repair the harm they suffer from the offense. Restorative justice has three major principles [80]: (1) provide support and healing for victims, (2) help offenders realize the consequences of their wrong-doing and repair the harm, and (3) encourage communities to provide support for both victims and offenders and to heal collectively. Multiple levels of communities can be primarily affected by harm, including the local community where the harm occurs or the broader society [80]. Our study addresses victims and offenders in the Overwatch gaming community.

2.2.2 Restorative Justice Practices.

Restorative justice has been successfully applied in a myriad of settings, such as the criminal justice system, schools, and workplaces [123, 130]. When formalized within an organization, restorative justice processes can be embedded within a punitive justice system to use on selected types of harm cases [123, 130]. Those who do not want to proceed using this approach or cannot reach a consensus during the restorative justice process are directed to the punitive justice system [6] for redress.
A widely used practice in restorative justice is the victim-offender conference [132]. Here, victims and offenders meet with a restorative justice facilitator to discuss three core questions: (1) what has happened? (2) who has been affected and how? and (3) what is needed to repair the harm? The facilitator mediates this process to ensure that victims and offenders have equal footing and helps move the parties toward reaching a consensus. Other forms of restorative justice meetings, such as family-group conferences, include additional community members, such as family and friends of victims and offenders [6, 132]. These practices embed values and principles of restorative justice to meet the needs of all parties involved, including the victim, offender, and community members. In this article, we use victim-offender conferences as an exemplar to inquire about participants’ preferences for online restorative justice practices.
Restorative justice practitioners emphasize the importance of preparation in advance of the victim-offender conference. A restorative justice facilitator first meets separately with the offender and the victim before the conference in a process called a pre-conference [132]. During these meetings, the facilitator introduces the restorative justice framework to the victim and offender and asks them the same questions they will be asked during the conference. After these meetings, the victim-offender conference happens only when both parties agree to meet voluntarily to repair the harm. Additionally, the facilitator acts as the gatekeeper to determine whether victims and offenders can meet to reach a desired outcome without causing more harm. In both pre- and victim-offender conferencing, the facilitator does not make decisions for victims or offenders about how to address the harm but guides them to reflect on the harm through the restorative justice framework [15]. For this research, the first author received training and acted as a facilitator in pre-conferencing with the participants to ask about their attitudes toward engaging in a restorative justice process to address online harm.

2.2.3 Restorative Justice Outcomes.

Restorative justice does not encourage punishment as the desired outcome. Instead, it focuses on the obligation to repair harm and heal those hurt [132]. Possible outcomes of a restorative justice conference include an apology from the offender or an action plan that the offender will carry forward, e.g., doing community service or attending an anger management course [27]. This framework acknowledges that it can be difficult, or in some cases even impossible, to fully restore the situation or repair the damage. However, symbolic steps, including acknowledgment of impact and an apology, can help victims heal and offenders learn and take accountability [90]. When restorative justice is embedded within a punitive system, its outcome can inform the punishment decision in some cases [123]. For example, when victims and offenders can reach a consensual outcome in a restorative justice process, offenders may receive a reduction or exemption from a punitive process. In other cases, a restorative justice process may run in parallel and have little influence on the punitive process [123].
Here, an example may be instructive. A high school in Minnesota was dealing with problems of drug and alcohol abuse. The school held a conference that gathered the offender (a student who had used drugs on school grounds), victims (affected students), and community members (faculty and staff). The offender first shared her story and the reasons for her actions. She also took the opportunity to ask for forgiveness. Other members of the conference then expressed how they were affected by the offender’s behavior and jointly discussed solutions. The outcome was that the offender became aware of the effects of her actions and agreed to go through periodic checks to monitor her continued sobriety [65].
Though practicing restorative justice benefited the affected parties in the case above, does it always succeed? Latimer et al. empirically analyzed existing literature on the effectiveness of restorative justice. They found that in-person restorative justice programs successfully reduced offender recidivism and increased victims’ satisfaction with the process and the outcome [75]. However, these positive findings were tempered by the self-selection bias inherent in restorative justice practices. Since it is a voluntary process, those who choose it may benefit more than others [75]. Restorative justice also requires commitment at the administrative level [41] and time and labor for the parties involved [132]. In this research, we examine the potential of applying restorative justice to online settings for addressing harm. The unique characteristics of online communities—such as anonymity [74], lack of social cues [31], and weak social ties [49]—create new challenges and opportunities for adapting restorative justice practices online.
Implementing restorative justice values to address online harm is gaining traction in the research community. Blackwell et al. first introduced restorative justice in the content moderation context [12]. Schoenebeck et al. conducted a large-scale survey study, showing that restorative approaches, such as using apologies to mitigate harm, were strongly supported by participants [106]. West proposed that education may be more effective than punishment for content moderation at scale [128].
In the context of online gaming communities, Kou argued that permanent bans produce stereotypes of the most toxic community members and are ineffective over the long term; he recommended that online communities use a restorative justice lens instead of dispensing bans to re-contextualize toxicity and reform members into becoming well-behaved contributors [69]. Hasinoff and Schneider examined the tension between online platforms’ emphasis on scalability and the principles of restorative and transformative justice, which prioritize tailored and individualized harm solutions. They argued that “subsidiarity,” the principle that local social units should have meaningful autonomy within larger systems, may address this tension [48]. We build on this rich line of work by using the Overwatch gaming community as a case study to explore the benefits and challenges of restorative justice for different stakeholders.

3 Background

To provide context for our methods and results, we briefly review the two platforms we study, Overwatch and Discord.

3.1 Overwatch and Its Moderation Practices

Overwatch is a real-time, team-based video game developed and published by Blizzard Entertainment.3 It assigns players to two opposing teams of six. Gamers play in the first-person shooter view and can select from over 30 heroes with unique skills. They pair up with random players if they enter the game alone, but they can also choose to pair with selected teammates. During each game, players communicate through the built-in voice chat and text chat functions, but some players also use Discord as an alternative. All players are expected to comply with a set of rules laid out in Blizzard’s code of conduct [14]. For example, these rules instruct, “You may not use language that could be offensive or vulgar to others,” and “We expect our players to treat each other with respect and promote an enjoyable environment.”
Blizzard hires commercial content moderators, who are paid company employees, to regulate its games [94]. Though the company shows users a small set of moderation rules in its code of conduct, it is likely that the company has an internal set of more detailed moderation guidelines to help moderators make their decisions [94]. While moderators do not monitor live games, they handle reports from victims by reviewing game replays and take moderation actions if they determine that users have violated platform rules. Typically, offenders receive a voice chat ban or a temporary or permanent account ban. Offenders receive the decision notification, usually without a detailed explanation. Victims who report the incident usually receive a notice that an action has been taken, but they are not told what the action is.

3.2 Discord and Its Moderation Practices

Discord is a popular instant messaging platform that is widely used by Overwatch gamers. On Discord, users can create their own communities, called servers, that contain both text and voice channels for real-time discussions. At the time of this study, more than 2,000 Discord servers were active under the tag “#Overwatch.” Overwatch gamers use these servers to discuss the game, find teammates, and organize Overwatch matches.
Discord moderators are volunteer end-users who regulate their communities and screen posts for inappropriate content [62]. Each community creates its own set of moderation rules. Moderators can sanction users by removing their posts, muting them, or banning them either temporarily or permanently. Since moderators and users both have access to public channels in real-time, moderators can actively monitor harm cases on those channels as they occur. They cannot access private channels, but users can report harmful incidents to moderators through private messages. Some communities use automated moderation tools to detect posts containing inappropriate keywords and issue automatic warnings to posters [62]. We contribute to the study of Discord moderation by showing how different stakeholders perceive and engage with cases of online harm.

3.3 The Overwatch Gaming Community Spans Overwatch and Discord

Like many online communities, the community we study spans multiple platforms [36]: Overwatch, the gaming platform, and Discord, a discussion platform. Though Overwatch pairs up random players if they enter alone, Overwatch gamers frequently use Discord to discuss the game, stay connected, and communicate with one another. Harm can occur both for players with pre-existing social connections and those who are strangers to one another. Since volunteer moderators participate in the Discord community and are often gamers, they may also be friends with a victim and/or offender in a harm case. In addition, a harm case may initiate in Overwatch but extend to Discord, or vice versa. Our research investigates harm cases on Overwatch, Discord, or both platforms and includes players with diverse social relationships.

4 Methods

In total, we interviewed 23 participants from the Overwatch gaming community for this study (Table 1). To understand Overwatch gamers’ perspectives on the restorative justice process, we interviewed victims (those harmed), offenders (those who harm),4 and Discord moderators who dealt with the cases being discussed. Some participants fall into more than one of these three groups. We could not include Overwatch moderators in our study because they are commercial content moderators [95] who remain anonymous and constrained by non-disclosure agreements.
Table 1.
 AgeGenderRace/ ethnicityEducationCountryVictimOffenderModeratorFacilitator
P125FemaleAsianMaster’s degreeUSx x 
P220MaleWhiteBachelor’s degreeUK  x 
P324Non-binaryWhiteSome collegeUK  x 
P420sFemaleWhiteAssociate degreeCA  x 
P526FemaleMixedMaster’s degreeUS   x
P627MaleFulaMaster’s degreeUS   x
P718Some collegeUK  x 
P819FemaleWhiteSome collegeUKx   
P920MaleWhiteSome collegeUSx   
P1020MaleHispanicSome collegeN/A  x 
P1121FemaleWhiteAssociate degreeUSx   
P1218FemaleMixedLess than high schoolUSx   
P1319FemaleWhiteHigh school graduateCAx   
P1424FemaleWhiteBachelor’s degreeCA  x 
P1518WhiteLess than high schoolNA  x 
P1625TransgenderWhiteMaster’s degreeUSx x 
P1718MaleAsianSome collegeUSxx  
P1823MaleWhiteBachelor’s degreeUKx x 
P1930MaleWhiteMaster’s degreeUKxxx 
P2018MaleAsianSome collegeCAxxx 
P2118MaleWhiteLess than high schoolUSxx  
P2220MaleWhiteN/AUKxxx 
P2318MaleWhiteHigh school graduateIrelandx   
P2418MaleBerbersLess than high schoolAlgeriaxxx 
P2524FemaleWhiteBachelor’s degreeCAx x 
Table 1. Participants’ Demographic Information
We recruited participants using surveys for Overwatch or Discord users who (1) have been harmed or (2) have been banned or warned. Additionally, we recruited moderators on Discord. We show here the demographic details of each participant and their self-identified role (victim, offender, moderator, or facilitator; marked by “x”) in the harm cases they discussed with us. Note that a single person may have multiple roles in a harm case or across different cases. We also recruited two restorative justice facilitators.
Participants’ gender and race/ethnicity are self-identified.
Restorative justice primarily evolved through practice rather than as an academic discipline. To more deeply understand how its principles might be applied online, we conducted two additional expert interviews with facilitators from a restorative justice center at the University of California, Berkeley. Further, the first author attended 30 hours of restorative justice training courses to learn how it is practiced in local communities. These interviews and the training helped ground our research in restorative justice values and practices.

4.1 Recruitment

We recruited participants using a combination of convenience sampling and snowball sampling [10, 96]. First, we joined multiple Discord communities focused on Overwatch and reached out to moderators, sending private messages to request an interview. After building rapport with moderators through interviews, we asked their permission to publish recruitment surveys in their communities to find victims and offenders of harm. Some moderators referred us to their fellow moderators for interviews and invited us to other Overwatch Discord communities they were involved in. In total, we recruited participants from five Overwatch Discord “server” communities. In addition, we recruited two facilitators for expert interviews from a training program the first author participated in. We recruited victims and offenders separately through two surveys. In the survey for victims, we described our recruitment criteria as people who have experienced online harm on Overwatch Discord or in an Overwatch game. During the interviews, some victims referred us to their friends who have experienced harm or have been banned in the Overwatch gaming community, and we included them as participants. For the second survey, we did not describe participants as “offenders” or “people who have caused harm” since prior research suggests that people may not want to associate themselves with those categories, especially when there has been no opportunity to discuss what has happened [58, 60]. Therefore, we described the recruitment criteria as people who have been warned or banned on Overwatch Discord or in the game.
In the recruitment surveys, we asked participants to briefly describe a harm case they had experienced. We selected participants from this survey based on the time order of their replies. Additionally, we conducted preliminary data analysis to categorize the types of harm (e.g., on Discord vs Overwatch; between friends/strangers; within the moderation team/between end-users/between end-users and moderators). We then prioritized participants who had experienced different types of harm for interviews. Table 1 describes the demographic information of our participants.

4.2 Interview Procedure

Through interviews with victims and offenders, we wanted to understand both their current experiences with harm cases and their perspectives on a restorative justice process for those cases. We adapted our interview questions from restorative justice pre-conferences, where facilitators meet one-on-one with victims and offenders to solicit their perceptions on using a restorative justice process to address harm [132]. During the first stage of the interviews, we asked participants questions about the harm case they had experienced or caused, including how it was handled, its impact, and the need to address the harm.
During the second stage, we introduced participants to restorative justice principles and the victim-offender conferences. We focused on these conferences because they constitute a widely used practice that embeds the core restorative justice principles. We included frequently posed questions in preparation for victim-offender conferences, such as what information they would like to convey in the conference and their expectations and concerns regarding the process. If time allowed, we asked participants to reflect on more than one harmful incident. Further, some victims and offenders were also moderators of the community. We first inquired about the harm with their primary self-identified roles and then asked about their secondary role(s).
Our interviews with Discord moderators were intended to assess how they deal with harm in their communities and their attitudes toward using restorative justice to repair the harm. Additionally, since the facilitator is essential in offline restorative justice practices, we explored the possibility of creating a corresponding role in online scenarios. Since moderators have the closest currently existing role to a facilitator, we sought to learn their perspectives on assuming this role. During interviews with them, we first asked about the harm cases they had handled in their communities and their decision rationales. We then introduced the idea of a victim-offender conference and asked them to (1) reflect on potentially using it as an alternative approach for addressing harm cases they had handled in their communities and (2) share their thoughts about serving as restorative justice facilitators on those cases.
It is challenging to elicit people’s perspectives on a hypothetical process or a process they lack previous knowledge about. To make restorative justice concepts more concrete, we asked participants to imagine a restorative justice process based on actual harm cases they had experienced or handled. Additionally, we answered their follow-up questions and corrected any misconceptions we identified in our discussions. We continued analyzing our interview data as we recruited and interviewed more participants. We ceased recruiting when our analysis reached theoretical saturation [21].
We conducted two expert interviews with restorative justice facilitators to elicit their insights about using restorative justice in online settings. We introduced these facilitators to Discord and Overwatch moderation mechanisms and described examples of how harm cases were handled based on our interviews with victims and offenders. Here, we stayed close to our raw data and described the harm cases through the perspectives of our participants. The facilitators evaluated the current moderation practices through the restorative justice lens and envisioned the future of restorative justice on Discord and Overwatch. We did not intend to reach theoretical saturation for this population [21]; we still incorporated these interviews because doing so provided valuable insights on how these cases could be alternatively handled in a restorative justice context [104].
We conducted our interviews from February to July 2020. The interviews lasted one to two hours each, and participants received compensation from $25 to $50 US dollars based on interview duration. We conducted 21 interviews using Discord’s voice call function, and two participants (P1, P13) chose to be interviewed using Discord’s text chat. Before each interview, we negotiated interview time, which was based on participants’ availability and the number of harm cases they wanted to share during the interview. We also conducted two in-person interviews with the facilitators (P5, P6). Our study was approved by the Institutional Review Board (IRB) at the University of California, Berkeley.

4.3 Data Analysis

We conducted interpretive data analysis on our interview transcripts [21]. We began with a round of initial coding [102], applying short phrases as codes to our data line-by-line to keep the codes close to the data. Examples of first-level codes included “impact of harm,” “creating new account,” and “banning.” Next, we conducted focused coding [102] by identifying frequently occurring themes and forming higher-level descriptions; second-level codes included “notion of justice” and “sense of community.” Coding was done iteratively, where the first author frequently moved between interview transcripts and codes and discussed emergent themes with other authors. After these initial and focused coding rounds, we applied restorative justice principles and values as a lens to guide our interpretations. Finally, we established connections between our themes to arrive at our findings, which we categorized according to participants’ roles (offenders, victims, moderators, and facilitators). We coded the expert interviews using the same code book we used for other interviews since it helped us compare facilitators’ views of harm to other participants’ opinions during the analysis.

4.4 Methodological Limitations

We used convenience and snowball sampling to recruit participants in a single gaming community. In addition, our research examines harm from an interpersonal perspective. As a result, the experiences participants report and the solutions we design may not be representative of, or applicable to, all forms of gaming communities or online harm. Further, though Overwatch has a multinational and multicultural user base, most participants were English speakers from Western cultures. Finally, our recruitment method did not let us recruit offenders who were neither warned nor banned.
As researchers, we lack a full picture of what occurred in the harm cases we studied. To this end, we adopted Whitney Phillips’s approach [89], which involved observing how our participants presented themselves to us and drawing conclusions from their performance, which may have been choreographed. As a result, we present our findings as subjective perspectives of Overwatch members rather than as objective truths. During the interviews, all victims believed that they were the party that had received harm; all offenders believed they had caused harm, but some assumed only partial accountability since they were also harmed in the process. We analyzed our interviews based on these self-described roles and views of participants in each harm case.

4.5 Positionality Statement

As researchers working in the sensitive space of understanding online harm and imagining interventions that may help address it, we briefly reflect on our position on this topic. All article authors feel deeply concerned that online harm is a persistent social problem that disproportionately affects marginalized and vulnerable people [34]. Some authors also come from marginalized groups and are survivors of harm. Our prior research on online harassment has helped us understand the inherent limitations of existing moderation mechanisms, which has shaped our desire to look for alternative approaches. We celebrate the growing interest in applying alternative justice theories [8, 54, 67, 81, 127], e.g., racial justice [106] and transformative justice [25] to reduce harm to victims, offenders, communities, and societies. We are particularly enthused by the success of restorative justice in addressing offline harm and its potential to provide agency and care for vulnerable groups who are often ignored and harmed in a punitive justice model.
While we embrace the values of restorative justice and see its potential, we do not seek to advocate for restorative justice in this work uncritically. We all live in cultures where punitive justice is the dominant approach to addressing harm, and we have actively worked to learn about restorative justice through research and by practicing its values in our own lives. This education and experience have shaped our interpretations of and understanding of the implications of our findings. At the same time, we acknowledge that restorative justice processes alone cannot address deep-rooted, systemic cultural and social issues such as racism and sexism (see transformative justice [25]).
We recognize that restorative justice is voluntary, and forcing this process may further harm the victims. Thus, we attempt to present an impartial account of our participants’ perspectives on the victim-offender conference and restorative justice principles. Designing and implementing tools for restorative justice practices, and experimenting with these practices in online communities, are crucial areas for further understanding the utility of restorative justice in the online context.

5 Findings: Current Moderation Models Through A Restorative Justice Lens

We use the three restorative justice principles (see Section 1) to understand the experiences of victims, offenders, and the community during instances of online harm within the current content moderation landscape. With our sampling methods, we included two victim-offender pairs in our study — P17 and P25 as well as P22 and P24. To illustrate our findings, we present a relevant harm case from our data in each section.
We warn readers that the following sections contain offensive ideas and language. However, we believe that such sensitive content can help readers understand the nature of online harm.

5.1 Victims Have Limited Means to Heal from Past Harm or Stop the Continuation of Harm

Restorative justice centers on victim needs. In offline restorative justice processes, victims usually share their stories, receive emotional support, and provide input on what is needed to repair the harm. The process aims at helping them heal from harm and stop harm continuation [132]. Our analysis indicates that the main tool available to online victims to address harm is reporting their offenders to moderators or the moderation system. However, it also shows that reporting such incidents does not effectively address their need for emotional healing or prevent future harm.
To demonstrate the harm participants receive in online gaming communities, we first relate an example in Case 1.
Case 1
P13 teamed up with two previously unknown players to play an Overwatch game. She said hello in the voice chat but instantly regretted it: “These guys started saying “Is that a girl damn are you cute,” and making jokes.” P13 chose to remain silent, but she soon received a lurid threat: “They got harsh and said, “If you don’t respond I will take out my dick and slap you.””
P13 confronted those players, but the players “all acted like it wasn’t a big deal.” They continued to make jokes about her and complained about her gaming skills after they lost the first round. P13 decided to leave the game but continued monitoring the in-game chat: “They were complaining about me, saying girls shouldn’t play games. It makes me feel so nervous, I started to cry.
Several participants told us that they frequently experience harm, such as offensive name-calling or sexual harassment, on Overwatch or Discord. We find that such cases are often related to structural issues such as sexism, misogyny, transphobia, and homophobia—offenders often target their victims’ gender or gender identity. Victims also received negative comments about their gaming skills. Women and gender non-conforming gamers sometimes experience compounding harms due to both of these patterns. P11 offered some examples of the comments she had heard while gaming:
“Go make me a sandwich, calling me a bitch, telling me that I should go make food for my husband, I probably have kids and cats [...] getting called bitch, slut, whore, cunt, just every derogatory name for women in the book.”
On both Overwatch and Discord, the main tool for victims to address harm is reporting to the moderation system or the moderators. However, victims often felt left out of the decision-making process after reporting a harm case and did not find the process helpful for healing from the harm. Several participants told us that after reporting, they either do not hear back or receive a vague message that indicates the moderation decision but offers no procedural details: “[The moderation system] just tells me that “something” happened, and my report led to that happening” (P11). P9 believes that the moderation system is intentionally opaque to gamers: “They don’t want players understand[ing] the system.”
In addition, though reporting the harm may result in punishment of the offenders, it does not directly help victims heal from the emotional impact of the harm. P11 talked about the emotional toll of the incident: “I’ve cried about it a few times [...and I told my friend], “I don’t even feel like I want to live in a world where there are people like that.””
Many victims indicated that though they tried to report their offenders, they continued to be harmed in the community, where a culture of harm is prevalent. Even if they do not encounter the same offender again, incidents of harm are so frequent that they come across new offenders repeatedly: “At least 5 out of 10 games, I get somebody saying some sort of crude comment about sexually harassing me” (P25). When those experiences of harm accumulate, victims anticipate being frequently subjected to harm, and they accept that there are no effective ways to address it. P9 said, “[Harm] happens frequently enough that you just get used to it.” P11 said, “I feel pretty helpless that I have to endure that every time I play a game.” P12 noted that reporting itself becomes labor intensive: “[Harm] happens so often. I wouldn’t want to report every single person I’ve talked to.”
Victims indicated that they need to consciously avoid harm in the community, which impacts their gaming experiences. For example, some participants stopped using voice chat out of fear that it would reveal their gender, even though communication with teammates is important for winning. P13 fears engaging with strangers after having experienced ongoing harmful behavior: “I will never play a game without a friend I trust, or talk in the game without my friends around because I don’t trust that there will be even one person that will defend me when it gets bad.” Some participants also leave a game or Discord community where they have experienced harm.
In sum, these findings show that the intensity and frequency of online harm can substantially impair users’ online experiences and cause long-lasting emotional damage. Although the current moderation systems on platforms like Overwatch and Discord offer socio-technical mechanisms like reporting to address cases of online harm, the goals, and the lack of transparency and follow-up inherent in these mechanisms, do little to meet victims’ need to heal. In addition, the current reporting mechanisms do not substantially reduce offenses within the gaming community because they do not effectively change the culture: even when participants can avoid the original offender, they continue to be harmed by new ones.

5.2 Offenders Are Not Supported in Learning the Impact of Their Actions or Taking Accountability

In a restorative justice view, offenders can address harm by acknowledging their wrongdoing, making amends, and changing their future behaviors. Through practices like victim-offender conferencing, offenders learn the impact of their actions by listening to the victim’s side of the story. Afterward, they can repair the harm and learn through actions, such as acknowledging harm (e.g., apologizing), taking anger management courses, or doing community service [27, 132].
As our participants report, current moderation approaches often embody graduated sanctions [70]. That is, moderation teams in Overwatch and Discord tend not to ban offenders permanently after their first offense. Instead, they use more lenient punishment first to give offenders a chance to change their behaviors. Our interviews show that many Discord moderators have elements of a restorative mindset: they want to help offenders learn their wrongdoing and change their behavior by giving them second chances and providing explanations for their sanctions. Several Discord moderators explained their rationale for graduated sanctions as “giving people a second chance.” As moderator P14 noted, “We don’t want [the moderation decision] to be a surprise [...] and we will actually really want to encourage them to improve.” Some Discord moderators also provide explanations of their moderation decisions. Such messages typically explain the rules and consequences of breaking them and can be posted by a moderator or a pre-configured bot.
Though some moderators have a restorative mindset, the punitive moderation system and rule-based explanations they rely on may fail to provide learning and support for offenders. They may not effectively stop the perpetuation of harm. We show an example through Case 2.
Case 2
P14, a Discord moderator, described a case where moderators in her community gave an offender multiple chances to reform her behavior, but the offender was reluctant to change.
A transgender woman in P14’s community had negative experiences with cisgender men in her life and ranted publicly in the community about her hatred of all men. Though the majority of the community are female and LGBTQ gamers, P14 strives to create an environment that is friendly to its cisgender male members: “[Cisgender men] are not allowed to say horrible things about the female and trans plus members, [so] in return we expect the same courtesy.”
The moderators warned the offending user several times. P14 reflected, “One of our mods would go in and be like, “Hey, just so you know, this is not okay here. I’m going to remove your message and just don’t do it again.” She would respond with, “Oh yeah, got to protect the cishet [ed. cisgender and heterosexual] men from the trans people.””
This user was temporarily banned after several warnings. She then messaged moderators, telling them that they “don’t know anything about oppression” (P14). The moderators defended themselves, arguing that they understood oppression, and directed her to the community rules. Subsequently, the moderators stopped engaging with this member.
The temporary ban caused the transgender woman to lose her gender pronoun tag, which demonstrated her gender identity to the community. When she realized this had happened, she swore at the moderators, leading them to permanently ban her from the community.
In case 2, the moderators made efforts to negotiate with the transgender woman, hoping that she would change by giving her multiple chances and providing explanations for their sanctions. However, their explanations mainly emphasize that her actions violated the rules and were prohibited, and the moderation decisions are all punitive. We argue that this model directs offenders’ attention away from the victims and does not further their understanding of the impacts of their actions on the victim. Instead, the punishment, warnings, and restating of rules position the offenders against the moderation system and in defense of their own behavior.
Restorative justice recognizes that often, offenders of harm may have been the victims of harm in other cases. As activist Mariame Kaba says: “No one enters violence the first time by committing it.” [46]. In case 2, the harm that the transgender woman caused is a reaction to the harm she experienced elsewhere from cisgender men. Similarly, several offenders in our sample revealed how they had been constantly harmed in gaming communities: “I have been called countless names that are vulgar, offensive, and stuff like that” (P21). Restorative justice practices assume that although the past harms experienced by offenders do not absolve them of their responsibility for committing an offense, it can be hard to stop offensive behavior without addressing their sense of victimization through support and healing. Punishment, on the contrary, usually reinforces the sense of victimization [132].
The moderators in case 2 stopped the transgender woman from sharing her story. Instead, she received a denial, which worsened the damage and led her to defend herself more aggressively. In addition, when harm such as discrimination is a systemic issue in the gaming community or the broader society, stopping the perpetration of harm may require work to change the culture in addition to work to change individual offenders. Facilitator P6 explained how a restorative justice approach would strive to provide support and look at the root cause instead of punishing a particular offender:
“Rather than just zooming in on that one case and trying to blame this individual for that action, acknowledging that [they were] also harmed by the community and pushed to act in that way [...] it’s a much harder conversation to engage with, but then, that takes away from saying that person is wrong. It’s more about [...] how do we understand this collectively, and then, how do we address this collectively? ”
We also find that when punishment is used as the only tool to stop offenses, it loses its function when offenders are not actually punished. Several moderators, victims, and offenders mentioned the convenience of using alternative accounts once one account has been moderated: “It is so easy to make new accounts in Discord that “reporting” people don’t really work” (P1). P17 used to conduct multiple offenses in games. He pointed out the fallacy of regulating with banning while there is no cost to creating and using another account in the game:
“Everyone right now is happy with that system because, if I was a normal player [...] I heard that [the offenders] got banned, I would be like, “Wonderful, they got banned. I’m never going to see them again,” but in actuality, when I got banned I’m going to say, “I don’t give a fuck, I’m just going to log into my second account.”
Despite these shortcomings, we found evidence that the current punitive approach contributed to maintaining the health of community content and stopped the continuation of harm. Moderators we interviewed told us that some offenders would stop their misbehavior after receiving a warning or would not return to the same community after being banned. However, when punishment is the only means of addressing harm, it cannot always be effective.

5.3 Gaming Communities Create Challenges for Victims and Offenders to Address Harm

Harmful actions disrupt relationships within a community and potentially affect all its members. Thus, it is important for community members to acknowledge the harm and participate in redressing it collectively [132]. Restorative justice defines a micro-community as the secondary victims who are affected by the harm and the people who can support the victims and offenders to address the harm (e.g., family or friends) [80]. In online gaming communities, the micro-community includes not only victims, offenders, and moderators but also bystanders, friends of victims and offenders, and gamers, more generally.
However, we found that the relevant stakeholders had no shared sense of community during many instances of online harm. As a result, the harm was often considered “someone else’s problem” and remained unaddressed. Additionally, when micro-community members got involved, they often created secondary effects that further harmed victims and offenders.

5.3.1 Victims, Offenders, and Moderators Do Not Have a Shared Sense of Community.

Restorative justice appeals to the mutual obligations and responsibilities of all community members to each other as necessary to address harm. However, in current moderation systems, we found that victims, offenders, and moderators lack a shared sense of community.
Many victims in our sample relate to and show care toward other gamers the offenders might harm. For example, P11 said, “I usually am just sad not for myself really, but just that other people have to deal with those people.” Some victims even care about their offenders and what may have led them to perpetrate harm. However, there is a lack of shared sense of connection from the offenders, which makes it hard for them to care about their victims and may lead them to fail to see the impact of their actions on others. The anonymous and ephemeral nature of online conversations deters offenders from relating or caring about the community or the victims. P17 used to harm other gamers but has since reformed himself. Reflecting on his previous mindset as an offender, he pointed out: “[The victims’] day is legitimately ruined because of what [offenders] said, but these people aren’t going to think about what they said twice [...] They’re not going to reflect because it doesn’t affect them.”
Additionally, Overwatch randomly pairs up gamers when they do not join as a team. As a result, the offenders do not need to interact with their victims after a game has finished. This absolves offenders from feeling accountable for their actions and in fact creates an environment where repeating harm comes at no cost. P21, who has attacked others in Overwatch games, noted: “In a game where you know somebody for 20 minutes and you see that they’re bad so you flame [insult] them, and then you just move on. You never see them again.”
As mentioned in Section 5.2, many gamers we interviewed have multiple online identities in games and do not feel particularly attached to one. P17 used to attack others online using his anonymous accounts. He reflected, “No one can recognize what accounts we’re on so we can do whatever we want.” Given the limited information and social cues online, it is harder for people to find mutual relationships and connections. P21 used it to justify offenders’ behavior:
“In games you don’t know these people. You don’t know their personalities. You don’t know their backstory [...] All you know is that they’re doing bad [at gaming], and so that’s what you use against them. And, there’s no way of fixing that. And if you’re getting offended by some of these words, then just mute them.”
On many Discord servers, moderators regulate a confined, public space — the general chat. Several moderators told us they do not intervene in harm cases that happen outside this general chat. Moderator P4 noted: “For the most part, we just tell people we can’t moderate things that happen outside of our community, and at that point it’s on them to block people.” For harm cases in the general chat, moderators do not help offenders and victims mitigate problems; instead, they punish the offender or ask both parties to resolve problems themselves. Moderator P19 would move contentious conversations to a private space: “I’ll delete all messages that they’ve put through to each other, put them both into the group chat, ban them from talking in general or whatever and keep them in this private chat and get it all solved in there.”
In a restorative justice view, a sense of community is essential for offenders to care about the harm they cause to the victims and may even stop them from conducting harm in the first place. However, we found that the anonymous, ephemeral nature of online interactions makes offenders careless about breaking relationships with victims. Additionally, the unclear distinction between public and private online spaces creates challenges for moderators to support victims, and the ensuing lack of support may leave victims vulnerable.

5.3.2 The Community Creates Secondary Harm Against the Victims.

In some offline practices of restorative justice conferencing [132], community members who care about the harm that occurred can choose to join the conference. Community members can support victims by listening to their stories, acknowledging the harm that occurred, and providing emotional support [132]. Analyzing our interview data, we found that due to the absence of such a victim-centered structure online, community members often create secondary harm. Though it takes courage for victims to share their stories, finding an audience that supports and cares about them can be difficult. Instead, victims may be challenged, humiliated, or blamed, which creates secondary harm.
P12 was sexually harassed by a male member in the gaming community when she was underage. She chose to reveal the member’s behavior during a voice chat when other community members were present: “I felt like more people should know about it since most of the people in [the Discord community] are underage.” Though P12 expected to get support from her friends, she received multiple challenges from them after this revelation: “Some people were on my side [...] but some other people were on his side and said I just wanted attention, and I should have known he was just joking.” Later, the man denied all accusations and accused P12 instead. According to P12, “He was trying to make me look like a bad person.” P12 felt so unsupported and hurt by the community’s reaction that she eventually chose to leave it.
Community members often ignore the harms they witness, which can further fuel harm. P13 was a victim herself and has witnessed harm in the Overwatch community. She talked about how she was disappointed by the bystanders who did not speak up for the victims:
“Most people just think that the game is over; there is no point in trying to help even your teammates, and think they will just end the game and try again another game. Since it is online, most people don’t realize that it can actually hurt people what someone says.”
Further, community members can feed into harm by rewarding offenders with attention and positive reactions. P17 talked about why he used to harm intentionally to get attention from friends:
“All my friends found it funny. The reason I said those things was not for [the victim ...] It’s for them [my friends] to laugh. It’s for the reaction. It’s the adrenaline rush of being the center of attention, you know?”
These findings suggest that community members are not neutral bystanders when harm occurs. They may create secondary harm for victims by ignoring the harm, challenging their stories, or encouraging offenders’ behavior. In contrast, restorative justice enables the community to build a culture of responsibility and care, where community members collectively prevent and address harm.

5.3.3 The Community has a Punitive Mindset Toward Offenders.

Restorative justice encourages the community to facilitate change in offenders’ behaviors. Community members can share how the harm has impacted them, express willingness to support offenders through the change, and acknowledge any changes offenders have made. However, in punitive settings, community members who disagree with the offenders’ behavior want to punish them, as is the case in online moderation systems. Case 3 shows an example of such an occurrence and its effects in Overwatch.
Case 3
P17 and his male friend were gaming as a team with a female player, P25, whom they started to attack with sexist comments after losing the first round. P25 decided to record this incident and posted it on Twitter after the attack continued for a while: “I knew that it was something that a lot of women and a lot of people deal with daily almost within the community. I wanted to be able to show just how bad it can be.” Many people showed empathy and support for P25, and to date, the tweet has received more than 100 retweets and 600 likes.
After P25 reported P17 and his friend through Overwatch, their accounts were temporarily banned. However, for P17, the account banning was not the most severe punishment; rather, it was the subsequent harassment and damage to his reputation due to P25’s public revelation. People located P17’s multiple social media accounts. He reflected, “I would get random messages throughout the day saying, “You should kill yourself,” and death threats like, “I’m going to come. I will.”” He abandoned his previous online identity completely: “I deleted my Twitter, deleted Instagram, deleted Discord, changed all my Overwatch account names.” P17 later apologized to P25 and has since changed his behavior. However, P17 believed the incident changed his career path as a 16-year-old: “To be honest, I think that was the main reason I didn’t try to pursue to go pro in Overwatch harder [...] because of how much I had to do to get back the reputation.”
P17 suffered bans and community condemnation, which stopped his offensive online behavior. However, no one supported him in taking accountability and changing after learning the impact of his actions except for one gamer friend who reached out to him. This friend suggested that P17 interact more with people offline because he seemed emotionally detached from his online victims. At that time, P17 used to play video games all day. P17 took that advice, and as time went on, he began realizing the impact of his actions on others: “[By] talking to real people and interacting with them face to face so you can see their emotions [...] now I can imagine them [people I attacked in games] sitting at their computer and just like crying [...] and that’s why I don’t say these things.”
In Case 3, victim P25 had a positive experience when sharing her experiences of harm on Twitter. However, most other gamers decided to punish P17 instead of telling him how his actions caused harm. As we mentioned in Section 5.2, the punishment only further harmed P17: “It’s just anxiety. That’s what’s constantly going through your head.” This punishment stopped P17 from offending again, but he did not learn to care about his victims until his friend reached out and supported him.
Additionally, though P17 apologized to P25, stopped those behaviors, and learned the impact of his action afterward, he did not have a chance to demonstrate his change to the community. He left the community and abandoned his career goal to become a professional gamer. This is not what his victim, P25, had wished would happen. She wanted the offender to have a chance to learn and demonstrate his change: “Nobody’s perfect, everybody makes mistakes. We’re all human [...] I don’t think that just because you did one bad thing a year ago means that you just don’t have a chance anymore.” The community response also runs counter to restorative justice views, which maintain that offenders will be welcomed back to the community after their reform [132].
In general, we found that some community members have a sense of responsibility and care about the victims and the harm that occurs inside their community. However, how they address harm is largely shaped by the punitive justice values prevalent in the gaming community and broader society. Their actions of punishing the offenders can create further harm in the community and do not help offenders who may choose to reform themselves.

6 Findings: Online Restorative Justice Possibilities and Challenges

We now discuss how the various gaming community stakeholders responded to the idea of using one common restorative justice practice, namely, a victim-offender conference, to address harm in online gaming communities. Our goal was not to measure the participants’ binary response regarding their interest in attending the victim-offender conference or use it as a direct measure of the potential of online restorative justice practices. Rather, we sought to identify the essential prerequisites and potential obstacles associated with designing and implementing new forms of online restorative justice by encouraging participants to reflect on their needs and concerns as related in the victim-offender conference.

6.1 Victims’ Needs and Concerns for a Restorative Justice Process

Many victims wanted to join a victim-offender conference if the offenders were willing to repair the harm. Concurrently, they had concerns and doubts about the process, especially about offenders’ readiness to attend the conference. In our sample, no participants mentioned punishing offenders as a desired outcome of the victim-offender conference. We describe these needs and concerns below.

6.1.1 Some Victims Want to Understand and Communicate with Offenders.

Some victims wanted to understand why offenders harmed them and communicate to the offenders how they were hurt. Victims observed that the harm against them occurred unexpectedly, and they could not rationalize the offenders’ behavior, which resulted in a need to understand it. P19 said, “I just want to know what goes through their head at that point in time.” P11 expressed her confusion and frustration:
“I just don’t understand the motive and don’t understand the reasoning. Do they have girlfriends? [...] Or their sisters? [...] Because, I’m a sister, I’m a girlfriend, I have men in my life that I love and they would never do that to me. So why would you do it to someone else’s loved one? I don’t know.”
Several victims wanted to tell the offenders specifics of how they were hurt. P23 said, “First and foremost, I would express my feelings, how I felt and how that hurt me.” Some victims hoped their sharing would help offenders realize the impact of their actions. P13 said, “Maybe hearing how hurt and scared I felt or feel, they would change their perspective.” As someone who had once harmed others, P17 believed that learning how victims feel about their harm was essential to his change: “If I knew how people felt in games when I made fun of them, I would probably never make fun of them.”

6.1.2 Some Victims Want an Apology, an Acknowledgment of Mistakes, and a Promise of Change from the Offenders.

When we asked victims what they needed to repair the harm, several mentioned that they would like the offenders to realize their mistakes and issue an apology to start their emotional healing. As P13 said, “Just their realization of what was wrong with a small apology would make me happy.”
Other victims hoped that the offenders would change their behavior and stop offending. For some victims, causing offenders to reform their behavior was paramount:
“I wouldn’t want him to do anything personally. I just want him to understand what he did wrong and try to fix it so it doesn’t happen again with me or another person.” (P12).
“I don’t really care to see their ranks [in games] drop or anything; I just want them to change.” (P13)

6.1.3 Some Victims have Concerns About Whether Offenders are Willing to Repair the Harm.

During interviews, several victims were willing to join restorative justice meetings if the offenders were ready to repair the harm. However, they were concerned about whether the offenders they encountered would meet this condition. Several victims had already attempted to reach out to offenders to resolve the issue but were ignored or dismissed by them. These victims did not think the offenders would be open to participating in a restorative justice conference. P22 said, “I wouldn’t be opposed to speaking to him [the offender] again, but I mean, from previous history, it’s going to be difficult to speak to him or be able to trust him again because of his actions in the past.” Some victims also questioned whether the offenders would genuinely want to repair the harm even if they consented to join the conference. P21 worried that the conference is a “get out of jail free card.” He said, “People could be as offensive as they want, then turn around and be like, “Oh, I’m so sorry. I won’t do it again.””
Most victims who shared these concerns did not want to meet with offenders if they could not ensure there was a genuine desire to repair the harm. One exception was P13, who had a strong preference for offenders to hear her voice even if they might be unaccommodating: “Even if they [offenders] are aggressive, as long as I had someone I trusted there, I would think it would have the best chance at an outcome [for the offender] to hear my voice.” In restorative justice, victims and offenders meet only when both parties are willing to repair the harm. The facilitator also acts as the gatekeeper to ensure that further harm is unlikely before victims and offenders can meet [15].

6.1.4 Some Victims Believe the Harm They Experienced is Systemic and Addressing it Requires Long-term Efforts.

While several victims hoped that the victim-offender conference could reform offenders’ behavior, others thought the problems they faced had systemic roots that could not be addressed in a single meeting. P16, a transgender woman who was harmed by someone she believed was transphobic, did not want to meet with the offender because she did not believe the meeting could solve the problem:
“I mean, because it’s being trans, there’s just so much systematic oppression to it [...] It just takes time, and it takes education. It takes advocacy [...] I believe he [the offender] is somebody who probably will change if he’s given the right education, the right information, but it’s going to take time and it’s going to take people becoming more and more accepting of trans people. That’s not just a fix that’s going to be fixed easily through Discord.”
P25, a female gamer who was verbally attacked while playing Overwatch (Case 3), believed the offender’s aggressive behavior was likely shaped by more than just what happened in the video game. As a result, she thought the problem could not be solved within gaming communities alone:
“It’s more probably deeper rooted than just the video game. It probably seeps into family issues to schooling issues to the environment they’ve grown up in, etc. You can’t really help that without doing more and being there in person.”

6.1.5 Some Victims Want to Move Past the Harm.

Some victims noted that the harm had already occurred and wanted to move past it. Several were emotionally exhausted by the harm. P20 already felt disappointed by the offender’s reaction when he tried to negotiate: “I don’t think it’s worth trying to regain her [the offender] as a friend [...] because things like this can happen again due to the rash behavior.” P16 described her feelings as “I’m just kind of over it.”
Some victims thought the harm they experienced was trivial and that they were not severely impacted by it, or that they could reduce the emotional harm through their own efforts. As a result, they wanted to move on with their lives. P8 said, “If it’s more personal, then they should maybe apologize. But [...] in this case it’s not particularly serious.” P9 gave offenders the benefit of the doubt and felt that moving past the event was an easier option for him:
“Maybe the people that were talking inappropriately were just having a bad day or something, or maybe they were genuinely not a good person and rude. But either way, it just seems easier for me to move past it by literally moving past it.”
In sum, most victims have needs other than punishing the offenders. However, they want offenders to engage with such needs only when they are sure that offenders genuinely wish to address the harm. In addition, many victims feel that restorative justice approaches might not be sufficient or ideal to address systemic online harm.

6.2 Offenders’ Interest in a Restorative Justice Process

During the interviews, we simulated pre-conference practices of restorative justice by asking offenders a series of questions to help them reflect on their experiences of harm and discuss their willingness to join a victim-offender conference. We now describe offenders’ thoughts on the process.

6.2.1 Offenders May Want to Repair Their Relationships with the Victims.

One offender (P22) wanted to join the victim-offender conference to repair the relationship with the victim (P24). He admitted that he was emotional when committing the harm:
“I have since learned to control my temper as I have gotten older. However, it can still be a problem from time to time. I have an extremely competitive and stubborn personality, so when something doesn’t exactly go to plan, it can be difficult for myself to accept and get over it.”
P22 wanted to issue an apology to P24. They were organizing an Overwatch tournament together in a Discord community. P22 hoped that the meeting could help him maintain a professional relationship with the victim: “I would literally just settle for being civil with one another.”

6.2.2 Some Offenders Prefer the Punitive Process Over Restorative Justice.

Several offenders (e.g., P17, P21) agreed that they should take full accountability for the harm they had caused. However, they preferred to receive punishment for their actions rather than join a victim-offender conference. In Case 3, P17 was banned for verbally attacking P25 and lost his reputation in the Overwatch community after P25 posted the video footage online. Though he acknowledged his mistake and apologized to P25 privately after the offense, he did not think he would have attended a victim-offender conference at that time. He described the process as “boring”:
“I’m not going to go in with an open mind so nothing will get done anyways [...] Would an immature 16 year old teenager who’s rebellious against his mom, he doesn’t want to do the freaking dishes, do you think he’ll want to sit in a call or a meeting with the person that he just harassed for the last 20 minutes and figure it out? My answer is no.” (P17)
We found that these offenders’ notion of justice aligns with the tenets of punitive justice. They believed they deserved punishment and did not trust the restorative justice process to help them achieve a better outcome. P17 said, “I definitely fucked up a little bit, so I deserve the punishment.” P21 was banned for insulting others in a game. He was aware of his wrong-doing and expected to get punished even if he joined the meeting, so he wanted to go through the punishment directly: “If you know for a fact you’re in the wrong, then there’s no point in even talking, because you’re going to get banned anyway.”
Both P17 and P21 were under-aged when they committed the harm. Facilitator P5, who works in a middle school, noted that the resistance P17 and P21 expressed is common in offline practices. P5 argued that the end goal of restorative justice is not to punish offenders but to help them take accountability for their actions and change future behavior. The pre-conference is a chance for facilitators to introduce restorative justice and talk about how it may benefit the offender. In school settings, P5 noted that the support for restorative justice from community members, including parents and teachers, encourages teenagers to be more open to the process:
“There’s parents’ support [...and] there’s so much research around how restorative justice works in a school and is incredibly beneficial for the students; it’s an incredible healing process for the entire school community that a lot of [school staff] are willing to buy in.” (P5)

6.2.3 Some Offenders Do Not Fully Acknowledge Their Role in the Harm.

Several offenders wanted to use victim meetings as a chance to justify their behavior. Those offenders believed that their victims behaved improperly or had hurt their interests in the first place. For example, P20 argued that the supposed victim lied about the situation, and he was wrongfully banned for defending himself. He hoped the victim could come to the meeting without preparation so he could challenge her by surprise: “They’ll present the evidence right on the spot so that she doesn’t have time to think or lie or really erase any evidence.” This group of offenders agreed that they should take partial accountability for the harm they caused. However, they primarily wanted to use the meeting to hold the other party accountable and/or alleviate their own punishment:
“Why is it I’m being reported and banned after saying one thing, but yet you’re not getting any punishments [for saying many]?”(P19)
As noted in Section 5.2, the current moderation systems lacks a means to hold offenders who use multiple anonymous accounts accountable for their actions. When offenders can conduct harm without receiving any consequence that they care about, participation in restorative justice becomes additional labor instead of an alternative to receiving punishment. As P21 said, “People aren’t going to apologize. This isn’t the real world [...] If you get banned, they’ll just go play another game.”
In sum, we found one offender in our entire sample who wanted to repair his relationship with the victim, but the mindset of most offenders aligns with punitive justice. Offenders who acknowledge their wrongdoings think that they deserve punishment and do not believe that restorative justice can lead to a better outcome. Those who do not fully acknowledge their wrongdoings want to appropriate the victim-offender conference into a punitive process that holds victims accountable. These views reflect offenders’ emphasis on the consequences of their actions on themselves rather than on harm reparation.

6.3 Moderators’ Views on Implementing Restorative Justice Process in Their Communities

We next discuss Discord moderators’ attitudes toward implementing a restorative justice process in their communities. Most moderators agreed with the values of restorative justice, and some of their moderation practices overlapped with its practices. At the same time, they expressed concerns about adapting the moderation process to the restorative justice model. Additionally, several moderators had already attempted restorative justice in their communities but received push-back and challenges from other moderators and community members.

6.3.1 There are Elements of Restorative Justice in the Current Moderation Practice on Discord, but for Different Purposes.

We find that both moderators and facilitators talk with victims and offenders when handling harm cases, but the issues addressed and the end goals differ. In offline restorative justice practices, facilitators talk with victims or offenders in pre-conferences, where they ask questions to determine what is needed to repair the harm [90]. Some Discord moderators also speak to victims and offenders before making a moderation decision, but their goal is to make informed decisions on how to punish offenders.
A restorative justice pre-conference happens after fact-finding and focuses on emotions, impact, and the need to repair harm, and the facilitator shows support and empathy throughout the process [15, 123]. On the other hand, the conversation by moderators focuses on facts and evidence, with the moderators acting as judges. As moderator P3 described, “We will go and speak to whoever was reporting them [offenders] and speak to people who were involved, and try and get a feel for what actually happened and make a call from there.”
In addition to the process of talking with victims and offenders, both restorative justice and the current moderation system aim at understanding offender behaviors beyond the current harm case of interest. Restorative justice situates offenders in their life stories. As noted in Case 2, one life story would be that the transgender woman who offended cisgender men in the Discord community had had negative experiences with cisgender men in her life. A life story can help offenders find their triggers of harm, and the community could then provide support to help them heal [65]. On the other hand, Discord moderators keep logs of past offenses for all users in the community. When an offender commits harm, the moderators review the logs as a reference to determine the proper punishment. As Discord moderator P2 explained, “We keep logs of all moderation actions that have been taken against any individuals. And so we always check those before handing out any issues [moderation decisions].” Because of the graduated sanctions mechanism, the punishment is often heavier for offenders with past offenses.

6.3.2 Moderators’ Power May Hinder Restorative Justice Process.

Though we may think that the moderators’ role is closest to that of facilitators, we find that the power moderators hold may impede restorative justice process. Moderator P7 has a work background in restorative justice, and he tried to implement pre-conferencing with offenders in his Discord community. He believed that he failed to reach desired outcomes with offenders because of his position of power. P7 banned an offender for cheating about his game rank to win. He conducted a pre-conference with him, where he wanted the offender to learn the impact of his actions on the victims he cheated on. However, the offender expressed the wish to get the ban revoked by offering professional Overwatch courses to P7 as an Overwatch coach. P7 was disturbed by the answer: “It was concerning, right? Because he’s answering in a way to try and please me.” P7 believed that the removal of power from online facilitators is essential for authentic sharing:
“I don’t want that kind of attitude of when people go into a performance they think of like, “Oh, I have to do well now, because they [moderators] have the decision-making power to remove me from [the Discord server].””
Facilitators also pointed out that while moderators have the final right to interpret what has happened and who is right or wrong, restorative justice practices seek to give agency to victims and offenders. Moderators’ power may create prejudices against victims or offenders and reduce their agency in the restorative justice process:
“As facilitators I’m never like, “Oh I take side with this story.” I’m always multi-partial, I hold the stories and then it’s up for the individuals to figure out what’s right to move forward.” (P5)
Though removing punitive power from facilitators may be important, several moderators showed reluctance to let go of power. They worried that giving users agency may lead to unfair and biased outcomes because users may pursue an outcome that aligns with their own interests:
“They [the victims] are going to use it as kind of a tool to punish people that they don’t like.” (P25)
“[The community members in the conference] may initiate a witch hunt or just try and protect their friend group.” (P3)
These moderators’ concerns are valid: it requires labor and skill from facilitators to address those potential issues. In offline practices, the facilitator is an essential role that maintains a power balance between victims and offenders, for instance, by ensuring they have equal opportunities to express their opinions. [90].

6.3.3 Some Moderators Think the Labor of Restorative Justice is Disproportional to its Gains.

The pre-conference and conference processes require significant labor from facilitators [90]. Though a facilitator is usually paid in offline settings such as schools and prisons [63, 65], Discord moderators are volunteers. Several moderators expressed concerns about the labor required to implement restorative justice. Moderator P7 said, “A lot of [moderators] are volunteers and so the easy option is to just mute people or temp ban people or permanently ban people.” Being a facilitator also requires knowledge of the restorative justice practice. P3 thinks they would need to receive additional training to become a facilitator: “We’re not qualified for this [...] I don’t know how we could provide support, or how to make sure that if we give the support, it’s beneficial to them.”
We found that many moderators are devoted to maintaining a healthy community environment, and some spend hours handling a single harm case. Moderator P14 gave an example: “To ban somebody, we actually can have about five or six hours worth of meetings [...] to make sure that our punishment fits the crime essentially.” They were concerned with the restorative justice process because they were unsure whether the extra labor would make any changes. Several moderators believed that users who re-offend multiple times have malicious intentions, so it is not worth spending more effort on them and helping them change. P1 said, “[Those harms] aimed at our identity (women, queer, trans, etc.) are often by people who enjoy calling us names. So, I don’t really see the point of giving these people more opportunities to be prejudiced bigots.” Similarly, P20 thinks offenders “can’t be as civil” in an online restorative justice conference compared to offline: “I do agree that having a talk and communicating would be nicer, but [...] it generally doesn’t go as well as in real life.”

6.3.4 Some Moderators Experienced Challenges Moving From Individual Restorative Justice Practices to Institutional Buy-in.

Several participants in our sample had an education or work background related to restorative justice and had attempted restorative justice practices in their gaming community. As noted in Section 6.3.2, P7 conducted a pre-conference with an offender on the Discord server he moderates. P25, an Overwatch gamer, facilitated one victim-offender conference with two friends who fought during an Overwatch game. In the conference, the two friends acknowledged the impact and apologized to each other.
P7 and P25 independently initiated the restorative justice process. However, P7 believed in the importance of engaging the moderation team and other community members to practice restorative justice: “It takes more than one person to effectively execute restorative justice. For me, I have had a lot of roadblocks when it comes to trying to implement the system [alone].” In offline scenarios, buy-in at the institutional level (e.g., schools, neighborhoods, workplaces) is important. Institutions can officially establish restorative justice as an alternative to the established punitive justice system [63] and have resources to hire facilitators and remove or mitigate punishment for offenders who successfully pursue the restorative justice process [65]. Community members also gradually familiarize themselves with restorative justice and support victims and offenders in the process [7, 65].
It is difficult for an online community to endorse restorative justice when the established culture and systems are punitive, and no examples demonstrate its effectiveness. As a moderator, P7 promoted restorative justice in his moderation team but failed: “You’ve got a traditional model and there’s no real examples to demonstrate the capabilities of this [restorative justice].” P18, a head moderator in P7’s team, reflected on people’s reactions to P7: “He (P7) kind of just mentions it [restorative justice], tries to explain it and then everyone gets confused and they kind of step back.” Similarly, P8 tried to promote an alternative moderation system in Overwatch but could not get support from the gamers and Overwatch staff she talked to. She shared how people responded: “[People said that] the current system worked. It wasn’t perfect, but it worked, and our system was new and untested.” Despite such impressions, it is important to consider how we evaluate models for responding to interpersonal harm. For instance, have the victims’ needs been met? Was harm repeated? Do offenders recognize the impact of their actions?
In our interviews, we found it hard for people to imagine alternatives to the current moderation system. When we asked participants about their needs beyond having offenders banned or we introduced restorative justice to them, they often found it difficult to imagine the alternatives:
“It’s not something I’ve thought about before.” (P2)
“I’m not really sure what else you could do. Banned is the thing that everybody’s always done.” (P9)
In sum, many moderators spend time implementing practices that, on the surface, resemble restorative justice (e.g., talking with victims and offenders) but serve punitive purposes. In addition, the prevalence of intentional harm and the wide adoption of punitive models create challenges for communities that want to adopt a restorative justice approach.

7 Discussion

As noted throughout the article, current content moderation systems predominantly address online harm using a punitive approach. Analyzing through an alternate lens of restorative justice values, we found cases where this approach does not effectively meet the needs of victims or offenders of harm and can even further perpetuate the harm. Restorative justice provides a set of principles and practices that have the potential to address these issues. Table 2 compares our sample’s punitive content moderation approach with a restorative justice one. As the table shows, the latter provides alternative ways to achieve justice from the perspectives of victims, offenders, and community members, who are often absent in current content moderation.
Table 2.
 Punitive content moderation approachRestorative justice approach
VictimsVictims are left out of the moderation process.Addresses the victim’s needs for reparation of harm, such as support and healing.
OffendersOffenders receive rule-centered moderation explanations and punishment.Encourages offenders to learn the impact of their actions and take responsibility to make reparations.
Community membersCommunity members may further harm or punish victims and offenders.Engages community members in supporting victims and offenders and heal collectively.
Table 2. Comparison of Punitive Content Moderation with an Approach Based on Restorative Justice
Applying restorative justice practices to the governance of online social spaces is not straightforward. Victims, offenders, and moderators currently have unmet needs and concerns about the process. In this section, we first discuss the challenges of implementing restorative justice. We then offer possible ways for interested online communities to implement the approach. Finally, we reflect on the relationship between punitive content moderation approaches and restorative justice.

7.1 Challenges in Implementing Restorative Justice Online

We now explore the potential structural, cultural, and resource-related challenges of implementing restorative justice online. We also discuss some possible ways to address them.

7.1.1 Labor of Restorative Justice.

Restorative justice practices and the process of shifting to them will likely require significant labor from online communities. Restorative justice conferencing can be time intensive for all stakeholders involved. Before victims and offenders can meet, facilitators must negotiate with each party in pre-conferences, sometimes through multiple rounds of discussion. The collective meeting involves a sequence of procedures, including sharing, negotiating, and reaching a consensus.
Stakeholders, in particular facilitators, must expend emotional as well as cognitive labor. In offline practices, facilitators are usually paid by host organizations such as schools [41]. However, the voluntary nature of moderation on social media sites like Discord means that online facilitators may be asked to do additional unpaid work. This issue can be particularly salient when moderators already expend extensive labor with a growing community [32, 94, 113]. Labor is also involved in training the facilitators. Unlike punitive justice, restorative justice is not currently a societal norm that people can experience and learn about on a daily basis. Aspiring facilitators need additional training to learn restorative justice principles and practices and implement them successfully to avoid creating additional harm.
We estimate that resources for addressing the aforementioned labor needs could be attained in both a top-down and a bottom-up fashion. A top-down process could require resources from companies that host online communities. There is precedent for platforms making such investments; in recent years, social media companies such as Discord have hosted training for its community moderators.5 A bottom-up process could engage users with preexisting knowledge about restorative justice to first introduce the process in their communities and gradually expand the restorative culture and practice from there. In our sample, two moderators attempted or practiced online restorative justice within their own communities; they showed enthusiasm for expanding its practice to other online gaming communities. It is possible that resources from companies and practitioners could collectively begin to address the labor problem.
Additionally, implementing online restorative justice requires reallocationing labor instead of merely adding labor. We found that many moderators we interviewed already practiced different elements of restorative justice. Some aim at supporting victims and give offenders a second chance but do not have the proper tools or procedures to achieve that. Other moderators have practices that embed elements of restorative justice, such as talking with offenders and victims. Rather than necessarily requiring new procedures, restorative justice requires a shift of purpose in existing processes – from determining the point of offense to caring for victims and offenders.
Importantly, if online restorative justice could stop the perpetuation of harm more frequently than punitive justice, it could reduce the need for moderation labor in the long term. While research has shown that offline restorative justice has successfully reduced re-offense rates [27, 75, 84], evaluating the effectiveness of restorative justice practices in online communities is an important area for future work.

7.1.2 Individuals’ Understanding of Justice Aligns with the Existing Punitive Justice Model.

Although people have needs that the current system does not address, we found that their mindsets and understanding of potential options often align with what the current system can already achieve through its punitive approach. As our research shows, many moderators and victims think that punishing offenders is the most or best they can do, and some offenders also expect to receive punishment. Community members also further perpetuate the punishment paradigm. This mindset is not only a result of the gaming community’s culture; it is pervasive throughout society, including in prisons, workplaces, and schools [37, 68].
Given the lack of sufficient information about and experiences with restorative justice, people may misunderstand and mistrust this new justice model. Some offenders in our interviews still expected to receive bans after the process, but restorative justice usually serves as an alternative to punishment, not an addition to it. Some participants wanted to implement alternative justice models in their own communities but received resistance from users who argued that the current moderation system works for them while disregarding its limitations for others. We found that such perspectives usually lead to quick rejections of the notion of implementing restorative justice online.
Before people can imagine and implement alternative justice frameworks to address harm, they must be aware of them. Crucial steps in this direction are information and education. Helping people understand the diversity of restorative justice processes and how their aim is restoration instead of punishment may address their doubts and open more opportunities for change. This is especially important since an incomplete understanding of restorative justice may cause harm in its implementation. For example, enforcing “successful outcomes” may disempower victims and result in disingenuous responses from offenders. Adapting restorative justice to online communities may require changes in the format and procedure of how harm is handled, but prioritizing its core values should help avoid additional unintentional harm.
Restorative justice has developed and rapidly evolved in worldwide practice [26, 132]. Future research can build on and expand restorative justice beyond the three principles and the victim-offender conference. In addition, people need to experience it to understand it, adapt it to their needs, and learn about its effectiveness. By experiencing it offline, several participants in our sample came to see it as a natural tool adaptable to online communities.
In future work, we plan to collaborate with online communities to implement and test restorative justice practices. We want to pilot online restorative justice sessions and run survey studies to understand the types of harm and types of processes most likely to benefit from these practices. Building on that research, we aim at providing more precise empirical guidelines about how restorative justice can be embedded in moderation systems based on the socio-technical affordances of various online communities. To manifest a future of restorative justice, “We will figure it out by working to get there.” [64].
While this research focuses on the restorative justice approach, it is not the only alternative and has its limitations. As some of our participants mentioned, many harms are rooted in structural issues such as sexism and racism. Transformative justice [25] and community accountability [82] are frameworks of justice developed to address such issues. Procedural justice emphasizes the importance of a fair moderation decision-making process [52] and is a key objective in restorative justice practices [4]. Future work should explore the potential of these different justice frameworks to address online harm.

7.1.3 Offender Accountability.

Another challenge may be in motivating offenders to take accountability for their wrongdoing – a persistent moderation problem regardless of the justice model implemented. In our interviews, we found that given the finite scope of moderation in many contexts and the limits in technical affordances of online communities, offenders can often easily avoid punishment. The harm may happen in a place without moderation or clear rules of moderation, e.g., when harm occurs during a private Discord chat or across multiple platforms. Some participants also noted that having multiple identities/accounts in Overwatch or Discord is easy. Thus, when punishment is ineffective, punitive justice may also lose its effectiveness.
Punishment is not the only form of holding offenders accountable. Restorative justice believes that people are also connected through relationships. Our interview data, as well as restorative justice literature, suggest the importance of a sense of community. If offenders perceive themselves as members of a shared community with victims, they will be more likely to take accountability for addressing harm [65, 90]. However, in our interviews, we find that there exists a lack of sense of community. Offenders may not expect to meet victims again or hide behind multiple anonymous accounts. Moderators typically moderate a confined space of general chat, which can leave harm unaddressed in any place outside.
Therefore, it is vital that we inquire into what accountability means to the community and how to hold people accountable within the current moderation system. If offenders can simply avoid any punitive consequences of conducting harm and do not feel a sense of belonging to the community where harm occurs, it would be challenging to engage them in any process –punitive or restorative– that holds them accountable.

7.1.4 Emotion Sharing and Communication in Restorative Justice.

The limited modes of interaction and the often anonymous member participation in online platforms may influence the effectiveness of restorative justice. Many online interactions are restricted to text or voice, prohibiting victims and offenders from sharing emotions, and may give rise to disingenuous participation. Emotional engagement by victims and offenders is essential for empathy and remorse [114]. Face-to-face sharing lets victims and offenders see each other’s body language and facial expressions.
Implementing offline restorative justice includes a series of physical structures and meeting procedures to elicit genuine, equal sharing. For example, participants sit in a circle; individuals who speak hold a talking stick; there are rituals at the beginning of the conference to build connections among participants and mark the circle as a unique space for change [90]. Those rituals for emotion sharing are hard to replicate in the online space. For example, if an offender messages an apology through text, it can be harder to discern a genuine apology from a disingenuous one.
The issue of computer-mediated communication and emotion-sharing has been long discussed in the HCI and CSCW literature. In recent years, increasingly more advanced technologies have been developed to facilitate civic engagement and communication in online systems. For example, Kriplean et al. built a platform, ConsiderIt, to support reflective interpersonal deliberation [71]. REASON (Rapid Evidence Aggregation Supporting Optimal Negotiation) is a Java applet developed to improve information pooling and arrive at consensus decisions [55]. Many researchers have attempted to model human-like characteristics and emotional awareness in chatbots [1, 112].
In the context of the restorative justice approach, Hughes has developed a tool, Keeper, for implementing online restorative justice circles [53]. Such existing systems can be leveraged to develop advanced tools that facilitate emotion-sharing and communication in online restorative justice processes. Online platforms such as Overwatch and Discord could add such tools to improve emotional sharing and communication, necessary conditions for implementing restorative justice.

7.2 Applying Restorative Justice in Online Moderation

We now discuss possible ways to adapt current moderation practices to implement restorative justice online. While we have used victim-offender conferencing as a vehicle to interrogate the opportunities and challenges of implementation, restorative justice includes a set of values and practices that extend beyond the conference. It is important to meet online communities and platforms where they are and design restorative justice mechanisms based on their resources, culture, and socio-technical affordances. For example, compared to Overwatch, Discord provides more flexibility in communication and ways of maintaining social connection after a game. Some Discord servers have a greater sense of community and moderation resources than others. We suggest that online communities or platforms can begin with partial restorative justice practices that involve only a few stakeholders (e.g., pre-conferencing) or adapt some moderation practices to embed restorative justice values (e.g., moderation explanations) and implement victim-offender conferencing when its preconditions are met.

7.2.1 Embed Restorative Justice Language in Moderation Explanations.

Moderators can embed restorative justice language in explanations of moderation decisions. Prior work has shown that providing such explanations can reduce re-offense in the same community. However, many moderated offenders dismiss such explanations and continue re-offending [56, 57]. Our research shows that explanations often focus on describing community guidelines or presenting community norms to the users, not encouraging offenders to reflect on the impact of their actions. These explanations usually indicate the actual or potential punitive consequence, which may direct offenders’ attention to the moderation system instead of their victims.
We suggest a shift in language from a rule-based explanation to an explanation that highlights the possible impact offenders may cause to victims and supports the offender in taking accountability. Facilitator P5 provided an example of how she would communicate with the offenders if she were an online facilitator: “This post had this emotional impact [...] this is how you’ve caused harm. This is the feedback from the community, and we want to work with you to create change.”
Victims are often left out of the moderation process in the online communities we studied. However, some victims want information on the moderation process being used and results of moderator interventions. While prior research has discussed how moderation transparency prevents offenders from re-offending [57], our work highlights the importance of providing information for victims in the moderation process since they are the ones with needs for support and healing. We suggest that a note of care may help victims feel heard and validated and help them recover.

7.2.2 Restorative Justice Conferences with Victims or Offenders.

In our interviews, we found that some victims or offenders may not be available or willing to meet and have a conversation to address the harm. When a collective conference is not possible, it is possible to apply a partial restorative justice process that includes offenders or victims alone [73, 80]. Some Discord moderators already talk with victims and offenders before making moderation decisions, a practice similar to restorative justice pre-conferencing. In offline pre-conferencing, facilitators ask questions to help victims and offenders reflect on the harm, providing them with opportunities for healing and learning.
We propose that pre-conferencing provides opportunities to meet some of the needs our participants identified. For example, when an offender wants to apologize to the victim, the victim may not want to meet the offender but communicate their feelings through the facilitator.
In offline restorative justice, pre-conferencing is a preparation step for a potential victim-offender conference. Thus, if both victim and offender are willing to meet with each other in the pre-conference, the moderators can organize a victim-offender conference, where both parties share their perspectives and collectively decide how to address the harm. Moderators have the responsibility to maintain a safe space for sharing, for example, to halt the process when harm seems likely to happen, ensure a power balance between the victim and the offender, and work with participants to establish agreements of behavior and values to adhere to throughout the process [15]. In cases where restorative justice does not succeed, preventing the continuation of harm must be prioritized. Because facilitating these processes is difficult, we discuss the challenges and opportunities for moderators to facilitate in the following section.
A victim-support group is another form of restorative justice conferencing [29]. In our sample, many victims indicated a need to receive emotional support. We propose that online communities offer a space for victims to share the harm they experienced and receive support. Systems with similar goals have previously been built in online spaces. For example, Hollaback is a platform that allows victims of street harassment to empower each other through storytelling [30]. Heartmob enables victims to describe their experiences of harm and solicit advice and emotional support from volunteer Heartmobers [13]. These platforms offer templates for how victim support systems can be built. However, support in these platforms is distant from where harm occurs, and it is also important to think about how online communities can support victims by motivating community members to provide support.

7.3 Situating Restorative Justice in the Moderation Landscape

We have illustrated possible ways to implement restorative justice in the current moderation system. Yet, we do not seek to advocate for restorative justice as a wholesale replacement of the current moderation approach. We propose that restorative justice goals and practices be embedded as part of an overall governance structure in online spaces.
Restorative justice conferencing should be used in select cases only because it is effective only when it is voluntary and the parties involved are committed to engaging. Our findings show that individuals have different conceptualizations of justice. Schoenebeck et al. also found that people’s preferences for punitive or restorative outcomes vary with their identity or social media behaviors [106]. It is thus important to attend to victims’ and offenders’ preferences in individual harm cases.
A larger governance structure should also take into account what to do if restorative justice processes fail. For instance, if it is determined at the pre-conference stage that a restorative justice approach cannot be applied, actions such as removing access by muting or banning might be used to prevent the continuation of harm. This is consistent with offline restorative justice practices, where the community or court clearly defines the types of cases that go through a restorative justice process and the action to take if a restorative justice process is not possible [41, 123].
We estimate that whether an online community decides to apply restorative justice is a value-related question. While restorative justice and punitive justice processes share a goal of addressing harm and preventing its perpetuation, they have different processes and orientations toward when justice is achieved. Content moderation—closer to a punitive justice approach—addresses harm by limiting the types of content that remain on the site and punishing offenders in proportion to their offense. In contrast, restorative justice aims at assuring that victims’ needs are met, and offenders learn, repair harm, and stop offending in the future. Thus, the primary reason for applying restorative justice in moderation is not to achieve the current goals of effectively removing inappropriate content and punishing offenders but to benefit online communities using restorative justice values and goals.
Seering et al. found that community moderators have diverse values regarding what moderation should achieve [108]. While some have a more punitive mindset and hope to be a “governor,” others align more with restorative justice values and hope to be “facilitator” or “gardener.” Thus, online communities must reflect on their values and goals and decide on what mechanisms (e.g., punitive or restorative) help realize those values. Recent research has argued that social media platforms are responsible for incorporating ethical values in moderation instead of merely optimizing to achieve commercial goals [50]. In particular, some researchers have proposed values and goals that align with restorative justice, such as centering victims’ needs in addressing harm [13, 106], democracy [107], and education [128]. Our work adds to this line of research and envisions how restorative justice may benefit online communities in addressing some severe forms of online harm, such as harassment.
Finally, communities should be cautious about expecting or enforcing a positive outcome. Enforcing forgiveness from victims or expecting a change in offenders’ behavior may undermine victims’ needs and put them in a vulnerable place for forgiveness or induce a disingenuous response from offenders [5]. Online communities should allow for partial success or no success without enforcing the ideal outcome, especially at the early stage of implementation when there are insufficient resources or commitments. Instead, they may focus on how victims, offenders, and the entire community could holistically benefit from the process.

8 Conclusion

In this research, we interviewed victims, offenders, and moderators in the Overwatch gaming community. We presented case studies that identified opportunities and challenges for using restorative justice in addressing online harm and discussed possible ways to embed a restorative justice approach in the current content moderation landscape. Much remains to be done to explore what restorative justice may look like in an online community context and when and how to implement this approach. We hope this work offers a valuable guide for designers, volunteers, activists, and other scholars to experiment with restorative justice and related approaches in online communities.

Footnotes

4
We use the terms “victim” and “offender” for brevity and to clarify participants’ roles in specific harm cases. We recognize and agree with calls for eradicating the use of these labels over the long term. Some restorative justice practitioners believe these labels and their meanings are rooted in the punitive justice system, and transformation to restorative justice requires transforming our language. Using “offenders” may imply that people are “inherently bad” and deserve the condemnation of society [17], while using “victims” may feel disempowering to some and deny the agency victims should have in restorative justice [40]. Less popular alternatives to these terms include “the person who caused harm” or “perpetrator” and the “person who has been harmed” or “survivor.” Accordingly, the language of “victim-offender conference” has also shifted to “restorative justice circle.” However, such alternative terms may also not align with the self-image of harmed participants, as we found during our data collection. For this early-stage research, we retain the original terminology to be consistent with the language participants used.
5
Discord Moderator Academy. https://discord.com/moderation.

References

[1]
Achini Adikari, Daswin De Silva, Damminda Alahakoon, and Xinghuo Yu. 2019. A cognitive model for emotion awareness in industrial Chatbots. In Proceedings of the 2019 IEEE 17th International Conference on Industrial Informatics. IEEE, 183–186.
[2]
Sonam Adinolf and Selen Turkay. 2018. Toxic behaviors in esports games: Player perceptions and coping strategies. In Proceedings of the 2018 Annual Symposium on Computer-human Interaction in Play Companion Extended Abstracts. 365–372.
[3]
Hunt Allcott and Matthew Gentzkow. 2017. Social media and fake news in the 2016 election. Journal of Economic Perspectives 31, 2(2017), 211–236. DOI:
[4]
Geoffrey C. Barnes, Jordan M. Hyatt, Caroline M. Angel, Heather Strang, and Lawrence W. Sherman. 2015. Are restorative justice conferences more fair than criminal courts? Comparing levels of observed procedural justice in the reintegrative shaming experiments (RISE). Criminal Justice Policy Review 26, 2 (2015), 103–130.
[5]
Gordon Bazemore and Mara Schiff. 2015. Restorative Community Justice: Repairing Harm and Transforming Communities. Routledge.
[6]
S. Gordon Bazemore. 2001. A Comparison of Four Restorative Conferencing Models. US Department of Justice, Office of Justice Programs, Office of Juvenile....
[7]
S. Gordon Bazemore and Lode Walgrave. 1999. Restorative Juvenile Justice: Repairing the Harm of Youth Crime. Criminal Justice Press Monsey, NY.
[8]
Derek Bell. 2008. And We Are Not Saved: The Elusive Quest for Racial Justice. Basic Books.
[9]
Nicole A. Beres, Julian Frommel, Elizabeth Reid, Regan L. Mandryk, and Madison Klarkowski. 2021. Don’t you know that you’re toxic: Normalization of toxicity in online gaming. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1–15.
[10]
Patrick Biernacki and Dan Waldorf. 1981. Snowball sampling: problems and techniques of chain referral sampling. Sociological Methods & Research 10, 2(1981), 141–163. DOI:Publisher: SAGE PublicationsSage CA: Los Angeles, CA.
[11]
Reuben Binns, Michael Veale, Max Van Kleek, and Nigel Shadbolt. 2017. Like trainer, like bot? Inheritance of bias in algorithmic content moderation. In Proceedings of the International Conference on Social Informatics. Springer, 405–415.
[12]
Lindsay Blackwell, Tianying Chen, Sarita Schoenebeck, and Cliff Lampe. 2018. When online harassment is perceived as justified. In Proceedings of the 12th International AAAI Conference on Web and Social Media.
[13]
Lindsay Blackwell, Jill P. Dimond, Sarita Schoenebeck, and Cliff Lampe. 2017. Classification and its consequences for online harassment: Design insights from heartMob. PACMHCI 1, CSCW (2017), 24–1.
[14]
Blizzard. 2020. Blizzard’s In-Game Code of Conduct. Retrieved 25 June 2023 from https://us.battle.net/support/en/article/42673.
[15]
Jane Bolitho and Jasmine Bruce. 2017. Science, art and alchemy: Best practice in facilitating restorative justice. Contemporary Justice Review 20, 3 (2017), 336–362.
[16]
Shannon A. Bowen. 2013. Using classic social media cases to distill ethical guidelines for digital engagement. Journal of Mass Media Ethics 28, 2 (2013), 119–133.
[17]
Lynn S. Branham. 2019. Eradicating the Label “Offender” from the Lexicon of Restorative Practices and Criminal Justice. Retrieved 25 June 2023 from http://wakeforestlawreview.com/2019/08/eradicating-the-label-offender-from-the-lexicon-of-restorative-practices-and-criminal-justice/.
[18]
Caitlin Ring Carlson and Luc S. Cousineau. 2020. Are you sure you want to view this community? Exploring the ethics of reddit’s quarantine practice. Journal of Media Ethics 35, 4 (2020), 202–213.
[19]
Eshwar Chandrasekharan, Shagun Jhaver, Amy Bruckman, and Eric Gilbert. 2022. Quarantined! Examining the effects of a community-wide moderation intervention on reddit. ACM Transactions on Computer-Human Interaction 29, 4(2022), 26 pages. DOI:
[20]
Eshwar Chandrasekharan, Mattia Samory, Shagun Jhaver, Hunter Charvat, Amy Bruckman, Cliff Lampe, Jacob Eisenstein, and Eric Gilbert. 2018. The internet’s hidden rules: An empirical study of reddit norm violations at micro, meso, and macro scales. Proceedings of the ACM on Human-Computer Interaction 2, CSCW (2018), 32.
[21]
Kathy Charmaz. 2006. Constructing Grounded Theory: A Practical Guide Through Qualitative Analysis. Sage.
[22]
Justin Cheng, Cristian Danescu-Niculescu-Mizil, and Jure Leskovec. 2015. Antisocial behavior in online discussion communities. In Proceedings of the 9th International AAAI Conference on Web and Social Media.
[23]
Danielle Keats Citron. 2015. Addressing cyber harassment: An overview of hate crimes in cyberspace. Social Science Research Network, Rochester, NY. Retrieved from https://papers.ssrn.com/abstract=2932358.
[24]
Danielle Keats Citron and Mary Anne Franks. 2014. Criminalizing Revenge Porn. Wake Forest Law Review, 49 (2014), 345–383. https://papers.ssrn.com/sol3/papers.cfm?abstract_id=2368946#.
[25]
Erin Daly. 2001. Transformative justice: Charting a path to reconciliation. International Legal Perspectives 12, 2001-2002 (2001), 73–184. https://heinonline.org/HOL/P?h=hein.journals/intlegp12&i=77.
[26]
Kathleen Daly and Hennessey Hayes. 2001. Restorative Justice and Conferencing in Australia. Trends and Issues in Crime and Criminal Justice 186, (2001).
[27]
Kathleen Daly. 2002. Restorative justice: The real story. Punishment & Society 4, 1(2002), 55–79. DOI:
[28]
Lina Dencik, Arne Hintz, and Jonathan Cable. 2019. Towards data justice. In Bigo, Didier, Engin Isin, and Evelyn Ruppert. Data politics: Worlds, subjects, rights. Taylor & Francis, 2019.
[29]
James Dignan. 2004. Understanding Victims and Restorative Justice. McGraw-Hill Education (UK).
[30]
Jill P. Dimond, Michaelanne Dye, Daphne LaRose, and Amy S. Bruckman. 2013. Hollaback! The role of storytelling online in a social movement organization. In Proceedings of the 2013 Conference on Computer Supported Cooperative Work. 477–490.
[31]
Judith S. Donath. 2002. Identity and deception in the virtual community. In Proceedings of the Communities in Cyberspace. Routledge, 37–68.
[32]
Bryan Dosono and Bryan Semaan. 2019. Moderation practices as emotional labor in sustaining online communities: The case of AAPI identity work on Reddit. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. 1–13.
[33]
Maeve Duggan. 2014. Online Harassment. Pew Research Center.
[34]
Maeve Duggan. 2017. Online Harassment 2017. Retrieved 25 June 2023 from https://www.pewresearch.org/internet/2017/07/11/online-harassment-2017/.
[35]
Maeve Duggan and Aaron Smith. 2013. 6% of online adults are reddit users. (July 2013). Retrieved June 26, 2023 from https://www.pewresearch.org/internet/2013/07/03/6-of-online-adults-are-reddit-users/.
[36]
Casey Fiesler and Brianna Dym. 2020. Moving across lands: Online platform migration in fandom communities. Proc. ACM Hum.-Comput. Interact. 4, CSCW1(2020), 25 pages. DOI:
[37]
Michel Foucault. 2012. Discipline and Punish: The Birth of the Prison. Vintage.
[38]
Jesse Fox and Wai Yen Tang. 2017. Women’s experiences with general and sexual harassment in online video games: Rumination, organizational responsiveness, withdrawal, and coping strategies. New Media & Society 19, 8 (2017), 1290–1307.
[39]
David Garland. 2012. Punishment and Modern Society: A Study in Social Theory. University of Chicago Press.
[40]
Theo Gavrielides. 2017. Collapsing the labels ‘victim’ and ‘offender’ in the Victims’ Directive and the paradox of restorative justice. Restorative Justice 5, 3 (2017), 368–381.
[41]
Theo Gavrielides. 2017. Restorative Justice: Ideals and Realities. Routledge.
[42]
Tarleton Gillespie. 2018. Custodians of the Internet: Platforms, Content Moderation, and the Hidden Decisions that Shape Social Media. Yale University Press.
[43]
Robert Gorwa. 2019. What is platform governance? Information, Communication & Society 22, 6 (2019), 854–871. DOI:
[44]
Nitesh Goyal, Leslie Park, and Lucy Vasserman. 2022. “You have to prove the threat is real”: Understanding the needs of Female Journalists and Activists to Document and Report Online Harassment. In Proceedings of the CHI Conference on Human Factors in Computing Systems. 1–17.
[45]
Kishonna L. Gray. 2012. Intersecting oppressions and online communities: Examining the experiences of women of color in Xbox Live. Information, Communication & Society 15, 3 (2012), 411–428.
[46]
Why Is This Happening?2019. Thinking about how to abolish prisons with Mariame Kaba: Podcast; transcript. Retrieved 25 June 2023 from https://www.nbcnews.com/think/opinion/thinking-about-how-abolish-prisons-mariame-kaba-podcast-transcript-ncna992721.
[47]
A. Hasinoff, A. Gibson, and N. Salehi. 2020. The promise of restorative justice in addressing online harm, Brookings Institution. United States of America. Retrieved 25 Jun 2023 from https://policycommons.net/artifacts/4138898/the-promise-of-restorative-justice-in-addressing-online-harm/4947425/.
[48]
Amy A. Hasinoff and Nathan Schneider. 2022. From scalability to subsidiarity in addressing online harm. Social Media+ Society 8, 3 (2022), 20563051221126041.
[49]
Caroline Haythornthwaite. 2002. Strong, weak, and latent ties and the impact of new media. The Information Society 18, 5 (2002), 385–401.
[50]
Natali Helberger, Jo Pierson, and Thomas Poell. 2018. Governing online platforms: From contested to cooperative responsibility. The Information Society 34, 1 (2018), 1–14.
[51]
Zorah Hilvert-Bruce and James T. Neill. 2020. I’m just trolling: The role of normative beliefs in aggressive behaviour in online gaming. Computers in Human Behavior 102, (2020), 303–311. DOI:
[52]
Youyang Hou, Cliff Lampe, Maximilian Bulinski, and James J. Prescott. 2017. Factors in fairness and emotion in online case resolution systems. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems. 2511–2522.
[53]
Maggie Hughes and Deb Roy. 2020. Keeper: An online synchronous conversation environment informed by in-person facilitation practices. In Proceedings of the Conference Companion Publication of the 2020 on Computer Supported Cooperative Work and Social Computing. 275–279.
[54]
Jon Hurwitz and Mark Peffley. 2005. Explaining the great racial divide: Perceptions of fairness in the US criminal justice system. The Journal of Politics 67, 3 (2005), 762–783.
[55]
Joshua E. Introne. 2008. Adaptive Mediation in Groupware. Brandeis University.
[56]
Shagun Jhaver, Darren Scott Appling, Eric Gilbert, and Amy Bruckman. 2019. “Did you suspect the post would be removed?”: Understanding user reactions to content removals on reddit. Proceedings of the ACM on Human-Computer Interaction 3, CSCW(2019), 33 pages. DOI:
[57]
Shagun Jhaver, Amy Bruckman, and Eric Gilbert. 2019. Does transparency in moderation really matter? User behavior after content removal explanations on reddit. Proceedings of the ACM on Human-Computer Interaction 3, CSCW(2019), 27 pages. DOI:
[58]
Shagun Jhaver, Larry Chan, and Amy Bruckman. 2018. The view from the other side: The border between controversial speech and harassment on kotaku in action. (February 2018). Retrieved June 25, 2023 from https://firstmonday.org/ojs/index.php/fm/article/view/8232.
[59]
Shagun Jhaver, Quan Ze Chen, Detlef Knauss, and Amy X. Zhang. 2022. Designing word filter tools for creator-led comment moderation. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems.Association for Computing Machinery, New York, NY, 21 pages. DOI:
[60]
Shagun Jhaver, Sucheta Ghoshal, Amy Bruckman, and Eric Gilbert. 2018. Online harassment and content moderation: The case of blocklists. ACM Transactions on Computer-Human Interaction 25, 2(2018), 33 pages. DOI:
[61]
Shagun Jhaver, Alice Qian Zhang, Quan Ze Chan, Nikhila Natarajan, Ruotong Wang, and Amy X. Zhang. 2023. Personalizing content moderation on social media: User perspectives on moderation choices, interface design, and labor. Proceedings of the ACM on Human-Computer InteractionCSCW (2023), 33 pages.
[62]
Jialun Aaron Jiang, Charles Kiene, Skyler Middler, Jed R. Brubaker, and Casey Fiesler. 2019. Moderation challenges in voice-based online communities on discord. Proceedings of the ACM on Human-Computer Interaction 3, CSCW (2019), 1–23. Publisher: ACM New York, NY.
[63]
Gerry Johnstone and Daniel Van Ness. 2013. Handbook of Restorative Justice. Routledge. Google-Books-ID: U2UQBAAAQBAJ.
[64]
Mariame Kaba. 2021. We Do This Til We Free Us: Abolitionist Organizing and Transforming Justice. Haymarket Books.
[65]
David R. Karp and Beau Breslin. 2001. Restorative justice in school communities. Youth & Society 33, 2(2001), 249–272. DOI:
[66]
Charles Kiene and Benjamin Mako Hill. 2020. Who uses bots? A statistical analysis of bot usage in moderation teams. In Proceedings of the Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems. 1–8.
[67]
Mimi E. Kim. 2018. From carceral feminism to transformative justice: Women-of-color feminism and alternatives to incarceration. Journal of Ethnic & Cultural Diversity in Social Work 27, 3 (2018), 219–233. DOI:. arXiv:https://doi.org/10.1080/15313204.2018.1474827
[68]
Anna King and Shadd Maruna. 2009. Is a conservative just a liberal who has been mugged? Exploring the origins of punitive views. Punishment & Society 11, 2 (2009), 147–169.
[69]
Yubo Kou. 2021. Punishment and its discontents: An analysis of permanent ban in an online game community. Proceedings of the ACM on Human-Computer Interaction 5, CSCW2 (2021), 1–21.
[70]
Robert E. Kraut and Paul Resnick. 2012. Building Successful Online Communities: Evidence-based Social Design. Mit Press.
[71]
Travis Kriplean, Jonathan Morgan, Deen Freelon, Alan Borning, and Lance Bennett. [n. d.]. Supporting reflective public thought with considerit. In Proceedings of the ACM 2012 conference on Computer Supported Cooperative Work.ACM, 265–274. DOI:http://dl.acm.org/citation.cfm?doid=2145204.2145249
[72]
Etienne G. Krug, James A. Mercy, Linda L. Dahlberg, and Anthony B. Zwi. 2002. The world report on violence and health. The lancet 360, 9339 (2002), 1083–1088.
[73]
Shih-Ya Kuo, Dennis Longmire, and Steven J. Cuvelier. 2010. An empirical assessment of the process of restorative justice. Journal of Criminal Justice 38, 3 (2010), 318–328.
[74]
Noam Lapidot-Lefler and Azy Barak. 2012. Effects of anonymity, invisibility, and lack of eye-contact on toxic online disinhibition. Computers in Human Behavior 28, 2 (2012), 434–443.
[75]
Jeff Latimer, Craig Dowden, and Danielle Muise. 2005. The effectiveness of restorative justice practices: A meta-analysis. The Prison Journal 85, 2(2005), 127–144. DOI:
[76]
Anti-Defamation League. 2019. Free to Play: Hate, Harassment and Positive Social Experiences in Online Games. (July 2019). Retrieved June 25, 2023 from https://www.adl.org/resources/report/free-play-hate-harassment-and-positive-social-experiences-online-games.
[77]
Amanda Lenhart, Michele Ybarra, Kathryn Zickuhr, and Myeshia Price-Feeney. 2016. Online Harassment, Digital Abuse, and Cyberstalking in America. Data and Society Research Institute.
[78]
Kaitlin Mahar, Amy X. Zhang, and David Karger. 2018. Squadbox: A tool to combat email harassment using friendsourced moderation. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems.ACM, New York, NY,586:1–586:13. DOI:. event-place: Montreal QC, Canada.
[79]
Alice E. Marwick. 2021. Morally motivated networked harassment as normative reinforcement. Social Media+ Society 7, 2 (2021), 20563051211021378.
[80]
Paul McCold. 2000. Toward a holistic vision of restorative juvenile justice: A reply to the maximalist model. Contemporary Justice Review 3, 4 (2000), 357–414.
[81]
Jon’a F. Meyer. 1998. History repeats itself: Restorative justice in native american communities. Journal of Contemporary Criminal Justice 14, 1 (1998), 42–57.
[82]
Sassy Molyneux, Martin Atela, Vibian Angwenyi, and Catherine Goodman. 2012. Community accountability at peripheral health facilities: A review of the empirical literature and development of a conceptual framework. Health Policy and Planning 27, 7 (2012), 541–554.
[83]
C. K. Monge and TC O’Brien. 2022. Effects of individual toxic behavior on team performance in league of legends. Media Psychology 25, 1 (2022), 82–105.
[84]
Allison Morris, Gabrielle M. Maxwell, and Jeremy P. Robertson. 1993. Giving victims a voice: A new zealand experiment. The Howard Journal of Criminal Justice 32, 4 (1993), 304–321.
[85]
Garrett Morrow, Briony Swire-Thompson, Jessica Montgomery Polny, Matthew Kopec, and John P. Wihbey. 2022. The emerging science of content labeling: Contextualizing social media content moderation. Journal of the Association for Information Science and Technology 73, 10 (2022), 1365–1386.
[86]
Tyler Musgrave, Alia Cummings, and Sarita Schoenebeck. 2022. Experiences of harm, healing, and joy among black women and femmes on social media. In Proceedings of the CHI Conference on Human Factors in Computing Systems. 1–17.
[87]
Sarah Myers West. 2018. Censored, suspended, shadowbanned: User interpretations of content moderation on social media platforms. New Media & Society 20, 11 (2018), 4366–4383.
[88]
Angela Nagle. 2017. Kill All Normies: Online Culture Wars From 4chan and Tumblr to Trump and the Alt-right. John Hunt Publishing.
[89]
Whitney Phillips. 2015. This is Why we Can’t have Nice Things: Mapping the Relationship Between Online Trolling and Mainstream Culture. MIT Press.
[90]
Kay Pranis. 2015. Little book of circle processes: A new/old approach to peacemaking. Simon.
[91]
Elizabeth Reid. 1999. Hierarchy and power. Communities in Cyberspace (1999), 107–133.
[92]
Elizabeth Reid, Regan L. Mandryk, Nicole A. Beres, Madison Klarkowski, and Julian Frommel. 2022. Feeling good and in control: In-game tools to support targets of toxicity. Proceedings of the ACM on Human-Computer Interaction 6, CHI PLAY (2022), 1–27.
[93]
Elizabeth Reid, Regan L. Mandryk, Nicole A. Beres, Madison Klarkowski, and Julian Frommel. 2022. “Bad vibrations”: Sensing toxicity from in-game audio features. IEEE Transactions on Games 14, 4 (2022), 558–568.
[94]
Sarah T. Roberts. 2019. Behind the Acreen. Yale University Press.
[95]
Sarah T. Roberts. 2018. Digital detritus: ‘Error’ and the logic of opacity in social media content moderation. (March 2018). Retrieved June 25, 2023 from https://firstmonday.org/ojs/index.php/fm/article/view/8283.
[96]
Oliver C. Robinson. 2014. Sampling in interview-based qualitative research: A theoretical and practical guide. Qualitative Research in Psychology 11, 1 (2014), 25–41. Publisher: Taylor & Francis.
[97]
Aja Romano. 2017. How the alt-right uses internet trolling to confuse you into dismissing its ideology. Retrieved 25 June 2023 from https://www.vox.com/2016/11/23/13659634/alt-right-trolling.
[98]
Jon Ronson. 2015. How one stupid tweet blew up Justine Sacco’s life. (February 2015). Retrieved June 25, 2023 from https://www.nytimes.com/2015/02/15/magazine/how-one-stupid-tweet-ruined-justine-saccos-life.html.
[99]
Amanda Phillips and Bonnie Ruberg. 2018. Not Gay as in Happy: Queer Resistance and Video Games (Introduction). Game Studies 18, 2018 (2018), 1–2.
[100]
Maria Ruotsalainen and Usva Friman. 2018. “There are no women and they all play mercy”: Understanding and explaining (the lack of) women’s presence in esports and competitive gaming. In Proceedings of the Conference of Digital Games Research Association. Digital Games Research Association.
[101]
Omar Ruvalcaba, Jeffrey Shulze, Angela Kim, Sara R. Berzenski, and Mark P. Otten. 2018. Women’s experiences in eSports: Gendered differences in peer and spectator feedback during competitive video game play. Journal of Sport and Social Issues 42, 4 (2018), 295–311.
[102]
Johnny Saldaña. 2013. The Coding Manual for Qualitative Researchers+ Qualitative Data Analysis: A Methods Sourcebook. Sage Publications.
[103]
Niloufar Salehi. 2020. Do No Harm. Retrieved 25 June 2023 from https://logicmag.io/care/do-no-harm/.
[104]
Niloufar Salehi, Jaime Teevan, Shamsi Iqbal, and Ece Kamar. 2017. Communicating context to the crowd for complex writing tasks. In Proceedings of the 2017 ACM Conference on Computer Supported Cooperative Work and Social Computing. 1890–1901.
[105]
santaclaraprinciples.org.2022. Santa Clara principles on transparency and accountability in content moderation. Retrieved 25 June 2023 from https://santaclaraprinciples.org/.
[106]
Sarita Schoenebeck, Oliver L. Haimson, and Lisa Nakamura. 2021. Drawing from justice theories to support targets of online harassment. New Media & Society 23, 5 (2021), 1278–1300.
[107]
Joseph Seering. 2020. Reconsidering self-moderation: The role of research in supporting community-based models for online content moderation. Proceedings of the ACM on Human-Computer Interaction 4, CSCW2 (2020), 1–28.
[108]
Joseph Seering, Geoff Kaufman, and Stevie Chancellor. 2022. Metaphors in moderation. New Media & Society 24, 3 (2022), 621–640.
[109]
Joseph Seering, Robert Kraut, and Laura Dabbish. 2017. Shaping pro and anti-social behavior on twitch through moderation and example-setting. In Proceedings of the 2017 ACM Conference on Computer Supported Cooperative Work and Social Computing. 111–125.
[110]
Joseph Seering, Tony Wang, Jina Yoon, and Geoff Kaufman. 2019. Moderator engagement and community development in the age of algorithms. New Media & Society 21, 7(2019), 1417–1443. DOI:
[111]
Orit Shaer, Lauren Westendorf, Nicholas A. Knouf, and Claudia Pederson. 2017. Understanding gaming perceptions and experiences in a women’s college community. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems. 1544–1557.
[112]
Timo Spring, Jacky Casas, Karl Daher, Elena Mugellini, and Omar Abou Khaled. 2019. Empathic response generation in chatbots. In Proceedings of the SwissText.
[113]
Miriah Steiger, Timir J. Bharucha, Sukrit Venkatagiri, Martin J. Riedl, and Matthew Lease. 2021. The Psychological Well-Being of Content Moderators: The Emotional Labor of Commercial Moderation and Avenues for Improving Support. Association for Computing Machinery, New York, NY. DOI:
[114]
Heather Strang and Lawrence W. Sherman. 2003. Repairing the harm: Victims and restorative justice. Utah Law Review 2003, 1 (2003), 15–42. https://heinonline.org/HOL/P?h=hein.journals/utahlr2003&i=25.
[115]
Sharifa Sultana, Mitrasree Deb, Ananya Bhattacharjee, Shaid Hasan, S. M. Raihanul Alam, Trishna Chakraborty, Prianka Roy, Samira Fairuz Ahmed, Aparna Moitra, M. Ashraful Amin, A. K. M. Najmul Islam, and Syed Ishtiaque Ahmed. 2021. “Unmochon”: A tool to combat online sexual harassment over facebook messenger. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1–18.
[116]
Nicolas Suzor. 2018. Digital constitutionalism: Using the rule of law to evaluate the legitimacy of governance by platforms. Social Media + Society 4, 3 (2018), 2056305118787812.
[117]
Nicolas Suzor, Tess Van Geelen, and Sarah Myers West. 2018. Evaluating the legitimacy of platform governance: A review of research and a shared research agenda. International Communication Gazette 80, 4 (2018), 385–400.
[118]
Nicolas P. Suzor, Sarah Myers West, Andrew Quodling, and Jillian York. 2019. What do we mean when we talk about transparency? Toward meaningful transparency in commercial content moderation. International Journal of Communication 13 (2019), 18.
[119]
Zenon Szablowinski. 2008. Punitive justice and restorative justice as social reconciliation. The Heythrop Journal 49, 3 (2008), 405–422.
[120]
Kurt Thomas, Patrick Gage Kelley, Sunny Consolvo, Patrawat Samermit, and Elie Bursztein. 2022. “It’s common and a part of being a content creator”: Understanding how creators experience and cope with hate and harassment online. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems. 1–15.
[121]
Alexandra To, Wenxia Sweeney, Jessica Hammer, and Geoff Kaufman. 2020. “They just don’t get it”: Towards social technologies for coping with interpersonal racism. Proceedings of the ACM on Human-Computer Interaction 4, CSCW1 (2020), 1–29.
[122]
Daniel Van Ness and Karen Heetderks Strong. 2014. Restoring Justice: An Introduction to Restorative Justice. Routledge.
[123]
Daniel W. Van Ness. 2016. An overview of restorative justice around the world. (2016). https://biblioteca.cejamericas.org/bitstream/handle/2015/4767/dan_van_ness_final_paper.pdf?sequence=1&isAllowed=y.
[124]
Jonathan Vanian. 2017. Twitter Toughens Rules on Nudity and Revenge Porn | Fortune. Retrieved 25 June 2023 from http://fortune.com/2017/10/27/nudity-revenge-porn-twitter/.
[125]
Emily A. Vogels. 2021. The state of online harassment. Retrieved June 25, 2023 from https://www.pewresearch.org/internet/2021/01/13/the-state-of-online-harassment/.
[126]
Charlie Warzel. 2016. A honeypot for assholes: Inside twitters 10-year failure to stop harassment. (August 2016). Retrieved June 23, 2023 from https://www.buzzfeednews.com/article/charliewarzel/a-honeypot-for-assholes-inside-twitters-10-year-failure-to-s.
[127]
Michael Wenzel, Tyler G. Okimoto, Norman T. Feather, and Michael J. Platow. 2008. Retributive and restorative justice. Law and Human Behavior 32, 5 (2008), 375–389.
[128]
Sarah Myers West. 2018. Censored, suspended, shadowbanned: User interpretations of content moderation on social media platforms. New Media & Society 20, 11 (2018), 4366–4383.
[129]
Donghee Yvette Wohn. 2019. Volunteer moderators in twitch micro communities: How they get involved, the roles they play, and the emotional labor they experience. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. 1–13.
[130]
William R. Wood and Masahiro Suzuki. 2016. Four challenges in the future of restorative justice. Victims & Offenders 11, 1 (2016), 149–172.
[131]
Sijia Xiao, Coye Cheshire, and Niloufar Salehi. 2022. Sensemaking, support, safety, retribution, transformation: A restorative justice approach to understanding adolescents’ needs for addressing online harm. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems. 1–15.
[132]
Howard Zehr. 2015. The Little Book of Restorative Justice: Revised and Updated. Simon and Schuster.

Cited By

View all
  • (2024)A playbook for mapping adolescent interactions with misinformation to perceptions of online harmHarvard Kennedy School Misinformation Review10.37016/mr-2020-161Online publication date: 30-Oct-2024
  • (2024)Repairing the harm: Toward an algorithmic reparations approach to hate speech content moderationBig Data & Society10.1177/2053951724124533311:2Online publication date: 22-Apr-2024
  • (2024)Chillbot: Content Moderation in the BackchannelProceedings of the ACM on Human-Computer Interaction10.1145/36869418:CSCW2(1-26)Online publication date: 8-Nov-2024
  • Show More Cited By

Index Terms

  1. Addressing Interpersonal Harm in Online Gaming Communities: The Opportunities and Challenges for a Restorative Justice Approach

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Transactions on Computer-Human Interaction
    ACM Transactions on Computer-Human Interaction  Volume 30, Issue 6
    December 2023
    424 pages
    ISSN:1073-0516
    EISSN:1557-7325
    DOI:10.1145/3623488
    Issue’s Table of Contents
    This work is licensed under a Creative Commons Attribution International 4.0 License.

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 25 September 2023
    Online AM: 08 June 2023
    Accepted: 11 May 2023
    Revised: 02 May 2023
    Received: 02 November 2022
    Published in TOCHI Volume 30, Issue 6

    Check for updates

    Author Tags

    1. Content moderation
    2. online harassment
    3. alternative justice
    4. Discord
    5. Overwatch

    Qualifiers

    • Research-article

    Funding Sources

    • National Science Foundation

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)4,273
    • Downloads (Last 6 weeks)449
    Reflects downloads up to 10 Nov 2024

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)A playbook for mapping adolescent interactions with misinformation to perceptions of online harmHarvard Kennedy School Misinformation Review10.37016/mr-2020-161Online publication date: 30-Oct-2024
    • (2024)Repairing the harm: Toward an algorithmic reparations approach to hate speech content moderationBig Data & Society10.1177/2053951724124533311:2Online publication date: 22-Apr-2024
    • (2024)Chillbot: Content Moderation in the BackchannelProceedings of the ACM on Human-Computer Interaction10.1145/36869418:CSCW2(1-26)Online publication date: 8-Nov-2024
    • (2024)Designing Interactive and Immersive Socio-Technical Tools to Support Communication and Empowerment for Immigrant CommunitiesCompanion Publication of the 2024 Conference on Computer-Supported Cooperative Work and Social Computing10.1145/3678884.3682052(31-34)Online publication date: 11-Nov-2024
    • (2024)Reflecting on the Relational: Youth-Centered Approaches to Living with TechnologyCompanion Publication of the 2024 Conference on Computer-Supported Cooperative Work and Social Computing10.1145/3678884.3682047(27-30)Online publication date: 11-Nov-2024
    • (2024)Towards Ethical AI Moderation in Multiplayer GamesProceedings of the ACM on Human-Computer Interaction10.1145/36771098:CHI PLAY(1-30)Online publication date: 15-Oct-2024
    • (2024)The Toxic Cost of Cheap Usernames: Re-Applying Friedman and Resnick's Framework in Competitive Video Game NetworksGames: Research and Practice10.1145/36699362:2(1-17)Online publication date: 8-Aug-2024
    • (2024)Combating Islamophobia: Compromise, Community, and Harmony in Mitigating Harmful Online ContentACM Transactions on Social Computing10.1145/36415107:1-4(1-32)Online publication date: 27-Apr-2024
    • (2024)Fighting for Their Voice: Understanding Indian Muslim Women's Responses to Networked HarassmentProceedings of the ACM on Human-Computer Interaction10.1145/36410058:CSCW1(1-24)Online publication date: 26-Apr-2024
    • (2024)Support in Short Form: Investigating TikTok Comments on Videos with #HarassmentExtended Abstracts of the CHI Conference on Human Factors in Computing Systems10.1145/3613905.3650849(1-8)Online publication date: 11-May-2024
    • Show More Cited By

    View Options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Get Access

    Login options

    Full Access

    Media

    Figures

    Other

    Tables

    Share

    Share

    Share this Publication link

    Share on social media